qgallouedec HF Staff commited on
Commit
052e709
·
verified ·
1 Parent(s): 1b30f27

Upload LlavaNextForConditionalGeneration

Browse files
Files changed (3) hide show
  1. config.json +1 -3
  2. generation_config.json +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -55,11 +55,10 @@
55
  "vocab_size": 32064
56
  },
57
  "tie_word_embeddings": false,
58
- "transformers_version": "4.57.3",
59
  "use_image_newline_parameter": true,
60
  "vision_config": {
61
  "attention_dropout": 0.0,
62
- "embed_dim": 64,
63
  "hidden_act": "quick_gelu",
64
  "hidden_size": 16,
65
  "image_size": 336,
@@ -71,7 +70,6 @@
71
  "num_attention_heads": 4,
72
  "num_channels": 3,
73
  "num_hidden_layers": 2,
74
- "num_key_value_heads": 2,
75
  "patch_size": 14,
76
  "projection_dim": 768,
77
  "vocab_size": 32000
 
55
  "vocab_size": 32064
56
  },
57
  "tie_word_embeddings": false,
58
+ "transformers_version": "4.56.2",
59
  "use_image_newline_parameter": true,
60
  "vision_config": {
61
  "attention_dropout": 0.0,
 
62
  "hidden_act": "quick_gelu",
63
  "hidden_size": 16,
64
  "image_size": 336,
 
70
  "num_attention_heads": 4,
71
  "num_channels": 3,
72
  "num_hidden_layers": 2,
 
73
  "patch_size": 14,
74
  "projection_dim": 768,
75
  "vocab_size": 32000
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.57.3"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.56.2"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e8a6c86af0bf399304a803b129d2235c9ed6f3cc95231374b6b8182294b4842
3
  size 5399728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:825b5459899334f14044e7723b95848796988ac45751408dc9d79b200da27677
3
  size 5399728