Upload LlavaForConditionalGeneration

#3
by qgallouedec HF Staff - opened
Files changed (4) hide show
  1. config.json +2 -4
  2. generation_config.json +1 -1
  3. model.safetensors +2 -2
  4. tokenizer.json +8 -1
config.json CHANGED
@@ -38,10 +38,9 @@
38
  "vocab_size": 32064
39
  },
40
  "tie_word_embeddings": false,
41
- "transformers_version": "4.57.3",
42
  "vision_config": {
43
  "attention_dropout": 0.0,
44
- "embed_dim": 64,
45
  "hidden_act": "quick_gelu",
46
  "hidden_size": 16,
47
  "image_size": 336,
@@ -52,8 +51,7 @@
52
  "model_type": "clip_vision_model",
53
  "num_attention_heads": 4,
54
  "num_channels": 3,
55
- "num_hidden_layers": 2,
56
- "num_key_value_heads": 2,
57
  "patch_size": 14,
58
  "projection_dim": 768,
59
  "vocab_size": 32000
 
38
  "vocab_size": 32064
39
  },
40
  "tie_word_embeddings": false,
41
+ "transformers_version": "4.56.2",
42
  "vision_config": {
43
  "attention_dropout": 0.0,
 
44
  "hidden_act": "quick_gelu",
45
  "hidden_size": 16,
46
  "image_size": 336,
 
51
  "model_type": "clip_vision_model",
52
  "num_attention_heads": 4,
53
  "num_channels": 3,
54
+ "num_hidden_layers": 3,
 
55
  "patch_size": 14,
56
  "projection_dim": 768,
57
  "vocab_size": 32000
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 32001,
6
- "transformers_version": "4.57.3"
7
  }
 
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 32001,
6
+ "transformers_version": "4.56.2"
7
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:538a84a8c3d25c5ba153e44ebe56d359fa5f92f4479c7e89f168e31570f47576
3
- size 4855832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d35b769340aa6eaee4b650a39f71198cd2a292271a2eacdf7f92cc6f1e94a26f
3
+ size 5130552
tokenizer.json CHANGED
@@ -1,7 +1,14 @@
1
  {
2
  "version": "1.0",
3
  "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
  "truncation": null,
4
+ "padding": {
5
+ "strategy": "BatchLongest",
6
+ "direction": "Left",
7
+ "pad_to_multiple_of": null,
8
+ "pad_id": 32001,
9
+ "pad_type_id": 0,
10
+ "pad_token": "<pad>"
11
+ },
12
  "added_tokens": [
13
  {
14
  "id": 0,