qgallouedec HF Staff commited on
Commit
b0d3026
·
verified ·
1 Parent(s): 3d5a777

Upload SmolVLMForConditionalGeneration

Browse files
Files changed (3) hide show
  1. config.json +5 -3
  2. generation_config.json +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -2,6 +2,7 @@
2
  "architectures": [
3
  "SmolVLMForConditionalGeneration"
4
  ],
 
5
  "image_token_id": 49190,
6
  "model_type": "smolvlm",
7
  "pad_token_id": 128002,
@@ -14,11 +15,13 @@
14
  ],
15
  "attention_bias": false,
16
  "attention_dropout": 0.0,
 
17
  "head_dim": 64,
18
  "hidden_act": "silu",
19
  "hidden_size": 16,
20
  "initializer_range": 0.02,
21
  "intermediate_size": 8192,
 
22
  "max_position_embeddings": 8192,
23
  "mlp_bias": false,
24
  "model_type": "llama",
@@ -106,7 +109,6 @@
106
  "rms_norm_eps": 1e-05,
107
  "rope_scaling": null,
108
  "rope_theta": 130000,
109
- "torch_dtype": "bfloat16",
110
  "transformers.js_config": {
111
  "kv_cache_dtype": {
112
  "fp16": "float16",
@@ -118,12 +120,12 @@
118
  "vocab_size": 49280
119
  },
120
  "tie_word_embeddings": false,
121
- "torch_dtype": "bfloat16",
122
- "transformers_version": "4.56.0.dev0",
123
  "use_cache": false,
124
  "use_reentrant_checkpointing": false,
125
  "vision_config": {
126
  "attention_dropout": 0.0,
 
127
  "hidden_act": "gelu_pytorch_tanh",
128
  "hidden_size": 16,
129
  "image_size": 384,
 
2
  "architectures": [
3
  "SmolVLMForConditionalGeneration"
4
  ],
5
+ "dtype": "bfloat16",
6
  "image_token_id": 49190,
7
  "model_type": "smolvlm",
8
  "pad_token_id": 128002,
 
15
  ],
16
  "attention_bias": false,
17
  "attention_dropout": 0.0,
18
+ "dtype": "bfloat16",
19
  "head_dim": 64,
20
  "hidden_act": "silu",
21
  "hidden_size": 16,
22
  "initializer_range": 0.02,
23
  "intermediate_size": 8192,
24
+ "layer_types": null,
25
  "max_position_embeddings": 8192,
26
  "mlp_bias": false,
27
  "model_type": "llama",
 
109
  "rms_norm_eps": 1e-05,
110
  "rope_scaling": null,
111
  "rope_theta": 130000,
 
112
  "transformers.js_config": {
113
  "kv_cache_dtype": {
114
  "fp16": "float16",
 
120
  "vocab_size": 49280
121
  },
122
  "tie_word_embeddings": false,
123
+ "transformers_version": "4.57.0.dev0",
 
124
  "use_cache": false,
125
  "use_reentrant_checkpointing": false,
126
  "vision_config": {
127
  "attention_dropout": 0.0,
128
+ "embed_dim": 32,
129
  "hidden_act": "gelu_pytorch_tanh",
130
  "hidden_size": 16,
131
  "image_size": 384,
generation_config.json CHANGED
@@ -3,6 +3,6 @@
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 128002,
6
- "transformers_version": "4.56.0.dev0",
7
  "use_cache": false
8
  }
 
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 128002,
6
+ "transformers_version": "4.57.0.dev0",
7
  "use_cache": false
8
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76fe0e0e6ec46fb0d7b81da1c30fb06aca991774e4378a41a7c0383f567981cf
3
  size 5402800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08c9b5478b3a81be135acd5aac1e8aff8d375fbb27daa8d5228161dbd17c76bd
3
  size 5402800