qgallouedec HF Staff commited on
Commit
2cb3712
·
verified ·
1 Parent(s): b4eeae1

Upload PaliGemmaForConditionalGeneration

Browse files
Files changed (2) hide show
  1. config.json +18 -9
  2. model.safetensors +2 -2
config.json CHANGED
@@ -2,10 +2,14 @@
2
  "architectures": [
3
  "PaliGemmaForConditionalGeneration"
4
  ],
 
 
5
  "hidden_size": 2048,
6
- "image_token_index": 256000,
 
7
  "model_type": "paligemma",
8
- "projection_dim": 8,
 
9
  "text_config": {
10
  "attention_bias": false,
11
  "attention_dropout": 0.0,
@@ -14,32 +18,37 @@
14
  "hidden_activation": null,
15
  "hidden_size": 16,
16
  "initializer_range": 0.02,
17
- "intermediate_size": 32,
18
  "max_position_embeddings": 8192,
19
  "model_type": "gemma",
20
  "num_attention_heads": 4,
21
  "num_hidden_layers": 2,
22
- "num_image_tokens": 196,
23
  "num_key_value_heads": 2,
24
  "rms_norm_eps": 1e-06,
25
  "rope_theta": 10000.0,
 
26
  "use_cache": true,
27
- "vocab_size": 258522
28
  },
29
- "torch_dtype": "float32",
30
  "transformers_version": "4.56.0.dev0",
31
  "vision_config": {
32
  "attention_dropout": 0.0,
33
  "hidden_act": "gelu_pytorch_tanh",
34
  "hidden_size": 16,
35
  "image_size": 224,
36
- "intermediate_size": 32,
37
  "layer_norm_eps": 1e-06,
38
  "model_type": "siglip_vision_model",
39
  "num_attention_heads": 4,
40
  "num_channels": 3,
41
  "num_hidden_layers": 2,
42
- "patch_size": 16,
43
- "projection_dim": 8
 
 
 
 
44
  }
45
  }
 
2
  "architectures": [
3
  "PaliGemmaForConditionalGeneration"
4
  ],
5
+ "bos_token_id": 2,
6
+ "eos_token_id": 1,
7
  "hidden_size": 2048,
8
+ "ignore_index": -100,
9
+ "image_token_index": 257152,
10
  "model_type": "paligemma",
11
+ "pad_token_id": 0,
12
+ "projection_dim": 2048,
13
  "text_config": {
14
  "attention_bias": false,
15
  "attention_dropout": 0.0,
 
18
  "hidden_activation": null,
19
  "hidden_size": 16,
20
  "initializer_range": 0.02,
21
+ "intermediate_size": 16384,
22
  "max_position_embeddings": 8192,
23
  "model_type": "gemma",
24
  "num_attention_heads": 4,
25
  "num_hidden_layers": 2,
26
+ "num_image_tokens": 256,
27
  "num_key_value_heads": 2,
28
  "rms_norm_eps": 1e-06,
29
  "rope_theta": 10000.0,
30
+ "torch_dtype": "float32",
31
  "use_cache": true,
32
+ "vocab_size": 257216
33
  },
34
+ "torch_dtype": "bfloat16",
35
  "transformers_version": "4.56.0.dev0",
36
  "vision_config": {
37
  "attention_dropout": 0.0,
38
  "hidden_act": "gelu_pytorch_tanh",
39
  "hidden_size": 16,
40
  "image_size": 224,
41
+ "intermediate_size": 4304,
42
  "layer_norm_eps": 1e-06,
43
  "model_type": "siglip_vision_model",
44
  "num_attention_heads": 4,
45
  "num_channels": 3,
46
  "num_hidden_layers": 2,
47
+ "num_image_tokens": 256,
48
+ "num_key_value_heads": 2,
49
+ "patch_size": 14,
50
+ "projection_dim": 2048,
51
+ "projector_hidden_act": "gelu_fast",
52
+ "vision_use_head": false
53
  }
54
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5601312b285dba1d671591395f2e624a08b1dd014d9686fe0b0e2bf329afa9c9
3
- size 17048992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62bd764abe98bd832ce30f426108b70e2790384596a9ea8257a192d16b219966
3
+ size 12250448