prince-canuma commited on
Commit
004ef05
·
verified ·
1 Parent(s): 5ee8b2f

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -8,8 +8,6 @@
8
  "output_hidden_states": false,
9
  "torchscript": false,
10
  "torch_dtype": "bfloat16",
11
- "use_bfloat16": false,
12
- "tf_legacy_loss": false,
13
  "pruned_heads": {},
14
  "tie_word_embeddings": true,
15
  "chunk_size_feed_forward": 0,
@@ -18,6 +16,25 @@
18
  "cross_attention_hidden_size": null,
19
  "add_cross_attention": false,
20
  "tie_encoder_decoder": false,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  "max_length": 20,
22
  "min_length": 0,
23
  "do_sample": false,
@@ -43,27 +60,10 @@
43
  "exponential_decay_length_penalty": null,
44
  "suppress_tokens": null,
45
  "begin_suppress_tokens": null,
46
- "architectures": null,
47
- "finetuning_task": null,
48
- "id2label": {
49
- "0": "LABEL_0",
50
- "1": "LABEL_1"
51
- },
52
- "label2id": {
53
- "LABEL_0": 0,
54
- "LABEL_1": 1
55
- },
56
- "tokenizer_class": null,
57
- "prefix": null,
58
- "bos_token_id": null,
59
- "pad_token_id": null,
60
- "eos_token_id": null,
61
- "sep_token_id": null,
62
- "decoder_start_token_id": null,
63
- "task_specific_params": null,
64
- "problem_type": null,
65
  "_name_or_path": "",
66
  "model_type": "gemma3n_audio",
 
 
67
  "input_feat_size": 128,
68
  "hidden_size": 1536,
69
  "rms_norm_eps": 1e-06,
@@ -3940,8 +3940,6 @@
3940
  "output_hidden_states": false,
3941
  "torchscript": false,
3942
  "torch_dtype": "bfloat16",
3943
- "use_bfloat16": false,
3944
- "tf_legacy_loss": false,
3945
  "pruned_heads": {},
3946
  "tie_word_embeddings": true,
3947
  "chunk_size_feed_forward": 0,
@@ -3950,6 +3948,25 @@
3950
  "cross_attention_hidden_size": null,
3951
  "add_cross_attention": false,
3952
  "tie_encoder_decoder": false,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3953
  "max_length": 20,
3954
  "min_length": 0,
3955
  "do_sample": false,
@@ -3975,27 +3992,10 @@
3975
  "exponential_decay_length_penalty": null,
3976
  "suppress_tokens": null,
3977
  "begin_suppress_tokens": null,
3978
- "architectures": null,
3979
- "finetuning_task": null,
3980
- "id2label": {
3981
- "0": "LABEL_0",
3982
- "1": "LABEL_1"
3983
- },
3984
- "label2id": {
3985
- "LABEL_0": 0,
3986
- "LABEL_1": 1
3987
- },
3988
- "tokenizer_class": null,
3989
- "prefix": null,
3990
- "bos_token_id": 2,
3991
- "pad_token_id": 0,
3992
- "eos_token_id": 1,
3993
- "sep_token_id": null,
3994
- "decoder_start_token_id": null,
3995
- "task_specific_params": null,
3996
- "problem_type": null,
3997
  "_name_or_path": "",
3998
  "model_type": "gemma3n_text",
 
 
3999
  "vocab_size": 262400,
4000
  "vocab_size_per_layer_input": 262144,
4001
  "max_position_embeddings": 32768,
@@ -4142,7 +4142,7 @@
4142
  "top_k": 50,
4143
  "top_p": 1.0,
4144
  "torchscript": false,
4145
- "transformers_version": "4.53.2",
4146
  "typical_p": 1.0,
4147
  "use_bfloat16": false,
4148
  "vision_config": {
@@ -4150,8 +4150,6 @@
4150
  "output_hidden_states": false,
4151
  "torchscript": false,
4152
  "torch_dtype": "bfloat16",
4153
- "use_bfloat16": false,
4154
- "tf_legacy_loss": false,
4155
  "pruned_heads": {},
4156
  "tie_word_embeddings": true,
4157
  "chunk_size_feed_forward": 0,
@@ -4160,6 +4158,17 @@
4160
  "cross_attention_hidden_size": null,
4161
  "add_cross_attention": false,
4162
  "tie_encoder_decoder": false,
 
 
 
 
 
 
 
 
 
 
 
4163
  "max_length": 20,
4164
  "min_length": 0,
4165
  "do_sample": false,
@@ -4185,17 +4194,6 @@
4185
  "exponential_decay_length_penalty": null,
4186
  "suppress_tokens": null,
4187
  "begin_suppress_tokens": null,
4188
- "architectures": null,
4189
- "finetuning_task": null,
4190
- "tokenizer_class": null,
4191
- "prefix": null,
4192
- "bos_token_id": null,
4193
- "pad_token_id": null,
4194
- "eos_token_id": null,
4195
- "sep_token_id": null,
4196
- "decoder_start_token_id": null,
4197
- "task_specific_params": null,
4198
- "problem_type": null,
4199
  "_name_or_path": "",
4200
  "label_names": [
4201
  "LABEL_0",
@@ -4203,6 +4201,8 @@
4203
  ],
4204
  "model_type": "gemma3n_vision",
4205
  "num_classes": 2,
 
 
4206
  "initializer_range": 0.02,
4207
  "do_pooling": false,
4208
  "model_args": null,
 
8
  "output_hidden_states": false,
9
  "torchscript": false,
10
  "torch_dtype": "bfloat16",
 
 
11
  "pruned_heads": {},
12
  "tie_word_embeddings": true,
13
  "chunk_size_feed_forward": 0,
 
16
  "cross_attention_hidden_size": null,
17
  "add_cross_attention": false,
18
  "tie_encoder_decoder": false,
19
+ "architectures": null,
20
+ "finetuning_task": null,
21
+ "id2label": {
22
+ "0": "LABEL_0",
23
+ "1": "LABEL_1"
24
+ },
25
+ "label2id": {
26
+ "LABEL_0": 0,
27
+ "LABEL_1": 1
28
+ },
29
+ "task_specific_params": null,
30
+ "problem_type": null,
31
+ "tokenizer_class": null,
32
+ "prefix": null,
33
+ "bos_token_id": null,
34
+ "pad_token_id": null,
35
+ "eos_token_id": null,
36
+ "sep_token_id": null,
37
+ "decoder_start_token_id": null,
38
  "max_length": 20,
39
  "min_length": 0,
40
  "do_sample": false,
 
60
  "exponential_decay_length_penalty": null,
61
  "suppress_tokens": null,
62
  "begin_suppress_tokens": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
63
  "_name_or_path": "",
64
  "model_type": "gemma3n_audio",
65
+ "tf_legacy_loss": false,
66
+ "use_bfloat16": false,
67
  "input_feat_size": 128,
68
  "hidden_size": 1536,
69
  "rms_norm_eps": 1e-06,
 
3940
  "output_hidden_states": false,
3941
  "torchscript": false,
3942
  "torch_dtype": "bfloat16",
 
 
3943
  "pruned_heads": {},
3944
  "tie_word_embeddings": true,
3945
  "chunk_size_feed_forward": 0,
 
3948
  "cross_attention_hidden_size": null,
3949
  "add_cross_attention": false,
3950
  "tie_encoder_decoder": false,
3951
+ "architectures": null,
3952
+ "finetuning_task": null,
3953
+ "id2label": {
3954
+ "0": "LABEL_0",
3955
+ "1": "LABEL_1"
3956
+ },
3957
+ "label2id": {
3958
+ "LABEL_0": 0,
3959
+ "LABEL_1": 1
3960
+ },
3961
+ "task_specific_params": null,
3962
+ "problem_type": null,
3963
+ "tokenizer_class": null,
3964
+ "prefix": null,
3965
+ "bos_token_id": 2,
3966
+ "pad_token_id": 0,
3967
+ "eos_token_id": 1,
3968
+ "sep_token_id": null,
3969
+ "decoder_start_token_id": null,
3970
  "max_length": 20,
3971
  "min_length": 0,
3972
  "do_sample": false,
 
3992
  "exponential_decay_length_penalty": null,
3993
  "suppress_tokens": null,
3994
  "begin_suppress_tokens": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3995
  "_name_or_path": "",
3996
  "model_type": "gemma3n_text",
3997
+ "tf_legacy_loss": false,
3998
+ "use_bfloat16": false,
3999
  "vocab_size": 262400,
4000
  "vocab_size_per_layer_input": 262144,
4001
  "max_position_embeddings": 32768,
 
4142
  "top_k": 50,
4143
  "top_p": 1.0,
4144
  "torchscript": false,
4145
+ "transformers_version": "4.54.0.dev0",
4146
  "typical_p": 1.0,
4147
  "use_bfloat16": false,
4148
  "vision_config": {
 
4150
  "output_hidden_states": false,
4151
  "torchscript": false,
4152
  "torch_dtype": "bfloat16",
 
 
4153
  "pruned_heads": {},
4154
  "tie_word_embeddings": true,
4155
  "chunk_size_feed_forward": 0,
 
4158
  "cross_attention_hidden_size": null,
4159
  "add_cross_attention": false,
4160
  "tie_encoder_decoder": false,
4161
+ "architectures": null,
4162
+ "finetuning_task": null,
4163
+ "task_specific_params": null,
4164
+ "problem_type": null,
4165
+ "tokenizer_class": null,
4166
+ "prefix": null,
4167
+ "bos_token_id": null,
4168
+ "pad_token_id": null,
4169
+ "eos_token_id": null,
4170
+ "sep_token_id": null,
4171
+ "decoder_start_token_id": null,
4172
  "max_length": 20,
4173
  "min_length": 0,
4174
  "do_sample": false,
 
4194
  "exponential_decay_length_penalty": null,
4195
  "suppress_tokens": null,
4196
  "begin_suppress_tokens": null,
 
 
 
 
 
 
 
 
 
 
 
4197
  "_name_or_path": "",
4198
  "label_names": [
4199
  "LABEL_0",
 
4201
  ],
4202
  "model_type": "gemma3n_vision",
4203
  "num_classes": 2,
4204
+ "tf_legacy_loss": false,
4205
+ "use_bfloat16": false,
4206
  "initializer_range": 0.02,
4207
  "do_pooling": false,
4208
  "model_args": null,
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:709e3fa64911e5a753395f0e18181471794f239baffd7dac455f4befcb326df1
3
- size 5364004839
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35481571d6ab17a60515c4378e3f3c171edb0e7fbe908e8db9d0f49f2838a544
3
+ size 5364004911
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ffa02c7d3402c85b43798b966bb1a9da3b7ff1b2312909770a3321f3327d9ba9
3
  size 455053642
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62ae1980a669cca260a3a3a4a3efc4285d861b9be2df4c7da98c8ac61420d1f7
3
  size 455053642