Upload GitForCausalLM
Browse files- config.json +2 -1
- pytorch_model.bin +2 -2
config.json
CHANGED
|
@@ -17,6 +17,7 @@
|
|
| 17 |
"model_type": "git",
|
| 18 |
"num_attention_heads": 12,
|
| 19 |
"num_hidden_layers": 6,
|
|
|
|
| 20 |
"pad_token_id": 0,
|
| 21 |
"position_embedding_type": "absolute",
|
| 22 |
"tie_word_embeddings": false,
|
|
@@ -97,7 +98,7 @@
|
|
| 97 |
"top_p": 1.0,
|
| 98 |
"torch_dtype": null,
|
| 99 |
"torchscript": false,
|
| 100 |
-
"transformers_version": "4.
|
| 101 |
"typical_p": 1.0,
|
| 102 |
"use_bfloat16": false
|
| 103 |
},
|
|
|
|
| 17 |
"model_type": "git",
|
| 18 |
"num_attention_heads": 12,
|
| 19 |
"num_hidden_layers": 6,
|
| 20 |
+
"num_image_with_embedding": null,
|
| 21 |
"pad_token_id": 0,
|
| 22 |
"position_embedding_type": "absolute",
|
| 23 |
"tie_word_embeddings": false,
|
|
|
|
| 98 |
"top_p": 1.0,
|
| 99 |
"torch_dtype": null,
|
| 100 |
"torchscript": false,
|
| 101 |
+
"transformers_version": "4.26.0.dev0",
|
| 102 |
"typical_p": 1.0,
|
| 103 |
"use_bfloat16": false
|
| 104 |
},
|
pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e75e4c723e2ff445797b48db5d19bbe92ede848837abab5f4f43917e3db2bf20
|
| 3 |
+
size 708756315
|