fix `_name_or_path` in config.json (#3)
Browse files- update config.json (0a86922a8af6074096d8b9439fd6458c27b9e71c)
- add requiements (eee24f04b87735f3deaf4df649f1dd39fe176585)
Co-authored-by: LAin <[email protected]>
- README.md +6 -0
- config.json +2 -2
- requirements.txt +2 -0
README.md
CHANGED
|
@@ -36,6 +36,12 @@ The training setup was `4xA100's 80GB` and took ~6 hours to pretrain and ~13 hou
|
|
| 36 |
---
|
| 37 |
|
| 38 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 39 |
## Training:
|
| 40 |
Training code Released !!!
|
| 41 |
https://github.com/tensoic/Cerule
|
|
|
|
| 36 |
---
|
| 37 |
|
| 38 |
|
| 39 |
+
## Installing requirements
|
| 40 |
+
```
|
| 41 |
+
pip install -qr https://huggingface.co/Tensoic/Cerule-v0.1/resolve/main/requirements.txt
|
| 42 |
+
```
|
| 43 |
+
|
| 44 |
+
|
| 45 |
## Training:
|
| 46 |
Training code Released !!!
|
| 47 |
https://github.com/tensoic/Cerule
|
config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "Tensoic/Cerule",
|
| 3 |
"architectures": [
|
| 4 |
"CeruleGemmaForCausalLM"
|
| 5 |
],
|
|
@@ -34,7 +34,7 @@
|
|
| 34 |
"tokenizer_model_max_length": 2048,
|
| 35 |
"tokenizer_padding_side": "right",
|
| 36 |
"torch_dtype": "bfloat16",
|
| 37 |
-
"transformers_version": "4.39.
|
| 38 |
"tune_mm_mlp_adapter": false,
|
| 39 |
"use_cache": true,
|
| 40 |
"use_mm_proj": true,
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "Tensoic/Cerule-v0.1",
|
| 3 |
"architectures": [
|
| 4 |
"CeruleGemmaForCausalLM"
|
| 5 |
],
|
|
|
|
| 34 |
"tokenizer_model_max_length": 2048,
|
| 35 |
"tokenizer_padding_side": "right",
|
| 36 |
"torch_dtype": "bfloat16",
|
| 37 |
+
"transformers_version": "4.39.1",
|
| 38 |
"tune_mm_mlp_adapter": false,
|
| 39 |
"use_cache": true,
|
| 40 |
"use_mm_proj": true,
|
requirements.txt
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
transformers>=4.39.1
|
| 2 |
+
flash_attn
|