Saving weights and logs of step 10
Browse files- sinhala-gpt2/config.json → config.json +0 -0
- config.py +1 -1
- sinhala-gpt2/events.out.tfevents.1625744111.t1v-n-0b4afe6c-w-0.150137.3.v2 → events.out.tfevents.1625851395.t1v-n-0b4afe6c-w-0.251724.3.v2 +2 -2
- flax_model.msgpack +3 -0
- run_clm_flax.py +0 -1
- si_tokenizer.py +1 -1
- sinhala-gpt2/flax_model.msgpack +0 -0
- sinhala-gpt2/tokenizer.json → tokenizer.json +0 -0
sinhala-gpt2/config.json → config.json
RENAMED
|
File without changes
|
config.py
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
from transformers import GPT2Config
|
| 2 |
|
| 3 |
-
model_dir = "
|
| 4 |
|
| 5 |
config = GPT2Config.from_pretrained("gpt2", resid_pdrop=0.0, embd_pdrop=0.0, attn_pdrop=0.0)
|
| 6 |
config.save_pretrained(model_dir)
|
|
|
|
| 1 |
from transformers import GPT2Config
|
| 2 |
|
| 3 |
+
model_dir = "./sihala-gpt2"
|
| 4 |
|
| 5 |
config = GPT2Config.from_pretrained("gpt2", resid_pdrop=0.0, embd_pdrop=0.0, attn_pdrop=0.0)
|
| 6 |
config.save_pretrained(model_dir)
|
sinhala-gpt2/events.out.tfevents.1625744111.t1v-n-0b4afe6c-w-0.150137.3.v2 → events.out.tfevents.1625851395.t1v-n-0b4afe6c-w-0.251724.3.v2
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e497ce08418104de7651e94eb9dffb54daac6d986bc927df823d3b9d8a78ae45
|
| 3 |
+
size 40
|
flax_model.msgpack
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:694fdb364154b16f66f557226c6f0006187a1fe305f6696be427aa8e7f634748
|
| 3 |
+
size 497764120
|
run_clm_flax.py
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
/home/Keshan/transformers/examples/flax/language-modeling/run_clm_flax.py
|
|
|
|
|
|
si_tokenizer.py
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
from datasets import load_dataset
|
| 2 |
from tokenizers import trainers, Tokenizer, normalizers, ByteLevelBPETokenizer
|
| 3 |
|
| 4 |
-
model_dir = "
|
| 5 |
|
| 6 |
# load dataset
|
| 7 |
dataset = load_dataset("mc4", "si", split="train")
|
|
|
|
| 1 |
from datasets import load_dataset
|
| 2 |
from tokenizers import trainers, Tokenizer, normalizers, ByteLevelBPETokenizer
|
| 3 |
|
| 4 |
+
model_dir = "sinhala-gpt2" # ${MODEL_DIR}
|
| 5 |
|
| 6 |
# load dataset
|
| 7 |
dataset = load_dataset("mc4", "si", split="train")
|
sinhala-gpt2/flax_model.msgpack
DELETED
|
File without changes
|
sinhala-gpt2/tokenizer.json → tokenizer.json
RENAMED
|
File without changes
|