Upload tokenizer
Browse files- merges.txt +1 -1
- tokenizer.json +0 -1
merges.txt
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
-
#version: 0.2
|
| 2 |
Ġ Ġ
|
| 3 |
ĠĠ ĠĠ
|
| 4 |
ĠĠĠĠ ĠĠĠĠ
|
|
|
|
| 1 |
+
#version: 0.2 - Trained by `huggingface/tokenizers`
|
| 2 |
Ġ Ġ
|
| 3 |
ĠĠ ĠĠ
|
| 4 |
ĠĠĠĠ ĠĠĠĠ
|
tokenizer.json
CHANGED
|
@@ -205,7 +205,6 @@
|
|
| 205 |
"continuing_subword_prefix": null,
|
| 206 |
"end_of_word_suffix": null,
|
| 207 |
"fuse_unk": false,
|
| 208 |
-
"byte_fallback": false,
|
| 209 |
"vocab": {
|
| 210 |
"<|endoftext|>": 0,
|
| 211 |
"<fim_prefix>": 1,
|
|
|
|
| 205 |
"continuing_subword_prefix": null,
|
| 206 |
"end_of_word_suffix": null,
|
| 207 |
"fuse_unk": false,
|
|
|
|
| 208 |
"vocab": {
|
| 209 |
"<|endoftext|>": 0,
|
| 210 |
"<fim_prefix>": 1,
|