Upload tokenizer_config.json with huggingface_hub
Browse files- tokenizer_config.json +2 -2
    	
        tokenizer_config.json
    CHANGED
    
    | @@ -268,7 +268,7 @@ | |
| 268 | 
             
                "<|video_pad|>"
         | 
| 269 | 
             
              ],
         | 
| 270 | 
             
              "bos_token": null,
         | 
| 271 | 
            -
              "chat_template": "{ | 
| 272 | 
             
              "clean_up_tokenization_spaces": false,
         | 
| 273 | 
             
              "eos_token": "<|im_end|>",
         | 
| 274 | 
             
              "errors": "replace",
         | 
| @@ -277,4 +277,4 @@ | |
| 277 | 
             
              "split_special_tokens": false,
         | 
| 278 | 
             
              "tokenizer_class": "Qwen2Tokenizer",
         | 
| 279 | 
             
              "unk_token": null
         | 
| 280 | 
            -
            }
         | 
|  | |
| 268 | 
             
                "<|video_pad|>"
         | 
| 269 | 
             
              ],
         | 
| 270 | 
             
              "bos_token": null,
         | 
| 271 | 
            +
              "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n'}}{% if message['content'] is string %}{{ message['content'] }}{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' %}{{ '<image>\n' }}{% elif content['type'] == 'video' %}{{ '<video>\n' }}{% elif content['type'] == 'text' %}{{ content['text'] }}{% endif %}{% endfor %}{% endif %}{{'<|im_end|>\n'}}{% endfor %}{% if add_generation_prompt %}{{'<|im_start|>assistant\n' }}{% endif %}",
         | 
| 272 | 
             
              "clean_up_tokenization_spaces": false,
         | 
| 273 | 
             
              "eos_token": "<|im_end|>",
         | 
| 274 | 
             
              "errors": "replace",
         | 
|  | |
| 277 | 
             
              "split_special_tokens": false,
         | 
| 278 | 
             
              "tokenizer_class": "Qwen2Tokenizer",
         | 
| 279 | 
             
              "unk_token": null
         | 
| 280 | 
            +
            }
         | 

