Fix prompt template and EOS
Browse files- config.json +1 -1
- generation_config.json +1 -1
- tokenizer_config.json +2 -2
    	
        config.json
    CHANGED
    
    | @@ -6,7 +6,7 @@ | |
| 6 | 
             
              "attention_bias": false,
         | 
| 7 | 
             
              "attention_dropout": 0.0,
         | 
| 8 | 
             
              "bos_token_id": 128000,
         | 
| 9 | 
            -
              "eos_token_id":  | 
| 10 | 
             
              "hidden_act": "silu",
         | 
| 11 | 
             
              "hidden_size": 8192,
         | 
| 12 | 
             
              "initializer_range": 0.02,
         | 
|  | |
| 6 | 
             
              "attention_bias": false,
         | 
| 7 | 
             
              "attention_dropout": 0.0,
         | 
| 8 | 
             
              "bos_token_id": 128000,
         | 
| 9 | 
            +
              "eos_token_id": 128257,
         | 
| 10 | 
             
              "hidden_act": "silu",
         | 
| 11 | 
             
              "hidden_size": 8192,
         | 
| 12 | 
             
              "initializer_range": 0.02,
         | 
    	
        generation_config.json
    CHANGED
    
    | @@ -2,6 +2,6 @@ | |
| 2 | 
             
              "_from_model_config": true,
         | 
| 3 | 
             
              "bos_token_id": 128000,
         | 
| 4 | 
             
              "do_sample": true,
         | 
| 5 | 
            -
              "eos_token_id":  | 
| 6 | 
             
              "transformers_version": "4.40.1"
         | 
| 7 | 
             
            }
         | 
|  | |
| 2 | 
             
              "_from_model_config": true,
         | 
| 3 | 
             
              "bos_token_id": 128000,
         | 
| 4 | 
             
              "do_sample": true,
         | 
| 5 | 
            +
              "eos_token_id": 128257,
         | 
| 6 | 
             
              "transformers_version": "4.40.1"
         | 
| 7 | 
             
            }
         | 
    	
        tokenizer_config.json
    CHANGED
    
    | @@ -2066,9 +2066,9 @@ | |
| 2066 | 
             
                }
         | 
| 2067 | 
             
              },
         | 
| 2068 | 
             
              "bos_token": "<|begin_of_text|>",
         | 
| 2069 | 
            -
              "chat_template": "{ | 
| 2070 | 
             
              "clean_up_tokenization_spaces": true,
         | 
| 2071 | 
            -
              "eos_token": "<| | 
| 2072 | 
             
              "model_input_names": [
         | 
| 2073 | 
             
                "input_ids",
         | 
| 2074 | 
             
                "attention_mask"
         | 
|  | |
| 2066 | 
             
                }
         | 
| 2067 | 
             
              },
         | 
| 2068 | 
             
              "bos_token": "<|begin_of_text|>",
         | 
| 2069 | 
            +
              "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nBelow is a conversation between a curious user and a helpful AI assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content']}}{% if (loop.last and add_generation_prompt) or not loop.last %}{{ '<|im_end|>' + '\n'}}{% endif %}{% endfor %}{% if add_generation_prompt and messages[-1]['role'] != 'assistant' %}{{ '<|im_start|>assistant\n' }}{% endif %}",
         | 
| 2070 | 
             
              "clean_up_tokenization_spaces": true,
         | 
| 2071 | 
            +
              "eos_token": "<|im_end|>",
         | 
| 2072 | 
             
              "model_input_names": [
         | 
| 2073 | 
             
                "input_ids",
         | 
| 2074 | 
             
                "attention_mask"
         | 
