Set BOS and EOS to <|im_start|> and <|im_end|> respectively (#1)
Browse files- Set BOS and EOS to <|im_start|> and <|im_end|> respectively (2ede38ff39509acb6f95f3efc75add0547e13838)
Co-authored-by: Ashwin Devaraj <[email protected]>
- config.json +3 -3
    	
        config.json
    CHANGED
    
    | @@ -3,7 +3,7 @@ | |
| 3 | 
             
                "ArcticForCausalLM"
         | 
| 4 | 
             
              ],
         | 
| 5 | 
             
              "attention_dropout": 0,
         | 
| 6 | 
            -
              "bos_token_id":  | 
| 7 | 
             
              "enable_expert_tensor_parallelism": false,
         | 
| 8 | 
             
              "enc_index": [
         | 
| 9 | 
             
                0,
         | 
| @@ -42,7 +42,7 @@ | |
| 42 | 
             
                33,
         | 
| 43 | 
             
                34
         | 
| 44 | 
             
              ],
         | 
| 45 | 
            -
              "eos_token_id":  | 
| 46 | 
             
              "hidden_act": "silu",
         | 
| 47 | 
             
              "hidden_size": 7168,
         | 
| 48 | 
             
              "initializer_range": 0.02,
         | 
| @@ -72,4 +72,4 @@ | |
| 72 | 
             
              "use_cache": true,
         | 
| 73 | 
             
              "use_residual": true,
         | 
| 74 | 
             
              "vocab_size": 32000
         | 
| 75 | 
            -
            }
         | 
|  | |
| 3 | 
             
                "ArcticForCausalLM"
         | 
| 4 | 
             
              ],
         | 
| 5 | 
             
              "attention_dropout": 0,
         | 
| 6 | 
            +
              "bos_token_id": 31998,
         | 
| 7 | 
             
              "enable_expert_tensor_parallelism": false,
         | 
| 8 | 
             
              "enc_index": [
         | 
| 9 | 
             
                0,
         | 
|  | |
| 42 | 
             
                33,
         | 
| 43 | 
             
                34
         | 
| 44 | 
             
              ],
         | 
| 45 | 
            +
              "eos_token_id": 31999,
         | 
| 46 | 
             
              "hidden_act": "silu",
         | 
| 47 | 
             
              "hidden_size": 7168,
         | 
| 48 | 
             
              "initializer_range": 0.02,
         | 
|  | |
| 72 | 
             
              "use_cache": true,
         | 
| 73 | 
             
              "use_residual": true,
         | 
| 74 | 
             
              "vocab_size": 32000
         | 
| 75 | 
            +
            }
         | 

