Upload SafeCoder LoRA model
Browse files- adapter_config.json +6 -6
- adapter_model.bin +2 -2
    	
        adapter_config.json
    CHANGED
    
    | @@ -1,6 +1,6 @@ | |
| 1 | 
             
            {
         | 
| 2 | 
             
              "auto_mapping": null,
         | 
| 3 | 
            -
              "base_model_name_or_path": "Qwen/Qwen2.5-Coder- | 
| 4 | 
             
              "bias": "none",
         | 
| 5 | 
             
              "fan_in_fan_out": false,
         | 
| 6 | 
             
              "inference_mode": true,
         | 
| @@ -14,14 +14,14 @@ | |
| 14 | 
             
              "r": 16,
         | 
| 15 | 
             
              "revision": null,
         | 
| 16 | 
             
              "target_modules": [
         | 
| 17 | 
            -
                "q_proj",
         | 
| 18 | 
            -
                "gate_proj",
         | 
| 19 | 
            -
                "v_proj",
         | 
| 20 | 
             
                "lm_head",
         | 
| 21 | 
            -
                "k_proj",
         | 
| 22 | 
             
                "up_proj",
         | 
|  | |
|  | |
|  | |
| 23 | 
             
                "o_proj",
         | 
| 24 | 
            -
                "down_proj"
         | 
|  | |
| 25 | 
             
              ],
         | 
| 26 | 
             
              "task_type": "CAUSAL_LM"
         | 
| 27 | 
             
            }
         | 
|  | |
| 1 | 
             
            {
         | 
| 2 | 
             
              "auto_mapping": null,
         | 
| 3 | 
            +
              "base_model_name_or_path": "Qwen/Qwen2.5-Coder-3B-Instruct",
         | 
| 4 | 
             
              "bias": "none",
         | 
| 5 | 
             
              "fan_in_fan_out": false,
         | 
| 6 | 
             
              "inference_mode": true,
         | 
|  | |
| 14 | 
             
              "r": 16,
         | 
| 15 | 
             
              "revision": null,
         | 
| 16 | 
             
              "target_modules": [
         | 
|  | |
|  | |
|  | |
| 17 | 
             
                "lm_head",
         | 
|  | |
| 18 | 
             
                "up_proj",
         | 
| 19 | 
            +
                "q_proj",
         | 
| 20 | 
            +
                "k_proj",
         | 
| 21 | 
            +
                "v_proj",
         | 
| 22 | 
             
                "o_proj",
         | 
| 23 | 
            +
                "down_proj",
         | 
| 24 | 
            +
                "gate_proj"
         | 
| 25 | 
             
              ],
         | 
| 26 | 
             
              "task_type": "CAUSAL_LM"
         | 
| 27 | 
             
            }
         | 
    	
        adapter_model.bin
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
            -
            size  | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:b3f52d34ea1ee084a8ca92ef459e83adaf43e9603c86ce187f1da0b501e245c2
         | 
| 3 | 
            +
            size 129753589
         |