system
		
	commited on
		
		
					Update README.md
Browse files
    	
        README.md
    CHANGED
    
    | @@ -1,41 +1,23 @@ | |
| 1 | 
             
            ---
         | 
| 2 | 
             
            base_model:
         | 
| 3 | 
            -
            -  | 
| 4 | 
            -
            - Blazgo/temp-model-for-2-mini-007-final
         | 
| 5 | 
             
            library_name: transformers
         | 
| 6 | 
             
            tags:
         | 
| 7 | 
            -
            -  | 
| 8 | 
            -
            -  | 
| 9 | 
            -
             | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 10 | 
             
            ---
         | 
| 11 | 
            -
             | 
| 12 | 
            -
             | 
| 13 | 
            -
             | 
| 14 | 
            -
             | 
| 15 | 
            -
            ## Merge Details
         | 
| 16 | 
            -
            ### Merge Method
         | 
| 17 | 
            -
             | 
| 18 | 
            -
            This model was merged using the [SLERP](https://en.wikipedia.org/wiki/Slerp) merge method.
         | 
| 19 | 
            -
             | 
| 20 | 
            -
            ### Models Merged
         | 
| 21 | 
            -
             | 
| 22 | 
            -
            The following models were included in the merge:
         | 
| 23 | 
            -
            * [Blazgo/temp-model-for-2-mini-008-final](https://huggingface.co/Blazgo/temp-model-for-2-mini-008-final)
         | 
| 24 | 
            -
            * [Blazgo/temp-model-for-2-mini-007-final](https://huggingface.co/Blazgo/temp-model-for-2-mini-007-final)
         | 
| 25 | 
            -
             | 
| 26 | 
            -
            ### Configuration
         | 
| 27 | 
            -
             | 
| 28 | 
            -
            The following YAML configuration was used to produce this model:
         | 
| 29 | 
            -
             | 
| 30 | 
            -
            ```yaml
         | 
| 31 |  | 
| 32 | 
            -
             | 
| 33 | 
            -
              - model: Blazgo/temp-model-for-2-mini-007-final
         | 
| 34 | 
            -
              - model: Blazgo/temp-model-for-2-mini-008-final
         | 
| 35 | 
            -
            merge_method: slerp
         | 
| 36 | 
            -
            base_model: Blazgo/temp-model-for-2-mini-007-final
         | 
| 37 | 
            -
            dtype: bfloat16
         | 
| 38 | 
            -
            parameters:
         | 
| 39 | 
            -
              t: [0, 0.5, 1, 0.5, 0] # V shaped curve: Hermes for input & output, WizardMath in the middle layers
         | 
| 40 |  | 
| 41 | 
            -
             | 
|  | 
|  | |
| 1 | 
             
            ---
         | 
| 2 | 
             
            base_model:
         | 
| 3 | 
            +
            - deepseek-ai/DeepSeek-R1
         | 
|  | |
| 4 | 
             
            library_name: transformers
         | 
| 5 | 
             
            tags:
         | 
| 6 | 
            +
            - reasoning
         | 
| 7 | 
            +
            - R1
         | 
| 8 | 
            +
            - 1M
         | 
| 9 | 
            +
            - fast
         | 
| 10 | 
            +
            - Deca
         | 
| 11 | 
            +
            - Deca-AI
         | 
| 12 | 
            +
            - Deca-2
         | 
| 13 | 
            +
            - Qwen
         | 
| 14 | 
            +
            license: other
         | 
| 15 | 
             
            ---
         | 
| 16 | 
            +
            The Deca 2 family of models is currently in BETA
         | 
| 17 | 
            +
            
         | 
| 18 | 
            +
            The Deca 2 family of models, currently in BETA, is built on cutting-edge architectures like DeepSeek R1, and Qwen 2, delivering extraordinary performance. With a focus on insane speed and high efficiency, Deca 2 is revolutionizing text generation and setting new standards in the industry. It also comes with a **1 million** context window.
         | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 19 |  | 
| 20 | 
            +
            As more capabilities are added, Deca 2 will evolve into a more powerful, any-to-any model in the future. While it’s focused on text generation for now, its foundation is designed to scale, bringing even more advanced functionalities to come.
         | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 21 |  | 
| 22 | 
            +
            * **2/14 Realase:**
         | 
| 23 | 
            +
            * Enhanced Instruction Following
         | 
