Update README.md
Browse files
README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
tags:
|
3 |
- long-cot-reasoning
|
4 |
- transformers
|
5 |
-
- mamba2
|
6 |
- llms
|
7 |
- chain-of-thought
|
8 |
license: apache-2.0
|
@@ -19,46 +19,54 @@ library_name: transformers
|
|
19 |
|
20 |

|
21 |
|
22 |
-
# **Sphinx:
|
23 |
-
|
24 |
-
- **Developed by:** Daemontatox
|
25 |
-
- **License:** Apache-2.0
|
26 |
-
- **Base Model:** Fine-tuned from `unsloth/qwen2.5-14b-instruct-bnb-4bit`
|
27 |
-
- **Accelerated by:** [Unsloth Framework](https://github.com/unslothai/unsloth)
|
28 |
-
- **TRL-Optimized:** Integrated with Huggingface's TRL library for enhanced performance.
|
29 |
-
|
30 |
-
## **Overview**
|
31 |
-
Sphinx is a state-of-the-art Long Chain-of-Thought (CoT) reasoning model designed to address complex, multi-step reasoning tasks with precision and clarity. Built on the Qwen2.5 architecture, Sphinx excels in generating coherent, logical thought processes while maintaining high levels of interpretability and explainability.
|
32 |
-
|
33 |
-
> _"Decoding complexity into clarity."_
|
34 |
-
|
35 |
-
### **Key Features**
|
36 |
-
- **Enhanced CoT Reasoning:** Fine-tuned for generating multi-step solutions with deep logical consistency.
|
37 |
-
- **Efficient Performance:** Powered by Unsloth, achieving 2x faster training without compromising accuracy.
|
38 |
-
- **4-bit Quantization:** Optimized for resource-constrained environments while maintaining robust performance.
|
39 |
-
- **Multi-Task Versatility:** Excels in diverse domains, including mathematical proofs, legal reasoning, and advanced scientific problem-solving.
|
40 |
-
- **TRL Integration:** Employs reinforcement learning to improve generation quality through continuous feedback loops.
|
41 |
-
|
42 |
-
## **Model Details**
|
43 |
-
### **Architecture**
|
44 |
-
- **Base Model:** Qwen2.5-14B
|
45 |
-
- **Parameters:** 14 billion
|
46 |
-
- **Quantization:** 4-bit precision using BitsAndBytes (bnb).
|
47 |
-
- **Token Window:** Supports long-form inputs with a context window of up to 16k tokens, ideal for extensive reasoning tasks.
|
48 |
-
|
49 |
-
### **Training Details**
|
50 |
-
- **Frameworks:** Huggingface Transformers + TRL + Unsloth.
|
51 |
-
- **Data Sources:** Curated datasets emphasizing reasoning tasks, including academic, legal, and logical contexts.
|
52 |
-
- **Optimization:** LoRA for parameter-efficient fine-tuning; RLHF for enhanced response alignment.
|
53 |
-
|
54 |
-
### **Capabilities**
|
55 |
-
1. **Long-CoT Generation:** Capable of breaking down and solving complex, multi-layered problems.
|
56 |
-
2. **Explainable AI (XAI):** Provides clear, step-by-step reasoning for outputs.
|
57 |
-
3. **Customizability:** Easily adaptable to niche reasoning tasks via lightweight fine-tuning.
|
58 |
-
|
59 |
-
## **Applications**
|
60 |
-
- **Academic Research:** Generating detailed, structured analyses for scientific problems.
|
61 |
-
- **Legal Assistance:** Drafting and explaining multi-step legal arguments.
|
62 |
-
- **STEM Education:** Guiding students through intricate mathematical and logical problems.
|
63 |
-
- **Cognitive AI Systems:** Seamless integration into systems requiring transparent decision-making.
|
64 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
tags:
|
3 |
- long-cot-reasoning
|
4 |
- transformers
|
5 |
+
- mamba2 # Consider updating if this isn't the architecture
|
6 |
- llms
|
7 |
- chain-of-thought
|
8 |
license: apache-2.0
|
|
|
19 |
|
20 |

|
21 |
|
22 |
+
# **Sphinx: The Apex of Logical Deduction and Chain-of-Thought Reasoning**
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
23 |
|
24 |
+
- **Developed by:** Daemontatox
|
25 |
+
- **License:** Apache-2.0
|
26 |
+
- **Base Model:** Fine-tuned from `unsloth/qwen2.5-14b-instruct-bnb-4bit`
|
27 |
+
- **Accelerated by:** [Unsloth Framework](https://github.com/unslothai/unsloth)
|
28 |
+
- **TRL-Optimized:** Integrated with Huggingface's TRL library for enhanced performance in logical reasoning.
|
29 |
+
|
30 |
+
## **Unveiling Sphinx: Master of Reasoned Thought**
|
31 |
+
|
32 |
+
Sphinx is a cutting-edge Long Chain-of-Thought (CoT) reasoning model meticulously crafted to unravel complex challenges requiring rigorous logical analysis. Built upon the robust foundation of the Qwen2.5 architecture, Sphinx excels at constructing coherent, step-by-step thought processes, providing unparalleled insight into its reasoning and ensuring clarity in its conclusions.
|
33 |
+
|
34 |
+
> _"Where complexity yields to logical clarity."_
|
35 |
+
|
36 |
+
### **Core Strengths: Reasoning, Logic, and CoT**
|
37 |
+
|
38 |
+
- **Unrivaled Chain-of-Thought (CoT) Mastery:** Engineered for dissecting intricate problems, Sphinx meticulously constructs each step of its reasoning, offering a transparent and verifiable pathway to the solution.
|
39 |
+
- **Deep Logical Reasoning Capabilities:** Sphinx is adept at navigating complex logical structures, drawing valid inferences and forming sound conclusions through multi-layered analysis.
|
40 |
+
- **Exceptional Reasoning Fidelity:** Fine-tuned to maintain the highest standards of logical consistency, Sphinx delivers outputs that are not only correct but also demonstrably well-reasoned.
|
41 |
+
- **Efficient Long-Context Reasoning:** Leveraging the power of Unsloth, Sphinx processes extensive information efficiently, maintaining logical coherence across extended reasoning chains.
|
42 |
+
- **Explainable AI through Transparent Logic:** Sphinx's inherent CoT approach provides explicit and understandable reasoning, making its decision-making process transparent and trustworthy.
|
43 |
+
|
44 |
+
## **Model Architecture and Fine-tuning for Logical Prowess**
|
45 |
+
|
46 |
+
### **Architectural Foundation**
|
47 |
+
|
48 |
+
- **Base Model:** Qwen2.5-14B - Renowned for its strong general language understanding, forming a solid basis for specialized reasoning.
|
49 |
+
- **Parameters:** 14 billion - Providing the capacity to model intricate reasoning patterns.
|
50 |
+
- **Quantization:** 4-bit precision using BitsAndBytes (bnb) - Optimizing for accessibility without sacrificing logical reasoning accuracy.
|
51 |
+
- **Extended Reasoning Window:** Supports inputs up to 16k tokens, crucial for accommodating the detailed context required for complex logical deductions.
|
52 |
+
|
53 |
+
### **Training Methodology: Honing Logical Acumen**
|
54 |
+
|
55 |
+
- **Frameworks:** Huggingface Transformers + TRL + Unsloth - A powerful combination for efficient training and reinforcement learning.
|
56 |
+
- **Data Sources:** A meticulously curated collection of datasets specifically designed to challenge and refine logical reasoning skills, encompassing academic, legal, and formal logic domains.
|
57 |
+
- **Optimization Strategies:**
|
58 |
+
- **LoRA (Low-Rank Adaptation):** Enabling parameter-efficient fine-tuning, focusing on adapting the model for superior logical inference.
|
59 |
+
- **Reinforcement Learning from Human Feedback (RLHF):** Guiding the model towards generating more logically sound and human-aligned reasoning steps.
|
60 |
+
|
61 |
+
## **Sphinx's Reasoning Toolkit: Capabilities in Action**
|
62 |
+
|
63 |
+
1. **Masterful Long-CoT Generation:** Deconstructs and conquers multi-layered problems by constructing detailed, logically interconnected reasoning sequences.
|
64 |
+
2. **Explanatory Power through Logic:** Provides clear, step-by-step logical derivations for its outputs, enhancing trust and understanding.
|
65 |
+
3. **Adaptable Logical Framework:** Easily tailored to specialized reasoning tasks through targeted fine-tuning, enabling application in diverse logical domains.
|
66 |
+
|
67 |
+
## **Unlocking Potential: Applications Driven by Logic**
|
68 |
+
|
69 |
+
- **Advanced Academic Research:** Generating in-depth, logically structured analyses for complex scientific and philosophical inquiries.
|
70 |
+
- **Robust Legal Reasoning Assistance:** Constructing and articulating multi-step legal arguments with precision and logical rigor.
|
71 |
+
- **Transformative STEM Education:** Guiding learners through intricate mathematical and logical problems with clear, step-by-step explanations.
|
72 |
+
- **Transparent Cognitive AI Systems:** Powering AI systems where explainability and logical justification are paramount for decision-making.
|