klei1 commited on
Commit
babb460
·
verified ·
1 Parent(s): c852883

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +97 -9
README.md CHANGED
@@ -1,21 +1,109 @@
1
  ---
2
- base_model: unsloth/gemma-3-27b-it-unsloth-bnb-4bit
3
  tags:
4
  - text-generation-inference
5
- - transformers
6
- - unsloth
 
7
  - gemma3
 
 
 
 
 
 
8
  license: apache-2.0
9
  language:
10
- - sq
 
 
 
 
 
 
11
  ---
12
 
13
- # Uploaded finetuned model
14
 
15
- - **Developed by:** klei1
 
 
16
  - **License:** apache-2.0
17
- - **Finetuned from model :** unsloth/gemma-3-27b-it-unsloth-bnb-4bit
 
 
18
 
19
- This gemma3 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
20
 
21
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: bleta-logjike-27b
3
  tags:
4
  - text-generation-inference
5
+ - llama.cpp
6
+ - gguf
7
+ - albanian
8
  - gemma3
9
+ - reasoning
10
+ - logical-reasoning
11
+ - grpo
12
+ - gsm8k
13
+ - mathematics
14
+ - llm
15
  license: apache-2.0
16
  language:
17
+ - al
18
+ inference:
19
+ parameters:
20
+ temperature: 0.7
21
+ top_p: 0.95
22
+ top_k: 64
23
+ max_new_tokens: 512
24
  ---
25
 
26
+ # Bleta-Logjike 27B Albanian Logical Reasoning Model (GGUF)
27
 
28
+ ## Model Description
29
+ - **Developed by:** klei aliaj
30
+ - **Model type:** Bleta-Logjike 27B optimized for Albanian logical reasoning
31
  - **License:** apache-2.0
32
+ - **Format:** GGUF 8-bit quantized for llama.cpp
33
+ - **Language:** Albanian
34
+ - **Base architecture:** Based on Gemma 3 27B
35
 
36
+ This model is a GGUF quantized version of the Bleta-Logjike 27B model, specifically optimized for logical reasoning tasks in the Albanian language. Bleta is an Albanian adaptation based on Google's Gemma 3 architecture, with this version focused on enhancing logical reasoning and problem-solving capabilities.
37
 
38
+ ## Capabilities & Features
39
+
40
+ ### Logical Reasoning Focus
41
+ This Albanian language model excels at:
42
+
43
+ 1. Logical analysis and deduction in Albanian
44
+ 2. Step-by-step problem solving
45
+ 3. Structured reasoning for complex problems
46
+ 4. Understanding logical relationships and dependencies
47
+ 5. Mathematical reasoning for grade-school level problems
48
+
49
+ ### GGUF Quantization Benefits
50
+ - **Efficient inference:** Optimized for use with llama.cpp and similar frameworks
51
+ - **Reduced memory usage:** 8-bit quantization substantially reduces RAM requirements
52
+ - **Faster inference:** More efficient processing for consumer hardware
53
+ - **Compatible with:** llama.cpp, Jan AI, LM Studio, and other GGUF-compatible applications
54
+
55
+ ### Albanian Language Optimization
56
+ - Native support for Albanian grammar and vocabulary
57
+ - Understanding of Albanian cultural context
58
+ - Handling of Albanian-specific logical expressions and constructs
59
+
60
+ ## Training Methodology
61
+
62
+ ### GRPO Approach
63
+ This model was fine-tuned using Generative Rejection Policy Optimization (GRPO), a reinforcement learning technique that trains models to optimize for specific reward functions. GRPO allows the model to learn from feedback on its generated responses, improving reasoning quality over time by:
64
+
65
+ 1. Generating multiple candidate responses
66
+ 2. Evaluating responses against specific reward criteria
67
+ 3. Learning to prefer high-quality reasoning patterns
68
+ 4. Optimizing for step-by-step problem solving
69
+
70
+ ### GSM8K Dataset
71
+ The training utilized the GSM8K (Grade School Math 8K) dataset, which contains over 8,000 high-quality grade school math problems, requiring step-by-step reasoning to solve. The dataset provides:
72
+
73
+ - Diverse mathematical problem types
74
+ - Multi-step reasoning challenges
75
+ - Clear step-by-step solutions
76
+ - Grade-school level complexity
77
+
78
+ This dataset was adapted for Albanian language training to ensure the model can handle mathematical reasoning tasks in Albanian.
79
+
80
+ ## Technical Specifications
81
+
82
+ ### Model Architecture
83
+ - 27B parameters
84
+ - Based on Gemma 3 architecture with Albanian adaptations
85
+ - 128K context window
86
+ - QK normalization
87
+ - 5 sliding + 1 global attention pattern
88
+ - 1024 sliding window attention
89
+
90
+ ### Usage Requirements
91
+ - Recommended minimum 16GB RAM for inference
92
+ - Compatible with CPU inference but GPU recommended
93
+ - Works with llama.cpp and compatible UIs
94
+
95
+ ## Limitations
96
+
97
+ The current model is an 8-bit quantized version of the 27B parameter model. This quantization offers advantages in terms of size and speed, but comes with some limitations:
98
+
99
+ - Reduced precision compared to the original 16-bit or 32-bit model
100
+ - May exhibit occasional numerical instabilities in complex reasoning chains
101
+ - While optimized for logical reasoning in Albanian, complex or ambiguous problems may produce inconsistent results
102
+ - As with all language models, it may occasionally hallucinate or provide incorrect information
103
+ - Performance may vary depending on the complexity and clarity of the input prompts
104
+
105
+ ## Acknowledgments
106
+ - Google for developing the Gemma 3 architecture
107
+ - llama.cpp team for the GGUF format and inference engine
108
+ - OpenAI for the GSM8K dataset
109
+ - Hugging Face for their TRL library and GRPO implementation