Luth-0.6B-Instruct / README.md
MaxLSB's picture
Update README.md
3469064 verified
|
raw
history blame
1.45 kB
---
library_name: transformers
license: apache-2.0
datasets:
- kurakurai/luth-sft
language:
- fr
- en
base_model:
- Qwen/Qwen3-0.6B
pipeline_tag: text-generation
---
![Kurakura AI Logo](media/logo_kurakura.png)
# Luth-0.6B-Instruct
**Luth-0.6B-Instruct** is a French fine-tuned version of [Qwen3-0.6B](https://huggingface.co/Qwen/Qwen3-0.6B), trained on the [Luth-SFT](https://huggingface.co/datasets/kurakurai/luth-sft) dataset. The model has drastically improved its French capabilities in instruction following, math, and general knowledge. Additionally, its English capabilities have remained stable and have even increased in some areas.
## Model Details
Luth was trained using full fine-tuning on the Luth-SFT dataset with [Axolotl](https://github.com/axolotl-ai-cloud/axolotl). The resulting model was then merged with the base Qwen3-0.6B model. This process successfully retained the model's English capabilities while improving its performance on nearly all selected benchmarks in both French and English.
## Benchmark Results
**French Evaluation:**
![French Evaluation](media/french_evaluation.png)
**English Evaluation:**
![English Evaluation](media/english_evaluation.png)
## Citation
```bibtex
@misc{luth2025kurakurai,
title = {Luth-0.6B},
author = {Kurakura AI Team},
year = {2025},
howpublished = {\url{https://huggingface.co/kurakurai/Luth-0.6B}},
note = {Qwen3-0.6B fine-tuned on French datasets}
}
```