cli-lora-tinyllama / README.md
Harish2002's picture
Update README.md
8315e48 verified
|
raw
history blame
1.11 kB
---
license: mit
tags:
- tinyllama
- lora
- cli
- fine-tuning
- qna
- transformers
- peft
library_name: transformers
datasets:
- custom
language: en
model_type: causal-lm
---
# 🔧 CLI LoRA-TinyLlama
A fine-tuned version of [TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0) on a custom dataset of command-line Q&A, using **LoRA** (Low-Rank Adaptation). Built for fast, accurate help on common CLI topics.
---
## 🧩 Base Model
- Model: `TinyLlama/TinyLlama-1.1B-Chat-v1.0`
- Fine-Tuning Method: [LoRA](https://arxiv.org/abs/2106.09685)
- Libraries Used: `transformers`, `peft`, `datasets`, `accelerate`
---
## 📚 Dataset
- Custom dataset with **150+ Q&A pairs** covering:
- `git`, `bash`, `grep`, `tar`, `venv`
- Raw file: `cli_questions.json`
- Tokenized version: `tokenized_dataset/`
---
## 🛠️ Training Configuration
```python
from peft import LoraConfig
base_model = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
lora_config = LoraConfig(
r=16,
lora_alpha=32,
lora_dropout=0.1,
bias="none",
task_type="CAUSAL_LM"
)