chansung/merged_ds_coding
Viewer • Updated • 60.6k • 146 • 18
How to use chansung/coding_llamaduo_result3 with PEFT:
from peft import PeftModel
from transformers import AutoModelForCausalLM
base_model = AutoModelForCausalLM.from_pretrained("google/gemma-7b")
model = PeftModel.from_pretrained(base_model, "chansung/coding_llamaduo_result3")This model is a fine-tuned version of google/gemma-7b on the chansung/merged_ds_coding dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss |
|---|---|---|---|
| 0.987 | 1.0 | 82 | 1.2808 |
| 0.6859 | 2.0 | 164 | 1.1719 |
| 0.5836 | 3.0 | 246 | 1.1480 |
| 0.5178 | 4.0 | 328 | 1.1717 |
| 0.4668 | 5.0 | 410 | 1.2044 |
| 0.3955 | 6.0 | 492 | 1.3252 |
| 0.3233 | 7.0 | 574 | 1.4225 |
| 0.2669 | 8.0 | 656 | 1.6119 |
| 0.2591 | 9.0 | 738 | 1.7353 |
| 0.2367 | 10.0 | 820 | 1.7502 |
Base model
google/gemma-7b