Text Generation
Transformers
Safetensors
mistral
Generated from Trainer
conversational
text-generation-inference
dvilasuero HF Staff commited on
Commit
4ac327d
·
verified ·
1 Parent(s): 9e0a2f2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -10
README.md CHANGED
@@ -33,17 +33,9 @@ This model is a fine-tuned version of [argilla/zephyr-7b-spin-iter2-v0](https://
33
  [argilla/10k_prompts_SPIN_iter3_zephyr_top](https://huggingface.co/datasets/argilla/10k_prompts_SPIN_iter3_zephyr_top) and the
34
  [argilla/10k_prompts_SPIN_iter2_zephyr_top](https://huggingface.co/datasets/argilla/10k_prompts_SPIN_iter2_zephyr_top) dataset.
35
 
36
- It achieves the following results on the evaluation set:
37
- - Loss: 0.1099
38
- - Rewards/real: -2.9181
39
- - Rewards/generated: -29.6970
40
- - Rewards/accuracies: 0.9271
41
- - Rewards/margins: 26.7789
42
- - Logps/generated: -702.4378
43
- - Logps/real: -278.1470
44
- - Logits/generated: -2.8177
45
- - Logits/real: -2.8051
46
 
 
47
 
48
 
49
  ## MT-Bench results
 
33
  [argilla/10k_prompts_SPIN_iter3_zephyr_top](https://huggingface.co/datasets/argilla/10k_prompts_SPIN_iter3_zephyr_top) and the
34
  [argilla/10k_prompts_SPIN_iter2_zephyr_top](https://huggingface.co/datasets/argilla/10k_prompts_SPIN_iter2_zephyr_top) dataset.
35
 
36
+ Check [this repo](https://github.com/argilla-io/distilabel-spin-dibt) for full reproducible code using the original SPIN implementation and distilabel.
 
 
 
 
 
 
 
 
 
37
 
38
+ If you want to contribute to high quality datasets like this, contribute to the [DIBT prompt collective initiative](https://huggingface.co/spaces/DIBT/prompt-collective-dashboard).
39
 
40
 
41
  ## MT-Bench results