| 
							 | 
						--- | 
					
					
						
						| 
							 | 
						language: | 
					
					
						
						| 
							 | 
						- en | 
					
					
						
						| 
							 | 
						license: apache-2.0 | 
					
					
						
						| 
							 | 
						base_model: | 
					
					
						
						| 
							 | 
						- mistralai/Mistral-Small-24B-Base-2501 | 
					
					
						
						| 
							 | 
						datasets: | 
					
					
						
						| 
							 | 
						- anthracite-org/kalo-opus-instruct-22k-no-refusal | 
					
					
						
						| 
							 | 
						model-index: | 
					
					
						
						| 
							 | 
						- name: Not-So-Small-Alpaca-24B | 
					
					
						
						| 
							 | 
						  results: | 
					
					
						
						| 
							 | 
						  - task: | 
					
					
						
						| 
							 | 
						      type: text-generation | 
					
					
						
						| 
							 | 
						      name: Text Generation | 
					
					
						
						| 
							 | 
						    dataset: | 
					
					
						
						| 
							 | 
						      name: IFEval (0-Shot) | 
					
					
						
						| 
							 | 
						      type: HuggingFaceH4/ifeval | 
					
					
						
						| 
							 | 
						      args: | 
					
					
						
						| 
							 | 
						        num_few_shot: 0 | 
					
					
						
						| 
							 | 
						    metrics: | 
					
					
						
						| 
							 | 
						    - type: inst_level_strict_acc and prompt_level_strict_acc | 
					
					
						
						| 
							 | 
						      value: 62.44 | 
					
					
						
						| 
							 | 
						      name: strict accuracy | 
					
					
						
						| 
							 | 
						    source: | 
					
					
						
						| 
							 | 
						      url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=SaisExperiments/Not-So-Small-Alpaca-24B | 
					
					
						
						| 
							 | 
						      name: Open LLM Leaderboard | 
					
					
						
						| 
							 | 
						  - task: | 
					
					
						
						| 
							 | 
						      type: text-generation | 
					
					
						
						| 
							 | 
						      name: Text Generation | 
					
					
						
						| 
							 | 
						    dataset: | 
					
					
						
						| 
							 | 
						      name: BBH (3-Shot) | 
					
					
						
						| 
							 | 
						      type: BBH | 
					
					
						
						| 
							 | 
						      args: | 
					
					
						
						| 
							 | 
						        num_few_shot: 3 | 
					
					
						
						| 
							 | 
						    metrics: | 
					
					
						
						| 
							 | 
						    - type: acc_norm | 
					
					
						
						| 
							 | 
						      value: 33.02 | 
					
					
						
						| 
							 | 
						      name: normalized accuracy | 
					
					
						
						| 
							 | 
						    source: | 
					
					
						
						| 
							 | 
						      url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=SaisExperiments/Not-So-Small-Alpaca-24B | 
					
					
						
						| 
							 | 
						      name: Open LLM Leaderboard | 
					
					
						
						| 
							 | 
						  - task: | 
					
					
						
						| 
							 | 
						      type: text-generation | 
					
					
						
						| 
							 | 
						      name: Text Generation | 
					
					
						
						| 
							 | 
						    dataset: | 
					
					
						
						| 
							 | 
						      name: MATH Lvl 5 (4-Shot) | 
					
					
						
						| 
							 | 
						      type: hendrycks/competition_math | 
					
					
						
						| 
							 | 
						      args: | 
					
					
						
						| 
							 | 
						        num_few_shot: 4 | 
					
					
						
						| 
							 | 
						    metrics: | 
					
					
						
						| 
							 | 
						    - type: exact_match | 
					
					
						
						| 
							 | 
						      value: 18.05 | 
					
					
						
						| 
							 | 
						      name: exact match | 
					
					
						
						| 
							 | 
						    source: | 
					
					
						
						| 
							 | 
						      url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=SaisExperiments/Not-So-Small-Alpaca-24B | 
					
					
						
						| 
							 | 
						      name: Open LLM Leaderboard | 
					
					
						
						| 
							 | 
						  - task: | 
					
					
						
						| 
							 | 
						      type: text-generation | 
					
					
						
						| 
							 | 
						      name: Text Generation | 
					
					
						
						| 
							 | 
						    dataset: | 
					
					
						
						| 
							 | 
						      name: GPQA (0-shot) | 
					
					
						
						| 
							 | 
						      type: Idavidrein/gpqa | 
					
					
						
						| 
							 | 
						      args: | 
					
					
						
						| 
							 | 
						        num_few_shot: 0 | 
					
					
						
						| 
							 | 
						    metrics: | 
					
					
						
						| 
							 | 
						    - type: acc_norm | 
					
					
						
						| 
							 | 
						      value: 14.54 | 
					
					
						
						| 
							 | 
						      name: acc_norm | 
					
					
						
						| 
							 | 
						    source: | 
					
					
						
						| 
							 | 
						      url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=SaisExperiments/Not-So-Small-Alpaca-24B | 
					
					
						
						| 
							 | 
						      name: Open LLM Leaderboard | 
					
					
						
						| 
							 | 
						  - task: | 
					
					
						
						| 
							 | 
						      type: text-generation | 
					
					
						
						| 
							 | 
						      name: Text Generation | 
					
					
						
						| 
							 | 
						    dataset: | 
					
					
						
						| 
							 | 
						      name: MuSR (0-shot) | 
					
					
						
						| 
							 | 
						      type: TAUR-Lab/MuSR | 
					
					
						
						| 
							 | 
						      args: | 
					
					
						
						| 
							 | 
						        num_few_shot: 0 | 
					
					
						
						| 
							 | 
						    metrics: | 
					
					
						
						| 
							 | 
						    - type: acc_norm | 
					
					
						
						| 
							 | 
						      value: 12.09 | 
					
					
						
						| 
							 | 
						      name: acc_norm | 
					
					
						
						| 
							 | 
						    source: | 
					
					
						
						| 
							 | 
						      url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=SaisExperiments/Not-So-Small-Alpaca-24B | 
					
					
						
						| 
							 | 
						      name: Open LLM Leaderboard | 
					
					
						
						| 
							 | 
						  - task: | 
					
					
						
						| 
							 | 
						      type: text-generation | 
					
					
						
						| 
							 | 
						      name: Text Generation | 
					
					
						
						| 
							 | 
						    dataset: | 
					
					
						
						| 
							 | 
						      name: MMLU-PRO (5-shot) | 
					
					
						
						| 
							 | 
						      type: TIGER-Lab/MMLU-Pro | 
					
					
						
						| 
							 | 
						      config: main | 
					
					
						
						| 
							 | 
						      split: test | 
					
					
						
						| 
							 | 
						      args: | 
					
					
						
						| 
							 | 
						        num_few_shot: 5 | 
					
					
						
						| 
							 | 
						    metrics: | 
					
					
						
						| 
							 | 
						    - type: acc | 
					
					
						
						| 
							 | 
						      value: 29.94 | 
					
					
						
						| 
							 | 
						      name: accuracy | 
					
					
						
						| 
							 | 
						    source: | 
					
					
						
						| 
							 | 
						      url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=SaisExperiments/Not-So-Small-Alpaca-24B | 
					
					
						
						| 
							 | 
						      name: Open LLM Leaderboard | 
					
					
						
						| 
							 | 
						--- | 
					
					
						
						| 
							 | 
						 | 
					
					
						
						| 
							 | 
						### This model uses *Alpaca* | 
					
					
						
						| 
							 | 
						
 | 
					
					
						
						| 
							 | 
						This is a lora finetune of [mistralai/Mistral-Small-24B-Base-2501](https://huggingface.co/mistralai/Mistral-Small-24B-Base-2501) using [anthracite-org/kalo-opus-instruct-22k-no-refusal](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-22k-no-refusal) for ~6M tokens | 
					
					
						
						| 
							 | 
						# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) | 
					
					
						
						| 
							 | 
						Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/SaisExperiments__Not-So-Small-Alpaca-24B-details) | 
					
					
						
						| 
							 | 
						
 | 
					
					
						
						| 
							 | 
						|      Metric       |Value| | 
					
					
						
						| 
							 | 
						|-------------------|----:| | 
					
					
						
						| 
							 | 
						|Avg.               |28.35| | 
					
					
						
						| 
							 | 
						|IFEval (0-Shot)    |62.44| | 
					
					
						
						| 
							 | 
						|BBH (3-Shot)       |33.02| | 
					
					
						
						| 
							 | 
						|MATH Lvl 5 (4-Shot)|18.05| | 
					
					
						
						| 
							 | 
						|GPQA (0-shot)      |14.54| | 
					
					
						
						| 
							 | 
						|MuSR (0-shot)      |12.09| | 
					
					
						
						| 
							 | 
						|MMLU-PRO (5-shot)  |29.94| | 
					
					
						
						| 
							 | 
						
 | 
					
					
						
						| 
							 | 
						
 |