| 
							 | 
						--- | 
					
					
						
						| 
							 | 
						base_model: | 
					
					
						
						| 
							 | 
						- Equall/Saul-7B-Instruct-v1 | 
					
					
						
						| 
							 | 
						- mistralai/Mistral-7B-Instruct-v0.3 | 
					
					
						
						| 
							 | 
						tags: | 
					
					
						
						| 
							 | 
						- merge | 
					
					
						
						| 
							 | 
						- mergekit | 
					
					
						
						| 
							 | 
						- lazymergekit | 
					
					
						
						| 
							 | 
						- Equall/Saul-7B-Instruct-v1 | 
					
					
						
						| 
							 | 
						- mistralai/Mistral-7B-Instruct-v0.3 | 
					
					
						
						| 
							 | 
						--- | 
					
					
						
						| 
							 | 
						 | 
					
					
						
						| 
							 | 
						# MistralSaulInstruct | 
					
					
						
						| 
							 | 
						
 | 
					
					
						
						| 
							 | 
						MistralSaulInstruct is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): | 
					
					
						
						| 
							 | 
						* [Equall/Saul-7B-Instruct-v1](https://huggingface.co/Equall/Saul-7B-Instruct-v1) | 
					
					
						
						| 
							 | 
						* [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) | 
					
					
						
						| 
							 | 
						
 | 
					
					
						
						| 
							 | 
						## 🧩 Configuration | 
					
					
						
						| 
							 | 
						
 | 
					
					
						
						| 
							 | 
						```yaml | 
					
					
						
						| 
							 | 
						slices: | 
					
					
						
						| 
							 | 
						  - sources: | 
					
					
						
						| 
							 | 
						      - model: Equall/Saul-7B-Instruct-v1 | 
					
					
						
						| 
							 | 
						        layer_range: [0, 32] | 
					
					
						
						| 
							 | 
						      - model: mistralai/Mistral-7B-Instruct-v0.3 | 
					
					
						
						| 
							 | 
						        layer_range: [0, 32] | 
					
					
						
						| 
							 | 
						merge_method: slerp | 
					
					
						
						| 
							 | 
						base_model: mistralai/Mistral-7B-Instruct-v0.3 | 
					
					
						
						| 
							 | 
						parameters: | 
					
					
						
						| 
							 | 
						  t: | 
					
					
						
						| 
							 | 
						    - filter: self_attn | 
					
					
						
						| 
							 | 
						      value: [0, 0.5, 0.3, 0.7, 1] | 
					
					
						
						| 
							 | 
						    - filter: mlp | 
					
					
						
						| 
							 | 
						      value: [1, 0.5, 0.7, 0.3, 0] | 
					
					
						
						| 
							 | 
						    - value: 0.5 | 
					
					
						
						| 
							 | 
						dtype: bfloat16 | 
					
					
						
						| 
							 | 
						``` | 
					
					
						
						| 
							 | 
						
 | 
					
					
						
						| 
							 | 
						## 💻 Usage | 
					
					
						
						| 
							 | 
						
 | 
					
					
						
						| 
							 | 
						```python | 
					
					
						
						| 
							 | 
						!pip install -qU transformers accelerate | 
					
					
						
						| 
							 | 
						 | 
					
					
						
						| 
							 | 
						from transformers import AutoTokenizer | 
					
					
						
						| 
							 | 
						import transformers | 
					
					
						
						| 
							 | 
						import torch | 
					
					
						
						| 
							 | 
						 | 
					
					
						
						| 
							 | 
						model = "jonathanbb/MistralSaulInstruct" | 
					
					
						
						| 
							 | 
						messages = [{"role": "user", "content": "What is a large language model?"}] | 
					
					
						
						| 
							 | 
						 | 
					
					
						
						| 
							 | 
						tokenizer = AutoTokenizer.from_pretrained(model) | 
					
					
						
						| 
							 | 
						prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) | 
					
					
						
						| 
							 | 
						pipeline = transformers.pipeline( | 
					
					
						
						| 
							 | 
						    "text-generation", | 
					
					
						
						| 
							 | 
						    model=model, | 
					
					
						
						| 
							 | 
						    torch_dtype=torch.float16, | 
					
					
						
						| 
							 | 
						    device_map="auto", | 
					
					
						
						| 
							 | 
						) | 
					
					
						
						| 
							 | 
						 | 
					
					
						
						| 
							 | 
						outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) | 
					
					
						
						| 
							 | 
						print(outputs[0]["generated_text"]) | 
					
					
						
						| 
							 | 
						``` |