update bar charts
Browse files
README.md
CHANGED
|
@@ -54,11 +54,12 @@ print((tokenizer.decode(outputs[0])))
|
|
| 54 |
Zamba2-1.2B-Instruct achieves leading instruction-following and multi-turn chat performance for a model of its size and matches strong models significantly larger. For instance, Zamba2-1.2B-Instruct outperforms Gemma2-2B-Instruct, a very strong model over 2x its size.
|
| 55 |
|
| 56 |
<center>
|
| 57 |
-
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/
|
| 58 |
</center>
|
| 59 |
|
| 60 |
-
|
| 61 |
-
|
|
|
|
| 62 |
| **Zamba2-1.2B-Instruct** | 1.2B | **59.53** | **41.45** |
|
| 63 |
| Gemma2-2B-Instruct | 2.7B | 51.69 | 42.20 |
|
| 64 |
| H2O-Danube-1.8B-Chat | 1.6B | 49.78 | 27.95 |
|
|
@@ -69,7 +70,7 @@ Zamba2-1.2B-Instruct achieves leading instruction-following and multi-turn chat
|
|
| 69 |
Moreover, due to its unique hybrid SSM architecture, Zamba2-1.2B-Instruct achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer-based models.
|
| 70 |
|
| 71 |
<center>
|
| 72 |
-
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/
|
| 73 |
</center>
|
| 74 |
|
| 75 |
|
|
|
|
| 54 |
Zamba2-1.2B-Instruct achieves leading instruction-following and multi-turn chat performance for a model of its size and matches strong models significantly larger. For instance, Zamba2-1.2B-Instruct outperforms Gemma2-2B-Instruct, a very strong model over 2x its size.
|
| 55 |
|
| 56 |
<center>
|
| 57 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/ceOUHVeJPhBgwTDCsR9Y6.png" width="900"/>
|
| 58 |
</center>
|
| 59 |
|
| 60 |
+
|
| 61 |
+
| Model | Size | Aggregate MT-Bench | IFEval |
|
| 62 |
+
|:-------------:|:----:|:-------------:|:----:|
|
| 63 |
| **Zamba2-1.2B-Instruct** | 1.2B | **59.53** | **41.45** |
|
| 64 |
| Gemma2-2B-Instruct | 2.7B | 51.69 | 42.20 |
|
| 65 |
| H2O-Danube-1.8B-Chat | 1.6B | 49.78 | 27.95 |
|
|
|
|
| 70 |
Moreover, due to its unique hybrid SSM architecture, Zamba2-1.2B-Instruct achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer-based models.
|
| 71 |
|
| 72 |
<center>
|
| 73 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/tQ-j1krA634EfTU1Lp3E7.png" width="700" alt="Zamba performance">
|
| 74 |
</center>
|
| 75 |
|
| 76 |
|