Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
base_model:
|
4 |
+
- openai/gpt-oss-20b
|
5 |
+
tags:
|
6 |
+
- mlx
|
7 |
+
- gpt
|
8 |
+
- openai
|
9 |
+
- chatGPT
|
10 |
+
---
|
11 |
+
|
12 |
+
# NexaAI/gpt-oss-20b-MLX-4bit
|
13 |
+
|
14 |
+
## Quickstart
|
15 |
+
|
16 |
+
Run them directly with [nexa-sdk](https://github.com/NexaAI/nexa-sdk) installed
|
17 |
+
In nexa-sdk CLI:
|
18 |
+
|
19 |
+
```bash
|
20 |
+
NexaAI/gpt-oss-20b-MLX-4bit
|
21 |
+
```
|
22 |
+
|
23 |
+
## Overview
|
24 |
+
This is a 4-bit quantized version of the OpenAI GPT OSS 20B model, optimized for Apple Silicon using the MLX framework. The model was successfully converted from the original `gpt_oss` architecture to MLX format using the development version of `mlx-lm` (v0.26.3).
|
25 |
+
|
26 |
+
>[!TIP] Tip: For best results with tool calling and reasoning, update your LMStudio to the latest version (0.3.22)
|
27 |
+
|
28 |
+
|
29 |
+
## Reference
|
30 |
+
**Original model card**: [InferenceIllusionist/gpt-oss-20b-MLX-4bit](https://huggingface.co/InferenceIllusionist/gpt-oss-20b-MLX-4bit)
|