How to run

./llama.cpp/build/bin/llama-server -hf yarikdevcom/Intern-S1-mini-GGUF  --n-gpu-layers 99 --temp 0.8 --top-p 0.8 --top-k 50 --port 8999 --host 0.0.0.0 -fa
Downloads last month
20
GGUF
Model size
8.2B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

6-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for yarikdevcom/Intern-S1-mini-GGUF

Quantized
(5)
this model