oobabooga text-generation-webui
#11
by
Lyralei
- opened
Try the ggml-model-f16.gguf version, because gpt-oss-20b itself is 4-bit quantized, and quantizing it to 4-bit again will slightly degrade the quality.
Try the ggml-model-f16.gguf version, because gpt-oss-20b itself is 4-bit quantized, and quantizing it to 4-bit again will slightly degrade the quality.