GGUF uploaded now + Chat template Fixes!
pinned🤗
🔥
4
27
#2 opened 19 days ago
by
shimmyshimmer

[solved] Setup high reasoning mode
1
#27 opened 8 days ago
by
Maria99934

Problems with FP32 model
2
#25 opened 8 days ago
by
YardWeasel
Feature Request: Disable reasoning
👀
1
3
#22 opened 12 days ago
by
SomAnon
Speed differences for different quants
👍
1
2
#21 opened 12 days ago
by
leonardlin

New Chat Template Fixes as of Aug 8, 2025:
🧠
🤗
1
#19 opened 16 days ago
by
shimmyshimmer

ollama load error
➕
👍
7
10
#17 opened 17 days ago
by
kwangtek
Failed to use with vLLM
➕
3
#16 opened 17 days ago
by
chengorange1
failed to read tensor info
4
#15 opened 17 days ago
by
valid-name1

Error installing model
2
#13 opened 18 days ago
by
nototon
Absurd sizes.
3
#12 opened 18 days ago
by
ZeroWw
Giving me error with llama-cpp-python
👍
1
1
#11 opened 18 days ago
by
divyanshu-k
Is the BF16 gguf any different from the F16 one? (speed/accuracy)
6
#10 opened 18 days ago
by
CHNtentes
Tool calling broken
3
#5 opened 19 days ago
by
AekDevDev

Wow, amazing response time
🤗
➕
9
7
#1 opened 19 days ago
by
AlexPradas