New updates: Correct system prompt, Tool calling, more fixes & llama.cpp!
pinned🚀
❤️
3
#7 opened about 1 month ago
by
shimmyshimmer

Quality compare in IQ4_NL (582Gb RAM) with Q5_K_XLARGE (735Gb RAM) on $150 ancient Xeon PC from 2014
1
#17 opened 10 days ago
by
krustik
Update README.md
#16 opened 25 days ago
by
sudo-xda
Amazing quality in such low Q4 on 2014 ANCIENT Xeon CPU with just shy 582Gb RAM
❤️
2
3
#15 opened about 1 month ago
by
krustik
Really appreciate the work you put into this.🤍
❤️
🔥
3
1
#14 opened about 1 month ago
by
deep-div

Slow Token Generation on A100
2
#13 opened about 1 month ago
by
kingabzpro

144gb vram and 256gb ram
1
#12 opened about 1 month ago
by
fuutott

The correct eos_token_id value for Kimi-K2-Instruct
2
#11 opened about 1 month ago
by
anikifoss
Update the instructions on requirements
2
#10 opened about 1 month ago
by
segmond
Model link at the bottom is broken
3
#9 opened about 1 month ago
by
Ray9821
Good llama.cpp -ot offloading parameter for 24 GB / 32 GB cards?
1
#5 opened about 1 month ago
by
qaraleza
Q5_K_M vs Q5_K_L vs Q5_K_XL
1
#4 opened about 1 month ago
by
ChuckMcSneed

Trouble running Q5_K_M With Llama.cpp
6
#3 opened about 1 month ago
by
simusid