Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
免费去水印
Log In
Sign Up
Building on HF
116.7
TFLOPS
8
5
8
Ed Addario
PRO
eaddario
Follow
Tibbnak's profile picture
ngxson's profile picture
bartowski's profile picture
101 followers
·
31 following
EAddario
AI & ML interests
None yet
Recent Activity
posted
an
update
about 13 hours ago
Experimental global target bits‑per‑weight quantization of ServiceNow-AI/Apriel-1.6-15b-Thinker and zai-org/GLM-4.6V-Flash Unlike standard llama.cpp quantizations that rely on fixed type heuristics (e.g., Q4_K_M), the Target BPW approach optimizes per-tensor precision where it matters the most, and produces high quality models that meet a precise global file size target. Key Advantages: - VRAM Maximization: Can generate high quality models sized exactly to fit hardware constraints (e.g., fitting the model into exactly 24GB VRAM). - Data-Driven Precision: Quantization mix is determined by actual weight error sensitivity rather than hardcoded rules, often yielding better PPL/KLD size trade-offs. Full benchmarks (PPL, KLD, ARC, MMLU, etc.) and methodology in the models' cards https://huggingface.co/eaddario/Apriel-1.6-15b-Thinker-GGUF https://huggingface.co/eaddario/GLM-4.6V-Flash-GGUF
updated
a model
about 14 hours ago
eaddario/GLM-4.6V-Flash-GGUF
published
a model
about 18 hours ago
eaddario/GLM-4.6V-Flash-GGUF
View all activity
Organizations
eaddario
's datasets
1
Sort: Recently updated
eaddario/imatrix-calibration
Updated
May 24
•
13.2k
•
26
×
🎉 Free Image Generator Now Available!
Totally Free + Zero Barriers + No Login Required
Visit Now