Performance Gaps in Multi-view Clustering under the Nested Matrix-Tensor Model Paper • 2402.10677 • Published Feb 16, 2024
Investigating Regularization of Self-Play Language Models Paper • 2404.04291 • Published Apr 4, 2024 • 1
Do Vision and Language Encoders Represent the World Similarly? Paper • 2401.05224 • Published Jan 10, 2024
Falcon-H1: A Family of Hybrid-Head Language Models Redefining Efficiency and Performance Paper • 2507.22448 • Published 26 days ago • 64
Falcon-H1: A Family of Hybrid-Head Language Models Redefining Efficiency and Performance Paper • 2507.22448 • Published 26 days ago • 64
Running 3.11k 3.11k The Ultra-Scale Playbook 🌌 The ultimate guide to training LLM on large GPU Clusters
view article Article Falcon-Edge: A series of powerful, universal, fine-tunable 1.58bit language models. By tiiuae and 9 others • May 15 • 35
Running on CPU Upgrade 13.5k 13.5k Open LLM Leaderboard 🏆 Track, rank and evaluate open LLMs and chatbots
view article Article 🐺🐦⬛ LLM Comparison/Test: DeepSeek-V3, QVQ-72B-Preview, Falcon3 10B, Llama 3.3 70B, Nemotron 70B in my updated MMLU-Pro CS benchmark By wolfram • Jan 2 • 41
Falcon3 Collection Falcon3 family of Open Foundation Models is a set of pretrained and instruct LLMs ranging from 1B to 10B parameters. • 40 items • Updated Jul 23 • 86