Falcon-H1: A Family of Hybrid-Head Language Models Redefining Efficiency and Performance Paper • 2507.22448 • Published 26 days ago • 64
Falcon-H1 Collection Falcon-H1 Family of Hybrid-Head Language Models (Transformer-SSM), including 0.5B, 1.5B, 1.5B-Deep, 3B, 7B, and 34B (pretrained & instruction-tuned). • 38 items • Updated 25 days ago • 52
view article Article Falcon-H1: A Family of Hybrid-Head Language Models Redefining Efficiency and Performance By tiiuae and 5 others • May 21 • 34
view article Article Falcon-Arabic: A Breakthrough in Arabic Language Models By tiiuae and 7 others • May 21 • 34
Falcon Edge series Collection A series of powerful, universal and fine-tunable small Language Models • 7 items • Updated Jul 23 • 22
Falcon3 Collection Falcon3 family of Open Foundation Models is a set of pretrained and instruct LLMs ranging from 1B to 10B parameters. • 40 items • Updated Jul 23 • 86