Upload README.md with huggingface_hub
Browse files
README.md
ADDED
|
@@ -0,0 +1,174 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
tags:
|
| 4 |
+
- robotics
|
| 5 |
+
- vla
|
| 6 |
+
- knowledge-distillation
|
| 7 |
+
- model-compression
|
| 8 |
+
- edge-deployment
|
| 9 |
+
- action-chunking
|
| 10 |
+
- multi-teacher
|
| 11 |
+
datasets:
|
| 12 |
+
- lerobot/pusht
|
| 13 |
+
- lerobot/libero
|
| 14 |
+
language:
|
| 15 |
+
- en
|
| 16 |
+
library_name: forge
|
| 17 |
+
pipeline_tag: robotics
|
| 18 |
+
---
|
| 19 |
+
|
| 20 |
+
# FORGE-Nano: Compressed VLA for Real-Time Robot Control
|
| 21 |
+
|
| 22 |
+
<p align="center">
|
| 23 |
+
<strong>7B VLA teacher → <1B student → 14.1 fps on edge GPU</strong>
|
| 24 |
+
</p>
|
| 25 |
+
|
| 26 |
+
## What is FORGE?
|
| 27 |
+
|
| 28 |
+
**FORGE** (Fast Optimized Robot Generation Engine) is a model distillation pipeline that takes any 7B+ Vision-Language-Action (VLA) model and compresses it to **<2GB for real-time edge deployment** on NVIDIA Jetson and Apple Silicon.
|
| 29 |
+
|
| 30 |
+
Part of the **ANIMA** agentic robotics AI stack by [Robot Flow Labs](https://robotflowlabs.com).
|
| 31 |
+
|
| 32 |
+
## Architecture
|
| 33 |
+
|
| 34 |
+
```
|
| 35 |
+
Teacher (7B VLA)
|
| 36 |
+
|
|
| 37 |
+
v
|
| 38 |
+
[SigLIP-SO400M] ---> [Bridge Attention] ---> [Qwen2.5-0.5B + LoRA] ---> [Action Head]
|
| 39 |
+
(frozen) (64 queries, 4L) (rank=32/64) (diffusion/flow)
|
| 40 |
+
472.3M params 39.7M params ~494M params ~1.7M params
|
| 41 |
+
```
|
| 42 |
+
|
| 43 |
+
**Total: 967.9M params** (495.6M trainable, 472.3M frozen)
|
| 44 |
+
|
| 45 |
+
## Benchmark Results (4x NVIDIA L4 24GB)
|
| 46 |
+
|
| 47 |
+
### Student Variants
|
| 48 |
+
| Variant | Params | FP32 fps | FP16 fps | FP16 Speedup | Training Loss Reduction |
|
| 49 |
+
|---------|--------|----------|----------|--------------|------------------------|
|
| 50 |
+
| Nano (diffusion, LoRA=32) | 967.9M | 7.9 | **11.0** | 1.39x | 67.0% |
|
| 51 |
+
| Nano (diffusion, LoRA=64) | 972.3M | 7.9 | 10.8 | 1.37x | **76.9%** |
|
| 52 |
+
| Nano (flow, LoRA=32) | 967.9M | **8.2** | **12.6** | **1.54x** | 85.8% |
|
| 53 |
+
| Small (diffusion) | 2097.7M | 6.2 | 9.9 | -- | -- |
|
| 54 |
+
| Small (flow) | 2097.7M | 6.1 | **11.3** | -- | -- |
|
| 55 |
+
|
| 56 |
+
### Full Pipeline: Build -> Train -> Prune -> Deploy
|
| 57 |
+
| Configuration | Post-Prune Params | FP32 fps | FP16 fps | Loss Reduction |
|
| 58 |
+
|---------------|-------------------|----------|----------|----------------|
|
| 59 |
+
| Diffusion + p75 + INT4 | 830.8M | 10.0 | 12.0 | 41.4% |
|
| 60 |
+
| Flow + p50 + INT4 | **739.3M** | **14.1** | 7.8 | 76.3% |
|
| 61 |
+
| LoRA-64 + p90 + INT4 | 880.8M | 9.1 | 11.2 | **86.3%** |
|
| 62 |
+
| **Flow + LoRA-64 + p60** | **774.1M** | **12.7** | **14.1** | 75.7% |
|
| 63 |
+
| No prune + INT8 | 922.2M | 8.1 | 11.0 | 59.4% |
|
| 64 |
+
|
| 65 |
+
### Multi-GPU Scaling
|
| 66 |
+
| GPUs | FP32 b=16 | FP16 b=32 |
|
| 67 |
+
|------|-----------|-----------|
|
| 68 |
+
| 1 GPU | 9.3 fps | **33.6 fps** |
|
| 69 |
+
| 2 GPU | 13.5 fps | -- |
|
| 70 |
+
| 4 GPU | **13.6 fps** | 31.6 fps |
|
| 71 |
+
|
| 72 |
+
### Multi-Teacher Distillation
|
| 73 |
+
- **5 teachers** fit across 2 GPUs (22.7 GB total)
|
| 74 |
+
- Router learns optimal teacher weighting in <50 steps
|
| 75 |
+
- Best config: balanced (alpha_task=0.3) achieves **76.1% loss reduction**
|
| 76 |
+
- Supports: OpenVLA-7B, RDT2-FM, SmolVLA, BitVLA, Pi0
|
| 77 |
+
|
| 78 |
+
### Pruning Results
|
| 79 |
+
| Pruning Ratio | Layers | Params | FP32 fps | Speedup |
|
| 80 |
+
|---------------|--------|--------|----------|---------|
|
| 81 |
+
| No prune | 24 | 967.9M | 7.9 | 1.0x |
|
| 82 |
+
| 90% keep | 18 | 880.8M | 9.1 | 1.15x |
|
| 83 |
+
| 75% keep | 15 | 830.8M | 10.0 | 1.27x |
|
| 84 |
+
| 60% keep | 11 | 774.1M | 12.7 | **1.61x** |
|
| 85 |
+
| 50% keep | 9 | 739.3M | **14.1** | **1.78x** |
|
| 86 |
+
|
| 87 |
+
## Recommended Configurations
|
| 88 |
+
|
| 89 |
+
### Production (Edge Deployment)
|
| 90 |
+
```yaml
|
| 91 |
+
variant: nano
|
| 92 |
+
action_head: flow
|
| 93 |
+
lora_rank: 64
|
| 94 |
+
prune_ratio: 0.60
|
| 95 |
+
target_bits: 4
|
| 96 |
+
# Result: 774M params, FP16 14.1 fps, <600MB INT4
|
| 97 |
+
```
|
| 98 |
+
|
| 99 |
+
### Quality-First
|
| 100 |
+
```yaml
|
| 101 |
+
variant: nano
|
| 102 |
+
action_head: diffusion
|
| 103 |
+
lora_rank: 32
|
| 104 |
+
prune_ratio: 0.75
|
| 105 |
+
target_bits: 8
|
| 106 |
+
# Result: 830M params, 92.3% loss reduction
|
| 107 |
+
```
|
| 108 |
+
|
| 109 |
+
## Key Findings
|
| 110 |
+
|
| 111 |
+
1. **Flow matching head is 15% faster** than diffusion at FP16 inference (12.6 vs 11.0 fps)
|
| 112 |
+
2. **LoRA rank=64 trains 10% better** than rank=32 (76.9% vs 67.0% loss reduction) with negligible speed cost
|
| 113 |
+
3. **Aggressive pruning works**: 50% layer removal still produces a functional model at 14.1 fps
|
| 114 |
+
4. **FP16 autocast gives 1.4-1.5x speedup** for free — always use it in production
|
| 115 |
+
5. **Multi-teacher routing converges fast**: Router learns to weight teachers optimally in <50 steps
|
| 116 |
+
|
| 117 |
+
## Supported Teachers
|
| 118 |
+
|
| 119 |
+
| Teacher | Type | Params | Chunk Size |
|
| 120 |
+
|---------|------|--------|------------|
|
| 121 |
+
| OpenVLA-7B | Token-AR | 7.6B | H=1 |
|
| 122 |
+
| RDT2-FM | Diffusion | 1.2B | H=8 |
|
| 123 |
+
| SmolVLA | Parallel | 0.5B | H=1 |
|
| 124 |
+
| BitVLA | Quantized | 5.9B | H=1 |
|
| 125 |
+
| Pi0 | Flow | 3.0B | H=4 |
|
| 126 |
+
|
| 127 |
+
## Supported Robots
|
| 128 |
+
|
| 129 |
+
| Robot | DoF | Action Head | Horizon | Control Rate |
|
| 130 |
+
|-------|-----|-------------|---------|-------------|
|
| 131 |
+
| Franka Panda | 7 | Flow | H=8 | 20 Hz |
|
| 132 |
+
| ALOHA (bimanual) | 14 | Chunk | H=16 | 50 Hz |
|
| 133 |
+
| xArm | 6 | Flow | H=4 | 100 Hz |
|
| 134 |
+
| UR5e | 6 | Flow | H=4 | 125 Hz |
|
| 135 |
+
|
| 136 |
+
## Pipeline
|
| 137 |
+
|
| 138 |
+
```
|
| 139 |
+
Teacher Labels -> Knowledge Distillation -> Layer Pruning -> Quantization -> Edge Export
|
| 140 |
+
(HDF5) (LoRA + Bridge) (Chunk-aware) (INT4/INT8) (TRT/ONNX/MLX)
|
| 141 |
+
```
|
| 142 |
+
|
| 143 |
+
## Usage
|
| 144 |
+
|
| 145 |
+
```bash
|
| 146 |
+
pip install anima-forge
|
| 147 |
+
|
| 148 |
+
# Full pipeline
|
| 149 |
+
forge pipeline --device cuda --variant nano --steps 5000
|
| 150 |
+
|
| 151 |
+
# Auto-detect model dimensions
|
| 152 |
+
forge autosense --model-dir /path/to/models
|
| 153 |
+
|
| 154 |
+
# Benchmark
|
| 155 |
+
forge benchmark run --device cuda
|
| 156 |
+
|
| 157 |
+
# Deploy
|
| 158 |
+
forge serve --port 8000
|
| 159 |
+
```
|
| 160 |
+
|
| 161 |
+
## Citation
|
| 162 |
+
|
| 163 |
+
```bibtex
|
| 164 |
+
@software{forge2026,
|
| 165 |
+
title={FORGE: Fast Optimized Robot Generation Engine},
|
| 166 |
+
author={Robot Flow Labs},
|
| 167 |
+
year={2026},
|
| 168 |
+
url={https://github.com/RobotFlow-Labs/anima-forge-distillation-pipeline}
|
| 169 |
+
}
|
| 170 |
+
```
|
| 171 |
+
|
| 172 |
+
## License
|
| 173 |
+
|
| 174 |
+
Apache 2.0
|