google/switch-base-256
Updated
•
81
•
4
This release included various MoE (Mixture of expert) models, based on the T5 architecture . The base models use from 8 to 256 experts.
Totally Free + Zero Barriers + No Login Required