| # Quantized Dia 1.6B (INT8) | |
| This is a dynamic int8 quantized version of [nari-labs/Dia-1.6B](https://huggingface.co/nari-labs/Dia-1.6B). | |
| It uses dynamic quantization for lighter deployment and faster inference. | |
| Original model: **float16**, ~6.4GB | |
| Quantized model: **int8 dynamic**, ~6.4GB | |
| Uploaded by [RobertAgee](https://github.com/RobertAgee) and [RobAgrees](https://huggingface.co/RobAgrees. | |
| > Quantized automatically with PyTorch dynamic quantization in Google Colab. |