# Quantized Dia 1.6B (INT8) | |
This is a dynamic int8 quantized version of [nari-labs/Dia-1.6B](https://huggingface.co/nari-labs/Dia-1.6B). | |
It uses dynamic quantization for lighter deployment and faster inference. | |
Original model: **float16**, ~6.4GB | |
Quantized model: **int8 dynamic**, ~6.4GB | |
Uploaded by [RobertAgee](https://github.com/RobertAgee) and [RobAgrees](https://huggingface.co/RobAgrees. | |
> Quantized automatically with PyTorch dynamic quantization in Google Colab. |