File size: 489 Bytes
ee8e7fd |
1 2 3 4 5 6 7 8 9 10 11 |
# Quantized Dia 1.6B (INT8)
This is a dynamic int8 quantized version of [nari-labs/Dia-1.6B](https://huggingface.co/nari-labs/Dia-1.6B).
It uses dynamic quantization for lighter deployment and faster inference.
Original model: **float16**, ~6.4GB
Quantized model: **int8 dynamic**, ~6.4GB
Uploaded by [RobertAgee](https://github.com/RobertAgee) and [RobAgrees](https://huggingface.co/RobAgrees.
> Quantized automatically with PyTorch dynamic quantization in Google Colab. |