File size: 489 Bytes
ee8e7fd
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
# Quantized Dia 1.6B (INT8)

This is a dynamic int8 quantized version of [nari-labs/Dia-1.6B](https://huggingface.co/nari-labs/Dia-1.6B).
It uses dynamic quantization for lighter deployment and faster inference.

Original model: **float16**, ~6.4GB  
Quantized model: **int8 dynamic**, ~6.4GB  

Uploaded by [RobertAgee](https://github.com/RobertAgee) and [RobAgrees](https://huggingface.co/RobAgrees.

> Quantized automatically with PyTorch dynamic quantization in Google Colab.