Update README.md
Browse files
README.md
CHANGED
@@ -26,10 +26,12 @@ torch 2.7
|
|
26 |
|
27 |
https://github.com/nunchaku-tech/deepcompressor/blob/main/examples/diffusion/README.md
|
28 |
|
29 |
-
`(deepcompressor) python -m deepcompressor.app.diffusion.ptq examples/diffusion/configs/model/flux.1-kontext-dev.yaml examples/diffusion/configs/svdquant/nvfp4.yaml`
|
30 |
-
|
31 |
Model Path: https://github.com/nunchaku-tech/deepcompressor/issues/70#issuecomment-2788155233
|
32 |
|
|
|
|
|
|
|
|
|
33 |
# Dependencies
|
34 |
https://github.com/Dao-AILab/flash-attention
|
35 |
|
|
|
26 |
|
27 |
https://github.com/nunchaku-tech/deepcompressor/blob/main/examples/diffusion/README.md
|
28 |
|
|
|
|
|
29 |
Model Path: https://github.com/nunchaku-tech/deepcompressor/issues/70#issuecomment-2788155233
|
30 |
|
31 |
+
Save model: `--save-model true` or `--save-model /PATH/TO/CHECKPOINT/DIR`
|
32 |
+
|
33 |
+
Example: `python -m deepcompressor.app.diffusion.ptq examples/diffusion/configs/model/flux.1-kontext-dev.yaml examples/diffusion/configs/svdquant/nvfp4.yaml`
|
34 |
+
|
35 |
# Dependencies
|
36 |
https://github.com/Dao-AILab/flash-attention
|
37 |
|