Upload ./README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -12,7 +12,7 @@ Language: **English** | [**中文**](https://huggingface.co/Tencent-Hunyuan/Tens
|
|
| 12 |
|
| 13 |
We provide a TensorRT version of [HunyuanDiT](https://github.com/Tencent/HunyuanDiT) for inference acceleration
|
| 14 |
(faster than flash attention). One can convert the torch model to TensorRT model using the following steps based on
|
| 15 |
-
**TensorRT-
|
| 16 |
|
| 17 |
> ⚠️ Important Reminder (Suggestion for testing the TensorRT acceleration version):
|
| 18 |
> We recommend users to test the TensorRT version on NVIDIA GPUs with Compute Capability >= 8.0,(For example, RTX4090,
|
|
@@ -68,12 +68,22 @@ ln -s ${REMOTE_PATH} ./ckpts/t2i/model_trt/engine/model_onnx.plan
|
|
| 68 |
|
| 69 |
If you are using a different GPU, you can build the engine using the following command.
|
| 70 |
|
|
|
|
|
|
|
| 71 |
```shell
|
| 72 |
# Build the TensorRT engine. By default, it will read the `ckpts` folder in the current directory.
|
| 73 |
sh trt/build_engine.sh
|
| 74 |
```
|
| 75 |
|
| 76 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 77 |
|
| 78 |
### 4. Run the inference using the TensorRT model.
|
| 79 |
|
|
|
|
| 12 |
|
| 13 |
We provide a TensorRT version of [HunyuanDiT](https://github.com/Tencent/HunyuanDiT) for inference acceleration
|
| 14 |
(faster than flash attention). One can convert the torch model to TensorRT model using the following steps based on
|
| 15 |
+
**TensorRT-10.1.0.27** and **cuda (11.7 or 11.8)**.
|
| 16 |
|
| 17 |
> ⚠️ Important Reminder (Suggestion for testing the TensorRT acceleration version):
|
| 18 |
> We recommend users to test the TensorRT version on NVIDIA GPUs with Compute Capability >= 8.0,(For example, RTX4090,
|
|
|
|
| 68 |
|
| 69 |
If you are using a different GPU, you can build the engine using the following command.
|
| 70 |
|
| 71 |
+
##### Hunyuan-DiT v1.2
|
| 72 |
+
|
| 73 |
```shell
|
| 74 |
# Build the TensorRT engine. By default, it will read the `ckpts` folder in the current directory.
|
| 75 |
sh trt/build_engine.sh
|
| 76 |
```
|
| 77 |
|
| 78 |
+
##### Using Previous versions, Hunyuan-DiT <= v1.1
|
| 79 |
+
```shell
|
| 80 |
+
# v1.1
|
| 81 |
+
sh trt/build_engine.sh 1.1
|
| 82 |
+
# v1.0
|
| 83 |
+
sh trt/build_engine.sh 1.0
|
| 84 |
+
```
|
| 85 |
+
|
| 86 |
+
Finally, if you see the output like `&&&& PASSED TensorRT.trtexec [TensorRT v10100]`, the engine is built successfully.
|
| 87 |
|
| 88 |
### 4. Run the inference using the TensorRT model.
|
| 89 |
|