Update README.md
Browse files
README.md
CHANGED
|
@@ -15,7 +15,7 @@ base_model:
|
|
| 15 |
|
| 16 |
# Qwen2.5-VL-72B-Instruct-Pointer-AWQ
|
| 17 |
|
| 18 |
-
|
| 19 |
|
| 20 |
# Qwen2.5-VL-72B-Instruct
|
| 21 |
<a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;">
|
|
|
|
| 15 |
|
| 16 |
# Qwen2.5-VL-72B-Instruct-Pointer-AWQ
|
| 17 |
|
| 18 |
+
Since the official `Qwen/Qwen2.5-VL-72B-Instruct-AWQ` doesn't work with tensor parallel on vllm yet, this model fixed the issue and supports `--tensor-parallel` with 2, 4 or 8 GPUs. Use `vllm==0.7.3`.
|
| 19 |
|
| 20 |
# Qwen2.5-VL-72B-Instruct
|
| 21 |
<a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;">
|