Update README.md
Browse files
README.md
CHANGED
@@ -42,19 +42,7 @@ curl -o https://raw.githubusercontent.com/microsoft/onnxruntime-genai/refs/heads
|
|
42 |
python model-chat.py -m /path/to/cuda-int4-rtn-block-32/ -e cuda --chat_template "<|begin▁of▁sentence|><|User|>{input}<|Assistant|>"
|
43 |
```
|
44 |
|
45 |
-
For DirectML:
|
46 |
|
47 |
-
```bash
|
48 |
-
# Download the model directly using the Hugging Face CLI
|
49 |
-
huggingface-cli download onnxruntime/DeepSeek-R1-Distill-ONNX --include 'deepseek-r1-distill-qwen-1.5B/directml/*' --local-dir .
|
50 |
-
|
51 |
-
# Install the DirectML package of ONNX Runtime GenAI
|
52 |
-
pip install onnxruntime-genai-directml
|
53 |
-
|
54 |
-
# Please adjust the model directory (-m) accordingly
|
55 |
-
curl -o https://raw.githubusercontent.com/microsoft/onnxruntime-genai/refs/heads/main/examples/python/model-chat.py
|
56 |
-
python model-chat.py -m /path/to/directml-int4-rtn-block-32/ -e dml --chat_template "<|begin▁of▁sentence|><|User|>{input}<|Assistant|>"
|
57 |
-
```
|
58 |
|
59 |
## ONNX Models
|
60 |
Here are some of the optimized configurations we have added:
|
|
|
42 |
python model-chat.py -m /path/to/cuda-int4-rtn-block-32/ -e cuda --chat_template "<|begin▁of▁sentence|><|User|>{input}<|Assistant|>"
|
43 |
```
|
44 |
|
|
|
45 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
46 |
|
47 |
## ONNX Models
|
48 |
Here are some of the optimized configurations we have added:
|