parinitarahi commited on
Commit
f192828
·
verified ·
1 Parent(s): e7066a8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -12
README.md CHANGED
@@ -42,19 +42,7 @@ curl -o https://raw.githubusercontent.com/microsoft/onnxruntime-genai/refs/heads
42
  python model-chat.py -m /path/to/cuda-int4-rtn-block-32/ -e cuda --chat_template "<|begin▁of▁sentence|><|User|>{input}<|Assistant|>"
43
  ```
44
 
45
- For DirectML:
46
 
47
- ```bash
48
- # Download the model directly using the Hugging Face CLI
49
- huggingface-cli download onnxruntime/DeepSeek-R1-Distill-ONNX --include 'deepseek-r1-distill-qwen-1.5B/directml/*' --local-dir .
50
-
51
- # Install the DirectML package of ONNX Runtime GenAI
52
- pip install onnxruntime-genai-directml
53
-
54
- # Please adjust the model directory (-m) accordingly
55
- curl -o https://raw.githubusercontent.com/microsoft/onnxruntime-genai/refs/heads/main/examples/python/model-chat.py
56
- python model-chat.py -m /path/to/directml-int4-rtn-block-32/ -e dml --chat_template "<|begin▁of▁sentence|><|User|>{input}<|Assistant|>"
57
- ```
58
 
59
  ## ONNX Models
60
  Here are some of the optimized configurations we have added:
 
42
  python model-chat.py -m /path/to/cuda-int4-rtn-block-32/ -e cuda --chat_template "<|begin▁of▁sentence|><|User|>{input}<|Assistant|>"
43
  ```
44
 
 
45
 
 
 
 
 
 
 
 
 
 
 
 
46
 
47
  ## ONNX Models
48
  Here are some of the optimized configurations we have added: