Update README.md
Browse files
README.md
CHANGED
@@ -21,9 +21,6 @@ pipeline_tag: text-generation
|
|
21 |
## Model Details
|
22 |
deepseek-ai/DeepSeek-R1-Distill-Llama-8B quantized to ONNX GenAI INT4 with Microsoft DirectML optimization.<br>
|
23 |
Output is reformatted that each sentence starts at new line to improve readability.<br>
|
24 |
-
Output will start with COTS/reasoning.<br>
|
25 |
-
|
26 |
-
In the tokenizer_config.json, the "unk_token" value is changed from null to ""
|
27 |
<pre>
|
28 |
...
|
29 |
vNewDecoded = tokenizer_stream.decode(new_token)
|
@@ -33,6 +30,8 @@ print(vNewDecoded, end='', flush=True)
|
|
33 |
vPreviousDecoded = vNewDecoded
|
34 |
...
|
35 |
</pre>
|
|
|
|
|
36 |
|
37 |
### Model Description
|
38 |
meta-llama/Meta-Llama-3.1-8B-Instruct quantized to ONNX GenAI INT4 with Microsoft DirectML optimization<br>
|
|
|
21 |
## Model Details
|
22 |
deepseek-ai/DeepSeek-R1-Distill-Llama-8B quantized to ONNX GenAI INT4 with Microsoft DirectML optimization.<br>
|
23 |
Output is reformatted that each sentence starts at new line to improve readability.<br>
|
|
|
|
|
|
|
24 |
<pre>
|
25 |
...
|
26 |
vNewDecoded = tokenizer_stream.decode(new_token)
|
|
|
30 |
vPreviousDecoded = vNewDecoded
|
31 |
...
|
32 |
</pre>
|
33 |
+
Output will start with COTS/reasoning.<br>
|
34 |
+
In the tokenizer_config.json, the "unk_token" value is changed from null to ""
|
35 |
|
36 |
### Model Description
|
37 |
meta-llama/Meta-Llama-3.1-8B-Instruct quantized to ONNX GenAI INT4 with Microsoft DirectML optimization<br>
|