RangiLyu commited on
Commit
293e42b
·
verified ·
1 Parent(s): a3ded57

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -200,13 +200,13 @@ The minimum hardware requirements for deploying Intern-S1 series models are:
200
 
201
  You can utilize one of the following LLM inference frameworks to create an OpenAI compatible server:
202
 
203
- #### [lmdeploy(>=0.9.2)](https://github.com/InternLM/lmdeploy)
204
 
205
  ```bash
206
  lmdeploy serve api_server internlm/Intern-S1 --reasoning-parser intern-s1 --tool-call-parser intern-s1 --tp 8
207
  ```
208
 
209
- #### [vllm](https://github.com/vllm-project/vllm)
210
 
211
  ```bash
212
  vllm serve internlm/Intern-S1 --tensor-parallel-size 8 --trust-remote-code
 
200
 
201
  You can utilize one of the following LLM inference frameworks to create an OpenAI compatible server:
202
 
203
+ #### [lmdeploy (>=0.9.2)](https://github.com/InternLM/lmdeploy)
204
 
205
  ```bash
206
  lmdeploy serve api_server internlm/Intern-S1 --reasoning-parser intern-s1 --tool-call-parser intern-s1 --tp 8
207
  ```
208
 
209
+ #### [vllm (>=0.10.1)](https://github.com/vllm-project/vllm)
210
 
211
  ```bash
212
  vllm serve internlm/Intern-S1 --tensor-parallel-size 8 --trust-remote-code