zR
commited on
Commit
·
3526756
1
Parent(s):
bcf026a
fix readme
Browse files- README.md +5 -2
- README_en.md +5 -2
README.md
CHANGED
|
@@ -45,7 +45,10 @@ GLM-4-9B 是智谱 AI 推出的最新一代预训练模型 GLM-4 系列中的开
|
|
| 45 |
|
| 46 |
```python
|
| 47 |
import torch
|
| 48 |
-
from
|
|
|
|
|
|
|
|
|
|
| 49 |
|
| 50 |
device = "cuda"
|
| 51 |
|
|
@@ -62,7 +65,7 @@ inputs = tokenizer.apply_chat_template([{"role": "user", "content": query}],
|
|
| 62 |
|
| 63 |
inputs = inputs.to(device)
|
| 64 |
model = AutoModelForCausalLM.from_pretrained(
|
| 65 |
-
"
|
| 66 |
torch_dtype=torch.bfloat16,
|
| 67 |
low_cpu_mem_usage=True,
|
| 68 |
trust_remote_code=True
|
|
|
|
| 45 |
|
| 46 |
```python
|
| 47 |
import torch
|
| 48 |
+
from transformers import (
|
| 49 |
+
AutoModelForCausalLM,
|
| 50 |
+
AutoTokenizer,
|
| 51 |
+
)
|
| 52 |
|
| 53 |
device = "cuda"
|
| 54 |
|
|
|
|
| 65 |
|
| 66 |
inputs = inputs.to(device)
|
| 67 |
model = AutoModelForCausalLM.from_pretrained(
|
| 68 |
+
"THUDM/glm-4-9b-chat-1m",
|
| 69 |
torch_dtype=torch.bfloat16,
|
| 70 |
low_cpu_mem_usage=True,
|
| 71 |
trust_remote_code=True
|
README_en.md
CHANGED
|
@@ -38,13 +38,16 @@ Use the transformers backend for inference:
|
|
| 38 |
|
| 39 |
```python
|
| 40 |
import torch
|
| 41 |
-
from transformers import
|
|
|
|
|
|
|
|
|
|
| 42 |
|
| 43 |
device = "cuda"
|
| 44 |
|
| 45 |
tokenizer = AutoTokenizer.from_pretrained("THUDM/glm-4-9b-chat-1m",trust_remote_code=True)
|
| 46 |
|
| 47 |
-
query = "
|
| 48 |
|
| 49 |
inputs = tokenizer.apply_chat_template([{"role": "user", "content": query}],
|
| 50 |
add_generation_prompt=True,
|
|
|
|
| 38 |
|
| 39 |
```python
|
| 40 |
import torch
|
| 41 |
+
from transformers import (
|
| 42 |
+
AutoModelForCausalLM,
|
| 43 |
+
AutoTokenizer,
|
| 44 |
+
)
|
| 45 |
|
| 46 |
device = "cuda"
|
| 47 |
|
| 48 |
tokenizer = AutoTokenizer.from_pretrained("THUDM/glm-4-9b-chat-1m",trust_remote_code=True)
|
| 49 |
|
| 50 |
+
query = "你好"
|
| 51 |
|
| 52 |
inputs = tokenizer.apply_chat_template([{"role": "user", "content": query}],
|
| 53 |
add_generation_prompt=True,
|