Update README.md
Browse files
README.md
CHANGED
@@ -11,6 +11,7 @@ Please follow the license of the original model.
|
|
11 |
|
12 |
## How To Use
|
13 |
### INT4 Inference
|
|
|
14 |
```python
|
15 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
16 |
import transformers
|
@@ -61,6 +62,28 @@ for i, prompt in enumerate(prompts):
|
|
61 |
print(f"Prompt: {prompt}")
|
62 |
print(f"Generated: {decoded_outputs[i]}")
|
63 |
"""
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
64 |
Prompt: 9.11和9.8哪个数字大
|
65 |
Generated: 9.11 和 9.8 相比,**9.11 更大**。
|
66 |
- 9.11 可以理解为 9.11
|
@@ -100,7 +123,7 @@ Here are the key points about the company:
|
|
100 |
"""
|
101 |
|
102 |
### Generate the model
|
103 |
-
|
104 |
```python
|
105 |
import torch
|
106 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
11 |
|
12 |
## How To Use
|
13 |
### INT4 Inference
|
14 |
+
Due to kernel issues, potential overflow/underflow has been observed on CUDA. For accuracy, it is recommended to use CPU.
|
15 |
```python
|
16 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
17 |
import transformers
|
|
|
62 |
print(f"Prompt: {prompt}")
|
63 |
print(f"Generated: {decoded_outputs[i]}")
|
64 |
"""
|
65 |
+
GPU result:
|
66 |
+
Prompt: 9.11和9.8哪个数字大
|
67 |
+
Generated: 9.11 和 9.8 相比,**9.11 更大**。
|
68 |
+
|
69 |
+
- 9.11 可以理解为 9.11
|
70 |
+
# 1. 概述
|
71 |
+
|
72 |
+
## 1.1 什么是Spring
|
73 |
+
|
74 |
+
Spring是一个开源框架,它由Rod Johnson创建。它是为了解决企业应用开发的复杂性而创建的。Spring使用基本的JavaBean来完成以前只可能由EJB完成的事情。然而,Spring的用途不仅限于服务器端的开发。从简单性、可测试性和松耦合的角度而言,任何Java应用都可以从Spring中受益。
|
75 |
+
|
76 |
+
**目的:**解决企业应用开发的复杂性
|
77 |
+
|
78 |
+
**功能:**使用基本的JavaBean代替EJB,并提供了更多的企业应用功能
|
79 |
+
|
80 |
+
**范围:**任何Java应用
|
81 |
+
|
82 |
+
Spring是一个轻量级控制反转(IoC)和面向切面(AOP)的容器框架。
|
83 |
+
|
84 |
+
## 1.
|
85 |
+
--------------------------------------------------
|
86 |
+
CPU result:
|
87 |
Prompt: 9.11和9.8哪个数字大
|
88 |
Generated: 9.11 和 9.8 相比,**9.11 更大**。
|
89 |
- 9.11 可以理解为 9.11
|
|
|
123 |
"""
|
124 |
|
125 |
### Generate the model
|
126 |
+
Mian branch is required if the model is fp8 and the device supports fp8 https://github.com/intel/auto-round
|
127 |
```python
|
128 |
import torch
|
129 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|