Commit
·
ff0d44e
1
Parent(s):
bfc0c05
update readme regarding FA2
Browse files
README.md
CHANGED
|
@@ -105,7 +105,7 @@ from transformers import AutoProcessor
|
|
| 105 |
|
| 106 |
model_id = "microsoft/Phi-3-vision-128k-instruct"
|
| 107 |
|
| 108 |
-
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="cuda", trust_remote_code=True, torch_dtype="auto")
|
| 109 |
|
| 110 |
processor = AutoProcessor.from_pretrained(model_id, trust_remote_code=True)
|
| 111 |
|
|
@@ -217,24 +217,6 @@ Note that by default, the Phi-3-Vision-128K model uses flash attention, which re
|
|
| 217 |
* NVIDIA A6000
|
| 218 |
* NVIDIA H100
|
| 219 |
|
| 220 |
-
### Running on Windows or without flash attention
|
| 221 |
-
To enable the model on these enviroment here are steps that you may consider to follow:
|
| 222 |
-
|
| 223 |
-
Step 1: comment flash attention import code in modeling_phi3_v.py from line 52 to line 56.
|
| 224 |
-
```python
|
| 225 |
-
# if is_flash_attn_2_available():
|
| 226 |
-
# from flash_attn import flash_attn_func, flash_attn_varlen_func
|
| 227 |
-
# from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
|
| 228 |
-
|
| 229 |
-
# _flash_supports_window_size = "window_size" in list(inspect.signature(flash_attn_func).parameters)
|
| 230 |
-
```
|
| 231 |
-
|
| 232 |
-
Step 2: change _"_attn_implementation"_ from _"flash_attention_2"_ to _"eager"_ in config.json or disable flash attention when you create the model as below.
|
| 233 |
-
|
| 234 |
-
```python
|
| 235 |
-
model = AutoModelForCausalLM.from_pretrained('microsoft/Phi-3-vision-128k-instruct', device_map="cuda", trust_remote_code=True, torch_dtype="auto", _attn_implementation="eager")
|
| 236 |
-
```
|
| 237 |
-
|
| 238 |
## License
|
| 239 |
|
| 240 |
The model is licensed under the [MIT license](https://huggingface.co/microsoft/Phi-3-vision-128k-instruct/resolve/main/LICENSE).
|
|
|
|
| 105 |
|
| 106 |
model_id = "microsoft/Phi-3-vision-128k-instruct"
|
| 107 |
|
| 108 |
+
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="cuda", trust_remote_code=True, torch_dtype="auto", attn_implementation='eager') # use attn_implementation='flash_attention_2' to enable flash attention
|
| 109 |
|
| 110 |
processor = AutoProcessor.from_pretrained(model_id, trust_remote_code=True)
|
| 111 |
|
|
|
|
| 217 |
* NVIDIA A6000
|
| 218 |
* NVIDIA H100
|
| 219 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 220 |
## License
|
| 221 |
|
| 222 |
The model is licensed under the [MIT license](https://huggingface.co/microsoft/Phi-3-vision-128k-instruct/resolve/main/LICENSE).
|