Update modeling_dots_vision.py

#19
by chenj123 - opened
  1. 添加 flash_atten fallback_to_earge , 解决强依赖flash_atten.
  2. 添加两个 attention_backend , 用于显存受限 或者 ascend 910b 的推理
redmoe-ai-v1 changed pull request status to merged

Sign up or log in to comment