We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
背景:在本地部署ModelScope-Agent-7B,机器为nvidia的A100,速度特别慢,chat一次平均耗时18秒 已经按照[https://modelscope.cn/models/iic/ModelScope-Agent-7B/summary的步骤安装了flash-attention==2.3.5、layer_norm、rotary-embedding-torch==0.5.3] 启动ModelScope-Agent-7B 还是报:Warning: import flash_attn rotary fail, please install FlashAttention rotary to get higher efficiency https://github.com/Dao-AILab/flash-attention/tree/main/csrc/rotary
The text was updated successfully, but these errors were encountered:
帮忙看看呢
Sorry, something went wrong.
您好,最近modelscope-agent更新了版本,建议使用最新版本,关于本地部署的问题,可以参考这个下面这个关于无外网环境部署的方案 https://github.com/modelscope/modelscope-agent/pull/307/files
suluyana
No branches or pull requests
背景:在本地部署ModelScope-Agent-7B,机器为nvidia的A100,速度特别慢,chat一次平均耗时18秒
已经按照[https://modelscope.cn/models/iic/ModelScope-Agent-7B/summary的步骤安装了flash-attention==2.3.5、layer_norm、rotary-embedding-torch==0.5.3]
启动ModelScope-Agent-7B
还是报:Warning: import flash_attn rotary fail, please install FlashAttention rotary to get higher efficiency https://github.com/Dao-AILab/flash-attention/tree/main/csrc/rotary
The text was updated successfully, but these errors were encountered: