-
Notifications
You must be signed in to change notification settings - Fork 243
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
support mistral and llava_mistral in turbomind #1579
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Tested OK with pipeline.
May resolve conflicts. |
What is the merge plan? since it has conflicts with the other two VL PRs |
After other PRs merged |
Auto backend UT failed. |
Lines 294 to 296 in 7758151
should add |
多轮对话时,模版跟llava有一些不一样,不确定是否重要。 lmdeploy: https://github.com/haotian-liu/LLaVA/blob/main/llava/conversation.py#L350-L359 |
I found in the interactive inference mode, the model generate " " at the end of its answer. |
Thanks for your contribution and we appreciate it a lot. The following instructions would make your pull request more healthy and more easily receiving feedbacks. If you do not understand some items, don't worry, just make the pull request and seek help from maintainers.
Motivation
resolve #1573
Note that slide window attention is not supported yet.
Tests