Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

微调完成后使用llama_factory的vllm和qwen官方的vllm部署方式启动返回的不一样 #1241

Closed
2 tasks done
lxb0425 opened this issue May 8, 2024 · 2 comments
Closed
2 tasks done

Comments

@lxb0425
Copy link

lxb0425 commented May 8, 2024

是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?

  • 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions

该问题是否在FAQ中有解答? | Is there an existing answer for this in FAQ?

  • 我已经搜索过FAQ | I have searched FAQ

当前行为 | Current Behavior

大佬 我使用llama_factory 微调成功后
使用llama_factory 的vllm与使用qwen官方文档推荐的vllm方式部署
返回不一样
llama_factory vllm部署的返回都很正常 从没出过问题
千问官方vllm部署的 总是有些问题 回复的效果很差 几乎乱回答 如下图
image

大概什么原因啊

期望行为 | Expected Behavior

期望返回都很正常

复现方法 | Steps To Reproduce

No response

运行环境 | Environment

- OS:
- Python: 3.10
- Transformers:
- PyTorch:
- CUDA 12.2
- vllm 0.3.3

备注 | Anything else?

No response

@jklj077
Copy link
Contributor

jklj077 commented May 9, 2024

Hi, can you clarify on the difference between deployment of "vllm from llama_factory" and "vllm from Qwen's official documentation"?

Based on the shared screenshot, it appears that you are using a custom frontend. As vllm is not fully compatible with Qwen(1.0) models (unaware of the chat template and the stop token ids), the frontend has to at least pass stop_token_ids to the API created by vllm. Or, you could use fastchat+vllm as introduced in the README. If you are using Qwen1.5, plain vllm should work fine.

@jklj077
Copy link
Contributor

jklj077 commented May 22, 2024

As Qwen1.0 is no longer actively maintained, we kindly ask to you migrate to Qwen1.5 and direct your related question there. Thanks for you cooperation.

@jklj077 jklj077 closed this as not planned Won't fix, can't repro, duplicate, stale May 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants