Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

请问llm_type为llama时,模型权重文件必须为model/stable-vicuna-13B.ggml.q4_2.bin吗?权重文件是否可以为llama2?以及策略是否支持fp16? #534

Open
15229684931 opened this issue Apr 30, 2024 · 0 comments

Comments

@15229684931
Copy link

我在npu上进行测试时,迁移了wenda的代码,在适配llama时,config.yaml如下配置
llama: path: "/home/ch/stable-vicuna-13B-HF" strategy: "npu fp16"
image
但是在web端推理时,出现了回答不全的现象
231

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant