Issues: InternLM/lmdeploy
[Benchmark] benchmarks on different cuda architecture with mo...
#815
opened Dec 11, 2023 by
lvhan028
Open
6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] change h_input_length_buf_ before synchronization
#1584
opened May 11, 2024 by
mengmeexix
2 tasks done
[Feature] Support for LLaVA-NeXT Qwen1.5-110, Qwen1.5-72B, LLaMA3-8B
awaiting response
#1583
opened May 11, 2024 by
Iven2132
[Feature] 是否支持enc-dec类型模型中decoder的persistent batch
awaiting response
#1581
opened May 10, 2024 by
Oldpan
[Bug] 使用docker部署internlm/internlm-xcomposer-vl-7b和internlm/internlm-xcomposer2-vl-7b均报错
#1577
opened May 10, 2024 by
ye7love7
2 tasks done
[Bug] Error deploying HuggingFace model llava-v1.6-mistral-7b with lmdeploy: Unrecognized model type llava_mistral
awaiting response
#1573
opened May 10, 2024 by
enogalaxies
1 of 2 tasks
[Docs] Can we have an up-to-date version of the persistent batching gif?
#1564
opened May 9, 2024 by
youkaichao
[Bug] Qwen1.5-32B在做awq量化并convert的时候会报错
awaiting response
#1560
opened May 8, 2024 by
dingjingzhen
2 tasks
Why did the model not understand the previous conversations?
#1551
opened May 7, 2024 by
Iven2132
2 tasks done
[Bug] set logprobs = true and top_logprobs = 5 in restful server. The number of top logrobs is 4 which is unexpected.
#1548
opened May 6, 2024 by
zhulinJulia24
2 tasks
[Bug] llama3和Chinese-LLaMA-Alpaca-3 无论是pipeline方式,还是lmdeploy chat方式都无法生成结果(llama2都是好的)
#1538
opened May 1, 2024 by
zhanghui-china
2 tasks
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.