Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] Support for LLaVA-NeXT Qwen1.5-110, Qwen1.5-72B, LLaMA3-8B #1583

Open
Iven2132 opened this issue May 11, 2024 · 13 comments
Open

[Feature] Support for LLaVA-NeXT Qwen1.5-110, Qwen1.5-72B, LLaMA3-8B #1583

Iven2132 opened this issue May 11, 2024 · 13 comments

Comments

@Iven2132
Copy link

Motivation

It outperforms existing open-source models like Intern-VL-1.5

https://llava-vl.github.io/blog/2024-05-10-llava-next-stronger-llms/#:~:text=Live%20Demo-,Benchmark%20Results,-Results%20with%20LMMs

Related resources

https://llava-vl.github.io/blog/2024-05-10-llava-next-stronger-llms/

Additional context

No response

@White-Friday
Copy link

请问现在是否已经支持LLaVA-NeXT Qwen1.5-110, Qwen1.5-72B, LLaMA3-8B了吗

@lvhan028
Copy link
Collaborator

Yes. They are supported.

@lvhan028
Copy link
Collaborator

It probably has problem when the model is large.
#1563 explains the reason

@Iven2132
Copy link
Author

Yes. They are supported.

@lvhan028 So I can use the LLaVA-NeXT Qwen1.5-72B, LLaMA3-8B?

@lvhan028
Copy link
Collaborator

Sorry, my bad.
It probably needs to make some changes like PR #1579 does.
I didn't find the checkpoints. Could you share the huggingface repo_id?

@Iven2132
Copy link
Author

Sorry, my bad. It probably needs to make some changes like PR #1579 does. I didn't find the checkpoints. Could you share the huggingface repo_id?

lmms-lab/llama3-llava-next-8b: https://huggingface.co/lmms-lab/llama3-llava-next-8b

lmms-lab/llava-next-72b: https://huggingface.co/lmms-lab/llava-next-72b

lmms-lab/llava-next-110b: https://huggingface.co/lmms-lab/llava-next-110b

Please add system prompt feature too

@babla9
Copy link

babla9 commented May 13, 2024

Do you know what minimum GPU memory requirements would be to serve these VL models? Thanks!

@Iven2132
Copy link
Author

Do you know what minimum GPU memory requirements would be to serve these VL models? Thanks!

A100s are best if you want you can run these on Modal, they are giving $30 free credits

@babla9
Copy link

babla9 commented May 15, 2024

Do you know what minimum GPU memory requirements would be to serve these VL models? Thanks!

A100s are best if you want you can run these on Modal, they are giving $30 free credits

Thanks, for LLaVA, do you know what the minimum configuration would be? eg 1 x A100 40gb? Would V100s work (eg 2xV100)? Anyway I can serve quantized models using V100s via lmdeploy?

@Iven2132
Copy link
Author

Do you know what minimum GPU memory requirements would be to serve these VL models? Thanks!

A100s are best if you want you can run these on Modal, they are giving $30 free credits

Thanks, for LLaVA, do you know what the minimum configuration would be? eg 1 x A100 40gb? Would V100s work (eg 2xV100)? Anyway I can serve quantized models using V100s via lmdeploy?

I'm not sure but you can run the llava-next 8b model with one A100

@Iven2132
Copy link
Author

@lvhan028 Will it support Llava-next?

@lvhan028
Copy link
Collaborator

Regarding llava-next, models in https://huggingface.co/collections/liuhaotian/llava-16-65b9e40155f60fd046a5ccf2 have already been supported except https://huggingface.co/liuhaotian/llava-v1.6-mistral-7b

We will support the following in June

lmms-lab/llama3-llava-next-8b: https://huggingface.co/lmms-lab/llama3-llava-next-8b
lmms-lab/llava-next-72b: https://huggingface.co/lmms-lab/llava-next-72b
lmms-lab/llava-next-110b: https://huggingface.co/lmms-lab/llava-next-110b

Copy link

This issue is marked as stale because it has been marked as invalid or awaiting response for 7 days without any further response. It will be closed in 5 days if the stale label is not removed or if there is no further response.

@github-actions github-actions bot added the Stale label May 24, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

4 participants