Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for Larger Models #3

Open
scblaze opened this issue Apr 14, 2023 · 2 comments
Open

Support for Larger Models #3

scblaze opened this issue Apr 14, 2023 · 2 comments

Comments

@scblaze
Copy link

scblaze commented Apr 14, 2023

It would be great if there was a way to use this with the 13B, 30B or 60B LLaMa model sizes.

@zetavg
Copy link
Owner

zetavg commented Apr 14, 2023

In theory, it will work by specifying a larger LLaMA base model via the --base_model flag, e.g. --base_model=decapoda-research/llama-13b-hf, then select a LoRA model that's trained on top of that base model (such as chansung/alpaca-lora-13b). However, I still need to test it. If you had the chance to try it first, your sharing of how it works would be appreciated! 🚀

BTW I think I'll be adding the ability to switch between base models without restarting the app and support non-llama models in the near future.

Update 2023/4/20: The ability of switching between base models has been added.

@monydochev
Copy link

I can confirm that it working with llama-13b-h, it use 93% from 80GB VRAM A100, Lora trained successfully and working

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants