Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LoRA Support #1204

Open
4 tasks
Quentin-Anthony opened this issue Apr 23, 2024 · 1 comment
Open
4 tasks

LoRA Support #1204

Quentin-Anthony opened this issue Apr 23, 2024 · 1 comment
Assignees
Labels
feature request New feature or request

Comments

@Quentin-Anthony
Copy link
Member

Our finetuning focus has been a bit lacking lately. Let's change that, starting with LoRA!

  • Run a finetuning baseline with a Pythia model that saturates VRAM, measure its TFLOPs/VRAM usage in wandb and link it here. Choose a small finetuning dataset for this that requires little compute to converge without being trivial.
  • Add prototype LoRA support to gpt-neox
  • Compare to baseline in step 1 and ensure TFLOP/VRAM changes make sense
  • Compare to baseline in step 1 and ensure loss is maintained

As discussed in Discord @mkerin

@Quentin-Anthony Quentin-Anthony added the feature request New feature or request label Apr 23, 2024
@mkerin
Copy link
Contributor

mkerin commented Apr 23, 2024

Thanks for creating the feature request! As discussed - I'm happy to take this one on :) I expect to start work on it late this week / early next week.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants