Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What is the data format to LoRA-fine-tune Vicuna? #4

Open
DavidFarago opened this issue May 11, 2023 · 0 comments
Open

What is the data format to LoRA-fine-tune Vicuna? #4

DavidFarago opened this issue May 11, 2023 · 0 comments

Comments

@DavidFarago
Copy link

Since https://github.com/lm-sys/FastChat/ does not publish its data, but mentions it "enhanced the training scripts provided by Alpaca to better handle multi-round conversations and long sequences", I looked at ShareGPT Vicuna datasets on Huggingface, and they contain conversations.

Now I see in this repo, data/merge_sample.json is used as data_path for the script supervised_finetune.py, but it contains Aplaca-like instruction, input, output triples.

Can we use supervised_finetune.py to fine-tune on conversations, e.g. in the format as the ShareGPT Vicuna datasets on Huggingface? If so, have you tried such a fine-tuning? If not, do you know of some repo that offers Vicuna fine-tuning based on conversations? Do you think supervised_finetune.py can be adapted easily to allow fine-tuning based on conversations?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant