-
Notifications
You must be signed in to change notification settings - Fork 33
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fine-tune other models #8
Comments
That should be possible in principle, but some of the code might be model specific now. Could you point me to the
I could look into that. |
I'm relatively new to AI development, but I've interested in a fine-tuned version of Mistral Orca. It's available here: Mistral 7B OpenOrca on Hugging Face. However, it seems like this model is in a Hugging Face format, which may not be directly compatible with the code, yes? You can find the original weights for the Mistral 7B model here: Original Weights for Mistral 7B. |
I tried to find a method for converting hf weights to pytorch, but nothing came up. |
Looking at https://huggingface.co/mistralai/Mistral-7B-v0.1/blob/main/pytorch_model.bin.index.json it should be possible to modify the loading to make it work. Need some updates to the loader code though. |
Do you have any suggestions for getting started? I want to put this into action, despite the fact that there will be a lot to learn :D |
@Gincioks - I'm not entirely sure about the best way, but probably here's how I'd do it:
|
Currently I facing with this error:
Update: |
could you share your code somewhere? Maybe a branch in your forked repo? |
Yes, yes, I will share the code, I made too many changes so I will start new repo. Also I was able get generation working perfectly. Now will do the same with finetuning. |
yeah, i think doing that in the forked version might be good option. thank you for looking into this! |
Hey, this is a new repository: https://github.com/Gincioks/PicoTuner. I intend to utilize this as a package in another project, so I created a small cli for easier use. |
Hello,
Can we apply this method to fine-tune models other than llamas and codellama, such as mistral 7b?
Many thanks in advance!
The text was updated successfully, but these errors were encountered: