Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Convert HF Llama Checkpoints to Neox Checkpoints #994

Open
sxthunder opened this issue Jul 10, 2023 · 1 comment
Open

Convert HF Llama Checkpoints to Neox Checkpoints #994

sxthunder opened this issue Jul 10, 2023 · 1 comment
Labels
feature request New feature or request

Comments

@sxthunder
Copy link

Hello, I am excited that gpt-neox now support llama model. However, the script in tools/convert_raw_llama_weights_to_neox.py only support origin llama weight. Considering the large number of users currently using Huggingface, would it be possible to provide a script for converting the Huggingface Llama model into Neox?

In my experiments, training speed and memory usage in gpt-neox is much better than other language model framework, even if training by Lora. So I want to use gpt-neox to train if it support the model.

@sxthunder sxthunder added the feature request New feature or request label Jul 10, 2023
@StellaAthena
Copy link
Member

@haileyschoelkopf You were the one who ported LLaMA2 right? Do you think it would be easy to adapt that script to porting LLaMA1?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants