Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[QUESTION] Does Megatron-Core supports LLAMA models? #803

Open
noob-ctrl opened this issue May 3, 2024 · 5 comments
Open

[QUESTION] Does Megatron-Core supports LLAMA models? #803

noob-ctrl opened this issue May 3, 2024 · 5 comments

Comments

@noob-ctrl
Copy link

Does Megatron-Core supports LLAMA models?

@ethanhe42
Copy link
Member

yes

@noob-ctrl
Copy link
Author

@ethanhe42 When transformer-impl is local, it reports the following error:
AssertionError: (RMSNorm) is not supported in FusedLayerNorm when instantiating FusedLayerNorm when instantiating TransformerLayer
When transformer-impl is transformer_engine, the following code does not seem to define RMSNorm?
image
So do I need to make any changes when I want to use llama?

@ethanhe42
Copy link
Member

You need to use mcore models. local is deprecating

@noob-ctrl
Copy link
Author

@ethanhe42 When transformer-impl is set to transformer_engine, the following code does not seem to define RMSNorm?
image

@ethanhe42
Copy link
Member

It's handled by TEnorm

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants