Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LLAMA 3 #679

Open
tiger2014 opened this issue Apr 20, 2024 · 10 comments
Open

LLAMA 3 #679

tiger2014 opened this issue Apr 20, 2024 · 10 comments

Comments

@tiger2014
Copy link

Hi Team,

LLAMA 3 can be used in this frame?

Thanks

@KirillianAmu
Copy link

it does function right now, but in my experience rather poorly, likely due to the weird prompt formatting that llama 3 has.

@adammikulis
Copy link

If llama3 is being overly verbose, add "<|eot_id|>" to the AntiPrompts. I've had good results with it so far and it seems to have more personality than Mistral Instruct v0.2.

@imxcstar
Copy link

If llama3 is being overly verbose, add "<|eot_id|>" to the AntiPrompts. I've had good results with it so far and it seems to have more personality than Mistral Instruct v0.2.

I added <|eot_id|>, but it seems to have no effect

@AsakusaRinne AsakusaRinne added this to the v0.12.0 milestone Apr 24, 2024
@martindevans
Copy link
Collaborator

I'm not sure if our antiprompt detection will properly handle special tokens like that. I know there's special case handling for EOS in some places. That could be a good improvement to the antiprompt processing for someone to make.

@AsakusaRinne
Copy link
Collaborator

Hi, #708 has added an example of LLaMA3 chat session. I'll appreciate it if you would like to try it and report problem to us if any. To run it, please pull the latest code of master branch and run the example project. :)

@adammikulis
Copy link

Will PR #6920 from llama.cpp resolve this issue?

@AsakusaRinne
Copy link
Collaborator

@adammikulis I'll update the native library binaries following ggerganov/llama.cpp#6920. Besides you could also try the current master branch, which has already provided an example of LLaMA3.

@dlyz
Copy link
Contributor

dlyz commented Apr 30, 2024

I think x == EOS should be replaced everywhere with a llama_token_is_eog(x) that should be true for all end-of-something tokens. Antiprompt for <|eot_id|> won't help cause special tokens are not rendered.

ggerganov/llama.cpp#6745 (comment)

@AsakusaRinne
Copy link
Collaborator

I think x == EOS should be replaced everywhere with a llama_token_is_eog(x) that should be true for all end-of-something tokens. Antiprompt for <|eot_id|> won't help cause special tokens are not rendered.

That's right, we'll add such things soon. For now you could use "�" as anti-prompt as a temporary resolution.

@martindevans
Copy link
Collaborator

I think x == EOS should be replaced everywhere with a llama_token_is_eog(x)

This has actually been done in PR #712 (along with updating the binaries).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

7 participants