Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Performence on terminal is very different within llamachat #24

Open
Darkhell101 opened this issue Apr 22, 2023 · 7 comments
Open

Performence on terminal is very different within llamachat #24

Darkhell101 opened this issue Apr 22, 2023 · 7 comments
Assignees
Labels
bug Something isn't working
Milestone

Comments

@Darkhell101
Copy link

I find the performence when using llama.cpp in terminal is better in llamachat.
The answer speed in terminal (interactive mode) is clearly faster on the same cpu usage level.
It's a little strange. Maybe it can be solved.

@alexrozanski
Copy link
Owner

thanks @fsrmlgd -- one thing might be to try tweaking the number of CPU cores used for prediction in LlamaChat (in Settings > General) if this is different to llama.cpp

@Darkhell101
Copy link
Author

Thanks for your reply. The app is very helpful.
here is my test for two conditions.

llamachat.mp4
warp.mov

As you can see the cat on top bar, It seems ridiculous. I change nothing about default settings.

@Darkhell101
Copy link
Author

If you need, some models can be directly used on PC are listed here.
https://huggingface.co/Pi3141
https://huggingface.co/eachadea
So make one easy to find a suitable version.

@alexrozanski
Copy link
Owner

thanks @fsrmlgd, will take a look into this!

@alexrozanski alexrozanski self-assigned this Apr 23, 2023
@alexrozanski alexrozanski added the bug Something isn't working label Apr 23, 2023
@alexrozanski alexrozanski added this to the v1.3 milestone Apr 23, 2023
@spacecowgoesmoo
Copy link

spacecowgoesmoo commented May 1, 2023

I set up vicuna7b on an m1 Mac mini and am seeing the same thing. Text in llamachat generates very slowly just like in the video above. Mac was at 75% cpu usage.

(and this is fine if llamas are still too heavy for low end systems, but that doesn’t seem like the case here.)

@alexrozanski
Copy link
Owner

@spacecowgoesmoo thanks, still getting round to this!

@Freccia
Copy link

Freccia commented Aug 11, 2023

Hello, here to say I experience the same issue. On terminal it works smoothly, but the GUI makes it very slow.
I am currently running on a Macbook Pro M1 with 16Gb of RAM.

btw amazing app design !

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
Status: No status
Development

No branches or pull requests

4 participants