-
Notifications
You must be signed in to change notification settings - Fork 261
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support local LLMs such as CodeLLaMa (llama-2 based) #26
base: main
Are you sure you want to change the base?
Conversation
Hi! Thanks a lot for this PR! Let me try to run codellama-2 locally if it's not too complicated, to see if I can test the code and make it work! |
You are welcome! It takes a bit of resources (I am running it on an A100 GPU!) . You could try with oobabooga in llama.cpp mode (read: on CPUs / mixed GPU+CPU) in case you don't have the resources. Note that sometimes codellama does not return proper JSON yet. That's something which probably could be addressed with things like langchain's output parsers. |
Hi! So, I've started playing with the code, to check if the UI works, etc. Seems to be good, but just trying to send a request with the default config value ( |
Let me check... it might be that you discovered a developer hack which was still in there. |
how to adapt to https://huggingface.co/Phind/Phind-CodeLlama-34B-v2 |
Tested. It works! Thanks! |
Hi!
I contributed a tiny set of patches so that gepetto can also access locally running LLMs such as CodeLLama (or see here).
Ping me pls. in case you want to test it, I can demo it if needed.
Have a good week!