You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hey, very exciting project, I feel like I wanted to build something like this for a while now.
I've recently started playing with local models using ollama (hosted by a small computer on my network).
Wondering if it is possible to connect this to ollama? (maybe I'm asking if it is possible to use this without the router that requires nvidia hardware)
Also, beyond installation instructions and high level design, is there a wiki where it is explained how individual parts work?
Thanks!
The text was updated successfully, but these errors were encountered:
Hello, thanks for the kudos on the project. I just took a quick look at Ollama and it could be integrated in the same way the software integrates with other external chat APIs (like OpenAI), but it would have to be done at the python worker level and that would not solve the issue of the Nvidia requirement. I am currently brainstorming how to remove the Nvidia requirement (real dependency is on ExLlama for lora hot swapping) but have yet to come up with a solution. I will be creating more verbose documentation and hosting it somewhere in the near future to gives a detailed explanation of the induvial parts of the stack.
Hey, very exciting project, I feel like I wanted to build something like this for a while now.
I've recently started playing with local models using ollama (hosted by a small computer on my network).
Wondering if it is possible to connect this to ollama? (maybe I'm asking if it is possible to use this without the router that requires nvidia hardware)
Also, beyond installation instructions and high level design, is there a wiki where it is explained how individual parts work?
Thanks!
The text was updated successfully, but these errors were encountered: