Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add OpenELM #3910

Open
3Samourai opened this issue Apr 25, 2024 · 7 comments
Open

Add OpenELM #3910

3Samourai opened this issue Apr 25, 2024 · 7 comments
Labels
model request Model requests

Comments

@3Samourai
Copy link

Apple released several open source LLMs that are designed to run on-device.
Huggingface Link

@3Samourai 3Samourai added the model request Model requests label Apr 25, 2024
@thinkverse
Copy link

thinkverse commented Apr 25, 2024

Not supported in llama.cpp yet, there's an issue for it, ggerganov/llama.cpp#6868, labeled as a good first issue, if someone with C++ and Python experience wants to tackle it. 👍

@mertbozkir
Copy link

Interesting, I haven't seen this issue, and was trying to upload this model. 🫣

@userforsource
Copy link

What is the requirement for llama.cpp.
As far as i see ollama does not depend on llama.cpp

1 similar comment
@userforsource
Copy link

What is the requirement for llama.cpp.
As far as i see ollama does not depend on llama.cpp

@thinkverse
Copy link

As far as i see ollama does not depend on llama.cpp

Ollama makes heavy use of llama.ccp, it's the backend Ollama uses. When you start Ollama, it starts a llama.cpp server. When you chat with an LLM using Ollama is forwards it to the llama.cpp server.

func NewLlamaServer(gpus gpu.GpuInfoList, model string, ggml *GGML, adapters, projectors []string, opts api.Options) (LlamaServer, error) {

ollama/server/routes.go

Lines 1315 to 1320 in 2bed629

if err := runner.llama.Completion(c.Request.Context(), llm.CompletionRequest{
Prompt: prompt,
Format: req.Format,
Images: images,
Options: opts,
}, fn); err != nil {

You can see the llama.cpp submodude under https://github.com/ollama/ollama/tree/main/llm.

@userforsource
Copy link

ggerganov/llama.cpp#6960

@thinkverse
Copy link

Update on the OpenELM support, a draft PR has been opened ggerganov/llama.cpp#6986, and @joshcarp is looking for anyone to help out. I'm certain it would be appreciated if anyone who has experience with C++, Python, or something related can help. 👍

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
model request Model requests
Projects
None yet
Development

No branches or pull requests

4 participants