New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Crash in hipDriverGetVersion on windows #4094
Comments
Ollama is a client server architecture. My suspicion is you're setting this flag in the client, not the server. On windows, typically you set this as a system environment variable. See https://github.com/ollama/ollama/blob/main/docs/faq.md#setting-environment-variables-on-windows That said, it shouldn't crash when running on the GPU. Can you share the server log for your crash scenario? https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#how-to-troubleshoot-issues |
Certainly, here is the translation of your issue into English:In fact, the server crash was caused by my old GPU (Radeon520). I noticed that when I did not disable it, as soon as I used the command:
When I manually disable it (Radeon520), ollama can successfully load the model and run the model using the CPU. The relevant log information is as follows:
So after reading the documentation, I thought by setting the environment variable |
You didn't mention what version you were running. On versions before 0.1.33 we didn't handle spaces and quotes on the OLLAMA_LLM_LIBRARY variable properly, so it's possible you may have included quotes and that could explain it not working. Please give 0.1.33 a try and if it still isn't respecting your OLLAMA_LLM_LIBRARY setting, share you server.log so we can see more details. It may be helpful to set OLLAMA_DEBUg=1 as well |
I have updated to the latest version, here is the relevant log information, |
Thanks for the server log @ggjk616 It looks like we're crashing while trying to call an AMD Driver API to check the version via
Can you share some more information about your system? Which version of Windows? Home/Pro? Is your AMD Driver up to date? Do other GPU apps work correctly on your GPU? |
After reading your reply, I checked my drivers and indeed the issue was caused by the drivers not being the latest version. After updating the drivers, I was able to load the model without any issues. Thank you very much for your help! However, I'm still a bit curious as to why setting the OLLAMA_LLM_LIBRARY environment variable didn't work. You can now close this issue, and once again, thank you for your assistance! |
The next release should have better parsing of quotes and spaces around our env vars. |
What is the issue?
Can you help me,In the documentation, I noticed the following statement: “You can set OLLAMA_LLM_LIBRARY to any of the available LLM libraries to bypass autodetection, so for example, if you have a CUDA card, but want to force the CPU LLM library with AVX2 vector support, use:
OLLAMA_LLM_LIBRARY="cpu_avx2" ollama serve”
But After setting OLLAMA_LLM_LIBRARY=“cpu_avx2”, the program still detects my GPU when loading the model, resulting in an error: Error: Post “https://127.0.0.1:11434/api/chat”: read tcp 127.0.0.1:56915->127.0.0.1:11434: wsarecv: An existing connection was forcibly closed by the remote host.
OS
Windows
GPU
AMD
CPU
Intel
Ollama version
No response
The text was updated successfully, but these errors were encountered: