-
Notifications
You must be signed in to change notification settings - Fork 5.4k
Issues: ollama/ollama
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
ollama creat -f Modelfile doesn't process utf-8 encoding correctly
bug
Something isn't working
#5005
opened Jun 12, 2024 by
MGdesigner
please add support for rk3588 NPU
feature request
New feature or request
#5001
opened Jun 12, 2024 by
pine64noob
Regression on ollama docker images >=0.1.33-rocm: rocBLAS does not find secondary GPU for Tensile
amd
Issues relating to AMD GPUs and ROCm
bug
Something isn't working
#5000
opened Jun 12, 2024 by
icodeforyou-dot-net
Error "transferring model data " when creating a model
bug
Something isn't working
#4998
opened Jun 12, 2024 by
tigerkin89
Apple Silicon macs with 8GB or 16GB slow down when loading larger models
bug
Something isn't working
#4996
opened Jun 12, 2024 by
jmorganca
AI Models stop working after few user only messages.
bug
Something isn't working
#4993
opened Jun 12, 2024 by
TheUntitledGoose
error pulling llama2 manifest
bug
Something isn't working
#4992
opened Jun 12, 2024 by
adityapandit1798
First value different on CUDA/ROCM when setting Issues relating to AMD GPUs and ROCm
bug
Something isn't working
nvidia
Issues relating to Nvidia GPUs and CUDA
seed
amd
#4990
opened Jun 12, 2024 by
jmorganca
Failed to acquire semaphore" error="context canceled"
bug
Something isn't working
#4989
opened Jun 12, 2024 by
travisgu
CUDA error: out of memory - Phi-3 Mini 128k prompted with 20k+ tokens on 4GB GPU
bug
Something isn't working
memory
#4985
opened Jun 11, 2024 by
kozuch
Ollama not using GPU after OS Reboot
bug
Something isn't working
#4984
opened Jun 11, 2024 by
lukasmwerner
SIGSEGV during ollama serve cgo execution CUDA
amd
Issues relating to AMD GPUs and ROCm
bug
Something isn't working
#4982
opened Jun 11, 2024 by
7910f6ba7ee4
Error Pulling any model - "Error: pull model manifest: 200: stream error: stream ID 3; NO_ERROR; received from peer"
bug
Something isn't working
#4981
opened Jun 11, 2024 by
ziptron
qwen2-72b start to output gibberish at some point if i set num_ctx to 8192
bug
Something isn't working
#4977
opened Jun 11, 2024 by
Mikhael-Danilov
Is RTX 4070 and not RTX 4070ti supported - ambigous documentation
bug
Something isn't working
#4975
opened Jun 11, 2024 by
thinkrapido
OLLAMA_MODEL_DIR is not reflecting on MacOS
bug
Something isn't working
#4973
opened Jun 11, 2024 by
yusufaly
How to disallow the use of both gpu and cpu
feature request
New feature or request
#4971
opened Jun 11, 2024 by
xiaohanglei
Error pulling Quen2 models: unknown pre-tokenizer type: 'qwen2'
bug
Something isn't working
#4969
opened Jun 11, 2024 by
agilebean
API Silently Truncates Conversation
bug
Something isn't working
#4967
opened Jun 10, 2024 by
flu0r1ne
Previous Next
ProTip!
Type g p on any issue or pull request to go back to the pull request listing page.