We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
加入调用Ollama模型时对keep_alive参数的设置选项
ollama默认在5分钟未使用后会将模型从显存中卸载,导致下次调用时会经历较长时间的加载过程,较为影响使用体验。
Ollama在api调用时提供了keep_alive选项可以设置保留时长或永不卸载选项,希望可以加入设置页面让用户自主设置。详见Ollama关于API调用的文档: https://github.com/ollama/ollama/blob/c02db93243353855b983db2a1562a02b57e66db1/docs/faq.md#how-do-i-keep-a-model-loaded-in-memory-or-make-it-unload-immediately
No response
The text was updated successfully, but these errors were encountered:
好的٩(•̤̀ᵕ•̤́๑)ᵒᵏᵎᵎᵎᵎ
Sorry, something went wrong.
最新版已支持 嘻嘻
!!!!!!!
大佬好快!!!!
大佬nb!!!
No branches or pull requests
Search before asking
I searched in the issues and found nothing similar.
feature
加入调用Ollama模型时对keep_alive参数的设置选项
描述 Motivation
ollama默认在5分钟未使用后会将模型从显存中卸载,导致下次调用时会经历较长时间的加载过程,较为影响使用体验。
Solution
Ollama在api调用时提供了keep_alive选项可以设置保留时长或永不卸载选项,希望可以加入设置页面让用户自主设置。详见Ollama关于API调用的文档:
https://github.com/ollama/ollama/blob/c02db93243353855b983db2a1562a02b57e66db1/docs/faq.md#how-do-i-keep-a-model-loaded-in-memory-or-make-it-unload-immediately
还有其他内容吗?Anything else?
No response
你是否愿意提交一份 PR?Are you willing to submit a PR?
The text was updated successfully, but these errors were encountered: