-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Failed to get max tokens for LLM with name XXX (Ollama provider) #1342
Comments
Hi, Same error i am getting using the latest build. I am using the llama2. |
This is fixed by setting the value of GEN_AI_MAX_TOKENS |
@gargmukku07 which value did you set for llama2? |
just wanted to add my findings I was getting this error 05/28/2024 11:34:41 PM utils.py 328 : Failed to get max tokens for LLM with name azuregpt35turbo. Defaulting to 4096. turns out you need to set a variable for the GEN_AI_MAX_TOKENS this is due to this section of code in backend/danswer/llm/utils.py ) -> int:
|
Hi,
I'm facing the following issue when trying to chat with Ollama:
Then Danswer answer :)
My
.env
:Ollama is up and running, tested from inside
danswer-stack-api_server
container withcurl http://host.docker.internal:11434/api/tags
(obviously I had to install curl first).BTW, Danswer seems to replay the request once in case of error.
The text was updated successfully, but these errors were encountered: