We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Wondering if streaming output is supported? Or are there any results about the time to first token and time per output token? Thanks.
The text was updated successfully, but these errors were encountered:
Copying the code of def stream_chat(self, ....) to modeling_internvl_chat.py from modeling_internlm2.py https://huggingface.co/OpenGVLab/InternVL-Chat-V1-5/blob/main/modeling_internlm2.py and make very small changes , I implement it and verify that is usefull.
def stream_chat(self, ....)
modeling_internvl_chat.py
modeling_internlm2.py
Sorry, something went wrong.
No branches or pull requests
Wondering if streaming output is supported? Or are there any results about the time to first token and time per output token? Thanks.
The text was updated successfully, but these errors were encountered: