-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Triton Server OpenVINO backend not working with Tensorflow saved models #7200
Comments
I found out that in the Triton image there are 2 versions of OpenVINO, and one of them is missing libraries from OpenVINO:
So this problem most likely affects also TF Lite, PaddlePaddle & Pytorch model formats. Culprit is most likely here: |
@tanmayv25 for vis. |
@atobiszei The openVINO backend in Triton does not support models saved in savedModel format. Read about Triton's OpenVINO backend here: https://github.com/triton-inference-server/openvino_backend?tab=readme-ov-file#openvino-backend You'd have to convert savedModel using model optimizer tool into OpenVINO IR model (.xml and .bin files). Then place these files into the model directory instead of TF savedmodel dir. |
@tanmayv25 When I removed ONNX backend from Triton image & tuned shape parameters in config it worked fine. |
Thanks for the correction. It seems the feature to load savedmodel has been added recently. This could also help us installing different OV between OV and ONNXRuntime backends. |
Description
Triton is unable to load models with Tensorflow saved model format with OpenVINO backend.
Triton Information
What version of Triton are you using?
23.10,23.11,23.12,24.03,24.04 don't work.
Are you using the Triton container or did you build it yourself?
Triton container
To Reproduce
Basically follow:
https://github.com/triton-inference-server/tutorials/tree/main/Quick_Deploy/TensorFlow but change backend to OpenVINO
Model config:
Command and logs:
Expected behavior
Load the model.
The text was updated successfully, but these errors were encountered: