You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
v1beta1Controllers no runtime found to support predictor with model type: {nvidia-nim-llm 0xc002c46540}
Based on my understanding of the docs, if I specify the same model format/version as supported in my runtime as I list in my predictor, KServe should select it for use. I was not expecting to get this error. If I manually specify the name of the runtime in my yaml it works fine, but I would like to just specify the modelFormat name/version.
Running this through Kubeflow v1.8.0 and on the latest KServe v1.12.1.
The text was updated successfully, but these errors were encountered:
Also the version here means for the model framework version like pytorch version or tensorrt version, I think what you specified is the nim serving time version which we have another field named “runtimeVersion”. It is not the reason why this fails though but just want to clarify.
/kind bug
I created a ClusterServingRuntime that looks like this:
And an InferenceService like this:
And I get this error:
Based on my understanding of the docs, if I specify the same model format/version as supported in my runtime as I list in my predictor, KServe should select it for use. I was not expecting to get this error. If I manually specify the name of the runtime in my yaml it works fine, but I would like to just specify the modelFormat name/version.
Running this through Kubeflow v1.8.0 and on the latest KServe v1.12.1.
The text was updated successfully, but these errors were encountered: