You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Check for Resource Limits and Quotas:
Ensure that your Kubernetes cluster has sufficient resources (CPU, memory) available and that no resource quotas are being exceeded. Preemptible VMs can also be a source of instability if they are being reclaimed by the cloud provider. Review Logs:
Check the logs of the InferenceService and the associated pods (kserve-container and aioli-logger) for any errors or warnings. This can provide insights into what might be causing the transition status to remain in progress.
/kind bug
What steps did you take and what happened:
Deployed inferenceservice iris-classifier-deployment:
After running it for some time, it could get into a transition "InProgress" status and could not get out of it.
At the same time, the deployed inference service is functional:
What did you expect to happen:
Expect the model transition status UpToDate:
What's the InferenceService yaml:
[To help us debug please run
kubectl get isvc $name -n $namespace -oyaml
and paste the output]Anything else you would like to add:
[Miscellaneous information that will assist in solving the issue.]
Could be related to using preemptible VMs to run the Kubernetes cluster. But we had also seen it happen creating new pod for canary rollout.
Environment:
kubectl version
):/etc/os-release
):The text was updated successfully, but these errors were encountered: