You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Endpoints for a service are not getting updated on scale down operation or pod deletes. This leaves a lot of incorrect values in the endpoints. The propagates to the public service as well.
% kubectl -n detection get endpoints my-app-00112-private
NAME ENDPOINTS AGE
my-app-00112-private 10.32.101.40:9091,10.32.101.41:9091,10.32.101.43:9091 + 5997 more... 136m
% kubectl -n detection get deploy my-app-00112-deployment
NAME READY UP-TO-DATE AVAILABLE AGE
my-app-00112-deployment 2/2 2 2 136m
In the SKS update process, it is the private service Endpoints that are feeding SKS. Is there any plan to read from EnpointSlices (stable since 1.21) and move away from the legacy Endpoints? From the docs:
The EndpointSlice API is the recommended replacement for Endpoints.
We've just been affected by this in our environment on knative 1.16 in Google Cloud - for reference for people experiencing this in GKE, although the current stable channel is 1.30.5, it is 1.30.6 and above that contains the fix.
(and can confirm that once the fix is in, the endpoints behave normally again)
What version of Knative?
Expected Behavior
endpoints should update properly
Actual Behavior
Endpoints for a service are not getting updated on scale down operation or pod deletes. This leaves a lot of incorrect values in the endpoints. The propagates to the public service as well.
I was able to get logs like this from SKS:
Steps to Reproduce the Problem
This happens with all our
ksvc
that scale up and then down or have pods removed (via delete / evict).The text was updated successfully, but these errors were encountered: