-
Notifications
You must be signed in to change notification settings - Fork 8.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GRPC GOAWAY #11118
Comments
This issue is currently awaiting triage. If Ingress contributors determines this is a relevant issue, they will accept it by applying the The Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
/remove-kind bug
/triage needs-information |
What type of GCP load balancer are you using? Can you try port forwarding to the ingress service and see if the issue persists? |
Client -> TCP L4 -> NGINX Ingress Controller -> App With: Client -> TCP L4 -> App Connection holds open for 24 hrs |
Morning, tested with TCP L4 LB removed and port forwarded to ingress-controller k8s svc. We still observe the same behaviour and connections are issued a GOAWAY after 5-10 minutes.
@longwuyuan this isn't straightforward to achieve as the app contains IP that we'd need to strip from the image which will take considerable time. We'll start the process on this but will take time so was hoping to at least start the conversation on this. |
@rcng6514 , thanks, Would you know if there is a chart on artifacthub.io or a image in hub.docker.com that can be used for reproduce ? |
This is stale, but we won't close it automatically, just bare in mind the maintainers may be busy with other tasks and will reach your issue ASAP. If you have any question or request to prioritize this, please reach |
I'm observing GOAWAY after ~5-10 minutes, shorter periods dependent on NGINX load |
|
@rcng6514 thanks
|
This is stale, but we won't close it automatically, just bare in mind the maintainers may be busy with other tasks and will reach your issue ASAP. If you have any question or request to prioritize this, please reach |
Apologies for the delay, we've since upgraded the controller so we're now running v1.10.1.
This only seems to be on some of our higher utilised GKE clusters.
|
If you provide ;
then anyone can try to reproduce on a minikube or a kind cluster. In the ingress yaml posted above, the use of annotations like limit-connections and upstream-vhost, just throw UN-necessary complications for testing long-living gRPC stream, I would not use that. |
What happened:
We have an application with GRPC streams working on GKE using an Ingress Cluster.
We have a use case where we want to open a long lived grpc stream between my GRPC server(GKE) and Client should send data every second for infinite period of time.
To achieve this use case I have designed my code in a way, that I never call OnCompleted method from GRPC Server java implementation, so that my stream remains open for infinite period of time. When I call my grpc method from Client, data transfer starts successfully for some time and streams run fine. However after few minutes(infrequent intervals) connection is automatically terminated giving below error message:-
UNAVAILABLE: Connection closed after GOAWAY. HTTP/2 error code: NO_ERROR
Time period for this error is not fixed, however this occurs after around 5 minutes of success data transfer between the client and the server(GKE).
We have tried various properties and timeouts to increase the longevity for streams (attached below is annotations attempted) however, we haven't found anything concrete on it.
Below is the ingress configuration annotations we are using
What you expected to happen:
Either annotations are respected or there is a misunderstanding in how we can make the above requirement possible
Not sure, we've exhausted all avenues
NGINX Ingress controller version (exec into the pod and run nginx-ingress-controller --version.):
Kubernetes version (use
kubectl version
):v1.27.9-gke.1092000
Environment:
Cloud provider or hardware configuration: GKE
OS (e.g. from /etc/os-release): cos_containerd
Kernel (e.g.
uname -a
):How was the ingress-nginx-controller installed:
$ helm template --values values.yaml --namespace example --version 4.9.0 ingress-nginx ingress-nginx/ingress-nginx > manifests.yaml
Then reference manifests.yaml as a resource in Kustomization with env specific patches for naming/annotations + lables
kubectl describe ingressclasses
kubectl describe ...
of any custom configmap(s) created and in useHow to reproduce this issue:
Vanilla ingress-nginx install in k8s.
Simple GRPC app with long lived connection
This happens across multiple applications on the cluster and across multiple envs so its a specific instance of this. We've read and tried the following: https://kubernetes.github.io/ingress-nginx/examples/grpc/#notes-on-using-responserequest-streams
However GOAWAYs continue to occur. Removing ingress-nginx and routing via NodePort, application held open the connection for 24+ hrs
--->
Anything else we need to know:
The text was updated successfully, but these errors were encountered: