Status Update
Comments
kk...@google.com <kk...@google.com>
ma...@google.com <ma...@google.com> #2
Hello,
Thanks for reaching out to us!
The Product Engineering Team has been made aware of your feature request, and will address it in due course. Though we can't provide an ETA on feature requests nor guarantee their implementation, rest assured that your feedback is always taken very seriously, as it allows us to improve our products. Thank you for your trust and continued support to improve Google Cloud Platform products.
In case you want to report a new issue, please do not hesitate to create a new
Thanks & Regards,
Manish Bavireddy.
Google Cloud Support
ma...@google.com <ma...@google.com>
jo...@gmail.com <jo...@gmail.com> #3
We are stuck with the same problem.
We have an open case with Google Support about this and, among others, they have redirected us here.
We have tried the following workaround:
START THE WORKARROUND
As per temporary fix the recommended way to match on port would be at the Gateway level.In the following example, the Route will only match incoming traffic on port 80 because it's attaching to the "http" Gateway listener.
apiVersion:
kind: Gateway
metadata:
name: my-gateway
spec:
gatewayClassName: gke-l7-global-external-managed
listeners:
- name: http
protocol: HTTP
port: 80
- name: https
protocol: HTTPS
port: 443
---
apiVersion:
kind: HTTPRoute
metadata:
name: http-store-80
spec:
parentRefs:
- name: my-gateway
sectionName: http
hostnames:
- "
rules:
backendRefs:
- name: my-service
port: 8080
END THE WORKARROUND
But the workarround has not worked, the problem persists.
We post this comment to see what the status of the Issue is and to see if there are other workarrounds available.
Thanks & Regards,
Joan Cholvi.
Mercadona
ba...@google.com <ba...@google.com>
ba...@google.com <ba...@google.com> #4
Hi,
Thanks for your response.
The information has been shared with the Product Team and further updates will be provided in this thread.
According to the
Please note that the Issue Tracker is primarily meant for reporting bugs and requesting new features. If you have any additional issues or concerns, please don’t hesitate to create a new thread on the
Thanks
Description
Problem you have encountered:
Updating the node-pool labels triggers node recreation immediately, even if the cluster is out of the maintenance window. It can be confusing, as some customer configuration changes respect maintenance window, but this one doesn't.
What you expected to happen:
This should be documented in official documentation.https://cloud.google.com/kubernetes-engine/docs/how-to/creating-managing-labels#update-node-pool-labels
Steps to reproduce:
Changing labels on any node-pool will trigger its recreation.
Other information (workarounds you have tried, documentation consulted, etc):
N/A