You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# -- You may not want to expose the metrics port on production deployments.
# If you want to access it from outside your cluster,
# use `kubectl port-forward` or create a secure ingress
expose:
default: false
# -- The exposed port for this service
exposedPort: 9100
# -- The port protocol (TCP/UDP)
protocol: TCP
The default value for graceTimeout is 10 seconds according to https://doc.traefik.io/traefik/routing/entrypoints/#lifecycle , which means most people don't notice this bug. However, we needed to increase graceTimeout for long lived connections so for long periods of time traefik becomes completely unmonitored, and appears down (up == 0) to our prometheus.
What did you see instead?
I saw a bug
What is your environment & configuration?
Traefik helm chart + kube + prometheus + long lived connections with graceTimeout set to a long value (hours).
Additional Information
No response
The text was updated successfully, but these errors were encountered:
At first glance, it seems more a configuration enhancement or warning to display than a real bug, but let's dig it.
Would you please share values showing this issue you encountered ?
Welcome!
What version of the Traefik's Helm Chart are you using?
v32.0.0
What version of Traefik are you using?
default from the v32.0.0 helm chart
What did you do?
I noticed that traefik is marked
up == 0
whilekube_pod_status_ready{condition="true"} == 0
I debugged it to the helm chart not setting
requestAcceptGraceTimeout
properly for the metrics entrypoint/port (docs: https://doc.traefik.io/traefik/routing/entrypoints/#lifecycle). It should be set here by default:traefik-helm-chart/traefik/values.yaml
Lines 707 to 719 in 7a13fc8
The default value for
graceTimeout
is 10 seconds according to https://doc.traefik.io/traefik/routing/entrypoints/#lifecycle , which means most people don't notice this bug. However, we needed to increasegraceTimeout
for long lived connections so for long periods of time traefik becomes completely unmonitored, and appears down (up == 0
) to our prometheus.What did you see instead?
I saw a bug
What is your environment & configuration?
Traefik helm chart + kube + prometheus + long lived connections with
graceTimeout
set to a long value (hours).Additional Information
No response
The text was updated successfully, but these errors were encountered: