Replies: 2 comments
-
same boat here on rke2
|
Beta Was this translation helpful? Give feedback.
-
How did you configure your cluster to use WIth regards to failures you're seeing when the coredns pod is not on the same node as the test pod, this is usually caused by something dropping the vxlan packets that traffic between pods and services that are on different nodes. Confirm that the vxlan port is open between all the nodes in your cluster, and that you're not affected by any common known issues: |
Beta Was this translation helpful? Give feedback.
-
We're currently testing several RKE2 clusters on top of CentOS 8 Stream, each with 3 manager nodes and 3 worker nodes. The clusters work well after applying the required CentOS 8 fixes, except for CoreDNS. However, whenever a pod needs to look up another pod's cluster IP address in CoreDNS, CoreDNS responds with host not found (NXDOMAIN).
We've gone through https://kubernetes.io/docs/tasks/administer-cluster/dns-debugging-resolution/ to find the problem, but logs show DNS requests coming in, and apparently DNS responses arriving at the requesting pods (i.e. the applications that did the DNS lookup fail instantly, and not after some time-out).
What can we do to resolve this issue?
Some additional information:
So CoreDNS responds correctly to everything except hostnames that exist within the kubernetes cluster ("No answer").
Doing three nslookups like above on
git.rancher.io
,kubernetes
, andkubernetes.default
results in the following wireshark output:UPDATE: this problem stopped occurring as soon as I restarted
dnsutils
, and it got scheduled on a different node, it started working... However, on another cluster, wherednsutils
was already scheduled on a different node, the problem remained. It appears that the problem occurs only when the DNS client is scheduled on a specific node, but that specific node could be any node...Beta Was this translation helpful? Give feedback.
All reactions