-
Notifications
You must be signed in to change notification settings - Fork 603
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Container is not getting additional interface on k8s node reboot #1387
Comments
Have you checked the /etc/cni/net.d dir to see if cilium deleted the multus config? |
Many thanks for the tip! I found that Multus automatically generates 00-multus.conf in /etc/cni/net.d dir and automatically deletes it on shutdown. |
Well, reboot was due to OOM)) So what was actually happening in our environment:
Same thing will happen if you just delete multus pod and pod with additional interface at the same time. |
I have also seen pods come up with their multus interfaces missing, also around reboots. I use calico. In my case the pods look like they are running normally (they bind 0.0.0.0, so don't notice anything is wrong). It feels like it only started happening in the 0.4 series, but I couldn't say when. From your description, this is a simple race that could happen to anyone at any time not just due to OOM. And that tracks with my older nodes being more of a problem. If upstream don't like your patch / don't have resources to help, can we get calico/cilium to write its cni to a different folder ( |
@Jc2k unfortunately, I didn't find the confDir parameter in the current multus master. It seems that the docs are a bit outdated. I can only see cniConfigDir, but it looks like we can't split multus config and any other CNI configs:
|
Hi! We're using Multus with Cilium to supply some of our apps with additional interfaces.
Example of CR:
The problem is that after node reboot our container is not getting additional interface from Multus, instead we can the see the following errors in Multus logs:
To resolve this issue, we need to restart the pod that uses the Multus additional interface.
Additionally, if we simply kill the container within a pod, it will be restored with the additional interface, but there will be no log messages from Multus. It appears that Cilium is handling all the work.
When using Multus with Flannel, we didn't experience these issues.
Can you please advise on what we might be doing wrong?
The text was updated successfully, but these errors were encountered: