-
Notifications
You must be signed in to change notification settings - Fork 281
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Protect against starting rke2-server "by accident" on rke2-agent nodes #1590
Comments
For the RPM packages at least, we only install the unit for one - whichever is selected during install: Lines 25 to 27 in cfa99d2
The two units also conflict with each other, so you can't start them at the same time:
|
But "systemctl stop rke2-agent; systemctl start rke2-server" does work and has caused a big problem.... maybe we could get a "type=[agent|server]" option for config.yaml? |
Just accidentally ran "systemctl start rke2-server" on all of my agent nodes and Bad Things happened. Still working on recovering. This is v1.21.4+rke2r2. Installed via "curl -sfL https://get.rke2.io | sh -" method. |
I had to completely blow away and recreate the cluster. I didn't capture all the logs or issues or things I tried. But suffice to say, doing this is very bad and should be more actively prevented. |
In the mean time, I'm adding this to my Ansible deploy scripts for agent nodes:
|
This repository uses a bot to automatically label issues which have not had any activity (commit/comment/label) for 180 days. This helps us manage the community issues better. If the issue is still relevant, please add a comment to the issue so the bot can remove the label and we know it is still valid. If it is no longer relevant (or possibly fixed in the latest release), the bot will automatically close the issue in 14 days. Thank you for your contributions. |
Is your feature request related to a problem? Please describe.
We have started rke2-server instead of rke2-agent "by accident" on a system. This i.e. caused one more etcd to be created and this also caused a change in the cluster CIDR settings for kube-proxy, because on agents the settings for CIRD were not included in the config.yaml.
Describe the solution you'd like
We should have a config option "agent or server" in the config.yaml and have only one single service.
Or we should detect during server or agent start, that the "other" one has been enabled or similar..
Describe alternatives you've considered
Do not make "human" mistakes ;-).
The text was updated successfully, but these errors were encountered: