-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Nebula resiliancey #5335
Comments
@porscheme I have some questions about your scenario:
|
Thanks @MegaByte875 for the reply
[Porsche] Each one of the component has its own separate K8s node pool and subnet. We are using Azure Standard_L16as_v3 SKU. Auto scale disabled
[Porsche] No, we are not using any PDB since Nebula official docs doesn't mention about it. Should we use PDB, can you point me to any Nebula docs?
[Porsche] Yes, Azure Standard_L16as_v3 SKU comes with two 2 TB SSD NVMe disks attached to the VM
[Porsche] Yes before upgrade, the partition leader does exists on the storage node and it was balanced with other storage nodes. But during upgrade, leaders on the storage node become zero (SHOW HOSTS) |
Azure patched our K8S cluster today and the Nebula cluster was down.
|
Here is an implementation plan, I wish will help you:
|
Cluster config
metad: 3
Storaged: 3
metad: 5
Each of the storage node has 2 X 2TB SSD NVMe Disks
Space Config
VID: String (Length 20)
Partition Number: 200
Replica Factor: 3
Our cluster is running in Azure. We enabled auto patching & upgrade (Kubernetes upgrade). Often times manual intervention is required when VM is stuck in upgrade.
The text was updated successfully, but these errors were encountered: