From f3eafb923e2772823c7fadf75708a411980bddcb Mon Sep 17 00:00:00 2001 From: Liuxiaozhen12 <82579298+Liuxiaozhen12@users.noreply.github.com> Date: Thu, 5 Aug 2021 19:36:18 +0800 Subject: [PATCH] Update scale-tidb-using-tiup.md Co-authored-by: TomShawn <41534398+TomShawn@users.noreply.github.com> --- scale-tidb-using-tiup.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/scale-tidb-using-tiup.md b/scale-tidb-using-tiup.md index 0444705d919da..eb776a50e5a96 100644 --- a/scale-tidb-using-tiup.md +++ b/scale-tidb-using-tiup.md @@ -249,7 +249,7 @@ If you want to remove a TiKV node from the `10.0.1.5` host, take the following s > **Note:** > -> The PD Client in TiKV caches the list of PD nodes. The current version of TiKV has a mechanism to automatically and regularly update PD nodes, which can help avoid that the list of PD nodes cached by TiKV is expired. However, after expanding new PD, you should try to avoid directly shrinking all PD nodes existing before the expansion. If necessary, before making all the previously existing PD nodes offline, make sure to switch the PD leader to the newly expanded PD nodes. +> The PD Client in TiKV caches the list of PD nodes. The current version of TiKV has a mechanism to automatically and regularly update PD nodes, which can help mitigate the issue of an expired list of PD nodes cached by TiKV. However, after scaling out PD, you should try to avoid directly removing all PD nodes at once that exist before the scaling. If necessary, before making all the previously existing PD nodes offline, make sure to switch the PD leader to a newly added PD node. 1. View the node ID information: