0
0
0
0
专栏/.../

TiDB迁移、升级与案例分享(TiDB v4.0.11 → v6.5.1)

 xingzhenxiang  发表于  2023-04-17

环境描述

Cluster version:    v4.0.11

部署情况

节点名称 部署情况
Node1 1tidb+1tikv+1pd
Node2 1tidb+1tikv+1pd+监控组件
Node3 1tidb+1tikv+1pd
Node4 4tikv
Node5 4tikv

数据大小

no-alt

主要任务

  1. 升级版本到v6.5.1
  2. 替换老旧机器Node1-Node3,并优化部署结构

升级测试

配置文件是否需要编辑

以源配置文件减配为3tidb+3pd+3tikv安装v4.0.11然后升级到v6.5.1。

这样可以验证配置文件是否有需要修改变更之处

监控迁移测试

生产系统从没有遇到过监控组件的迁移测试,为了保证过程安全可控,在测试环境进行监控组件迁移测试。

主要命令:

tiup cluster scale-in tidb-test -N XXX.XXX.XXX.XXX:9093,XXX.XXX.XXX.XXX:3000,XXX.XXX.XXX.XXX:9090
tiup cluster check  ./single-monitor.yaml  --user root -p (check因为tiup版本不同,方式有所不同)
tiup cluster scale-out tidb-test single-monitor.yaml  --user root -p

遇到问题:迁移后部分数据不显示

迁移前显示内容:

no-alt

迁移后显示内容:

no-alt

原因和解决方法:测试环境,新的监控组件所在服务器时间比正常晚了5分钟左右,时间同步正常

带数据升级测试

按照使用方提供的标准从生产环境导入部分数据,然后程序接入测试环境,做一次升级测试,并观察是否有不适应的地方

生产机器资源调整

tikv server通过扩容缩容的方式迁移

主要命令:

tiup cluster check  ./scale-out20230301.yaml  --user root -p
tiup cluster scale-out <cluster-name> scale-out20230301.yaml  --user root -p
tiup cluster scale-in <cluster-name>  --node XXX.XXX.XXX.60:20160,XXX.XXX.XXX.61:20160,XXX.XXX.XXX.62:20160

回收过程监控变化:no-alt

遇到问题:display没有Tombstone,监控还是有Tombstone

解决方法:

curl -X DELETE {pd_leader_ip}:2379/pd/api/v1/stores/remove-tombstone

tidb server通过扩容缩容的方式迁移

由于tidb是无状态节点,直接扩缩容

主要命令:

tiup cluster check  ./tidb-scale-out2023031701.yaml  --user root -p
tiup cluster scale-out <cluster-name> tidb-scale-out2023031701.yaml  --user root -p
tiup cluster scale-in <cluster-name>  --node XXX.XXX.XXX.XXX:3306

pd server通过扩容缩容的方式迁移

pd server 迁移官网注意提示:

no-alt

主要命令:

tiup cluster check  ./pd-scale-out2023031601.yaml  --user root -p
tiup cluster scale-out <cluster-name> pd-scale-out2023031601.yaml  --user root -p
tiup cluster scale-in <cluster-name> --node XXX.XXX.XXX.60:2379

原来leader 缩容:

我这里是通过reload的方式让主切换到新的机器上去的,可以先尝试reload,如果没有自动切换,再尝试手动切换leader

no-alt

监控组件迁移

步骤和测试场景一样

调整后部署情况

节点名称 部署情况
Node4 4tikv
Node5 4tikv
Node6 4tikv
Node7 4tikv
Node8 4tikv
Node9 1tidb+1pd
Node10 1tidb+1pd+监控组件
Node11 1tidb+1pd

生产升级

先升级 TiUP和cluster 版本(建议 tiup 版本不低于 1.11.0

tiup update --self
tiup --version
tiup update cluster
tiup cluster --version

检查当前集群的健康状况

tiup cluster check <cluster-name> --cluster

no-alt

其他确认项

变更配置文件,根据测试修改需要变更的参数 检查当前集群没有 DDL 和 Backup 在执行 表记录数据情况统计,方便升级完成验证

执行不停机升级升级

tiup cluster upgrade <cluster-name> v6.5.1

[09:52:11][tidb@Node1 ~]$ tiup cluster check <cluster-name> --cluster
[09:52:11]tiup is checking updates for component cluster ...
[09:52:11]Starting component `cluster`: /home/tidb/.tiup/components/cluster/v1.11.3/tiup-cluster check <cluster-name> --cluster
[09:52:13]Run command on XXX.XXX.XXX.68(sudo:false): /tmp/tiup/bin/insight
[09:52:13]Run command on XXX.XXX.XXX.67(sudo:false): /tmp/tiup/bin/insight
[09:52:13]Run command on XXX.XXX.XXX.98(sudo:false): /tmp/tiup/bin/insight
[09:52:13]Run command on XXX.XXX.XXX.99(sudo:false): /tmp/tiup/bin/insight
[09:52:13]Run command on XXX.XXX.XXX.97(sudo:false): /tmp/tiup/bin/insight
[09:52:14]Run command on XXX.XXX.XXX.103(sudo:false): /tmp/tiup/bin/insight
[09:52:15]Run command on XXX.XXX.XXX.102(sudo:false): /tmp/tiup/bin/insight
[09:52:15]Run command on XXX.XXX.XXX.101(sudo:false): /tmp/tiup/bin/insight
[09:52:17]Run command on XXX.XXX.XXX.97(sudo:false): cat /etc/security/limits.conf
[09:52:17]Run command on XXX.XXX.XXX.98(sudo:false): cat /etc/security/limits.conf
[09:52:17]Run command on XXX.XXX.XXX.99(sudo:false): cat /etc/security/limits.conf
[09:52:17]Run command on XXX.XXX.XXX.97(sudo:true): sysctl -a
[09:52:17]Run command on XXX.XXX.XXX.98(sudo:true): sysctl -a
[09:52:17]Run command on XXX.XXX.XXX.67(sudo:false): cat /etc/security/limits.conf
[09:52:17]Run command on XXX.XXX.XXX.99(sudo:true): sysctl -a
[09:52:17]Run command on XXX.XXX.XXX.68(sudo:false): cat /etc/security/limits.conf
[09:52:18]Run command on XXX.XXX.XXX.67(sudo:true): sysctl -a
[09:52:18]Run command on XXX.XXX.XXX.68(sudo:true): sysctl -a
[09:52:19]Run command on XXX.XXX.XXX.101(sudo:false): cat /etc/security/limits.conf
[09:52:20]Run command on XXX.XXX.XXX.102(sudo:false): cat /etc/security/limits.conf
[09:52:20]Run command on XXX.XXX.XXX.103(sudo:false): cat /etc/security/limits.conf
[09:52:20]Run command on XXX.XXX.XXX.101(sudo:true): sysctl -a
[09:52:20]Run command on XXX.XXX.XXX.103(sudo:true): sysctl -a
[09:52:20]Run command on XXX.XXX.XXX.102(sudo:true): sysctl -a
[09:52:22]Node            Check         Result  Message
[09:52:22]----            -----         ------  -------
[09:52:22]XXX.XXX.XXX.103  os-version    Pass    OS is CentOS Linux 7 (Core) 7.9.2009
[09:52:22]XXX.XXX.XXX.103  cpu-governor  Warn    Unable to determine current CPU frequency governor policy
[09:52:22]XXX.XXX.XXX.103  thp           Pass    THP is disabled
[09:52:22]XXX.XXX.XXX.103  timezone      Pass    time zone is the same as the first PD machine: Asia/Shanghai
[09:52:22]XXX.XXX.XXX.103  permission    Pass    /tikv3/tidb-deploy/tikv-20162 is writable
[09:52:22]XXX.XXX.XXX.103  permission    Pass    /tikv2/tidb-data/tikv-20161 is writable
[09:52:22]XXX.XXX.XXX.103  permission    Pass    /tikv4/tidb-data/tikv-20163 is writable
[09:52:22]XXX.XXX.XXX.103  permission    Pass    /tikv3/tidb-data/tikv-20162 is writable
[09:52:22]XXX.XXX.XXX.103  permission    Pass    /tikv1/tidb-deploy/tikv-20160 is writable
[09:52:22]XXX.XXX.XXX.103  permission    Pass    /tikv1/tidb-data/tikv-20160 is writable
[09:52:22]XXX.XXX.XXX.103  permission    Pass    /tikv2/tidb-deploy/tikv-20161 is writable
[09:52:22]XXX.XXX.XXX.103  permission    Pass    /tikv4/tidb-deploy/tikv-20163 is writable
[09:52:22]XXX.XXX.XXX.103  network       Pass    network speed of bond0 is 20000MB
[09:52:22]XXX.XXX.XXX.103  network       Pass    network speed of em1 is 10000MB
[09:52:22]XXX.XXX.XXX.103  network       Pass    network speed of em2 is 10000MB
[09:52:22]XXX.XXX.XXX.103  selinux       Pass    SELinux is disabled
[09:52:22]XXX.XXX.XXX.103  command       Pass    numactl: policy: default
[09:52:22]XXX.XXX.XXX.103  cpu-cores     Pass    number of CPU cores / threads: 72
[09:52:22]XXX.XXX.XXX.103  memory        Pass    memory size is 131072MB
[09:52:22]XXX.XXX.XXX.97   permission    Pass    /tidb-deploy/pd-2379 is writable
[09:52:22]XXX.XXX.XXX.97   permission    Pass    /tidb-data/pd-2379 is writable
[09:52:22]XXX.XXX.XXX.97   permission    Pass    /tidb-deploy/tidb-3306 is writable
[09:52:22]XXX.XXX.XXX.97   os-version    Pass    OS is CentOS Linux 7 (Core) 7.9.2009
[09:52:22]XXX.XXX.XXX.97   cpu-cores     Pass    number of CPU cores / threads: 56
[09:52:22]XXX.XXX.XXX.97   memory        Pass    memory size is 131072MB
[09:52:22]XXX.XXX.XXX.97   network       Pass    network speed of p2p2 is 1000MB
[09:52:22]XXX.XXX.XXX.97   network       Pass    network speed of bond0 is 20000MB
[09:52:22]XXX.XXX.XXX.97   network       Pass    network speed of em1 is 10000MB
[09:52:22]XXX.XXX.XXX.97   network       Pass    network speed of em2 is 10000MB
[09:52:22]XXX.XXX.XXX.97   network       Pass    network speed of p2p1 is 1000MB
[09:52:22]XXX.XXX.XXX.97   thp           Pass    THP is disabled
[09:52:22]XXX.XXX.XXX.97   command       Pass    numactl: policy: default
[09:52:22]XXX.XXX.XXX.97   cpu-governor  Warn    Unable to determine current CPU frequency governor policy
[09:52:22]XXX.XXX.XXX.97   selinux       Pass    SELinux is disabled
[09:52:22]XXX.XXX.XXX.98   timezone      Pass    time zone is the same as the first PD machine: Asia/Shanghai
[09:52:22]XXX.XXX.XXX.98   os-version    Pass    OS is CentOS Linux 7 (Core) 7.9.2009
[09:52:22]XXX.XXX.XXX.98   selinux       Pass    SELinux is disabled
[09:52:22]XXX.XXX.XXX.98   permission    Pass    /tidb-data/alertmanager-9093 is writable
[09:52:22]XXX.XXX.XXX.98   permission    Pass    /tidb-deploy/pd-2379 is writable
[09:52:22]XXX.XXX.XXX.98   permission    Pass    /tidb-data/pd-2379 is writable
[09:52:22]XXX.XXX.XXX.98   permission    Pass    /tidb-deploy/tidb-3306 is writable
[09:52:22]XXX.XXX.XXX.98   permission    Pass    /tidb-deploy/prometheus-9090 is writable
[09:52:22]XXX.XXX.XXX.98   permission    Pass    /tidb-deploy/grafana-3000 is writable
[09:52:22]XXX.XXX.XXX.98   permission    Pass    /tidb-deploy/alertmanager-9093 is writable
[09:52:22]XXX.XXX.XXX.98   permission    Pass    /tidb-data/prometheus-9090 is writable
[09:52:22]XXX.XXX.XXX.98   cpu-cores     Pass    number of CPU cores / threads: 56
[09:52:22]XXX.XXX.XXX.98   cpu-governor  Warn    Unable to determine current CPU frequency governor policy
[09:52:22]XXX.XXX.XXX.98   memory        Pass    memory size is 131072MB
[09:52:22]XXX.XXX.XXX.98   network       Pass    network speed of em2 is 10000MB
[09:52:22]XXX.XXX.XXX.98   network       Pass    network speed of p1p1 is 1000MB
[09:52:22]XXX.XXX.XXX.98   network       Pass    network speed of p1p2 is 1000MB
[09:52:22]XXX.XXX.XXX.98   network       Pass    network speed of p2p1 is 1000MB
[09:52:22]XXX.XXX.XXX.98   network       Pass    network speed of p2p2 is 1000MB
[09:52:22]XXX.XXX.XXX.98   network       Pass    network speed of bond0 is 20000MB
[09:52:22]XXX.XXX.XXX.98   network       Pass    network speed of em1 is 10000MB
[09:52:22]XXX.XXX.XXX.98   thp           Pass    THP is disabled
[09:52:22]XXX.XXX.XXX.98   command       Pass    numactl: policy: default
[09:52:22]XXX.XXX.XXX.99   os-version    Pass    OS is CentOS Linux 7 (Core) 7.9.2009
[09:52:22]XXX.XXX.XXX.99   memory        Pass    memory size is 131072MB
[09:52:22]XXX.XXX.XXX.99   selinux       Pass    SELinux is disabled
[09:52:22]XXX.XXX.XXX.99   timezone      Pass    time zone is the same as the first PD machine: Asia/Shanghai
[09:52:22]XXX.XXX.XXX.99   permission    Pass    /tidb-deploy/pd-2379 is writable
[09:52:22]XXX.XXX.XXX.99   permission    Pass    /tidb-data/pd-2379 is writable
[09:52:22]XXX.XXX.XXX.99   permission    Pass    /tidb-deploy/tidb-3306 is writable
[09:52:22]XXX.XXX.XXX.99   cpu-cores     Pass    number of CPU cores / threads: 56
[09:52:22]XXX.XXX.XXX.99   cpu-governor  Warn    Unable to determine current CPU frequency governor policy
[09:52:22]XXX.XXX.XXX.99   network       Pass    network speed of p2p1 is 1000MB
[09:52:22]XXX.XXX.XXX.99   network       Pass    network speed of p2p2 is 1000MB
[09:52:22]XXX.XXX.XXX.99   network       Pass    network speed of bond0 is 20000MB
[09:52:22]XXX.XXX.XXX.99   network       Pass    network speed of em1 is 10000MB
[09:52:22]XXX.XXX.XXX.99   network       Pass    network speed of em2 is 10000MB
[09:52:22]XXX.XXX.XXX.99   network       Pass    network speed of p1p1 is 1000MB
[09:52:22]XXX.XXX.XXX.99   network       Pass    network speed of p1p2 is 1000MB
[09:52:22]XXX.XXX.XXX.99   thp           Pass    THP is disabled
[09:52:22]XXX.XXX.XXX.99   command       Pass    numactl: policy: default
[09:52:22]XXX.XXX.XXX.67   os-version    Pass    OS is CentOS Linux 7 (Core) 7.7.1908
[09:52:22]XXX.XXX.XXX.67   memory        Pass    memory size is 131072MB
[09:52:22]XXX.XXX.XXX.67   network       Pass    network speed of em2 is 1000MB
[09:52:22]XXX.XXX.XXX.67   network       Pass    network speed of em3 is 1000MB
[09:52:22]XXX.XXX.XXX.67   network       Pass    network speed of em4 is 1000MB
[09:52:22]XXX.XXX.XXX.67   network       Pass    network speed of p1p1 is 1000MB
[09:52:22]XXX.XXX.XXX.67   network       Pass    network speed of bond0 is 1000MB
[09:52:22]XXX.XXX.XXX.67   network       Pass    network speed of em1 is 1000MB
[09:52:22]XXX.XXX.XXX.67   network       Pass    network speed of p1p4 is 1000MB
[09:52:22]XXX.XXX.XXX.67   network       Pass    network speed of p1p2 is 1000MB
[09:52:22]XXX.XXX.XXX.67   network       Pass    network speed of p1p3 is 1000MB
[09:52:22]XXX.XXX.XXX.67   selinux       Pass    SELinux is disabled
[09:52:22]XXX.XXX.XXX.67   command       Pass    numactl: policy: default
[09:52:22]XXX.XXX.XXX.67   timezone      Pass    time zone is the same as the first PD machine: Asia/Shanghai
[09:52:22]XXX.XXX.XXX.67   permission    Pass    /tikv1/tidb-deploy/tikv-20160 is writable
[09:52:22]XXX.XXX.XXX.67   permission    Pass    /tikv1/tidb-data/tikv-20160 is writable
[09:52:22]XXX.XXX.XXX.67   permission    Pass    /tikv2/tidb-data/tikv-20161 is writable
[09:52:22]XXX.XXX.XXX.67   permission    Pass    /tikv3/tidb-deploy/tikv-20162 is writable
[09:52:22]XXX.XXX.XXX.67   permission    Pass    /tikv4/tidb-deploy/tikv-20163 is writable
[09:52:22]XXX.XXX.XXX.67   permission    Pass    /tikv3/tidb-data/tikv-20162 is writable
[09:52:22]XXX.XXX.XXX.67   permission    Pass    /tikv4/tidb-data/tikv-20163 is writable
[09:52:22]XXX.XXX.XXX.67   permission    Pass    /tikv2/tidb-deploy/tikv-20161 is writable
[09:52:22]XXX.XXX.XXX.67   thp           Pass    THP is disabled
[09:52:22]XXX.XXX.XXX.67   cpu-cores     Pass    number of CPU cores / threads: 56
[09:52:22]XXX.XXX.XXX.67   cpu-governor  Warn    Unable to determine current CPU frequency governor policy
[09:52:22]XXX.XXX.XXX.68   network       Pass    network speed of p1p2 is 1000MB
[09:52:22]XXX.XXX.XXX.68   network       Pass    network speed of bond0 is 1000MB
[09:52:22]XXX.XXX.XXX.68   network       Pass    network speed of em1 is 1000MB
[09:52:22]XXX.XXX.XXX.68   network       Pass    network speed of em2 is 1000MB
[09:52:22]XXX.XXX.XXX.68   network       Pass    network speed of em3 is 1000MB
[09:52:22]XXX.XXX.XXX.68   network       Pass    network speed of em4 is 1000MB
[09:52:22]XXX.XXX.XXX.68   network       Pass    network speed of p1p1 is 1000MB
[09:52:22]XXX.XXX.XXX.68   network       Pass    network speed of p1p3 is 1000MB
[09:52:22]XXX.XXX.XXX.68   network       Pass    network speed of p1p4 is 1000MB
[09:52:22]XXX.XXX.XXX.68   command       Pass    numactl: policy: default
[09:52:22]XXX.XXX.XXX.68   timezone      Pass    time zone is the same as the first PD machine: Asia/Shanghai
[09:52:22]XXX.XXX.XXX.68   permission    Pass    /tikv2/tidb-data/tikv-20161 is writable
[09:52:22]XXX.XXX.XXX.68   permission    Pass    /tikv3/tidb-data/tikv-20162 is writable
[09:52:22]XXX.XXX.XXX.68   permission    Pass    /tikv1/tidb-data/tikv-20160 is writable
[09:52:22]XXX.XXX.XXX.68   permission    Pass    /tikv4/tidb-deploy/tikv-20163 is writable
[09:52:22]XXX.XXX.XXX.68   permission    Pass    /tikv4/tidb-data/tikv-20163 is writable
[09:52:22]XXX.XXX.XXX.68   permission    Pass    /tikv2/tidb-deploy/tikv-20161 is writable
[09:52:22]XXX.XXX.XXX.68   permission    Pass    /tikv3/tidb-deploy/tikv-20162 is writable
[09:52:22]XXX.XXX.XXX.68   permission    Pass    /tikv1/tidb-deploy/tikv-20160 is writable
[09:52:22]XXX.XXX.XXX.68   os-version    Pass    OS is CentOS Linux 7 (Core) 7.7.1908
[09:52:22]XXX.XXX.XXX.68   cpu-governor  Warn    Unable to determine current CPU frequency governor policy
[09:52:22]XXX.XXX.XXX.68   memory        Pass    memory size is 131072MB
[09:52:22]XXX.XXX.XXX.68   cpu-cores     Pass    number of CPU cores / threads: 56
[09:52:22]XXX.XXX.XXX.68   selinux       Pass    SELinux is disabled
[09:52:22]XXX.XXX.XXX.68   thp           Pass    THP is disabled
[09:52:22]XXX.XXX.XXX.101  permission    Pass    /tikv1/tidb-deploy/tikv-20160 is writable
[09:52:22]XXX.XXX.XXX.101  permission    Pass    /tikv2/tidb-deploy/tikv-20161 is writable
[09:52:22]XXX.XXX.XXX.101  permission    Pass    /tikv3/tidb-deploy/tikv-20162 is writable
[09:52:22]XXX.XXX.XXX.101  permission    Pass    /tikv4/tidb-deploy/tikv-20163 is writable
[09:52:22]XXX.XXX.XXX.101  permission    Pass    /tikv1/tidb-data/tikv-20160 is writable
[09:52:22]XXX.XXX.XXX.101  permission    Pass    /tikv4/tidb-data/tikv-20163 is writable
[09:52:22]XXX.XXX.XXX.101  permission    Pass    /tikv2/tidb-data/tikv-20161 is writable
[09:52:22]XXX.XXX.XXX.101  permission    Pass    /tikv3/tidb-data/tikv-20162 is writable
[09:52:22]XXX.XXX.XXX.101  os-version    Pass    OS is CentOS Linux 7 (Core) 7.9.2009
[09:52:22]XXX.XXX.XXX.101  cpu-governor  Warn    Unable to determine current CPU frequency governor policy
[09:52:22]XXX.XXX.XXX.101  network       Pass    network speed of em3 is 1000MB
[09:52:22]XXX.XXX.XXX.101  network       Pass    network speed of em4 is 1000MB
[09:52:22]XXX.XXX.XXX.101  network       Pass    network speed of p1p2 is 1000MB
[09:52:22]XXX.XXX.XXX.101  network       Pass    network speed of p2p4 is 10000MB
[09:52:22]XXX.XXX.XXX.101  network       Pass    network speed of bond0 is 20000MB
[09:52:22]XXX.XXX.XXX.101  network       Pass    network speed of em2 is 10000MB
[09:52:22]XXX.XXX.XXX.101  network       Pass    network speed of p1p1 is 1000MB
[09:52:22]XXX.XXX.XXX.101  network       Pass    network speed of p2p2 is 10000MB
[09:52:22]XXX.XXX.XXX.101  network       Pass    network speed of em1 is 10000MB
[09:52:22]XXX.XXX.XXX.101  selinux       Pass    SELinux is disabled
[09:52:22]XXX.XXX.XXX.101  command       Pass    numactl: policy: default
[09:52:22]XXX.XXX.XXX.101  timezone      Pass    time zone is the same as the first PD machine: Asia/Shanghai
[09:52:22]XXX.XXX.XXX.101  memory        Pass    memory size is 131072MB
[09:52:22]XXX.XXX.XXX.101  thp           Pass    THP is disabled
[09:52:22]XXX.XXX.XXX.101  cpu-cores     Pass    number of CPU cores / threads: 56
[09:52:22]XXX.XXX.XXX.102  timezone      Pass    time zone is the same as the first PD machine: Asia/Shanghai
[09:52:22]XXX.XXX.XXX.102  os-version    Pass    OS is CentOS Linux 7 (Core) 7.9.2009
[09:52:22]XXX.XXX.XXX.102  cpu-cores     Pass    number of CPU cores / threads: 56
[09:52:22]XXX.XXX.XXX.102  selinux       Pass    SELinux is disabled
[09:52:22]XXX.XXX.XXX.102  thp           Pass    THP is disabled
[09:52:22]XXX.XXX.XXX.102  command       Pass    numactl: policy: default
[09:52:22]XXX.XXX.XXX.102  permission    Pass    /tikv1/tidb-deploy/tikv-20160 is writable
[09:52:22]XXX.XXX.XXX.102  permission    Pass    /tikv4/tidb-data/tikv-20163 is writable
[09:52:22]XXX.XXX.XXX.102  permission    Pass    /tikv3/tidb-data/tikv-20162 is writable
[09:52:22]XXX.XXX.XXX.102  permission    Pass    /tikv2/tidb-data/tikv-20161 is writable
[09:52:22]XXX.XXX.XXX.102  permission    Pass    /tikv1/tidb-data/tikv-20160 is writable
[09:52:22]XXX.XXX.XXX.102  permission    Pass    /tikv3/tidb-deploy/tikv-20162 is writable
[09:52:22]XXX.XXX.XXX.102  permission    Pass    /tikv2/tidb-deploy/tikv-20161 is writable
[09:52:22]XXX.XXX.XXX.102  permission    Pass    /tikv4/tidb-deploy/tikv-20163 is writable
[09:52:22]XXX.XXX.XXX.102  cpu-governor  Warn    Unable to determine current CPU frequency governor policy
[09:52:22]XXX.XXX.XXX.102  memory        Pass    memory size is 131072MB
[09:52:22]XXX.XXX.XXX.102  network       Pass    network speed of p2p3 is 10000MB
[09:52:22]XXX.XXX.XXX.102  network       Pass    network speed of bond0 is 20000MB
[09:52:22]XXX.XXX.XXX.102  network       Pass    network speed of em1 is 10000MB
[09:52:22]XXX.XXX.XXX.102  network       Pass    network speed of em2 is 10000MB
[09:52:22]XXX.XXX.XXX.102  network       Pass    network speed of em4 is 1000MB
[09:52:22]XXX.XXX.XXX.102  network       Pass    network speed of p1p2 is 1000MB
[09:52:22]XXX.XXX.XXX.102  network       Pass    network speed of em3 is 1000MB
[09:52:22]XXX.XXX.XXX.102  network       Pass    network speed of p1p1 is 1000MB
[09:52:22]XXX.XXX.XXX.102  network       Pass    network speed of p2p4 is 10000MB
[09:52:22]Checking region status of the cluster <cluster-name>...
[09:52:22]All regions are healthy.
[09:53:35][tidb@Node1 ~]$ tiup cluster upgrade <cluster-name> v6.5.1
[09:53:35]tiup is checking updates for component cluster ...
[09:53:35]Starting component `cluster`: /home/tidb/.tiup/components/cluster/v1.11.3/tiup-cluster upgrade <cluster-name> v6.5.1
[09:53:35]Before the upgrade, it is recommended to read the upgrade guide at https://docs.pingcap.com/tidb/stable/upgrade-tidb-using-tiup and finish the preparation steps.
[09:53:35]This operation will upgrade tidb v4.0.11 cluster <cluster-name> to v6.5.1.
[09:53:59]Do you want to continue? [y/N]:(default=N) y
[09:53:59]Upgrading cluster...
[09:53:59]+ [ Serial ] - SSHKeySet: privateKey=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/ssh/id_rsa, publicKey=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/ssh/id_rsa.pub
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.67
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.98
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.67
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.99
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.67
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.67
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.97
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.101
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.101
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.68
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.68
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.102
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.68
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.102
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.102
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.101
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.103
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.102
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.68
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.103
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.101
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.97
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.103
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.103
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.99
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.98
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.98
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.98
[09:53:59]+ [Parallel] - UserSSH: user=tidb, host=XXX.XXX.XXX.98
[09:53:59]+ [ Serial ] - Download: component=grafana, version=v6.5.1, os=linux, arch=amd64
[09:53:59]+ [ Serial ] - Download: component=tidb, version=v6.5.1, os=linux, arch=amd64
[09:53:59]+ [ Serial ] - Download: component=prometheus, version=v6.5.1, os=linux, arch=amd64
[09:53:59]+ [ Serial ] - Download: component=pd, version=v6.5.1, os=linux, arch=amd64
[09:53:59]+ [ Serial ] - Download: component=tikv, version=v6.5.1, os=linux, arch=amd64
[09:54:06]+ [ Serial ] - Download: component=alertmanager, version=, os=linux, arch=amd64
[09:54:21]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.67, directories='/tikv2/tidb-data/tikv-20161'
[09:54:21]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.98, directories='/tidb-data/pd-2379'
[09:54:21]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.99, directories='/tidb-data/pd-2379'
[09:54:21]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.67, directories='/tikv1/tidb-data/tikv-20160'
[09:54:21]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.97, directories='/tidb-data/pd-2379'
[09:54:22]+ [ Serial ] - BackupComponent: component=pd, currentVersion=v4.0.11, remote=XXX.XXX.XXX.98:/tidb-deploy/pd-2379
[09:54:22]+ [ Serial ] - BackupComponent: component=pd, currentVersion=v4.0.11, remote=XXX.XXX.XXX.99:/tidb-deploy/pd-2379
[09:54:22]+ [ Serial ] - BackupComponent: component=pd, currentVersion=v4.0.11, remote=XXX.XXX.XXX.97:/tidb-deploy/pd-2379
[09:54:22]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.67:/tikv2/tidb-deploy/tikv-20161
[09:54:22]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.67:/tikv1/tidb-deploy/tikv-20160
[09:54:22]+ [ Serial ] - CopyComponent: component=pd, version=v6.5.1, remote=XXX.XXX.XXX.98:/tidb-deploy/pd-2379 os=linux, arch=amd64
[09:54:22]+ [ Serial ] - CopyComponent: component=pd, version=v6.5.1, remote=XXX.XXX.XXX.99:/tidb-deploy/pd-2379 os=linux, arch=amd64
[09:54:22]+ [ Serial ] - CopyComponent: component=pd, version=v6.5.1, remote=XXX.XXX.XXX.97:/tidb-deploy/pd-2379 os=linux, arch=amd64
[09:54:22]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.67:/tikv2/tidb-deploy/tikv-20161 os=linux, arch=amd64
[09:54:22]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.67:/tikv1/tidb-deploy/tikv-20160 os=linux, arch=amd64
[09:54:24]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.99, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/pd-2379.service, deploy_dir=/tidb-deploy/pd-2379, data_dir=[/tidb-data/pd-2379], log_dir=/tidb-deploy/pd-2379/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:25]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.98, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/pd-2379.service, deploy_dir=/tidb-deploy/pd-2379, data_dir=[/tidb-data/pd-2379], log_dir=/tidb-deploy/pd-2379/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:25]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.97, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/pd-2379.service, deploy_dir=/tidb-deploy/pd-2379, data_dir=[/tidb-data/pd-2379], log_dir=/tidb-deploy/pd-2379/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:26]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.67, directories='/tikv3/tidb-data/tikv-20162'
[09:54:27]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.67, directories='/tikv4/tidb-data/tikv-20163'
[09:54:27]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.68, directories='/tikv1/tidb-data/tikv-20160'
[09:54:28]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.67:/tikv3/tidb-deploy/tikv-20162
[09:54:28]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.68:/tikv1/tidb-deploy/tikv-20160
[09:54:28]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.67:/tikv4/tidb-deploy/tikv-20163
[09:54:28]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.67:/tikv3/tidb-deploy/tikv-20162 os=linux, arch=amd64
[09:54:29]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.68:/tikv1/tidb-deploy/tikv-20160 os=linux, arch=amd64
[09:54:29]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.67:/tikv4/tidb-deploy/tikv-20163 os=linux, arch=amd64
[09:54:31]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.67, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20160.service, deploy_dir=/tikv1/tidb-deploy/tikv-20160, data_dir=[/tikv1/tidb-data/tikv-20160], log_dir=/tikv1/tidb-deploy/tikv-20160/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:31]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.67, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20161.service, deploy_dir=/tikv2/tidb-deploy/tikv-20161, data_dir=[/tikv2/tidb-data/tikv-20161], log_dir=/tikv2/tidb-deploy/tikv-20161/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:35]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.68, directories='/tikv2/tidb-data/tikv-20161'
[09:54:35]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.68, directories='/tikv3/tidb-data/tikv-20162'
[09:54:36]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.68:/tikv2/tidb-deploy/tikv-20161
[09:54:36]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.68:/tikv3/tidb-deploy/tikv-20162
[09:54:37]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.68:/tikv2/tidb-deploy/tikv-20161 os=linux, arch=amd64
[09:54:37]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.68:/tikv3/tidb-deploy/tikv-20162 os=linux, arch=amd64
[09:54:37]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.67, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20162.service, deploy_dir=/tikv3/tidb-deploy/tikv-20162, data_dir=[/tikv3/tidb-data/tikv-20162], log_dir=/tikv3/tidb-deploy/tikv-20162/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:38]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.67, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20163.service, deploy_dir=/tikv4/tidb-deploy/tikv-20163, data_dir=[/tikv4/tidb-data/tikv-20163], log_dir=/tikv4/tidb-deploy/tikv-20163/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:39]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.68, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20160.service, deploy_dir=/tikv1/tidb-deploy/tikv-20160, data_dir=[/tikv1/tidb-data/tikv-20160], log_dir=/tikv1/tidb-deploy/tikv-20160/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:41]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.68, directories='/tikv4/tidb-data/tikv-20163'
[09:54:42]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.68:/tikv4/tidb-deploy/tikv-20163
[09:54:42]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.101, directories='/tikv1/tidb-data/tikv-20160'
[09:54:42]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.68:/tikv4/tidb-deploy/tikv-20163 os=linux, arch=amd64
[09:54:43]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.101:/tikv1/tidb-deploy/tikv-20160
[09:54:43]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.101, directories='/tikv2/tidb-data/tikv-20161'
[09:54:43]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.101:/tikv1/tidb-deploy/tikv-20160 os=linux, arch=amd64
[09:54:44]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.101:/tikv2/tidb-deploy/tikv-20161
[09:54:45]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.68, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20162.service, deploy_dir=/tikv3/tidb-deploy/tikv-20162, data_dir=[/tikv3/tidb-data/tikv-20162], log_dir=/tikv3/tidb-deploy/tikv-20162/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:45]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.68, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20161.service, deploy_dir=/tikv2/tidb-deploy/tikv-20161, data_dir=[/tikv2/tidb-data/tikv-20161], log_dir=/tikv2/tidb-deploy/tikv-20161/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:45]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.101:/tikv2/tidb-deploy/tikv-20161 os=linux, arch=amd64
[09:54:49]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.101, directories='/tikv3/tidb-data/tikv-20162'
[09:54:49]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.101, directories='/tikv4/tidb-data/tikv-20163'
[09:54:50]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.68, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20163.service, deploy_dir=/tikv4/tidb-deploy/tikv-20163, data_dir=[/tikv4/tidb-data/tikv-20163], log_dir=/tikv4/tidb-deploy/tikv-20163/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:50]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.101:/tikv3/tidb-deploy/tikv-20162
[09:54:50]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.101:/tikv4/tidb-deploy/tikv-20163
[09:54:50]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.101:/tikv3/tidb-deploy/tikv-20162 os=linux, arch=amd64
[09:54:50]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.101:/tikv4/tidb-deploy/tikv-20163 os=linux, arch=amd64
[09:54:52]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.101, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20160.service, deploy_dir=/tikv1/tidb-deploy/tikv-20160, data_dir=[/tikv1/tidb-data/tikv-20160], log_dir=/tikv1/tidb-deploy/tikv-20160/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:53]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.101, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20161.service, deploy_dir=/tikv2/tidb-deploy/tikv-20161, data_dir=[/tikv2/tidb-data/tikv-20161], log_dir=/tikv2/tidb-deploy/tikv-20161/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:53]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.102, directories='/tikv1/tidb-data/tikv-20160'
[09:54:54]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.102:/tikv1/tidb-deploy/tikv-20160
[09:54:55]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.102:/tikv1/tidb-deploy/tikv-20160 os=linux, arch=amd64
[09:54:56]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.102, directories='/tikv2/tidb-data/tikv-20161'
[09:54:57]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.102:/tikv2/tidb-deploy/tikv-20161
[09:54:57]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.102, directories='/tikv3/tidb-data/tikv-20162'
[09:54:57]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.102:/tikv2/tidb-deploy/tikv-20161 os=linux, arch=amd64
[09:54:58]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.101, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20162.service, deploy_dir=/tikv3/tidb-deploy/tikv-20162, data_dir=[/tikv3/tidb-data/tikv-20162], log_dir=/tikv3/tidb-deploy/tikv-20162/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:58]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.102:/tikv3/tidb-deploy/tikv-20162
[09:54:58]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.101, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20163.service, deploy_dir=/tikv4/tidb-deploy/tikv-20163, data_dir=[/tikv4/tidb-data/tikv-20163], log_dir=/tikv4/tidb-deploy/tikv-20163/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:54:58]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.102:/tikv3/tidb-deploy/tikv-20162 os=linux, arch=amd64
[09:55:01]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.102, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20160.service, deploy_dir=/tikv1/tidb-deploy/tikv-20160, data_dir=[/tikv1/tidb-data/tikv-20160], log_dir=/tikv1/tidb-deploy/tikv-20160/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:02]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.102, directories='/tikv4/tidb-data/tikv-20163'
[09:55:03]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.103, directories='/tikv1/tidb-data/tikv-20160'
[09:55:03]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.102:/tikv4/tidb-deploy/tikv-20163
[09:55:03]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.102:/tikv4/tidb-deploy/tikv-20163 os=linux, arch=amd64
[09:55:03]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.103:/tikv1/tidb-deploy/tikv-20160
[09:55:04]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.102, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20161.service, deploy_dir=/tikv2/tidb-deploy/tikv-20161, data_dir=[/tikv2/tidb-data/tikv-20161], log_dir=/tikv2/tidb-deploy/tikv-20161/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:04]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.103:/tikv1/tidb-deploy/tikv-20160 os=linux, arch=amd64
[09:55:05]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.103, directories='/tikv2/tidb-data/tikv-20161'
[09:55:05]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.102, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20162.service, deploy_dir=/tikv3/tidb-deploy/tikv-20162, data_dir=[/tikv3/tidb-data/tikv-20162], log_dir=/tikv3/tidb-deploy/tikv-20162/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:06]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.103:/tikv2/tidb-deploy/tikv-20161
[09:55:06]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.103:/tikv2/tidb-deploy/tikv-20161 os=linux, arch=amd64
[09:55:08]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.103, directories='/tikv3/tidb-data/tikv-20162'
[09:55:08]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.103:/tikv3/tidb-deploy/tikv-20162
[09:55:09]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.103:/tikv3/tidb-deploy/tikv-20162 os=linux, arch=amd64
[09:55:10]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.103, directories='/tikv4/tidb-data/tikv-20163'
[09:55:10]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.102, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20163.service, deploy_dir=/tikv4/tidb-deploy/tikv-20163, data_dir=[/tikv4/tidb-data/tikv-20163], log_dir=/tikv4/tidb-deploy/tikv-20163/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:10]+ [ Serial ] - BackupComponent: component=tikv, currentVersion=v4.0.11, remote=XXX.XXX.XXX.103:/tikv4/tidb-deploy/tikv-20163
[09:55:11]+ [ Serial ] - CopyComponent: component=tikv, version=v6.5.1, remote=XXX.XXX.XXX.103:/tikv4/tidb-deploy/tikv-20163 os=linux, arch=amd64
[09:55:12]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.103, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20160.service, deploy_dir=/tikv1/tidb-deploy/tikv-20160, data_dir=[/tikv1/tidb-data/tikv-20160], log_dir=/tikv1/tidb-deploy/tikv-20160/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:13]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.103, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20161.service, deploy_dir=/tikv2/tidb-deploy/tikv-20161, data_dir=[/tikv2/tidb-data/tikv-20161], log_dir=/tikv2/tidb-deploy/tikv-20161/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:14]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.97, directories=''
[09:55:14]+ [ Serial ] - BackupComponent: component=tidb, currentVersion=v4.0.11, remote=XXX.XXX.XXX.97:/tidb-deploy/tidb-3306
[09:55:15]+ [ Serial ] - CopyComponent: component=tidb, version=v6.5.1, remote=XXX.XXX.XXX.97:/tidb-deploy/tidb-3306 os=linux, arch=amd64
[09:55:16]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.103, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20162.service, deploy_dir=/tikv3/tidb-deploy/tikv-20162, data_dir=[/tikv3/tidb-data/tikv-20162], log_dir=/tikv3/tidb-deploy/tikv-20162/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:18]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.97, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tidb-3306.service, deploy_dir=/tidb-deploy/tidb-3306, data_dir=[], log_dir=/tidb-deploy/tidb-3306/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:18]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.103, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tikv-20163.service, deploy_dir=/tikv4/tidb-deploy/tikv-20163, data_dir=[/tikv4/tidb-data/tikv-20163], log_dir=/tikv4/tidb-deploy/tikv-20163/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:19]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.98, directories=''
[09:55:19]+ [ Serial ] - BackupComponent: component=tidb, currentVersion=v4.0.11, remote=XXX.XXX.XXX.98:/tidb-deploy/tidb-3306
[09:55:20]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.99, directories=''
[09:55:20]+ [ Serial ] - BackupComponent: component=tidb, currentVersion=v4.0.11, remote=XXX.XXX.XXX.99:/tidb-deploy/tidb-3306
[09:55:20]+ [ Serial ] - CopyComponent: component=tidb, version=v6.5.1, remote=XXX.XXX.XXX.98:/tidb-deploy/tidb-3306 os=linux, arch=amd64
[09:55:20]+ [ Serial ] - CopyComponent: component=tidb, version=v6.5.1, remote=XXX.XXX.XXX.99:/tidb-deploy/tidb-3306 os=linux, arch=amd64
[09:55:21]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.98, directories='/tidb-data/prometheus-9090'
[09:55:22]+ [ Serial ] - BackupComponent: component=prometheus, currentVersion=v4.0.11, remote=XXX.XXX.XXX.98:/tidb-deploy/prometheus-9090
[09:55:22]+ [ Serial ] - CopyComponent: component=prometheus, version=v6.5.1, remote=XXX.XXX.XXX.98:/tidb-deploy/prometheus-9090 os=linux, arch=amd64
[09:55:22]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.98, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tidb-3306.service, deploy_dir=/tidb-deploy/tidb-3306, data_dir=[], log_dir=/tidb-deploy/tidb-3306/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:23]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.99, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/tidb-3306.service, deploy_dir=/tidb-deploy/tidb-3306, data_dir=[], log_dir=/tidb-deploy/tidb-3306/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:24]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.98, directories=''
[09:55:24]+ [ Serial ] - BackupComponent: component=grafana, currentVersion=v4.0.11, remote=XXX.XXX.XXX.98:/tidb-deploy/grafana-3000
[09:55:24]+ [ Serial ] - Mkdir: host=XXX.XXX.XXX.98, directories='/tidb-data/alertmanager-9093'
[09:55:24]+ [ Serial ] - CopyComponent: component=grafana, version=v6.5.1, remote=XXX.XXX.XXX.98:/tidb-deploy/grafana-3000 os=linux, arch=amd64
[09:55:25]+ [ Serial ] - BackupComponent: component=alertmanager, currentVersion=v4.0.11, remote=XXX.XXX.XXX.98:/tidb-deploy/alertmanager-9093
[09:55:25]+ [ Serial ] - CopyComponent: component=alertmanager, version=, remote=XXX.XXX.XXX.98:/tidb-deploy/alertmanager-9093 os=linux, arch=amd64
[09:55:26]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.98, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/prometheus-9090.service, deploy_dir=/tidb-deploy/prometheus-9090, data_dir=[/tidb-data/prometheus-9090], log_dir=/tidb-deploy/prometheus-9090/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:26]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.98, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/alertmanager-9093.service, deploy_dir=/tidb-deploy/alertmanager-9093, data_dir=[/tidb-data/alertmanager-9093], log_dir=/tidb-deploy/alertmanager-9093/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:27]+ [ Serial ] - InitConfig: cluster=<cluster-name>, user=tidb, host=XXX.XXX.XXX.98, path=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache/grafana-3000.service, deploy_dir=/tidb-deploy/grafana-3000, data_dir=[], log_dir=/tidb-deploy/grafana-3000/log, cache_dir=/home/tidb/.tiup/storage/cluster/clusters/<cluster-name>/config-cache
[09:55:34]+ [ Serial ] - UpgradeCluster
[09:55:34]Upgrading component pd
[09:55:34]        Restarting instance XXX.XXX.XXX.98:2379
[09:55:36]        Restart instance XXX.XXX.XXX.98:2379 success
[09:55:38]        Restarting instance XXX.XXX.XXX.99:2379
[09:55:40]        Restart instance XXX.XXX.XXX.99:2379 success
[09:55:49]        Restarting instance XXX.XXX.XXX.97:2379
[09:55:51]        Restart instance XXX.XXX.XXX.97:2379 success
[09:55:53]Upgrading component tikv
[09:55:53]        Evicting 7109 leaders from store XXX.XXX.XXX.67:20160...
[09:55:53]          Still waitting for 7109 store leaders to transfer...
[09:55:55]          Still waitting for 7109 store leaders to transfer...
[09:55:57]          Still waitting for 7109 store leaders to transfer...
[09:56:00]          Still waitting for 7109 store leaders to transfer...
[09:56:02]          Still waitting for 7109 store leaders to transfer...
[09:56:04]          Still waitting for 5342 store leaders to transfer...
[09:56:06]          Still waitting for 5342 store leaders to transfer...
[09:56:08]          Still waitting for 5342 store leaders to transfer...
[09:56:10]          Still waitting for 5342 store leaders to transfer...
[09:56:12]          Still waitting for 4140 store leaders to transfer...
[09:56:14]          Still waitting for 4140 store leaders to transfer...
[09:56:16]          Still waitting for 4140 store leaders to transfer...
[09:56:18]          Still waitting for 4140 store leaders to transfer...
[09:56:21]          Still waitting for 4140 store leaders to transfer...
[09:56:23]          Still waitting for 2942 store leaders to transfer...
[09:56:25]          Still waitting for 2942 store leaders to transfer...
[09:56:27]          Still waitting for 2942 store leaders to transfer...
[09:56:29]          Still waitting for 2942 store leaders to transfer...
[09:56:31]          Still waitting for 2942 store leaders to transfer...
[09:56:33]          Still waitting for 1810 store leaders to transfer...
[09:56:35]          Still waitting for 1810 store leaders to transfer...
[09:56:38]          Still waitting for 1810 store leaders to transfer...
[09:56:40]          Still waitting for 1810 store leaders to transfer...
[09:56:42]          Still waitting for 1810 store leaders to transfer...
[09:56:44]          Still waitting for 488 store leaders to transfer...
[09:56:46]          Still waitting for 488 store leaders to transfer...
[09:56:48]          Still waitting for 488 store leaders to transfer...
[09:56:50]          Still waitting for 488 store leaders to transfer...
[09:56:52]        Restarting instance XXX.XXX.XXX.67:20160
[09:57:17]        Restart instance XXX.XXX.XXX.67:20160 success
[09:57:17]        Evicting 7469 leaders from store XXX.XXX.XXX.67:20161...
[09:57:17]          Still waitting for 7469 store leaders to transfer...
[09:57:19]          Still waitting for 7469 store leaders to transfer...
[09:57:21]          Still waitting for 7469 store leaders to transfer...
[09:57:23]          Still waitting for 7469 store leaders to transfer...
[09:57:26]          Still waitting for 5324 store leaders to transfer...
[09:57:28]          Still waitting for 5324 store leaders to transfer...
[09:57:30]          Still waitting for 5324 store leaders to transfer...
[09:57:32]          Still waitting for 5324 store leaders to transfer...
[09:57:34]          Still waitting for 5324 store leaders to transfer...
[09:57:36]          Still waitting for 2513 store leaders to transfer...
[09:57:38]          Still waitting for 2513 store leaders to transfer...
[09:57:40]          Still waitting for 2513 store leaders to transfer...
[09:57:42]          Still waitting for 2513 store leaders to transfer...
[09:57:45]          Still waitting for 2513 store leaders to transfer...
[09:57:47]        Restarting instance XXX.XXX.XXX.67:20161
[09:58:11]        Restart instance XXX.XXX.XXX.67:20161 success
[09:58:11]        Evicting 7481 leaders from store XXX.XXX.XXX.67:20162...
[09:58:11]          Still waitting for 7481 store leaders to transfer...
[09:58:13]          Still waitting for 7481 store leaders to transfer...
[09:58:15]          Still waitting for 7481 store leaders to transfer...
[09:58:18]          Still waitting for 5796 store leaders to transfer...
[09:58:20]          Still waitting for 5796 store leaders to transfer...
[09:58:22]          Still waitting for 5796 store leaders to transfer...
[09:58:24]          Still waitting for 5796 store leaders to transfer...
[09:58:26]          Still waitting for 5796 store leaders to transfer...
[09:58:28]          Still waitting for 2902 store leaders to transfer...
[09:58:30]          Still waitting for 2902 store leaders to transfer...
[09:58:32]          Still waitting for 2902 store leaders to transfer...
[09:58:34]          Still waitting for 2902 store leaders to transfer...
[09:58:37]          Still waitting for 2902 store leaders to transfer...
[09:58:39]          Still waitting for 22 store leaders to transfer...
[09:58:41]          Still waitting for 22 store leaders to transfer...
[09:58:43]          Still waitting for 22 store leaders to transfer...
[09:58:45]          Still waitting for 22 store leaders to transfer...
[09:58:47]          Still waitting for 22 store leaders to transfer...
[09:58:49]        Restarting instance XXX.XXX.XXX.67:20162
[09:59:15]        Restart instance XXX.XXX.XXX.67:20162 success
[09:59:15]        Evicting 7480 leaders from store XXX.XXX.XXX.67:20163...
[09:59:15]          Still waitting for 7480 store leaders to transfer...
[09:59:17]          Still waitting for 7480 store leaders to transfer...
[09:59:20]          Still waitting for 7480 store leaders to transfer...
[09:59:22]          Still waitting for 5687 store leaders to transfer...
[09:59:24]          Still waitting for 5687 store leaders to transfer...
[09:59:26]          Still waitting for 5687 store leaders to transfer...
[09:59:28]          Still waitting for 5687 store leaders to transfer...
[09:59:30]          Still waitting for 5687 store leaders to transfer...
[09:59:32]          Still waitting for 2842 store leaders to transfer...
[09:59:34]          Still waitting for 2842 store leaders to transfer...
[09:59:36]          Still waitting for 2842 store leaders to transfer...
[09:59:39]          Still waitting for 2842 store leaders to transfer...
[09:59:41]          Still waitting for 2842 store leaders to transfer...
[09:59:43]          Still waitting for 7 store leaders to transfer...
[09:59:45]          Still waitting for 7 store leaders to transfer...
[09:59:47]          Still waitting for 7 store leaders to transfer...
[09:59:49]          Still waitting for 7 store leaders to transfer...
[09:59:51]          Still waitting for 7 store leaders to transfer...
[09:59:53]        Restarting instance XXX.XXX.XXX.67:20163
[10:00:19]        Restart instance XXX.XXX.XXX.67:20163 success
[10:00:19]        Evicting 7480 leaders from store XXX.XXX.XXX.68:20160...
[10:00:19]          Still waitting for 7480 store leaders to transfer...
[10:00:22]          Still waitting for 7063 store leaders to transfer...
[10:00:24]          Still waitting for 7063 store leaders to transfer...
[10:00:26]          Still waitting for 7063 store leaders to transfer...
[10:00:28]          Still waitting for 7063 store leaders to transfer...
[10:00:30]          Still waitting for 7063 store leaders to transfer...
[10:00:32]          Still waitting for 4180 store leaders to transfer...
[10:00:34]          Still waitting for 4180 store leaders to transfer...
[10:00:36]          Still waitting for 4180 store leaders to transfer...
[10:00:38]          Still waitting for 4180 store leaders to transfer...
[10:00:41]          Still waitting for 4180 store leaders to transfer...
[10:00:43]          Still waitting for 1318 store leaders to transfer...
[10:00:45]          Still waitting for 1318 store leaders to transfer...
[10:00:47]          Still waitting for 1318 store leaders to transfer...
[10:00:49]          Still waitting for 1318 store leaders to transfer...
[10:00:51]        Restarting instance XXX.XXX.XXX.68:20160
[10:01:16]        Restart instance XXX.XXX.XXX.68:20160 success
[10:01:16]        Evicting 7481 leaders from store XXX.XXX.XXX.68:20161...
[10:01:16]          Still waitting for 7481 store leaders to transfer...
[10:01:18]          Still waitting for 7481 store leaders to transfer...
[10:01:20]          Still waitting for 7481 store leaders to transfer...
[10:01:22]          Still waitting for 7481 store leaders to transfer...
[10:01:24]          Still waitting for 7481 store leaders to transfer...
[10:01:26]          Still waitting for 4703 store leaders to transfer...
[10:01:29]          Still waitting for 4703 store leaders to transfer...
[10:01:31]          Still waitting for 4703 store leaders to transfer...
[10:01:33]          Still waitting for 4703 store leaders to transfer...
[10:01:35]          Still waitting for 4703 store leaders to transfer...
[10:01:37]          Still waitting for 1823 store leaders to transfer...
[10:01:39]          Still waitting for 1823 store leaders to transfer...
[10:01:41]          Still waitting for 1823 store leaders to transfer...
[10:01:43]          Still waitting for 1823 store leaders to transfer...
[10:01:45]          Still waitting for 1823 store leaders to transfer...
[10:01:48]        Restarting instance XXX.XXX.XXX.68:20161
[10:02:13]        Restart instance XXX.XXX.XXX.68:20161 success
[10:02:13]        Evicting 7481 leaders from store XXX.XXX.XXX.68:20162...
[10:02:13]          Still waitting for 7481 store leaders to transfer...
[10:02:16]          Still waitting for 7481 store leaders to transfer...
[10:02:18]          Still waitting for 6401 store leaders to transfer...
[10:02:20]          Still waitting for 6401 store leaders to transfer...
[10:02:22]          Still waitting for 6401 store leaders to transfer...
[10:02:24]          Still waitting for 6401 store leaders to transfer...
[10:02:26]          Still waitting for 6401 store leaders to transfer...
[10:02:27]
[10:02:28]          Still waitting for 3513 store leaders to transfer...
[10:02:30]          Still waitting for 3513 store leaders to transfer...
[10:02:32]          Still waitting for 3513 store leaders to transfer...
[10:02:35]          Still waitting for 3513 store leaders to transfer...
[10:02:35]
[10:02:36]
[10:02:37]          Still waitting for 3513 store leaders to transfer...
[10:02:39]          Still waitting for 647 store leaders to transfer...
[10:02:41]          Still waitting for 647 store leaders to transfer...
[10:02:43]          Still waitting for 647 store leaders to transfer...
[10:02:45]          Still waitting for 647 store leaders to transfer...
[10:02:47]        Restarting instance XXX.XXX.XXX.68:20162
[10:03:02]
[10:03:12]        Restart instance XXX.XXX.XXX.68:20162 success
[10:03:12]        Evicting 7482 leaders from store XXX.XXX.XXX.68:20163...
[10:03:12]          Still waitting for 7482 store leaders to transfer...
[10:03:14]          Still waitting for 7191 store leaders to transfer...
[10:03:16]          Still waitting for 7191 store leaders to transfer...
[10:03:18]          Still waitting for 7191 store leaders to transfer...
[10:03:20]          Still waitting for 7191 store leaders to transfer...
[10:03:23]          Still waitting for 7191 store leaders to transfer...
[10:03:25]          Still waitting for 4395 store leaders to transfer...
[10:03:27]          Still waitting for 4395 store leaders to transfer...
[10:03:29]          Still waitting for 4395 store leaders to transfer...
[10:03:31]          Still waitting for 4395 store leaders to transfer...
[10:03:33]          Still waitting for 1569 store leaders to transfer...
[10:03:35]          Still waitting for 1569 store leaders to transfer...
[10:03:37]          Still waitting for 1569 store leaders to transfer...
[10:03:39]          Still waitting for 1569 store leaders to transfer...
[10:03:42]          Still waitting for 1569 store leaders to transfer...
[10:03:44]        Restarting instance XXX.XXX.XXX.68:20163
[10:04:04]
[10:04:08]        Restart instance XXX.XXX.XXX.68:20163 success
[10:04:09]        Evicting 7472 leaders from store XXX.XXX.XXX.101:20160...
[10:04:09]          Still waitting for 7472 store leaders to transfer...
[10:04:11]          Still waitting for 7472 store leaders to transfer...
[10:04:13]          Still waitting for 6740 store leaders to transfer...
[10:04:15]          Still waitting for 6740 store leaders to transfer...
[10:04:17]          Still waitting for 6740 store leaders to transfer...
[10:04:19]          Still waitting for 6740 store leaders to transfer...
[10:04:21]          Still waitting for 3869 store leaders to transfer...
[10:04:23]          Still waitting for 3869 store leaders to transfer...
[10:04:25]          Still waitting for 3869 store leaders to transfer...
[10:04:27]          Still waitting for 3869 store leaders to transfer...
[10:04:30]          Still waitting for 3869 store leaders to transfer...
[10:04:32]          Still waitting for 949 store leaders to transfer...
[10:04:34]          Still waitting for 949 store leaders to transfer...
[10:04:36]          Still waitting for 949 store leaders to transfer...
[10:04:38]          Still waitting for 949 store leaders to transfer...
[10:04:40]          Still waitting for 949 store leaders to transfer...
[10:04:42]        Restarting instance XXX.XXX.XXX.101:20160
[10:05:05]        Restart instance XXX.XXX.XXX.101:20160 success
[10:05:05]        Evicting 7474 leaders from store XXX.XXX.XXX.101:20161...
[10:05:05]          Still waitting for 7474 store leaders to transfer...
[10:05:07]          Still waitting for 7474 store leaders to transfer...
[10:05:09]          Still waitting for 7474 store leaders to transfer...
[10:05:11]          Still waitting for 7474 store leaders to transfer...
[10:05:14]          Still waitting for 7474 store leaders to transfer...
[10:05:16]          Still waitting for 4891 store leaders to transfer...
[10:05:18]          Still waitting for 4891 store leaders to transfer...
[10:05:20]          Still waitting for 4891 store leaders to transfer...
[10:05:22]          Still waitting for 4891 store leaders to transfer...
[10:05:24]          Still waitting for 4891 store leaders to transfer...
[10:05:26]          Still waitting for 1999 store leaders to transfer...
[10:05:28]          Still waitting for 1999 store leaders to transfer...
[10:05:30]          Still waitting for 1999 store leaders to transfer...
[10:05:32]          Still waitting for 1999 store leaders to transfer...
[10:05:34]        Restarting instance XXX.XXX.XXX.101:20161
[10:05:57]        Restart instance XXX.XXX.XXX.101:20161 success
[10:05:57]        Evicting 7483 leaders from store XXX.XXX.XXX.101:20162...
[10:05:57]          Still waitting for 7483 store leaders to transfer...
[10:06:00]          Still waitting for 7483 store leaders to transfer...
[10:06:02]          Still waitting for 7483 store leaders to transfer...
[10:06:04]          Still waitting for 6184 store leaders to transfer...
[10:06:06]          Still waitting for 6184 store leaders to transfer...
[10:06:08]          Still waitting for 6184 store leaders to transfer...
[10:06:10]          Still waitting for 6184 store leaders to transfer...
[10:06:12]          Still waitting for 3305 store leaders to transfer...
[10:06:14]          Still waitting for 3305 store leaders to transfer...
[10:06:16]          Still waitting for 3305 store leaders to transfer...
[10:06:18]          Still waitting for 3305 store leaders to transfer...
[10:06:20]          Still waitting for 3305 store leaders to transfer...
[10:06:23]          Still waitting for 409 store leaders to transfer...
[10:06:25]          Still waitting for 409 store leaders to transfer...
[10:06:27]          Still waitting for 409 store leaders to transfer...
[10:06:29]          Still waitting for 409 store leaders to transfer...
[10:06:31]          Still waitting for 409 store leaders to transfer...
[10:06:33]        Restarting instance XXX.XXX.XXX.101:20162
[10:06:55]        Restart instance XXX.XXX.XXX.101:20162 success
[10:06:55]        Evicting 7483 leaders from store XXX.XXX.XXX.101:20163...
[10:06:55]          Still waitting for 7483 store leaders to transfer...
[10:06:57]          Still waitting for 7441 store leaders to transfer...
[10:06:59]          Still waitting for 7441 store leaders to transfer...
[10:07:01]          Still waitting for 7441 store leaders to transfer...
[10:07:03]          Still waitting for 7441 store leaders to transfer...
[10:07:05]          Still waitting for 4564 store leaders to transfer...
[10:07:07]          Still waitting for 4564 store leaders to transfer...
[10:07:09]          Still waitting for 4564 store leaders to transfer...
[10:07:12]          Still waitting for 4564 store leaders to transfer...
[10:07:14]          Still waitting for 4564 store leaders to transfer...
[10:07:16]          Still waitting for 1694 store leaders to transfer...
[10:07:18]          Still waitting for 1694 store leaders to transfer...
[10:07:20]          Still waitting for 1694 store leaders to transfer...
[10:07:22]          Still waitting for 1694 store leaders to transfer...
[10:07:24]          Still waitting for 1694 store leaders to transfer...
[10:07:26]        Restarting instance XXX.XXX.XXX.101:20163
[10:07:49]        Restart instance XXX.XXX.XXX.101:20163 success
[10:07:49]        Evicting 7478 leaders from store XXX.XXX.XXX.102:20160...
[10:07:49]          Still waitting for 7478 store leaders to transfer...
[10:07:51]          Still waitting for 7478 store leaders to transfer...
[10:07:53]          Still waitting for 7478 store leaders to transfer...
[10:07:55]          Still waitting for 5945 store leaders to transfer...
[10:07:58]          Still waitting for 5945 store leaders to transfer...
[10:08:00]          Still waitting for 5945 store leaders to transfer...
[10:08:02]          Still waitting for 5945 store leaders to transfer...
[10:08:04]          Still waitting for 5945 store leaders to transfer...
[10:08:06]          Still waitting for 3029 store leaders to transfer...
[10:08:08]          Still waitting for 3029 store leaders to transfer...
[10:08:10]          Still waitting for 3029 store leaders to transfer...
[10:08:12]          Still waitting for 3029 store leaders to transfer...
[10:08:14]          Still waitting for 3029 store leaders to transfer...
[10:08:16]          Still waitting for 138 store leaders to transfer...
[10:08:18]          Still waitting for 138 store leaders to transfer...
[10:08:21]          Still waitting for 138 store leaders to transfer...
[10:08:23]          Still waitting for 138 store leaders to transfer...
[10:08:25]        Restarting instance XXX.XXX.XXX.102:20160
[10:08:46]        Restart instance XXX.XXX.XXX.102:20160 success
[10:08:46]        Evicting 7477 leaders from store XXX.XXX.XXX.102:20161...
[10:08:47]          Still waitting for 7477 store leaders to transfer...
[10:08:49]          Still waitting for 7477 store leaders to transfer...
[10:08:51]          Still waitting for 7477 store leaders to transfer...
[10:08:53]          Still waitting for 5774 store leaders to transfer...
[10:08:55]          Still waitting for 5774 store leaders to transfer...
[10:08:57]          Still waitting for 5774 store leaders to transfer...
[10:08:59]          Still waitting for 5774 store leaders to transfer...
[10:09:01]          Still waitting for 5774 store leaders to transfer...
[10:09:03]          Still waitting for 2885 store leaders to transfer...
[10:09:05]          Still waitting for 2885 store leaders to transfer...
[10:09:07]          Still waitting for 2885 store leaders to transfer...
[10:09:10]          Still waitting for 2885 store leaders to transfer...
[10:09:12]          Still waitting for 2885 store leaders to transfer...
[10:09:14]          Still waitting for 24 store leaders to transfer...
[10:09:16]          Still waitting for 24 store leaders to transfer...
[10:09:18]          Still waitting for 24 store leaders to transfer...
[10:09:20]          Still waitting for 24 store leaders to transfer...
[10:09:22]          Still waitting for 24 store leaders to transfer...
[10:09:24]        Restarting instance XXX.XXX.XXX.102:20161
[10:09:47]        Restart instance XXX.XXX.XXX.102:20161 success
[10:09:47]        Evicting 7484 leaders from store XXX.XXX.XXX.102:20162...
[10:09:47]          Still waitting for 7484 store leaders to transfer...
[10:09:49]          Still waitting for 7484 store leaders to transfer...
[10:09:51]          Still waitting for 7484 store leaders to transfer...
[10:09:54]          Still waitting for 7484 store leaders to transfer...
[10:09:56]          Still waitting for 5564 store leaders to transfer...
[10:09:58]          Still waitting for 5564 store leaders to transfer...
[10:10:00]          Still waitting for 5564 store leaders to transfer...
[10:10:02]          Still waitting for 5564 store leaders to transfer...
[10:10:04]          Still waitting for 2748 store leaders to transfer...
[10:10:06]          Still waitting for 2748 store leaders to transfer...
[10:10:08]          Still waitting for 2748 store leaders to transfer...
[10:10:10]          Still waitting for 2748 store leaders to transfer...
[10:10:12]          Still waitting for 2748 store leaders to transfer...
[10:10:14]        Restarting instance XXX.XXX.XXX.102:20162
[10:10:36]        Restart instance XXX.XXX.XXX.102:20162 success
[10:10:36]        Evicting 7483 leaders from store XXX.XXX.XXX.102:20163...
[10:10:36]          Still waitting for 7483 store leaders to transfer...
[10:10:38]          Still waitting for 7483 store leaders to transfer...
[10:10:40]          Still waitting for 7483 store leaders to transfer...
[10:10:42]          Still waitting for 7483 store leaders to transfer...
[10:10:45]          Still waitting for 5573 store leaders to transfer...
[10:10:47]          Still waitting for 5573 store leaders to transfer...
[10:10:49]          Still waitting for 5573 store leaders to transfer...
[10:10:51]          Still waitting for 5573 store leaders to transfer...
[10:10:53]          Still waitting for 5573 store leaders to transfer...
[10:10:55]          Still waitting for 3276 store leaders to transfer...
[10:10:57]          Still waitting for 3276 store leaders to transfer...
[10:10:59]          Still waitting for 3276 store leaders to transfer...
[10:11:01]          Still waitting for 3276 store leaders to transfer...
[10:11:03]          Still waitting for 396 store leaders to transfer...
[10:11:05]          Still waitting for 396 store leaders to transfer...
[10:11:08]          Still waitting for 396 store leaders to transfer...
[10:11:10]          Still waitting for 396 store leaders to transfer...
[10:11:12]          Still waitting for 396 store leaders to transfer...
[10:11:14]        Restarting instance XXX.XXX.XXX.102:20163
[10:11:28]
[10:11:29]
[10:11:37]        Restart instance XXX.XXX.XXX.102:20163 success
[10:11:37]        Evicting 7476 leaders from store XXX.XXX.XXX.103:20160...
[10:11:37]          Still waitting for 7476 store leaders to transfer...
[10:11:39]          Still waitting for 7476 store leaders to transfer...
[10:11:41]          Still waitting for 6360 store leaders to transfer...
[10:11:43]          Still waitting for 6360 store leaders to transfer...
[10:11:45]          Still waitting for 6360 store leaders to transfer...
[10:11:47]          Still waitting for 6360 store leaders to transfer...
[10:11:49]          Still waitting for 6360 store leaders to transfer...
[10:11:51]          Still waitting for 3458 store leaders to transfer...
[10:11:53]          Still waitting for 3458 store leaders to transfer...
[10:11:56]          Still waitting for 3458 store leaders to transfer...
[10:11:58]          Still waitting for 3458 store leaders to transfer...
[10:12:00]          Still waitting for 3458 store leaders to transfer...
[10:12:02]          Still waitting for 572 store leaders to transfer...
[10:12:04]          Still waitting for 572 store leaders to transfer...
[10:12:06]          Still waitting for 572 store leaders to transfer...
[10:12:08]          Still waitting for 572 store leaders to transfer...
[10:12:10]          Still waitting for 572 store leaders to transfer...
[10:12:12]        Restarting instance XXX.XXX.XXX.103:20160
[10:12:37]        Restart instance XXX.XXX.XXX.103:20160 success
[10:12:37]        Evicting 7476 leaders from store XXX.XXX.XXX.103:20161...
[10:12:37]          Still waitting for 7476 store leaders to transfer...
[10:12:39]          Still waitting for 7476 store leaders to transfer...
[10:12:41]          Still waitting for 6714 store leaders to transfer...
[10:12:43]          Still waitting for 6714 store leaders to transfer...
[10:12:46]          Still waitting for 6714 store leaders to transfer...
[10:12:48]          Still waitting for 6714 store leaders to transfer...
[10:12:50]          Still waitting for 6714 store leaders to transfer...
[10:12:52]          Still waitting for 3853 store leaders to transfer...
[10:12:54]          Still waitting for 3853 store leaders to transfer...
[10:12:56]          Still waitting for 3853 store leaders to transfer...
[10:12:58]          Still waitting for 3853 store leaders to transfer...
[10:13:00]          Still waitting for 966 store leaders to transfer...
[10:13:02]          Still waitting for 966 store leaders to transfer...
[10:13:04]          Still waitting for 966 store leaders to transfer...
[10:13:06]          Still waitting for 966 store leaders to transfer...
[10:13:08]          Still waitting for 966 store leaders to transfer...
[10:13:11]        Restarting instance XXX.XXX.XXX.103:20161
[10:13:36]        Restart instance XXX.XXX.XXX.103:20161 success
[10:13:36]        Evicting 7485 leaders from store XXX.XXX.XXX.103:20162...
[10:13:36]          Still waitting for 7485 store leaders to transfer...
[10:13:38]          Still waitting for 7485 store leaders to transfer...
[10:13:40]          Still waitting for 7485 store leaders to transfer...
[10:13:42]          Still waitting for 5883 store leaders to transfer...
[10:13:44]          Still waitting for 5883 store leaders to transfer...
[10:13:47]          Still waitting for 5883 store leaders to transfer...
[10:13:49]          Still waitting for 5883 store leaders to transfer...
[10:13:51]          Still waitting for 5883 store leaders to transfer...
[10:13:53]          Still waitting for 3024 store leaders to transfer...
[10:13:55]          Still waitting for 3024 store leaders to transfer...
[10:13:57]          Still waitting for 3024 store leaders to transfer...
[10:13:59]          Still waitting for 3024 store leaders to transfer...
[10:14:01]          Still waitting for 3024 store leaders to transfer...
[10:14:03]          Still waitting for 137 store leaders to transfer...
[10:14:05]          Still waitting for 137 store leaders to transfer...
[10:14:07]          Still waitting for 137 store leaders to transfer...
[10:14:09]          Still waitting for 137 store leaders to transfer...
[10:14:12]          Still waitting for 137 store leaders to transfer...
[10:14:14]        Restarting instance XXX.XXX.XXX.103:20162
[10:14:38]        Restart instance XXX.XXX.XXX.103:20162 success
[10:14:38]        Evicting 7486 leaders from store XXX.XXX.XXX.103:20163...
[10:14:38]          Still waitting for 7486 store leaders to transfer...
[10:14:40]          Still waitting for 7486 store leaders to transfer...
[10:14:43]          Still waitting for 6568 store leaders to transfer...
[10:14:45]          Still waitting for 6568 store leaders to transfer...
[10:14:47]          Still waitting for 6568 store leaders to transfer...
[10:14:49]          Still waitting for 6568 store leaders to transfer...
[10:14:51]          Still waitting for 6568 store leaders to transfer...
[10:14:53]          Still waitting for 3666 store leaders to transfer...
[10:14:55]          Still waitting for 3666 store leaders to transfer...
[10:14:57]          Still waitting for 3666 store leaders to transfer...
[10:14:59]          Still waitting for 3666 store leaders to transfer...
[10:15:01]          Still waitting for 3666 store leaders to transfer...
[10:15:03]          Still waitting for 780 store leaders to transfer...
[10:15:06]          Still waitting for 780 store leaders to transfer...
[10:15:08]          Still waitting for 780 store leaders to transfer...
[10:15:10]          Still waitting for 780 store leaders to transfer...
[10:15:12]        Restarting instance XXX.XXX.XXX.103:20163
[10:15:36]        Restart instance XXX.XXX.XXX.103:20163 success
[10:15:36]Upgrading component tidb
[10:15:36]        Restarting instance XXX.XXX.XXX.97:3306
[10:16:31]        Restart instance XXX.XXX.XXX.97:3306 success
[10:16:31]        Restarting instance XXX.XXX.XXX.98:3306
[10:16:34]        Restart instance XXX.XXX.XXX.98:3306 success
[10:16:34]        Restarting instance XXX.XXX.XXX.99:3306
[10:16:37]        Restart instance XXX.XXX.XXX.99:3306 success
[10:16:37]Upgrading component prometheus
[10:16:37]        Restarting instance XXX.XXX.XXX.98:9090
[10:16:38]        Restart instance XXX.XXX.XXX.98:9090 success
[10:16:38]Upgrading component grafana
[10:16:38]        Restarting instance XXX.XXX.XXX.98:3000
[10:16:41]        Restart instance XXX.XXX.XXX.98:3000 success
[10:16:41]Upgrading component alertmanager
[10:16:41]        Restarting instance XXX.XXX.XXX.98:9093
[10:16:41]        Restart instance XXX.XXX.XXX.98:9093 success
[10:16:41]Stopping component node_exporter
[10:16:41]        Stopping instance XXX.XXX.XXX.99
[10:16:41]        Stopping instance XXX.XXX.XXX.102
[10:16:41]        Stopping instance XXX.XXX.XXX.103
[10:16:41]        Stopping instance XXX.XXX.XXX.98
[10:16:41]        Stopping instance XXX.XXX.XXX.68
[10:16:41]        Stopping instance XXX.XXX.XXX.101
[10:16:41]        Stopping instance XXX.XXX.XXX.67
[10:16:41]        Stopping instance XXX.XXX.XXX.97
[10:16:42]        Stop XXX.XXX.XXX.103 success
[10:16:42]        Stop XXX.XXX.XXX.98 success
[10:16:42]        Stop XXX.XXX.XXX.102 success
[10:16:42]        Stop XXX.XXX.XXX.101 success
[10:16:42]        Stop XXX.XXX.XXX.67 success
[10:16:42]        Stop XXX.XXX.XXX.97 success
[10:16:42]        Stop XXX.XXX.XXX.99 success
[10:16:42]        Stop XXX.XXX.XXX.68 success
[10:16:42]Stopping component blackbox_exporter
[10:16:42]        Stopping instance XXX.XXX.XXX.99
[10:16:42]        Stopping instance XXX.XXX.XXX.102
[10:16:42]        Stopping instance XXX.XXX.XXX.98
[10:16:42]        Stopping instance XXX.XXX.XXX.97
[10:16:42]        Stopping instance XXX.XXX.XXX.103
[10:16:42]        Stopping instance XXX.XXX.XXX.67
[10:16:42]        Stopping instance XXX.XXX.XXX.101
[10:16:42]        Stopping instance XXX.XXX.XXX.68
[10:16:42]        Stop XXX.XXX.XXX.98 success
[10:16:42]        Stop XXX.XXX.XXX.103 success
[10:16:42]        Stop XXX.XXX.XXX.99 success
[10:16:42]        Stop XXX.XXX.XXX.67 success
[10:16:42]        Stop XXX.XXX.XXX.97 success
[10:16:42]        Stop XXX.XXX.XXX.102 success
[10:16:42]        Stop XXX.XXX.XXX.68 success
[10:16:42]        Stop XXX.XXX.XXX.101 success
[10:16:42]Starting component node_exporter
[10:16:42]        Starting instance XXX.XXX.XXX.99
[10:16:42]        Starting instance XXX.XXX.XXX.98
[10:16:42]        Starting instance XXX.XXX.XXX.103
[10:16:42]        Starting instance XXX.XXX.XXX.101
[10:16:42]        Starting instance XXX.XXX.XXX.67
[10:16:42]        Starting instance XXX.XXX.XXX.97
[10:16:42]        Starting instance XXX.XXX.XXX.102
[10:16:42]        Starting instance XXX.XXX.XXX.68
[10:16:43]        Start XXX.XXX.XXX.98 success
[10:16:43]        Start XXX.XXX.XXX.99 success
[10:16:43]        Start XXX.XXX.XXX.67 success
[10:16:43]        Start XXX.XXX.XXX.68 success
[10:16:43]        Start XXX.XXX.XXX.97 success
[10:16:44]        Start XXX.XXX.XXX.103 success
[10:16:44]        Start XXX.XXX.XXX.101 success
[10:16:44]        Start XXX.XXX.XXX.102 success
[10:16:44]Starting component blackbox_exporter
[10:16:44]        Starting instance XXX.XXX.XXX.99
[10:16:44]        Starting instance XXX.XXX.XXX.68
[10:16:44]        Starting instance XXX.XXX.XXX.67
[10:16:44]        Starting instance XXX.XXX.XXX.98
[10:16:44]        Starting instance XXX.XXX.XXX.97
[10:16:44]        Starting instance XXX.XXX.XXX.101
[10:16:44]        Starting instance XXX.XXX.XXX.103
[10:16:44]        Starting instance XXX.XXX.XXX.102
[10:16:45]        Start XXX.XXX.XXX.98 success
[10:16:45]        Start XXX.XXX.XXX.99 success
[10:16:45]        Start XXX.XXX.XXX.68 success
[10:16:45]        Start XXX.XXX.XXX.97 success
[10:16:45]        Start XXX.XXX.XXX.67 success
[10:16:46]        Start XXX.XXX.XXX.103 success
[10:16:46]        Start XXX.XXX.XXX.102 success
[10:16:46]        Start XXX.XXX.XXX.101 success
[10:16:46]Upgraded cluster `<cluster-name>` successfully
[10:18:22][tidb@Node1 ~]$  tiup cluster display <cluster-name>       
[10:18:22]tiup is checking updates for component cluster ...
[10:18:22]Starting component `cluster`: /home/tidb/.tiup/components/cluster/v1.11.3/tiup-cluster display <cluster-name>
[10:18:22]Cluster type:       tidb
[10:18:22]Cluster name:       <cluster-name>
[10:18:22]Cluster version:    v6.5.1
[10:18:22]Deploy user:        tidb
[10:18:22]SSH type:           builtin
[10:18:22]Dashboard URL:      http://XXX.XXX.XXX.97:2379/dashboard
[10:18:22]Grafana URL:        http://XXX.XXX.XXX.98:3000
[10:18:22]ID                    Role          Host            Ports        OS/Arch       Status  Data Dir                             Deploy Dir
[10:18:22]--                    ----          ----            -----        -------       ------  --------                             ----------
[10:18:22]XXX.XXX.XXX.98:9093    alertmanager  XXX.XXX.XXX.98   9093/9094    linux/x86_64  Up      /tidb-data/alertmanager-9093  /tidb-deploy/alertmanager-9093
[10:18:22]XXX.XXX.XXX.98:3000    grafana       XXX.XXX.XXX.98   3000         linux/x86_64  Up      -                                    /tidb-deploy/grafana-3000
[10:18:22]XXX.XXX.XXX.97:2379    pd            XXX.XXX.XXX.97   2379/2380    linux/x86_64  Up|UI   /tidb-data/pd-2379            /tidb-deploy/pd-2379
[10:18:22]XXX.XXX.XXX.98:2379    pd            XXX.XXX.XXX.98   2379/2380    linux/x86_64  Up|L    /tidb-data/pd-2379            /tidb-deploy/pd-2379
[10:18:22]XXX.XXX.XXX.99:2379    pd            XXX.XXX.XXX.99   2379/2380    linux/x86_64  Up      /tidb-data/pd-2379            /tidb-deploy/pd-2379
[10:18:22]XXX.XXX.XXX.98:9090    prometheus    XXX.XXX.XXX.98   9090/12020   linux/x86_64  Up      /tidb-data/prometheus-9090    /tidb-deploy/prometheus-9090
[10:18:22]XXX.XXX.XXX.97:3306    tidb          XXX.XXX.XXX.97   3306/10080   linux/x86_64  Up      -                                    /tidb-deploy/tidb-3306
[10:18:22]XXX.XXX.XXX.98:3306    tidb          XXX.XXX.XXX.98   3306/10080   linux/x86_64  Up      -                                    /tidb-deploy/tidb-3306
[10:18:22]XXX.XXX.XXX.99:3306    tidb          XXX.XXX.XXX.99   3306/10080   linux/x86_64  Up      -                                    /tidb-deploy/tidb-3306
[10:18:22]XXX.XXX.XXX.101:20160  tikv          XXX.XXX.XXX.101  20160/20180  linux/x86_64  Up      /tikv1/tidb-data/tikv-20160   /tikv1/tidb-deploy/tikv-20160
[10:18:22]XXX.XXX.XXX.101:20161  tikv          XXX.XXX.XXX.101  20161/20181  linux/x86_64  Up      /tikv2/tidb-data/tikv-20161   /tikv2/tidb-deploy/tikv-20161
[10:18:22]XXX.XXX.XXX.101:20162  tikv          XXX.XXX.XXX.101  20162/20182  linux/x86_64  Up      /tikv3/tidb-data/tikv-20162   /tikv3/tidb-deploy/tikv-20162
[10:18:22]XXX.XXX.XXX.101:20163  tikv          XXX.XXX.XXX.101  20163/20183  linux/x86_64  Up      /tikv4/tidb-data/tikv-20163   /tikv4/tidb-deploy/tikv-20163
[10:18:22]XXX.XXX.XXX.102:20160  tikv          XXX.XXX.XXX.102  20160/20180  linux/x86_64  Up      /tikv1/tidb-data/tikv-20160   /tikv1/tidb-deploy/tikv-20160
[10:18:22]XXX.XXX.XXX.102:20161  tikv          XXX.XXX.XXX.102  20161/20181  linux/x86_64  Up      /tikv2/tidb-data/tikv-20161   /tikv2/tidb-deploy/tikv-20161
[10:18:22]XXX.XXX.XXX.102:20162  tikv          XXX.XXX.XXX.102  20162/20182  linux/x86_64  Up      /tikv3/tidb-data/tikv-20162   /tikv3/tidb-deploy/tikv-20162
[10:18:22]XXX.XXX.XXX.102:20163  tikv          XXX.XXX.XXX.102  20163/20183  linux/x86_64  Up      /tikv4/tidb-data/tikv-20163   /tikv4/tidb-deploy/tikv-20163
[10:18:22]XXX.XXX.XXX.103:20160  tikv          XXX.XXX.XXX.103  20160/20180  linux/x86_64  Up      /tikv1/tidb-data/tikv-20160   /tikv1/tidb-deploy/tikv-20160
[10:18:22]XXX.XXX.XXX.103:20161  tikv          XXX.XXX.XXX.103  20161/20181  linux/x86_64  Up      /tikv2/tidb-data/tikv-20161   /tikv2/tidb-deploy/tikv-20161
[10:18:22]XXX.XXX.XXX.103:20162  tikv          XXX.XXX.XXX.103  20162/20182  linux/x86_64  Up      /tikv3/tidb-data/tikv-20162   /tikv3/tidb-deploy/tikv-20162
[10:18:22]XXX.XXX.XXX.103:20163  tikv          XXX.XXX.XXX.103  20163/20183  linux/x86_64  Up      /tikv4/tidb-data/tikv-20163   /tikv4/tidb-deploy/tikv-20163
[10:18:22]XXX.XXX.XXX.67:20160   tikv          XXX.XXX.XXX.67   20160/20180  linux/x86_64  Up      /tikv1/tidb-data/tikv-20160   /tikv1/tidb-deploy/tikv-20160
[10:18:22]XXX.XXX.XXX.67:20161   tikv          XXX.XXX.XXX.67   20161/20181  linux/x86_64  Up      /tikv2/tidb-data/tikv-20161   /tikv2/tidb-deploy/tikv-20161
[10:18:22]XXX.XXX.XXX.67:20162   tikv          XXX.XXX.XXX.67   20162/20182  linux/x86_64  Up      /tikv3/tidb-data/tikv-20162   /tikv3/tidb-deploy/tikv-20162
[10:18:22]XXX.XXX.XXX.67:20163   tikv          XXX.XXX.XXX.67   20163/20183  linux/x86_64  Up      /tikv4/tidb-data/tikv-20163   /tikv4/tidb-deploy/tikv-20163
[10:18:22]XXX.XXX.XXX.68:20160   tikv          XXX.XXX.XXX.68   20160/20180  linux/x86_64  Up      /tikv1/tidb-data/tikv-20160   /tikv1/tidb-deploy/tikv-20160
[10:18:22]XXX.XXX.XXX.68:20161   tikv          XXX.XXX.XXX.68   20161/20181  linux/x86_64  Up      /tikv2/tidb-data/tikv-20161   /tikv2/tidb-deploy/tikv-20161
[10:18:22]XXX.XXX.XXX.68:20162   tikv          XXX.XXX.XXX.68   20162/20182  linux/x86_64  Up      /tikv3/tidb-data/tikv-20162   /tikv3/tidb-deploy/tikv-20162
[10:18:22]XXX.XXX.XXX.68:20163   tikv          XXX.XXX.XXX.68   20163/20183  linux/x86_64  Up      /tikv4/tidb-data/tikv-20163   /tikv4/tidb-deploy/tikv-20163
[10:18:22]Total nodes: 29

我把升级命令在screen里面执行的,怕本地与机房直接网络有问题影响升级过程

升级后验证

新统计表数据和以前的表记录数据相比较

中控机迁移

no-alt

遇到问题

auto analyze table 出现失败的情况

show analyze status;

no-alt

解决方法: 手工执行了,速度确实快,我的命令如下:

停止自动收集
set global tidb_auto_analyze_end_time ='01:00 +0000'; 即时生效
set global tidb_max_auto_analyze_time =600; 即时生效,在设置之前启动的超过这个时间也被kill;

设置并发参数
set global tidb_build_stats_concurrency=8; 这个变量用来设置 ANALYZE 语句执行时并发度。

执行手工收集
analyze table `Table_schema`.`Table_name` 手工速度比自动快的多,15分钟已经比自动3个小时执行的数据多了

PD 日志报错["transport: Got too many pings from the client, closing the connection."]

目前看不影响数据和集群使用,已经报官方排查

0
0
0
0

版权声明:本文为 TiDB 社区用户原创文章,遵循 CC BY-NC-SA 4.0 版权协议,转载请附上原文出处链接和本声明。

评论
暂无评论