暂无图片
暂无图片
暂无图片
暂无图片
暂无图片

国产数据库|TiDB 7.5 实验测试环境搭建及小插曲处理

616
1、前  言

随着国产数据库技术的发展进步,我发现身边越来越多DBA同学开始华丽转型。近期我很幸运地参与到多款数据库技术支持工作,包括:达梦、人大金仓、TDSQL、TBase、TiDB、HotDB等。我每次入门学习一款新产品,都会尝试在自己的笔记本搭建测试环境。本期,我将为大家分享下“TiDB 7.5 实验测试环境搭建及小插曲处理”。

2、TiDB简介

TiDB 是平凯星辰 PingCAP 公司自主设计、研发的开源分布式关系型数据库,是一款同时支持在线事务处理与在线分析处理 (Hybrid Transactional and Analytical Processing, HTAP) 的融合型分布式数据库产品,具备水平扩容或者缩容、金融级高可用、实时 HTAP、云原生的分布式数据库、兼容 MySQL 协议和 MySQL 生态等重要特性。目标是为用户提供一站式 OLTP (Online Transactional Processing)、OLAP (Online Analytical Processing)、HTAP 解决方案。TiDB 适合高可用、强一致要求较高、数据规模较大等各种应用场景。

TiDB 7.5.0 为长期支持版本 (Long-Term Support Release, LTS)。

2024 年 3 月 29 日,TiDB 发布 8.0.0 DMR 版本。如果你感兴趣,可去官网查看相关内容:https://docs.pingcap.com/zh/tidb/v8.0/release-8.0.0

3、TiDB学习资料

TiDB课程 https://cn.pingcap.com/courses-catalog/

TiDB社区 https://tidb.net/

TiDB文档 https://docs.pingcap.com/zh/

【TiDB 社区智慧合集】史上最全的 158篇 TiDB 用户实践合集来啦!千万不要错过!

https://tidb.net/blog/87fb0073

4、单机模拟部署生产集群

适用场景:希望用单台 Linux 服务器,体验 TiDB 最小的完整拓扑的集群,并模拟生产环境下的部署步骤。只要你的主机可以联网,便可通过单台 Linux 服务器,体验 TiDB 最小的完整拓扑的集群,并模拟生产环境下的部署步骤,可以快速搭建出一套基础测试集群。

本节介绍如何参照 TiUP 最小拓扑的一个 YAML 文件部署 TiDB 集群。TiDB 是一个分布式系统。最基础的 TiDB 测试集群通常由 2 个 TiDB 实例、3 个 TiKV 实例、3 个 PD 实例和可选的 TiFlash 实例构成。部署步骤如下:

准备环境

开始部署 TiDB 集群前,准备一台部署主机,确保其软件满足需求:

(1)推荐安装 CentOS 7.3 及以上版本

(2)部署需要使用部署主机的 root 用户及密码

(3)部署主机关闭防火墙或者开放 TiDB 集群的节点间所需端口

(4)运行环境可以支持互联网访问,用于下载 TiDB 及相关软件安装包

最小规模的 TiDB 集群拓扑包含以下实例:

实例
个数IP配置
TiKV
3

192.168.198.132

192.168.198.132

192.168.198.132

避免端口和目录冲突
PD1192.168.198.132默认端口和全局目录配置
TiFlash1192.168.198.132默认端口和全局目录配置
Monitor
1192.168.198.132默认端口和全局目录配置

下载并安装 TiUP:

    curl --proto '=https' --tlsv1.2 -sSf https://tiup-mirrors.pingcap.com/install.sh | sh


    Successfully set mirror to https://tiup-mirrors.pingcap.com
    Detected shell: bash
    Shell profile: root/.bash_profile
    /root/.bash_profile has been modified to add tiup to PATH
    open a new terminal or source root/.bash_profile to use it
    Installed path: root/.tiup/bin/tiup
    ===============================================
    Have a try: tiup playground
    ===============================================

    声明全局环境变量:

      --source ${your_shell_profile}
      source root/.bash_profile

      安装 TiUP 的 cluster 组件:

        tiup cluster

        如果机器已经安装 TiUP cluster,需要更新软件版本:

          tiup update --self && tiup update cluster

          由于模拟多机部署,需要通过 root 用户调大 sshd 服务的连接数限制。

          修改 etc/ssh/sshd_config 将 MaxSessions 调至 20。

          重启 sshd 服务。

            service sshd restart

            创建并启动集群

            按下面的配置模板,编辑配置文件,命名为 topo.yaml,其中:

            user: "tidb":表示通过 tidb 系统用户(部署会自动创建)来做集群的内部管理,默认使用 22 端口通过 ssh 登录目标机器

            replication.enable-placement-rules:设置这个 PD 参数来确保 TiFlash 正常运行

            host:设置为本部署主机的 IP

            配置模板如下:

              # # Global variables are applied to all deployments and used as the default value of
              # # the deployments if a specific deployment value is missing.
              global:
              user: "tidb"
              ssh_port: 22
              deploy_dir: "/tidb-deploy"
              data_dir: "/tidb-data"


              # # Monitored variables are applied to all the machines.
              monitored:
              node_exporter_port: 9100
              blackbox_exporter_port: 9115


              server_configs:
              tidb:
              instance.tidb_slow_log_threshold: 300
              tikv:
              readpool.storage.use-unified-pool: false
              readpool.coprocessor.use-unified-pool: true
              pd:
              replication.enable-placement-rules: true
              replication.location-labels: ["host"]
              tiflash:
              logger.level: "info"


              pd_servers:
              - host: 192.168.198.132


              tidb_servers:
              - host: 192.168.198.132


              tikv_servers:
              - host: 192.168.198.132
              port: 20160
              status_port: 20180
              config:
              server.labels: { host: "logic-host-1" }


              - host: 192.168.198.132
              port: 20161
              status_port: 20181
              config:
              server.labels: { host: "logic-host-2" }


              - host: 192.168.198.132
              port: 20162
              status_port: 20182
              config:
              server.labels: { host: "logic-host-3" }


              tiflash_servers:
              - host: 192.168.198.132


              monitoring_servers:
              - host: 192.168.198.132


              grafana_servers:
              - host: 192.168.198.132

              执行集群部署命令:

                --tiup cluster deploy <cluster-name> <version> ./topo.yaml --user root -p
                tiup cluster deploy sjkywzd v7.5.1 ./topo.yaml --user root -p

                参数 <cluster-name> 表示设置集群名称

                参数 <version> 表示设置集群版本,例如 v7.5.1。可以通过 tiup list tidb 命令来查看当前支持部署的 TiDB 版本

                参数 -p 表示在连接目标机器时使用密码登录

                注意

                如果主机通过密钥进行 SSH 认证,请使用 -i 参数指定密钥文件路径,-i 与 -p 不可同时使用。

                按照引导,输入”y”及 root 密码,来完成部署:

                Do you want to continue? [y/N]:  y

                Input SSH password:

                  [root@tidb ~]# tiup cluster deploy sjkywzd v7.5.1 ./topo.yaml --user root -p
                  Input SSH password:


                  + Detect CPU Arch Name
                  - Detecting node 192.168.198.132 Arch info ... Done


                  + Detect CPU OS Name
                  - Detecting node 192.168.198.132 OS info ... Done
                  Please confirm your topology:
                  Cluster type: tidb
                  Cluster name: sjkywzd
                  Cluster version: v7.5.1
                  Role Host Ports OS/Arch Directories
                  ---- ---- ----- ------- -----------
                  pd 192.168.198.132 2379/2380 linux/x86_64 tidb-deploy/pd-2379,/tidb-data/pd-2379
                  tikv 192.168.198.132 20160/20180 linux/x86_64 tidb-deploy/tikv-20160,/tidb-data/tikv-20160
                  tikv 192.168.198.132 20161/20181 linux/x86_64 tidb-deploy/tikv-20161,/tidb-data/tikv-20161
                  tikv 192.168.198.132 20162/20182 linux/x86_64 tidb-deploy/tikv-20162,/tidb-data/tikv-20162
                  tidb 192.168.198.132 4000/10080 linux/x86_64 /tidb-deploy/tidb-4000
                  tiflash 192.168.198.132 9000/8123/3930/20170/20292/8234 linux/x86_64 /tidb-deploy/tiflash-9000,/tidb-data/tiflash-9000
                  prometheus 192.168.198.132 9090/12020 linux/x86_64 /tidb-deploy/prometheus-9090,/tidb-data/prometheus-9090
                  grafana 192.168.198.132 3000 linux/x86_64 /tidb-deploy/grafana-3000
                  Attention:
                  1. If the topology is not what you expected, check your yaml file.
                  2. Please confirm there is no port/directory conflicts in same host.
                  Do you want to continue? [y/N]: (default=N) y
                  + Generate SSH keys ... Done
                  + Download TiDB components
                  - Download pd:v7.5.1 (linux/amd64) ... Done
                  - Download tikv:v7.5.1 (linux/amd64) ... Done
                  - Download tidb:v7.5.1 (linux/amd64) ... Done
                  - Download tiflash:v7.5.1 (linux/amd64) ... Done
                  - Download prometheus:v7.5.1 (linux/amd64) ... Done
                  - Download grafana:v7.5.1 (linux/amd64) ... Done
                  - Download node_exporter: (linux/amd64) ... Done
                  - Download blackbox_exporter: (linux/amd64) ... Done
                  + Initialize target host environments
                  - Prepare 192.168.198.132:22 ... Done
                  + Deploy TiDB instance
                  - Copy pd -> 192.168.198.132 ... Done
                  - Copy tikv -> 192.168.198.132 ... Done
                  - Copy tikv -> 192.168.198.132 ... Done
                  - Copy tikv -> 192.168.198.132 ... Done
                  - Copy tidb -> 192.168.198.132 ... Done
                  - Copy tiflash -> 192.168.198.132 ... Done
                  - Copy prometheus -> 192.168.198.132 ... Done
                  - Copy grafana -> 192.168.198.132 ... Done
                  - Deploy node_exporter -> 192.168.198.132 ... Done
                  - Deploy blackbox_exporter -> 192.168.198.132 ... Done
                  + Copy certificate to remote host
                  + Init instance configs
                  - Generate config pd -> 192.168.198.132:2379 ... Done
                  - Generate config tikv -> 192.168.198.132:20160 ... Done
                  - Generate config tikv -> 192.168.198.132:20161 ... Done
                  - Generate config tikv -> 192.168.198.132:20162 ... Done
                  - Generate config tidb -> 192.168.198.132:4000 ... Done
                  - Generate config tiflash -> 192.168.198.132:9000 ... Done
                  - Generate config prometheus -> 192.168.198.132:9090 ... Done
                  - Generate config grafana -> 192.168.198.132:3000 ... Done
                  + Init monitor configs
                  - Generate config node_exporter -> 192.168.198.132 ... Done
                  - Generate config blackbox_exporter -> 192.168.198.132 ... Done
                  Enabling component pd
                  Enabling instance 192.168.198.132:2379
                  Enable instance 192.168.198.132:2379 success
                  Enabling component tikv
                  Enabling instance 192.168.198.132:20162
                  Enabling instance 192.168.198.132:20160
                  Enabling instance 192.168.198.132:20161
                  Enable instance 192.168.198.132:20161 success
                  Enable instance 192.168.198.132:20160 success
                  Enable instance 192.168.198.132:20162 success
                  Enabling component tidb
                  Enabling instance 192.168.198.132:4000
                  Enable instance 192.168.198.132:4000 success
                  Enabling component tiflash
                  Enabling instance 192.168.198.132:9000
                  Enable instance 192.168.198.132:9000 success
                  Enabling component prometheus
                  Enabling instance 192.168.198.132:9090
                  Enable instance 192.168.198.132:9090 success
                  Enabling component grafana
                  Enabling instance 192.168.198.132:3000
                  Enable instance 192.168.198.132:3000 success
                  Enabling component node_exporter
                  Enabling instance 192.168.198.132
                  Enable 192.168.198.132 success
                  Enabling component blackbox_exporter
                  Enabling instance 192.168.198.132
                  Enable 192.168.198.132 success
                  Cluster `sjkywzd` deployed successfully, you can start it with command: `tiup cluster start sjkywzd --init`


                  首次启动集群:

                    tiup cluster start sjkywzd --init

                    访问集群:

                    安装 MySQL 客户端。如果已安装 MySQL 客户端则可跳过这一步骤。

                      yum -y install mysql
                      yum install numactl

                      访问 TiDB 数据库,密码为空:

                        mysql -h 192.168.198.132 -P 4000 -u root

                        访问 TiDB 的 Grafana 监控:

                        通过 http://{grafana-ip}:3000 访问集群 Grafana 监控页面,默认用户名和密码均为 admin。

                          http://192.168.198.132:3000/login

                          访问 TiDB 的 Dashboard:

                          通过 http://{pd-ip}:2379/dashboard 访问集群 TiDB Dashboard 监控页面,默认用户名为 root,密码为空。

                          执行以下命令确认当前已经部署的集群列表:

                            tiup cluster list
                            [root@tidb log]# tiup cluster list
                            Name User Version Path PrivateKey
                            ---- ---- ------- ---- ----------
                            sjkywzd  tidb  v7.5.1   /root/.tiup/storage/cluster/clusters/sjkywzd  /root/.tiup/storage/cluster/clusters/sjkywzd/ssh/id_rsa

                            执行以下命令tiup cluster display sjkywzd查看集群的拓扑结构和状态:

                              --tiup cluster display <cluster-name>
                              [root@tidb log]# tiup cluster display sjkywzd
                              Cluster type: tidb
                              Cluster name: sjkywzd
                              Cluster version: v7.5.1
                              Deploy user: tidb
                              SSH type: builtin
                              Dashboard URL: http://192.168.198.132:2379/dashboard
                              Grafana URL: http://192.168.198.132:3000
                              ID Role Host Ports OS/Arch Status Data Dir Deploy Dir
                              -- ---- ---- ----- ------- ------ -------- ----------
                              192.168.198.132:3000 grafana 192.168.198.132 3000 linux/x86_64 Up - /tidb-deploy/grafana-3000
                              192.168.198.132:2379 pd 192.168.198.132 2379/2380 linux/x86_64 Up|L|UI /tidb-data/pd-2379 /tidb-deploy/pd-2379
                              192.168.198.132:9090 prometheus 192.168.198.132 9090/12020 linux/x86_64 Up /tidb-data/prometheus-9090 /tidb-deploy/prometheus-9090
                              192.168.198.132:4000 tidb 192.168.198.132 4000/10080 linux/x86_64 Up - /tidb-deploy/tidb-4000
                              192.168.198.132:9000 tiflash 192.168.198.132 9000/8123/3930/20170/20292/8234 linux/x86_64 Up /tidb-data/tiflash-9000 /tidb-deploy/tiflash-9000
                              192.168.198.132:20160 tikv 192.168.198.132 20160/20180 linux/x86_64 Up /tidb-data/tikv-20160 /tidb-deploy/tikv-20160
                              192.168.198.132:20161 tikv 192.168.198.132 20161/20181 linux/x86_64 Up /tidb-data/tikv-20161 /tidb-deploy/tikv-20161
                              192.168.198.132:20162 tikv 192.168.198.132 20162/20182 linux/x86_64 Up /tidb-data/tikv-20162 /tidb-deploy/tikv-20162
                              Total nodes: 8


                              5、清除集群数据

                              此操作会关闭所有服务,并清空其数据目录或/和日志目录,并且无法恢复,需要谨慎操作

                              清空集群所有服务的数据,但保留日志:

                                tiup cluster clean ${cluster-name} --data

                                清空集群所有服务的日志,但保留数据:

                                  tiup cluster clean ${cluster-name} --log

                                  清空集群所有服务的数据和日志:

                                    tiup cluster clean ${cluster-name} --all

                                    清空 Prometheus 以外的所有服务的日志和数据:

                                      tiup cluster clean ${cluster-name} --all --ignore-role prometheus

                                      清空节点 172.16.13.11:9000 以外的所有服务的日志和数据:

                                        tiup cluster clean ${cluster-name} --all --ignore-node 172.16.13.11:9000

                                        清空部署在 172.16.13.12 以外的所有服务的日志和数据:

                                          tiup cluster clean ${cluster-name} --all --ignore-node 172.16.13.12

                                          6、销毁集群

                                          销毁集群操作会关闭服务,清空数据目录和部署目录,并且无法恢复,需要谨慎操作

                                            tiup cluster destroy ${cluster-name}

                                            7、小插曲

                                            由于之前使用tiup playgroup安装过tidb集群,没有把tidb socket清理干净,导致后面使用tidb用户启动tidb出现权限不足错误。即:集群安装过程执行tiup cluster start sjkywzd --init出现了以下报错,tidb节点状态为Down。

                                              # tiup cluster start sjkywzd --init
                                              中间部分省略
                                              Started cluster `sjkywzd` successfully
                                              Failed to set root password of TiDB database to 'd+Qi61vhT@2*8b^A09'
                                              Error: dial tcp 192.168.198.132:4000connectconnection refused
                                              Verbose debug logs has been written to /root/.tiup/logs/tiup-cluster-debug-2024-04-26-16-08-55.log.


                                              检查防火墙处于关闭状态
                                              查看tidb进程是否存在
                                              ps -ef|grep tidb-server


                                              检查 4000端口是否被占用
                                              ss -lnp|grep 4000


                                              检查tidb运行日志
                                              cd /tidb-deploy/tidb-4000/log
                                              tail -20 tidb.log
                                              看到/tmp/tidb-4000.sock文件没权限删除。
                                              [2024/04/26 16:34:31.146 +08:00] [FATAL] [main.go:887] ["failed to create the server"]
                                              [error="failed to cleanup stale Unix socket file /tmp/tidb-4000.sock:
                                              dial unix /tmp/tidb-4000.sock: connect: permission denied"] [


                                              手动删除:rm -rf /tmp/tidb-4000.sock


                                              手动修改root密码
                                              mysql -h 192.168.198.132 -P 4000 -u root
                                              use mysql;
                                              select host,user from user;
                                              alter user 'root'@'%' identified by 'TidbAdmin51##';
                                              mysql -h 192.168.198.132 -P 4000 -u root -p

                                              常用命令:

                                              手动拉起tidb服务:tiup cluster start sjkywzd -R tidb

                                              集群环境全面检查:tiup cluster check --cluster sjkywzd

                                                [root@tidb ~]# tiup cluster check --cluster sjkywzd
                                                + Download necessary tools
                                                - Downloading check tools for linux/amd64 ... Done
                                                + Collect basic system information
                                                + Collect basic system information
                                                - Getting system info of 192.168.198.132:22 ... Done
                                                + Check time zone
                                                - Checking node 192.168.198.132 ... Done
                                                + Check system requirements
                                                + Check system requirements
                                                + Check system requirements
                                                - Checking node 192.168.198.132 ... Done
                                                - Checking node 192.168.198.132 ... Done
                                                - Checking node 192.168.198.132 ... Done
                                                - Checking node 192.168.198.132 ... Done
                                                - Checking node 192.168.198.132 ... Done
                                                - Checking node 192.168.198.132 ... Done
                                                - Checking node 192.168.198.132 ... Done
                                                - Checking node 192.168.198.132 ... Done
                                                - Checking node 192.168.198.132 ... Done
                                                + Cleanup check files
                                                - Cleanup check files on 192.168.198.132:22 ... Done
                                                Node Check Result Message
                                                ---- ----- ------ -------
                                                192.168.198.132 permission Pass /tidb-data/prometheus-9090 is writable
                                                192.168.198.132 permission Pass /tidb-data/tikv-20161 is writable
                                                192.168.198.132 permission Pass /tidb-data/tikv-20160 is writable
                                                192.168.198.132 permission Pass /tidb-deploy/tiflash-9000 is writable
                                                192.168.198.132 permission Pass /tidb-deploy/tikv-20161 is writable
                                                192.168.198.132 permission Pass /tidb-deploy/pd-2379 is writable
                                                192.168.198.132 permission Pass /tidb-deploy/prometheus-9090 is writable
                                                192.168.198.132 permission Pass /tidb-deploy/tikv-20160 is writable
                                                192.168.198.132 permission Pass /tidb-data/tikv-20162 is writable
                                                192.168.198.132 permission Pass /tidb-data/pd-2379 is writable
                                                192.168.198.132 permission Pass /tidb-deploy/grafana-3000 is writable
                                                192.168.198.132 permission Pass /tidb-data/tiflash-9000 is writable
                                                192.168.198.132 permission Pass /tidb-deploy/tikv-20162 is writable
                                                192.168.198.132 permission Pass /tidb-deploy/tidb-4000 is writable
                                                192.168.198.132 network Pass network speed of ens33 is 1000MB
                                                192.168.198.132 cpu-governor Warn Unable to determine current CPU frequency governor policy
                                                192.168.198.132 memory Pass memory size is 8192MB
                                                192.168.198.132 sysctl Fail fs.file-max = 787310, should be greater than 1000000
                                                192.168.198.132 sysctl Fail net.core.somaxconn = 128, should be greater than 32768
                                                192.168.198.132 sysctl Fail net.ipv4.tcp_syncookies = 1, should be 0
                                                192.168.198.132 sysctl Fail vm.swappiness = 30, should be 0
                                                192.168.198.132 selinux Pass SELinux is disabled
                                                192.168.198.132 command Fail numactl not usable, bash: numactl: command not found
                                                192.168.198.132 os-version Pass OS is CentOS Linux 7 (Core) 7.6.1810
                                                192.168.198.132 disk Warn mount point / does not have 'noatime' option set
                                                192.168.198.132 limits Fail soft limit of 'nofile' for user 'tidb' is not set or too low
                                                192.168.198.132 limits Fail hard limit of 'nofile' for user 'tidb' is not set or too low
                                                192.168.198.132 limits Fail soft limit of 'stack' for user 'tidb' is not set or too low
                                                192.168.198.132 thp Fail THP is enabled, please disable it for best performance
                                                192.168.198.132 cpu-cores Pass number of CPU cores / threads: 4
                                                192.168.198.132 swap Warn swap is enabled, please disable it for best performance
                                                192.168.198.132 disk Fail multiple components tikv:/tidb-data/tikv-20160,tikv:/tidb-data/tikv-20161,tikv:/tidb-data/tikv-20162,tiflash:/tidb-data/tiflash-9000 are using the same partition 192.168.198.132:/ as data dir
                                                Checking region status of the cluster sjkywzd...
                                                All regions are healthy.


                                                查找并杀掉占用进程:isof -i:2380

                                                快速修复不合规项:tiup cluster check --apply --cluster sjkywzd

                                                  [root@tidb ~]# tiup cluster check --apply --cluster sjkywzd
                                                  + Download necessary tools
                                                  - Downloading check tools for linux/amd64 ... Done
                                                  + Collect basic system information
                                                  + Collect basic system information
                                                  - Getting system info of 192.168.198.132:22 ... Done
                                                  + Check time zone
                                                  - Checking node 192.168.198.132 ... Done
                                                  + Check system requirements
                                                  + Check system requirements
                                                  + Check system requirements
                                                  - Checking node 192.168.198.132 ... Done
                                                  - Checking node 192.168.198.132 ... Done
                                                  - Checking node 192.168.198.132 ... Done
                                                  - Checking node 192.168.198.132 ... Done
                                                  - Checking node 192.168.198.132 ... Done
                                                  - Checking node 192.168.198.132 ... Done
                                                  - Checking node 192.168.198.132 ... Done
                                                  - Checking node 192.168.198.132 ... Done
                                                  - Checking node 192.168.198.132 ... Done
                                                  + Cleanup check files
                                                  - Cleanup check files on 192.168.198.132:22 ... Done
                                                  Node Check Result Message
                                                  ---- ----- ------ -------
                                                  192.168.198.132 limits Fail will try to set 'tidb soft nofile 1000000'
                                                  192.168.198.132 limits Fail will try to set 'tidb hard nofile 1000000'
                                                  192.168.198.132 limits Fail will try to set 'tidb soft stack 10240'
                                                  192.168.198.132 sysctl Fail will try to set 'fs.file-max = 1000000'
                                                  192.168.198.132 sysctl Fail will try to set 'net.core.somaxconn = 32768'
                                                  192.168.198.132 sysctl Fail will try to set 'net.ipv4.tcp_syncookies = 0'
                                                  192.168.198.132 sysctl Fail will try to set 'vm.swappiness = 0'
                                                  192.168.198.132 selinux Pass SELinux is disabled
                                                  192.168.198.132 command Fail numactl not usable, bash: numactl: command not found, auto fixing not supported
                                                  192.168.198.132 disk Warn mount point / does not have 'noatime' option set, auto fixing not supported
                                                  192.168.198.132 disk Fail multiple components tikv:/tidb-data/tikv-20160,tikv:/tidb-data/tikv-20161,tikv:/tidb-data/tikv-20162,tiflash:/tidb-data/tiflash-9000 are using the same partition 192.168.198.132:/ as data dir, auto fixing not supported
                                                  192.168.198.132 permission Pass /tidb-deploy/prometheus-9090 is writable
                                                  192.168.198.132 permission Pass /tidb-data/tiflash-9000 is writable
                                                  192.168.198.132 permission Pass /tidb-deploy/tikv-20160 is writable
                                                  192.168.198.132 permission Pass /tidb-deploy/pd-2379 is writable
                                                  192.168.198.132 permission Pass /tidb-deploy/grafana-3000 is writable
                                                  192.168.198.132 permission Pass /tidb-data/tikv-20161 is writable
                                                  192.168.198.132 permission Pass /tidb-deploy/tiflash-9000 is writable
                                                  192.168.198.132 permission Pass /tidb-data/tikv-20162 is writable
                                                  192.168.198.132 permission Pass /tidb-data/pd-2379 is writable
                                                  192.168.198.132 permission Pass /tidb-deploy/tikv-20161 is writable
                                                  192.168.198.132 permission Pass /tidb-data/tikv-20160 is writable
                                                  192.168.198.132 permission Pass /tidb-data/prometheus-9090 is writable
                                                  192.168.198.132 permission Pass /tidb-deploy/tikv-20162 is writable
                                                  192.168.198.132 permission Pass /tidb-deploy/tidb-4000 is writable
                                                  192.168.198.132 cpu-cores Pass number of CPU cores / threads: 4
                                                  192.168.198.132 memory Pass memory size is 8192MB
                                                  192.168.198.132 swap Warn will try to disable swap, please also check /etc/fstab manually
                                                  192.168.198.132 network Pass network speed of ens33 is 1000MB
                                                  192.168.198.132 thp Fail will try to disable THP, please check again after reboot
                                                  192.168.198.132 os-version Pass OS is CentOS Linux 7 (Core) 7.6.1810
                                                  192.168.198.132 cpu-governor Warn Unable to determine current CPU frequency governor policy, auto fixing not supported


                                                  + Try to apply changes to fix failed checks
                                                  + Try to apply changes to fix failed checks
                                                  - Applying changes on 192.168.198.132 ... Done
                                                  Checking region status of the cluster sjkywzd...
                                                  All regions are healthy.
                                                  8、参考链接
                                                    https://docs.pingcap.com/zh/tidb/v8.0/release-8.0.0
                                                    https://tidb.net/blog/94c2b44c
                                                    https://docs.pingcap.com/zh/tidb/dev/quick-start-with-tidb

                                                    全文完,希望可以帮到正在阅读的你,如果觉得有帮助,可以分享给你身边的朋友,同事,你关心谁就分享给谁,一起学习共同进步~~~

                                                    文章转载自数据库运维之道,如果涉嫌侵权,请发送邮件至:contact@modb.pro进行举报,并提供相关证据,一经查实,墨天轮将立刻删除相关内容。

                                                    评论