GaussDB T 1.0.2分布式集群部署故障总结

之前安装GaussDB T 1.0.2分布式集群的时候,安装过程中会报segmentation fault错误,如下:

[[email protected] ~]$ gs_install -X /mnt/Huawei/db/clusterconfig.xml
Parsing the configuration file.
Check preinstall on every node.
Successfully checked preinstall on every node.
Creating the backup directory.
Successfully created the backup directory.
Check the time difference between hosts in the cluster.
Installing the cluster.
Check install cluster condition.
Successfully check install cluster condition.
Installing applications on all nodes.
Successfully installed APP.
Initializing cluster instances
Initializing ETCD instance.
Distribute etcd communication keys.
Successfully distribute etcd communication keys.
Check the status of ETCD cluster.
Successfully initialize ETCD instance.
Segmentation fault

经分析排查,原因是在安装Python3的时候,没有指定目录,导致它的一些依赖库和系统自带的Python2库文件冲突导致的。使用下面的方法在各个节点重新安装Python3即可:

[[email protected] ~]# mkdir /usr/local/python373
[[email protected] ~]# tar -xzf /mnt/Python-3.7.3.tgz
[[email protected] ~]# cd Python-3.7.3/
[[email protected] ~]# ./configure --prefix=/usr/local/python373&& make && make install
编辑/etc/profile,加入以下内容:
[[email protected] ~]# vi /etc/profile
export PATH=${PATH}:/usr/local/python373/bin
[[email protected] ~]# source /etc/profile
[[email protected] ~]# python3 -V
Python 3.7.3

再次执行gs_install脚本,如下:

[[email protected] ~]$ gs_install -X /tmp/clusterconfig3C3D.xml
Parsing the configuration file.
Check preinstall on every node.
Successfully checked preinstall on every node.
Creating the backup directory.
Successfully created the backup directory.
Check the time difference between hosts in the cluster.
Installing the cluster.
Check install cluster condition.
Successfully check install cluster condition.
Installing applications on all nodes.
Successfully installed APP.
Initializing cluster instances
Initializing ETCD instance.
Distribute etcd communication keys.
Successfully distribute etcd communication keys.
Check the status of ETCD cluster.
Successfully initialize ETCD instance.
Initializing database instance.
..........3130s
Initializing cluster instances is completed.
.Configuring standby datanode.
...............2475s
Successfully configure datanode.
.Stop database instance.
......36s
Successfully stop database instance.
.Configuring.
Configuring the cluster.
Successfully configuring the cluster.
Configuration is completed.
Load cluster configuration file.
Start cm agent.
Successfully start cm agent and ETCD in cluster.
Warning: Auto failover switch closed.
Starting the cluster.
==============================================
.........39s
Successfully starting the cluster.
==============================================
.Registration settings have been skipped.
[[email protected] ~]$ gs_om -t status
Set output to terminal.
--------------------------------------------------------------------Cluster Status--------------------------------------------------------------------
az_state :      single_az
cluster_state : Normal
balanced :      true
----------------------------------------------------------------------AZ Status-----------------------------------------------------------------------
AZ:AZ1                ROLE:primary            STATUS:ONLINE
---------------------------------------------------------------------Host Status----------------------------------------------------------------------
HOST:hwd01            AZ:AZ1                  STATUS:ONLINE       IP:192.168.120.22
HOST:hwd02            AZ:AZ1                  STATUS:ONLINE       IP:192.168.120.23
HOST:hwd03            AZ:AZ1                  STATUS:ONLINE       IP:192.168.120.24
----------------------------------------------------------------Cluster Manager Status----------------------------------------------------------------
INSTANCE:CM1          ROLE:primary            STATUS:ONLINE       HOST:hwd01            ID:601
INSTANCE:CM2          ROLE:slave              STATUS:ONLINE       HOST:hwd02            ID:602
INSTANCE:CM3          ROLE:slave              STATUS:ONLINE       HOST:hwd03            ID:603
---------------------------------------------------------------------ETCD Status----------------------------------------------------------------------
INSTANCE:ETCD1        ROLE:leader             STATUS:ONLINE       HOST:hwd01            ID:701      PORT:2379         DataDir:/opt/huawei/gaussdb/data_etcd1/data
INSTANCE:ETCD2        ROLE:follower           STATUS:ONLINE       HOST:hwd02            ID:702      PORT:2379         DataDir:/opt/huawei/gaussdb/data_etcd1/data
INSTANCE:ETCD3        ROLE:follower           STATUS:ONLINE       HOST:hwd03            ID:703      PORT:2379         DataDir:/opt/huawei/gaussdb/data_etcd1/data
----------------------------------------------------------------------CN Status-----------------------------------------------------------------------
INSTANCE:cn_401       ROLE:no role            STATUS:ONLINE       HOST:hwd01            ID:401      PORT:8000         DataDir:/opt/huawei/gaussdb/data/data_cn
INSTANCE:cn_402       ROLE:no role            STATUS:ONLINE       HOST:hwd02            ID:402      PORT:8000         DataDir:/opt/huawei/gaussdb/data/data_cn
INSTANCE:cn_403       ROLE:no role            STATUS:ONLINE       HOST:hwd03            ID:403      PORT:8000         DataDir:/opt/huawei/gaussdb/data/data_cn
----------------------------------------------------------------------GTS Status----------------------------------------------------------------------
INSTANCE:GTS1         ROLE:primary            STATUS:ONLINE       HOST:hwd01            ID:441      PORT:7000         DataDir:/opt/huawei/gaussdb/data/gts
INSTANCE:GTS2         ROLE:standby            STATUS:ONLINE       HOST:hwd02            ID:442      PORT:7000         DataDir:/opt/huawei/gaussdb/data/gts
---------------------------------------------------------Instances Status in Group (group_1)----------------------------------------------------------
INSTANCE:DB1_1        ROLE:primary            STATUS:ONLINE       HOST:hwd01            ID:1        PORT:40000        DataDir:/opt/huawei/gaussdb/data_db/dn1
INSTANCE:DB1_2        ROLE:standby            STATUS:ONLINE       HOST:hwd02            ID:2        PORT:40042        DataDir:/opt/huawei/gaussdb/data_db/dn1
INSTANCE:DB1_3        ROLE:standby            STATUS:ONLINE       HOST:hwd03            ID:3        PORT:40021        DataDir:/opt/huawei/gaussdb/data_db/dn1
---------------------------------------------------------Instances Status in Group (group_2)----------------------------------------------------------
INSTANCE:DB2_6        ROLE:standby            STATUS:ONLINE       HOST:hwd01            ID:6        PORT:40021        DataDir:/opt/huawei/gaussdb/data_db/dn2
INSTANCE:DB2_4        ROLE:primary            STATUS:ONLINE       HOST:hwd02            ID:4        PORT:40000        DataDir:/opt/huawei/gaussdb/data_db/dn2
INSTANCE:DB2_5        ROLE:standby            STATUS:ONLINE       HOST:hwd03            ID:5        PORT:40042        DataDir:/opt/huawei/gaussdb/data_db/dn2
---------------------------------------------------------Instances Status in Group (group_3)----------------------------------------------------------
INSTANCE:DB3_8        ROLE:standby            STATUS:ONLINE       HOST:hwd01            ID:8        PORT:40042        DataDir:/opt/huawei/gaussdb/data_db/dn3
INSTANCE:DB3_9        ROLE:standby            STATUS:ONLINE       HOST:hwd02            ID:9        PORT:40021        DataDir:/opt/huawei/gaussdb/data_db/dn3
INSTANCE:DB3_7        ROLE:primary            STATUS:ONLINE       HOST:hwd03            ID:7        PORT:40000        DataDir:/opt/huawei/gaussdb/data_db/dn3
-----------------------------------------------------------------------Manage IP----------------------------------------------------------------------
HOST:hwd01            IP:192.168.120.22
HOST:hwd02            IP:192.168.120.23
HOST:hwd03            IP:192.168.120.24
-------------------------------------------------------------------Query Action Info------------------------------------------------------------------
HOSTNAME: hwd01     TIME: 2020-04-10 07:40:20.544385
------------------------------------------------------------------------Float Ip------------------------------------------------------------------
HOST:hwd01    DB1_1:192.168.120.22    IP:
HOST:hwd02    DB2_4:192.168.120.23    IP:
HOST:hwd03    DB3_7:192.168.120.24    IP:

原文地址:https://blog.51cto.com/candon123/2486415

时间: 2024-10-25 15:38:45

GaussDB T 1.0.2分布式集群部署故障总结的相关文章

Kubernetes+Etcd-v1.7.0 + CA 分布式集群部署

kubernetes 1.7.0 + flannel 二进制部署   kubernetes 1.7.0 + flannel 基于 二进制 文件部署 本地化 kube-apiserver, kube-controller-manager , kube-scheduler (1).环境说明 k8s-master-1: 192.168.54.12 k8s-node1: 192.168.54.13 k8s-node2:   192.168.54.14 (2).初始化环境 hostnamectl --st

GaussDB T分布式集群部署以及升级指南

本文用四节点部署GaussDB T 1.0.1分布式集群,部署完成后再将其升级到1.0.2版本(直接安装1.0.2版本,在安装过程中会遇到segment fault报错,目前尚未解决).前期操作系统准备工作参考之前的几篇文章. 1.部署分布式集群 1.1 节点信息 各节点信息如下表所示: 1.2 集群参数文件 根据实际情况修改集群参数,或者通过database manager工具生成,内容如下: [[email protected] db]# vi clusterconfig.xml <?xml

redis3.0.2 分布式集群安装详细步骤

redis3.0.2 分布式集群安装详细步骤 --(centos5.8 X64系统) 版本历史 时间 版本 说明 编写者 2015-06-5 1.0 redis3.0.2 分布式集群安装详细步骤 csc 一: redis cluster介绍篇 1:redis cluster的现状 目前redis支持的cluster特性(已亲测): 1):节点自动发现 2):slave->master 选举,集群容错 3):Hot resharding:在线分片 4):进群管理:cluster xxx 5):基于

solr 集群(SolrCloud 分布式集群部署步骤)

SolrCloud 分布式集群部署步骤 安装软件包准备 apache-tomcat-7.0.54 jdk1.7 solr-4.8.1 zookeeper-3.4.5 注:以上软件都是基于 Linux 环境的 64位 软件,以上软件请到各自的官网下载. 服务器准备 为搭建这个集群,准备三台服务器,分别为 192.168.0.2 -- master 角色192.168.0.3 -- slave 角色192.168.0.4 -- slave 角色 搭建基础环境 安装 jdk1.7 - 这个大家都会安装

solrCloud 4.9 分布式集群部署及注意事项

环境搭建 一.zookeeper 参考:http://blog.chinaunix.net/uid-25135004-id-4214399.html 现有4台机器 10.14.2.201 10.14.2.202 10.14.2.203 10.14.2.204 安装zookeeper集群 在所有机器上进行 1.下载安装包解压 tar xvf zookeeper-3.4.5.tar.gz -C /export/ cd /export/ ln -s zookeeper-3.4.5 zookeeper

超详细从零记录Hadoop2.7.3完全分布式集群部署过程

超详细从零记录Ubuntu16.04.1 3台服务器上Hadoop2.7.3完全分布式集群部署过程.包含,Ubuntu服务器创建.远程工具连接配置.Ubuntu服务器配置.Hadoop文件配置.Hadoop格式化.启动.(首更时间2016年10月27日) 主机名/hostname IP 角色 hadoop1 192.168.193.131 ResourceManager/NameNode/SecondaryNameNode hadoop2 192.168.193.132 NodeManager/

Kubernetes v1.14.0 之 kube-apiserver集群部署

kube-apiserver集群准备 1.kube-apiserver 服务器配置 对外ip 内网ip cpu 内存 硬盘 192.168.3.10 172.172.1.1 64 256 1T 192.168.3.11 172.172.1.2 64 256 1T 192.168.3.12 172.172.1.3 64 256 1T 192.168.3.13 172.172.1.4 64 256 1T 192.168.3.14 172.172.1.5 64 256 1T 2.kube-apiser

分布式实时日志系统(四) 环境搭建之centos 6.4下hbase 1.0.1 分布式集群搭建

一.hbase简介 HBase是一个开源的非关系型分布式数据库(NoSQL),它参考了谷歌的BigTable建模,实现的编程语言为 Java.它是Apache软件基金会的Hadoop项目的一部分,运行于HDFS文件系统之上,为 Hadoop 提供类似于BigTable 规模的服务.因此,它可以容错地存储海量稀疏的数据.HBase在列上实现了BigTable论文提到的压缩算法.内存操作和布隆过滤器.HBase的表能够作为MapReduce任务的输入和输出,可以通过Java API来存取数据,也可以

hadoop-2.7.1+zookeeper-3.4.8+hbase-1.2.1+apache-hive-2.0.0完全分布式集群

网上有的hadoop,zookeeper,hbase,apache-hive相关文档,具体大家可以百度具体了解,本文不多做介绍,本文从最基本的环境搭建入手,纯环境搭建. 搭建次环境唯一要考虑就是软件相互之间的兼容性,及服务器内存(最好大点,虚拟机2G以上),当然也可以通过调启动参数,调小启动内存大小.具体文章会适当介绍. 一.环境准备 1.查看系统版本 cat /etc/issue CentOS release 6.8 (Final) 2.软件准备 软件自行官网下载,但自己要统一规划方便后面操作