一. 下载软件

下载地址
在搜索栏中输入关键词进行搜索

选择对应的版本,点击下载(需要登录账号)

上传到/opt目录下

二. 安装

root用户登录master机器操作

#解压
[root@mdw opt]# unzip greenplum-cc-web-6.0.0-rhel7_x86_64.zip
Archive:  greenplum-cc-web-6.7.1-gp6-rhel7-x86_64.zipcreating: greenplum-cc-web-6.7.1-gp6-rhel7-x86_64/inflating: greenplum-cc-web-6.7.1-gp6-rhel7-x86_64/gpccinstall-6.0.0
[root@mdw opt]# ls
greenplum-cc-web-6.7.1-gp6-rhel7-x86_64  greenplum-cc-web-6.7.1-gp6-rhel7-x86_64.zip  greenplum-cc-web-6.7.1-gp6-rhel7-x86_64.rpm  rh
[root@mdw opt]##目录赋权限给gpadmin
[root@mdw opt]# chown -R gpadmin:gpadmin /opt/greenplum-cc-web-6.7.1-gp6-rhel7-x86_64#切换用户安装
[root@mdw greenplum-cc-web-6.7.1-gp6-rhel7-x86_64]# su - gpadmin
[gpadmin@mdw ~]$ cd /opt/greenplum-cc-web-6.7.1-gp6-rhel7-x86_64/
[gpadmin@mdw greenplum-cc-web-6.7.1-gp6-rhel7-x86_64]$ ./gpccinstall-6.0.0...
#同意
Do you agree to the Pivotal Greenplum Command Center End User License Agreement? Yy/Nn (Default=Y)
y#指定安装路径greenplum
Where would you like to install Greenplum Command Center? (Default=/usr/local)
/greenplum#默认回车
What would you like to name this installation of Greenplum Command Center? (Default=gpcc)#默认,回车
What port would you like gpcc webserver to use? (Default=28080)#输入n
Would you like enable SSL? Yy/Nn (Default=N)
n#输入2
Please choose a display language (Default=English)
1.  English
2.  Chinese
3.  Korean
4.  Russian
5.  Japanese
2INSTALLATION IN PROGRESS...
RELOADING pg_hba.conf. PLEASE WAIT ...********************************************************************************
*                                                                              *
* INSTALLATION COMPLETED SUCCESSFULLY                                          *
*                                                                              *
* Source the gpcc_path.sh or add it to your bashrc file to use gpcc command    *
* utility.                                                                     *
*                                                                              *
* To see the GPCC web UI, you must first start the GPCC webserver.             *
*                                                                              *
* To start the GPCC webserver on the current host, run gpcc start.             *
*                                                                              *
********************************************************************************To manage Command Center, use the gpcc command-line utility.
Usage:gpcc [OPTIONS] <command>Application Options:-v, --version   Show Greenplum Command Center version--settings  Print the current configuration settingsHelp Options:-h, --help      Show this help messageAvailable commands:help        Print list of commandskrbdisable  Disables kerberos authenticationkrbenable   Enables kerberos authenticationstart       Starts Greenplum Command Center webserver and metrics collection agentswith [-W] option to force password prompt for GPDB user gpmon [optional]status      Print agent statuswith [-W] option to force password prompt for GPDB user gpmon [optional]stop        Stops Greenplum Command Center webserver and metrics collection agentswith [-W] option to force password prompt for GPDB user gpmon [optional][gpadmin@mdw greenplum-cc-web-6.0.0-rhel7_x86_64]$

三. 设置环境变量

gpadmin用户在maser机器操作

cat >> /home/gpadmin/.bash_profile << EOF
source  /greenplum/greenplum-cc-web-6.0.0/gpcc_path.sh
EOF# 生效环境变量
source /home/gpadmin/.bash_profile# 新建包含standby主机,segment主机的文件,无多余的空格和换行
[gpadmin@mdw ~]$ vi /tmp/standby_seg_hosts
#添加如下内容
smdw
sdw1
sdw2保存退出#分发
gpscp -f /tmp/standby_seg_hosts ~/.bash_profile =:~

四. 修改配置

vi /data/master/gpseg-1/pg_hba.conf#添加如下内容
host    gpperfmon gpmon 0.0.0.0/0 md5#host    all     gpmon   0.0.0.0/0 trustgpscp -f /tmp/hostlist /data/master/gpseg-1/postgresql.conf =/data/master/gpseg-1/postgresql.conf[gpadmin@mdw gpseg-1]$ scp postgresql.conf gpadmin@192.168.0.113:/data/master/gpseg-1
postgresql.conf                                                                                                                                                                100%   23KB  13.8MB/s   00:00
[gpadmin@mdw gpseg-1]$ scp pg_hba.conf gpadmin@192.168.0.113:/data/master/gpseg-1
pg_hba.conf                                                                                                                                                                    100% 4956     4.1MB/s   00:00
[gpadmin@mdw gpseg-1]$

修改postgres.conf

cd /data/master/gpseg-1
vi postgresql.conf
#添加如下内容
gp_enable_gpperfmon=on
gpperfmon_port=8888
gp_external_enable_exec=on
gpperfmon_log_alert_level=warning保存退出gpscp -f /tmp/standby_seg_hosts /data/master/gpseg-1/postgresql.conf =/data/master/gpseg-1/postgresql.conf

五. 启动

gpadmin用户在master机器操作

#先刷新下配置pg_hba.conf[gpadmin@mdw gpseg-1]$ gpstop -u
20220703:03:28:57:037778 gpstop:mdw:gpadmin-[INFO]:-Starting gpstop with args: -u
20220703:03:28:57:037778 gpstop:mdw:gpadmin-[INFO]:-Gathering information and validating the environment...
20220703:03:28:57:037778 gpstop:mdw:gpadmin-[INFO]:-Obtaining Greenplum Master catalog information
20220703:03:28:57:037778 gpstop:mdw:gpadmin-[INFO]:-Obtaining Segment details from master...
20220703:03:28:57:037778 gpstop:mdw:gpadmin-[INFO]:-Greenplum Version: 'postgres (Greenplum Database) 6.18.2 build commit:1242aadf0137d3b26ee42c80e579e78bd7a805c7'
20220703:03:28:57:037778 gpstop:mdw:gpadmin-[INFO]:-Signalling all postmaster processes to reload
.# 启动
[gpadmin@mdw gpseg-1]$ gpcc start
2022-07-03 03:29:00 Starting the gpcc agents and webserver...
2022-07-03 03:29:10 Agent successfully started on 4/4 hosts
2022-07-03 03:29:10 View Greenplum Command Center at http://mdw:28080
[gpadmin@mdw gpseg-1]$# 访问
#将http://mdw:28080中的mdw替换成实际的ip地址,用浏览器打开
http://192.168.0.110:28080

gpadmin用户修改gpmon的密码

[gpadmin@mdw gpseg-1]$ psql
psql (9.4.24)
Type "help" for help.gpdb=# alter user gpmon with password 'gpmon';
ALTER ROLE
gpdb=# \q
[gpadmin@mdw gpseg-1]$

设置密码文件

格式
hostname:port:database:username:password

[gpadmin@mdw gpseg-1]$ more /home/gpadmin/.pgpass[gpadmin@mdw gpseg-1]$vi /home/gpadmin/.pgpass
将密码修改成
*:5432:gpperfmon:gpmon:gpmon# 分发
gpscp -f /tmp/hostlist /home/gpadmin/.pgpass =~

用户名密码:gpmon/gpmon,登录

六. 补充

gpdb=# select * from pg_authid;rolname       | rolsuper | rolinherit | rolcreaterole | rolcreatedb | rolcatupdate | rolcanlogin | rolreplication | rolconnlimit |             rolpassword             | rolvaliduntil | rolresqueue | rol
createrextgpfd | rolcreaterexthttp | rolcreatewextgpfd | rolresgroup
---------------------+----------+------------+---------------+-------------+--------------+-------------+----------------+--------------+-------------------------------------+---------------+-------------+----
---------------+-------------------+-------------------+-------------gpadmin             | t        | t          | t             | t           | t            | t           | t              |           -1 | md5d115c26178447dc7986c5e67afcd710b |               |        6055 | t| t                 | t                 |        6438gpcc_basic          | f        | t          | f             | f           | f            | f           | f              |           -1 |                                     |               |        6055 | f| f                 | f                 |        6437gpcc_operator       | f        | t          | f             | f           | f            | f           | f              |           -1 |                                     |               |        6055 | f| f                 | f                 |        6437gpcc_operator_basic | f        | t          | f             | f           | f            | f           | f              |           -1 |                                     |               |        6055 | f| f                 | f                 |        6437gpmon               | t        | t          | f             | t           | t            | t           | f              |           -1 | md5271fed4320b7ac3aae3ac625420cf73f |               |        6055 | f| f                 | f                 |        6438
(5 rows)gpdb=#

错误:password authentication failed for user “gpmon”

[gpadmin@mdw gpseg-1]$ gpcc stop
2022-07-03 10:40:58 Stopping the gpcc agents and webserver...
2022-07-03 10:40:58 pq: password authentication failed for user "gpmon"
[gpadmin@mdw gpseg-1]$

psql进入修改密码无效
解决:手动修改vi /home/gpadmin/.pgpass。下面命令生效

[gpadmin@mdw gpseg-1]$ gpcc stop
2022-07-03 10:55:34 Stopping the gpcc agents and webserver...
2022-07-03 10:55:35 GPCC webserver and metrics collection agents have been stopped. Use gpcc start to start them again
[gpadmin@mdw gpseg-1]$[gpadmin@mdw gpseg-1]$ gpcc start
2022-07-03 10:56:31 Starting the gpcc agents and webserver...
2022-07-03 10:56:35 Agent successfully started on 4/4 hosts
2022-07-03 10:56:35 View Greenplum Command Center at http://mdw:28080
[gpadmin@mdw gpseg-1]$

错误2:镜像mirror节点故障


检查镜像状态

[gpadmin@mdw gpseg-1]$ gpstate -m
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-Starting gpstate with args: -m
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-local Greenplum Version: 'postgres (Greenplum Database) 6.18.2 build commit:1242aadf0137d3b26ee42c80e579e78bd7a805c7'
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-master Greenplum Version: 'PostgreSQL 9.4.24 (Greenplum Database 6.18.2 build commit:1242aadf0137d3b26ee42c80e579e78bd7a805c7) on x86_64-unknown-linux-gnu, compiled by gcc (GCC) 6.4.0, 64-bit compiled on Nov 10 2021 00:05:30'
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-Obtaining Segment details from master...
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:--------------------------------------------------------------
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:--Current GPDB mirror list and status
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:--Type = Group
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:--------------------------------------------------------------
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-   Mirror   Datadir               Port    Status              Data Status
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-   sdw2     /data/mirror/gpseg0   37000   Acting as Primary   Not In Sync
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-   sdw2     /data/mirror/gpseg1   37001   Acting as Primary   Not In Sync
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-   sdw2     /data/mirror/gpseg2   37002   Acting as Primary   Not In Sync
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-   sdw2     /data/mirror/gpseg3   37003   Acting as Primary   Not In Sync
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-   sdw1     /data/mirror/gpseg4   37000   Acting as Primary   Not In Sync
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-   sdw1     /data/mirror/gpseg5   37001   Acting as Primary   Not In Sync
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-   sdw1     /data/mirror/gpseg6   37002   Acting as Primary   Not In Sync
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:-   sdw1     /data/mirror/gpseg7   37003   Acting as Primary   Not In Sync
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[INFO]:--------------------------------------------------------------
20220703:13:02:28:020990 gpstate:mdw:gpadmin-[WARNING]:-8 segment(s) configured as mirror(s) are acting as primaries
20220703:13:02:28:020990 gpst

执行重启数据库

gpstop -m -M fast
gpstart

[gpadmin@mdw ~]$ gpstop -m -M fast
20220703:13:05:15:022266 gpstop:mdw:gpadmin-[INFO]:-Starting gpstop with args: -m -M fast
20220703:13:05:15:022266 gpstop:mdw:gpadmin-[INFO]:-Gathering information and validating the environment...
20220703:13:05:15:022266 gpstop:mdw:gpadmin-[INFO]:-Obtaining Greenplum Master catalog information
20220703:13:05:15:022266 gpstop:mdw:gpadmin-[INFO]:-Obtaining Segment details from master...
20220703:13:05:15:022266 gpstop:mdw:gpadmin-[INFO]:-Greenplum Version: 'postgres (Greenplum Database) 6.18.2 build commit:1242aadf0137d3b26ee42c80e579e78bd7a805c7'Continue with master-only shutdown Yy|Nn (default=N):
> y
20220703:13:05:20:022266 gpstop:mdw:gpadmin-[INFO]:-Commencing Master instance shutdown with mode='fast'
20220703:13:05:20:022266 gpstop:mdw:gpadmin-[INFO]:-Master segment instance directory=/data/master/gpseg-1
20220703:13:05:20:022266 gpstop:mdw:gpadmin-[INFO]:-Attempting forceful termination of any leftover master process
20220703:13:05:20:022266 gpstop:mdw:gpadmin-[INFO]:-Terminating processes for segment /data/master/gpseg-1
[gpadmin@mdw ~]$
[gpadmin@mdw ~]$ gpstart
20220703:13:07:12:023074 gpstart:mdw:gpadmin-[INFO]:-Starting gpstart with args:
20220703:13:07:12:023074 gpstart:mdw:gpadmin-[INFO]:-Gathering information and validating the environment...
20220703:13:07:12:023074 gpstart:mdw:gpadmin-[INFO]:-Greenplum Binary Version: 'postgres (Greenplum Database) 6.18.2 build commit:1242aadf0137d3b26ee42c80e579e78bd7a805c7'
20220703:13:07:12:023074 gpstart:mdw:gpadmin-[INFO]:-Greenplum Catalog Version: '301908232'
20220703:13:07:12:023074 gpstart:mdw:gpadmin-[INFO]:-Starting Master instance in admin mode
20220703:13:07:12:023074 gpstart:mdw:gpadmin-[INFO]:-Obtaining Greenplum Master catalog information
20220703:13:07:12:023074 gpstart:mdw:gpadmin-[INFO]:-Obtaining Segment details from master...
20220703:13:07:12:023074 gpstart:mdw:gpadmin-[INFO]:-Setting new master era
20220703:13:07:12:023074 gpstart:mdw:gpadmin-[INFO]:-Master Started...
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-Shutting down master
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[WARNING]:-Skipping startup of segment marked down in configuration: on sdw1 directory /data/primary/gpseg0 <<<<<
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[WARNING]:-Skipping startup of segment marked down in configuration: on sdw1 directory /data/primary/gpseg1 <<<<<
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[WARNING]:-Skipping startup of segment marked down in configuration: on sdw1 directory /data/primary/gpseg2 <<<<<
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[WARNING]:-Skipping startup of segment marked down in configuration: on sdw1 directory /data/primary/gpseg3 <<<<<
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[WARNING]:-Skipping startup of segment marked down in configuration: on sdw2 directory /data/primary/gpseg4 <<<<<
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[WARNING]:-Skipping startup of segment marked down in configuration: on sdw2 directory /data/primary/gpseg5 <<<<<
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[WARNING]:-Skipping startup of segment marked down in configuration: on sdw2 directory /data/primary/gpseg6 <<<<<
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[WARNING]:-Skipping startup of segment marked down in configuration: on sdw2 directory /data/primary/gpseg7 <<<<<
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:---------------------------
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-Master instance parameters
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:---------------------------
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-Database                 = template1
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-Master Port              = 5432
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-Master directory         = /data/master/gpseg-1
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-Timeout                  = 600 seconds
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-Master standby start     = On
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:---------------------------------------
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-Segment instances that will be started
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:---------------------------------------
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-   Host   Datadir               Port    Role
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-   sdw2   /data/mirror/gpseg0   37000   Primary
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-   sdw2   /data/mirror/gpseg1   37001   Primary
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-   sdw2   /data/mirror/gpseg2   37002   Primary
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-   sdw2   /data/mirror/gpseg3   37003   Primary
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-   sdw1   /data/mirror/gpseg4   37000   Primary
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-   sdw1   /data/mirror/gpseg5   37001   Primary
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-   sdw1   /data/mirror/gpseg6   37002   Primary
20220703:13:07:13:023074 gpstart:mdw:gpadmin-[INFO]:-   sdw1   /data/mirror/gpseg7   37003   PrimaryContinue with Greenplum instance startup Yy|Nn (default=N):
> y
20220703:13:07:23:023074 gpstart:mdw:gpadmin-[INFO]:-Commencing parallel primary and mirror segment instance startup, please wait...
....
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[INFO]:-Process results...
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[INFO]:-----------------------------------------------------
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[INFO]:-   Successful segment starts                                            = 8
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[INFO]:-   Failed segment starts                                                = 0
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[WARNING]:-Skipped segment starts (segments are marked down in configuration)   = 8   <<<<<<<<
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[INFO]:-----------------------------------------------------
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[INFO]:-Successfully started 8 of 8 segment instances, skipped 8 other segments
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[INFO]:-----------------------------------------------------
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[WARNING]:-****************************************************************************
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[WARNING]:-There are 8 segment(s) marked down in the database
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[WARNING]:-To recover from this current state, review usage of the gprecoverseg
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[WARNING]:-management utility which will recover failed segment instance databases.
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[WARNING]:-****************************************************************************
20220703:13:07:27:023074 gpstart:mdw:gpadmin-[INFO]:-Starting Master instance mdw directory /data/master/gpseg-1
20220703:13:07:28:023074 gpstart:mdw:gpadmin-[INFO]:-Command pg_ctl reports Master mdw instance active
20220703:13:07:28:023074 gpstart:mdw:gpadmin-[INFO]:-Connecting to dbname='template1' connect_timeout=15
20220703:13:07:28:023074 gpstart:mdw:gpadmin-[INFO]:-Starting standby master
20220703:13:07:28:023074 gpstart:mdw:gpadmin-[INFO]:-Checking if standby master is running on host: smdw  in directory: /data/master/gpseg-1
20220703:13:07:30:023074 gpstart:mdw:gpadmin-[WARNING]:-Number of segments not attempted to start: 8
20220703:13:07:30:023074 gpstart:mdw:gpadmin-[INFO]:-Check status of database with gpstate utility
[gpadmin@mdw ~]$

重启后仍然失败

七. 启动后mirro失败修复

修复方式一(无效)

修复segment节点,需要重启greenplum集群

将primary正常的拷贝到对应失败的mirror中

1.停止集群

gpadmin用户操作master机器

#停止集群
gpstop -M fast

#####2.记录节点信息

#看下信息
gpstart -m
gpstate -c

3. 停止集群,拷贝数据

#停止
gpstop -M fast
输入y
# 将primary的数据拷贝到对应的mirror
# 依次执行
scp -r sdw1:/data/primary/gpseg0  sdw2:/data/mirror
# 由于显示内容太多,-q静默方式
scp -r sdw1:/data/primary/gpseg0  sdw2:/data/mirror
scp -r sdw1:/data/primary/gpseg1  sdw2:/data/mirror
scp -r sdw1:/data/primary/gpseg2  sdw2:/data/mirror
scp -r sdw1:/data/primary/gpseg3  sdw2:/data/mirror
scp -r sdw2:/data/primary/gpseg4  sdw1:/data/mirror
scp -r sdw2:/data/primary/gpseg5  sdw1:/data/mirror
scp -r sdw2:/data/primary/gpseg6  sdw1:/data/mirror
scp -r sdw2:/data/primary/gpseg7  sdw1:/data/mirror

很多文件
-q 参数可以静默方式,无效

可以看到segment服务器mirror目录,拷贝后文件夹时间已经变动

4.仅启动master

# master only模式
gpstart -m[gpadmin@mdw ~]$ PGOPTIONS='-c gp_session_role=utility' psql postgrespostgres=# set allow_system_table_mods='true';
psql (9.4.24)
Type "help" for help.postgres=# create table gp_segment_configuration_bak_0703 as select * from gp_segment_configuration;
update gp_segment_configuration set mode =  's' where mode <>'s';
update gp_segment_configuration set status ='u' where status <> 'u';\q#关闭master
[gpadmin@mdw ~]$ gpstop -m -M fast




修改前

修改后

6.启动集群

#启动
gpstart -a

7.查看状态

#查看状态
gpstate


仍然不同步

8. 监控

[gpadmin@mdw ~]$ gpcc stop
2022-07-03 13:56:46 Stopping the gpcc agents and webserver...
2022-07-03 13:56:48 GPCC webserver and metrics collection agents have been stopped. Use gpcc start to start them again
[gpadmin@mdw ~]$ gpcc start
2022-07-03 13:56:51 Starting the gpcc agents and webserver...
2022-07-03 13:56:56 Agent successfully started on 4/4 hosts
2022-07-03 13:56:56 View Greenplum Command Center at http://mdw:28080
[gpadmin@mdw ~]$

仍然错误

9. gprecoverseg修复

执行gprecoverseg命令


[gpadmin@mdw ~]$ gprecoverseg
20220703:14:23:41:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Starting gprecoverseg with args:
20220703:14:23:41:055380 gprecoverseg:mdw:gpadmin-[INFO]:-local Greenplum Version: 'postgres (Greenplum Database) 6.18.2 build commit:1242aadf0137d3b26ee42c80e579e78bd7a805c7'
20220703:14:23:41:055380 gprecoverseg:mdw:gpadmin-[INFO]:-master Greenplum Version: 'PostgreSQL 9.4.24 (Greenplum Database 6.18.2 build commit:1242aadf0137d3b26ee42c80e579e78bd7a805c7) on x86_64-unknown-linux-gnu, compiled by gcc (GCC) 6.4.0, 64-bit compiled on Nov 10 2021 00:05:30'
20220703:14:23:41:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Obtaining Segment details from master...
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Heap checksum setting is consistent between master and the segments that are candidates for recoverseg
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Greenplum instance recovery parameters
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Recovery type              = Standard
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Recovery 1 of 8
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Synchronization mode                 = Incremental
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance host                 = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance address              = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance directory            = /data/primary/gpseg0
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance port                 = 36000
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance host        = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance address     = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance directory   = /data/mirror/gpseg0
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance port        = 37000
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Target                      = in-place
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Recovery 2 of 8
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Synchronization mode                 = Incremental
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance host                 = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance address              = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance directory            = /data/primary/gpseg1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance port                 = 36001
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance host        = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance address     = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance directory   = /data/mirror/gpseg1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance port        = 37001
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Target                      = in-place
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Recovery 3 of 8
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Synchronization mode                 = Incremental
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance host                 = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance address              = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance directory            = /data/primary/gpseg2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance port                 = 36002
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance host        = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance address     = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance directory   = /data/mirror/gpseg2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance port        = 37002
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Target                      = in-place
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Recovery 4 of 8
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Synchronization mode                 = Incremental
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance host                 = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance address              = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance directory            = /data/primary/gpseg3
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance port                 = 36003
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance host        = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance address     = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance directory   = /data/mirror/gpseg3
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance port        = 37003
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Target                      = in-place
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Recovery 5 of 8
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Synchronization mode                 = Incremental
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance host                 = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance address              = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance directory            = /data/primary/gpseg4
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance port                 = 36000
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance host        = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance address     = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance directory   = /data/mirror/gpseg4
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance port        = 37000
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Target                      = in-place
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Recovery 6 of 8
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Synchronization mode                 = Incremental
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance host                 = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance address              = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance directory            = /data/primary/gpseg5
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance port                 = 36001
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance host        = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance address     = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance directory   = /data/mirror/gpseg5
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance port        = 37001
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Target                      = in-place
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Recovery 7 of 8
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Synchronization mode                 = Incremental
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance host                 = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance address              = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance directory            = /data/primary/gpseg6
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance port                 = 36002
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance host        = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance address     = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance directory   = /data/mirror/gpseg6
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance port        = 37002
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Target                      = in-place
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Recovery 8 of 8
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Synchronization mode                 = Incremental
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance host                 = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance address              = sdw2
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance directory            = /data/primary/gpseg7
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Failed instance port                 = 36003
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance host        = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance address     = sdw1
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance directory   = /data/mirror/gpseg7
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Source instance port        = 37003
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:-   Recovery Target                      = in-place
20220703:14:23:42:055380 gprecoverseg:mdw:gpadmin-[INFO]:----------------------------------------------------------Continue with segment recovery procedure Yy|Nn (default=N):
> y
20220703:14:23:51:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Starting to create new pg_hba.conf on primary segments
20220703:14:23:56:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Successfully modified pg_hba.conf on primary segments to allow replication connections
20220703:14:23:56:055380 gprecoverseg:mdw:gpadmin-[INFO]:-8 segment(s) to recover
20220703:14:23:56:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Ensuring 8 failed segment(s) are stopped
20220703:14:23:57:055380 gprecoverseg:mdw:gpadmin-[INFO]:-57638: /data/primary/gpseg0
20220703:14:23:58:055380 gprecoverseg:mdw:gpadmin-[INFO]:-57639: /data/primary/gpseg1
20220703:14:24:00:055380 gprecoverseg:mdw:gpadmin-[INFO]:-57636: /data/primary/gpseg2
20220703:14:24:01:055380 gprecoverseg:mdw:gpadmin-[INFO]:-57633: /data/primary/gpseg3
20220703:14:24:03:055380 gprecoverseg:mdw:gpadmin-[INFO]:-54324: /data/primary/gpseg4
20220703:14:24:04:055380 gprecoverseg:mdw:gpadmin-[INFO]:-54325: /data/primary/gpseg5
20220703:14:24:05:055380 gprecoverseg:mdw:gpadmin-[INFO]:-54310: /data/primary/gpseg6
20220703:14:24:07:055380 gprecoverseg:mdw:gpadmin-[INFO]:-54306: /data/primary/gpseg7
.
20220703:14:24:08:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Ensuring that shared memory is cleaned up for stopped segments
20220703:14:24:08:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Updating configuration with new mirrors
20220703:14:24:08:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Updating mirrors
20220703:14:24:08:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Running pg_rewind on failed segments
sdw1 (dbid 2): no rewind required
sdw1 (dbid 3): no rewind required
sdw1 (dbid 4): no rewind required
sdw1 (dbid 5): no rewind required
sdw2 (dbid 6): no rewind required
sdw2 (dbid 7): no rewind required
sdw2 (dbid 8): no rewind required
sdw2 (dbid 9): no rewind required
20220703:14:24:13:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Starting mirrors
20220703:14:24:13:055380 gprecoverseg:mdw:gpadmin-[INFO]:-era is 8ee3ebb3041964e3_220703142102
20220703:14:24:13:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Commencing parallel segment instance startup, please wait...
......
20220703:14:24:19:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Process results...
20220703:14:24:19:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Triggering FTS probe
20220703:14:24:20:055380 gprecoverseg:mdw:gpadmin-[INFO]:-********************************
20220703:14:24:20:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Segments successfully recovered.
20220703:14:24:20:055380 gprecoverseg:mdw:gpadmin-[INFO]:-********************************
20220703:14:24:20:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Recovered mirror segments need to sync WAL with primary segments.
20220703:14:24:20:055380 gprecoverseg:mdw:gpadmin-[INFO]:-Use 'gpstate -e' to check progress of WAL sync remaining bytes
[gpadmin@mdw ~]$

10. 查看状态

gpstate -m
此时还是未同步

重启

gpstop -M fast
gpstart

修复方式二 (无效)

1. 找到对应的故障节点

gpstate -m

2. 登录对应节点集群,停止对应的实例

如gpadmin用户登录sdw1

pg_ctl -D /data/mirror/gpseg4 stop -m fast
pg_ctl -D /data/mirror/gpseg5 stop -m fast
pg_ctl -D /data/mirror/gpseg6 stop -m fast
pg_ctl -D /data/mirror/gpseg7 stop -m fast


如gpadmin用户登录sdw2

pg_ctl -D /data/mirror/gpseg0 stop -m fast
pg_ctl -D /data/mirror/gpseg1 stop -m fast
pg_ctl -D /data/mirror/gpseg2 stop -m fast
pg_ctl -D /data/mirror/gpseg3 stop -m fast

3. 查看状态

gpstate

gpstate -e

4. 关闭集群

[gpadmin@mdw ~]$ gpstop -a -M fast

5. 启动

# restrict方式
gpstart -a -R

6. 修复故障

必须注意数据库恢复过程中主备节点切换问题

gprecoverseg -a

6.1 错误Failed to start segment

20220703:15:10:36:071273 gprecoverseg:mdw:gpadmin-[WARNING]:-Failed to start segment.  The fault prober will shortly mark it as down. Segment: sdw1:/data/primary/gpseg2:content=2:dbid=4:role=m:preferred_role=p:mode=n:status=d: REASON: PG_CTL failed.
20220703:15:10:36:071273 gprecoverseg:mdw:gpadmin-[WARNING]:-Failed to start segment.  The fault prober will shortly mark it as down. Segment: sdw1:/data/primary/gpseg0:content=0:dbid=2:role=m:preferred_role=p:mode=n:status=d: REASON: PG_CTL failed.
20220703:15:10:36:071273 gprecoverseg:mdw:gpadmin-[WARNING]:-Failed to start segment.  The fault prober will shortly mark it as down. Segment: sdw1:/data/primary/gpseg1:content=1:dbid=3:role=m:preferred_role=p:mode=n:status=d: REASON: PG_CTL failed.
20220703:15:10:36:071273 gprecoverseg:mdw:gpadmin-[WARNING]:-Failed to start segment.  The fault prober will shortly mark it as down. Segment: sdw1:/data/primary/gpseg3:content=3:dbid=5:role=m:preferred_role=p:mode=n:status=d: REASON: PG_CTL failed.
20220703:15:10:36:071273 gprecoverseg:mdw:gpadmin-[WARNING]:-Failed to start segment.  The fault prober will shortly mark it as down. Segment: sdw2:/data/primary/gpseg4:content=4:dbid=6:role=m:preferred_role=p:mode=n:status=d: REASON: PG_CTL failed.
20220703:15:10:36:071273 gprecoverseg:mdw:gpadmin-[WARNING]:-Failed to start segment.  The fault prober will shortly mark it as down. Segment: sdw2:/data/primary/gpseg7:content=7:dbid=9:role=m:preferred_role=p:mode=n:status=d: REASON: PG_CTL failed.
20220703:15:10:36:071273 gprecoverseg:mdw:gpadmin-[WARNING]:-Failed to start segment.  The fault prober will shortly mark it as down. Segment: sdw2:/data/primary/gpseg6:content=6:dbid=8:role=m:preferred_role=p:mode=n:status=d: REASON: PG_CTL failed.
20220703:15:10:36:071273 gprecoverseg:mdw:gpadmin-[WARNING]:-Failed to start segment.  The fault prober will shortly mark it as down. Segment: sdw2:/data/primary/gpseg5:content=5:dbid=7:role=m:preferred_role=p:mode=n:status=d: REASON: PG_CTL failed.

6.2 找失败的segment


查看是否存在backup_label文件,存在可以删除

[gpadmin@sdw2 mirror]$ cd /data/mirror
[gpadmin@sdw2 mirror]$ ls gpseg*/backup*
gpseg0/backup_label.old  gpseg1/backup_label.old  gpseg2/backup_label.old  gpseg3/backup_label.old

7. 重启

cd ~gpstop -a -rgprecoverseg -o ./recover_file
gprecoverseg -i ./recover_file#恢复完成后交换
gprecoverseg  -r
gpstate -m

SQL

1.查询gp_segment_configuration

[gpadmin@mdw ~]$ psql -c "SELECT * FROM gp_segment_configuration";dbid | content | role | preferred_role | mode | status | port  | hostname | address |       datadir
------+---------+------+----------------+------+--------+-------+----------+---------+----------------------18 |      -1 | m    | m              | s    | u      |  5432 | smdw     | smdw    | /data/master/gpseg-11 |      -1 | p    | p              | s    | u      |  5432 | mdw      | mdw     | /data/master/gpseg-112 |       2 | p    | m              | n    | u      | 37002 | sdw2     | sdw2    | /data/mirror/gpseg24 |       2 | m    | p              | n    | d      | 36002 | sdw1     | sdw1    | /data/primary/gpseg210 |       0 | p    | m              | n    | u      | 37000 | sdw2     | sdw2    | /data/mirror/gpseg02 |       0 | m    | p              | n    | d      | 36000 | sdw1     | sdw1    | /data/primary/gpseg011 |       1 | p    | m              | n    | u      | 37001 | sdw2     | sdw2    | /data/mirror/gpseg13 |       1 | m    | p              | n    | d      | 36001 | sdw1     | sdw1    | /data/primary/gpseg113 |       3 | p    | m              | n    | u      | 37003 | sdw2     | sdw2    | /data/mirror/gpseg35 |       3 | m    | p              | n    | d      | 36003 | sdw1     | sdw1    | /data/primary/gpseg314 |       4 | p    | m              | n    | u      | 37000 | sdw1     | sdw1    | /data/mirror/gpseg46 |       4 | m    | p              | n    | d      | 36000 | sdw2     | sdw2    | /data/primary/gpseg415 |       5 | p    | m              | n    | u      | 37001 | sdw1     | sdw1    | /data/mirror/gpseg57 |       5 | m    | p              | n    | d      | 36001 | sdw2     | sdw2    | /data/primary/gpseg516 |       6 | p    | m              | n    | u      | 37002 | sdw1     | sdw1    | /data/mirror/gpseg68 |       6 | m    | p              | n    | d      | 36002 | sdw2     | sdw2    | /data/primary/gpseg617 |       7 | p    | m              | n    | u      | 37003 | sdw1     | sdw1    | /data/mirror/gpseg79 |       7 | m    | p              | n    | d      | 36003 | sdw2     | sdw2    | /data/primary/gpseg7
(18 rows)

Greenplum6.x监控软件搭建相关推荐

  1. 企业级监控软件Zabbix搭建部署之使用mutt+msmtp配置Zabbix邮件报警

    企业级监控软件Zabbix搭建部署之使用mutt+msmtp配置Zabbix邮件报警 [root@Zabbix-Server ~]# wget http://nchc.dl.sourceforge.n ...

  2. Linux监控平台搭建( zabbix监控)

    2019独角兽企业重金招聘Python工程师标准>>> Linux监控平台搭建( zabbix监控) 一.Linux监控平台介绍 1.监控存在的原因 站点出了问题,没有人知道,等用户 ...

  3. android wifi视频监控软件,WiFi环境下Android智能视频监控系统研究与实现

    摘要: 在互联网飞速发展和移动互联网强势崛起的时代,科技产品服务于普通生活是新兴行业必然的发展趋势;监控系统是物联网时代各个领域必然争取的可控制系统.随着无线技术和移动终端设备的高歌猛进,移动终端智能 ...

  4. 【最佳实践】分布式数据库对接开源监控软件

    Grafana+InfluxDB+Telegraf+SequoiaDB Grafana作为一款强大的开源的监控软件,可以进行灵活的报表定制与性能监控.用户可以通过时序数据库InfluxDB作为数据源为 ...

  5. 京东价格监控软件开发技术探讨十四:电商领域区块链融合模式采集的探讨

    区块链及人工智能是这两年比较火的概念,在和圈内的朋友讨论时,就有如何将区块链融入电商项目中的探讨. 在开始之前,我们需要先明白区块链是什么? 说起区块链,我们首先想起的肯定是比特币.以太币这些炒的火热 ...

  6. Linux监控平台搭建Zabbix(资源)

    2019独角兽企业重金招聘Python工程师标准>>> 自动化监控: 1.cobbler实现自动装机 2.saltstack实现工程自动化配置 3.kubernetes实现容器自动化 ...

  7. 免杀远程监控软件大盘点

    谈到网络安全问题,就没法不谈***联盟.而说到***,当然离不开***联盟至爱的远程监控技术.世界上的远程控制软件多如牛毛,远程控制控水平也参差不齐.远程监控水平低的达不到用户的心理需求,远程监控水平 ...

  8. Linux下常用的系统监控软件

    转载:http://blog.chinaunix.net/uid-25266990-id-3380929.html Linux下常用的系统监控软件有Nagios.Cacti.Zabbix.Monit等 ...

  9. lepus天兔数据库监控系搭建

    一.lepus简介 简洁.直观.强大的开源数据库监控系统,MySQL/Oracle/MongoDB/Redis一站式性能监控,让数据库监控更简单.官网:http://www.lepus.cc/ 二.l ...

最新文章

  1. 独家 | 在浏览器中使用TensorFlow.js和Python构建机器学习模型(附代码)
  2. ewebeditor 5.2 列目录漏洞
  3. socket如何定义端口号才能避免和其他程序冲突?
  4. webpack打包css文件
  5. Oracle 生成随机密码
  6. 优雅地记录Python程序日志2:模块组件化日志记录器
  7. put url带参数_避免自己写的 url 被diss!建议看看这篇RESTful API简明教程!
  8. 学习笔记:区块链概念入门
  9. GCC: libgcc的用途以及交叉编译
  10. 95-136-041-源码-Operator-AbstractStreamOperator
  11. LeetCode题目总结-滑窗法
  12. 最大熵模型The Maximum Entropy:学习
  13. 计算机组成原理完整学习笔记(二):系统总线
  14. BeEF-XSS详细使用教程
  15. Transform.up与Vector3.up
  16. Android自定义星星评分控件
  17. elasticsearch学习之字段统计
  18. Elasticsearch 跨机房灾备方案实战(一) —— 消息队列实现双写
  19. 算法的时间复杂度的计算
  20. subline修改html模板,sublime Text 3 emmet自定义HTML模板的方法

热门文章

  1. 技术分享 | binlog 实用解析工具 my2sql
  2. SAP FI CO模块常用事务代码
  3. 第三只眼看财务-留存收益与职业道德
  4. 键盘哪个键是锁定计算机,笔记本键盘锁定键在哪_笔记本电脑的“键盘锁”是哪一个键-win7之家...
  5. 斗音视频制作技巧和分享技巧
  6. 运用awk提取日志文件中的IP地址
  7. Latex同时合并表格的多行多列
  8. 网易云信 UI 开发
  9. 线性规划-pulp-复杂矩阵
  10. 信息系统项目管理师必背核心考点(十)信息系统规划