Cannot use /dev/sdc: device is rejected by filter config报错处理
找到lvm配置文件:vim /etc/lvm/lvm.conf,搜索找到devices字段中filter
# Configuration option devices/issue_discards.# Issue discards to PVs that are no longer used by an LV.# Discards are sent to an LV's underlying physical volumes when the LV# is no longer using the physical volumes' space, e.g. lvremove,# lvreduce. Discards inform the storage that a region is no longer# used. Storage that supports discards advertise the protocol-specific# way discards should be issued by the kernel (TRIM, UNMAP, or# WRITE SAME with UNMAP bit set). Not all storage will support or# benefit from discards, but SSDs and thinly provisioned LUNs# generally do. If enabled, discards will only be issued if both the# storage and kernel provide support.issue_discards = 0# Configuration option devices/allow_changes_with_duplicate_pvs.# Allow VG modification while a PV appears on multiple devices.# When a PV appears on multiple devices, LVM attempts to choose the# best device to use for the PV. If the devices represent the same# underlying storage, the choice has minimal consequence. If the# devices represent different underlying storage, the wrong choice# can result in data loss if the VG is modified. Disabling this# setting is the safest option because it prevents modifying a VG# or activating LVs in it while a PV appears on multiple devices.# Enabling this setting allows the VG to be used as usual even with# uncertain devices.allow_changes_with_duplicate_pvs = 0# Configuration option devices/allow_mixed_block_sizes.# Allow PVs in the same VG with different logical block sizes.# When allowed, the user is responsible to ensure that an LV is# using PVs with matching block sizes when necessary.allow_mixed_block_sizes = 0
#filter = ["a|^/dev/disk/by-id/lvm-pv-uuid-IrtXBh-SUfm-t6C6-We8L-WB5I-byci-cpBYCK$|", "r|.*|"]
filter = ["a|^/dev/disk/by-id/lvm-pv-uuid-IrtXBh-SUfm-t6C6-We8L-WB5I-byci-cpBYCK$|", "a|^/dev/disk/by-id/lvm-pv-uuid-nAG78p-lcmT-D5Bt-6w6Z-feKS-cWzG-6YZJjn$|", "a|sdc|","a|sdd|","r|.*|"]
}
找到global_filter一行发现,此filter允许(a)/dev/sda7,拒绝(r)其他所有此目录下名称(如上,此目录为dir=/dev下),
bingo,问题就出在这里,/dev/sdb/在这里被过滤掉了,因此不搜索。
解决之道就出来了,在此行【】中添加:“a|sdb1|”,注意不能用/dev/sdb1,因为/dev/已经有了,就是上面说的dir=/dev/。
之后再执行pvcreate命令,通过,问题解决。
[root@ovirt106 gluster_bricks]# fdisk /dev/sdcWelcome to fdisk (util-linux 2.32.1).
Changes will remain in memory only, until you decide to write them.
Be careful before using the write command.Command (m for help): mHelp:DOS (MBR)a toggle a bootable flagb edit nested BSD disklabelc toggle the dos compatibility flagGenericd delete a partitionF list free unpartitioned spacel list known partition typesn add a new partitionp print the partition tablet change a partition typev verify the partition tablei print information about a partitionMiscm print this menuu change display/entry unitsx extra functionality (experts only)ScriptI load disk layout from sfdisk script fileO dump disk layout to sfdisk script fileSave & Exitw write table to disk and exitq quit without saving changesCreate a new labelg create a new empty GPT partition tableG create a new empty SGI (IRIX) partition tableo create a new empty DOS partition tables create a new empty Sun partition tableCommand (m for help): d
Selected partition 1
Partition 1 has been deleted.Command (m for help): p
Disk /dev/sdc: 836 GiB, 897648164864 bytes, 1753219072 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: dos
Disk identifier: 0x34fb55a9Command (m for help): w
The partition table has been altered.
Calling ioctl() to re-read partition table.
Syncing disks.[root@ovirt106 gluster_bricks]# pvcreate /dev/sdc
WARNING: dos signature detected on /dev/sdc at offset 510. Wipe it? [y/n]: yWiping dos signature on /dev/sdc.Physical volume "/dev/sdc" successfully created.
[root@ovirt106 gluster_bricks]#
[root@node103 ~]# pvcreate /dev/sdcCannot use /dev/sdc: device is rejected by filter config
[root@node103 ~]# Cannot use /dev/sdc: device is rejected by filter config
-bash: Cannot: command not found
[root@node103 ~]# vi /etc/lvm/lvm.conf
[root@node103 ~]# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 558.9G 0 disk
├─gluster_vg_sda-gluster_lv_engine 253:17 0 100G 0 lvm /gluster_bricks/engine
├─gluster_vg_sda-gluster_thinpool_gluster_vg_sda_tmeta 253:18 0 3G 0 lvm
│ └─gluster_vg_sda-gluster_thinpool_gluster_vg_sda-tpool 253:20 0 452.9G 0 lvm
│ ├─gluster_vg_sda-gluster_thinpool_gluster_vg_sda 253:21 0 452.9G 1 lvm
│ └─gluster_vg_sda-gluster_lv_data 253:22 0 480G 0 lvm /gluster_bricks/data
└─gluster_vg_sda-gluster_thinpool_gluster_vg_sda_tdata 253:19 0 452.9G 0 lvm└─gluster_vg_sda-gluster_thinpool_gluster_vg_sda-tpool 253:20 0 452.9G 0 lvm├─gluster_vg_sda-gluster_thinpool_gluster_vg_sda 253:21 0 452.9G 1 lvm└─gluster_vg_sda-gluster_lv_data 253:22 0 480G 0 lvm /gluster_bricks/data
sdb 8:16 0 838.4G 0 disk
sdc 8:32 0 838.4G 0 disk
sdd 8:48 0 278.9G 0 disk
├─sdd1 8:49 0 1G 0 part /boot
└─sdd2 8:50 0 277.9G 0 part├─onn-pool00_tmeta 253:0 0 1G 0 lvm│ └─onn-pool00-tpool 253:2 0 206G 0 lvm│ ├─onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 253:3 0 169G 0 lvm /│ ├─onn-pool00 253:10 0 206G 1 lvm│ ├─onn-var_log_audit 253:11 0 2G 0 lvm /var/log/audit│ ├─onn-var_log 253:12 0 8G 0 lvm /var/log│ ├─onn-var_crash 253:13 0 10G 0 lvm /var/crash│ ├─onn-var 253:14 0 15G 0 lvm /var│ ├─onn-tmp 253:15 0 1G 0 lvm /tmp│ └─onn-home 253:16 0 1G 0 lvm /home├─onn-pool00_tdata 253:1 0 206G 0 lvm│ └─onn-pool00-tpool 253:2 0 206G 0 lvm│ ├─onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 253:3 0 169G 0 lvm /│ ├─onn-pool00 253:10 0 206G 1 lvm│ ├─onn-var_log_audit 253:11 0 2G 0 lvm /var/log/audit│ ├─onn-var_log 253:12 0 8G 0 lvm /var/log│ ├─onn-var_crash 253:13 0 10G 0 lvm /var/crash│ ├─onn-var 253:14 0 15G 0 lvm /var│ ├─onn-tmp 253:15 0 1G 0 lvm /tmp│ └─onn-home 253:16 0 1G 0 lvm /home└─onn-swap 253:4 0 15.7G 0 lvm [SWAP]
sr0 11:0 1 1024M 0 rom
[root@node103 ~]# pvcreate /dev/sdb
WARNING: xfs signature detected on /dev/sdb at offset 0. Wipe it? [y/n]: yWiping xfs signature on /dev/sdb.Physical volume "/dev/sdb" successfully created.
[root@node103 ~]# pvcreate /dev/sdc
WARNING: xfs signature detected on /dev/sdc at offset 0. Wipe it? [y/n]: yWiping xfs signature on /dev/sdc.Physical volume "/dev/sdc" successfully created.
[root@node103 ~]#
vgcreate
[root@node101 gluster_bricks]# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 838.4G 0 disk
sdb 8:16 0 838.4G 0 disk
sdc 8:32 0 558.9G 0 disk
├─gluster_vg_sdc-gluster_lv_engine 253:14 0 100G 0 lvm /gluster_bricks/engine
├─gluster_vg_sdc-gluster_thinpool_gluster_vg_sdc_tmeta 253:15 0 3G 0 lvm
│ └─gluster_vg_sdc-gluster_thinpool_gluster_vg_sdc-tpool 253:17 0 452.9G 0 lvm
│ ├─gluster_vg_sdc-gluster_thinpool_gluster_vg_sdc 253:18 0 452.9G 1 lvm
│ └─gluster_vg_sdc-gluster_lv_data 253:19 0 480G 0 lvm /gluster_bricks/data
└─gluster_vg_sdc-gluster_thinpool_gluster_vg_sdc_tdata 253:16 0 452.9G 0 lvm└─gluster_vg_sdc-gluster_thinpool_gluster_vg_sdc-tpool 253:17 0 452.9G 0 lvm├─gluster_vg_sdc-gluster_thinpool_gluster_vg_sdc 253:18 0 452.9G 1 lvm└─gluster_vg_sdc-gluster_lv_data 253:19 0 480G 0 lvm /gluster_bricks/data
sdd 8:48 0 278.9G 0 disk
├─sdd1 8:49 0 1G 0 part /boot
└─sdd2 8:50 0 277.9G 0 part├─onn-pool00_tmeta 253:0 0 1G 0 lvm│ └─onn-pool00-tpool 253:2 0 206G 0 lvm│ ├─onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 253:3 0 169G 0 lvm /│ ├─onn-pool00 253:7 0 206G 1 lvm│ ├─onn-var_log_audit 253:8 0 2G 0 lvm /var/log/audit│ ├─onn-var_log 253:9 0 8G 0 lvm /var/log│ ├─onn-var_crash 253:10 0 10G 0 lvm /var/crash│ ├─onn-var 253:11 0 15G 0 lvm /var│ ├─onn-tmp 253:12 0 1G 0 lvm /tmp│ └─onn-home 253:13 0 1G 0 lvm /home├─onn-pool00_tdata 253:1 0 206G 0 lvm│ └─onn-pool00-tpool 253:2 0 206G 0 lvm│ ├─onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 253:3 0 169G 0 lvm /│ ├─onn-pool00 253:7 0 206G 1 lvm│ ├─onn-var_log_audit 253:8 0 2G 0 lvm /var/log/audit│ ├─onn-var_log 253:9 0 8G 0 lvm /var/log│ ├─onn-var_crash 253:10 0 10G 0 lvm /var/crash│ ├─onn-var 253:11 0 15G 0 lvm /var│ ├─onn-tmp 253:12 0 1G 0 lvm /tmp│ └─onn-home 253:13 0 1G 0 lvm /home└─onn-swap 253:4 0 15.7G 0 lvm [SWAP]
sr0 11:0 1 1024M 0 rom
[root@node101 gluster_bricks]# vgcreate vmstore /dev/sda /dev/sdbVolume group "vmstore" successfully created
[root@node101 gluster_bricks]#
创建LV
[root@ovirt106 gluster_bricks]# blkid
/dev/mapper/onn-var_crash: UUID="de2ea7ce-29f3-441a-8b8d-47bbcbca2b59" BLOCK_SIZE="512" TYPE="xfs"
/dev/sda2: UUID="IrtXBh-SUfm-t6C6-We8L-WB5I-byci-cpBYCK" TYPE="LVM2_member" PARTUUID="d77c4977-02"
/dev/sdb: UUID="nAG78p-lcmT-D5Bt-6w6Z-feKS-cWzG-6YZJjn" TYPE="LVM2_member"
/dev/sda1: UUID="084af988-29ed-4350-ad70-1d2f5afb8d1f" BLOCK_SIZE="512" TYPE="xfs" PARTUUID="d77c4977-01"
/dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1: UUID="5d49651c-c911-4a4e-94a1-d2df627a1d89" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-swap: UUID="98ff7321-fcb6-45c9-a68d-70ebc264d09f" TYPE="swap"
/dev/mapper/gluster_vg_sdb-gluster_lv_engine: UUID="52d64ba1-3064-4334-a84a-a12078f7b32b" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/gluster_vg_sdb-gluster_lv_data: UUID="51401db2-bf3a-4aba-af15-130c2d497d94" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var_log_audit: UUID="bc108545-59d6-4f3b-bf0c-cc5777be2a3c" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var_log: UUID="78fd9245-76a1-472f-8b71-2a167429ddae" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var: UUID="0e00bbf7-2d8c-45c4-a814-a7be1e04f0bd" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-tmp: UUID="9192cbbb-e48b-4b1f-8885-a681d190a363" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-home: UUID="14416f3f-0b8b-479e-af22-51f8fbf8dc2b" BLOCK_SIZE="512" TYPE="xfs"
/dev/sdc: UUID="Q3GgE2-MYmY-p5ZP-3NTz-YgA2-mJZX-Pptxst" TYPE="LVM2_member"
/dev/sdd: UUID="V003Vh-9mYN-0H2C-znGd-Xvuw-5xu9-32AoAx" TYPE="LVM2_member"
[root@ovirt106 gluster_bricks]# lvcreate -l 100%VG -n lv_vmstore vmstore
WARNING: LVM2_member signature detected on /dev/vmstore/lv_vmstore at offset 536. Wipe it? [y/n]: yWiping LVM2_member signature on /dev/vmstore/lv_vmstore.Logical volume "lv_vmstore" created.
[root@ovirt106 gluster_bricks]# mkfs.xfs /dev/vmstore/lv_vmstore
meta-data=/dev/vmstore/lv_vmstore isize=512 agcount=4, agsize=109575680 blks= sectsz=512 attr=2, projid32bit=1= crc=1 finobt=1, sparse=1, rmapbt=0= reflink=1 bigtime=0 inobtcount=0
data = bsize=4096 blocks=438302720, imaxpct=5= sunit=0 swidth=0 blks
naming =version 2 bsize=4096 ascii-ci=0, ftype=1
log =internal log bsize=4096 blocks=214015, version=2= sectsz=512 sunit=0 blks, lazy-count=1
realtime =none extsz=4096 blocks=0, rtextents=0
格式化LV,查看uuid
[root@node103 ~]# pvcreate /dev/sdcCannot use /dev/sdc: device is rejected by filter config
[root@node103 ~]# Cannot use /dev/sdc: device is rejected by filter config
-bash: Cannot: command not found
[root@node103 ~]# vi /etc/lvm/lvm.conf
[root@node103 ~]# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 558.9G 0 disk
├─gluster_vg_sda-gluster_lv_engine 253:17 0 100G 0 lvm /gluster_bricks/engine
├─gluster_vg_sda-gluster_thinpool_gluster_vg_sda_tmeta 253:18 0 3G 0 lvm
│ └─gluster_vg_sda-gluster_thinpool_gluster_vg_sda-tpool 253:20 0 452.9G 0 lvm
│ ├─gluster_vg_sda-gluster_thinpool_gluster_vg_sda 253:21 0 452.9G 1 lvm
│ └─gluster_vg_sda-gluster_lv_data 253:22 0 480G 0 lvm /gluster_bricks/data
└─gluster_vg_sda-gluster_thinpool_gluster_vg_sda_tdata 253:19 0 452.9G 0 lvm└─gluster_vg_sda-gluster_thinpool_gluster_vg_sda-tpool 253:20 0 452.9G 0 lvm├─gluster_vg_sda-gluster_thinpool_gluster_vg_sda 253:21 0 452.9G 1 lvm└─gluster_vg_sda-gluster_lv_data 253:22 0 480G 0 lvm /gluster_bricks/data
sdb 8:16 0 838.4G 0 disk
sdc 8:32 0 838.4G 0 disk
sdd 8:48 0 278.9G 0 disk
├─sdd1 8:49 0 1G 0 part /boot
└─sdd2 8:50 0 277.9G 0 part├─onn-pool00_tmeta 253:0 0 1G 0 lvm│ └─onn-pool00-tpool 253:2 0 206G 0 lvm│ ├─onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 253:3 0 169G 0 lvm /│ ├─onn-pool00 253:10 0 206G 1 lvm│ ├─onn-var_log_audit 253:11 0 2G 0 lvm /var/log/audit│ ├─onn-var_log 253:12 0 8G 0 lvm /var/log│ ├─onn-var_crash 253:13 0 10G 0 lvm /var/crash│ ├─onn-var 253:14 0 15G 0 lvm /var│ ├─onn-tmp 253:15 0 1G 0 lvm /tmp│ └─onn-home 253:16 0 1G 0 lvm /home├─onn-pool00_tdata 253:1 0 206G 0 lvm│ └─onn-pool00-tpool 253:2 0 206G 0 lvm│ ├─onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 253:3 0 169G 0 lvm /│ ├─onn-pool00 253:10 0 206G 1 lvm│ ├─onn-var_log_audit 253:11 0 2G 0 lvm /var/log/audit│ ├─onn-var_log 253:12 0 8G 0 lvm /var/log│ ├─onn-var_crash 253:13 0 10G 0 lvm /var/crash│ ├─onn-var 253:14 0 15G 0 lvm /var│ ├─onn-tmp 253:15 0 1G 0 lvm /tmp│ └─onn-home 253:16 0 1G 0 lvm /home└─onn-swap 253:4 0 15.7G 0 lvm [SWAP]
sr0 11:0 1 1024M 0 rom
[root@node103 ~]# pvcreate /dev/sdb
WARNING: xfs signature detected on /dev/sdb at offset 0. Wipe it? [y/n]: yWiping xfs signature on /dev/sdb.Physical volume "/dev/sdb" successfully created.
[root@node103 ~]# pvcreate /dev/sdc
WARNING: xfs signature detected on /dev/sdc at offset 0. Wipe it? [y/n]: yWiping xfs signature on /dev/sdc.Physical volume "/dev/sdc" successfully created.
[root@node103 ~]# ^C
[root@node103 ~]# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 558.9G 0 disk
├─gluster_vg_sda-gluster_lv_engine 253:17 0 100G 0 lvm /gluster_bricks/engine
├─gluster_vg_sda-gluster_thinpool_gluster_vg_sda_tmeta 253:18 0 3G 0 lvm
│ └─gluster_vg_sda-gluster_thinpool_gluster_vg_sda-tpool 253:20 0 452.9G 0 lvm
│ ├─gluster_vg_sda-gluster_thinpool_gluster_vg_sda 253:21 0 452.9G 1 lvm
│ └─gluster_vg_sda-gluster_lv_data 253:22 0 480G 0 lvm /gluster_bricks/data
└─gluster_vg_sda-gluster_thinpool_gluster_vg_sda_tdata 253:19 0 452.9G 0 lvm└─gluster_vg_sda-gluster_thinpool_gluster_vg_sda-tpool 253:20 0 452.9G 0 lvm├─gluster_vg_sda-gluster_thinpool_gluster_vg_sda 253:21 0 452.9G 1 lvm└─gluster_vg_sda-gluster_lv_data 253:22 0 480G 0 lvm /gluster_bricks/data
sdb 8:16 0 838.4G 0 disk
sdc 8:32 0 838.4G 0 disk
sdd 8:48 0 278.9G 0 disk
├─sdd1 8:49 0 1G 0 part /boot
└─sdd2 8:50 0 277.9G 0 part├─onn-pool00_tmeta 253:0 0 1G 0 lvm│ └─onn-pool00-tpool 253:2 0 206G 0 lvm│ ├─onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 253:3 0 169G 0 lvm /│ ├─onn-pool00 253:10 0 206G 1 lvm│ ├─onn-var_log_audit 253:11 0 2G 0 lvm /var/log/audit│ ├─onn-var_log 253:12 0 8G 0 lvm /var/log│ ├─onn-var_crash 253:13 0 10G 0 lvm /var/crash│ ├─onn-var 253:14 0 15G 0 lvm /var│ ├─onn-tmp 253:15 0 1G 0 lvm /tmp│ └─onn-home 253:16 0 1G 0 lvm /home├─onn-pool00_tdata 253:1 0 206G 0 lvm│ └─onn-pool00-tpool 253:2 0 206G 0 lvm│ ├─onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 253:3 0 169G 0 lvm /│ ├─onn-pool00 253:10 0 206G 1 lvm│ ├─onn-var_log_audit 253:11 0 2G 0 lvm /var/log/audit│ ├─onn-var_log 253:12 0 8G 0 lvm /var/log│ ├─onn-var_crash 253:13 0 10G 0 lvm /var/crash│ ├─onn-var 253:14 0 15G 0 lvm /var│ ├─onn-tmp 253:15 0 1G 0 lvm /tmp│ └─onn-home 253:16 0 1G 0 lvm /home└─onn-swap 253:4 0 15.7G 0 lvm [SWAP]
sr0 11:0 1 1024M 0 rom
[root@node103 ~]# vgcreate vmstore /dev/sdb /dev/sdcVolume group "vmstore" successfully created
[root@node103 ~]# vg
-bash: vg: command not found
[root@node103 ~]# fdisk -l
Disk /dev/sdd: 278.9 GiB, 299439751168 bytes, 584843264 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: dos
Disk identifier: 0xe6ac2d12Device Boot Start End Sectors Size Id Type
/dev/sdd1 * 2048 2099199 2097152 1G 83 Linux
/dev/sdd2 2099200 584843263 582744064 277.9G 8e Linux LVMDisk /dev/sda: 558.9 GiB, 600127266816 bytes, 1172123568 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/sdb: 838.4 GiB, 900185481216 bytes, 1758174768 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/sdc: 838.4 GiB, 900185481216 bytes, 1758174768 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1: 169 GiB, 181466562560 bytes, 354426880 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-swap: 15.7 GiB, 16840130560 bytes, 32890880 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/onn-var_log_audit: 2 GiB, 2147483648 bytes, 4194304 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var_log: 8 GiB, 8589934592 bytes, 16777216 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var_crash: 10 GiB, 10737418240 bytes, 20971520 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var: 15 GiB, 16106127360 bytes, 31457280 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-tmp: 1 GiB, 1073741824 bytes, 2097152 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-home: 1 GiB, 1073741824 bytes, 2097152 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/gluster_vg_sda-gluster_lv_engine: 100 GiB, 107374182400 bytes, 209715200 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/gluster_vg_sda-gluster_lv_data: 480 GiB, 515396075520 bytes, 1006632960 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 131072 bytes / 262144 bytes
[root@node103 ~]# lvcreate -l 100%VG -n lv_store vmstoreLogical volume "lv_store" created.
[root@node103 ~]# fdisk -l
Disk /dev/sdd: 278.9 GiB, 299439751168 bytes, 584843264 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: dos
Disk identifier: 0xe6ac2d12Device Boot Start End Sectors Size Id Type
/dev/sdd1 * 2048 2099199 2097152 1G 83 Linux
/dev/sdd2 2099200 584843263 582744064 277.9G 8e Linux LVMDisk /dev/sda: 558.9 GiB, 600127266816 bytes, 1172123568 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/sdb: 838.4 GiB, 900185481216 bytes, 1758174768 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/sdc: 838.4 GiB, 900185481216 bytes, 1758174768 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1: 169 GiB, 181466562560 bytes, 354426880 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-swap: 15.7 GiB, 16840130560 bytes, 32890880 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/onn-var_log_audit: 2 GiB, 2147483648 bytes, 4194304 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var_log: 8 GiB, 8589934592 bytes, 16777216 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var_crash: 10 GiB, 10737418240 bytes, 20971520 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var: 15 GiB, 16106127360 bytes, 31457280 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-tmp: 1 GiB, 1073741824 bytes, 2097152 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-home: 1 GiB, 1073741824 bytes, 2097152 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/gluster_vg_sda-gluster_lv_engine: 100 GiB, 107374182400 bytes, 209715200 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/gluster_vg_sda-gluster_lv_data: 480 GiB, 515396075520 bytes, 1006632960 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 131072 bytes / 262144 bytesDisk /dev/mapper/vmstore-lv_store: 1.7 TiB, 1800363048960 bytes, 3516334080 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
[root@node103 ~]# blkid
/dev/mapper/onn-var_crash: UUID="00050fc2-5319-46da-b6c2-b0557f602845" BLOCK_SIZE="512" TYPE="xfs"
/dev/sdd2: UUID="DWXIm1-1QSa-qU1q-0urm-1sRl-BEiz-NaemIW" TYPE="LVM2_member" PARTUUID="e6ac2d12-02"
/dev/sdd1: UUID="840d0f88-c5f0-4a98-a7dd-73ebf347a85b" BLOCK_SIZE="512" TYPE="xfs" PARTUUID="e6ac2d12-01"
/dev/sda: UUID="kMvSz0-GoE5-Q0vf-qrjM-3H5m-NCLg-fSTZb8" TYPE="LVM2_member"
/dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1: UUID="8bef5e42-0fd1-4595-8b3c-079127a2c237" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-swap: UUID="18284c2e-3596-4596-b0e8-59924944c417" TYPE="swap"
/dev/mapper/onn-var_log_audit: UUID="928237b1-7c78-42d0-afca-dfd362225740" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var_log: UUID="cb6aafa3-9a97-4dc5-81f6-d2da3b0fdcd9" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var: UUID="47b2e20c-f623-404d-9a12-3b6a31b71096" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-tmp: UUID="599eda13-d184-4a59-958c-93e37bc6fdaa" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-home: UUID="ef67a11a-c786-42d9-89d0-ee518b50fce5" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/gluster_vg_sda-gluster_lv_engine: UUID="3b99a374-baa4-4783-8da8-8ecbfdb227a4" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/gluster_vg_sda-gluster_lv_data: UUID="eb5ef30a-cb48-4491-99c6-3e6c6bfe3f69" BLOCK_SIZE="512" TYPE="xfs"
/dev/sdb: UUID="m8ZIeU-j2eU-8ZMK-q52W-Zcjd-hqij-JSG8XF" TYPE="LVM2_member"
/dev/sdc: UUID="pHlCKc-wvGK-4DKX-m5IF-IXig-BZJM-YqX38s" TYPE="LVM2_member"
[root@node103 ~]# lsblk -f /dev/mapper/vmstore-lv_store
NAME FSTYPE LABEL UUID MOUNTPOINT
vmstore-lv_store
[root@node103 ~]# lsblk /dev/mapper/vmstore-lv_store -f
NAME FSTYPE LABEL UUID MOUNTPOINT
vmstore-lv_store
[root@node103 ~]# mkfs.xfs /dev/mapper/vmstore-lv_store
meta-data=/dev/mapper/vmstore-lv_store isize=512 agcount=4, agsize=109885440 blks= sectsz=512 attr=2, projid32bit=1= crc=1 finobt=1, sparse=1, rmapbt=0= reflink=1 bigtime=0 inobtcount=0
data = bsize=4096 blocks=439541760, imaxpct=5= sunit=0 swidth=0 blks
naming =version 2 bsize=4096 ascii-ci=0, ftype=1
log =internal log bsize=4096 blocks=214620, version=2= sectsz=512 sunit=0 blks, lazy-count=1
realtime =none extsz=4096 blocks=0, rtextents=0
[root@node103 ~]# lsblk /dev/mapper/vmstore-lv_store -f
NAME FSTYPE LABEL UUID MOUNTPOINT
vmstore-lv_store xfs 0ab91915-a144-4e81-bca9-996daf719c5c
[root@node103 ~]# blkid
/dev/mapper/onn-var_crash: UUID="00050fc2-5319-46da-b6c2-b0557f602845" BLOCK_SIZE="512" TYPE="xfs"
/dev/sdd2: UUID="DWXIm1-1QSa-qU1q-0urm-1sRl-BEiz-NaemIW" TYPE="LVM2_member" PARTUUID="e6ac2d12-02"
/dev/sdd1: UUID="840d0f88-c5f0-4a98-a7dd-73ebf347a85b" BLOCK_SIZE="512" TYPE="xfs" PARTUUID="e6ac2d12-01"
/dev/sda: UUID="kMvSz0-GoE5-Q0vf-qrjM-3H5m-NCLg-fSTZb8" TYPE="LVM2_member"
/dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1: UUID="8bef5e42-0fd1-4595-8b3c-079127a2c237" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-swap: UUID="18284c2e-3596-4596-b0e8-59924944c417" TYPE="swap"
/dev/mapper/onn-var_log_audit: UUID="928237b1-7c78-42d0-afca-dfd362225740" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var_log: UUID="cb6aafa3-9a97-4dc5-81f6-d2da3b0fdcd9" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var: UUID="47b2e20c-f623-404d-9a12-3b6a31b71096" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-tmp: UUID="599eda13-d184-4a59-958c-93e37bc6fdaa" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-home: UUID="ef67a11a-c786-42d9-89d0-ee518b50fce5" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/gluster_vg_sda-gluster_lv_engine: UUID="3b99a374-baa4-4783-8da8-8ecbfdb227a4" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/gluster_vg_sda-gluster_lv_data: UUID="eb5ef30a-cb48-4491-99c6-3e6c6bfe3f69" BLOCK_SIZE="512" TYPE="xfs"
/dev/sdb: UUID="m8ZIeU-j2eU-8ZMK-q52W-Zcjd-hqij-JSG8XF" TYPE="LVM2_member"
/dev/sdc: UUID="pHlCKc-wvGK-4DKX-m5IF-IXig-BZJM-YqX38s" TYPE="LVM2_member"
/dev/mapper/vmstore-lv_store: UUID="0ab91915-a144-4e81-bca9-996daf719c5c" BLOCK_SIZE="512" TYPE="xfs"
[root@node103 ~]#
创建 gluster文件存储
[root@ovirt106 /]# gluster volume create gsvmstore replica 2 node101.com:/gluster_bricks/xgvmstore node103.com:/gluster_bricks/xgvm store ovirt106.com:/gluster_bricks/xgvmstore force
number of bricks is not a multiple of replica countUsage:
volume create <NEW-VOLNAME> [stripe <COUNT>] [[replica <COUNT> [arbiter <COUNT>]]|[replica 2 thin-arbiter 1]] [disperse [<COUNT>]] [disperse-data <COUNT>] [redundancy <COUNT>] [transport <tcp|rdma|tcp,rdma>] <NEW-BRICK> <TA-BRICK>... [force][root@ovirt106 /]# gluster volume create gsvmstore replica 2 node101.com:/gluster_bricks/xgvmstore node103.com:/gluster_bricks/xgvm store ovirt106.com:/gluster_bricks/xgvmstore force
number of bricks is not a multiple of replica countUsage:
volume create <NEW-VOLNAME> [stripe <COUNT>] [[replica <COUNT> [arbiter <COUNT>]]|[replica 2 thin-arbiter 1]] [disperse [<COUNT>]] [disperse-data <COUNT>] [redundancy <COUNT>] [transport <tcp|rdma|tcp,rdma>] <NEW-BRICK> <TA-BRICK>... [force][root@ovirt106 /]# gluster peer status
Number of Peers: 2Hostname: node101.com
Uuid: 8e61e896-60a0-44b5-8136-2477715d1dcd
State: Peer in Cluster (Connected)Hostname: node103.com
Uuid: 8695cb7c-7813-4139-beb6-6eed940b6ff7
State: Peer in Cluster (Connected)
[root@ovirt106 /]# gluster volume create gs_store replica 2 node101.com:/gluster_bricks/xgvmstore node103.com:/gluster_bricks/xgvmstore ovirt106.com:/gluster_bricks/xgvmstore
Replica 2 volumes are prone to split-brain. Use Arbiter or Replica 3 to avoid this. See: http://docs.gluster.org/en/latest/Administrator%20Guide/Split%20brain%20and%20ways%20to%20deal%20with%20it/.
Do you still want to continue?(y/n) y
number of bricks is not a multiple of replica countUsage:
volume create <NEW-VOLNAME> [stripe <COUNT>] [[replica <COUNT> [arbiter <COUNT>]]|[replica 2 thin-arbiter 1]] [disperse [<COUNT>]] [disperse-data <COUNT>] [redundancy <COUNT>] [transport <tcp|rdma|tcp,rdma>] <NEW-BRICK> <TA-BRICK>... [force][root@ovirt106 /]# gluster volume create gs_store replica 3 node101.com:/gluster_bricks/xgvmstore node103.com:/gluster_bricks/xgvmstore ovirt106.com:/gluster_bricks/xgvmstore
volume create: gs_store: failed: The brick ovirt106.com:/gluster_bricks/xgvmstore is a mount point. Please create a sub-directory under the mount point and use that as the brick directory. Or use 'force' at the end of the command if you want to override this behavior.
[root@ovirt106 /]# df -h
Filesystem Size Used Avail Use% Mounted on
devtmpfs 7.4G 0 7.4G 0% /dev
tmpfs 7.5G 4.0K 7.5G 1% /dev/shm
tmpfs 7.5G 193M 7.3G 3% /run
tmpfs 7.5G 0 7.5G 0% /sys/fs/cgroup
/dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 38G 7.0G 31G 19% /
/dev/mapper/gluster_vg_sdb-gluster_lv_engine 100G 752M 100G 1% /gluster_bricks/engine
/dev/sda1 1014M 350M 665M 35% /boot
/dev/mapper/gluster_vg_sdb-gluster_lv_data 200G 1.5G 199G 1% /gluster_bricks/data
/dev/mapper/onn-home 1014M 40M 975M 4% /home
/dev/mapper/onn-var 15G 507M 15G 4% /var
/dev/mapper/onn-tmp 1014M 40M 975M 4% /tmp
/dev/mapper/onn-var_log 8.0G 292M 7.8G 4% /var/log
/dev/mapper/onn-var_crash 10G 105M 9.9G 2% /var/crash
/dev/mapper/onn-var_log_audit 2.0G 56M 2.0G 3% /var/log/audit
ovirt106.com:/data 200G 3.5G 197G 2% /rhev/data-center/mnt/glusterSD/ovirt106.com:_data
ovirt106.com:/engine 100G 1.8G 99G 2% /rhev/data-center/mnt/glusterSD/ovirt106.com:_engine
tmpfs 1.5G 0 1.5G 0% /run/user/0
/dev/mapper/vmstore-lv_vmstore 836G 5.9G 830G 1% /gluster_bricks/xgvmstore
[root@ovirt106 /]# df -HT
Filesystem Type Size Used Avail Use% Mounted on
devtmpfs devtmpfs 8.0G 0 8.0G 0% /dev
tmpfs tmpfs 8.0G 4.1k 8.0G 1% /dev/shm
tmpfs tmpfs 8.0G 202M 7.8G 3% /run
tmpfs tmpfs 8.0G 0 8.0G 0% /sys/fs/cgroup
/dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 xfs 41G 7.5G 33G 19% /
/dev/mapper/gluster_vg_sdb-gluster_lv_engine xfs 108G 788M 107G 1% /gluster_bricks/engine
/dev/sda1 xfs 1.1G 367M 697M 35% /boot
/dev/mapper/gluster_vg_sdb-gluster_lv_data xfs 215G 1.6G 214G 1% /gluster_bricks/data
/dev/mapper/onn-home xfs 1.1G 42M 1.1G 4% /home
/dev/mapper/onn-var xfs 17G 532M 16G 4% /var
/dev/mapper/onn-tmp xfs 1.1G 42M 1.1G 4% /tmp
/dev/mapper/onn-var_log xfs 8.6G 306M 8.3G 4% /var/log
/dev/mapper/onn-var_crash xfs 11G 110M 11G 2% /var/crash
/dev/mapper/onn-var_log_audit xfs 2.2G 58M 2.1G 3% /var/log/audit
ovirt106.com:/data fuse.glusterfs 215G 3.7G 211G 2% /rhev/data-center/mnt/glusterSD/ovirt106.com:_data
ovirt106.com:/engine fuse.glusterfs 108G 1.9G 106G 2% /rhev/data-center/mnt/glusterSD/ovirt106.com:_engine
tmpfs tmpfs 1.6G 0 1.6G 0% /run/user/0
/dev/mapper/vmstore-lv_vmstore xfs 898G 6.3G 891G 1% /gluster_bricks/xgvmstore
[root@ovirt106 /]# fdisk -l
Disk /dev/sdb: 557.8 GiB, 598931931136 bytes, 1169788928 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/sda: 100 GiB, 107374182400 bytes, 209715200 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: dos
Disk identifier: 0xd77c4977Device Boot Start End Sectors Size Id Type
/dev/sda1 * 2048 2099199 2097152 1G 83 Linux
/dev/sda2 2099200 209715199 207616000 99G 8e Linux LVMDisk /dev/sdc: 836 GiB, 897648164864 bytes, 1753219072 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/sdd: 836 GiB, 897648164864 bytes, 1753219072 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1: 37.7 GiB, 40479227904 bytes, 79060992 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-swap: 4.3 GiB, 4588568576 bytes, 8962048 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/gluster_vg_sdb-gluster_lv_engine: 100 GiB, 107374182400 bytes, 209715200 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/gluster_vg_sdb-gluster_lv_data: 200 GiB, 214748364800 bytes, 419430400 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 262144 bytes / 262144 bytesDisk /dev/mapper/onn-var_log_audit: 2 GiB, 2147483648 bytes, 4194304 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var_log: 8 GiB, 8589934592 bytes, 16777216 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var_crash: 10 GiB, 10737418240 bytes, 20971520 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var: 15 GiB, 16106127360 bytes, 31457280 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-tmp: 1 GiB, 1073741824 bytes, 2097152 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-home: 1 GiB, 1073741824 bytes, 2097152 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/vmstore-lv_vmstore: 1.6 TiB, 1795287941120 bytes, 3506421760 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
[root@ovirt106 /]# umount /gluster_bricks/xgvmstore/
[root@ovirt106 /]# lvremove /gluster_bricks/xgvmstore/"/gluster_bricks/xgvmstore/": Invalid path for Logical Volume.
[root@ovirt106 /]# lvremove lv_vmstoreVolume group "lv_vmstore" not foundCannot process volume group lv_vmstore
[root@ovirt106 /]# lvremove /dev/mapper/vmstore-lv_vmstore
Do you really want to remove active logical volume vmstore/lv_vmstore? [y/n]: yLogical volume "lv_vmstore" successfully removed.
[root@ovirt106 /]# lvcreate -l 100%VG -n lv_store vmstore
WARNING: xfs signature detected on /dev/vmstore/lv_store at offset 0. Wipe it? [y/n]: yWiping xfs signature on /dev/vmstore/lv_store.Logical volume "lv_store" created.
[root@ovirt106 /]# mkfs.xfs /dev/mapper/vmstore-lv_vmstore
Error accessing specified device /dev/mapper/vmstore-lv_vmstore: No such file or directory
Usage: mkfs.xfs
/* blocksize */ [-b size=num]
/* metadata */ [-m crc=0|1,finobt=0|1,uuid=xxx,rmapbt=0|1,reflink=0|1,inobtcount=0|1,bigtime=0|1]
/* data subvol */ [-d agcount=n,agsize=n,file,name=xxx,size=num,(sunit=value,swidth=value|su=num,sw=num|noalign),sectsize=num
/* force overwrite */ [-f]
/* inode size */ [-i log=n|perblock=n|size=num,maxpct=n,attr=0|1|2,projid32bit=0|1,sparse=0|1]
/* no discard */ [-K]
/* log subvol */ [-l agnum=n,internal,size=num,logdev=xxx,version=nsunit=value|su=num,sectsize=num,lazy-count=0|1]
/* label */ [-L label (maximum 12 characters)]
/* naming */ [-n size=num,version=2|ci,ftype=0|1]
/* no-op info only */ [-N]
/* prototype file */ [-p fname]
/* quiet */ [-q]
/* realtime subvol */ [-r extsize=num,size=num,rtdev=xxx]
/* sectorsize */ [-s size=num]
/* version */ [-V]devicename
<devicename> is required unless -d name=xxx is given.
<num> is xxx (bytes), xxxs (sectors), xxxb (fs blocks), xxxk (xxx KiB),xxxm (xxx MiB), xxxg (xxx GiB), xxxt (xxx TiB) or xxxp (xxx PiB).
<value> is xxx (512 byte blocks).
[root@ovirt106 /]# fdisk -l
Disk /dev/sdb: 557.8 GiB, 598931931136 bytes, 1169788928 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/sda: 100 GiB, 107374182400 bytes, 209715200 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: dos
Disk identifier: 0xd77c4977Device Boot Start End Sectors Size Id Type
/dev/sda1 * 2048 2099199 2097152 1G 83 Linux
/dev/sda2 2099200 209715199 207616000 99G 8e Linux LVMDisk /dev/sdc: 836 GiB, 897648164864 bytes, 1753219072 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/sdd: 836 GiB, 897648164864 bytes, 1753219072 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1: 37.7 GiB, 40479227904 bytes, 79060992 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-swap: 4.3 GiB, 4588568576 bytes, 8962048 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/gluster_vg_sdb-gluster_lv_engine: 100 GiB, 107374182400 bytes, 209715200 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/gluster_vg_sdb-gluster_lv_data: 200 GiB, 214748364800 bytes, 419430400 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 262144 bytes / 262144 bytesDisk /dev/mapper/onn-var_log_audit: 2 GiB, 2147483648 bytes, 4194304 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var_log: 8 GiB, 8589934592 bytes, 16777216 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var_crash: 10 GiB, 10737418240 bytes, 20971520 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var: 15 GiB, 16106127360 bytes, 31457280 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-tmp: 1 GiB, 1073741824 bytes, 2097152 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-home: 1 GiB, 1073741824 bytes, 2097152 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/vmstore-lv_store: 1.6 TiB, 1795287941120 bytes, 3506421760 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
[root@ovirt106 /]# mkfs.xfs /dev/mapper/vmstore-lv_store
meta-data=/dev/mapper/vmstore-lv_store isize=512 agcount=4, agsize=109575680 blks= sectsz=512 attr=2, projid32bit=1= crc=1 finobt=1, sparse=1, rmapbt=0= reflink=1 bigtime=0 inobtcount=0
data = bsize=4096 blocks=438302720, imaxpct=5= sunit=0 swidth=0 blks
naming =version 2 bsize=4096 ascii-ci=0, ftype=1
log =internal log bsize=4096 blocks=214015, version=2= sectsz=512 sunit=0 blks, lazy-count=1
realtime =none extsz=4096 blocks=0, rtextents=0
[root@ovirt106 /]# df -HT
Filesystem Type Size Used Avail Use% Mounted on
devtmpfs devtmpfs 8.0G 0 8.0G 0% /dev
tmpfs tmpfs 8.0G 4.1k 8.0G 1% /dev/shm
tmpfs tmpfs 8.0G 202M 7.8G 3% /run
tmpfs tmpfs 8.0G 0 8.0G 0% /sys/fs/cgroup
/dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 xfs 41G 7.5G 33G 19% /
/dev/mapper/gluster_vg_sdb-gluster_lv_engine xfs 108G 788M 107G 1% /gluster_bricks/engine
/dev/sda1 xfs 1.1G 367M 697M 35% /boot
/dev/mapper/gluster_vg_sdb-gluster_lv_data xfs 215G 1.6G 214G 1% /gluster_bricks/data
/dev/mapper/onn-home xfs 1.1G 42M 1.1G 4% /home
/dev/mapper/onn-var xfs 17G 532M 16G 4% /var
/dev/mapper/onn-tmp xfs 1.1G 42M 1.1G 4% /tmp
/dev/mapper/onn-var_log xfs 8.6G 306M 8.3G 4% /var/log
/dev/mapper/onn-var_crash xfs 11G 110M 11G 2% /var/crash
/dev/mapper/onn-var_log_audit xfs 2.2G 58M 2.1G 3% /var/log/audit
ovirt106.com:/data fuse.glusterfs 215G 3.7G 211G 2% /rhev/data-center/mnt/glusterSD/ovirt106.com:_data
ovirt106.com:/engine fuse.glusterfs 108G 1.9G 106G 2% /rhev/data-center/mnt/glusterSD/ovirt106.com:_engine
tmpfs tmpfs 1.6G 0 1.6G 0% /run/user/0
[root@ovirt106 /]# blkid
/dev/mapper/onn-var_crash: UUID="de2ea7ce-29f3-441a-8b8d-47bbcbca2b59" BLOCK_SIZE="512" TYPE="xfs"
/dev/sda2: UUID="IrtXBh-SUfm-t6C6-We8L-WB5I-byci-cpBYCK" TYPE="LVM2_member" PARTUUID="d77c4977-02"
/dev/sdb: UUID="nAG78p-lcmT-D5Bt-6w6Z-feKS-cWzG-6YZJjn" TYPE="LVM2_member"
/dev/sda1: UUID="084af988-29ed-4350-ad70-1d2f5afb8d1f" BLOCK_SIZE="512" TYPE="xfs" PARTUUID="d77c4977-01"
/dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1: UUID="5d49651c-c911-4a4e-94a1-d2df627a1d89" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-swap: UUID="98ff7321-fcb6-45c9-a68d-70ebc264d09f" TYPE="swap"
/dev/mapper/gluster_vg_sdb-gluster_lv_engine: UUID="52d64ba1-3064-4334-a84a-a12078f7b32b" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/gluster_vg_sdb-gluster_lv_data: UUID="51401db2-bf3a-4aba-af15-130c2d497d94" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var_log_audit: UUID="bc108545-59d6-4f3b-bf0c-cc5777be2a3c" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var_log: UUID="78fd9245-76a1-472f-8b71-2a167429ddae" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var: UUID="0e00bbf7-2d8c-45c4-a814-a7be1e04f0bd" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-tmp: UUID="9192cbbb-e48b-4b1f-8885-a681d190a363" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-home: UUID="14416f3f-0b8b-479e-af22-51f8fbf8dc2b" BLOCK_SIZE="512" TYPE="xfs"
/dev/sdc: UUID="Q3GgE2-MYmY-p5ZP-3NTz-YgA2-mJZX-Pptxst" TYPE="LVM2_member"
/dev/sdd: UUID="V003Vh-9mYN-0H2C-znGd-Xvuw-5xu9-32AoAx" TYPE="LVM2_member"
/dev/mapper/vmstore-lv_store: UUID="ae6d1e8a-669f-4716-abf4-c914651061fb" BLOCK_SIZE="512" TYPE="xfs"
[root@ovirt106 /]# vi /etc/fstab
[root@ovirt106 /]# mount -a
[root@ovirt106 /]# df -HT
Filesystem Type Size Used Avail Use% Mounted on
devtmpfs devtmpfs 8.0G 0 8.0G 0% /dev
tmpfs tmpfs 8.0G 4.1k 8.0G 1% /dev/shm
tmpfs tmpfs 8.0G 202M 7.8G 3% /run
tmpfs tmpfs 8.0G 0 8.0G 0% /sys/fs/cgroup
/dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 xfs 41G 7.5G 33G 19% /
/dev/mapper/gluster_vg_sdb-gluster_lv_engine xfs 108G 788M 107G 1% /gluster_bricks/engine
/dev/sda1 xfs 1.1G 367M 697M 35% /boot
/dev/mapper/gluster_vg_sdb-gluster_lv_data xfs 215G 1.6G 214G 1% /gluster_bricks/data
/dev/mapper/onn-home xfs 1.1G 42M 1.1G 4% /home
/dev/mapper/onn-var xfs 17G 532M 16G 4% /var
/dev/mapper/onn-tmp xfs 1.1G 42M 1.1G 4% /tmp
/dev/mapper/onn-var_log xfs 8.6G 306M 8.3G 4% /var/log
/dev/mapper/onn-var_crash xfs 11G 110M 11G 2% /var/crash
/dev/mapper/onn-var_log_audit xfs 2.2G 58M 2.1G 3% /var/log/audit
ovirt106.com:/data fuse.glusterfs 215G 3.7G 211G 2% /rhev/data-center/mnt/glusterSD/ovirt106.com:_data
ovirt106.com:/engine fuse.glusterfs 108G 1.9G 106G 2% /rhev/data-center/mnt/glusterSD/ovirt106.com:_engine
tmpfs tmpfs 1.6G 0 1.6G 0% /run/user/0
/dev/mapper/vmstore-lv_store xfs 1.8T 13G 1.8T 1% /gluster_bricks/xgvmstore
[root@ovirt106 /]# gluster volume create gs_store replica 3 node101.com:/gluster_bricks/xgvmstore node103.com:/gluster_bricks/xgvmstore ovirt106.com:/gluster_bricks/xgvmstore force
volume create: gs_store: success: please start the volume to access data
[root@ovirt106 /]#
添加磁盘,显示没有权限
[root@ovirt106 /]# gluster volume create gs_store replica 3 node101.com:/gluster_bricks/xgvmstore node103.com:/gluster_bricks/xgvmstore ovirt106.com:/gluster_bricks/xgvmstore force
volume create: gs_store: success: please start the volume to access data
[root@ovirt106 /]# ^C
[root@ovirt106 /]# gluster volume list
data
engine
gs_store
[root@ovirt106 /]# cd /gluster_bricks/xgvmstore/
[root@ovirt106 xgvmstore]# ls
[root@ovirt106 xgvmstore]# gluseter volume start gs_store
-bash: gluseter: command not found
[root@ovirt106 xgvmstore]# gluster volume start gs_store
volume start: gs_store: success
[root@ovirt106 xgvmstore]# cd ..
[root@ovirt106 gluster_bricks]# l,s
-bash: l,s: command not found
[root@ovirt106 gluster_bricks]# ls
data engine xgvmstore
[root@ovirt106 gluster_bricks]# ll
total 0
drwxr-xr-x. 3 root root 18 May 31 11:15 data
drwxr-xr-x. 3 root root 20 May 31 11:14 engine
drwxr-xr-x. 3 root root 24 Jun 7 11:21 xgvmstore
[root@ovirt106 gluster_bricks]# cd engine/
[root@ovirt106 engine]# ll
total 8
drwxr-xr-x. 5 vdsm kvm 8192 Jun 7 11:29 engine
[root@ovirt106 engine]# cd engine/
[root@ovirt106 engine]# ls
5afdcc5b-9de7-49ea-af48-85575bbcc482
[root@ovirt106 engine]# ll
total 0
drwxr-xr-x. 4 vdsm kvm 34 Jun 6 14:54 5afdcc5b-9de7-49ea-af48-85575bbcc482
[root@ovirt106 engine]# cd 5afdcc5b-9de7-49ea-af48-85575bbcc482/
[root@ovirt106 5afdcc5b-9de7-49ea-af48-85575bbcc482]# ll
total 0
drwxr-xr-x. 2 vdsm kvm 89 Jun 6 14:55 dom_md
drwxr-xr-x. 4 vdsm kvm 94 Jun 6 14:59 images
[root@ovirt106 5afdcc5b-9de7-49ea-af48-85575bbcc482]# cd ..
[root@ovirt106 engine]# ls
5afdcc5b-9de7-49ea-af48-85575bbcc482
[root@ovirt106 engine]# ls
5afdcc5b-9de7-49ea-af48-85575bbcc482
[root@ovirt106 engine]# ..
-bash: ..: command not found
[root@ovirt106 engine]# ls
5afdcc5b-9de7-49ea-af48-85575bbcc482
[root@ovirt106 engine]# cd ..
[root@ovirt106 engine]# ls
engine
[root@ovirt106 engine]# cd ..
[root@ovirt106 gluster_bricks]# ls
data engine xgvmstore
[root@ovirt106 gluster_bricks]# cd data/
[root@ovirt106 data]# ll
total 8
drwxr-xr-x. 5 vdsm kvm 8192 Jun 7 11:30 data
[root@ovirt106 data]# cd ..
[root@ovirt106 gluster_bricks]# ll
total 0
drwxr-xr-x. 3 root root 18 May 31 11:15 data
drwxr-xr-x. 3 root root 20 May 31 11:14 engine
drwxr-xr-x. 3 root root 24 Jun 7 11:21 xgvmstore
[root@ovirt106 gluster_bricks]# cd xgvmstore/
[root@ovirt106 xgvmstore]# ls
[root@ovirt106 xgvmstore]# mkdir store
[root@ovirt106 xgvmstore]# ls
store
[root@ovirt106 xgvmstore]# gluster peer status
Number of Peers: 2Hostname: node101.com
Uuid: 8e61e896-60a0-44b5-8136-2477715d1dcd
State: Peer in Cluster (Connected)Hostname: node103.com
Uuid: 8695cb7c-7813-4139-beb6-6eed940b6ff7
State: Peer in Cluster (Connected)
[root@ovirt106 xgvmstore]# gluster info
unrecognized word: info (position 0)Usage: gluster [options] <help> <peer> <pool> <volume>Options:--help Shows the help information--version Shows the version--print-logdir Shows the log directory--print-statedumpdir Shows the state dump directory[root@ovirt106 xgvmstore]# gluster volume status
Status of volume: data
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick node101.com:/gluster_bricks/data/data 49154 0 Y 11362
Brick node103.com:/gluster_bricks/data/data 49152 0 Y 2963
Brick ovirt106.com:/gluster_bricks/data/dat
a 49152 0 Y 9180
Self-heal Daemon on localhost N/A N/A Y 9204
Self-heal Daemon on node103.com N/A N/A Y 3032
Self-heal Daemon on node101.com N/A N/A Y 2974Task Status of Volume data
------------------------------------------------------------------------------
There are no active volume tasksStatus of volume: engine
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick node101.com:/gluster_bricks/engine/en
gine 49155 0 Y 11373
Brick node103.com:/gluster_bricks/engine/en
gine 49153 0 Y 2988
Brick ovirt106.com:/gluster_bricks/engine/e
ngine 49153 0 Y 9193
Self-heal Daemon on localhost N/A N/A Y 9204
Self-heal Daemon on node101.com N/A N/A Y 2974
Self-heal Daemon on node103.com N/A N/A Y 3032Task Status of Volume engine
------------------------------------------------------------------------------
There are no active volume tasksStatus of volume: gs_store
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick node101.com:/gluster_bricks/xgvmstore 49152 0 Y 103067
Brick node103.com:/gluster_bricks/xgvmstore 49154 0 Y 54523
Brick ovirt106.com:/gluster_bricks/xgvmstor
e 49154 0 Y 165193
Self-heal Daemon on localhost N/A N/A Y 9204
Self-heal Daemon on node101.com N/A N/A Y 2974
Self-heal Daemon on node103.com N/A N/A Y 3032Task Status of Volume gs_store
------------------------------------------------------------------------------
There are no active volume tasks[root@ovirt106 xgvmstore]# ls
store
[root@ovirt106 xgvmstore]# rm store
rm: cannot remove 'store': Is a directory
[root@ovirt106 xgvmstore]# rm -rf store/
[root@ovirt106 xgvmstore]# ls
[root@ovirt106 xgvmstore]# cd ..
[root@ovirt106 gluster_bricks]# ls
data engine xgvmstore
[root@ovirt106 gluster_bricks]# chown -R vdsm:kvm /gluster_bricks/xgvmstore/
[root@ovirt106 gluster_bricks]# ll
total 0
drwxr-xr-x. 3 root root 18 May 31 11:15 data
drwxr-xr-x. 3 root root 20 May 31 11:14 engine
drwxr-xr-x. 3 vdsm kvm 24 Jun 7 11:38 xgvmstore
[root@ovirt106 gluster_bricks]# pwd
/gluster_bricks
[root@ovirt106 gluster_bricks]# cd data/
[root@ovirt106 data]# ll
total 8
drwxr-xr-x. 5 vdsm kvm 8192 Jun 7 11:40 data
[root@ovirt106 data]#
删除gluster卷,重建分布式卷
node status: OKSee `nodectl check` for more informationAdmin Console: https://172.16.100.106:9090/[root@ovirt106 ~]# df -h
Filesystem Size Used Avail Use% Mounted on
devtmpfs 7.4G 0 7.4G 0% /dev
tmpfs 7.5G 4.0K 7.5G 1% /dev/shm
tmpfs 7.5G 17M 7.5G 1% /run
tmpfs 7.5G 0 7.5G 0% /sys/fs/cgroup
/dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 38G 7.0G 31G 19% /
/dev/mapper/vmstore-lv_store 1.7T 12G 1.7T 1% /gluster_bricks/xgvmstore
/dev/mapper/gluster_vg_sdb-gluster_lv_engine 100G 99G 1.3G 99% /gluster_bricks/engine
/dev/sda1 1014M 350M 665M 35% /boot
/dev/mapper/gluster_vg_sdb-gluster_lv_data 200G 7.0G 193G 4% /gluster_bricks/data
/dev/mapper/onn-var 15G 624M 15G 5% /var
/dev/mapper/onn-home 1014M 40M 975M 4% /home
/dev/mapper/onn-var_crash 10G 105M 9.9G 2% /var/crash
/dev/mapper/onn-var_log 8.0G 298M 7.7G 4% /var/log
/dev/mapper/onn-tmp 1014M 40M 975M 4% /tmp
/dev/mapper/onn-var_log_audit 2.0G 56M 2.0G 3% /var/log/audit
tmpfs 1.5G 0 1.5G 0% /run/user/0
[root@ovirt106 ~]# blkid
/dev/mapper/onn-var_crash: UUID="de2ea7ce-29f3-441a-8b8d-47bbcbca2b59" BLOCK_SIZE="512" TYPE="xfs"
/dev/sda2: UUID="IrtXBh-SUfm-t6C6-We8L-WB5I-byci-cpBYCK" TYPE="LVM2_member" PARTUUID="d77c4977-02"
/dev/sda1: UUID="084af988-29ed-4350-ad70-1d2f5afb8d1f" BLOCK_SIZE="512" TYPE="xfs" PARTUUID="d77c4977-01"
/dev/sdb: UUID="nAG78p-lcmT-D5Bt-6w6Z-feKS-cWzG-6YZJjn" TYPE="LVM2_member"
/dev/sdc: UUID="Q3GgE2-MYmY-p5ZP-3NTz-YgA2-mJZX-Pptxst" TYPE="LVM2_member"
/dev/sdd: UUID="V003Vh-9mYN-0H2C-znGd-Xvuw-5xu9-32AoAx" TYPE="LVM2_member"
/dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1: UUID="5d49651c-c911-4a4e-94a1-d2df627a1d89" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-swap: UUID="98ff7321-fcb6-45c9-a68d-70ebc264d09f" TYPE="swap"
/dev/mapper/vmstore-lv_store: UUID="ae6d1e8a-669f-4716-abf4-c914651061fb" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/gluster_vg_sdb-gluster_lv_engine: UUID="52d64ba1-3064-4334-a84a-a12078f7b32b" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/gluster_vg_sdb-gluster_lv_data: UUID="51401db2-bf3a-4aba-af15-130c2d497d94" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var_log_audit: UUID="bc108545-59d6-4f3b-bf0c-cc5777be2a3c" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var_log: UUID="78fd9245-76a1-472f-8b71-2a167429ddae" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-var: UUID="0e00bbf7-2d8c-45c4-a814-a7be1e04f0bd" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-tmp: UUID="9192cbbb-e48b-4b1f-8885-a681d190a363" BLOCK_SIZE="512" TYPE="xfs"
/dev/mapper/onn-home: UUID="14416f3f-0b8b-479e-af22-51f8fbf8dc2b" BLOCK_SIZE="512" TYPE="xfs"
[root@ovirt106 ~]# fdisk -l
Disk /dev/sda: 100 GiB, 107374182400 bytes, 209715200 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: dos
Disk identifier: 0xd77c4977Device Boot Start End Sectors Size Id Type
/dev/sda1 * 2048 2099199 2097152 1G 83 Linux
/dev/sda2 2099200 209715199 207616000 99G 8e Linux LVMDisk /dev/sdb: 557.8 GiB, 598931931136 bytes, 1169788928 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/sdc: 836 GiB, 897648164864 bytes, 1753219072 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/sdd: 836 GiB, 897648164864 bytes, 1753219072 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1: 37.7 GiB, 40479227904 bytes, 79060992 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-swap: 4.3 GiB, 4588568576 bytes, 8962048 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/vmstore-lv_store: 1.6 TiB, 1795287941120 bytes, 3506421760 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/gluster_vg_sdb-gluster_lv_engine: 100 GiB, 107374182400 bytes, 209715200 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytesDisk /dev/mapper/gluster_vg_sdb-gluster_lv_data: 200 GiB, 214748364800 bytes, 419430400 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 262144 bytes / 262144 bytesDisk /dev/mapper/onn-var_log_audit: 2 GiB, 2147483648 bytes, 4194304 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var_log: 8 GiB, 8589934592 bytes, 16777216 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var_crash: 10 GiB, 10737418240 bytes, 20971520 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-var: 15 GiB, 16106127360 bytes, 31457280 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-tmp: 1 GiB, 1073741824 bytes, 2097152 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytesDisk /dev/mapper/onn-home: 1 GiB, 1073741824 bytes, 2097152 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 65536 bytes / 65536 bytes
[root@ovirt106 ~]# lvscanACTIVE '/dev/vmstore/lv_store' [1.63 TiB] inheritACTIVE '/dev/gluster_vg_sdb/gluster_thinpool_gluster_vg_sdb' [<198.00 GiB] inheritACTIVE '/dev/gluster_vg_sdb/gluster_lv_engine' [100.00 GiB] inheritACTIVE '/dev/gluster_vg_sdb/gluster_lv_data' [200.00 GiB] inheritACTIVE '/dev/onn/pool00' [<74.70 GiB] inheritACTIVE '/dev/onn/var_log_audit' [2.00 GiB] inheritACTIVE '/dev/onn/var_log' [8.00 GiB] inheritACTIVE '/dev/onn/var_crash' [10.00 GiB] inheritACTIVE '/dev/onn/var' [15.00 GiB] inheritACTIVE '/dev/onn/tmp' [1.00 GiB] inheritACTIVE '/dev/onn/home' [1.00 GiB] inheritinactive '/dev/onn/root' [<37.70 GiB] inheritACTIVE '/dev/onn/swap' [4.27 GiB] inheritinactive '/dev/onn/ovirt-node-ng-4.4.10.2-0.20220303.0' [<37.70 GiB] inheritACTIVE '/dev/onn/ovirt-node-ng-4.4.10.2-0.20220303.0+1' [<37.70 GiB] inherit
[root@ovirt106 ~]# pvscanPV /dev/sdc VG vmstore lvm2 [<836.00 GiB / 0 free]PV /dev/sdd VG vmstore lvm2 [<836.00 GiB / 0 free]PV /dev/sdb VG gluster_vg_sdb lvm2 [<300.00 GiB / 0 free]PV /dev/sda2 VG onn lvm2 [<99.00 GiB / 18.02 GiB free]Total: 4 [2.02 TiB] / in use: 4 [2.02 TiB] / in no VG: 0 [0 ]
[root@ovirt106 ~]# df -HT
Filesystem Type Size Used Avail Use% Mounted on
devtmpfs devtmpfs 8.0G 0 8.0G 0% /dev
tmpfs tmpfs 8.0G 4.1k 8.0G 1% /dev/shm
tmpfs tmpfs 8.0G 18M 8.0G 1% /run
tmpfs tmpfs 8.0G 0 8.0G 0% /sys/fs/cgroup
/dev/mapper/onn-ovirt--node--ng--4.4.10.2--0.20220303.0+1 xfs 41G 7.5G 33G 19% /
/dev/mapper/vmstore-lv_store xfs 1.8T 13G 1.8T 1% /gluster_bricks/xgvmstore
/dev/mapper/gluster_vg_sdb-gluster_lv_engine xfs 108G 107G 1.4G 99% /gluster_bricks/engine
/dev/sda1 xfs 1.1G 367M 697M 35% /boot
/dev/mapper/gluster_vg_sdb-gluster_lv_data xfs 215G 7.5G 208G 4% /gluster_bricks/data
/dev/mapper/onn-var xfs 17G 532M 16G 4% /var
/dev/mapper/onn-home xfs 1.1G 42M 1.1G 4% /home
/dev/mapper/onn-var_crash xfs 11G 110M 11G 2% /var/crash
/dev/mapper/onn-var_log xfs 8.6G 313M 8.3G 4% /var/log
/dev/mapper/onn-tmp xfs 1.1G 42M 1.1G 4% /tmp
/dev/mapper/onn-var_log_audit xfs 2.2G 59M 2.1G 3% /var/log/audit
tmpfs tmpfs 1.6G 0 1.6G 0% /run/user/0
ovirt106.com:/data fuse.glusterfs 215G 9.6G 206G 5% /rhev/data-center/mnt/glusterSD/ovirt106.com:_data
ovirt106.com:/engine fuse.glusterfs 108G 108G 229M 100% /rhev/data-center/mnt/glusterSD/ovirt106.com:_engine
[root@ovirt106 ~]# gluster volume gs_store
unrecognized word: gs_store (position 1)Usage: gluster [options] <help> <peer> <pool> <volume>Options:--help Shows the help information--version Shows the version--print-logdir Shows the log directory--print-statedumpdir Shows the state dump directory[root@ovirt106 ~]# gluster volume status
Status of volume: data
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick node101.com:/gluster_bricks/data/data 49154 0 Y 11362
Brick node103.com:/gluster_bricks/data/data 49152 0 Y 2963
Brick ovirt106.com:/gluster_bricks/data/dat
a 49152 0 Y 1979
Self-heal Daemon on localhost N/A N/A Y 2067
Self-heal Daemon on node101.com N/A N/A Y 2974
Self-heal Daemon on node103.com N/A N/A Y 3032Task Status of Volume data
------------------------------------------------------------------------------
There are no active volume tasksStatus of volume: engine
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick node101.com:/gluster_bricks/engine/en
gine 49155 0 Y 11373
Brick node103.com:/gluster_bricks/engine/en
gine 49153 0 Y 2988
Brick ovirt106.com:/gluster_bricks/engine/e
ngine 49153 0 Y 2007
Self-heal Daemon on localhost N/A N/A Y 2067
Self-heal Daemon on node101.com N/A N/A Y 2974
Self-heal Daemon on node103.com N/A N/A Y 3032Task Status of Volume engine
------------------------------------------------------------------------------
There are no active volume tasksVolume gs_store is not started[root@ovirt106 ~]# gluster volume gs_store
unrecognized word: gs_store (position 1)Usage: gluster [options] <help> <peer> <pool> <volume>Options:--help Shows the help information--version Shows the version--print-logdir Shows the log directory--print-statedumpdir Shows the state dump directory[root@ovirt106 ~]# gluster volume data
unrecognized word: data (position 1)Usage: gluster [options] <help> <peer> <pool> <volume>Options:--help Shows the help information--version Shows the version--print-logdir Shows the log directory--print-statedumpdir Shows the state dump directory[root@ovirt106 ~]# gluster volume data status
unrecognized word: data (position 1)Usage: gluster [options] <help> <peer> <pool> <volume>Options:--help Shows the help information--version Shows the version--print-logdir Shows the log directory--print-statedumpdir Shows the state dump directory[root@ovirt106 ~]# gluster volume delete gs_store
Deleting volume will erase all information about the volume. Do you want to continue? (y/n) y
volume delete: gs_store: success
[root@ovirt106 ~]# gluster volume data status
unrecognized word: data (position 1)Usage: gluster [options] <help> <peer> <pool> <volume>Options:--help Shows the help information--version Shows the version--print-logdir Shows the log directory--print-statedumpdir Shows the state dump directory[root@ovirt106 ~]# gluster volume infoVolume Name: data
Type: Replicate
Volume ID: 5db6c160-09c5-460b-8d1d-f79cb87dc73b
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: node101.com:/gluster_bricks/data/data
Brick2: node103.com:/gluster_bricks/data/data
Brick3: ovirt106.com:/gluster_bricks/data/data
Options Reconfigured:
performance.client-io-threads: on
nfs.disable: on
transport.address-family: inet
storage.fips-mode-rchecksum: on
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.low-prio-threads: 32
network.remote-dio: off
performance.strict-o-direct: on
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
cluster.choose-local: off
client.event-threads: 4
server.event-threads: 4
network.ping-timeout: 30
server.tcp-user-timeout: 20
server.keepalive-time: 10
server.keepalive-interval: 2
server.keepalive-count: 5
cluster.lookup-optimize: off
storage.owner-uid: 36
storage.owner-gid: 36
cluster.granular-entry-heal: enableVolume Name: engine
Type: Replicate
Volume ID: 47cb595e-018e-43b7-b387-26a1d83bb5ff
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: node101.com:/gluster_bricks/engine/engine
Brick2: node103.com:/gluster_bricks/engine/engine
Brick3: ovirt106.com:/gluster_bricks/engine/engine
Options Reconfigured:
performance.client-io-threads: on
nfs.disable: on
transport.address-family: inet
storage.fips-mode-rchecksum: on
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.low-prio-threads: 32
network.remote-dio: off
performance.strict-o-direct: on
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
cluster.choose-local: off
client.event-threads: 4
server.event-threads: 4
network.ping-timeout: 30
server.tcp-user-timeout: 20
server.keepalive-time: 10
server.keepalive-interval: 2
server.keepalive-count: 5
cluster.lookup-optimize: off
storage.owner-uid: 36
storage.owner-gid: 36
cluster.granular-entry-heal: enable
[root@ovirt106 ~]# gluster volume create gs_store node101.com:/gluster_bricks/xgvmstore node103.com:/gluster_bricks/xgvmstore ovirt106.com:/gluster_bricks/xgvmstore force
volume create: gs_store: success: please start the volume to access data
[root@ovirt106 ~]#
重启卷
[root@ovirt106 ~]# gluster volume start gs_store
volume start: gs_store: success
[root@ovirt106 ~]# gluster volume infoVolume Name: data
Type: Replicate
Volume ID: 5db6c160-09c5-460b-8d1d-f79cb87dc73b
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: node101.com:/gluster_bricks/data/data
Brick2: node103.com:/gluster_bricks/data/data
Brick3: ovirt106.com:/gluster_bricks/data/data
Options Reconfigured:
performance.client-io-threads: on
nfs.disable: on
transport.address-family: inet
storage.fips-mode-rchecksum: on
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.low-prio-threads: 32
network.remote-dio: off
performance.strict-o-direct: on
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
cluster.choose-local: off
client.event-threads: 4
server.event-threads: 4
network.ping-timeout: 30
server.tcp-user-timeout: 20
server.keepalive-time: 10
server.keepalive-interval: 2
server.keepalive-count: 5
cluster.lookup-optimize: off
storage.owner-uid: 36
storage.owner-gid: 36
cluster.granular-entry-heal: enableVolume Name: engine
Type: Replicate
Volume ID: 47cb595e-018e-43b7-b387-26a1d83bb5ff
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: node101.com:/gluster_bricks/engine/engine
Brick2: node103.com:/gluster_bricks/engine/engine
Brick3: ovirt106.com:/gluster_bricks/engine/engine
Options Reconfigured:
performance.client-io-threads: on
nfs.disable: on
transport.address-family: inet
storage.fips-mode-rchecksum: on
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.low-prio-threads: 32
network.remote-dio: off
performance.strict-o-direct: on
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
cluster.choose-local: off
client.event-threads: 4
server.event-threads: 4
network.ping-timeout: 30
server.tcp-user-timeout: 20
server.keepalive-time: 10
server.keepalive-interval: 2
server.keepalive-count: 5
cluster.lookup-optimize: off
storage.owner-uid: 36
storage.owner-gid: 36
cluster.granular-entry-heal: enableVolume Name: gs_store
Type: Distribute
Volume ID: e6ae7082-7ba6-4e02-9656-2f8a3792350c
Status: Started
Snapshot Count: 0
Number of Bricks: 3
Transport-type: tcp
Bricks:
Brick1: node101.com:/gluster_bricks/xgvmstore
Brick2: node103.com:/gluster_bricks/xgvmstore
Brick3: ovirt106.com:/gluster_bricks/xgvmstore
Options Reconfigured:
storage.fips-mode-rchecksum: on
transport.address-family: inet
nfs.disable: on
[root@ovirt106 ~]#
添加域
Cannot use /dev/sdc: device is rejected by filter config报错处理相关推荐
- 报错解决:Ubuntu插入硬盘不能挂载,Error mounting /dev/sdb1 at /media/×××/×××: unknown filesystem type ‘exfat’
报错解决:Ubuntu插入硬盘不能挂载,Error mounting /dev/sdb1 at /media/×××/×××: unknown filesystem type 'exfat' 报错 报 ...
- Device /dev/sdc excluded by a filter.
当做LVM逻辑卷,创建vg是报以下错误 [root@master ~]# vgextend myvo /dev/sdcDevice /dev/sdc excluded by a filter. 导致问 ...
- pvcreate报错:Device /dev/mapper/mpatha not found(or ignored by filtering). 2022_11_08
问题:pvcreate命令在使用多路径设备时失败,即使该设备存在,也会出现以下错误. pvcreate报错:Device /dev/mapper/mpatha not found(or ignored ...
- Linux格式化分区报错Could not start /dev/sda No such file or directory 解决办法
Linux查看已经分好的区 [root@linuxidc ~]# fdisk -l /dev/sda Disk /dev/sda: 21.5 GB, 21474836480 bytes 25 ...
- 使用fswebcam报错Error selecting input 0 VIDIOC_S_INPUT: Device or resource busy或者使用metion监控,画面为灰色
一,BUG解决 不知道大家使用树莓派的时候,有没有遇到过使用fswebcam报错Error selecting input 0 VIDIOC_S_INPUT: Device or resource b ...
- git pull遇到报错:! [rejected]xxx-> xxx (non-fast-forward)
本地做了修改之后再git pull遇到报错:! [rejected]xxx-> xxx (non-fast-forward) 这个时候我们有两种处理: 保留修改 丢弃修改 保留修改 法1. gi ...
- IDE报错 hdc_std version of the SDK does not match the hdcd version of the device.
IDE运行app报错:The hdc_std version of the SDK does not match the hdcd version of the device. 遇到这个问题主要是 h ...
- Linux 格式化磁盘,制作文件系统 => 报错: dev/sdb is NOT identified.
前言 在学习 ARM S5PV210 开发板的时候, SD 卡烧录时总是失败,报错信息: dev/sdb is NOT identified. 经过实践,发现格式化 SD 卡可以解决该问题:或者使用命 ...
- UMOUNT NAS报错No space left on device和was not found in /proc/mounts
UMOUNT NAS报错No space left on device和was not found in /proc/mounts NAS_IP:/sysidc 表示NAS的存储设备. [root@m ...
最新文章
- Nature子刊:涵盖20多万个基因组的人体肠道微生物参考基因组集
- go语言buffio与继承
- VTK:结构化网格之StructuredGrid
- Linux开发板网线直连电脑配置方法
- Java多线程编程系列-多线程基础
- Android Studio中删除所有的断点
- python 函数图解_Python函数说明(一)
- 各种免费在线工具收集
- 错误一例:expected expression before } token
- vep文件转换mp4翻录教程
- 假装写在了开头qwq
- LSTM之父最新长文:现代AI和深度学习发展史
- 据结构学习(冒泡、选择、插入、快速排....
- RuntimeError: CUDA out of memory. Tried to allocate 1.54 GiB...
- 前端知识总结之基础知识
- Retrofit2.0初始化
- android5 网络请求时 FileNotFoundException android兼容性原因
- 智慧社区如何设计 智慧社区解决方案
- 5G新机难救董明珠“手机梦”
- Android App运行时签名校验
热门文章
- 问题解决:尝试解决maven依赖找不到的n种姿势
- 十进制小数转为二进制小数
- iis7设置html支持asp,IIS7.0 设置技巧 (如何支持asp?)
- python 对角矩阵_numpy创建单位矩阵和对角矩阵的实例
- 手机上怎么打印资料?手机接收的文件怎么打印
- BS、CS分别是什么,BS、CS有什么区别和联系
- 伊登工业云:云计算打通“机器换人”最关键一环
- 域环境SID相同到底有什么影响?
- VScode 插件中 package.json 文件字段详解
- 浙大计算机保研面试,保研浙大计算机后总结