-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Description
Description of problem:
Installing GlusterFS 8, 9, or 10, all have issues starting volumes on restart
The exact command to reproduce the issue:
dnf install centos-release-gluster9 -y
nano /etc/yum.repos.d/CentOS-Gluster-9.repo
baseurl=https://dl.rockylinux.org/vault/centos/8.5.2111/storage/x86_64/gluster-9/
dnf install glusterfs glusterfs-libs glusterfs-server -y
systemctl enable --now glusterfsd glusterd
firewall-cmd --add-service=glusterfs --permanent
firewall-cmd --reload
gluster volume create storage1 storage1.sgtcoder.com:/data/brick1/storage1
gluster volume start storage1
gluster volume status storage1
nano /etc/fstab
## GLUSTER ##
127.0.0.1:/storage1 /mnt/storage glusterfs defaults,_netdev,acl 0 0
systemctl daemon-reload && mount -a
The full output of the command that failed:
Once you reboot, you have to
yes | /usr/sbin/gluster volume stop storage1
/usr/sbin/gluster volume start storage1
/bin/mount -a -t glusterfs
No matter what I do the volumes wont go back online until I force stop the "fake" one multiple times until it says it's unmounted. Then start the volume. Then mount the volume. I tried creating a startup service and everything. I read every single documentation out there and tried hundreds of different ways and different virtual machines and nothing ever seems to work
**Expected results:**
<details>
Every reboot I would expect it to come back online
**Mandatory info:**
**- The output of the `gluster volume info` command**:
[root@storage1 ~]# gluster volume info
Volume Name: storage1
Type: Distribute
Volume ID: 955b7fa1-ff1c-41a2-b467-fffffffff
Status: Started
Snapshot Count: 0
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: storage1:/data/brick1/storage1
Options Reconfigured:
storage.fips-mode-rchecksum: on
transport.address-family: inet
**- The output of the `gluster volume status` command**:
[root@storage1 ~]# gluster volume status
Status of volume: storage1
Gluster process TCP Port RDMA Port Online Pid
Brick storage1:/data/brick1/st
orage1 57447 0 Y 381245
**- The output of the `gluster volume heal` command**:
[root@storage1 ~]# gluster volume heal storage1
Launching heal operation to perform index self heal on volume storage1 has been unsuccessful:
Self-heal-daemon is disabled. Heal will not be triggered on volume storage1
**- Provide logs present on following locations of client and server nodes -
/var/log/glusterfs/
**- Is there any crash ? Provide the backtrace and coredump
No Crash
**Additional info:**
</details>
**- The operating system / glusterfs version**:
AlmaLinux 8 and GlusterFS 10
**Note: Please hide any confidential data which you don't want to share in public like IP address, file name, hostname or any other configuration**