Can you give volume info? Looks like you are using 2 way replica.
On Wed, Sep 19, 2018 at 9:39 AM, Johan Karlsson <Johan.Karlsson@xxxxxx> wrote:
I have two servers setup with glusterFS in replica mode, a single volume exposed via a mountpoint. The servers are running Ubuntu 16.04 LTS
After a package upgrade + reboot of both servers, it was discovered that the data was completely gone. New data written on the volume via the mountpoint is replicated correctly, and gluster status/info commands states that everything is ok (no split brain scenario or any healing needed etc). But the previous data is completely gone, not even present on any of the bricks.
The following upgrade was done:
glusterfs-server:amd64 (4.1.0-ubuntu1~xenial3 -> 4.1.4-ubuntu1~xenial1)
glusterfs-client:amd64 (4.1.0-ubuntu1~xenial3 -> 4.1.4-ubuntu1~xenial1)
glusterfs-common:amd64 (4.1.0-ubuntu1~xenial3 -> 4.1.4-ubuntu1~xenial1)
The logs only show that connection between the servers was lost, which is expected.
I can't even determine if it was the package upgrade or the reboot that caused this issue, but I've tried to recreate the issue without success.
Any idea what could have gone wrong, or if I have done some wrong during the setup. For reference, this is how I've done the setup:
---
Add a separate disk with a single partition on both servers (/dev/sdb1)
Add gfs hostnames for direct communication without DNS, on both servers:
/etc/hosts
192.168.4.45 gfs01
192.168.4.46 gfs02
On gfs01, create a new LVM Volume Group:
vgcreate gfs01-vg /dev/sdb1
And on the gfs02:
vgcreate gfs02-vg /dev/sdb1
Create logical volumes named "brick" on the servers:
gfs01:
lvcreate -l 100%VG -n brick1 gfs01-vg
gfs02:
lvcreate -l 100%VG -n brick2 gfs02-vg
Format the volumes with ext4 filesystem:
gfs01:
mkfs.ext4 /dev/gfs01-vg/brick1
gfs02:
mkfs.ext4 /dev/gfs02-vg/brick2
Create a mountpoint for the bricks on the servers:
gfs01:
mkdir -p /glusterdata/brick1
gds02:
mkdir -p /glusterdata/brick2
Make a permanent mount on the servers:
gfs01:
/dev/gfs01-vg/brick1 /glusterdata/brick1 ext4 defaults 0 0
gfs02:
/dev/gfs02-vg/brick2 /glusterdata/brick2 ext4 defaults 0 0
Mount it:
mount -a
Create a gluster volume mount point on the bricks on the servers:
gfs01:
mkdir -p /glusterdata/brick1/gvol0
gfs02:
mkdir -p /glusterdata/brick2/gvol0
>From each server, peer probe the other one:
gluster peer probe gfs01
peer probe: success
gluster peer probe gfs02
peer probe: success
>From any single server, create the gluster volume as a "replica" with two nodes; gfs01 and gfs02:
gluster volume create gvol0 replica 2 gfs01:/glusterdata/brick1/gvol0 gfs02:/glusterdata/brick2/ gvol0
Start the volume:
gluster volume start gvol0
On each server, mount the gluster filesystem on the /filestore mount point:
gfs01:
mount -t glusterfs gfs01:/gvol0 /filestore
gfs02:
mount -t glusterfs gfs02:/gvol0 /filestore
Make the mount permanent on the servers:
/etc/fstab
gfs01:
gfs01:/gvol0 /filestore glusterfs defaults,_netdev 0 0
gfs02:
gfs02:/gvol0 /filestore glusterfs defaults,_netdev 0 0
---
Regards,
Johan Karlsson
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users