Hello,
I have a question about volume rebalancing and modify/change timestamps.
We're running Gluster 5.11 on CentOS 7.
We recently added an 8th node to our 7 node distrubute cluster. We ran
the necessary fix-layout and rebalance commands after adding the new
brick, and the storage usage balanced out as expected.
However, we had some unexpected behavior from our backup clients. We use
Tivoli Storage Manager (TSM) to backup this volume to tape, and we're
backing up from the volume mountpoint.
We basically saw a large number of files and directories (around the
number that got moved in the rebalance) get re-backed up despite the
files not changing... This pushed our backup footprint up nearly 40TB.
When investigating some of the files we saw that the Modify date hadn't
changed, but the "Change" time had. For directories the Modify and
Change dates were updated. This caused the backup client to think the
files had changed... See below:
[root@gluster01 ~]# stat
/storage/data/projects/comp_support/rat/data/basemaps/bc_16_0_0_0.png
File:
‘/storage/data/projects/comp_support/rat/data/basemaps/bc_16_0_0_0.png’
Size: 2587 Blocks: 6 IO Block: 131072 regular file
Device: 29h/41d Inode: 13389859243885309381 Links: 1
Access: (0664/-rw-rw-r--) Uid: (69618/bveerman) Gid: ( 50/ ftp)
Context: system_u:object_r:fusefs_t:s0
Access: 2020-03-30 10:32:32.326169725 -0700
Modify: 2014-11-24 17:16:57.000000000 -0800
Change: 2020-03-13 21:52:41.158610077 -0700
Birth: -
[root@gluster01 ~]# stat
/storage/data/projects/comp_support/rat/data/basemaps
File: ‘/storage/data/projects/comp_support/rat/data/basemaps’
Size: 4096 Blocks: 8 IO Block: 131072 directory
Device: 29h/41d Inode: 13774747307766344103 Links: 2
Access: (2775/drwxrwsr-x) Uid: (69618/bveerman) Gid: ( 50/ ftp)
Context: system_u:object_r:fusefs_t:s0
Access: 2020-04-01 03:20:58.644695834 -0700
Modify: 2020-03-14 00:51:31.120718996 -0700
Change: 2020-03-14 00:51:31.384725500 -0700
Birth: -
If we look at the files in TSM we find that they were backed up because
the Inode changed. Is this expected for rebalancing? Or is there
something else going on here?
Size Backup Date Mgmt Class
A/I File
---- ----------- ---------- --- ----
2,587 B 2020-03-16 12:14:11 DEFAULT A
/storage/data/projects/comp_support/rat/data/basemaps/bc_16_0_0_0.png
Modified: 2014-11-24 17:16:57 Accessed: 2020-03-13 21:52:41
Inode changed: 2020-03-13 21:52:41
Compression Type: None Encryption Type: None
Client-deduplicated: NO Migrated: NO Inode#: 809741765
Media Class: Library Volume ID: 0375 Restore Order:
00000000-00003684-00000000-0046F92D
2,587 B 2019-10-18 17:01:22 DEFAULT I
/storage/data/projects/comp_support/rat/data/basemaps/bc_16_0_0_0.png
Modified: 2014-11-24 17:16:57 Accessed: 2019-08-08 00:22:50
Inode changed: 2019-08-07 10:55:21
Compression Type: None Encryption Type: None
Client-deduplicated: NO Migrated: NO Inode#: 809741765
Media Class: Library Volume ID: 33040 Restore Order:
00000000-0000D9EB-00000000-000890E3
Volume details:
[root@gluster01 ~]# df -h /storage/
Filesystem Size Used Avail Use% Mounted on
10.0.231.50:/storage 291T 210T 82T 72% /storage
[root@gluster01 ~]# gluster --version
glusterfs 5.11
Repository revision: git://git.gluster.org/glusterfs.git
Copyright (c) 2006-2016 Red Hat, Inc. <https://www.gluster.org/>
GlusterFS comes with ABSOLUTELY NO WARRANTY.
It is licensed to you under your choice of the GNU Lesser
General Public License, version 3 or any later version (LGPLv3
or later), or the GNU General Public License, version 2 (GPLv2),
in all cases as published by the Free Software Foundation.
[root@gluster01 ~]# cat /proc/mounts | egrep "/storage|raid6-storage"
/dev/sda1 /mnt/raid6-storage xfs
rw,seclabel,relatime,attr2,inode64,noquota 0 0
10.0.231.50:/storage /storage fuse.glusterfs
rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072
0 0
[root@gluster01 ~]# cat /etc/fstab | egrep "/storage|raid6-storage"
UUID=104f089e-6171-4750-a592-d41759c67f0c /mnt/raid6-storage xfs
defaults 0 0
10.0.231.50:/storage /storage glusterfs
defaults,log-level=WARNING,backupvolfile-server=10.0.231.51 0 0
[root@gluster01 ~]# gluster volume info storage
Volume Name: storage
Type: Distribute
Volume ID: 6f95525a-94d7-4174-bac4-e1a18fe010a2
Status: Started
Snapshot Count: 0
Number of Bricks: 8
Transport-type: tcp
Bricks:
Brick1: 10.0.231.50:/mnt/raid6-storage/storage
Brick2: 10.0.231.51:/mnt/raid6-storage/storage
Brick3: 10.0.231.52:/mnt/raid6-storage/storage
Brick4: 10.0.231.53:/mnt/raid6-storage/storage
Brick5: 10.0.231.54:/mnt/raid6-storage/storage
Brick6: 10.0.231.55:/mnt/raid6-storage/storage
Brick7: 10.0.231.56:/mnt/raid6-storage/storage
Brick8: 10.0.231.57:/mnt/raid6-storage/storage
Options Reconfigured:
features.read-only: off
features.inode-quota: on
features.quota: on
performance.readdir-ahead: on
nfs.disable: on
geo-replication.indexing: on
geo-replication.ignore-pid-check: on
transport.address-family: inet
features.quota-deem-statfs: on
changelog.changelog: on
diagnostics.client-log-level: INFO
Thanks,
-Matthew
________
Community Meeting Calendar:
Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://bluejeans.com/441850968
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users