Gluster Volume Rebalance inode modify change times

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

I have a question about volume rebalancing and modify/change timestamps. We're running Gluster 5.11 on CentOS 7.

We recently added an 8th node to our 7 node distrubute cluster. We ran the necessary fix-layout and rebalance commands after adding the new brick, and the storage usage balanced out as expected.

However, we had some unexpected behavior from our backup clients. We use Tivoli Storage Manager (TSM) to backup this volume to tape, and we're backing up from the volume mountpoint.

We basically saw a large number of files and directories (around the number that got moved in the rebalance) get re-backed up despite the files not changing... This pushed our backup footprint up nearly 40TB.

When investigating some of the files we saw that the Modify date hadn't changed, but the "Change" time had. For directories the Modify and Change dates were updated. This caused the backup client to think the files had changed... See below:

[root@gluster01 ~]# stat /storage/data/projects/comp_support/rat/data/basemaps/bc_16_0_0_0.png   File: ‘/storage/data/projects/comp_support/rat/data/basemaps/bc_16_0_0_0.png’
  Size: 2587          Blocks: 6          IO Block: 131072 regular file
Device: 29h/41d    Inode: 13389859243885309381  Links: 1
Access: (0664/-rw-rw-r--)  Uid: (69618/bveerman)   Gid: ( 50/     ftp)
Context: system_u:object_r:fusefs_t:s0
Access: 2020-03-30 10:32:32.326169725 -0700
Modify: 2014-11-24 17:16:57.000000000 -0800
Change: 2020-03-13 21:52:41.158610077 -0700
 Birth: -

[root@gluster01 ~]# stat /storage/data/projects/comp_support/rat/data/basemaps
  File: ‘/storage/data/projects/comp_support/rat/data/basemaps’
  Size: 4096          Blocks: 8          IO Block: 131072 directory
Device: 29h/41d    Inode: 13774747307766344103  Links: 2
Access: (2775/drwxrwsr-x)  Uid: (69618/bveerman)   Gid: ( 50/     ftp)
Context: system_u:object_r:fusefs_t:s0
Access: 2020-04-01 03:20:58.644695834 -0700
Modify: 2020-03-14 00:51:31.120718996 -0700
Change: 2020-03-14 00:51:31.384725500 -0700
 Birth: -


If we look at the files in TSM we find that they were backed up because the Inode changed. Is this expected for rebalancing? Or is there something else going on here?


           Size        Backup Date                Mgmt Class           A/I File
           ----        ----------- ----------           --- ----
         2,587  B  2020-03-16 12:14:11 DEFAULT              A /storage/data/projects/comp_support/rat/data/basemaps/bc_16_0_0_0.png          Modified: 2014-11-24 17:16:57  Accessed: 2020-03-13 21:52:41  Inode changed: 2020-03-13 21:52:41          Compression Type: None  Encryption Type: None  Client-deduplicated: NO  Migrated: NO  Inode#: 809741765   Media Class: Library  Volume ID: 0375  Restore Order: 00000000-00003684-00000000-0046F92D          2,587  B  2019-10-18 17:01:22 DEFAULT              I /storage/data/projects/comp_support/rat/data/basemaps/bc_16_0_0_0.png          Modified: 2014-11-24 17:16:57  Accessed: 2019-08-08 00:22:50  Inode changed: 2019-08-07 10:55:21          Compression Type: None  Encryption Type: None  Client-deduplicated: NO  Migrated: NO  Inode#: 809741765   Media Class: Library  Volume ID: 33040  Restore Order: 00000000-0000D9EB-00000000-000890E3



Volume details:


[root@gluster01 ~]# df -h /storage/
Filesystem            Size  Used Avail Use% Mounted on
10.0.231.50:/storage  291T  210T   82T  72% /storage

[root@gluster01 ~]# gluster --version
glusterfs 5.11
Repository revision: git://git.gluster.org/glusterfs.git
Copyright (c) 2006-2016 Red Hat, Inc. <https://www.gluster.org/>
GlusterFS comes with ABSOLUTELY NO WARRANTY.
It is licensed to you under your choice of the GNU Lesser
General Public License, version 3 or any later version (LGPLv3
or later), or the GNU General Public License, version 2 (GPLv2),
in all cases as published by the Free Software Foundation.

[root@gluster01 ~]# cat /proc/mounts | egrep "/storage|raid6-storage"
/dev/sda1 /mnt/raid6-storage xfs rw,seclabel,relatime,attr2,inode64,noquota 0 0 10.0.231.50:/storage /storage fuse.glusterfs rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072 0 0

[root@gluster01 ~]# cat /etc/fstab | egrep "/storage|raid6-storage"
UUID=104f089e-6171-4750-a592-d41759c67f0c    /mnt/raid6-storage xfs defaults 0 0 10.0.231.50:/storage /storage glusterfs defaults,log-level=WARNING,backupvolfile-server=10.0.231.51 0 0

[root@gluster01 ~]# gluster volume info storage

Volume Name: storage
Type: Distribute
Volume ID: 6f95525a-94d7-4174-bac4-e1a18fe010a2
Status: Started
Snapshot Count: 0
Number of Bricks: 8
Transport-type: tcp
Bricks:
Brick1: 10.0.231.50:/mnt/raid6-storage/storage
Brick2: 10.0.231.51:/mnt/raid6-storage/storage
Brick3: 10.0.231.52:/mnt/raid6-storage/storage
Brick4: 10.0.231.53:/mnt/raid6-storage/storage
Brick5: 10.0.231.54:/mnt/raid6-storage/storage
Brick6: 10.0.231.55:/mnt/raid6-storage/storage
Brick7: 10.0.231.56:/mnt/raid6-storage/storage
Brick8: 10.0.231.57:/mnt/raid6-storage/storage
Options Reconfigured:
features.read-only: off
features.inode-quota: on
features.quota: on
performance.readdir-ahead: on
nfs.disable: on
geo-replication.indexing: on
geo-replication.ignore-pid-check: on
transport.address-family: inet
features.quota-deem-statfs: on
changelog.changelog: on
diagnostics.client-log-level: INFO

Thanks,
 -Matthew
________



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://bluejeans.com/441850968

Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users




[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux