So I have now followed your procedure and I have the feeling that it worked because I find the same amount of files on the geo-rep slave node as on the master node. But there where quite a lot of operation not permitted and no such file or directory warnings on the slave node geo-rep log file. Also on the master the geo-rep status detail command outputs 1335 under the FAILUREs column. After the resync I tried to add a new file and see if it appeared on the slave node but that did not work at all and the slave node shows the following warnings: [2016-02-26 20:20:08.833047] W [fuse-bridge.c:1989:fuse_create_cbk] 0-glusterfs-fuse: 1979732: /.gfid/3483d2fb-ab26-46f0-adc3-75391a96b369 => -1 (Operation not permitted) [2016-02-26 20:20:09.507067] W [MSGID: 114031] [client-rpc-fops.c:2971:client3_3_lookup_cbk] 0-woelkli-pro-geo-client-0: remote operation failed. Path: <gfid:b515e589-1302-483e-b5a8-a767b0c5b3d0> (b515e589-1302-483e-b5a8-a767b0c5b3d0) [No such file or directory] [2016-02-26 20:20:09.939786] W [fuse-bridge.c:1989:fuse_create_cbk] 0-glusterfs-fuse: 1979794: /.gfid/3483d2fb-ab26-46f0-adc3-75391a96b369 => -1 (Operation not permitted) [2016-02-26 20:20:11.053035] W [fuse-bridge.c:1989:fuse_create_cbk] 0-glusterfs-fuse: 1979831: /.gfid/3483d2fb-ab26-46f0-adc3-75391a96b369 => -1 (Operation not permitted) [2016-02-26 20:20:12.077440] W [fuse-bridge.c:1989:fuse_create_cbk] 0-glusterfs-fuse: 1979874: /.gfid/3483d2fb-ab26-46f0-adc3-75391a96b369 => -1 (Operation not permitted) [2016-02-26 20:20:13.129645] W [fuse-bridge.c:1989:fuse_create_cbk] 0-glusterfs-fuse: 1979917: /.gfid/3483d2fb-ab26-46f0-adc3-75391a96b369 => -1 (Operation not permitted) [2016-02-26 20:20:14.243425] W [fuse-bridge.c:1989:fuse_create_cbk] 0-glusterfs-fuse: 1979960: /.gfid/3483d2fb-ab26-46f0-adc3-75391a96b369 => -1 (Operation not permitted) [2016-02-26 20:20:15.267638] W [fuse-bridge.c:1989:fuse_create_cbk] 0-glusterfs-fuse: 1980003: /.gfid/3483d2fb-ab26-46f0-adc3-75391a96b369 => -1 (Operation not permitted) [2016-02-26 20:20:16.321249] W [fuse-bridge.c:1989:fuse_create_cbk] 0-glusterfs-fuse: 1980046: /.gfid/3483d2fb-ab26-46f0-adc3-75391a96b369 => -1 (Operation not permitted) [2016-02-26 20:20:17.433468] W [fuse-bridge.c:1989:fuse_create_cbk] 0-glusterfs-fuse: 1980089: /.gfid/3483d2fb-ab26-46f0-adc3-75391a96b369 => -1 (Operation not permitted) [2016-02-26 20:20:18.459165] W [fuse-bridge.c:1989:fuse_create_cbk] 0-glusterfs-fuse: 1980132: /.gfid/3483d2fb-ab26-46f0-adc3-75391a96b369 => -1 (Operation not permitted) Any idea what is wrong now? Also for the point number 3 in your steps I wasn't sure if I should delete the changelogs of all master nodes bricks but that's what I did so I delete the whole changelogs directory on my two nodes bricks. Finally for point number 5 of your steps I have delete the whole data as well as the .glusterfs and .trashcan directory on the volume of my slave node, I hope that was correct. Regards ML On Friday, February 26, 2016 5:54 AM, Aravinda <avishwan@xxxxxxxxxx> wrote: regards Aravinda On 02/26/2016 12:30 AM, ML mail wrote: > Hi Aravinda, > > Many thanks for the steps. I have a few questions about it: > > - in your point number 3, can I simply do an "rm -rf /my/brick/.glusterfs/changelogs" ? Yes > - in your point number 4, do I need to remove the attributes from every master nodes? Yes, from all the bricks of Master Volume. > - i am currently using GlusterFS 3.7.6, will this procedure work with version 3.7.6 or do I need version 3.7.8 or 3.7.9? It should work. > > > Regards > ML > > > > On Thursday, February 25, 2016 11:57 AM, Aravinda <avishwan@xxxxxxxxxx> wrote: > Steps to force Geo-rep to sync from beginning > > 1. Stop Geo-replication > 2. Disable the Changelog using `gluster volume set <MASTER VOLNAME> > changelog.changelog off` > 3. Delete all changelogs and htime files from Brick backend of Master > Volume, $BRICK/.glusterfs/changelogs > 4. Remove stime xattrs from all Brick root of Master Volume (setfattr -x > trusted.glusterfs.<MASTERVOL_ID>.<SLAVEVOL_ID>.stime) > Where ID is `gluster volume info <VOLNAME> | grep ID` > 5. Enable changelog again `gluster volume set <MASTER VOLNAME> > changelog.changelog on` > 5. Clean the Slave Volume by deleting all content. > 6. Start Geo-replication as usual, This will start syncing like fresh setup. > > You can try this re-setup with 3.7.9 release(which is expected anytime > before mid March), We are trying to merge Geo-rep patches related to > this issue for glusterfs-3.7.9 > > Geo-rep should cleanup this xattrs when session is deleted, We will work > on that fix in future releases > BUG: https://bugzilla.redhat.com/show_bug.cgi?id=1311926 > > regards > Aravinda > > > On 02/24/2016 09:59 PM, ML mail wrote: >> That would be great thank you. For me it is not an option to delete the volume on my master node (2 nodes, 1 brick per node). On the other hand no problem to delete the volume on the slave node which is only used for geo-rep. >> >> Regards >> ML >> >> >> >> On Wednesday, February 24, 2016 4:44 PM, Aravinda <avishwan@xxxxxxxxxx> wrote: >> We can provide workaround steps to resync from beginning without >> deleting Volume(s). >> >> I will send the Session reset details by tomorrow. >> >> regards >> Aravinda >> >> On 02/24/2016 09:08 PM, ML mail wrote: >>> That's right I saw already a few error messages mentioning "Device or resource busy" and was wondering what it was... >>> >>> You mean I have to delete the brick on my slave node, delete the volume on my slave node and finally re-create the volume on my slave node in order to start geo-replication from the beginning again? I do not have to touch or delete anything on the master node, right? >>> >>> >>> Regards >>> ML >>> >>> >>> >>> On Wednesday, February 24, 2016 3:07 PM, Milind Changire <mchangir@xxxxxxxxxx> wrote: >>> ML, >>> Since the fixes to geo-rep are yet to get into a release, >>> I can only suggest you to be a bit patient. >>> Also, since you are using logrotate to rotate logs, you >>> will most likely get into the "No such file or directory" >>> or "Device or resource busy" scenario on the slave again. >>> I'm not saying logrotate is at fault, I'm just saying that >>> that specific use case leads to an inconsistent gluster >>> state. >>> >>> Unfortunately, you cannot selectively purge the changelogs. >>> You will have to delete the volume and empty the bricks >>> and recreate the volume with the empty bricks to start >>> all over again. >>> >>> You can delete the volume with: >>> # gluster volume stop <volume name> >>> # gluster volume delete <volume name> >>> >>> -- >>> Milind >>> >>> >>> ----- Original Message ----- >>> From: "ML mail" <mlnospam@xxxxxxxxx> >>> To: "Milind Changire" <mchangir@xxxxxxxxxx> >>> Cc: "Gluster-users" <gluster-users@xxxxxxxxxxx> >>> Sent: Wednesday, February 24, 2016 4:44:27 PM >>> Subject: Re: geo-rep: remote operation failed - No such file or directory >>> >>> Thanks Milind again for your help. I understand now the concept and managed to set the required attribute for forcing the resyncing. That worked but unfortunately it is a never ending story, I fix stuff, start geo-rep it goes for a few more files and fails again. >>> >>> Now I think it will be easier to reset geo-replication and start from scratch again, luckily my volume is only 16 GB big as I am still experimenting. What would be the correct way to reset geo-rep? I don't want to remove the config but I would like to trash all the changelogs, delete the whole data on the slave and re-start geo-rep. How should I proceed? >>> >>> Regards >>> ML >>> >>> >>> >>> >>> On Wednesday, February 24, 2016 10:14 AM, Milind Changire <mchangir@xxxxxxxxxx> wrote: >>> 1. You could use the script at >>> https://gist.github.com/aravindavk/afb16813261794faa432 >>> to create a path from the gfid that you could cd to >>> i.e. for gfid c4b19f1c-cc18-4727-87a4-18de8fe0089e >>> >>> 2. yes, you have to recursively set the virtual xattr >>> on all entries in the directory tree >>> Also, remember to set a value as well >>> # setfattr -n glusterfs.geo-rep.trigger-sync -v 1 <file-path> >>> >>> Also, remember to set the virtual xattr via the volume >>> mount path and not the brick back-end path. >>> You should have geo-replication stopped when you are >>> setting the virtual xattr and start it when you are >>> done setting the xattr for the entire directory tree. >>> >>> -- >>> Milind >>> >>> >>> ----- Original Message ----- >>> From: "ML mail" <mlnospam@xxxxxxxxx> >>> To: "Milind Changire" <mchangir@xxxxxxxxxx> >>> Cc: "Gluster-users" <gluster-users@xxxxxxxxxxx> >>> Sent: Wednesday, February 24, 2016 1:46:11 PM >>> Subject: Re: geo-rep: remote operation failed - No such file or directory >>> >>> Thank you for explaining me how the symbolic linking works in the the .glusterfs directory. Now regarding your new instructions I have two questions: >>> >>> 1) How can I find out which directory "OC_DEFAULT_MODULE" on my master brick I should run the >>> setfattr command on? My problem here is that there are a lot of OC_DEFAULT_MODULE directories on my brick not just only a single one. >>> >>> >>> >>> 2) If I understand your last paragraph correctly, you want me to locate the correct OC_DEFAULT_MODULE directory and recursively use setfattr on each sub-directories and/or files inside that directory, is this correct? >>> >>> Regards >>> ML >>> >>> >>> >>> On Wednesday, February 24, 2016 7:29 AM, Milind Changire <mchangir@xxxxxxxxxx> wrote: >>> ML, >>> You just need to worry about the very first entry that you found with >>> the find command: >>> >>> $ find .glusterfs -name c4b19f1c-cc18-4727-87a4-18de8fe0089e -ls >>> 228215 0 lrwxrwxrwx 1 root root 66 Feb 19 08:52 .glusterfs/c4/b1/c4b19f1c-cc18-4727-87a4-18de8fe0089e -> ../../92/1b/921bfe8e-81ef-4579-b335-abfa2c7e6afb/OC_DEFAULT_MODULE >>> >>> Since the back-end entry is a symlink, it means that OC_DEFAULT_MODULE >>> is a directory on the master and it is missing on the slave. >>> If you try to recursively look at the parent gfids of each of the entries >>> then they will always point to symlinks since a directory is always >>> represented as a symlink at the glusterfs back-end, and you will follow >>> them up to the ROOT gfid. >>> >>> ----- >>> >>> Now, to get the OC_DEFAULT_MODULE directory replicated on the slave, >>> you will have to set the virtual xattr on the entire directory tree >>> in pre-order listing i.e. set the virtual xattr on the directory >>> starting at OC_DEFAULT_MODULE and then on the entries inside the >>> directory, and so on down the directory tree. >>> >>> -- >>> Milind >>> >>> >>> ----- Original Message ----- >>> From: "ML mail" <mlnospam@xxxxxxxxx> >>> To: "Milind Changire" <mchangir@xxxxxxxxxx> >>> Cc: "Gluster-users" <gluster-users@xxxxxxxxxxx> >>> Sent: Wednesday, February 24, 2016 12:25:26 AM >>> Subject: Re: geo-rep: remote operation failed - No such file or directory >>> >>> Hi Milind, >>> >>> Thanks for the instructions for forcing the data sync of a specific file. I was not able to do that as I have discovered something even more weird by trying to find out the concerned file by GFID with the find command as you suggested. Indeed it looks like I have a symbolic link pointing to another one and then to another and so on, as you can see below: >>> >>> $ find .glusterfs -name c4b19f1c-cc18-4727-87a4-18de8fe0089e -ls >>> 228215 0 lrwxrwxrwx 1 root root 66 Feb 19 08:52 .glusterfs/c4/b1/c4b19f1c-cc18-4727-87a4-18de8fe0089e -> ../../92/1b/921bfe8e-81ef-4579-b335-abfa2c7e6afb/OC_DEFAULT_MODULE >>> >>> $ ls -la 92/1b/921bfe8e-81ef-4579-b335-abfa2c7e6afb >>> lrwxrwxrwx 1 root root 79 Feb 19 08:52 92/1b/921bfe8e-81ef-4579-b335-abfa2c7e6afb -> ../../d7/9f/d79f2ebd-029c-4ac5-8074-5eef7ff21236/160201_File_1602_XX.xls >>> >>> >>> $ ls -la d7/9f/d79f2ebd-029c-4ac5-8074-5eef7ff21236 >>> lrwxrwxrwx 1 root root 53 Feb 15 07:34 d7/9f/d79f2ebd-029c-4ac5-8074-5eef7ff21236 -> ../../fd/ea/fdea1fc6-0f2a-43d2-8776-651cc6ea73e8/1602 >>> >>> >>> $ ls -la fd/ea/fdea1fc6-0f2a-43d2-8776-651cc6ea73e8 >>> lrwxrwxrwx 1 root root 55 Feb 15 07:29 fd/ea/fdea1fc6-0f2a-43d2-8776-651cc6ea73e8 -> ../../20/25/20253364-add8-4149-a7cf-cf46d237a45c/Banana >>> >>> >>> Is this normal? I somehow don't understand this weird structure of never ending symbolic links... or am I missing something? >>> >>> >>> Regards >>> ML >>> >>> >>> >>> On Tuesday, February 23, 2016 6:31 AM, Milind Changire <mchangir@xxxxxxxxxx> wrote: >>> ML, >>> You will have to search for the gfid c4b19f1c-cc18-4727-87a4-18de8fe0089e >>> at the master cluster brick back-ends and run the following command for >>> that specific file on the master cluster to force triggering a data sync [1] >>> >>> # setfattr -n glusterfs.geo-rep.trigger-sync <file-path> >>> >>> To search for the file at the brick back-end: >>> >>> # find /<path-to-brick>/.glusterfs -name c4b19f1c-cc18-4727-87a4-18de8fe0089e >>> >>> Once path to the file is found at any of the bricks, you can then use >>> the setfattr command described above. >>> >>> Reference: >>> [1] feature/changelog: Virtual xattr to trigger explicit sync in geo-rep >>> http://review.gluster.org/#/c/9337/ >>> -- >>> Milind >>> >>> >>> ----- Original Message ----- >>> From: "ML mail" <mlnospam@xxxxxxxxx> >>> To: "Milind Changire" <mchangir@xxxxxxxxxx> >>> Cc: "Gluster-users" <gluster-users@xxxxxxxxxxx> >>> Sent: Monday, February 22, 2016 9:10:56 PM >>> Subject: Re: geo-rep: remote operation failed - No such file or directory >>> >>> Hi Milind, >>> >>> Thanks for the suggestion, I did that for a few problematic files and it seems to continue but now I am stuck at the following error message on the slave: >>> >>> [2016-02-22 15:21:30.451133] W [MSGID: 114031] [client-rpc-fops.c:2971:client3_3_lookup_cbk] 0-myvolume-geo-client-0: remote operation failed. Path: <gfid:c4b19f1c-cc18-4727-87a4-18de8fe0089e> (c4b19f1c-cc18-4727-87a4-18de8fe0089e) [No such file or directory] >>> >>> As you can see this message does not include any file or directory name, so I can't go any delete that file or directory. Any other ideas how I may proceed here? >>> >>> Or maybe would it be easier if I delete the whole directory which I think is affected and start geo-rep from there? Or will this mess things up? >>> >>> Regards >>> ML >>> >>> >>> >>> On Monday, February 22, 2016 12:12 PM, Milind Changire <mchangir@xxxxxxxxxx> wrote: >>> ML, >>> You could try deleting problematic files on slave to recover geo-replication >>> from Faulty state. >>> >>> However, changelogs generated due to logrotate scenario will still cause >>> geo-replication to go into Faulty state frequently if geo-replication >>> fails and restarts. >>> >>> The patches mentioned in an earlier mail are being worked upon and finalized. >>> They will be available soon in a release which will avoid geo-replication >>> going into a Faulty state. >>> >>> -- >>> Milind >>> >>> >>> ----- Original Message ----- >>> From: "ML mail" <mlnospam@xxxxxxxxx> >>> To: "Milind Changire" <mchangir@xxxxxxxxxx>, "Gluster-users" <gluster-users@xxxxxxxxxxx> >>> Sent: Monday, February 22, 2016 1:27:14 PM >>> Subject: Re: geo-rep: remote operation failed - No such file or directory >>> >>> Hi Milind, >>> >>> Any news on this issue? I was wondering how can I fix and restart my geo-replication? Can I simply delete the problematic file(s) on my slave and restart geo-rep? >>> >>> Regards >>> ML >>> >>> >>> >>> >>> >>> On Wednesday, February 17, 2016 4:30 PM, ML mail <mlnospam@xxxxxxxxx> wrote: >>> >>> >>> Hi Milind, >>> >>> Thank you for your short analysis. Indeed that's exactly what happens, as soon as I restart geo-rep it replays the same over and over as it does not succeed. >>> >>> >>> Now regarding the sequence of the file management operations I am not totally sure how it works but I can tell you that we are using ownCloud v8.2.2 (www.owncloud.org) and as storage for this cloud software we use GlusterFS. So it is very probable that ownCloud works like that: when a user uploads a new file if first creates it with another temporary name which it then either renames or moves after successful upload. >>> >>> >>> I have the feeling this issue is related to my initial issue which I have reported earlier this month: >>> https://www.gluster.org/pipermail/gluster-users/2016-February/025176.html >>> >>> For now my question would be how do I get to restart geo-replication succesfully? >>> >>> Regards >>> ML >>> >>> >>> >>> On Wednesday, February 17, 2016 4:10 PM, Milind Changire <mchangir@xxxxxxxxxx> wrote: >>> >>> >>> As per the slave logs, there is an attempt to RENAME files >>> i.e. a .part file getting renamed to a name without the >>> .part suffix >>> >>> Just restarting geo-rep isn't going to help much if >>> you've already hit the problem. Since the last CHANGELOG >>> is replayed by geo-rep on a restart, you'll most probably >>> encounter the same log messages in the logs. >>> >>> Are the .part files CREATEd, RENAMEd and DELETEd with the >>> same name often? Are the operations somewhat in the following >>> sequence that happen on the geo-replication master cluster? >>> >>> CREATE f1.part >>> RENAME f1.part f1 >>> DELETE f1 >>> CREATE f1.part >>> RENAME f1.part f1 >>> ... >>> ... >>> >>> >>> If not, then it would help if you could send the sequence >>> of file management operations. >>> >>> -- >>> Milind >>> >>> >>> ----- Original Message ----- >>> From: "Kotresh Hiremath Ravishankar" <khiremat@xxxxxxxxxx> >>> To: "ML mail" <mlnospam@xxxxxxxxx> >>> Cc: "Gluster-users" <gluster-users@xxxxxxxxxxx>, "Milind Changire" <mchangir@xxxxxxxxxx> >>> Sent: Tuesday, February 16, 2016 6:28:21 PM >>> Subject: Re: geo-rep: remote operation failed - No such file or directory >>> >>> Ccing Milind, he would be able to help >>> >>> Thanks and Regards, >>> Kotresh H R >>> >>> ----- Original Message ----- >>>> From: "ML mail" <mlnospam@xxxxxxxxx> >>>> To: "Gluster-users" <gluster-users@xxxxxxxxxxx> >>>> Sent: Monday, February 15, 2016 4:41:56 PM >>>> Subject: geo-rep: remote operation failed - No such file or directory >>>> >>>> Hello, >>>> >>>> I noticed that the geo-replication of a volume has STATUS "Faulty" and while >>>> looking in the *.gluster.log file in >>>> /var/log/glusterfs/geo-replication-slaves/ on my slave I can see the >>>> following relevant problem: >>>> >>>> [2016-02-15 10:58:40.402516] I [rpc-clnt.c:1847:rpc_clnt_reconfig] >>>> 0-myvolume-geo-client-0: changing port to 49152 (from 0) >>>> [2016-02-15 10:58:40.403928] I [MSGID: 114057] >>>> [client-handshake.c:1437:select_server_supported_programs] >>>> 0-myvolume-geo-client-0: Using Program GlusterFS 3.3, Num (1298437), Version >>>> (330) >>>> [2016-02-15 10:58:40.404130] I [MSGID: 114046] >>>> [client-handshake.c:1213:client_setvolume_cbk] 0-myvolume-geo-client-0: >>>> Connected to myvolume-geo-client-0, attached to remote volume >>>> '/data/myvolume-geo/brick'. >>>> [2016-02-15 10:58:40.404150] I [MSGID: 114047] >>>> [client-handshake.c:1224:client_setvolume_cbk] 0-myvolume-geo-client-0: >>>> Server and Client lk-version numbers are not same, reopening the fds >>>> [2016-02-15 10:58:40.410150] I [fuse-bridge.c:5137:fuse_graph_setup] 0-fuse: >>>> switched to graph 0 >>>> [2016-02-15 10:58:40.410223] I [MSGID: 114035] >>>> [client-handshake.c:193:client_set_lk_version_cbk] 0-myvolume-geo-client-0: >>>> Server lk version = 1 >>>> [2016-02-15 10:58:40.410370] I [fuse-bridge.c:4030:fuse_init] >>>> 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.22 kernel >>>> 7.23 >>>> [2016-02-15 10:58:45.662416] I [MSGID: 109066] [dht-rename.c:1411:dht_rename] >>>> 0-myvolume-geo-dht: renaming >>>> /.gfid/94310944-7f8a-421d-a51f-1e23e28da9cc/Bild_03_Rosen.JPG-chunking-2242590604-0.FpKL3SIUb9vKHyjd.part >>>> (hash=myvolume-geo-client-0/cache=myvolume-geo-client-0) => >>>> /.gfid/94310944-7f8a-421d-a51f-1e23e28da9cc/Bild_03_Rosen.JPG-chunking-2242590604-0 >>>> (hash=myvolume-geo-client-0/cache=<nul>) >>>> [2016-02-15 10:58:45.665144] I [MSGID: 109066] [dht-rename.c:1411:dht_rename] >>>> 0-myvolume-geo-dht: renaming >>>> /.gfid/94310944-7f8a-421d-a51f-1e23e28da9cc/Bild_03_Rosen.JPG-chunking-2242590604-1.C6l0DEurb2y3Azw4.part >>>> (hash=myvolume-geo-client-0/cache=myvolume-geo-client-0) => >>>> /.gfid/94310944-7f8a-421d-a51f-1e23e28da9cc/Bild_03_Rosen.JPG-chunking-2242590604-1 >>>> (hash=myvolume-geo-client-0/cache=<nul>) >>>> [2016-02-15 10:58:45.749829] I [MSGID: 109066] [dht-rename.c:1411:dht_rename] >>>> 0-myvolume-geo-dht: renaming >>>> /.gfid/94310944-7f8a-421d-a51f-1e23e28da9cc/Bild_02_Pilz.JPG-chunking-628343631-0.ajEnSguUZ7EkzjzT.part >>>> (hash=myvolume-geo-client-0/cache=myvolume-geo-client-0) => >>>> /.gfid/94310944-7f8a-421d-a51f-1e23e28da9cc/Bild_02_Pilz.JPG-chunking-628343631-0 >>>> (hash=myvolume-geo-client-0/cache=<nul>) >>>> [2016-02-15 10:58:45.750225] W [MSGID: 114031] >>>> [client-rpc-fops.c:2971:client3_3_lookup_cbk] 0-myvolume-geo-client-0: >>>> remote operation failed. Path: >>>> /.gfid/94310944-7f8a-421d-a51f-1e23e28da9cc/Bild_02_Pilz.JPG-chunking-628343631-0.ajEnSguUZ7EkzjzT.part >>>> (9164caeb-740d-4429-a3bd-c85f40c35e11) [No such file or directory] >>>> [2016-02-15 10:58:45.750418] W [fuse-bridge.c:1777:fuse_rename_cbk] >>>> 0-glusterfs-fuse: 60: >>>> /.gfid/94310944-7f8a-421d-a51f-1e23e28da9cc/Bild_02_Pilz.JPG-chunking-628343631-0.ajEnSguUZ7EkzjzT.part >>>> -> >>>> /.gfid/94310944-7f8a-421d-a51f-1e23e28da9cc/Bild_02_Pilz.JPG-chunking-628343631-0 >>>> => -1 (Device or resource busy) >>>> [2016-02-15 10:58:45.767788] I [fuse-bridge.c:4984:fuse_thread_proc] 0-fuse: >>>> unmounting /tmp/gsyncd-aux-mount-bZ9SMt >>>> [2016-02-15 10:58:45.768063] W [glusterfsd.c:1236:cleanup_and_exit] >>>> (-->/lib/x86_64-linux-gnu/libpthread.so.0(+0x80a4) [0x7feb610820a4] >>>> -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x7feb626f45b5] >>>> -->/usr/sbin/glusterfs(cleanup_and_exit+0x59) [0x7feb626f4429] ) 0-: >>>> received signum (15), shutting down >>>> [2016-02-15 10:58:45.768093] I [fuse-bridge.c:5683:fini] 0-fuse: Unmounting >>>> '/tmp/gsyncd-aux-mount-bZ9SMt'. >>>> [2016-02-15 10:58:54.871855] I [dict.c:473:dict_get] >>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.6/xlator/system/posix-acl.so(posix_acl_setxattr_cbk+0x26) >>>> [0x7f8313dfb166] >>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.6/xlator/system/posix-acl.so(handling_other_acl_related_xattr+0x20) >>>> [0x7f8313dfb060] >>>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(dict_get+0x93) >>>> [0x7f831f3f40c3] ) 0-dict: !this || key=system.posix_acl_access [Invalid >>>> argument] >>>> [2016-02-15 10:58:54.871914] I [dict.c:473:dict_get] >>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.6/xlator/system/posix-acl.so(posix_acl_setxattr_cbk+0x26) >>>> [0x7f8313dfb166] >>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.6/xlator/system/posix-acl.so(handling_other_acl_related_xattr+0xb0) >>>> [0x7f8313dfb0f0] >>>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(dict_get+0x93) >>>> [0x7f831f3f40c3] ) 0-dict: !this || key=system.posix_acl_default [Invalid >>>> argument] >>>> >>>> This error gets repeated forever with always the same files. I tried to stop >>>> and restart the geo-rep on the master but still the same problem and geo >>>> replication does not proceed. Does anyone have an idea how to fix this? >>>> >>>> I am using GlusterFS 3.7.6 on Debian 8 with a two node replicate volume (1 >>>> brick per node) and one single off-site slave node for geo-rep. >>>> >>>> Regards >>>> ML >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users@xxxxxxxxxxx >>>> http://www.gluster.org/mailman/listinfo/gluster-users >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users@xxxxxxxxxxx >>> http://www.gluster.org/mailman/listinfo/gluster-users _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users