Issue with Gluster Quota

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



After further tests, it appears both now work after the update.  Seems
the attributes set while using the git source build on the 'brs'
directory were gummed up.  When I recreated the directory and remounted
with -o acl, ACLs worked and so did quota enforcement.  I'll keep
testing and post if anything else comes up.  So far, so good.

-Brian

Brian Smith
Senior Systems Administrator
IT Research Computing, University of South Florida
4202 E. Fowler Ave. ENB308
Office Phone: +1 813 974-1467
Organization URL: http://rc.usf.edu

On 07/18/2011 11:57 AM, Brian Smith wrote:
> Updated to the 3.2.2 release and found that Quotas do not work when
> using the POSIX-ACL translator.  In fact, after I disabled the ACLs, I
> had to remove the quota'd directory (presumably, to remove some
> attributes) and start over in order to get them to work.  Once I
> disabled ACLs and re-created my directory, quotas worked as expected.
> Is this a known limitation of using POSIX ACLs?  I happen to need both
> features, so that could pose an issue :)
> 
> -Brian
> 
> Brian Smith
> Senior Systems Administrator
> IT Research Computing, University of South Florida
> 4202 E. Fowler Ave. ENB308
> Office Phone: +1 813 974-1467
> Organization URL: http://rc.usf.edu
> 
> On 07/11/2011 12:48 PM, Brian Smith wrote:
>> According to the logs, the last commit was:
>>
>> commit 5c20eb3bbf870edadd22d06babb5d38dad222533
>> Author: shishir gowda <shishirng at gluster.com>
>> Date:   Tue Jul 5 03:41:51 2011 +0000
>>
>> [root at gluster1 glusterfs-3.2git]# gluster volume quota home list
>> 	path		  limit_set	     size
>> ----------------------------------------------------------------------------------
>> /brs                   10485760             81965056
>>
>> [root at gluster1 glusterfs-3.2git]# gluster volume info
>>
>> Volume Name: home
>> Type: Distribute
>> Status: Started
>> Number of Bricks: 2
>> Transport-type: tcp,rdma
>> Bricks:
>> Brick1: gluster1:/glusterfs/home
>> Brick2: gluster2:/glusterfs/home
>> Options Reconfigured:
>> features.limit-usage: /brs:10MB
>> features.quota: on
>>
>> -Brian
>>
>> Brian Smith
>> Senior Systems Administrator
>> IT Research Computing, University of South Florida
>> 4202 E. Fowler Ave. ENB308
>> Office Phone: +1 813 974-1467
>> Organization URL: http://rc.usf.edu
>>
>> On 07/11/2011 08:29 AM, Saurabh Jain wrote:
>>> Hello Brian,
>>>
>>>
>>>   I synced my gluster repository back to July 5th and tried quota on a certain dir of a distribute and the quota was implemeted properly on that, here are the logs,
>>>
>>>    [root at centos-qa-client-3 glusterfs]# /root/july6git/inst/sbin/gluster volume quota dist list
>>> 	path		  limit_set	     size
>>> ----------------------------------------------------------------------------------
>>> /dir                   10485760             10485760
>>>
>>>
>>> [root at centos-qa-client-3 glusterfs]# /root/july6git/inst/sbin/gluster volume info
>>>
>>> Volume Name: dist
>>> Type: Distribute
>>> Status: Started
>>> Number of Bricks: 2
>>> Transport-type: tcp,rdma
>>> Bricks:
>>> Brick1: 10.1.12.134:/mnt/dist
>>> Brick2: 10.1.12.135:/mnt/dist
>>> Options Reconfigured:
>>> features.limit-usage: /dir:10MB
>>> features.quota: on
>>> [root at centos-qa-client-3 glusterfs]# 
>>>
>>> requesting you to please inform us about the <commit id> to which your workspace is synced.
>>>
>>> Thanks,
>>> Saurabh
>>> ________________________________________
>>> From: gluster-users-bounces at gluster.org [gluster-users-bounces at gluster.org] on behalf of gluster-users-request at gluster.org [gluster-users-request at gluster.org]
>>> Sent: Friday, July 08, 2011 12:30 AM
>>> To: gluster-users at gluster.org
>>> Subject: Gluster-users Digest, Vol 39, Issue 13
>>>
>>> Send Gluster-users mailing list submissions to
>>>         gluster-users at gluster.org
>>>
>>> To subscribe or unsubscribe via the World Wide Web, visit
>>>         http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>> or, via email, send a message with subject or body 'help' to
>>>         gluster-users-request at gluster.org
>>>
>>> You can reach the person managing the list at
>>>         gluster-users-owner at gluster.org
>>>
>>> When replying, please edit your Subject line so it is more specific
>>> than "Re: Contents of Gluster-users digest..."
>>>
>>>
>>> Today's Topics:
>>>
>>>    1. Re: Issue with Gluster Quota (Brian Smith)
>>>    2. Re: Issues with geo-rep (Carl Chenet)
>>>
>>>
>>> ----------------------------------------------------------------------
>>>
>>> Message: 1
>>> Date: Thu, 07 Jul 2011 13:10:06 -0400
>>> From: Brian Smith <brs at usf.edu>
>>> Subject: Re: Issue with Gluster Quota
>>> To: gluster-users at gluster.org
>>> Message-ID: <4E15E86E.6030407 at usf.edu>
>>> Content-Type: text/plain; charset=ISO-8859-1
>>>
>>> Sorry about that.  I re-populated with an 82MB dump from dd:
>>>
>>> [root at gluster1 ~]# gluster volume quota home list
>>>         path              limit_set          size
>>> ----------------------------------------------------------------------------------
>>> /brs                   10485760             81965056
>>>
>>> [root at gluster1 ~]# getfattr -m . -d -e hex /glusterfs/home/brs
>>> getfattr: Removing leading '/' from absolute path names
>>> # file: glusterfs/home/brs
>>> security.selinux=0x726f6f743a6f626a6563745f723a66696c655f743a733000
>>> trusted.gfid=0x1bbcb9a08bf64406b440f3bb3ad334ed
>>> trusted.glusterfs.dht=0x00000001000000007fffffffffffffff
>>> trusted.glusterfs.quota.00000000-0000-0000-0000-000000000001.contri=0x0000000000006000
>>> trusted.glusterfs.quota.dirty=0x3000
>>> trusted.glusterfs.quota.size=0x0000000000006000
>>>
>>> [root at gluster2 ~]# getfattr -m . -d -e hex /glusterfs/home/brs
>>> getfattr: Removing leading '/' from absolute path names
>>> # file: glusterfs/home/brs
>>> security.selinux=0x726f6f743a6f626a6563745f723a66696c655f743a733000
>>> trusted.gfid=0x1bbcb9a08bf64406b440f3bb3ad334ed
>>> trusted.glusterfs.dht=0x0000000100000000000000007ffffffe
>>> trusted.glusterfs.quota.00000000-0000-0000-0000-000000000001.contri=0x0000000004e25000
>>> trusted.glusterfs.quota.dirty=0x3000
>>> trusted.glusterfs.quota.size=0x0000000004e25000
>>>
>>> Brian Smith
>>> Senior Systems Administrator
>>> IT Research Computing, University of South Florida
>>> 4202 E. Fowler Ave. ENB308
>>> Office Phone: +1 813 974-1467
>>> Organization URL: http://rc.usf.edu
>>>
>>> On 07/07/2011 04:50 AM, Mohammed Junaid wrote:
>>>>>
>>>>> [root at gluster1 ~]# getfattr -m . -d -e hex /glusterfs/home/brs
>>>>> getfattr: Removing leading '/' from absolute path names
>>>>> # file: glusterfs/home/brs
>>>>> security.selinux=0x726f6f743a6f626a6563745f723a66696c655f743a733000
>>>>> trusted.gfid=0x1bbcb9a08bf64406b440f3bb3ad334ed
>>>>> trusted.glusterfs.dht=0x00000001000000007fffffffffffffff
>>>>>
>>>>> trusted.glusterfs.quota.00000000-0000-0000-0000-000000000001.contri=0x0000000000006000
>>>>> trusted.glusterfs.quota.dirty=0x3000
>>>>> trusted.glusterfs.quota.size=0x0000000000006000
>>>>>
>>>>> and
>>>>>
>>>>> [root at gluster2 ~]# getfattr -m . -d -e hex /glusterfs/home/brs
>>>>> getfattr: Removing leading '/' from absolute path names
>>>>> # file: glusterfs/home/brs
>>>>> security.selinux=0x726f6f743a6f626a6563745f723a66696c655f743a733000
>>>>> trusted.gfid=0x1bbcb9a08bf64406b440f3bb3ad334ed
>>>>> trusted.glusterfs.dht=0x0000000100000000000000007ffffffe
>>>>>
>>>>> trusted.glusterfs.quota.00000000-0000-0000-0000-000000000001.contri=0x0000000000002000
>>>>> trusted.glusterfs.quota.dirty=0x3000
>>>>> trusted.glusterfs.quota.size=0x0000000000002000
>>>>
>>>>
>>>> trusted.glusterfs.quota.size=0x0000000000006000
>>>> trusted.glusterfs.quota.size=0x0000000000002000
>>>>
>>>> So, quota adds these values to calculate the size of the directory. They are
>>>> in hex so when I add them the value comes upto 32kb. So I suspect that you
>>>> have deleted some data from your volume. These values wont be helpful now.
>>>> Can you please re-run the same test and report the values when such a
>>>> problem occurs again.
>>>>
>>>
>>>
>>> ------------------------------
>>>
>>> Message: 2
>>> Date: Thu, 07 Jul 2011 19:56:46 +0200
>>> From: Carl Chenet <chaica at ohmytux.com>
>>> Subject: Re: Issues with geo-rep
>>> To: gluster-users at gluster.org
>>> Message-ID: <4E15F35E.1060404 at ohmytux.com>
>>> Content-Type: text/plain; charset=ISO-8859-1; format=flowed
>>>
>>> On 07/07/2011 15:25, Kaushik BV wrote:
>>>> Hi Chaica,
>>>>
>>>> This primarily means that the RPC communtication between the master
>>>> gsyncd module and slave gsyncd module is broken, this could happen to
>>>> various reasons. Check if it satisies all the pre-requisites:
>>>>
>>>> - If FUSE is installed in the machine, since Geo-replication module
>>>> mounts the GlusterFS volume using FUSE to sync data.
>>>> - If the Slave is a volume, check if the volume is started.
>>>> - If the Slave is a plain directory, check if the directory has been
>>>> created already with the desired permissions (Not applicable in your case)
>>>> - If Glusterfs 3.2 is not installed in the default location (in Master)
>>>> and has been prefixed to be installed in a custom location, configure
>>>> the *gluster-command*  for it to point to exact location.
>>>> - If Glusterfs 3.2 is not installed in the default location (in slave)
>>>> and has been prefixed to be installed in a custom location, configure
>>>> the *remote-gsyncd-command*  for it to point to exact place where gsyncd
>>>>   is located.
>>>> - locate the slave log and see if it has any anomalies.
>>>> - Passwordless SSH is set up properly between the host and the remote
>>>> machine ( Not applicable in your case)
>>>
>>> Ok the situation has slightly evolved. Now I do have a slave log and
>>> clearer error message on the master :
>>>
>>>
>>> [2011-07-07 19:53:16.258866] I [monitor(monitor):42:monitor] Monitor:
>>> ------------------------------------------------------------
>>> [2011-07-07 19:53:16.259073] I [monitor(monitor):43:monitor] Monitor:
>>> starting gsyncd worker
>>> [2011-07-07 19:53:16.332720] I [gsyncd:286:main_i] <top>: syncing:
>>> gluster://localhost:test-volume -> ssh://192.168.1.32::test-volume
>>> [2011-07-07 19:53:16.343554] D [repce:131:push] RepceClient: call
>>> 6302:140305661662976:1310061196.34 __repce_version__() ...
>>> [2011-07-07 19:53:20.931523] D [repce:141:__call__] RepceClient: call
>>> 6302:140305661662976:1310061196.34 __repce_version__ -> 1.0
>>> [2011-07-07 19:53:20.932172] D [repce:131:push] RepceClient: call
>>> 6302:140305661662976:1310061200.93 version() ...
>>> [2011-07-07 19:53:20.933662] D [repce:141:__call__] RepceClient: call
>>> 6302:140305661662976:1310061200.93 version -> 1.0
>>> [2011-07-07 19:53:20.933861] D [repce:131:push] RepceClient: call
>>> 6302:140305661662976:1310061200.93 pid() ...
>>> [2011-07-07 19:53:20.934525] D [repce:141:__call__] RepceClient: call
>>> 6302:140305661662976:1310061200.93 pid -> 10075
>>> [2011-07-07 19:53:20.957355] E [syncdutils:131:log_raise_exception]
>>> <top>: FAIL:
>>> Traceback (most recent call last):
>>>    File "/usr/lib/glusterfs/glusterfs/python/syncdaemon/gsyncd.py", line
>>> 102, in main
>>>      main_i()
>>>    File "/usr/lib/glusterfs/glusterfs/python/syncdaemon/gsyncd.py", line
>>> 293, in main_i
>>>      local.connect()
>>>    File "/usr/lib/glusterfs/glusterfs/python/syncdaemon/resource.py",
>>> line 379, in connect
>>>      raise RuntimeError("command failed: " + " ".join(argv))
>>> RuntimeError: command failed: /usr/sbin/glusterfs --xlator-option
>>> *-dht.assert-no-child-down=true -L DEBUG -l
>>> /var/log/glusterfs/geo-replication/test-volume/ssh%3A%2F%2Froot%40192.168.1.32%3Agluster%3A%2F%2F127.0.0.1%3Atest-volume.gluster.log
>>> -s localhost --volfile-id test-volume --client-pid=-1
>>> /tmp/gsyncd-aux-mount-hy6T_w
>>> [2011-07-07 19:53:20.960621] D [monitor(monitor):58:monitor] Monitor:
>>> worker seems to be connected (?? racy check)
>>> [2011-07-07 19:53:21.962501] D [monitor(monitor):62:monitor] Monitor:
>>> worker died in startup phase
>>>
>>> The command launched by glusterfs returns a 255 error shell code, which
>>> I belive means the command is terminated by a signal. On the slave log I
>>> have :
>>>
>>> [2011-07-07 19:54:49.571549] I [fuse-bridge.c:3218:fuse_thread_proc]
>>> 0-fuse: unmounting /tmp/gsyncd-aux-mount-z2Q2Hg
>>> [2011-07-07 19:54:49.572459] W [glusterfsd.c:712:cleanup_and_exit]
>>> (-->/lib/libc.so.6(clone+0x6d) [0x7f2c8998b02d]
>>> (-->/lib/libpthread.so.0(+0x68ba) [0x7f2c89c238ba]
>>> (-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xc5) [0x7f2c8a8f51b5])))
>>> 0-: received signum (15), shutting down
>>> [2011-07-07 19:54:51.280207] W [write-behind.c:3029:init]
>>> 0-test-volume-write-behind: disabling write-behind for first 0 bytes
>>> [2011-07-07 19:54:51.291669] I [client.c:1935:notify]
>>> 0-test-volume-client-0: parent translators are ready, attempting connect
>>> on transport
>>> [2011-07-07 19:54:51.292329] I [client.c:1935:notify]
>>> 0-test-volume-client-1: parent translators are ready, attempting connect
>>> on transport
>>> [2011-07-07 19:55:38.582926] I [rpc-clnt.c:1531:rpc_clnt_reconfig]
>>> 0-test-volume-client-0: changing port to 24009 (from 0)
>>> [2011-07-07 19:55:38.583456] I [rpc-clnt.c:1531:rpc_clnt_reconfig]
>>> 0-test-volume-client-1: changing port to 24009 (from 0)
>>>
>>> Bye,
>>> Carl Chenet
>>>
>>>
>>> ------------------------------
>>>
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>
>>>
>>> End of Gluster-users Digest, Vol 39, Issue 13
>>> *********************************************
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux