On 07/28/2016 07:56 PM, Niels de Vos wrote: > On Thu, Jul 28, 2016 at 05:58:15PM +0530, Mohammed Rafi K C wrote: >> >> On 07/27/2016 04:33 PM, Raghavendra G wrote: >>> >>> On Wed, Jul 27, 2016 at 10:29 AM, Mohammed Rafi K C >>> <rkavunga@xxxxxxxxxx <mailto:rkavunga@xxxxxxxxxx>> wrote: >>> >>> Thanks for your feedback. >>> >>> In fact meta xlator is loaded only on fuse mount, is there any >>> particular reason to not to use meta-autoload xltor for nfs server >>> and libgfapi ? >>> >>> >>> I think its because of lack of resources. I am not aware of any >>> technical reason for not using on NFSv3 server and gfapi. >> Cool. I will try to see how we can implement meta-autoliad feature for >> nfs-server and libgfapi. Once we have the feature in place, I will >> implement the cache memory display/flush feature using meta xlators. > In case you plan to have this ready in a month (before the end of > August), you should propose it as a 3.9 feature. Click the "Edir this > page on GitHub" link on the bottom of > https://www.gluster.org/community/roadmap/3.9/ :) I will do an assessment and will see if I can spend some time on this for 3.9 release. If so, I will add it into 3.9 feature page. Regards Rafi KC > > Thanks, > Niels > > >> Thanks for your valuable feedback. >> Rafi KC >> >>> >>> >>> Regards >>> >>> Rafi KC >>> >>> On 07/26/2016 04:05 PM, Niels de Vos wrote: >>>> On Tue, Jul 26, 2016 at 12:43:56PM +0530, Kaushal M wrote: >>>>> On Tue, Jul 26, 2016 at 12:28 PM, Prashanth Pai <ppai@xxxxxxxxxx> <mailto:ppai@xxxxxxxxxx> wrote: >>>>>> +1 to option (2) which similar to echoing into /proc/sys/vm/drop_caches >>>>>> >>>>>> -Prashanth Pai >>>>>> >>>>>> ----- Original Message ----- >>>>>>> From: "Mohammed Rafi K C" <rkavunga@xxxxxxxxxx> <mailto:rkavunga@xxxxxxxxxx> >>>>>>> To: "gluster-users" <gluster-users@xxxxxxxxxxx> <mailto:gluster-users@xxxxxxxxxxx>, "Gluster Devel" <gluster-devel@xxxxxxxxxxx> <mailto:gluster-devel@xxxxxxxxxxx> >>>>>>> Sent: Tuesday, 26 July, 2016 10:44:15 AM >>>>>>> Subject: Need a way to display and flush gluster cache ? >>>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> Gluster stack has it's own caching mechanism , mostly on client side. >>>>>>> But there is no concrete method to see how much memory are consuming by >>>>>>> gluster for caching and if needed there is no way to flush the cache memory. >>>>>>> >>>>>>> So my first question is, Do we require to implement this two features >>>>>>> for gluster cache? >>>>>>> >>>>>>> >>>>>>> If so I would like to discuss some of our thoughts towards it. >>>>>>> >>>>>>> (If you are not interested in implementation discussion, you can skip >>>>>>> this part :) >>>>>>> >>>>>>> 1) Implement a virtual xattr on root, and on doing setxattr, flush all >>>>>>> the cache, and for getxattr we can print the aggregated cache size. >>>>>>> >>>>>>> 2) Currently in gluster native client support .meta virtual directory to >>>>>>> get meta data information as analogues to proc. we can implement a >>>>>>> virtual file inside the .meta directory to read the cache size. Also we >>>>>>> can flush the cache using a special write into the file, (similar to >>>>>>> echoing into proc file) . This approach may be difficult to implement in >>>>>>> other clients. >>>>> +1 for making use of the meta-xlator. We should be making more use of it. >>>> Indeed, this would be nice. Maybe this can also expose the memory >>>> allocations like /proc/slabinfo. >>>> >>>> The io-stats xlator can dump some statistics to >>>> /var/log/glusterfs/samples/ and /var/lib/glusterd/stats/ . That seems to >>>> be acceptible too, and allows to get statistics from server-side >>>> processes without involving any clients. >>>> >>>> HTH, >>>> Niels >>>> >>>> >>>>>>> 3) A cli command to display and flush the data with ip and port as an >>>>>>> argument. GlusterD need to send the op to client from the connected >>>>>>> client list. But this approach would be difficult to implement for >>>>>>> libgfapi based clients. For me, it doesn't seems to be a good option. >>>>>>> >>>>>>> Your suggestions and comments are most welcome. >>>>>>> >>>>>>> Thanks to Talur and Poornima for their suggestions. >>>>>>> >>>>>>> Regards >>>>>>> >>>>>>> Rafi KC >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Gluster-devel mailing list >>>>>>> Gluster-devel@xxxxxxxxxxx <mailto:Gluster-devel@xxxxxxxxxxx> >>>>>>> http://www.gluster.org/mailman/listinfo/gluster-devel >>>>>>> >>>>>> _______________________________________________ >>>>>> Gluster-devel mailing list >>>>>> Gluster-devel@xxxxxxxxxxx <mailto:Gluster-devel@xxxxxxxxxxx> >>>>>> http://www.gluster.org/mailman/listinfo/gluster-devel >>>>> _______________________________________________ >>>>> Gluster-users mailing list >>>>> Gluster-users@xxxxxxxxxxx <mailto:Gluster-users@xxxxxxxxxxx> >>>>> http://www.gluster.org/mailman/listinfo/gluster-users >>>>> >>>>> >>>>> _______________________________________________ >>>>> Gluster-devel mailing list >>>>> Gluster-devel@xxxxxxxxxxx <mailto:Gluster-devel@xxxxxxxxxxx> >>>>> http://www.gluster.org/mailman/listinfo/gluster-devel >>> >>> _______________________________________________ >>> Gluster-devel mailing list >>> Gluster-devel@xxxxxxxxxxx <mailto:Gluster-devel@xxxxxxxxxxx> >>> http://www.gluster.org/mailman/listinfo/gluster-devel >>> >>> >>> >>> >>> -- >>> Raghavendra G _______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-devel