Yes , We need to similar to inotify/fanotity .
came through link http://docs.ceph.com/docs/master/dev/osd_internals/watch_notify/?highlight=notify#watch-notifySiva
On Thu, Jun 9, 2016 at 6:06 PM, Anand Bhat <anand.bhat@xxxxxxxxx> wrote:
I think you are looking for inotify/fanotify events for Ceph. Usually these are implemented for local file system. Ceph being a networked file system, it will not be easy to implement and will involve network traffic to generate events.Not sure it is in the plan though.Regards,Anand--On Wed, Jun 8, 2016 at 2:46 PM, John Spray <jspray@xxxxxxxxxx> wrote:On Wed, Jun 8, 2016 at 8:40 AM, siva kumar <85siva@xxxxxxxxx> wrote:
> Dear Team,
>
> We are using ceph storage & cephFS for mounting .
>
> Our configuration :
>
> 3 osd
> 3 monitor
> 4 clients .
> ceph version 0.94.5 (9764da52395923e0b32908d83a9f7304401fee43)
>
> We would like to get file change notifications like what is the event
> (ADDED, MODIFIED,DELETED) and for which file the event has occurred. These
> notifications should be sent to our server.
> How to get these notifications?
This isn't a feature that CephFS has right now. Still, I would be
interested to know what protocol/format your server would consume
these kinds of notifications in?
John
> Ultimately we would like to add our custom file watch notification hooks to
> ceph so that we can handle this notifications by our self .
>
> Additional Info :
>
> [test@ceph-zclient1 ~]$ ceph -s
>
>> cluster a8c92ae6-6842-4fa2-bfc9-8cdefd28df5c
>
> health HEALTH_WARN
> mds0: ceph-client1 failing to respond to cache pressure
> mds0: ceph-client2 failing to respond to cache pressure
> mds0: ceph-client3 failing to respond to cache pressure
> mds0: ceph-client4 failing to respond to cache pressure
> monmap e1: 3 mons at
> {ceph-zadmin=xxx.xxx.xxx.xxx:6789/0,ceph-zmonitor=xxx.xxx.xxx.xxx:6789/0,ceph-zmonitor1=xxx.xxx.xxx.xxx:6789/0}
> election epoch 16, quorum 0,1,2
> ceph-zadmin,ceph-zmonitor1,ceph-zmonitor
> mdsmap e52184: 1/1/1 up {0=ceph-zstorage1=up:active}
> osdmap e3278: 3 osds: 3 up, 3 in
> pgmap v5068139: 384 pgs, 3 pools, 518 GB data, 7386 kobjects
> 1149 GB used, 5353 GB / 6503 GB avail
> 384 active+clean
>
> client io 1259 B/s rd, 179 kB/s wr, 11 op/s
>
>
>
> Thanks,
> S.Sivakumar
>
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
----------------------------------------------------------------------------
Never say never.
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com