Re: 160 Thousand ceph-client.admin.*.asok files : Wired problem , never seen before

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Thank you for your reply

No its these nodes are just Ceph nodes , nothing shared with OpenSack. Infact i am not using openstack with this ceph cluster.

However i have configured Calamari on this cluster , but not sure if that has caused the problem. I also tried to remove calamari configurations and packages from entire cluster but still files are getting generated.

See the below trend, files are getting generated every 1 Minute 

srwxr-xr-x.  1 root root    0 Aug 10 09:56 ceph-client.admin.3104224.53185360.asok
srwxr-xr-x.  1 root root    0 Aug 10 09:57 ceph-client.admin.3104606.45857616.asok
srwxr-xr-x.  1 root root    0 Aug 10 09:58 ceph-client.admin.3104987.45837136.asok
srwxr-xr-x.  1 root root    0 Aug 10 09:59 ceph-client.admin.3105378.63122256.asok
srwxr-xr-x.  1 root root    0 Aug 10 10:00 ceph-client.admin.3105763.62139216.asok
srwxr-xr-x.  1 root root    0 Aug 10 10:01 ceph-client.admin.3106175.47811408.asok
srwxr-xr-x.  1 root root    0 Aug 10 10:02 ceph-client.admin.3106553.45525840.asok
srwxr-xr-x.  1 root root    0 Aug 10 10:03 ceph-client.admin.3106937.57162576.asok
srwxr-xr-x.  1 root root    0 Aug 10 10:04 ceph-client.admin.3107316.68578128.asok
srwxr-xr-x.  1 root root    0 Aug 10 10:05 ceph-client.admin.3107701.54725456.asok
srwxr-xr-x.  1 root root    0 Aug 10 10:06 ceph-client.admin.3108044.65157968.asok
srwxr-xr-x.  1 root root    0 Aug 10 10:07 ceph-client.admin.3108513.50350928.asok
srwxr-xr-x.  1 root root    0 Aug 10 10:08 ceph-client.admin.3108900.59702096.asok
srwxr-xr-x.  1 root root    0 Aug 10 10:09 ceph-client.admin.3109276.57756496.asok
drwxr-xr-x.  2 root root  13M Aug 10 10:09 .
[root@node1 ceph]# date
Mon Aug 10 10:09:57 EEST 2015
[root@node1 ceph]# 

Any help to fix this is appreciated. 

Regards
Vicky

On Mon, Aug 10, 2015 at 2:32 AM, Nigel Williams <nigel.williams@xxxxxxxxxxx> wrote:
On 10/08/2015 12:02 AM, Robert LeBlanc wrote:
> I'm guessing this is on an OpenStack node? There is a fix for this and I
> think it will come out in the next release. For now we have had to disable
> the admin sockets.

Do you know what triggers the fault? we've not seen it on Firefly+RBD for Openstack.


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux