Re: ceph-fuse auto down

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yes, when some ceph-fuse crash , the mount driver has gone, and can't remount . Reboot the server is the only way I can do.
But other client with ceph-fuse mount on them working well. Can writing / reading data on them.

ceph-fuse --version
ceph version 0.94.3 (95cefea9fd9ab740263bf8bb4796fd864d9afe2b)

ceph -s
cluster 0fddc8e0-9e64-4049-902a-2f0f6d531630
     health HEALTH_OK
     monmap e1: 3 mons at {ceph01=10.3.1.11:6789/0,ceph02=10.3.1.12:6789/0,ceph03=10.3.1.13:6789/0}
            election epoch 8, quorum 0,1,2 ceph01,ceph02,ceph03
     mdsmap e29: 1/1/1 up {0=ceph04=up:active}, 1 up:standby
     osdmap e26: 4 osds: 4 up, 4 in
      pgmap v94931: 320 pgs, 3 pools, 90235 MB data, 241 kobjects
            289 GB used, 1709 GB / 1999 GB avail
                 320 active+clean
  client io 1023 kB/s rd, 1210 kB/s wr, 72 op/s

2015-09-13 10:23 GMT+08:00 Shinobu Kinjo <skinjo@xxxxxxxxxx>:
Can you give us package version of ceph-fuse?

> Multi ceph-fuse crash just now today.

Did you just mount filesystem or was there any
activity on filesystem?

  e.g: writing / reading data

Can you give us output of on cluster side:

  ceph -s

Shinobu

----- Original Message -----
From: "谷枫" <feicheche@xxxxxxxxx>
To: "Shinobu Kinjo" <skinjo@xxxxxxxxxx>
Cc: "ceph-users" <ceph-users@xxxxxxxxxxxxxx>
Sent: Sunday, September 13, 2015 10:51:35 AM
Subject: Re: ceph-fuse auto down

sorry Shinobu,
I don't understand what's the means what you pasted.
Multi ceph-fuse crash just now today.
The ceph-fuse completely unusable for me now.
Maybe i must change the kernal mount with it.

2015-09-12 20:08 GMT+08:00 Shinobu Kinjo <skinjo@xxxxxxxxxx>:

> In _usr_bin_ceph-fuse.0.crash.client2.tar
>
> What I'm seeing now is:
>
>   3 Date: Sat Sep 12 06:37:47 2015
>  ...
>   6 ExecutableTimestamp: 1440614242
>  ...
>   7 ProcCmdline: ceph-fuse -k /etc/ceph.new/ceph.client.admin.keyring -m
> 10.3.1.11,10.3.1.12,10.3.1.13 /grdata
>  ...
>  30  7f32de7fe000-7f32deffe000 rw-p 00000000 00:00 0
>     [stack:17270]
>  ...
> 250  7f341021d000-7f3410295000 r-xp 00000000 fd:01 267219
>    /usr/lib/x86_64-linux-gnu/nss/libfreebl3.so
>  ...
> 255  7f341049b000-7f341054f000 r-xp 00000000 fd:01 266443
>    /usr/lib/x86_64-linux-gnu/libsqlite3.so.0.8.6
>  ...
> 260  7f3410754000-7f3410794000 r-xp 00000000 fd:01 267222
>    /usr/lib/x86_64-linux-gnu/nss/libsoftokn3.so
>  ...
> 266  7f3411197000-7f341119a000 r-xp 00000000 fd:01 264953
>    /usr/lib/x86_64-linux-gnu/libplds4.so
>  ...
> 271  7f341139f000-7f341159e000 ---p 00004000 fd:01 264955
>    /usr/lib/x86_64-linux-gnu/libplc4.so
>  ...
> 274  7f34115a0000-7f34115c5000 r-xp 00000000 fd:01 267214
>    /usr/lib/x86_64-linux-gnu/libnssutil3.so
>  ...
> 278  7f34117cb000-7f34117ce000 r-xp 00000000 fd:01 1189512
>     /lib/x86_64-linux-gnu/libdl-2.19.so
>  ...
> 287  7f3411d94000-7f3411daa000 r-xp 00000000 fd:01 1179825
>     /lib/x86_64-linux-gnu/libgcc_s.so.1
>  ...
> 294  7f34122b0000-7f3412396000 r-xp 00000000 fd:01 266069
>    /usr/lib/x86_64-linux-gnu/libstdc++.so.6.0.19
>  ...
> 458  State: D (disk sleep)
>  ...
> 359  VmPeak:     5250648 kB
> 360  VmSize:     4955592 kB
>  ...
>
> What were you trying to do?
>
> Shinobu
>
>

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux