That's expected behaviour. If RADOS can't make your writes safe by
replicating them (because no other OSD is available) then clients will
pause their writes. See the min_size setting on a pool.
John
On 16/06/2015 00:11, Matteo Dacrema wrote:
With 3.16.3 kernel it seems to be stable but I've discovered one new
issue.
If I take down one of the two osd node all the client stop to respond.
Here the output of ceph -s
ceph -s
cluster 2de7b17f-0a3e-4109-b878-c035dd2f7735
health HEALTH_WARN
256 pgs degraded
127 pgs stuck inactive
127 pgs stuck unclean
256 pgs undersized
recovery 1457662/2915324 objects degraded (50.000%)
4/8 in osds are down
clock skew detected on mon.cephmds01, mon.ceph-mon1
monmap e5: 3 mons at
{ceph-mon1=10.29.81.184:6789/0,cephmds01=10.29.81.161:6789/0,cephmds02=10.29.81.160:6789/0}
election epoch 64, quorum 0,1,2 cephmds02,cephmds01,ceph-mon1
mdsmap e176: 1/1/1 up {0=cephmds01=up:active}, 1 up:standby
osdmap e712: 8 osds: 4 up, 8 in
pgmap v420651: 256 pgs, 2 pools, 133 GB data, 1423 kobjects
289 GB used, 341 GB / 631 GB avail
1457662/2915324 objects degraded (50.000%)
256 undersized+degraded+peered
client io 86991 B/s wr, 0 op/s
When I take UP the node all clients resume to work.
Thanks,
Matteo
------------------------------------------------------------------------
*Da:* ceph-users <ceph-users-bounces@xxxxxxxxxxxxxx> per conto di
Matteo Dacrema <mdacrema@xxxxxxxx>
*Inviato:* lunedì 15 giugno 2015 12:37
*A:* John Spray; Lincoln Bryant; ceph-users
*Oggetto:* Re: CephFS client issue
Ok, I'll update kernel to 3.16.3 version and let you know.
Thanks,
Matteo
------------------------------------------------------------------------
*Da:* John Spray <john.spray@xxxxxxxxxx>
*Inviato:* lunedì 15 giugno 2015 10:51
*A:* Matteo Dacrema; Lincoln Bryant; ceph-users
*Oggetto:* Re: CephFS client issue
On 14/06/15 20:00, Matteo Dacrema wrote:
Hi Lincoln,
I'm using the kernel client.
Kernel version is: 3.13.0-53-generic
That's old by CephFS standards. It's likely that the issue you're
seeing is one of the known bugs (which were actually the motivation
for adding the warning message you're seeing).
John
--
Questo messaggio e' stato analizzato con Libra ESVA ed e' risultato
non infetto.
Clicca qui per segnalarlo come spam.
<http://esva01.enter.it/cgi-bin/learn-msg.cgi?id=637D840263.A210F>
Clicca qui per metterlo in blacklist
<http://esva01.enter.it/cgi-bin/learn-msg.cgi?blacklist=1&id=637D840263.A210F>
--
Questo messaggio e' stato analizzato con Libra ESVA ed e' risultato
non infetto.
Clicca qui per segnalarlo come spam.
<http://esva01.enter.it/cgi-bin/learn-msg.cgi?id=DBA3140262.AE1FB>
Clicca qui per metterlo in blacklist
<http://esva01.enter.it/cgi-bin/learn-msg.cgi?blacklist=1&id=DBA3140262.AE1FB>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com