Hi, After recent switch do default ``--stripe-count 1'' on image upload I have observed some strange thing - single import or deletion of the striped image may temporarily turn off entire cluster, literally(see log below). Of course next issued osd map fix the situation, but all in-flight operations experiencing a short freeze. This issue appears randomly in some import or delete operation, have not seen any other types causing this. Even if a nature of this bug laying completely in the client-osd interaction, may be ceph should develop a some foolproof actions even if complaining client have admin privileges? Almost for sure this should be reproduced within teuthology with rwx rights both on osds and mons at the client. And as I can see there is no problem on both physical and protocol layer for dedicated cluster interface on client machine. 2012-12-17 02:17:03.691079 mon.0 [INF] pgmap v2403268: 15552 pgs: 15552 active+clean; 931 GB data, 2927 GB used, 26720 GB / 29647 GB avail 2012-12-17 02:17:04.693344 mon.0 [INF] pgmap v2403269: 15552 pgs: 15552 active+clean; 931 GB data, 2927 GB used, 26720 GB / 29647 GB avail 2012-12-17 02:17:05.695742 mon.0 [INF] pgmap v2403270: 15552 pgs: 15552 active+clean; 931 GB data, 2927 GB used, 26720 GB / 29647 GB avail 2012-12-17 02:17:05.991900 mon.0 [INF] osd.0 10.5.0.10:6800/4907 failed (3 reports from 1 peers after 2012-12-17 02:17:29.991859 >= grace 20.000000) 2012-12-17 02:17:05.992017 mon.0 [INF] osd.1 10.5.0.11:6800/5011 failed (3 reports from 1 peers after 2012-12-17 02:17:29.991995 >= grace 20.000000) 2012-12-17 02:17:05.992139 mon.0 [INF] osd.2 10.5.0.12:6803/5226 failed (3 reports from 1 peers after 2012-12-17 02:17:29.992110 >= grace 20.000000) 2012-12-17 02:17:05.992240 mon.0 [INF] osd.3 10.5.0.13:6803/6054 failed (3 reports from 1 peers after 2012-12-17 02:17:29.992224 >= grace 20.000000) 2012-12-17 02:17:05.992330 mon.0 [INF] osd.4 10.5.0.14:6803/5792 failed (3 reports from 1 peers after 2012-12-17 02:17:29.992317 >= grace 20.000000) 2012-12-17 02:17:05.992420 mon.0 [INF] osd.5 10.5.0.15:6803/5564 failed (3 reports from 1 peers after 2012-12-17 02:17:29.992405 >= grace 20.000000) 2012-12-17 02:17:05.992515 mon.0 [INF] osd.7 10.5.0.17:6803/5902 failed (3 reports from 1 peers after 2012-12-17 02:17:29.992501 >= grace 20.000000) 2012-12-17 02:17:05.992607 mon.0 [INF] osd.8 10.5.0.10:6803/5338 failed (3 reports from 1 peers after 2012-12-17 02:17:29.992591 >= grace 20.000000) 2012-12-17 02:17:05.992702 mon.0 [INF] osd.10 10.5.0.12:6800/5040 failed (3 reports from 1 peers after 2012-12-17 02:17:29.992686 >= grace 20.000000) 2012-12-17 02:17:05.992793 mon.0 [INF] osd.11 10.5.0.13:6800/5748 failed (3 reports from 1 peers after 2012-12-17 02:17:29.992778 >= grace 20.000000) 2012-12-17 02:17:05.992891 mon.0 [INF] osd.12 10.5.0.14:6800/5459 failed (3 reports from 1 peers after 2012-12-17 02:17:29.992875 >= grace 20.000000) 2012-12-17 02:17:05.992980 mon.0 [INF] osd.13 10.5.0.15:6800/5235 failed (3 reports from 1 peers after 2012-12-17 02:17:29.992966 >= grace 20.000000) 2012-12-17 02:17:05.993081 mon.0 [INF] osd.16 10.5.0.30:6800/5585 failed (3 reports from 1 peers after 2012-12-17 02:17:29.993065 >= grace 20.000000) 2012-12-17 02:17:05.993184 mon.0 [INF] osd.17 10.5.0.31:6800/5578 failed (3 reports from 1 peers after 2012-12-17 02:17:29.993169 >= grace 20.000000) 2012-12-17 02:17:05.993274 mon.0 [INF] osd.18 10.5.0.32:6800/5097 failed (3 reports from 1 peers after 2012-12-17 02:17:29.993260 >= grace 20.000000) 2012-12-17 02:17:05.993367 mon.0 [INF] osd.19 10.5.0.33:6800/5109 failed (3 reports from 1 peers after 2012-12-17 02:17:29.993352 >= grace 20.000000) 2012-12-17 02:17:05.993464 mon.0 [INF] osd.20 10.5.0.34:6800/5125 failed (3 reports from 1 peers after 2012-12-17 02:17:29.993448 >= grace 20.000000) 2012-12-17 02:17:05.993554 mon.0 [INF] osd.21 10.5.0.35:6800/5183 failed (3 reports from 1 peers after 2012-12-17 02:17:29.993538 >= grace 20.000000) 2012-12-17 02:17:05.993644 mon.0 [INF] osd.22 10.5.0.36:6800/5202 failed (3 reports from 1 peers after 2012-12-17 02:17:29.993628 >= grace 20.000000) 2012-12-17 02:17:05.993740 mon.0 [INF] osd.23 10.5.0.37:6800/5252 failed (3 reports from 1 peers after 2012-12-17 02:17:29.993725 >= grace 20.000000) 2012-12-17 02:17:05.993831 mon.0 [INF] osd.24 10.5.0.30:6803/5758 failed (3 reports from 1 peers after 2012-12-17 02:17:29.993816 >= grace 20.000000) 2012-12-17 02:17:05.993924 mon.0 [INF] osd.25 10.5.0.31:6803/5748 failed (3 reports from 1 peers after 2012-12-17 02:17:29.993908 >= grace 20.000000) 2012-12-17 02:17:05.994018 mon.0 [INF] osd.26 10.5.0.32:6803/5275 failed (3 reports from 1 peers after 2012-12-17 02:17:29.994002 >= grace 20.000000) 2012-12-17 02:17:06.105315 mon.0 [INF] osdmap e24204: 32 osds: 4 up, 32 in 2012-12-17 02:17:06.051291 osd.6 [WRN] 1 slow requests, 1 included below; oldest blocked for > 30.947080 secs 2012-12-17 02:17:06.051299 osd.6 [WRN] slow request 30.947080 seconds old, received at 2012-12-17 02:16:35.042711: osd_op(client.2804602.0:20660 rbd_data.2a8fb612200854.00000000000000ec [write 1572864~278528] 6.b45f4c88) v4 currently waiting for sub ops -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html