And are the OSDs getting marked down during the outage? Are all the MONs still up? Jan > On 09 Jul 2015, at 13:20, Mallikarjun Biradar <mallikarjuna.biradar@xxxxxxxxx> wrote: > > I have size=2 & min_size=1 and IO is paused till all hosts com back. > > On Thu, Jul 9, 2015 at 4:41 PM, Jan Schermer <jan@xxxxxxxxxxx> wrote: >> What is the min_size setting for the pool? If you have size=2 and min_size=2, then all your data is safe when one replica is down, but the IO is paused. If you want to continue IO you need to set min_size=1. >> But be aware that a single failure after that causes you to lose all the data, you’d have to revert to the other replica if it comes up and works - no idea how that works in ceph but will likely be a PITA to do. >> >> Jan >> >>> On 09 Jul 2015, at 12:42, Mallikarjun Biradar <mallikarjuna.biradar@xxxxxxxxx> wrote: >>> >>> Hi all, >>> >>> Setup details: >>> Two storage enclosures each connected to 4 OSD nodes (Shared storage). >>> Failure domain is Chassis (enclosure) level. Replication count is 2. >>> Each host has allotted with 4 drives. >>> >>> I have active client IO running on cluster. (Random write profile with >>> 4M block size & 64 Queue depth). >>> >>> One of enclosure had power loss. So all OSD's from hosts that are >>> connected to this enclosure went down as expected. >>> >>> But client IO got paused. After some time enclosure & hosts connected >>> to it came up. >>> And all OSD's on that hosts came up. >>> >>> Till this time, cluster was not serving IO. Once all hosts & OSD's >>> pertaining to that enclosure came up, client IO resumed. >>> >>> >>> Can anybody help me why cluster not serving IO during enclosure >>> failure. OR its a bug? >>> >>> -Thanks & regards, >>> Mallikarjun Biradar >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@xxxxxxxxxxxxxx >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com