Hi, Thanks for the link. I unset the nodown config option and things did seem to improve, although we did still get a few reports from users about issues related filesystem (rbd) access, even after that action was taken. Thanks again, Shain > On Mar 13, 2017, at 2:43 AM, Alexandre DERUMIER <aderumier@xxxxxxxxx> wrote: > > Hi, > >>> Currently I have the. noout and nodown flags set while doing the maintenance work. > > you only need noout to avoid rebalancing > > see documentation: > http://docs.ceph.com/docs/kraken/rados/troubleshooting/troubleshooting-osd/ > "STOPPING W/OUT REBALANCING". > > > Your clients are hanging because of the no down flag > > > See this blog for no-out, no-down flags experiements > > https://www.sebastien-han.fr/blog/2013/04/17/some-ceph-experiments/ > > ----- Mail original ----- > De: "Shain Miley" <smiley@xxxxxxx> > À: "ceph-users" <ceph-users@xxxxxxxx> > Envoyé: Lundi 13 Mars 2017 04:58:08 > Objet: noout, nodown and blocked requests > > Hello, > One of the nodes in our 14 node cluster is offline and before I totally commit to fully removing the node from the cluster (there is a chance I can get the node back in working order in the next few days) I would like to run the cluster with that single node out for a few days. > > Currently I have the. noout and nodown flags set while doing the maintenance work. > > Some users are complaining about disconnects and other oddities when try to save and access files currently on the cluster. > > I am also seeing some blocked requests when viewing the cluster status (at this point I see 160 block requests spread over 15 to 20 osd’s). > > Currently I have a replication level of 3 on this pool and a min_size of 1. > > My question is this…is there a better method to use (other than using noout and nodown) in this scenario where I do not want data movement yet…but I do want the reads and writes to the cluster to to respond as normally as possible for the end users? > > Thanks in advance, > > Shain > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com