Thanks a lot, xiubo!!! this time we still restarted mds fix this due to user urgent need list /path/to/A/, i will try to mds debug log if we hit it again. Also, haven’t try flush mds journal before, any side effect to do this? This cephfs cluster is a production environment, we need very careful to do anything. And I read this bug details https://tracker.ceph.com/issues/50840 <https://tracker.ceph.com/issues/50840> , from pre-mail mentions it fixed in 15.2.17, we using ceps-deploy to deploy(upgrade) cephfs, seems it latest ceph version is 15.2.16? Will update if we can fixed after upgrade. Will try to flush ads journal option when we hit this bug next time(if no user urgent need list directory). Seems it can 100% recurrent these days. Thanks All! Thanks, zx > 2022年8月31日 15:23,Xiubo Li <xiubli@xxxxxxxxxx> 写道: > > > On 8/31/22 2:43 PM, zxcs wrote: >> Hi, experts >> >> we have a cephfs(15.2.13) cluster with kernel mount, and when we read from 2000+ processes to one ceph path(called /path/to/A/), then all of the process hung, and ls -lrth /path/to/A/ always stuck, but list other directory are health( /path/to/B/), >> >> health detail always report mds has slow request. And then we need to restart the mds fix this issue. >> >> How can we fix this without restart mds(restart mds always impact other users)? >> >> Any suggestions are welcome! Thanks a ton! >> >> from this dump_ops_in_flight: >> >> "description": "client_request(client.100807215:2856632 getattr AsLsXsFs #0x200978a3326 2022-08-31T09:36:30.444927+0800 caller_id=2049, caller_gid=2049})", >> "initiated_at": "2022-08-31T09:36:30.454570+0800", >> "age": 17697.012491966001, >> "duration": 17697.012805568, >> "type_data": { >> "flag_point": "dispatched", >> "reqid": "client. 100807215:2856632", >> "op_type": "client_request", >> "client_info": >> "client": "client.100807215", >> "tid": 2856632 >> "events": >> "time": "2022-08-31T09:36:30.454570+0800", >> "event": "initiated" >> >> "time": "2022-08-31T09:36:30.454572+0800", >> "event": "throttled" >> >> "time": "2022-08-31T09:36:30.454570+0800", >> "event": "header read" >> >> "time": "2022-08-31T09:36:30.454580+0800", >> 'event": "all_read" >> "time": "2022-08-31T09:36:30.454604+0800", >> "event": "dispatched" >> } >> > AFAIK there is no easy way to do this. At least we need to know why it gets stuck and where. From your above and the previous mail thread, it should stuck in getattr request and sounds like a smiliar issue with https://tracker.ceph.com/issues/50840 <https://tracker.ceph.com/issues/50840>. > > If it's not, it should be a new bug and could you create one tracker and provide the mds side debug logs. > > Maybe you can try to flush the mds journal to see what will happen ? > > - Xiubo > > >> >> Thanks, >> Xiong >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx <mailto:ceph-users@xxxxxxx> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx <mailto:ceph-users-leave@xxxxxxx> _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx