problems about deep-scrub on Luminous

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi guys,

In my production environment, when deep-srub runned on some empty pgs,
the osd will become down. And I found that the ops on rocksdb took too
long time, which will lead the thread become unhealthy. And no ping
request sended to the peer osds. Then the monitor got reports from the
peer osd and mark this osd down.

The following was the log from the primary osd.

 2019-02-27 11:00:00.604029 7f9592f84700 20 osd.92 pg_epoch: 13415
pg[5.1bf( empty local-lis/les=13413/13414 n=0 ec=175/175 lis/c
13413/13413 les/c/f 13414/13414/0 13413/13413/13413) [92,142,111] r=0
lpr=13413 crt=0'0 mlcod 0'0 active+clean+scrubbing+deep] scrub state
INACTIVE [MIN,MIN)
2019-02-27 11:00:00.604168 7f9592f84700 20 osd.92 pg_epoch: 13415
pg[5.1bf( empty local-lis/les=13413/13414 n=0 ec=175/175 lis/c
13413/13413 les/c/f 13414/13414/0 13413/13413/13413) [92,142,111] r=0
lpr=13413 crt=0'0 mlcod 0'0 active+clean+scrubbing+deep] scrub state
NEW_CHUNK [5:fd800000::::head,MIN)
2019-02-27 11:00:00.604189 7f9592f84700 15
bluestore(/var/lib/ceph/osd/ceph-92) collection_list 5.1bf_head start
#5:fd800000::::head#0 end GHMAX max 5
2019-02-27 11:00:00.604195 7f9592f84700 20
bluestore(/var/lib/ceph/osd/ceph-92) _collection_list range
0x7f7ffffffffffffff9fd800000 to 0x7f7ffffffffffffff9fe000000 and
0x7f8000000000000005fd800000 to 0x7f8000000000000005fe000000 start
#5:fd800000::::head#0
2019-02-27 11:00:00.611606 7f9592f84700 20
bluestore(/var/lib/ceph/osd/ceph-92) _collection_list pend
0x7f8000000000000005fe000000
2019-02-27 11:00:00.611617 7f9592f84700 30
bluestore(/var/lib/ceph/osd/ceph-92) _collection_list key
0x7f8000000000000005fd80000021213dfffffffffffffffeffffffffffffffff'o'
2019-02-27 11:00:00.611621 7f9592f84700 20
bluestore(/var/lib/ceph/osd/ceph-92) _collection_list oid
#5:fd800000::::head# end GHMAX
2019-02-27 11:00:26.142032 7f9592f84700 20
bluestore(/var/lib/ceph/osd/ceph-92) _collection_list key
0x7f800000000000001d04c0000021213dfffffffffffffffeffffffffffffffff'o'
>= GHMAX
2019-02-27 11:00:26.142077 7f9592f84700 10
bluestore(/var/lib/ceph/osd/ceph-92) collection_list 5.1bf_head start
GHMAX end GHMAX max 5 = 0, ls.size() = 1, next = GHMAX
2019-02-27 11:00:26.142245 7f9592f84700 20 osd.92 pg_epoch: 13415
pg[5.1bf( empty local-lis/les=13413/13414 n=0 ec=175/175 lis/c
13413/13413 les/c/f 13414/13414/0 13413/13413/13413) [92,142,111] r=0
lpr=13413 crt=0'0 mlcod 0'0 active+clean+scrubbing+deep] scrub state
WAIT_PUSHES [5:fd800000::::head,MAX)
2019-02-27 11:00:26.142258 7f9592f84700 20 osd.92 pg_epoch: 13415
pg[5.1bf( empty local-lis/les=13413/13414 n=0 ec=175/175 lis/c
13413/13413 les/c/f 13414/13414/0 13413/13413/13413) [92,142,111] r=0
lpr=13413 crt=0'0 mlcod 0'0 active+clean+scrubbing+deep] scrub state
WAIT_LAST_UPDATE [5:fd800000::::head,MAX)
2019-02-27 11:00:26.142267 7f9592f84700 20 osd.92 pg_epoch: 13415
pg[5.1bf( empty local-lis/les=13413/13414 n=0 ec=175/175 lis/c
13413/13413 les/c/f 13414/13414/0 13413/13413/13413) [92,142,111] r=0
lpr=13413 crt=0'0 mlcod 0'0 active+clean+scrubbing+deep] scrub state
BUILD_MAP [5:fd800000::::head,MAX)
2019-02-27 11:00:26.142286 7f9592f84700 15
bluestore(/var/lib/ceph/osd/ceph-92) collection_list 5.1bf_head start
#5:fd800000::::head# end #MAX# max 2147483647
2019-02-27 11:00:26.142292 7f9592f84700 20
bluestore(/var/lib/ceph/osd/ceph-92) _collection_list range
0x7f7ffffffffffffff9fd800000 to 0x7f7ffffffffffffff9fe000000 and
0x7f8000000000000005fd800000 to 0x7f8000000000000005fe000000 start
#5:fd800000::::head#
2019-02-27 11:00:26.142805 7f9592f84700 20
bluestore(/var/lib/ceph/osd/ceph-92) _collection_list pend
0x7f8000000000000005fe000000
2019-02-27 11:00:26.142815 7f9592f84700 30
bluestore(/var/lib/ceph/osd/ceph-92) _collection_list key
0x7f8000000000000005fd80000021213dfffffffffffffffeffffffffffffffff'o'
2019-02-27 11:00:26.142820 7f9592f84700 20
bluestore(/var/lib/ceph/osd/ceph-92) _collection_list oid
#5:fd800000::::head# end #MAX#
2019-02-27 11:00:51.796411 7f9592f84700 20
bluestore(/var/lib/ceph/osd/ceph-92) _collection_list key
0x7f800000000000001d04c0000021213dfffffffffffffffeffffffffffffffff'o'
>= #MAX#
2019-02-27 11:00:51.796447 7f9592f84700 10
bluestore(/var/lib/ceph/osd/ceph-92) collection_list 5.1bf_head start
#5:fd800000::::head# end #MAX# max 2147483647 = 0, ls.size() = 1, next
= GHMIN
2019-02-27 11:00:51.796564 7f9592f84700 20 osd.92 pg_epoch: 13415
pg[5.1bf( empty local-lis/les=13413/13414 n=0 ec=175/175 lis/c
13413/13413 les/c/f 13414/13414/0 13413/13413/13413) [92,142,111] r=0
lpr=13413 crt=0'0 mlcod 0'0 active+clean+scrubbing+deep] scrub state
WAIT_REPLICAS [5:fd800000::::head,MAX)

I export the bluefs using ceph-bluestore-tool, and its size is about
5GB which locate on sata-ssd.
I know little about rocksdb and cannot find the real reason. So what
kind of problem about rocksdb ?
Any suggestion, thanks!

best regards.



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux