Really slow cache-evict

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA256

We are testing out cache tiering, but when evicting the cache on an
idle cluster it is extremely slow (10 objects per minutes). Looking at
top some of the OSD processes are busy, but the disks are idle across
the cluster. I upped debug_osd to 20/20 and saw millions of messages
per object like:

2015-10-23 16:52:15.952561 7fa42a1fd700 20 osd.169 39454
should_share_map client.329656 10.208.16.31:0/1012854 39454
2015-10-23 16:52:15.952605 7fa42a1fd700 15 osd.169 39454 enqueue_op
0x287ba400 prio 63 cost 450560 latency 0.000999
osd_op(client.329656.0:165865996
rbd_data.14143b5faf8a00.000000000000041f [stat,set-alloc-hint
object_size 8388608 write_size 8388608,write 860160~450560]
12.549948fa ack+ondisk+wr
ite+known_if_redirected e39454) v5
2015-10-23 16:52:15.952642 7fa43fbeb700 10 osd.169 39454 dequeue_op
0x287ba400 prio 63 cost 450560 latency 0.001036
osd_op(client.329656.0:165865996
rbd_data.14143b5faf8a00.000000000000041f [stat,set-alloc-hint
object_size 8388608 write_size 8388608,write 860160~450560]
12.549948fa ack+ondisk+wr
ite+known_if_redirected e39454) v5 pg pg[12.fa( v 39422'12054
(39253'9006,39422'12054] local-les=39453 n=284 ec=31174 les/c
39453/39453 39452/39452/39442) [169,186,153,136] r=0 lpr=39452
crt=39422'12051 lcod 0'0 mlcod 0'0 active+clean]
2015-10-23 16:52:15.952676 7fa43fbeb700 20 osd.169 pg_epoch: 39454
pg[12.fa( v 39422'12054 (39253'9006,39422'12054] local-les=39453 n=284
ec=31174 les/c 39453/39453 39452/39452/39442) [169,186,153,136] r=0
lpr=39452 crt=39422'12051 lcod 0'0 mlcod 0'0 active+clean]
op_has_sufficient_caps pool=12
(ssd-pool ) owner=0 need_read_cap=1 need_write_cap=1
need_class_read_cap=0 need_class_write_cap=0 -> NO
2015-10-23 16:52:15.952688 7fa42a1fd700 20 osd.169 39454
should_share_map client.329656 10.208.16.31:0/1012854 39454
2015-10-23 16:52:15.952701 7fa43fbeb700 10 osd.169 39454 dequeue_op
0x287ba400 finish
2015-10-23 16:52:15.952729 7fa42a1fd700 15 osd.169 39454 enqueue_op
0x287c7500 prio 63 cost 4096 latency 0.000092
osd_op(client.329656.0:165866002
rbd_data.14143b5faf8a00.000000000000041f [stat,set-alloc-hint
object_size 8388608 write_size 8388608,write 6369280~4096] 12.549948fa
ack+ondisk+write
+known_if_redirected e39454) v5
2015-10-23 16:52:15.952762 7fa43d3e6700 10 osd.169 39454 dequeue_op
0x287c7500 prio 63 cost 4096 latency 0.000125
osd_op(client.329656.0:165866002
rbd_data.14143b5faf8a00.000000000000041f [stat,set-alloc-hint
object_size 8388608 write_size 8388608,write 6369280~4096] 12.549948fa
ack+ondisk+write
+known_if_redirected e39454) v5 pg pg[12.fa( v 39422'12054
(39253'9006,39422'12054] local-les=39453 n=284 ec=31174 les/c
39453/39453 39452/39452/39442) [169,186,153,136] r=0 lpr=39452
crt=39422'12051 lcod 0'0 mlcod 0'0 active+clean]
2015-10-23 16:52:15.952787 7fa43d3e6700 20 osd.169 pg_epoch: 39454
pg[12.fa( v 39422'12054 (39253'9006,39422'12054] local-les=39453 n=284
ec=31174 les/c 39453/39453 39452/39452/39442) [169,186,153,136] r=0
lpr=39452 crt=39422'12051 lcod 0'0 mlcod 0'0 active+clean]
op_has_sufficient_caps pool=12
(ssd-pool ) owner=0 need_read_cap=1 need_write_cap=1
need_class_read_cap=0 need_class_write_cap=0 -> NO
2015-10-23 16:52:15.952832 7fa43d3e6700 10 osd.169 39454 dequeue_op
0x287c7500 finish

It looks like maybe the OP is bouncing between threads or something
like that and never getting dispatched correctly. This is on
ceph version 0.94.4 (95292699291242794510b39ffde3f4df67898d3a)

I followed the directions at [1] in the writeback section for
disabling the cache tier. I did this on another cluster recently and
didn't see this problem (only took an hour or two to evict 700GB of
data). I saw a very slow evict a while back, but just chalked it up to
btrfs being dumb.

[1] http://docs.ceph.com/docs/master/rados/operations/cache-tiering/#removing-a-writeback-cache
- ----------------
Robert LeBlanc
PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
-----BEGIN PGP SIGNATURE-----
Version: Mailvelope v1.2.2
Comment: https://www.mailvelope.com

wsFcBAEBCAAQBQJWKr8iCRDmVDuy+mK58QAAl7oP/2bfWT/DUxjxkEGIXOnW
x2dk7hlTAkTAts7bRQn+vL31bcHAEoN+nJH5J39nt98Kh4rpZyHcVtJammYM
hYNsc97Hah+O2iqDYWxr4Vu27eyq5icCw0oFo9hbsA6lGPcrgeVSGbnXaFk0
+4u6OCovRDb6Rgfc6IAJoHrkulnPaWGaKWa1sfofdlMDJoA2U5cnX7ZvkRnM
unnKKtAjmeVCFbt/QmwFJVEjdNh0yZXN2qLSJUVwf4APgB4G2trupQ6kuisn
oGZdJP1DRtFn5IgoZQzB4ZA1FNydrVoQmrPNTAkCGQoUWegbmhxrXxsrfLV9
jTky2inv8NLdP46VyEVyvOqFD3WVRYIP+KvVcaNnL2awMClOgZ4uish+rLK/
Uyhcm9dfXgBz1Fm+PzdzpMwK8/lvk/UcfnBs6s5rELgjCQ/WPLMu7y1ZP9FY
DzLsrtMdbMGZA7MTt9CKAe9fZ+sPC3Se8slh6mN7M7Tm2Knsv6uyKLJ06imW
M7DMGVMi2A3I+qTE/DEA/D5MLcx+vJtLD57/a2j3dWaEiiBqaAsMA18UiwEX
bMaHAbFMWBrlhIVytP4IaA7dbiEaO6C90pyHHtyVf1etKDfP+erNRtHVW4JI
8wpG87Z34o6r92fkbb0LRH/TvD+WQFU4Q0OiL5JqclJKJP6TNUUhMmstad62
vmqk
=dO6U
-----END PGP SIGNATURE-----
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux