Hi, Dear ceph-devel
I met a issue: Ceph osd rm one osd cause 30% objects degraded.
Step 1:
#created a ssd root
ceph osd crush add-bucket ssd root
Step 2: Installed a osd.100 failed:
94 1 osd.94 up 1
95 1 osd.95 up 1
96 1 osd.96 up 1
97 1 osd.97 up 1
98 1 osd.98 up 1
99 1 osd.99 up 1
100 0 osd.100 down 0
Step 3: Installed a osd.101 again successfully. Move a host into root=ssd.
-12 1 root ssd
-13 1 host ssd-cephnode1
101 1 osd.101 up 1
-1 100 root default
-2 10 host cephnode1
0 1 osd.0 up 1
1 1 osd.1 up 1
2 1 osd.2 up 1
Step 4: Then I ceph osd rm 100, but the the ceph health turned into 30%
objects degraded. Then the io performance downgrade to very slow (1MB/s
each clients).
Anybody know what is the root cause? Or some suggestions to finger it out?
Thank you very much.
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html