Hi
I'm testing a Ceph cluster with "rados bench", it's an empty Cephadm
install that only has one pool device_health_metrics.
Create a pool with 1024 pg on the hdd devices(15 servers has HDDs and 13
has SSDs)
ceph osd pool create pool-ec32-isa-reed_sol_van-hdd 1024 1024 erasue
ec32-isa-reed_sol_van-hdd --autoscale-mode=off
I then run "rados bench" from the 13 SSD hosts at the same time.
rados bench -p pool-ec32-isa-reed_sol_van-hdd 600 write --no-cleanup
After just a few seconds "ceph -s" starts to reports degraded data
redundancy
Here is some examples during the 10 minutes testing period
Degraded data redundancy: 260/11856050 objects degraded (0.014%), 1
pg degraded
Degraded data redundancy: 260/1856050 objects degraded (0.014%), 1
pg degraded
Degraded data redundancy: 1 pg undersized
Degraded data redundancy: 1688/3316225 objects degraded (0.051%), 3
pgs degraded
Degraded data redundancy: 5457/7005845 objects degraded (0.078%), 3
pgs degraded, 9 pgs undersized
Degraded data redundancy: 1 pg undersized
Degraded data redundancy: 4161/7005845 objects degraded (0.059%), 3
pgs degraded
Degraded data redundancy: 4315/7005845 objects degraded (0.062%), 2
pgs degraded, 4 pgs undersized
So my question is, it normal that Ceph report degraded under normal use?
or do I have a problem somewhere that I need to investigate?
--
Kai Stian Olstad
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx