Hi,
I run a cluster with 7 OSD. The cluster has no much traffic on it. But every few days, I get a HEALTH_ERR, because of inconsistent PGs: root@Sam ~ # ceph status cluster: id: c4bfc288-8ba8-4c3a-b3a6-ed95503f50b7 health: HEALTH_ERR 3 scrub errors Possible data damage: 3 pgs inconsistent services: mon: 1 daemons, quorum mon1 mgr: ceph-osd1(active) mds: FS-1/1/1 up {0=ceph-osd1=up:active} osd: 11 osds: 8 up, 7 in rgw: 1 daemon active data: pools: 6 pools, 168 pgs objects: 901.8 k objects, 2.6 TiB usage: 7.9 TiB used, 7.4 TiB / 15 TiB avail pgs: 165 active+clean 3 active+clean+inconsistent io: client: 641 KiB/s wr, 0 op/s rd, 3 op/s wr root@Sam ~ # ceph health detail HEALTH_ERR 3 scrub errors; Possible data damage: 3 pgs inconsistent OSD_SCRUB_ERRORS 3 scrub errors PG_DAMAGED Possible data damage: 3 pgs inconsistent pg 5.1d is active+clean+inconsistent, acting [6,8,3] pg 5.20 is active+clean+inconsistent, acting [3,9,0] pg 5.4a is active+clean+inconsistent, acting [6,3,7] What's the reason for this problem? How can I analyse it? Cheers, Dimitri |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com