Paul Emmerich
Looking for help with your Ceph cluster? Contact us at https://croit.io
croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
I have tried ceph pg repair several times. It claims "instructing pg 2.798s0 on osd.41 to repair" but then nothing happens as far as I can tell. Any way of knowing if it's doing more?_______________________________________________On Sat, May 18, 2019 at 3:33 PM Brett Chancellor <bchancellor@xxxxxxxxxxxxxx> wrote:I would try the ceph pg repair. If you see the pg go into deep scrubbing, then back to inconsistent you probably have a bad drive. Find which of the drives in the pg are bad (pg query or go to the host and look through dmesg). Take that osd offline and mark it out. Once backfill is complete, it should clear up.On Sat, May 18, 2019, 6:05 PM Jorge Garcia <jgarcia@xxxxxxxxxxxx> wrote:We are testing a ceph cluster mostly using cephfs. We are using an erasure-code pool, and have been loading it up with data. Recently, we got a HEALTH_ERR response when we were querying the ceph status. We stopped all activity to the filesystem, and waited to see if the error would go away. It didn't. Then we tried a couple of suggestions from the internet (ceph pg repair, ceph pg scrub, ceph pg deep-scrub) to no avail. I'm not sure how to find out more information about what the problem is, and how to repair the filesystem to bring it back to normal health. Any suggestions?_______________________________________________Current status:# ceph -s
cluster:
id: 28ef32f1-4350-491b-9003-b19b9c3a2076
health: HEALTH_ERR
5 scrub errors
Possible data damage: 1 pg inconsistent
services:
mon: 3 daemons, quorum gi-cba-01,gi-cba-02,gi-cba-03
mgr: gi-cba-01(active), standbys: gi-cba-02, gi-cba-03
mds: backups-1/1/1 up {0=gi-cbmd=up:active}
osd: 87 osds: 87 up, 87 in
data:
pools: 2 pools, 4096 pgs
objects: 90.98 M objects, 134 TiB
usage: 210 TiB used, 845 TiB / 1.0 PiB avail
pgs: 4088 active+clean
5 active+clean+scrubbing+deep
2 active+clean+scrubbing
1 active+clean+inconsistent
# ceph health detail
HEALTH_ERR 5 scrub errors; Possible data damage: 1 pg inconsistent
OSD_SCRUB_ERRORS 5 scrub errors
PG_DAMAGED Possible data damage: 1 pg inconsistent
pg 2.798 is active+clean+inconsistent, acting [41,50,17,2,86,70,61]
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com