I have a cluster with 1 inconsistent pg. I have attempted the following steps with no luck. What should my next move be?
1. executed ceph health detail to determine what pg was inconsistent
[ceph-admin@admin libr-cluster]$ ceph health detail
HEALTH_ERR 1 pgs inconsistent; 1 scrub errors
pg 1.959 is active+clean+inconsistent, acting [69,252,127]
1 scrub errors
2. executed ceph pg repair 1.959
3. after nothing happening for quite a while I decided to dig into it a bit. What strikes me most odd is that all of these files seem to be consistent using size, md5, and sha256. It's also a little concerning that they are all 0 for size.
[root@osd5 ceph-admin]# rados list-inconsistent-pg cephfs_data
["1.959"]
[root@osd5 ceph-admin]# rados list-inconsistent-pg cephfs_metadata
[]
[root@osd5 ceph-admin]# rados list-inconsistent-pg rbd
[]
[root@osd5 ceph-admin]# rados list-inconsistent-pg vmware_ecpool
[]
[root@osd5 ceph-admin]# rados list-inconsistent-pg vmware_cache
[]
[root@osd5 ceph-admin]# rados list-inconsistent-obj 1.959 --format=json-pretty
{
"epoch": 178113,
"inconsistents": []
}
[root@osd5 ceph-admin]# grep -Hn 'ERR' /var/log/ceph/ceph-osd.69.log
[root@osd5 ceph]# zgrep -Hn 'ERR' ./ceph-osd.69.log-*
./ceph-osd.69.log-20170512.gz:717:2017-05-11 09:23:11.734142 7ff46cbe4700 -1 log_channel(cluster) log [ERR] : scrub 1.959 1:9a97a372:::10004313b01.00000004:head on disk size (0) does not match object info size (1417216) adjusted for ondisk to (1417216)
./ceph-osd.69.log-20170512.gz:785:2017-05-11 09:26:02.877409 7ff46a3df700 -1 log_channel(cluster) log [ERR] : 1.959 scrub 1 errors
[root@osd0 ceph]# grep -Hn 'ERR' ./ceph-osd.127.log
[root@osd0 ceph]# zgrep -Hn 'ERR' ./ceph-osd.127.log-*
[root@osd11 ceph-admin]# grep -Hn 'ERR' /var/log/ceph/ceph-osd.252.log
[root@osd11 ceph-admin]# zgrep -Hn 'ERR' /var/log/ceph/ceph-osd.252.log-*
[root@osd5 ceph]# find /var/lib/ceph/osd/ceph-69/current/1.959_head/ -name '10004313b01.00000004*' -ls
2737776487 0 -rw-r--r-- 1 ceph ceph 0 May 10 07:01 /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[root@osd5 ceph-admin]# md5sum /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
d41d8cd98f00b204e9800998ecf8427e /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[root@osd5 ceph-admin]# sha256sum /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[root@osd5 ceph-admin]# ls -l /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
-rw-r--r--. 1 ceph ceph 0 May 10 07:01 /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[root@osd0 ceph-admin]# find /var/lib/ceph/osd/ceph-127/current/1.959_head/ -name '10004313b01.00000004*' -ls
2684661064 0 -rw-r--r-- 1 ceph ceph 0 May 10 07:01 /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[root@osd0 ceph-admin]# md5sum /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
d41d8cd98f00b204e9800998ecf8427e /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[root@osd0 ceph-admin]# sha256sum /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[root@osd0 ceph-admin]# ls -l /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
-rw-r--r--. 1 ceph ceph 0 May 10 07:01 /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[root@osd11 ceph-admin]# find /var/lib/ceph/osd/ceph-252/current/1.959_head/ -name '10004313b01.00000004*' -ls
1650915243 0 -rw-r--r-- 1 ceph ceph 0 May 10 07:01 /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[root@osd11 ceph-admin]# md5sum /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
d41d8cd98f00b204e9800998ecf8427e /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[root@osd11 ceph-admin]# sha256sum /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[root@osd11 ceph-admin]# ls -l /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
-rw-r--r--. 1 ceph ceph 0 May 10 07:01 /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
[ceph-admin@admin libr-cluster]$ ceph status
cluster 6f91f60c-7bc0-4aaa-a136-4a90851fbe10
health HEALTH_ERR
1 pgs inconsistent
1 scrub errors
monmap e17: 5 mons at {mon0=10.124.103.60:6789/0,mon1=10.124.103.61:6789/0,mon2=10.124.103.62:6789/0,osd2=10.124.103.72:6789/0,osd3=10.124.103.73:6789/0}
election epoch 454, quorum 0,1,2,3,4 mon0,mon1,mon2,osd2,osd3
fsmap e7005: 1/1/1 up {0=mds0=up:active}, 1 up:standby
osdmap e178115: 235 osds: 235 up, 235 in
flags sortbitwise,require_jewel_osds
pgmap v21842842: 5892 pgs, 5 pools, 305 TB data, 119 Mobjects
917 TB used, 364 TB / 1282 TB avail
5863 active+clean
16 active+clean+scrubbing+deep
12 active+clean+scrubbing
1 active+clean+inconsistent
client io 4076 kB/s rd, 633 kB/s wr, 15 op/s rd, 58 op/s wr
Thanks for any advice!
-Brady Deetz
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com