pg marked inconsistent while appearing to be consistent

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I have a cluster with 1 inconsistent pg. I have attempted the following steps with no luck. What should my next move be?

1. executed ceph health detail to determine what pg was inconsistent
[ceph-admin@admin libr-cluster]$ ceph health detail
HEALTH_ERR 1 pgs inconsistent; 1 scrub errors
pg 1.959 is active+clean+inconsistent, acting [69,252,127]
1 scrub errors

2. executed ceph pg repair 1.959

3. after nothing happening for quite a while I decided to dig into it a bit. What strikes me most odd is that all of these files seem to be consistent using size, md5, and sha256. It's also a little concerning that they are all 0 for size.
[root@osd5 ceph-admin]# rados list-inconsistent-pg cephfs_data
["1.959"]
[root@osd5 ceph-admin]# rados list-inconsistent-pg cephfs_metadata
[]
[root@osd5 ceph-admin]# rados list-inconsistent-pg rbd
[]
[root@osd5 ceph-admin]# rados list-inconsistent-pg vmware_ecpool
[]
[root@osd5 ceph-admin]# rados list-inconsistent-pg vmware_cache
[]

[root@osd5 ceph-admin]# rados list-inconsistent-obj 1.959 --format=json-pretty
{
    "epoch": 178113,
    "inconsistents": []
}


[root@osd5 ceph-admin]# grep -Hn 'ERR' /var/log/ceph/ceph-osd.69.log

[root@osd5 ceph]# zgrep -Hn 'ERR' ./ceph-osd.69.log-*
./ceph-osd.69.log-20170512.gz:717:2017-05-11 09:23:11.734142 7ff46cbe4700 -1 log_channel(cluster) log [ERR] : scrub 1.959 1:9a97a372:::10004313b01.00000004:head on disk size (0) does not match object info size (1417216) adjusted for ondisk to (1417216)
./ceph-osd.69.log-20170512.gz:785:2017-05-11 09:26:02.877409 7ff46a3df700 -1 log_channel(cluster) log [ERR] : 1.959 scrub 1 errors


[root@osd0 ceph]# grep -Hn 'ERR' ./ceph-osd.127.log

[root@osd0 ceph]# zgrep -Hn 'ERR' ./ceph-osd.127.log-*


[root@osd11 ceph-admin]# grep -Hn 'ERR' /var/log/ceph/ceph-osd.252.log

[root@osd11 ceph-admin]# zgrep -Hn 'ERR' /var/log/ceph/ceph-osd.252.log-*




[root@osd5 ceph]# find /var/lib/ceph/osd/ceph-69/current/1.959_head/ -name '10004313b01.00000004*' -ls
2737776487    0 -rw-r--r--   1 ceph     ceph            0 May 10 07:01 /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1

[root@osd5 ceph-admin]# md5sum /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
d41d8cd98f00b204e9800998ecf8427e  /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1

[root@osd5 ceph-admin]# sha256sum /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855  /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1

[root@osd5 ceph-admin]# ls -l /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
-rw-r--r--. 1 ceph ceph 0 May 10 07:01 /var/lib/ceph/osd/ceph-69/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1


[root@osd0 ceph-admin]# find /var/lib/ceph/osd/ceph-127/current/1.959_head/ -name '10004313b01.00000004*' -ls
2684661064    0 -rw-r--r--   1 ceph     ceph            0 May 10 07:01 /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1

[root@osd0 ceph-admin]# md5sum /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
d41d8cd98f00b204e9800998ecf8427e  /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1

[root@osd0 ceph-admin]# sha256sum /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855  /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1

[root@osd0 ceph-admin]# ls -l /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
-rw-r--r--. 1 ceph ceph 0 May 10 07:01 /var/lib/ceph/osd/ceph-127/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1


[root@osd11 ceph-admin]# find /var/lib/ceph/osd/ceph-252/current/1.959_head/ -name '10004313b01.00000004*' -ls
1650915243    0 -rw-r--r--   1 ceph     ceph            0 May 10 07:01 /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1

[root@osd11 ceph-admin]# md5sum /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
d41d8cd98f00b204e9800998ecf8427e  /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1

[root@osd11 ceph-admin]# sha256sum /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855  /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1

[root@osd11 ceph-admin]# ls -l /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1
-rw-r--r--. 1 ceph ceph 0 May 10 07:01 /var/lib/ceph/osd/ceph-252/current/1.959_head/DIR_9/DIR_5/DIR_9/DIR_E/DIR_5/10004313b01.00000004__head_4EC5E959__1





[ceph-admin@admin libr-cluster]$ ceph status
    cluster 6f91f60c-7bc0-4aaa-a136-4a90851fbe10
     health HEALTH_ERR
            1 pgs inconsistent
            1 scrub errors
     monmap e17: 5 mons at {mon0=10.124.103.60:6789/0,mon1=10.124.103.61:6789/0,mon2=10.124.103.62:6789/0,osd2=10.124.103.72:6789/0,osd3=10.124.103.73:6789/0}
            election epoch 454, quorum 0,1,2,3,4 mon0,mon1,mon2,osd2,osd3
      fsmap e7005: 1/1/1 up {0=mds0=up:active}, 1 up:standby
     osdmap e178115: 235 osds: 235 up, 235 in
            flags sortbitwise,require_jewel_osds
      pgmap v21842842: 5892 pgs, 5 pools, 305 TB data, 119 Mobjects
            917 TB used, 364 TB / 1282 TB avail
                5863 active+clean
                  16 active+clean+scrubbing+deep
                  12 active+clean+scrubbing
                   1 active+clean+inconsistent
  client io 4076 kB/s rd, 633 kB/s wr, 15 op/s rd, 58 op/s wr

Thanks for any advice!

-Brady Deetz

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux