Re: High memory usage kills OSD while peering

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi again,
now every thing almost sorted out. we had a few inconsistent shards that were killing the OSDs when recovering, we fixed some of them by removing the bad shards, and some by starting other OSDs with good shards. what is stopping us now, is that one OSD had a corrupted leveldb and refuses to start. not sure how that hapened, but i asume is due to the many times the node/osd died from lack of memory. I am also not sure if we should continue the discussion here, or start a new thread.

the osd (262) is showing those logs upon start:

2017-08-26 17:07:17.915861 7fbd8e4cbd00  0 set uid:gid to 0:0 (:)
2017-08-26 17:07:17.915875 7fbd8e4cbd00 0 ceph version 12.1.4 (a5f84b37668fc8e03165aaf5cbb380c78e4deba4) luminous (rc), process (unknown), pid 26713 2017-08-26 17:07:17.927085 7fbd8e4cbd00 0 pidfile_write: ignore empty --pid-file 2017-08-26 17:07:17.951358 7fbd8e4cbd00 0 load: jerasure load: lrc load: isa 2017-08-26 17:07:17.951602 7fbd8e4cbd00 0 filestore(/var/lib/ceph/osd/ceph-262) backend xfs (magic 0x58465342) 2017-08-26 17:07:17.952164 7fbd8e4cbd00 0 filestore(/var/lib/ceph/osd/ceph-262) backend xfs (magic 0x58465342) 2017-08-26 17:07:17.952977 7fbd8e4cbd00 0 genericfilestorebackend(/var/lib/ceph/osd/ceph-262) detect_features: FIEMAP ioctl is disabled via 'filestore fiemap' config option 2017-08-26 17:07:17.952983 7fbd8e4cbd00 0 genericfilestorebackend(/var/lib/ceph/osd/ceph-262) detect_features: SEEK_DATA/SEEK_HOLE is disabled via 'filestore seek data hole' config option 2017-08-26 17:07:17.952985 7fbd8e4cbd00 0 genericfilestorebackend(/var/lib/ceph/osd/ceph-262) detect_features: splice() is disabled via 'filestore splice' config option 2017-08-26 17:07:17.953309 7fbd8e4cbd00 0 genericfilestorebackend(/var/lib/ceph/osd/ceph-262) detect_features: syncfs(2) syscall fully supported (by glibc and kernel) 2017-08-26 17:07:17.953797 7fbd8e4cbd00 0 xfsfilestorebackend(/var/lib/ceph/osd/ceph-262) detect_feature: extsize is disabled by conf 2017-08-26 17:07:17.954628 7fbd8e4cbd00 0 filestore(/var/lib/ceph/osd/ceph-262) start omap initiation 2017-08-26 17:07:17.957166 7fbd8e4cbd00 -1 filestore(/var/lib/ceph/osd/ceph-262) mount(1724): Error initializing leveldb : Corruption: error in middle of record

2017-08-26 17:07:17.957179 7fbd8e4cbd00 -1 osd.262 0 OSD:init: unable to mount object store 2017-08-26 17:07:17.957183 7fbd8e4cbd00 -1 ** ERROR: osd init failed: (1) Operation not permitted

ceph-objectstore-tool shows similar errors.

so, we figured it is only one OSD and we can go without it. we marked it lost, pgs started to peer and got active. but 5 remain in the incomplete state. and te pg query shows:

...
    "recovery_state": [
        {
            "name": "Started/Primary/Peering/Incomplete",
            "enter_time": "2017-08-26 22:59:03.044623",
            "comment": "not enough complete instances of this PG"
        },
        {
            "name": "Started/Primary/Peering",
            "enter_time": "2017-08-26 22:59:02.540748",
            "past_intervals": [
                {
                    "first": "959669",
                    "last": "1090812",
                    "all_participants": [
                        {
                            "osd": 258
                        },
                        {
                            "osd": 262
                        },
                        {
                            "osd": 338
                        },
                        {
                            "osd": 545
                        },
                        {
                            "osd": 549
                        }
                    ],
                    "intervals": [
                        {
                            "first": "964880",
                            "last": "964924",
                            "acting": "262"
                        },
                        {
                            "first": "978855",
                            "last": "978956",
                            "acting": "545"
                        },
                        {
                            "first": "989628",
                            "last": "989808",
                            "acting": "258"
                        },
                        {
                            "first": "992614",
                            "last": "992975",
                            "acting": "549"
                        },
                        {
                            "first": "1085148",
                            "last": "1090812",
                            "acting": "338"
                        }
                    ]
                }
            ],
            "probing_osds": [
                "258",
                "338",
                "545",
                "549"
            ],
            "down_osds_we_would_probe": [
                262
            ],
            "peering_blocked_by": [],
            "peering_blocked_by_detail": [
                {
                    "detail": "peering_blocked_by_history_les_bound"
                }
            ]
        },
...

not sure wat that detail "peering_blocked_by_history_les_bound" is, and not sure how to proceed. i googled it, came up with nothing useful. all the incomplete pgs have the same detail as the above and similar recovery state.

ceph pg ls | grep incomplete
18.54b 0 0 0 0 0 0 2739 2739 incomplete 2017-08-26 23:15:46.705071 46889'4277 1091150:314001 [332,253] 332 [332,253] 332 46889'4277 2017-08-04 03:15:58.381025 46889'4277 2017-07-29 06:47:30.337673

19.54a 5950 0 0 0 0 26108435266 3019 3019 incomplete 2017-08-26 23:15:46.705156 961411'873129 1091150:58116482 [332,253] 332 [332,253] 332 960118'872495 2017-08-04 03:12:33.647414 952850'868978 2017-07-02 15:53:08.565948 19.608 0 0 0 0 0 0 0 0 incomplete 2017-08-26 22:59:03.044649 0'0 1091150:428 [258,338] 258 [258,338] 258 960118'862299 2017-08-04 03:01:57.011411 958900'861456 2017-07-28 02:33:29.476119 19.8bb 0 0 0 0 0 0 0 0 incomplete 2017-08-26 22:59:02.946453 0'0 1091150:339 [260,331] 260 [260,331] 260 960114'866811 2017-08-03 04:51:42.117840 952850'864443 2017-07-08 02:48:37.958357 19.dd3 5864 0 0 0 0 25600089555 3094 3094 incomplete 2017-08-26 17:20:07.948285 961411'865657 1091150:72381143 [263,142] 263 [263,142] 263 960118'865078 2017-08-25 17:32:06.181006 960118'865078 2017-08-25 17:32:06.181006


I also noticed that some of those have 0 objects in them despite the dir in one of the osds have objects in it.
these pools are replica 2


thanks
ali
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux