Re: Help - Multiple OSD's Down

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The first OSD took 156Gb of Ram to boot.. :(

Is there a easy way to stop Mempool pulling so much memory.

On Wed, 5 Jan 2022 at 22:12, Mazzystr <mazzystr@xxxxxxxxx> wrote:

> and that is exactly why I run osds containerized with limited cpu and
> memory as well as "bluestore cache size", "osd memory target", and "mds
> cache memory limit".  Osd processes have become noisy neighbors in the last
> few versions.
>
>
>
> On Wed, Jan 5, 2022 at 1:47 PM Lee <lquince@xxxxxxxxx> wrote:
>
>> I'm not rushing,
>>
>> I have found the issue, Im am getting OOM errors as the OSD boots,
>> basically is starts to process the PG's and then the node runs out of
>> memory and the daemon kill's
>>
>> 2022-01-05 20:09:08 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:01.024+0000 7fce3c6bc700 10 osd.51 24448261 tick
>> 2022-01-05 20:09:10 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:01.060+0000 7fce3b441700 10 osd.51 24448261
>> tick_without_osd_lock
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:02.268+0000 7fce3c6bc700 10 osd.51 24448261 do_waiters --
>> start
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:09.544+0000 7fce3c6bc700 10 osd.51 24448261 do_waiters --
>> finish
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:10.260+0000 7fce1e407700  5 osd.51 24448261 heartbeat
>> osd_stat(store_statfs(0x2258948000/0x40000000/0x3a38800000, data
>> 0x17919fd8c4/0x179feb4000, compress 0x0/0x0/0x0, omap 0xc9773, meta
>> 0x3ff3688d), peers [] op hist [])
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.060+0000 7fce3c6bc700 20 osd.51 24448261 tick
>> last_purged_snaps_scrub 2022-01-04T22:29:39.121925+0000 next
>> 2022-01-05T22:29:39.121925+0000
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.104+0000 7fce1e407700 20 osd.51 24448261
>> check_full_status cur ratio 0.410072, physical ratio 0.410072, new state
>> none
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.108+0000 7fce34c34700 20
>> bluestore(/var/lib/ceph/osd/ceph-51) deferred_try_submit 0 osrs, 0 txcs
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.108+0000 7fce34c34700  5
>> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size:
>> 134217728 kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864
>> meta_used: 75234 data_alloc: 67108864 data_used: 0
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.160+0000 7fce34c34700 20
>> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size:
>> 134217728 kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864
>> meta_used: 75234 data_alloc: 67108864 data_used: 0
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.216+0000 7fce34c34700 20
>> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size:
>> 134217728 kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864
>> meta_used: 75234 data_alloc: 67108864 data_used: 0
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.264+0000 7fce34c34700 20
>> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size:
>> 134217728 kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864
>> meta_used: 75234 data_alloc: 67108864 data_used: 0
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.400+0000 7fce34c34700 20
>> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size:
>> 134217728 kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864
>> meta_used: 75234 data_alloc: 67108864 data_used: 0
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.536+0000 7fce34c34700 20
>> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size:
>> 134217728 kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864
>> meta_used: 75234 data_alloc: 67108864 data_used: 0
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.640+0000 7fce34c34700 20
>> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size:
>> 134217728 kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864
>> meta_used: 75234 data_alloc: 67108864 data_used: 0
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.644+0000 7fce1e407700  5 osd.51 24448261 heartbeat
>> osd_stat(store_statfs(0x2258948000/0x40000000/0x3a38800000, data
>> 0x17919fd8c4/0x179feb4000, compress 0x0/0x0/0x0, omap 0xc9773, meta
>> 0x3ff3688d), peers [] op hist [])
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.712+0000 7fce34c34700 20
>> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size:
>> 134217728 kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864
>> meta_used: 75234 data_alloc: 67108864 data_used: 0
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.688+0000 7fce1e407700 20 osd.51 24448261
>> check_full_status cur ratio 0.410072, physical ratio 0.410072, new state
>> none
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.480+0000 7fce3b441700 20
>> bluestore(/var/lib/ceph/osd/ceph-51) statfs
>> store_statfs(0x2258948000/0x40000000/0x3a38800000, data
>> 0x17919fd8c4/0x179feb4000, compress 0x0/0x0/0x0, omap 0xc9773, meta
>> 0x3ff3688d)
>> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:13.844+0000 7fce34c34700 20
>> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size:
>> 134217728 kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864
>> meta_used: 75234 data_alloc: 67108864 data_used: 0
>> 2022-01-05 20:09:14 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:14.016+0000 7fce34c34700 20
>> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size:
>> 134217728 kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864
>> meta_used: 75234 data_alloc: 67108864 data_used: 0
>> 2022-01-05 20:09:38 bb-ceph-enc-rm63-osd03-31 osd.51
>> 2022-01-05T20:09:14.104+0000 7fce3c6bc700 10 osd.51 24448261 tick
>> 2022-01-05 20:10:37 bb-ceph-enc-rm63-osd03-31 init.scope
>> ceph-osd@51.service:
>> Main process exited, code=killed, status=9/KILL
>> 2022-01-05 20:10:37 bb-ceph-enc-rm63-osd03-31 init.scope
>> ceph-osd@51.service:
>> Failed with result 'signal'.
>> 2022-01-05 20:10:47 bb-ceph-enc-rm63-osd03-31 init.scope
>> ceph-osd@51.service:
>> Scheduled restart job, restart counter is at 1.
>> 2022-01-05 20:10:47 bb-ceph-enc-rm63-osd03-31 init.scope Stopped Ceph
>> object storage daemon osd.51.
>>
>> I have just increased the RAM physically in one of the node's removed the
>> other OSD's physically for now and managed to get one of the 3 down to
>> come
>> up. Just stepping through each at the moment.
>>
>> Regards
>>
>> Lee
>>
>> On Wed, 5 Jan 2022 at 21:10, mhnx <morphinwithyou@xxxxxxxxx> wrote:
>>
>> > First of all, do not rush into bad decisions.
>> > Production is down and you wanna make it online but you should fix the
>> > problem and be sure first. If a second crash occurs in a healing state
>> > you will lose metadata.
>> > You don't need to debug first!
>> >
>> > You didn't mention your cluster status and we don't know what you have.
>> > We need some information;
>> > 1- ceph -s
>> > 2- ceph health detail
>> > 3- ceph df
>> > 4- tail /var/log/ceph/ceph-osd{crashed osd number}.log -n 1000
>> >
>> >
>> >
>> > Lee <lquince@xxxxxxxxx>, 5 Oca 2022 Çar, 23:14 tarihinde şunu yazdı:
>> > >
>> > > Looking for some help as this is production effecting..
>> > >
>> > > We run a 3 Node cluster with a mix of 5xSSD,15xSATA and 5xSAS in each
>> > node.
>> > > Running 15.2.15. All using DB/WAL on NVME SSD except the SSD's
>> > >
>> > > Earlier today I increased the PG num from 32 to 128 on one of our
>> pools,
>> > > due to the status complaining. pretty normally really. 2-3 mins in I
>> > > watched in horror as SSD based OSD's crashed on all 3 nodes, refusing
>> to
>> > > restart.
>> > >
>> > > I've set debug_bluefs and bluestore to 20 it will get so far and then
>> the
>> > > daemon fails.
>> > >
>> > > 2022-01-05 19:39:23 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:23.335+0000 7f2794383700 20
>> > > bluestore(/var/lib/ceph/osd/ceph-51) deferred_try_submit 0 osrs, 0
>> txcs
>> > > 2022-01-05 19:39:23 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:23.335+0000 7f2794383700  5
>> > > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > > 2022-01-05 19:39:23 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:23.387+0000 7f2794383700 20
>> > > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > > 2022-01-05 19:39:23 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:23.467+0000 7f2794383700 20
>> > > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > > 2022-01-05 19:39:24 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:23.979+0000 7f2794383700 20
>> > > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > > 2022-01-05 19:39:24 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:24.167+0000 7f2794383700 20
>> > > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > > 2022-01-05 19:39:24 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:24.271+0000 7f2794383700 20
>> > > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > > 2022-01-05 19:39:24 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:24.327+0000 7f2794383700 20
>> > > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > > 2022-01-05 19:39:32 bb-ceph-enc-rm63-osd03-31 init.scope
>> > ceph-osd@51.service:
>> > > Main process exited, code=killed, status=9/KILL
>> > > 2022-01-05 19:39:32 bb-ceph-enc-rm63-osd03-31 init.scope
>> > ceph-osd@51.service:
>> > > Failed with result 'signal'.
>> > > 2022-01-05 19:39:42 bb-ceph-enc-rm63-osd03-31 init.scope
>> > ceph-osd@51.service:
>> > > Scheduled restart job, restart counter is at 1.
>> > >
>> > > I've run
>> > > ceph-bluestore-tool bluefs-bdev-sizes --path /var/lib/ceph/osd/ceph-51
>> > > inferring bluefs devices from bluestore path
>> > > 1 : device size 0x3a38800000 : own 0x[1bf2200000~254300000] =
>> > 0x254300000 :
>> > > using 0x3fd10000(1021 MiB) : bluestore has 0x1d83400000(118 GiB)
>> > available
>> > >
>> > > Also fsck and repair all seems to be ok.
>> > >
>> > > The normal log looks like
>> > >
>> > > 2022-01-05 19:39:42 bb-ceph-enc-rm63-osd03-31 init.scope Starting Ceph
>> > > object storage daemon osd.51...
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.467+0000 7fca32943e00  0 set uid:gid to
>> 64045:64045
>> > > (ceph:ceph)
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.467+0000 7fca32943e00  0 ceph version 15.2.15
>> > > (2dfb18841cfecc2f7eb7eb2afd65986ca4d95985) octopus (stable), process
>> > > ceph-osd, pid 139577
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.467+0000 7fca32943e00  0 pidfile_write: ignore
>> empty
>> > > --pid-file
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev create path
>> > > /var/lib/ceph/osd/ceph-51/block type kernel
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev(0x55b4b234e000
>> > > /var/lib/ceph/osd/ceph-51/block) open path
>> > /var/lib/ceph/osd/ceph-51/block
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev(0x55b4b234e000
>> > > /var/lib/ceph/osd/ceph-51/block) open size 250056015872 (0x3a38800000,
>> > 233
>> > > GiB) block_size 4096 (4 KiB) non-rotational discard not supported
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.491+0000 7fca32943e00  1
>> > > bluestore(/var/lib/ceph/osd/ceph-51) _set_cache_sizes cache_size
>> > 1073741824
>> > > meta 0.4 kv 0.4 data 0.2
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev create path
>> > > /var/lib/ceph/osd/ceph-51/block type kernel
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > > /var/lib/ceph/osd/ceph-51/block) open path
>> > /var/lib/ceph/osd/ceph-51/block
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > > /var/lib/ceph/osd/ceph-51/block) open size 250056015872 (0x3a38800000,
>> > 233
>> > > GiB) block_size 4096 (4 KiB) non-rotational discard not supported
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bluefs add_block_device
>> > bdev 1
>> > > path /var/lib/ceph/osd/ceph-51/block size 233 GiB
>> > > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > > /var/lib/ceph/osd/ceph-51/block) close
>> > > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:47.067+0000 7fca32943e00  0 starting osd.51 osd_data
>> > > /var/lib/ceph/osd/ceph-51 /var/lib/ceph/osd/ceph-51/journal
>> > > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:47.159+0000 7fca32943e00  0 load: jerasure load: lrc
>> > load:
>> > > isa
>> > > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:47.159+0000 7fca32943e00  1 bdev create path
>> > > /var/lib/ceph/osd/ceph-51/block type kernel
>> > > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:47.159+0000 7fca32943e00  1 bdev(0x55b4b234e000
>> > > /var/lib/ceph/osd/ceph-51/block) open path
>> > /var/lib/ceph/osd/ceph-51/block
>> > > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:47.163+0000 7fca32943e00  1 bdev(0x55b4b234e000
>> > > /var/lib/ceph/osd/ceph-51/block) open size 250056015872 (0x3a38800000,
>> > 233
>> > > GiB) block_size 4096 (4 KiB) non-rotational discard not supported
>> > > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:47.163+0000 7fca32943e00  1
>> > > bluestore(/var/lib/ceph/osd/ceph-51) _set_cache_sizes cache_size
>> > 1073741824
>> > > meta 0.4 kv 0.4 data 0.2
>> > > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:47.163+0000 7fca32943e00  1 bdev(0x55b4b234e000
>> > > /var/lib/ceph/osd/ceph-51/block) close
>> > > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:48.619+0000 7fca32943e00  1
>> > > bluestore(/var/lib/ceph/osd/ceph-51) _open_alloc loaded 138 GiB in
>> 276582
>> > > extents available 129 GiB
>> > > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:48.619+0000 7fca32943e00  1 bluefs umount
>> > > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:48.619+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > > /var/lib/ceph/osd/ceph-51/block) close
>> > > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:48.803+0000 7fca32943e00  1 bdev create path
>> > > /var/lib/ceph/osd/ceph-51/block type kernel
>> > > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:48.803+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > > /var/lib/ceph/osd/ceph-51/block) open path
>> > /var/lib/ceph/osd/ceph-51/block
>> > > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:48.803+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > > /var/lib/ceph/osd/ceph-51/block) open size 250056015872 (0x3a38800000,
>> > 233
>> > > GiB) block_size 4096 (4 KiB) non-rotational discard not supported
>> > > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:48.803+0000 7fca32943e00  1 bluefs add_block_device
>> > bdev 1
>> > > path /var/lib/ceph/osd/ceph-51/block size 233 GiB
>> > > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:48.803+0000 7fca32943e00  1 bluefs mount
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.087+0000 7fca32943e00  1
>> > > bluestore(/var/lib/ceph/osd/ceph-51) _open_db opened rocksdb path db
>> > > options
>> > >
>> >
>> compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.087+0000 7fca32943e00  1
>> > > bluestore(/var/lib/ceph/osd/ceph-51) _upgrade_super from 4, latest 4
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.087+0000 7fca32943e00  1
>> > > bluestore(/var/lib/ceph/osd/ceph-51) _upgrade_super done
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.131+0000 7fca32943e00  0
>> > >  /build/ceph-15.2.15/src/cls/cephfs/cls_cephfs.cc:198: loading cephfs
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.131+0000 7fca32943e00  0
>> > >  /build/ceph-15.2.15/src/cls/hello/cls_hello.cc:312: loading cls_hello
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.135+0000 7fca32943e00  0 _get_class not permitted
>> to
>> > > load kvs
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.171+0000 7fca32943e00  0 _get_class not permitted
>> to
>> > > load lua
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.207+0000 7fca32943e00  0 _get_class not permitted
>> to
>> > > load queue
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.319+0000 7fca32943e00  0 _get_class not permitted
>> to
>> > > load sdk
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.319+0000 7fca32943e00  0 osd.51 24448261 crush map
>> > has
>> > > features 288514051259236352, adjusting msgr requires for clients
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.319+0000 7fca32943e00  0 osd.51 24448261 crush map
>> > has
>> > > features 288514051259236352 was 8705, adjusting msgr requires for mons
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.319+0000 7fca32943e00  0 osd.51 24448261 crush map
>> > has
>> > > features 3314933000852226048, adjusting msgr requires for osds
>> > > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:39:49.319+0000 7fca32943e00  1 osd.51 24448261
>> > > check_osdmap_features require_osd_release unknown -> octopus
>> > > 2022-01-05 19:41:25 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:41:24.999+0000 7fca32943e00  0 osd.51 24448261 load_pgs
>> > > opened 66 pgs
>> > > 2022-01-05 19:41:25 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:41:25.071+0000 7fca32943e00 -1 osd.51 24448261
>> > > log_to_monitors {default=true}
>> > > 2022-01-05 19:41:25 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:41:25.071+0000 7fca32943e00 -1 osd.51 24448261
>> > > log_to_monitors {default=true}
>> > > 2022-01-05 19:42:16 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:42:16.631+0000 7fca32943e00  0 osd.51 24448261 done with
>> > > init, starting boot process
>> > > 2022-01-05 19:42:16 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:42:16.631+0000 7fca32943e00  1 osd.51 24448261
>> start_boot
>> > > 2022-01-05 19:42:16 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:42:16.635+0000 7fca14615700  1 osd.51 pg_epoch: 24448130
>> > > pg[44.17( v 24448128'27126321 (24447767'27121032,24448128'27126321]
>> > > local-lis/les=24447864/24447865 n=2356 ec=4550661/4550661
>> > > lis/c=24447864/24447864 les/c/f=24447865/24447865/22709931
>> sis=24448130)
>> > > [51,48,15] r=0 lpr=24448130 pi=[24447864,24448130)/1
>> > crt=24448128'27126321
>> > > lcod 0'0 mlcod 0'0 unknown mbc={}] start_peering_interval up
>> [51,48,15]
>> > ->
>> > > [51,48,15], acting [51,48,15] -> [51,48,15], acting_primary 51 -> 51,
>> > > up_primary 51 -> 51, role 0 -> 0, features acting 4540138292840890367
>> > > upacting 4540138292840890367
>> > > 2022-01-05 19:42:16 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:42:16.635+0000 7fca13613700  1 osd.51 pg_epoch: 24448130
>> > > pg[44.1( v 24448129'31648690 (24447777'31643388,24448129'31648690]
>> > > local-lis/les=24447865/24447866 n=2314 ec=4550661/4550661
>> > > lis/c=24447865/24447865 les/c/f=24447866/24447866/22709931
>> sis=24448130)
>> > > [51,15,5] r=0 lpr=24448130 pi=[24447865,24448130)/1
>> crt=24448129'31648690
>> > > lcod 0'0 mlcod 0'0 unknown mbc={}] start_peering_interval up
>> [51,15,5] ->
>> > > [51,15,5], acting [51,15,5] -> [51,15,5], acting_primary 51 -> 51,
>> > > up_primary 51 -> 51, role 0 -> 0, features acting 4540138292840890367
>> > > upacting 4540138292840890367
>> > > 2022-01-05 19:42:16 bb-ceph-enc-rm63-osd03-31 osd.51
>> > > 2022-01-05T19:42:16.635+0000 7fca15617700  1 osd.51 pg_epoch: 24448130
>> > > pg[44.15( v 24448129'37939392 (24447777'37936883,24448129'37939392]
>> > > local-lis/les=24448118/24448119 n=2350 ec=4550661/4550661
>> > > lis/c=24448118/24448118 les/c/f=24448119/24448119/22709931
>> sis=24448130)
>> > > [5,14,51] r=2 lpr=24448130 pi=[24448118,24448130)/1
>> crt=24448129'37939392
>> > > lcod 0'0 mlcod 0'0 unknown mbc={}] start_peering_interval up
>> [5,14,51] ->
>> > > [5,14,51], acting [5,14,51] -> [5,14,51], acting_primary 5 -> 5,
>> > up_primary
>> > > 5 -> 5, role 2 -> 2, features acting 4540138292840890367 upacting
>> > > 4540138292840890367
>> > > 2022-01-05 19:42:51 bb-ceph-enc-rm63-osd03-31 init.scope
>> > ceph-osd@51.service:
>> > > Main process exited, code=killed, status=9/KILL
>> > > 2022-01-05 19:42:51 bb-ceph-enc-rm63-osd03-31 init.scope
>> > ceph-osd@51.service:
>> > > Failed with result 'signal'.
>> > > 2022-01-05 19:43:01 bb-ceph-enc-rm63-osd03-31 init.scope
>> > ceph-osd@51.service:
>> > > Scheduled restart job, restart counter is at 2.
>> > >
>> > >
>> > > The problem I have this has basically taken the production and
>> metadata
>> > SSD
>> > > pool's down fully and all 3 copies are offline. And I cannot find a
>> way
>> > to
>> > > find out what is causing these to crash.
>> > >
>> > > Kind Regards
>> > >
>> > > Lee
>> > > _______________________________________________
>> > > ceph-users mailing list -- ceph-users@xxxxxxx
>> > > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux