Re: bluefs _allocate unable to allocate

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Good morning,

i checked today the osd.8 and the log shows again the same error
bluefs _allocate unable to allocate 0x100000 on bdev 0, allocator name
bluefs-wal, allocator type hybrid, capacity 0xb40000000, block size
0x100000, free 0xff000, fragmentation 0, allocated 0x0

any idea why that could be?



Am Mi., 6. Okt. 2021 um 22:23 Uhr schrieb José H. Freidhof <
harald.freidhof@xxxxxxxxxxxxxx>:

> Hi Igor,
>
> today i repaired one osd node and all osd´s on the node, creating them new
> again....
> after that i waited for the rebalance/recovery process and the cluster was
> healthy after some hours..
>
> i notices that the osd.2 does not have any more this error in the log.
> but i noticed it now on the same node on osd.8... so i did the test that
> you suggested on osd.8
>
> it took nearly 20minutes to compact those db´s on the bluestore, but it
> helped... the problem on osd.8 is gone...
>
>
> *so the problem that i have with the alloc on the wal device seems to be
> random on different nodes and osd´s and looks like it comes, stay a while
> and disappears after a longer while...*
>
> here are the results that you suggested:
>
> root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluestore bluefs device info
> {
>     "dev": {
>         "device": "BDEV_WAL",
>         "total": 48318377984,
>         "free": 1044480,
>         "bluefs_used": 48317333504
>     },
>     "dev": {
>         "device": "BDEV_DB",
>         "total": 187904811008,
>         "free": 79842762752,
>         "bluefs_used": 108062048256
>     },
>     "dev": {
>         "device": "BDEV_SLOW",
>         "total": 6001172414464,
>         "free": 5510727389184,
>         "bluefs_used": 0,
>         "bluefs max available": 5508815847424
>     }
> }
> root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluefs stats
> 0 : device size 0xb3ffff000 : using 0xb3ff00000(45 GiB)
> 1 : device size 0x2bbfffe000 : using 0x1931500000(101 GiB)
> 2 : device size 0x57541c00000 : using 0x7235e3e000(457 GiB)
> RocksDBBlueFSVolumeSelector: wal_total:45902462976, db_total:178509578240,
> slow_total:5701113793740, db_avail:103884521472
> Usage matrix:
> DEV/LEV     WAL         DB          SLOW        *           *
> REAL        FILES
> LOG         304 MiB     7.9 GiB     0 B         0 B         0 B
> 9.7 MiB     1
> WAL         45 GiB      100 GiB     0 B         0 B         0 B
> 144 GiB     2319
> DB          0 B         276 MiB     0 B         0 B         0 B
> 249 MiB     47
> SLOW        0 B         0 B         0 B         0 B         0 B         0
> B         0
> TOTALS      45 GiB      109 GiB     0 B         0 B         0 B         0
> B         2367
> MAXIMUMS:
> LOG         304 MiB     7.9 GiB     0 B         0 B         0 B         20
> MiB
> WAL         45 GiB      149 GiB     0 B         0 B         0 B
> 192 GiB
> DB          0 B         762 MiB     0 B         0 B         0 B
> 738 MiB
> SLOW        0 B         0 B         0 B         0 B         0 B         0 B
> TOTALS      45 GiB      150 GiB     0 B         0 B         0 B         0 B
>
> ---
>
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.464+0000 7f4a9483a700  1 bluefs _allocate unable to
> allocate 0x400000 on bdev 0, allocator name bluefs-wal, allocator type
> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
> fragmentation 0, allocated 0x0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.472+0000 7f4a9483a700  1 bluefs _allocate unable to
> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
> fragmentation 0, allocated 0x0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.480+0000 7f4a9483a700  1 bluefs _allocate unable to
> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
> fragmentation 0, allocated 0x0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.500+0000 7f4a9483a700  1 bluefs _allocate unable to
> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
> fragmentation 0, allocated 0x0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.576+0000 7f4a9483a700  1 bluefs _allocate unable to
> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
> fragmentation 0, allocated 0x0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.624+0000 7f4a9483a700  1 bluefs _allocate unable to
> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
> fragmentation 0, allocated 0x0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.636+0000 7f4a9483a700  1 bluefs _allocate unable to
> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
> fragmentation 0, allocated 0x0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.884+0000 7f4a9483a700  1 bluefs _allocate unable to
> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
> fragmentation 0, allocated 0x0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.968+0000 7f4a9483a700  1 bluefs _allocate unable to
> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
> fragmentation 0, allocated 0x0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.992+0000 7f4a9483a700  4 rocksdb:
> [db_impl/db_impl_write.cc:1668] [L] New memtable created with log file:
> #13656. Immutable memtables: 1.
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.992+0000 7f4a9483a700  1 bluefs _allocate unable to
> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
> fragmentation 0, allocated 0x0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.992+0000 7f4a9483a700  1 bluefs _allocate unable to
> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
> fragmentation 0, allocated 0x0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb: (Original Log Time
> 2021/10/06-19:51:34.996331) [db_impl/db_impl_compaction_flush.cc:2198]
> Calling FlushMemTableToOutputFile with column family [L], flush slots
> available 1, compaction slots available 1, flush slots scheduled 1,
> compaction slots scheduled 0
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb: [flush_job.cc:321]
> [L] [JOB 8859] Flushing memtable with next log file: 13655
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb: [flush_job.cc:321]
> [L] [JOB 8859] Flushing memtable with next log file: 13656
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb: EVENT_LOG_v1
> {"time_micros": 1633549894998273, "job": 8859, "event": "flush_started",
> "num_memtables": 2, "num_entries": 3662, "num_deletes": 0,
> "total_data_size": 130482337, "memory_usage": 132976224, "flush_reason":
> "Write Buffer Full"}
> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb: [flush_job.cc:350]
> [L] [JOB 8859] Level-0 flush table #13657: started
> Oct 06 21:51:35 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:35.004+0000 7f4aab067700  4 rocksdb: EVENT_LOG_v1
> {"time_micros": 1633549895008271, "cf_name": "L", "job": 8859, "event":
> "table_file_creation", "file_number": 13657, "file_size": 2952537,
> "table_properties": {"data_size": 2951222, "index_size": 267,
> "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key":
> 0, "index_value_is_delta_encoded": 0, "filter_size": 197, "raw_key_size":
> 1120, "raw_average_key_size": 16, "raw_value_size": 2950151,
> "raw_average_value_size": 42145, "num_data_blocks": 9, "num_entries": 70,
> "num_deletions": 61, "num_merge_operands": 0, "num_range_deletions": 0,
> "format_version": 0, "fixed_key_len": 0, "filter_policy":
> "rocksdb.BuiltinBloomFilter", "column_family_name": "L",
> "column_family_id": 10, "comparator": "leveldb.BytewiseComparator",
> "merge_operator": "nullptr", "prefix_extractor_name": "nullptr",
> "property_collectors": "[]", "compression": "NoCompression",
> "compression_options": "window_bits=-14; level=32767; strategy=0;
> max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; ", "creation_time":
> 1633549854, "oldest_key_time": 1633549854, "file_creation_time":
> 1633549894}}
> Oct 06 21:51:35 cd88-ceph-osdh-01 bash[6328]: debug
> 2021-10-06T19:51:35.004+0000 7f4aab067700  4 rocksdb: [flush_job.cc:401]
> [L] [JOB 8859] Level-0 flush table #13657: 2952537 bytes OK
>
> ---
>
> root@cd88-ceph-osdh-01:~# ceph osd set noout
> root@cd88-ceph-osdh-01:~# ceph orch daemon stop osd.8
> root@cd88-ceph-osdh-01:~# ceph orch ps
> ...
> osd.7                                  cd133-ceph-osdh-01
>  running (4h)     44s ago    -    2738M    5528M  16.2.5     6933c2a0b7dd
>  8a98ae61f0eb
> osd.8                                  cd88-ceph-osdh-01
> stopped           5s ago    -        -    5528M  <unknown>  <unknown>
> <unknown>
> osd.9                                  cd133k-ceph-osdh-01
> running (3d)      5m ago    -    4673M    8966M  16.2.5     6933c2a0b7dd
>  0ff7584b1808
> ...
>
> ---
>
> root@cd88-ceph-osdh-01:~# ceph-kvstore-tool bluestore-kv
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/ compact
> 2021-10-06T21:53:50.559+0200 7f87bde3c240  0
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _open_db_and_around read-only:0 repair:0
> 2021-10-06T21:53:50.559+0200 7f87bde3c240  1 bdev(0x5644f056c800
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open path
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056c800
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open size
> 6001172414464 (0x57541c00000, 5.5 TiB) block_size 4096 (4 KiB) rotational
> discard not supported
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _set_cache_sizes cache_size 1073741824 meta 0.45 kv 0.45 data 0.06
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056cc00
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) open
> path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056cc00
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) open
> size 187904819200 (0x2bc0000000, 175 GiB) block_size 4096 (4 KiB)
> rotational discard not supported
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs add_block_device bdev
> 1 path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
> size 175 GiB
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056d000
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open path
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056d000
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open size
> 6001172414464 (0x57541c00000, 5.5 TiB) block_size 4096 (4 KiB) rotational
> discard not supported
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs add_block_device bdev
> 2 path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block size
> 5.5 TiB
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056d400
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) open
> path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056d400
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) open
> size 48318382080 (0xb40000000, 45 GiB) block_size 4096 (4 KiB)
> non-rotational discard supported
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs add_block_device bdev
> 0 path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
> size 45 GiB
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs mount
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs _init_alloc new, id 0,
> allocator name bluefs-wal, allocator type hybrid, capacity 0xb40000000,
> block size 0x100000
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs _init_alloc new, id 1,
> allocator name bluefs-db, allocator type hybrid, capacity 0x2bc0000000,
> block size 0x100000
> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs _init_alloc shared, id
> 2, capacity 0x57541c00000, block size 0x10000
> 2021-10-06T21:53:50.655+0200 7f87bde3c240  1 bluefs mount shared_bdev_used
> = 0
> 2021-10-06T21:53:50.655+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _prepare_db_environment set db_paths to db,178509578240
> db.slow,5701113793740
> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _open_db opened rocksdb path db options
> compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2,max_total_wal_size=1073741824
> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _open_super_meta old nid_max 167450
> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _open_super_meta old blobid_max 30720
> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _open_super_meta freelist_type bitmap
> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _open_super_meta ondisk_format 4 compat_ondisk_format 3
> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _open_super_meta min_alloc_size 0x1000
> 2021-10-06T22:01:33.347+0200 7f87bde3c240  1 freelist init
> 2021-10-06T22:01:33.347+0200 7f87bde3c240  1 freelist _read_cfg
> 2021-10-06T22:01:33.347+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _init_alloc opening allocation metadata
> 2021-10-06T22:01:41.031+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _init_alloc loaded 5.0 TiB in 37191 extents, allocator type hybrid,
> capacity 0x57541c00000, block size 0x1000, free 0x502f8f9a000,
> fragmentation 2.76445e-05
> 2021-10-06T22:01:41.039+0200 7f87bde3c240  1 bluefs umount
> 2021-10-06T22:01:41.043+0200 7f87bde3c240  1 bdev(0x5644f056d400
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) close
> 2021-10-06T22:01:43.623+0200 7f87bde3c240  1 bdev(0x5644f056cc00
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) close
> 2021-10-06T22:01:54.727+0200 7f87bde3c240  1 bdev(0x5644f056d000
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) close
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056d000
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) open
> path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056d000
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) open
> size 187904819200 (0x2bc0000000, 175 GiB) block_size 4096 (4 KiB)
> rotational discard not supported
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs add_block_device bdev
> 1 path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
> size 175 GiB
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056cc00
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open path
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056cc00
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open size
> 6001172414464 (0x57541c00000, 5.5 TiB) block_size 4096 (4 KiB) rotational
> discard not supported
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs add_block_device bdev
> 2 path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block size
> 5.5 TiB
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056d400
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) open
> path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056d400
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) open
> size 48318382080 (0xb40000000, 45 GiB) block_size 4096 (4 KiB)
> non-rotational discard supported
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs add_block_device bdev
> 0 path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
> size 45 GiB
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs mount
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs _init_alloc new, id 0,
> allocator name bluefs-wal, allocator type hybrid, capacity 0xb40000000,
> block size 0x100000
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs _init_alloc new, id 1,
> allocator name bluefs-db, allocator type hybrid, capacity 0x2bc0000000,
> block size 0x100000
> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs _init_alloc shared, id
> 2, capacity 0x57541c00000, block size 0x10000
> 2021-10-06T22:01:55.079+0200 7f87bde3c240  1 bluefs mount shared_bdev_used
> = 0
> 2021-10-06T22:01:55.079+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _prepare_db_environment set db_paths to db,178509578240
> db.slow,5701113793740
> 2021-10-06T22:09:36.519+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
> _open_db opened rocksdb path db options
> compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2,max_total_wal_size=1073741824
> 2021-10-06T22:09:54.067+0200 7f87bde3c240  1
> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/) umount
> 2021-10-06T22:09:54.079+0200 7f87bde3c240  1 bluefs umount
> 2021-10-06T22:09:54.079+0200 7f87bde3c240  1 bdev(0x5644f056d400
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) close
> 2021-10-06T22:09:56.612+0200 7f87bde3c240  1 bdev(0x5644f056d000
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) close
> 2021-10-06T22:10:07.520+0200 7f87bde3c240  1 bdev(0x5644f056cc00
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) close
> 2021-10-06T22:10:07.688+0200 7f87bde3c240  1 freelist shutdown
> 2021-10-06T22:10:07.692+0200 7f87bde3c240  1 bdev(0x5644f056c800
> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) close
>
> ---
>
> root@cd88-ceph-osdh-01:~# ceph orch daemon start osd.8
>
> ---
>
> root@cd88-ceph-osdh-01:/# ceph -s
>   cluster:
>     id:     462c44b4-eed6-11eb-8b2c-a1ad45f88a97
>     health: HEALTH_OK
>
>   services:
>     mon:         3 daemons, quorum
> cd133-ceph-mon-01,cd88-ceph-mon-01,cd133k-ceph-mon-01 (age 15h)
>     mgr:         cd133-ceph-mon-01.mzapob(active, since 15h), standbys:
> cd133k-ceph-mon-01.imikwh
>     osd:         60 osds: 60 up (since 2m), 60 in (since 3h)
>     rgw:         4 daemons active (2 hosts, 1 zones)
>     tcmu-runner: 10 portals active (2 hosts)
>
>   data:
>     pools:   6 pools, 361 pgs
>     objects: 2.46M objects, 8.0 TiB
>     usage:   33 TiB used, 304 TiB / 338 TiB avail
>     pgs:     361 active+clean
>
>   io:
>     client:   45 MiB/s rd, 50 MiB/s wr, 921 op/s rd, 674 op/s wr
>
>
>
> ---
>
> root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluestore bluefs device info
> {
>     "dev": {
>         "device": "BDEV_WAL",
>         "total": 48318377984,
>
> *        "free": 41354784768,*        "bluefs_used": 6963593216
>     },
>     "dev": {
>         "device": "BDEV_DB",
>         "total": 187904811008,
>         "free": 187302928384,
>         "bluefs_used": 601882624
>     },
>     "dev": {
>         "device": "BDEV_SLOW",
>         "total": 6001172414464,
>         "free": 5507531620352,
>         "bluefs_used": 0,
>         "bluefs max available": 5505566572544
>     }
> }
>
> ---
>
> root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluefs stats
> 0 : device size 0xb3ffff000 : using 0x1a0c00000(6.5 GiB)
> 1 : device size 0x2bbfffe000 : using 0x23e00000(574 MiB)
> 2 : device size 0x57541c00000 : using 0x72f0803000(460 GiB)
> RocksDBBlueFSVolumeSelector: wal_total:45902462976, db_total:178509578240,
> slow_total:5701113793740, db_avail:103884521472
> Usage matrix:
> DEV/LEV     WAL         DB          SLOW        *           *
> REAL        FILES
> LOG         12 MiB      18 MiB      0 B         0 B         0 B         10
> MiB      0
> WAL         6.5 GiB     0 B         0 B         0 B         0 B
> 6.4 GiB     102
> DB          0 B         573 MiB     0 B         0 B         0 B
> 557 MiB     22
> SLOW        0 B         0 B         0 B         0 B         0 B         0
> B         0
> TOTALS      6.5 GiB     591 MiB     0 B         0 B         0 B         0
> B         125
> MAXIMUMS:
> LOG         12 MiB      18 MiB      0 B         0 B         0 B         17
> MiB
> WAL         45 GiB      101 GiB     0 B         0 B         0 B
> 145 GiB
> DB          0 B         688 MiB     0 B         0 B         0 B
> 670 MiB
> SLOW        0 B         0 B         0 B         0 B         0 B         0 B
> TOTALS      45 GiB      101 GiB     0 B         0 B         0 B         0 B
>
> ----
>
>
> Here are the osd.2... the problem disapeared from alone
> very strange...
>
> root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluefs stats
> 0 : device size 0xb3ffff000 : using 0x7bcc00000(31 GiB)
> 1 : device size 0x2bbfffe000 : using 0x458c00000(17 GiB)
> 2 : device size 0x57541c00000 : using 0x5cd3665000(371 GiB)
> RocksDBBlueFSVolumeSelector: wal_total:45902462976, db_total:178509578240,
> slow_total:5701113793740, db_avail:103884521472
> Usage matrix:
> DEV/LEV     WAL         DB          SLOW        *           *
> REAL        FILES
> LOG         920 MiB     4.0 GiB     0 B         0 B         0 B         10
> MiB      1
> WAL         31 GiB      17 GiB      0 B         0 B         0 B         48
> GiB      765
> DB          0 B         193 MiB     0 B         0 B         0 B
> 175 MiB     30
> SLOW        0 B         0 B         0 B         0 B         0 B         0
> B         0
> TOTALS      32 GiB      21 GiB      0 B         0 B         0 B         0
> B         796
> MAXIMUMS:
> LOG         920 MiB     4.0 GiB     0 B         0 B         0 B         17
> MiB
> WAL         45 GiB      149 GiB     0 B         0 B         0 B
> 192 GiB
> DB          0 B         762 MiB     0 B         0 B         0 B
> 741 MiB
> SLOW        0 B         0 B         0 B         0 B         0 B         0 B
> TOTALS      45 GiB      153 GiB     0 B         0 B         0 B         0 B
> root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluestore bluefs device info
> {
>     "dev": {
>         "device": "BDEV_WAL",
>         "total": 48318377984,
>         "free": 15043915776,
>         "bluefs_used": 33274462208
>     },
>     "dev": {
>         "device": "BDEV_DB",
>         "total": 187904811008,
>         "free": 169235963904,
>         "bluefs_used": 18668847104
>     },
>     "dev": {
>         "device": "BDEV_SLOW",
>         "total": 6001172414464,
>         "free": 5602453327872,
>         "bluefs_used": 0,
>         "bluefs max available": 5600865222656
>     }
> }
>
>
>
>
>
>
> Am Mi., 6. Okt. 2021 um 18:11 Uhr schrieb Igor Fedotov <
> igor.fedotov@xxxxxxxx>:
>
>>
>> On 10/6/2021 4:25 PM, José H. Freidhof wrote:
>> > hi,
>> >
>> > no risk no fun 😂 okay
>> >   I have reset the settings you mentioned to standard.
>> >
>> > what you exactly mean with taking offline the osd? ceph orch daemon stop
>> > osd.2? or mark down?
>> "daemon stop" is enough. You  might want to set noout flag before that
>> though...
>> >
>> > for the command which path i use? you mean:
>> >
>> > bluestore-kv /var/lib/ceph/$fsid/osd.2 compact???
>> yep
>> >
>> >
>> > Igor Fedotov <ifedotov@xxxxxxx> schrieb am Mi., 6. Okt. 2021, 13:33:
>> >
>> >> On 10/6/2021 2:16 PM, José H. Freidhof wrote:
>> >>> Hi Igor,
>> >>>
>> >>> yes i have some osd settings set :-) here are my ceph config dump.
>> those
>> >>> settings are from a redhat document for bluestore devices
>> >>> maybe it is that setting causing this problem? "advanced
>> >>>    mon_compact_on_trim    false"???
>> >> OMG!!!
>> >>
>> >> No - mon_compact_on_trim has nothing to deal with bluestore.
>> >>
>> >> Highly likely it's bluestore_rocksdb_options which hurts...
>> >> Documentations tend to fall behind the best practices.... I would
>> >> strongly discourage you from using non-default settings unless it's
>> >> absolutely clear why this is necessary.
>> >>
>> >> Even at the first glance the following settings (just a few ones I'm
>> >> completely aware) are suboptimal/non-recommended:
>> >>
>> >> rocksdb_perf
>> >>
>> >> bluefs_sync_write
>> >>
>> >> bluefs_csum_type
>> >>
>> >>
>> >> Not to mention bluestore_rocksdb_options which hasn't got much adoption
>> >> so far and apparently greatly alters rocksdb behavior...
>> >>
>> >>
>> >> So I would suggest to revert rocksdb options back to default, run the
>> >> compaction and if it succeeds monitor the OSD for a while. Then if it
>> >> works fine - apply the same for others
>> >>
>> >>
>> >> Hope this helps,
>> >>
>> >> Igor
>> >>
>> >>
>> >>
>> >>> i will test it this afternoon... at the moment are everything semi
>> >>> prodcuctive and i need to repair one osd node.. because i think of
>> this
>> >>> reason the osds crashed on the node and the osd container crashes
>> with a
>> >>> dump while coming up now.
>> >>> need first to replicate all between all three nodes and then i can
>> take
>> >>> offline the osd.2.and test your command. i will inform you later...
>> >>>
>> >>> root@cd88-ceph-osdh-01:/# ceph config dump
>> >>> WHO                                               MASK
>> >>>    LEVEL     OPTION                                       VALUE
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>                                      RO
>> >>> global
>> >>>    advanced  leveldb_max_open_files                       131072
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    advanced  mon_compact_on_trim                          false
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    dev       ms_crc_data                                  false
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    advanced  osd_deep_scrub_interval
>> 1209600.000000
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    advanced  osd_max_scrubs                               16
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    advanced  osd_scrub_load_threshold                     0.010000
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    advanced  osd_scrub_max_interval
>>  1209600.000000
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    advanced  osd_scrub_min_interval                       86400.000000
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    advanced  perf                                         true
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    advanced  rbd_readahead_disable_after_bytes            0
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    advanced  rbd_readahead_max_bytes                      4194304
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    advanced  rocksdb_perf                                 true
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> global
>> >>>    advanced  throttler_perf_counter                       false
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     mon
>> >>> advanced  auth_allow_insecure_global_id_reclaim        false
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     mon
>> >>> advanced  cluster_network                              10.50.50.0/24
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>                                      *
>> >>>     mon
>> >>> advanced  mon_osd_down_out_interval                    300
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     mon
>> >>> advanced  public_network                               10.50.50.0/24
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>                                      *
>> >>>     mgr
>> >>> advanced  mgr/cephadm/container_init                   True
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>                                     *
>> >>>     mgr
>> >>> advanced  mgr/cephadm/device_enhanced_scan             true
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>                                     *
>> >>>     mgr
>> >>> advanced  mgr/cephadm/migration_current                2
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>                                      *
>> >>>     mgr
>> >>> advanced  mgr/cephadm/warn_on_stray_daemons            false
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>                                      *
>> >>>     mgr
>> >>> advanced  mgr/cephadm/warn_on_stray_hosts              false
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>                                      *
>> >>>     osd
>> >>> advanced  bluefs_sync_write                            true
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       bluestore_cache_autotune                     true
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       bluestore_cache_kv_ratio                     0.200000
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       bluestore_cache_meta_ratio                   0.800000
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       bluestore_cache_size                         2147483648
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       bluestore_cache_size_hdd                     2147483648
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> advanced  bluestore_csum_type                          none
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       bluestore_extent_map_shard_max_size          200
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       bluestore_extent_map_shard_min_size          50
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       bluestore_extent_map_shard_target_size       100
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> advanced  bluestore_rocksdb_options
>> >>>
>> >>
>>  compression=kNoCompression,max_write_buffer_number=64,min_write_buffer_number_to_merge=32,recycle_log_file_num=64,compaction_style=kCompactionStyleLevel,write_buffer_size=4MB,target_file_size_base=4MB,max_background_compactions=64,level0_file_num_compaction_trigger=64,level0_slowdown_writes_trigger=128,level0_stop_writes_trigger=256,max_bytes_for_level_base=6GB,compaction_threads=32,flusher_threads=8,compaction_readahead_size=2MB
>> >>>    *
>> >>>     osd
>> >>> advanced  mon_osd_cache_size                           1024
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       ms_crc_data                                  false
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> advanced  osd_map_share_max_epochs                     5
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> advanced  osd_max_backfills                            1
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       osd_max_pg_log_entries                       10
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       osd_memory_cache_min                         3000000000
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>   host:cd133-ceph-osdh-01
>> >>> basic     osd_memory_target                            5797322383
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>   host:cd133k-ceph-osdh-01
>> >>>    basic     osd_memory_target                            9402402385
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>>  host:cd88-ceph-osdh-01
>> >>>    basic     osd_memory_target                            5797322096
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> advanced  osd_memory_target_autotune                   true
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       osd_min_pg_log_entries                       10
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> advanced  osd_op_num_shards                            8
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>                                      *
>> >>>     osd
>> >>> advanced  osd_op_num_threads_per_shard                 2
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>                                      *
>> >>>     osd
>> >>> dev       osd_pg_log_dups_tracked                      10
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> dev       osd_pg_log_trim_min                          10
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> advanced  osd_recovery_max_active                      3
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> advanced  osd_recovery_max_single_start                1
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>     osd
>> >>> advanced  osd_recovery_sleep                           0.000000
>> >>>
>> >>>
>> >>> Am Mi., 6. Okt. 2021 um 12:55 Uhr schrieb Igor Fedotov <
>> ifedotov@xxxxxxx
>> >>> :
>> >>>
>> >>>> Jose,
>> >>>>
>> >>>> In fact 48GB is a way too much for WAL drive - usually the write
>> ahead
>> >> log
>> >>>> tend to be 2-4 GBs.
>> >>>>
>> >>>> But in your case it's ~150GB, while DB itself is very small
>> (146MB!!!):
>> >>>>
>> >>>> WAL         45 GiB      111 GiB     0 B         0 B         0 B
>> >>>> 154 GiB     2400
>> >>>>
>> >>>> DB          0 B         164 MiB     0 B         0 B         0 B
>> >>>> 146 MiB     30
>> >>>>
>> >>>>
>> >>>> which means that there are some issues with RocksDB's WAL processing,
>> >>>> which needs some troubleshooting...
>> >>>>
>> >>>> Curious if other OSDs are suffering from the same and whether you
>> have
>> >> any
>> >>>> custom settings for your OSD(s)?
>> >>>>
>> >>>> Additionally you might want to try the following command to compact
>> this
>> >>>> specific OSD manually and check if this would normalize the DB
>> layout -
>> >> the
>> >>>> majority of data has to be at DB level not WAL. Please share the
>> >> resulting
>> >>>> layout (reported by "ceph daemon osd.2 bluefs stats" command) after
>> the
>> >>>> compaction is fulfiled and OSD is restarted.
>> >>>>
>> >>>> The compaction command to be applied on an offline OSD:
>> >> "ceph-kvstore-tool
>> >>>> bluestore-kv <path-to-osd> compact"
>> >>>>
>> >>>> Even if the above works great please refrain from applying that
>> >> compaction
>> >>>> to every OSD - let's see how that "compacted" OSD evolves.Would WAL
>> grow
>> >>>> again or not?
>> >>>>
>> >>>> Thanks,
>> >>>>
>> >>>> Igor
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>> On 10/6/2021 1:35 PM, José H. Freidhof wrote:
>> >>>>
>> >>>> Hello Igor,
>> >>>>
>> >>>> yes the volume is  nvme wal partitions for the bluestore devicegroups
>> >> are
>> >>>> only 48gb each
>> >>>>
>> >>>> on each osd node are 1 nvme with 1tb splitted in 20 lvs with 48gb
>> (WAL)
>> >>>> on each osd node are 4 ssd with 1tb splitted in 5 lvs with 175gb
>> >> (rock.db)
>> >>>> on each osd node are 20 hdd with 5.5tb with 1 lvs (block.db)
>> >>>>
>> >>>> each blustore have 1 partition nvme,ssd and hdd like described in the
>> >>>> documentation
>> >>>>
>> >>
>> https://docs.ceph.com/en/latest/rados/configuration/bluestore-config-ref/
>> >>>> is this to small or can i adjust the max allocation on the wal nvme
>> >> device
>> >>>> in the ceph configuration?
>> >>>> i know that the ssd and nvme are to small for those 5.5tb disk...
>> its 1%
>> >>>> only ot the rotation disk.
>> >>>> i am new in ceph and still or always learning, but we are in a little
>> >>>> hurry because our other datastores are old and full.
>> >>>>
>> >>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluestore bluefs device
>> >> info
>> >>>> {
>> >>>>       "dev": {
>> >>>>           "device": "BDEV_WAL",
>> >>>>           "total": 48318377984,
>> >>>>           "free": 1044480,
>> >>>>           "bluefs_used": 48317333504
>> >>>>       },
>> >>>>       "dev": {
>> >>>>           "device": "BDEV_DB",
>> >>>>           "total": 187904811008,
>> >>>>           "free": 68757217280,
>> >>>>           "bluefs_used": 119147593728
>> >>>>       },
>> >>>>       "dev": {
>> >>>>           "device": "BDEV_SLOW",
>> >>>>           "total": 6001172414464,
>> >>>>           "free": 5624912359424,
>> >>>>           "bluefs_used": 0,
>> >>>>           "bluefs max available": 5624401231872
>> >>>>       }
>> >>>> }
>> >>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluefs stats
>> >>>> 0 : device size 0xb3ffff000 : using 0xb3ff00000(45 GiB)
>> >>>> 1 : device size 0x2bbfffe000 : using 0x1bbeb00000(111 GiB)
>> >>>> 2 : device size 0x57541c00000 : using 0x579b592000(350 GiB)
>> >>>> RocksDBBlueFSVolumeSelector: wal_total:45902462976,
>> >> db_total:178509578240,
>> >>>> slow_total:5701113793740, db_avail:103884521472
>> >>>> Usage matrix:
>> >>>> DEV/LEV     WAL         DB          SLOW        *           *
>> >>>> REAL        FILES
>> >>>> LOG         124 MiB     2.3 GiB     0 B         0 B         0 B
>> >>>> 7.5 MiB     1
>> >>>> WAL         45 GiB      111 GiB     0 B         0 B         0 B
>> >>>> 154 GiB     2400
>> >>>> DB          0 B         164 MiB     0 B         0 B         0 B
>> >>>> 146 MiB     30
>> >>>> SLOW        0 B         0 B         0 B         0 B         0 B
>> >>   0
>> >>>> B         0
>> >>>> TOTALS      45 GiB      113 GiB     0 B         0 B         0 B
>> >>   0
>> >>>> B         2431
>> >>>> MAXIMUMS:
>> >>>> LOG         124 MiB     2.3 GiB     0 B         0 B         0 B
>> >>   17
>> >>>> MiB
>> >>>> WAL         45 GiB      149 GiB     0 B         0 B         0 B
>> >>>> 192 GiB
>> >>>> DB          0 B         762 MiB     0 B         0 B         0 B
>> >>>> 741 MiB
>> >>>> SLOW        0 B         0 B         0 B         0 B         0 B
>> >>   0 B
>> >>>> TOTALS      45 GiB      150 GiB     0 B         0 B         0 B
>> >>   0 B
>> >>>> Am Mi., 6. Okt. 2021 um 11:45 Uhr schrieb Igor Fedotov <
>> >> ifedotov@xxxxxxx>:
>> >>>>> Hey Jose,
>> >>>>>
>> >>>>> it looks like your WAL volume is out of space which looks weird
>> given
>> >>>>> its capacity = 48Gb.
>> >>>>>
>> >>>>> Could you please share the output of the following commands:
>> >>>>>
>> >>>>> ceph daemon osd.N bluestore bluefs device info
>> >>>>>
>> >>>>> ceph daemon osd.N bluefs stats
>> >>>>>
>> >>>>>
>> >>>>> Thanks,
>> >>>>>
>> >>>>> Igor
>> >>>>>
>> >>>>>
>> >>>>> On 10/6/2021 12:24 PM, José H. Freidhof wrote:
>> >>>>>> Hello together
>> >>>>>>
>> >>>>>> we have a running ceph pacific 16.2.5 cluster and i found this
>> >> messages
>> >>>>> in
>> >>>>>> the service logs of the osd daemons.
>> >>>>>>
>> >>>>>> we have three osd nodes .. each node has 20osds as bluestore with
>> >>>>>> nvme/ssd/hdd
>> >>>>>>
>> >>>>>> is this a bug or maybe i have some settings wrong?
>> >>>>>>
>> >>>>>>
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:25.821+0000
>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>> bdev
>> >> 0,
>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>> >> 0xb40000000,
>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:29.857+0000
>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>> bdev
>> >> 0,
>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>> >> 0xb40000000,
>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.073+0000
>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x400000 on
>> bdev
>> >> 0,
>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>> >> 0xb40000000,
>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.405+0000
>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>> bdev
>> >> 0,
>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>> >> 0xb40000000,
>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.465+0000
>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>> bdev
>> >> 0,
>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>> >> 0xb40000000,
>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.529+0000
>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>> bdev
>> >> 0,
>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>> >> 0xb40000000,
>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>> >>>>>> 7f38eebd4700  4 rocksdb: [db_impl/db_impl_write.cc:1668] [L] New
>> >>>>> memtable
>> >>>>>> created with log file: #9588. Immutable memtables: 1.
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>> bdev
>> >> 0,
>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>> >> 0xb40000000,
>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>> >> 2021/10/06-09:17:30.547575)
>> >>>>>> [db_impl/db_impl_compaction_flush.cc:2198] Calling
>> >>>>>> FlushMemTableToOutputFile with column family [L], flush slots
>> >> available
>> >>>>> 1,
>> >>>>>> compaction slots available 1, flush slots scheduled 1, compaction
>> >> slots
>> >>>>>> scheduled 0
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>> >>>>>> 7f3905c02700  4 rocksdb: [flush_job.cc:321] [L] [JOB 5709] Flushing
>> >>>>>> memtable with next log file: 9587
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>> >>>>>> 7f3905c02700  4 rocksdb: [flush_job.cc:321] [L] [JOB 5709] Flushing
>> >>>>>> memtable with next log file: 9588
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>> >>>>>> 7f3905c02700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850547916,
>> >>>>>> "job": 5709, "event": "flush_started", "num_memtables": 2,
>> >>>>> "num_entries":
>> >>>>>> 4146, "num_deletes": 0, "total_data_size": 127203926,
>> "memory_usage":
>> >>>>>> 130479920, "flush_reason": "Write Buffer Full"}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>> >>>>>> 7f3905c02700  4 rocksdb: [flush_job.cc:350] [L] [JOB 5709] Level-0
>> >> flush
>> >>>>>> table #9589: started
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>> >>>>>> 7f3905c02700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850559292,
>> >>>>>> "cf_name": "L", "job": 5709, "event": "table_file_creation",
>> >>>>> "file_number":
>> >>>>>> 9589, "file_size": 3249934, "table_properties": {"data_size":
>> 3247855,
>> >>>>>> "index_size": 1031, "index_partitions": 0, "top_level_index_size":
>> 0,
>> >>>>>> "index_key_is_user_key": 0, "index_value_is_delta_encoded": 0,
>> >>>>>> "filter_size": 197, "raw_key_size": 1088, "raw_average_key_size":
>> 16,
>> >>>>>> "raw_value_size": 3246252, "raw_average_value_size": 47739,
>> >>>>>> "num_data_blocks": 36, "num_entries": 68, "num_deletions": 32,
>> >>>>>> "num_merge_operands": 0, "num_range_deletions": 0,
>> "format_version":
>> >> 0,
>> >>>>>> "fixed_key_len": 0, "filter_policy": "rocksdb.BuiltinBloomFilter",
>> >>>>>> "column_family_name": "L", "column_family_id": 10, "comparator":
>> >>>>>> "leveldb.BytewiseComparator", "merge_operator": "nullptr",
>> >>>>>> "prefix_extractor_name": "nullptr", "property_collectors": "[]",
>> >>>>>> "compression": "NoCompression", "compression_options":
>> >> "window_bits=-14;
>> >>>>>> level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0;
>> >>>>>> enabled=0; ", "creation_time": 1633511730, "oldest_key_time":
>> >>>>> 1633511730,
>> >>>>>> "file_creation_time": 1633511850}}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>> >>>>>> 7f3905c02700  4 rocksdb: [flush_job.cc:401] [L] [JOB 5709] Level-0
>> >> flush
>> >>>>>> table #9589: 3249934 bytes OK
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>> >> 2021/10/06-09:17:30.559362)
>> >>>>>> [memtable_list.cc:447] [L] Level-0 commit table #9589 started
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>> >> 2021/10/06-09:17:30.559583)
>> >>>>>> [memtable_list.cc:503] [L] Level-0 commit table #9589: memtable #1
>> >> done
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>> >> 2021/10/06-09:17:30.559586)
>> >>>>>> [memtable_list.cc:503] [L] Level-0 commit table #9589: memtable #2
>> >> done
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>> >> 2021/10/06-09:17:30.559601)
>> >>>>>> EVENT_LOG_v1 {"time_micros": 1633511850559593, "job": 5709,
>> "event":
>> >>>>>> "flush_finished", "output_compression": "NoCompression",
>> "lsm_state":
>> >>>>> [8,
>> >>>>>> 1, 0, 0, 0, 0, 0], "immutable_memtables": 0}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>> >> 2021/10/06-09:17:30.559638)
>> >>>>>> [db_impl/db_impl_compaction_flush.cc:205] [L] Level summary:
>> files[8 1
>> >>>>> 0 0
>> >>>>>> 0 0 0] max score 1.00
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1676] [L]
>> [JOB
>> >>>>> 5710]
>> >>>>>> Compacting 8@0 + 1@1 files to L1, score 1.00
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1680] [L]
>> >>>>> Compaction
>> >>>>>> start summary: Base version 3090 Base level 0, inputs:
>> [9589(3173KB)
>> >>>>>> 9586(4793KB) 9583(1876KB) 9580(194KB) 9576(6417KB) 9573(1078KB)
>> >>>>> 9570(405KB)
>> >>>>>> 9567(29KB)], [9564(1115KB)]
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850559956,
>> >>>>>> "job": 5710, "event": "compaction_started", "compaction_reason":
>> >>>>>> "LevelL0FilesNum", "files_L0": [9589, 9586, 9583, 9580, 9576, 9573,
>> >>>>> 9570,
>> >>>>>> 9567], "files_L1": [9564], "score": 1, "input_data_size": 19542092}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1349] [L]
>> [JOB
>> >>>>> 5710]
>> >>>>>> Generated table #9590: 36 keys, 3249524 bytes
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850582987,
>> >>>>>> "cf_name": "L", "job": 5710, "event": "table_file_creation",
>> >>>>> "file_number":
>> >>>>>> 9590, "file_size": 3249524, "table_properties": {"data_size":
>> 3247449,
>> >>>>>> "index_size": 1031, "index_partitions": 0, "top_level_index_size":
>> 0,
>> >>>>>> "index_key_is_user_key": 0, "index_value_is_delta_encoded": 0,
>> >>>>>> "filter_size": 197, "raw_key_size": 576, "raw_average_key_size":
>> 16,
>> >>>>>> "raw_value_size": 3246252, "raw_average_value_size": 90173,
>> >>>>>> "num_data_blocks": 36, "num_entries": 36, "num_deletions": 0,
>> >>>>>> "num_merge_operands": 0, "num_range_deletions": 0,
>> "format_version":
>> >> 0,
>> >>>>>> "fixed_key_len": 0, "filter_policy": "rocksdb.BuiltinBloomFilter",
>> >>>>>> "column_family_name": "L", "column_family_id": 10, "comparator":
>> >>>>>> "leveldb.BytewiseComparator", "merge_operator": "nullptr",
>> >>>>>> "prefix_extractor_name": "nullptr", "property_collectors": "[]",
>> >>>>>> "compression": "NoCompression", "compression_options":
>> >> "window_bits=-14;
>> >>>>>> level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0;
>> >>>>>> enabled=0; ", "creation_time": 1633471854, "oldest_key_time": 0,
>> >>>>>> "file_creation_time": 1633511850}}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1415] [L]
>> [JOB
>> >>>>> 5710]
>> >>>>>> Compacted 8@0 + 1@1 files to L1 => 3249524 bytes
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: (Original Log Time
>> >> 2021/10/06-09:17:30.583469)
>> >>>>>> [compaction/compaction_job.cc:760] [L] compacted to: files[0 1 0 0
>> 0 0
>> >>>>> 0]
>> >>>>>> max score 0.01, MB/sec: 846.1 rd, 140.7 wr, level 1, files in(8, 1)
>> >>>>> out(1)
>> >>>>>> MB in(17.5, 1.1) out(3.1), read-write-amplify(1.2)
>> write-amplify(0.2)
>> >>>>> OK,
>> >>>>>> records in: 376, records dropped: 340 output_compression:
>> >> NoCompression
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: (Original Log Time
>> >> 2021/10/06-09:17:30.583498)
>> >>>>>> EVENT_LOG_v1 {"time_micros": 1633511850583485, "job": 5710,
>> "event":
>> >>>>>> "compaction_finished", "compaction_time_micros": 23098,
>> >>>>>> "compaction_time_cpu_micros": 20039, "output_level": 1,
>> >>>>> "num_output_files":
>> >>>>>> 1, "total_output_size": 3249524, "num_input_records": 376,
>> >>>>>> "num_output_records": 36, "num_subcompactions": 1,
>> >> "output_compression":
>> >>>>>> "NoCompression", "num_single_delete_mismatches": 0,
>> >>>>>> "num_single_delete_fallthrough": 0, "lsm_state": [0, 1, 0, 0, 0, 0,
>> >> 0]}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850583615,
>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number": 9589}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850583648,
>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number": 9586}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850583675,
>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number": 9583}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850583709,
>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number": 9580}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850583739,
>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number": 9576}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850583769,
>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number": 9573}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850583804,
>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number": 9570}
>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>> >> 1633511850583835,
>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number": 9567}
>> >>>>>> _______________________________________________
>> >>>>>> ceph-users mailing list -- ceph-users@xxxxxxx
>> >>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >>>> --
>>
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux