Re: bluefs _allocate unable to allocate

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Igor

"Does single OSD startup (after if's experiencing "unable to allocate)
takes 20 mins as well?"
A: YES

Here the example log of the startup and recovery of a problematic osd.
https://paste.ubuntu.com/p/2WVJbg7cBy/

Here the example log of a problematic osd
https://paste.ubuntu.com/p/qbB6y7663f/

I found this post about a similar error and a bug in 16.2.4... we are
running 16.2.5...maybe the bug is not really fixed???
https://tracker.ceph.com/issues/50656
https://forum.proxmox.com/threads/ceph-16-2-pacific-cluster-crash.92367/



Am Mo., 11. Okt. 2021 um 11:53 Uhr schrieb Igor Fedotov <
igor.fedotov@xxxxxxxx>:

> hmm... so it looks like RocksDB still doesn't perform WAL cleanup during
> regular operation but applies it on OSD startup....
>
> Does single OSD startup (after if's experiencing "unable to allocate)
> takes 20 mins as well?
>
> Could you please share OSD log containing both that long startup and
> following (e.g. 1+ hour) regular operation?
>
> Preferable for OSD.2  (or whatever one which has been using default
> settings from the deployment).
>
>
> Thanks,
>
> Igor
>
>
> On 10/9/2021 12:18 AM, José H. Freidhof wrote:
>
> Hi Igor,
>
> "And was osd.2 redeployed AFTER settings had been reset to defaults ?"
> A: YES
>
> "Anything particular about current cluster use cases?"
> A: we are using it temporary as a iscsi target for a vmware esxi cluster
> with 6 hosts. We created two 10tb iscsi images/luns for vmware, because the
> other datastore are at 90%.
> We plan in the future, after ceph is working right, and stable to install
> openstack and kvm and we want to convert all vms into rbd images.
> Like i told you is a three osd nodes cluster with 32 cores and 256gb ram
> and two 10g bond network cards on a 10g network
>
> "E.g. is it a sort of regular usage (with load flukes and peak) or may be
> some permanently running stress load testing. The latter might tend to hold
> the resources and e.g. prevent from internal house keeping...
> A: Its a SAN for vmware and there are running 43 VMs at the moment... at
> the daytime is more stress on the disks because the people are working and
> in the afternoon the iops goes down because the users are at home
> noting speculative...
>
> There is something else that i noticed... if i reboot one osd with
> 20osds then it takes 20min to come up... if i tail the logs of the osd i
> can see a lot of " recovery log mode 2" on all osd
> after the 20min the osd comes one after one up and the waldb are small and
> no error in the logs about bluefs _allocate unable to allocate...
>
> it seems that the problem is rocking up after a longer time (12h)
>
>
> Am Fr., 8. Okt. 2021 um 15:24 Uhr schrieb Igor Fedotov <
> igor.fedotov@xxxxxxxx>:
>
>> And was osd.2 redeployed AFTER settings had been reset to defaults ?
>>
>> Anything particular about current cluster use cases?
>>
>> E.g. is it a sort of regular usage (with load flukes and peak) or may be
>> some permanently running stress load testing. The latter might tend to hold
>> the resources and e.g. prevent from internal house keeping...
>>
>> Igor
>>
>>
>> On 10/8/2021 12:16 AM, José H. Freidhof wrote:
>>
>> Hi Igor,
>>
>> yes the same problem is on osd.2
>>
>> we have 3 OSD Nodes... Each Node has 20 Bluestore OSDs ... in total we
>> have 60 OSDs
>> i checked right now one node... and 15 of 20 OSDs have this problem and
>> error in the log.
>>
>> the settings that you have complained some emails ago .. i have reverted
>> them to default.
>>
>> ceph.conf file:
>>
>> [global]
>>         fsid = 462c44b4-eed6-11eb-8b2c-a1ad45f88a97
>>         mon_host = [v2:10.50.50.21:3300/0,v1:10.50.50.21:6789/0] [v2:
>> 10.50.50.22:3300/0,v1:10.50.50.22:6789/0] [v2:
>> 10.50.50.20:3300/0,v1:10.50.50.20:6789/0]
>>         log file = /var/log/ceph/$cluster-$type-$id.log
>>         max open files = 131072
>>         mon compact on trim = False
>>         osd deep scrub interval = 137438953472
>>         osd max scrubs = 16
>>         osd objectstore = bluestore
>>         osd op threads = 2
>>         osd scrub load threshold = 0.01
>>         osd scrub max interval = 137438953472
>>         osd scrub min interval = 137438953472
>>         perf = True
>>         rbd readahead disable after bytes = 0
>>         rbd readahead max bytes = 4194304
>>         throttler perf counter = False
>>
>> [client]
>>         rbd cache = False
>>
>>
>> [mon]
>>         mon health preluminous compat = True
>>         mon osd down out interval = 300
>>
>> [osd]
>>         bluestore cache autotune = 0
>>         bluestore cache kv ratio = 0.2
>>         bluestore cache meta ratio = 0.8
>>         bluestore extent map shard max size = 200
>>         bluestore extent map shard min size = 50
>>         bluestore extent map shard target size = 100
>>         bluestore rocksdb options =
>> compression=kNoCompression,max_write_buffer_number=32,min_write_buffer_number_to_merge=2,recycle_log_file_num=32,compaction_style=kCompactionStyleLevel,write_buffer_size=67108864,target_file_size_base=67108864,max_background_compactions=31,level0_file_num_compaction_trigger=8,level0_slowdown_writes_trigger=32,level0_stop_writes_trigger=64,max_bytes_for_level_base=536870912,compaction_threads=32,max_bytes_for_level_multiplier=8,flusher_threads=8,compaction_readahead_size=2MB
>>         osd map share max epochs = 100
>>         osd max backfills = 5
>>         osd op num shards = 8
>>         osd op num threads per shard = 2
>>         osd min pg log entries = 10
>>         osd max pg log entries = 10
>>         osd pg log dups tracked = 10
>>         osd pg log trim min = 10
>>
>>
>>
>> root@cd133-ceph-osdh-01:~# ceph config dump
>> WHO                                               MASK
>>    LEVEL     OPTION                                       VALUE
>>
>>
>>
>>
>>
>>                                      RO
>> global
>>    basic     container_image
>> docker.io/ceph/ceph@sha256:829ebf54704f2d827de00913b171e5da741aad9b53c1f35ad59251524790eceb
>>
>>
>>
>>
>>                                        *
>> global
>>    advanced  leveldb_max_open_files                       131072
>>
>>
>>
>>
>> global
>>    advanced  mon_compact_on_trim                          false
>>
>>
>>
>>
>> global
>>    dev       ms_crc_data                                  false
>>
>>
>>
>>
>> global
>>    advanced  osd_deep_scrub_interval                      1209600.000000
>>
>>
>>
>>
>> global
>>    advanced  osd_max_scrubs                               16
>>
>>
>>
>>
>> global
>>    advanced  osd_scrub_load_threshold                     0.010000
>>
>>
>>
>>
>> global
>>    advanced  osd_scrub_max_interval                       1209600.000000
>>
>>
>>
>>
>> global
>>    advanced  osd_scrub_min_interval                       86400.000000
>>
>>
>>
>>
>> global
>>    advanced  perf                                         true
>>
>>
>>
>>
>> global
>>    advanced  rbd_readahead_disable_after_bytes            0
>>
>>
>>
>>
>> global
>>    advanced  rbd_readahead_max_bytes                      4194304
>>
>>
>>
>>
>> global
>>    advanced  throttler_perf_counter                       false
>>
>>
>>
>>
>>   mon
>>   advanced  auth_allow_insecure_global_id_reclaim        false
>>
>>
>>
>>
>>   mon
>>   advanced  cluster_network                              10.50.50.0/24
>>
>>
>>
>>
>>
>>                                        *
>>   mon
>>   advanced  mon_osd_down_out_interval                    300
>>
>>
>>
>>
>>   mon
>>   advanced  public_network                               10.50.50.0/24
>>
>>
>>
>>
>>
>>                                        *
>>   mgr
>>   advanced  mgr/cephadm/container_init                   True
>>
>>
>>
>>
>>
>>                                     *
>>   mgr
>>   advanced  mgr/cephadm/device_enhanced_scan             true
>>
>>
>>
>>
>>
>>                                     *
>>   mgr
>>   advanced  mgr/cephadm/migration_current                2
>>
>>
>>
>>
>>
>>                                      *
>>   mgr
>>   advanced  mgr/cephadm/warn_on_stray_daemons            false
>>
>>
>>
>>
>>
>>                                      *
>>   mgr
>>   advanced  mgr/cephadm/warn_on_stray_hosts              false
>>
>>
>>
>>
>>
>>                                      *
>>   mgr
>>   advanced  mgr/dashboard/10.50.50.21/server_addr
>>
>>
>>
>>
>>
>>                                         *
>>
>>
>>
>>                             *
>>   mgr
>>   advanced  mgr/dashboard/camdatadash/ssl_server_port    8443
>>
>>
>>
>>
>>
>>                                     *
>>   mgr
>>   advanced  mgr/dashboard/cd133-ceph-mon-01/server_addr
>>
>>
>>
>>
>>
>>                                     *
>>   mgr
>>   advanced  mgr/dashboard/dasboard/server_port           80
>>
>>
>>
>>
>>
>>                                     *
>>   mgr
>>   advanced  mgr/dashboard/dashboard/server_addr          10.251.133.161
>>
>>
>>
>>
>>
>>                                     *
>>   mgr
>>   advanced  mgr/dashboard/dashboard/ssl_server_port      8443
>>
>>
>>
>>
>>
>>                                     *
>>   mgr
>>   advanced  mgr/dashboard/server_addr                    0.0.0.0
>>
>>
>>
>>
>>
>>                                      *
>>   mgr
>>   advanced  mgr/dashboard/server_port                    8080
>>
>>
>>
>>
>>
>>                                     *
>>   mgr
>>   advanced  mgr/dashboard/ssl                            false
>>
>>
>>
>>
>>
>>                                      *
>>   mgr
>>   advanced  mgr/dashboard/ssl_server_port                8443
>>
>>
>>
>>
>>
>>                                     *
>>   mgr
>>   advanced  mgr/orchestrator/orchestrator                cephadm
>>
>>
>>
>>
>>   mgr
>>   advanced  mgr/prometheus/server_addr                   0.0.0.0
>>
>>
>>
>>
>>
>>                                      *
>>   mgr
>>   advanced  mgr/telemetry/channel_ident                  true
>>
>>
>>
>>
>>
>>                                     *
>>   mgr
>>   advanced  mgr/telemetry/enabled                        true
>>
>>
>>
>>
>>
>>                                     *
>>   mgr
>>   advanced  mgr/telemetry/last_opt_revision              3
>>
>>
>>
>>
>>
>>                                      *
>>   osd
>>   dev       bluestore_cache_autotune                     true
>>
>>
>>
>>
>>   osd
>>   dev       bluestore_cache_kv_ratio                     0.200000
>>
>>
>>
>>
>>   osd
>>   dev       bluestore_cache_meta_ratio                   0.800000
>>
>>
>>
>>
>>   osd
>>   dev       bluestore_cache_size                         2147483648
>>
>>
>>
>>
>>   osd
>>   dev       bluestore_cache_size_hdd                     2147483648
>>
>>
>>
>>
>>   osd
>>   dev       bluestore_extent_map_shard_max_size          200
>>
>>
>>
>>
>>   osd
>>   dev       bluestore_extent_map_shard_min_size          50
>>
>>
>>
>>
>>   osd
>>   dev       bluestore_extent_map_shard_target_size       100
>>
>>
>>
>>
>>   osd
>>   advanced  bluestore_rocksdb_options
>>  compression=kNoCompression,max_write_buffer_number=64,min_write_buffer_number_to_merge=32,recycle_log_file_num=64,compaction_style=kCompactionStyleLevel,write_buffer_size=4MB,target_file_size_base=4MB,max_background_compactions=64,level0_file_num_compaction_trigger=64,level0_slowdown_writes_trigger=128,level0_stop_writes_trigger=256,max_bytes_for_level_base=6GB,compaction_threads=32,flusher_threads=8,compaction_readahead_size=2MB
>>  *
>>   osd
>>   advanced  mon_osd_cache_size                           1024
>>
>>
>>
>>
>>   osd
>>   dev       ms_crc_data                                  false
>>
>>
>>
>>
>>   osd
>>   advanced  osd_map_share_max_epochs                     5
>>
>>
>>
>>
>>   osd
>>   advanced  osd_max_backfills                            1
>>
>>
>>
>>
>>   osd
>>   dev       osd_max_pg_log_entries                       10
>>
>>
>>
>>
>>   osd
>>   dev       osd_memory_cache_min                         3000000000
>>
>>
>>
>>
>>   osd                                             host:cd133-ceph-osdh-01
>>   basic     osd_memory_target                            5797322096
>>
>>
>>
>>
>>   osd
>> host:cd133k-ceph-osdh-01  basic     osd_memory_target
>>      9402402385
>>
>>
>>
>>   osd                                             host:cd88-ceph-osdh-01
>>    basic     osd_memory_target                            5797322096
>>
>>
>>
>>
>>   osd
>>   advanced  osd_memory_target_autotune                   true
>>
>>
>>
>>
>>   osd
>>   dev       osd_min_pg_log_entries                       10
>>
>>
>>
>>
>>   osd
>>   advanced  osd_op_num_shards                            8
>>
>>
>>
>>
>>
>>                                      *
>>   osd
>>   advanced  osd_op_num_threads_per_shard                 2
>>
>>
>>
>>
>>
>>                                      *
>>   osd
>>   dev       osd_pg_log_dups_tracked                      10
>>
>>
>>
>>
>>   osd
>>   dev       osd_pg_log_trim_min                          10
>>
>>
>>
>>
>>   osd
>>   advanced  osd_recovery_max_active                      3
>>
>>
>>
>>
>>   osd
>>   advanced  osd_recovery_max_single_start                1
>>
>>
>>
>>
>>   osd
>>   advanced  osd_recovery_sleep                           0.000000
>>
>>
>>
>>
>>   client
>>    advanced  rbd_cache                                    false
>>
>>
>>
>>
>>
>>
>>
>> Am Do., 7. Okt. 2021 um 19:27 Uhr schrieb Igor Fedotov <
>> igor.fedotov@xxxxxxxx>:
>>
>>> And does redeployed osd.2 expose the same issue (or at least DB/WAL
>>> disbalance) again? Were settings reverted to defaults for it as well?
>>>
>>>
>>> Thanks
>>>
>>> Igor
>>> On 10/7/2021 12:46 PM, José H. Freidhof wrote:
>>>
>>> Good morning,
>>>
>>> i checked today the osd.8 and the log shows again the same error
>>> bluefs _allocate unable to allocate 0x100000 on bdev 0, allocator name
>>> bluefs-wal, allocator type hybrid, capacity 0xb40000000, block size
>>> 0x100000, free 0xff000, fragmentation 0, allocated 0x0
>>>
>>> any idea why that could be?
>>>
>>> Am Mi., 6. Okt. 2021 um 22:23 Uhr schrieb José H. Freidhof <
>>> harald.freidhof@xxxxxxxxxxxxxx>:
>>>
>>>> Hi Igor,
>>>>
>>>> today i repaired one osd node and all osd´s on the node, creating them
>>>> new again....
>>>> after that i waited for the rebalance/recovery process and the cluster
>>>> was healthy after some hours..
>>>>
>>>> i notices that the osd.2 does not have any more this error in the log.
>>>> but i noticed it now on the same node on osd.8... so i did the test
>>>> that you suggested on osd.8
>>>>
>>>> it took nearly 20minutes to compact those db´s on the bluestore, but it
>>>> helped... the problem on osd.8 is gone...
>>>>
>>>>
>>>> *so the problem that i have with the alloc on the wal device seems to
>>>> be random on different nodes and osd´s and looks like it comes, stay
>>>> a while and disappears after a longer while... *
>>>>
>>>> here are the results that you suggested:
>>>>
>>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluestore bluefs device
>>>> info
>>>> {
>>>>     "dev": {
>>>>         "device": "BDEV_WAL",
>>>>         "total": 48318377984,
>>>>         "free": 1044480,
>>>>         "bluefs_used": 48317333504
>>>>     },
>>>>     "dev": {
>>>>         "device": "BDEV_DB",
>>>>         "total": 187904811008,
>>>>         "free": 79842762752,
>>>>         "bluefs_used": 108062048256
>>>>     },
>>>>     "dev": {
>>>>         "device": "BDEV_SLOW",
>>>>         "total": 6001172414464,
>>>>         "free": 5510727389184,
>>>>         "bluefs_used": 0,
>>>>         "bluefs max available": 5508815847424
>>>>     }
>>>> }
>>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluefs stats
>>>> 0 : device size 0xb3ffff000 : using 0xb3ff00000(45 GiB)
>>>> 1 : device size 0x2bbfffe000 : using 0x1931500000(101 GiB)
>>>> 2 : device size 0x57541c00000 : using 0x7235e3e000(457 GiB)
>>>> RocksDBBlueFSVolumeSelector: wal_total:45902462976,
>>>> db_total:178509578240, slow_total:5701113793740, db_avail:103884521472
>>>> Usage matrix:
>>>> DEV/LEV     WAL         DB          SLOW        *           *
>>>> REAL        FILES
>>>> LOG         304 MiB     7.9 GiB     0 B         0 B         0 B
>>>> 9.7 MiB     1
>>>> WAL         45 GiB      100 GiB     0 B         0 B         0 B
>>>> 144 GiB     2319
>>>> DB          0 B         276 MiB     0 B         0 B         0 B
>>>> 249 MiB     47
>>>> SLOW        0 B         0 B         0 B         0 B         0 B
>>>> 0 B         0
>>>> TOTALS      45 GiB      109 GiB     0 B         0 B         0 B
>>>> 0 B         2367
>>>> MAXIMUMS:
>>>> LOG         304 MiB     7.9 GiB     0 B         0 B         0 B
>>>> 20 MiB
>>>> WAL         45 GiB      149 GiB     0 B         0 B         0 B
>>>> 192 GiB
>>>> DB          0 B         762 MiB     0 B         0 B         0 B
>>>> 738 MiB
>>>> SLOW        0 B         0 B         0 B         0 B         0 B
>>>> 0 B
>>>> TOTALS      45 GiB      150 GiB     0 B         0 B         0 B
>>>> 0 B
>>>>
>>>> ---
>>>>
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.464+0000 7f4a9483a700  1 bluefs _allocate unable to
>>>> allocate 0x400000 on bdev 0, allocator name bluefs-wal, allocator type
>>>> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
>>>> fragmentation 0, allocated 0x0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.472+0000 7f4a9483a700  1 bluefs _allocate unable to
>>>> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
>>>> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
>>>> fragmentation 0, allocated 0x0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.480+0000 7f4a9483a700  1 bluefs _allocate unable to
>>>> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
>>>> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
>>>> fragmentation 0, allocated 0x0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.500+0000 7f4a9483a700  1 bluefs _allocate unable to
>>>> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
>>>> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
>>>> fragmentation 0, allocated 0x0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.576+0000 7f4a9483a700  1 bluefs _allocate unable to
>>>> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
>>>> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
>>>> fragmentation 0, allocated 0x0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.624+0000 7f4a9483a700  1 bluefs _allocate unable to
>>>> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
>>>> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
>>>> fragmentation 0, allocated 0x0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.636+0000 7f4a9483a700  1 bluefs _allocate unable to
>>>> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
>>>> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
>>>> fragmentation 0, allocated 0x0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.884+0000 7f4a9483a700  1 bluefs _allocate unable to
>>>> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
>>>> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
>>>> fragmentation 0, allocated 0x0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.968+0000 7f4a9483a700  1 bluefs _allocate unable to
>>>> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
>>>> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
>>>> fragmentation 0, allocated 0x0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.992+0000 7f4a9483a700  4 rocksdb:
>>>> [db_impl/db_impl_write.cc:1668] [L] New memtable created with log file:
>>>> #13656. Immutable memtables: 1.
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.992+0000 7f4a9483a700  1 bluefs _allocate unable to
>>>> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
>>>> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
>>>> fragmentation 0, allocated 0x0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.992+0000 7f4a9483a700  1 bluefs _allocate unable to
>>>> allocate 0x100000 on bdev 0, allocator name bluefs-wal, allocator type
>>>> hybrid, capacity 0xb40000000, block size 0x100000, free 0xff000,
>>>> fragmentation 0, allocated 0x0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb: (Original Log Time
>>>> 2021/10/06-19:51:34.996331) [db_impl/db_impl_compaction_flush.cc:2198]
>>>> Calling FlushMemTableToOutputFile with column family [L], flush slots
>>>> available 1, compaction slots available 1, flush slots scheduled 1,
>>>> compaction slots scheduled 0
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb: [flush_job.cc:321]
>>>> [L] [JOB 8859] Flushing memtable with next log file: 13655
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb: [flush_job.cc:321]
>>>> [L] [JOB 8859] Flushing memtable with next log file: 13656
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb: EVENT_LOG_v1
>>>> {"time_micros": 1633549894998273, "job": 8859, "event": "flush_started",
>>>> "num_memtables": 2, "num_entries": 3662, "num_deletes": 0,
>>>> "total_data_size": 130482337, "memory_usage": 132976224, "flush_reason":
>>>> "Write Buffer Full"}
>>>> Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb: [flush_job.cc:350]
>>>> [L] [JOB 8859] Level-0 flush table #13657: started
>>>> Oct 06 21:51:35 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:35.004+0000 7f4aab067700  4 rocksdb: EVENT_LOG_v1
>>>> {"time_micros": 1633549895008271, "cf_name": "L", "job": 8859, "event":
>>>> "table_file_creation", "file_number": 13657, "file_size": 2952537,
>>>> "table_properties": {"data_size": 2951222, "index_size": 267,
>>>> "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key":
>>>> 0, "index_value_is_delta_encoded": 0, "filter_size": 197, "raw_key_size":
>>>> 1120, "raw_average_key_size": 16, "raw_value_size": 2950151,
>>>> "raw_average_value_size": 42145, "num_data_blocks": 9, "num_entries": 70,
>>>> "num_deletions": 61, "num_merge_operands": 0, "num_range_deletions": 0,
>>>> "format_version": 0, "fixed_key_len": 0, "filter_policy":
>>>> "rocksdb.BuiltinBloomFilter", "column_family_name": "L",
>>>> "column_family_id": 10, "comparator": "leveldb.BytewiseComparator",
>>>> "merge_operator": "nullptr", "prefix_extractor_name": "nullptr",
>>>> "property_collectors": "[]", "compression": "NoCompression",
>>>> "compression_options": "window_bits=-14; level=32767; strategy=0;
>>>> max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; ", "creation_time":
>>>> 1633549854, "oldest_key_time": 1633549854, "file_creation_time":
>>>> 1633549894}}
>>>> Oct 06 21:51:35 cd88-ceph-osdh-01 bash[6328]: debug
>>>> 2021-10-06T19:51:35.004+0000 7f4aab067700  4 rocksdb: [flush_job.cc:401]
>>>> [L] [JOB 8859] Level-0 flush table #13657: 2952537 bytes OK
>>>>
>>>> ---
>>>>
>>>> root@cd88-ceph-osdh-01:~# ceph osd set noout
>>>> root@cd88-ceph-osdh-01:~# ceph orch daemon stop osd.8
>>>> root@cd88-ceph-osdh-01:~# ceph orch ps
>>>> ...
>>>> osd.7                                  cd133-ceph-osdh-01
>>>>  running (4h)     44s ago    -    2738M    5528M  16.2.5     6933c2a0b7dd
>>>>  8a98ae61f0eb
>>>> osd.8                                  cd88-ceph-osdh-01
>>>>   stopped           5s ago    -        -    5528M  <unknown>  <unknown>
>>>> <unknown>
>>>> osd.9                                  cd133k-ceph-osdh-01
>>>>   running (3d)      5m ago    -    4673M    8966M  16.2.5     6933c2a0b7dd
>>>>  0ff7584b1808
>>>> ...
>>>>
>>>> ---
>>>>
>>>> root@cd88-ceph-osdh-01:~# ceph-kvstore-tool bluestore-kv
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/ compact
>>>> 2021-10-06T21:53:50.559+0200 7f87bde3c240  0
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _open_db_and_around read-only:0 repair:0
>>>> 2021-10-06T21:53:50.559+0200 7f87bde3c240  1 bdev(0x5644f056c800
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open path
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056c800
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open size
>>>> 6001172414464 (0x57541c00000, 5.5 TiB) block_size 4096 (4 KiB) rotational
>>>> discard not supported
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _set_cache_sizes cache_size 1073741824 meta 0.45 kv 0.45 data 0.06
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056cc00
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) open
>>>> path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056cc00
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) open
>>>> size 187904819200 (0x2bc0000000, 175 GiB) block_size 4096 (4 KiB)
>>>> rotational discard not supported
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs add_block_device
>>>> bdev 1 path
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db size 175
>>>> GiB
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056d000
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open path
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056d000
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open size
>>>> 6001172414464 (0x57541c00000, 5.5 TiB) block_size 4096 (4 KiB) rotational
>>>> discard not supported
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs add_block_device
>>>> bdev 2 path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
>>>> size 5.5 TiB
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056d400
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) open
>>>> path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bdev(0x5644f056d400
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) open
>>>> size 48318382080 (0xb40000000, 45 GiB) block_size 4096 (4 KiB)
>>>> non-rotational discard supported
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs add_block_device
>>>> bdev 0 path
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal size 45
>>>> GiB
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs mount
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs _init_alloc new, id
>>>> 0, allocator name bluefs-wal, allocator type hybrid, capacity 0xb40000000,
>>>> block size 0x100000
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs _init_alloc new, id
>>>> 1, allocator name bluefs-db, allocator type hybrid, capacity 0x2bc0000000,
>>>> block size 0x100000
>>>> 2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs _init_alloc shared,
>>>> id 2, capacity 0x57541c00000, block size 0x10000
>>>> 2021-10-06T21:53:50.655+0200 7f87bde3c240  1 bluefs mount
>>>> shared_bdev_used = 0
>>>> 2021-10-06T21:53:50.655+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _prepare_db_environment set db_paths to db,178509578240
>>>> db.slow,5701113793740
>>>> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _open_db opened rocksdb path db options
>>>> compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2,max_total_wal_size=1073741824
>>>> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _open_super_meta old nid_max 167450
>>>> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _open_super_meta old blobid_max 30720
>>>> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _open_super_meta freelist_type bitmap
>>>> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _open_super_meta ondisk_format 4 compat_ondisk_format 3
>>>> 2021-10-06T22:01:32.715+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _open_super_meta min_alloc_size 0x1000
>>>> 2021-10-06T22:01:33.347+0200 7f87bde3c240  1 freelist init
>>>> 2021-10-06T22:01:33.347+0200 7f87bde3c240  1 freelist _read_cfg
>>>> 2021-10-06T22:01:33.347+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _init_alloc opening allocation metadata
>>>> 2021-10-06T22:01:41.031+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _init_alloc loaded 5.0 TiB in 37191 extents, allocator type hybrid,
>>>> capacity 0x57541c00000, block size 0x1000, free 0x502f8f9a000,
>>>> fragmentation 2.76445e-05
>>>> 2021-10-06T22:01:41.039+0200 7f87bde3c240  1 bluefs umount
>>>> 2021-10-06T22:01:41.043+0200 7f87bde3c240  1 bdev(0x5644f056d400
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) close
>>>> 2021-10-06T22:01:43.623+0200 7f87bde3c240  1 bdev(0x5644f056cc00
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) close
>>>> 2021-10-06T22:01:54.727+0200 7f87bde3c240  1 bdev(0x5644f056d000
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) close
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056d000
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) open
>>>> path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056d000
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) open
>>>> size 187904819200 (0x2bc0000000, 175 GiB) block_size 4096 (4 KiB)
>>>> rotational discard not supported
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs add_block_device
>>>> bdev 1 path
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db size 175
>>>> GiB
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056cc00
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open path
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056cc00
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) open size
>>>> 6001172414464 (0x57541c00000, 5.5 TiB) block_size 4096 (4 KiB) rotational
>>>> discard not supported
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs add_block_device
>>>> bdev 2 path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
>>>> size 5.5 TiB
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056d400
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) open
>>>> path /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bdev(0x5644f056d400
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) open
>>>> size 48318382080 (0xb40000000, 45 GiB) block_size 4096 (4 KiB)
>>>> non-rotational discard supported
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs add_block_device
>>>> bdev 0 path
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal size 45
>>>> GiB
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs mount
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs _init_alloc new, id
>>>> 0, allocator name bluefs-wal, allocator type hybrid, capacity 0xb40000000,
>>>> block size 0x100000
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs _init_alloc new, id
>>>> 1, allocator name bluefs-db, allocator type hybrid, capacity 0x2bc0000000,
>>>> block size 0x100000
>>>> 2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs _init_alloc shared,
>>>> id 2, capacity 0x57541c00000, block size 0x10000
>>>> 2021-10-06T22:01:55.079+0200 7f87bde3c240  1 bluefs mount
>>>> shared_bdev_used = 0
>>>> 2021-10-06T22:01:55.079+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _prepare_db_environment set db_paths to db,178509578240
>>>> db.slow,5701113793740
>>>> 2021-10-06T22:09:36.519+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
>>>> _open_db opened rocksdb path db options
>>>> compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2,max_total_wal_size=1073741824
>>>> 2021-10-06T22:09:54.067+0200 7f87bde3c240  1
>>>> bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/) umount
>>>> 2021-10-06T22:09:54.079+0200 7f87bde3c240  1 bluefs umount
>>>> 2021-10-06T22:09:54.079+0200 7f87bde3c240  1 bdev(0x5644f056d400
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal) close
>>>> 2021-10-06T22:09:56.612+0200 7f87bde3c240  1 bdev(0x5644f056d000
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db) close
>>>> 2021-10-06T22:10:07.520+0200 7f87bde3c240  1 bdev(0x5644f056cc00
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) close
>>>> 2021-10-06T22:10:07.688+0200 7f87bde3c240  1 freelist shutdown
>>>> 2021-10-06T22:10:07.692+0200 7f87bde3c240  1 bdev(0x5644f056c800
>>>> /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block) close
>>>>
>>>> ---
>>>>
>>>> root@cd88-ceph-osdh-01:~# ceph orch daemon start osd.8
>>>>
>>>> ---
>>>>
>>>> root@cd88-ceph-osdh-01:/# ceph -s
>>>>   cluster:
>>>>     id:     462c44b4-eed6-11eb-8b2c-a1ad45f88a97
>>>>     health: HEALTH_OK
>>>>
>>>>   services:
>>>>     mon:         3 daemons, quorum
>>>> cd133-ceph-mon-01,cd88-ceph-mon-01,cd133k-ceph-mon-01 (age 15h)
>>>>     mgr:         cd133-ceph-mon-01.mzapob(active, since 15h), standbys:
>>>> cd133k-ceph-mon-01.imikwh
>>>>     osd:         60 osds: 60 up (since 2m), 60 in (since 3h)
>>>>     rgw:         4 daemons active (2 hosts, 1 zones)
>>>>     tcmu-runner: 10 portals active (2 hosts)
>>>>
>>>>   data:
>>>>     pools:   6 pools, 361 pgs
>>>>     objects: 2.46M objects, 8.0 TiB
>>>>     usage:   33 TiB used, 304 TiB / 338 TiB avail
>>>>     pgs:     361 active+clean
>>>>
>>>>   io:
>>>>     client:   45 MiB/s rd, 50 MiB/s wr, 921 op/s rd, 674 op/s wr
>>>>
>>>>
>>>>
>>>> ---
>>>>
>>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluestore bluefs device
>>>> info
>>>> {
>>>>     "dev": {
>>>>         "device": "BDEV_WAL",
>>>>         "total": 48318377984,
>>>>
>>>> *        "free": 41354784768, *        "bluefs_used": 6963593216
>>>>     },
>>>>     "dev": {
>>>>         "device": "BDEV_DB",
>>>>         "total": 187904811008,
>>>>         "free": 187302928384,
>>>>         "bluefs_used": 601882624
>>>>     },
>>>>     "dev": {
>>>>         "device": "BDEV_SLOW",
>>>>         "total": 6001172414464,
>>>>         "free": 5507531620352,
>>>>         "bluefs_used": 0,
>>>>         "bluefs max available": 5505566572544
>>>>     }
>>>> }
>>>>
>>>> ---
>>>>
>>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluefs stats
>>>> 0 : device size 0xb3ffff000 : using 0x1a0c00000(6.5 GiB)
>>>> 1 : device size 0x2bbfffe000 : using 0x23e00000(574 MiB)
>>>> 2 : device size 0x57541c00000 : using 0x72f0803000(460 GiB)
>>>> RocksDBBlueFSVolumeSelector: wal_total:45902462976,
>>>> db_total:178509578240, slow_total:5701113793740, db_avail:103884521472
>>>> Usage matrix:
>>>> DEV/LEV     WAL         DB          SLOW        *           *
>>>> REAL        FILES
>>>> LOG         12 MiB      18 MiB      0 B         0 B         0 B
>>>> 10 MiB      0
>>>> WAL         6.5 GiB     0 B         0 B         0 B         0 B
>>>> 6.4 GiB     102
>>>> DB          0 B         573 MiB     0 B         0 B         0 B
>>>> 557 MiB     22
>>>> SLOW        0 B         0 B         0 B         0 B         0 B
>>>> 0 B         0
>>>> TOTALS      6.5 GiB     591 MiB     0 B         0 B         0 B
>>>> 0 B         125
>>>> MAXIMUMS:
>>>> LOG         12 MiB      18 MiB      0 B         0 B         0 B
>>>> 17 MiB
>>>> WAL         45 GiB      101 GiB     0 B         0 B         0 B
>>>> 145 GiB
>>>> DB          0 B         688 MiB     0 B         0 B         0 B
>>>> 670 MiB
>>>> SLOW        0 B         0 B         0 B         0 B         0 B
>>>> 0 B
>>>> TOTALS      45 GiB      101 GiB     0 B         0 B         0 B
>>>> 0 B
>>>>
>>>> ----
>>>>
>>>>
>>>> Here are the osd.2... the problem disapeared from alone
>>>> very strange...
>>>>
>>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluefs stats
>>>> 0 : device size 0xb3ffff000 : using 0x7bcc00000(31 GiB)
>>>> 1 : device size 0x2bbfffe000 : using 0x458c00000(17 GiB)
>>>> 2 : device size 0x57541c00000 : using 0x5cd3665000(371 GiB)
>>>> RocksDBBlueFSVolumeSelector: wal_total:45902462976,
>>>> db_total:178509578240, slow_total:5701113793740, db_avail:103884521472
>>>> Usage matrix:
>>>> DEV/LEV     WAL         DB          SLOW        *           *
>>>> REAL        FILES
>>>> LOG         920 MiB     4.0 GiB     0 B         0 B         0 B
>>>> 10 MiB      1
>>>> WAL         31 GiB      17 GiB      0 B         0 B         0 B
>>>> 48 GiB      765
>>>> DB          0 B         193 MiB     0 B         0 B         0 B
>>>> 175 MiB     30
>>>> SLOW        0 B         0 B         0 B         0 B         0 B
>>>> 0 B         0
>>>> TOTALS      32 GiB      21 GiB      0 B         0 B         0 B
>>>> 0 B         796
>>>> MAXIMUMS:
>>>> LOG         920 MiB     4.0 GiB     0 B         0 B         0 B
>>>> 17 MiB
>>>> WAL         45 GiB      149 GiB     0 B         0 B         0 B
>>>> 192 GiB
>>>> DB          0 B         762 MiB     0 B         0 B         0 B
>>>> 741 MiB
>>>> SLOW        0 B         0 B         0 B         0 B         0 B
>>>> 0 B
>>>> TOTALS      45 GiB      153 GiB     0 B         0 B         0 B
>>>> 0 B
>>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluestore bluefs device
>>>> info
>>>> {
>>>>     "dev": {
>>>>         "device": "BDEV_WAL",
>>>>         "total": 48318377984,
>>>>         "free": 15043915776,
>>>>         "bluefs_used": 33274462208
>>>>     },
>>>>     "dev": {
>>>>         "device": "BDEV_DB",
>>>>         "total": 187904811008,
>>>>         "free": 169235963904,
>>>>         "bluefs_used": 18668847104
>>>>     },
>>>>     "dev": {
>>>>         "device": "BDEV_SLOW",
>>>>         "total": 6001172414464,
>>>>         "free": 5602453327872,
>>>>         "bluefs_used": 0,
>>>>         "bluefs max available": 5600865222656
>>>>     }
>>>> }
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Am Mi., 6. Okt. 2021 um 18:11 Uhr schrieb Igor Fedotov <
>>>> igor.fedotov@xxxxxxxx>:
>>>>
>>>>>
>>>>> On 10/6/2021 4:25 PM, José H. Freidhof wrote:
>>>>> > hi,
>>>>> >
>>>>> > no risk no fun 😂 okay
>>>>> >   I have reset the settings you mentioned to standard.
>>>>> >
>>>>> > what you exactly mean with taking offline the osd? ceph orch daemon
>>>>> stop
>>>>> > osd.2? or mark down?
>>>>> "daemon stop" is enough. You  might want to set noout flag before that
>>>>> though...
>>>>> >
>>>>> > for the command which path i use? you mean:
>>>>> >
>>>>> > bluestore-kv /var/lib/ceph/$fsid/osd.2 compact???
>>>>> yep
>>>>> >
>>>>> >
>>>>> > Igor Fedotov <ifedotov@xxxxxxx> schrieb am Mi., 6. Okt. 2021, 13:33:
>>>>> >
>>>>> >> On 10/6/2021 2:16 PM, José H. Freidhof wrote:
>>>>> >>> Hi Igor,
>>>>> >>>
>>>>> >>> yes i have some osd settings set :-) here are my ceph config dump.
>>>>> those
>>>>> >>> settings are from a redhat document for bluestore devices
>>>>> >>> maybe it is that setting causing this problem? "advanced
>>>>> >>>    mon_compact_on_trim    false"???
>>>>> >> OMG!!!
>>>>> >>
>>>>> >> No - mon_compact_on_trim has nothing to deal with bluestore.
>>>>> >>
>>>>> >> Highly likely it's bluestore_rocksdb_options which hurts...
>>>>> >> Documentations tend to fall behind the best practices.... I would
>>>>> >> strongly discourage you from using non-default settings unless it's
>>>>> >> absolutely clear why this is necessary.
>>>>> >>
>>>>> >> Even at the first glance the following settings (just a few ones I'm
>>>>> >> completely aware) are suboptimal/non-recommended:
>>>>> >>
>>>>> >> rocksdb_perf
>>>>> >>
>>>>> >> bluefs_sync_write
>>>>> >>
>>>>> >> bluefs_csum_type
>>>>> >>
>>>>> >>
>>>>> >> Not to mention bluestore_rocksdb_options which hasn't got much
>>>>> adoption
>>>>> >> so far and apparently greatly alters rocksdb behavior...
>>>>> >>
>>>>> >>
>>>>> >> So I would suggest to revert rocksdb options back to default, run
>>>>> the
>>>>> >> compaction and if it succeeds monitor the OSD for a while. Then if
>>>>> it
>>>>> >> works fine - apply the same for others
>>>>> >>
>>>>> >>
>>>>> >> Hope this helps,
>>>>> >>
>>>>> >> Igor
>>>>> >>
>>>>> >>
>>>>> >>
>>>>> >>> i will test it this afternoon... at the moment are everything semi
>>>>> >>> prodcuctive and i need to repair one osd node.. because i think of
>>>>> this
>>>>> >>> reason the osds crashed on the node and the osd container crashes
>>>>> with a
>>>>> >>> dump while coming up now.
>>>>> >>> need first to replicate all between all three nodes and then i can
>>>>> take
>>>>> >>> offline the osd.2.and test your command. i will inform you later...
>>>>> >>>
>>>>> >>> root@cd88-ceph-osdh-01:/# ceph config dump
>>>>> >>> WHO                                               MASK
>>>>> >>>    LEVEL     OPTION                                       VALUE
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>                                      RO
>>>>> >>> global
>>>>> >>>    advanced  leveldb_max_open_files                       131072
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    advanced  mon_compact_on_trim                          false
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    dev       ms_crc_data                                  false
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    advanced  osd_deep_scrub_interval
>>>>> 1209600.000000
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    advanced  osd_max_scrubs                               16
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    advanced  osd_scrub_load_threshold                     0.010000
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    advanced  osd_scrub_max_interval
>>>>>  1209600.000000
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    advanced  osd_scrub_min_interval
>>>>>  86400.000000
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    advanced  perf                                         true
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    advanced  rbd_readahead_disable_after_bytes            0
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    advanced  rbd_readahead_max_bytes                      4194304
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    advanced  rocksdb_perf                                 true
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>> global
>>>>> >>>    advanced  throttler_perf_counter                       false
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     mon
>>>>> >>> advanced  auth_allow_insecure_global_id_reclaim        false
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     mon
>>>>> >>> advanced  cluster_network
>>>>> 10.50.50.0/24
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>                                      *
>>>>> >>>     mon
>>>>> >>> advanced  mon_osd_down_out_interval                    300
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     mon
>>>>> >>> advanced  public_network
>>>>> 10.50.50.0/24
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>                                      *
>>>>> >>>     mgr
>>>>> >>> advanced  mgr/cephadm/container_init                   True
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>                                     *
>>>>> >>>     mgr
>>>>> >>> advanced  mgr/cephadm/device_enhanced_scan             true
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>                                     *
>>>>> >>>     mgr
>>>>> >>> advanced  mgr/cephadm/migration_current                2
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>                                      *
>>>>> >>>     mgr
>>>>> >>> advanced  mgr/cephadm/warn_on_stray_daemons            false
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>                                      *
>>>>> >>>     mgr
>>>>> >>> advanced  mgr/cephadm/warn_on_stray_hosts              false
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>                                      *
>>>>> >>>     osd
>>>>> >>> advanced  bluefs_sync_write                            true
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       bluestore_cache_autotune                     true
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       bluestore_cache_kv_ratio                     0.200000
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       bluestore_cache_meta_ratio                   0.800000
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       bluestore_cache_size                         2147483648
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       bluestore_cache_size_hdd                     2147483648
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> advanced  bluestore_csum_type                          none
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       bluestore_extent_map_shard_max_size          200
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       bluestore_extent_map_shard_min_size          50
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       bluestore_extent_map_shard_target_size       100
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> advanced  bluestore_rocksdb_options
>>>>> >>>
>>>>> >>
>>>>>  compression=kNoCompression,max_write_buffer_number=64,min_write_buffer_number_to_merge=32,recycle_log_file_num=64,compaction_style=kCompactionStyleLevel,write_buffer_size=4MB,target_file_size_base=4MB,max_background_compactions=64,level0_file_num_compaction_trigger=64,level0_slowdown_writes_trigger=128,level0_stop_writes_trigger=256,max_bytes_for_level_base=6GB,compaction_threads=32,flusher_threads=8,compaction_readahead_size=2MB
>>>>> >>>    *
>>>>> >>>     osd
>>>>> >>> advanced  mon_osd_cache_size                           1024
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       ms_crc_data                                  false
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> advanced  osd_map_share_max_epochs                     5
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> advanced  osd_max_backfills                            1
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       osd_max_pg_log_entries                       10
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       osd_memory_cache_min                         3000000000
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>   host:cd133-ceph-osdh-01
>>>>> >>> basic     osd_memory_target                            5797322383
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>   host:cd133k-ceph-osdh-01
>>>>> >>>    basic     osd_memory_target
>>>>> 9402402385
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>>  host:cd88-ceph-osdh-01
>>>>> >>>    basic     osd_memory_target
>>>>> 5797322096
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> advanced  osd_memory_target_autotune                   true
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       osd_min_pg_log_entries                       10
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> advanced  osd_op_num_shards                            8
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>                                      *
>>>>> >>>     osd
>>>>> >>> advanced  osd_op_num_threads_per_shard                 2
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>                                      *
>>>>> >>>     osd
>>>>> >>> dev       osd_pg_log_dups_tracked                      10
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> dev       osd_pg_log_trim_min                          10
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> advanced  osd_recovery_max_active                      3
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> advanced  osd_recovery_max_single_start                1
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>
>>>>> >>>     osd
>>>>> >>> advanced  osd_recovery_sleep                           0.000000
>>>>> >>>
>>>>> >>>
>>>>> >>> Am Mi., 6. Okt. 2021 um 12:55 Uhr schrieb Igor Fedotov <
>>>>> ifedotov@xxxxxxx
>>>>> >>> :
>>>>> >>>
>>>>> >>>> Jose,
>>>>> >>>>
>>>>> >>>> In fact 48GB is a way too much for WAL drive - usually the write
>>>>> ahead
>>>>> >> log
>>>>> >>>> tend to be 2-4 GBs.
>>>>> >>>>
>>>>> >>>> But in your case it's ~150GB, while DB itself is very small
>>>>> (146MB!!!):
>>>>> >>>>
>>>>> >>>> WAL         45 GiB      111 GiB     0 B         0 B         0 B
>>>>> >>>> 154 GiB     2400
>>>>> >>>>
>>>>> >>>> DB          0 B         164 MiB     0 B         0 B         0 B
>>>>> >>>> 146 MiB     30
>>>>> >>>>
>>>>> >>>>
>>>>> >>>> which means that there are some issues with RocksDB's WAL
>>>>> processing,
>>>>> >>>> which needs some troubleshooting...
>>>>> >>>>
>>>>> >>>> Curious if other OSDs are suffering from the same and whether you
>>>>> have
>>>>> >> any
>>>>> >>>> custom settings for your OSD(s)?
>>>>> >>>>
>>>>> >>>> Additionally you might want to try the following command to
>>>>> compact this
>>>>> >>>> specific OSD manually and check if this would normalize the DB
>>>>> layout -
>>>>> >> the
>>>>> >>>> majority of data has to be at DB level not WAL. Please share the
>>>>> >> resulting
>>>>> >>>> layout (reported by "ceph daemon osd.2 bluefs stats" command)
>>>>> after the
>>>>> >>>> compaction is fulfiled and OSD is restarted.
>>>>> >>>>
>>>>> >>>> The compaction command to be applied on an offline OSD:
>>>>> >> "ceph-kvstore-tool
>>>>> >>>> bluestore-kv <path-to-osd> compact"
>>>>> >>>>
>>>>> >>>> Even if the above works great please refrain from applying that
>>>>> >> compaction
>>>>> >>>> to every OSD - let's see how that "compacted" OSD evolves.Would
>>>>> WAL grow
>>>>> >>>> again or not?
>>>>> >>>>
>>>>> >>>> Thanks,
>>>>> >>>>
>>>>> >>>> Igor
>>>>> >>>>
>>>>> >>>>
>>>>> >>>>
>>>>> >>>>
>>>>> >>>>
>>>>> >>>>
>>>>> >>>> On 10/6/2021 1:35 PM, José H. Freidhof wrote:
>>>>> >>>>
>>>>> >>>> Hello Igor,
>>>>> >>>>
>>>>> >>>> yes the volume is  nvme wal partitions for the bluestore
>>>>> devicegroups
>>>>> >> are
>>>>> >>>> only 48gb each
>>>>> >>>>
>>>>> >>>> on each osd node are 1 nvme with 1tb splitted in 20 lvs with 48gb
>>>>> (WAL)
>>>>> >>>> on each osd node are 4 ssd with 1tb splitted in 5 lvs with 175gb
>>>>> >> (rock.db)
>>>>> >>>> on each osd node are 20 hdd with 5.5tb with 1 lvs (block.db)
>>>>> >>>>
>>>>> >>>> each blustore have 1 partition nvme,ssd and hdd like described in
>>>>> the
>>>>> >>>> documentation
>>>>> >>>>
>>>>> >>
>>>>> https://docs.ceph.com/en/latest/rados/configuration/bluestore-config-ref/
>>>>> >>>> is this to small or can i adjust the max allocation on the wal
>>>>> nvme
>>>>> >> device
>>>>> >>>> in the ceph configuration?
>>>>> >>>> i know that the ssd and nvme are to small for those 5.5tb disk...
>>>>> its 1%
>>>>> >>>> only ot the rotation disk.
>>>>> >>>> i am new in ceph and still or always learning, but we are in a
>>>>> little
>>>>> >>>> hurry because our other datastores are old and full.
>>>>> >>>>
>>>>> >>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluestore bluefs
>>>>> device
>>>>> >> info
>>>>> >>>> {
>>>>> >>>>       "dev": {
>>>>> >>>>           "device": "BDEV_WAL",
>>>>> >>>>           "total": 48318377984,
>>>>> >>>>           "free": 1044480,
>>>>> >>>>           "bluefs_used": 48317333504
>>>>> >>>>       },
>>>>> >>>>       "dev": {
>>>>> >>>>           "device": "BDEV_DB",
>>>>> >>>>           "total": 187904811008,
>>>>> >>>>           "free": 68757217280,
>>>>> >>>>           "bluefs_used": 119147593728
>>>>> >>>>       },
>>>>> >>>>       "dev": {
>>>>> >>>>           "device": "BDEV_SLOW",
>>>>> >>>>           "total": 6001172414464,
>>>>> >>>>           "free": 5624912359424,
>>>>> >>>>           "bluefs_used": 0,
>>>>> >>>>           "bluefs max available": 5624401231872
>>>>> >>>>       }
>>>>> >>>> }
>>>>> >>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluefs stats
>>>>> >>>> 0 : device size 0xb3ffff000 : using 0xb3ff00000(45 GiB)
>>>>> >>>> 1 : device size 0x2bbfffe000 : using 0x1bbeb00000(111 GiB)
>>>>> >>>> 2 : device size 0x57541c00000 : using 0x579b592000(350 GiB)
>>>>> >>>> RocksDBBlueFSVolumeSelector: wal_total:45902462976,
>>>>> >> db_total:178509578240,
>>>>> >>>> slow_total:5701113793740, db_avail:103884521472
>>>>> >>>> Usage matrix:
>>>>> >>>> DEV/LEV     WAL         DB          SLOW        *           *
>>>>> >>>> REAL        FILES
>>>>> >>>> LOG         124 MiB     2.3 GiB     0 B         0 B         0 B
>>>>> >>>> 7.5 MiB     1
>>>>> >>>> WAL         45 GiB      111 GiB     0 B         0 B         0 B
>>>>> >>>> 154 GiB     2400
>>>>> >>>> DB          0 B         164 MiB     0 B         0 B         0 B
>>>>> >>>> 146 MiB     30
>>>>> >>>> SLOW        0 B         0 B         0 B         0 B         0 B
>>>>> >>   0
>>>>> >>>> B         0
>>>>> >>>> TOTALS      45 GiB      113 GiB     0 B         0 B         0 B
>>>>> >>   0
>>>>> >>>> B         2431
>>>>> >>>> MAXIMUMS:
>>>>> >>>> LOG         124 MiB     2.3 GiB     0 B         0 B         0 B
>>>>> >>   17
>>>>> >>>> MiB
>>>>> >>>> WAL         45 GiB      149 GiB     0 B         0 B         0 B
>>>>> >>>> 192 GiB
>>>>> >>>> DB          0 B         762 MiB     0 B         0 B         0 B
>>>>> >>>> 741 MiB
>>>>> >>>> SLOW        0 B         0 B         0 B         0 B         0 B
>>>>> >>   0 B
>>>>> >>>> TOTALS      45 GiB      150 GiB     0 B         0 B         0 B
>>>>> >>   0 B
>>>>> >>>> Am Mi., 6. Okt. 2021 um 11:45 Uhr schrieb Igor Fedotov <
>>>>> >> ifedotov@xxxxxxx>:
>>>>> >>>>> Hey Jose,
>>>>> >>>>>
>>>>> >>>>> it looks like your WAL volume is out of space which looks weird
>>>>> given
>>>>> >>>>> its capacity = 48Gb.
>>>>> >>>>>
>>>>> >>>>> Could you please share the output of the following commands:
>>>>> >>>>>
>>>>> >>>>> ceph daemon osd.N bluestore bluefs device info
>>>>> >>>>>
>>>>> >>>>> ceph daemon osd.N bluefs stats
>>>>> >>>>>
>>>>> >>>>>
>>>>> >>>>> Thanks,
>>>>> >>>>>
>>>>> >>>>> Igor
>>>>> >>>>>
>>>>> >>>>>
>>>>> >>>>> On 10/6/2021 12:24 PM, José H. Freidhof wrote:
>>>>> >>>>>> Hello together
>>>>> >>>>>>
>>>>> >>>>>> we have a running ceph pacific 16.2.5 cluster and i found this
>>>>> >> messages
>>>>> >>>>> in
>>>>> >>>>>> the service logs of the osd daemons.
>>>>> >>>>>>
>>>>> >>>>>> we have three osd nodes .. each node has 20osds as bluestore
>>>>> with
>>>>> >>>>>> nvme/ssd/hdd
>>>>> >>>>>>
>>>>> >>>>>> is this a bug or maybe i have some settings wrong?
>>>>> >>>>>>
>>>>> >>>>>>
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:25.821+0000
>>>>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>>>>> bdev
>>>>> >> 0,
>>>>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>>>>> >> 0xb40000000,
>>>>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated
>>>>> 0x0
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:29.857+0000
>>>>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>>>>> bdev
>>>>> >> 0,
>>>>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>>>>> >> 0xb40000000,
>>>>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated
>>>>> 0x0
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.073+0000
>>>>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x400000 on
>>>>> bdev
>>>>> >> 0,
>>>>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>>>>> >> 0xb40000000,
>>>>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated
>>>>> 0x0
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.405+0000
>>>>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>>>>> bdev
>>>>> >> 0,
>>>>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>>>>> >> 0xb40000000,
>>>>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated
>>>>> 0x0
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.465+0000
>>>>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>>>>> bdev
>>>>> >> 0,
>>>>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>>>>> >> 0xb40000000,
>>>>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated
>>>>> 0x0
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.529+0000
>>>>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>>>>> bdev
>>>>> >> 0,
>>>>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>>>>> >> 0xb40000000,
>>>>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated
>>>>> 0x0
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>>>>> >>>>>> 7f38eebd4700  4 rocksdb: [db_impl/db_impl_write.cc:1668] [L] New
>>>>> >>>>> memtable
>>>>> >>>>>> created with log file: #9588. Immutable memtables: 1.
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>>>>> >>>>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on
>>>>> bdev
>>>>> >> 0,
>>>>> >>>>>> allocator name bluefs-wal, allocator type hybrid, capacity
>>>>> >> 0xb40000000,
>>>>> >>>>>> block size 0x100000, free 0xff000, fragmentation 0, allocated
>>>>> 0x0
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>>>>> >> 2021/10/06-09:17:30.547575)
>>>>> >>>>>> [db_impl/db_impl_compaction_flush.cc:2198] Calling
>>>>> >>>>>> FlushMemTableToOutputFile with column family [L], flush slots
>>>>> >> available
>>>>> >>>>> 1,
>>>>> >>>>>> compaction slots available 1, flush slots scheduled 1,
>>>>> compaction
>>>>> >> slots
>>>>> >>>>>> scheduled 0
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: [flush_job.cc:321] [L] [JOB 5709]
>>>>> Flushing
>>>>> >>>>>> memtable with next log file: 9587
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: [flush_job.cc:321] [L] [JOB 5709]
>>>>> Flushing
>>>>> >>>>>> memtable with next log file: 9588
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850547916,
>>>>> >>>>>> "job": 5709, "event": "flush_started", "num_memtables": 2,
>>>>> >>>>> "num_entries":
>>>>> >>>>>> 4146, "num_deletes": 0, "total_data_size": 127203926,
>>>>> "memory_usage":
>>>>> >>>>>> 130479920, "flush_reason": "Write Buffer Full"}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: [flush_job.cc:350] [L] [JOB 5709]
>>>>> Level-0
>>>>> >> flush
>>>>> >>>>>> table #9589: started
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850559292,
>>>>> >>>>>> "cf_name": "L", "job": 5709, "event": "table_file_creation",
>>>>> >>>>> "file_number":
>>>>> >>>>>> 9589, "file_size": 3249934, "table_properties": {"data_size":
>>>>> 3247855,
>>>>> >>>>>> "index_size": 1031, "index_partitions": 0,
>>>>> "top_level_index_size": 0,
>>>>> >>>>>> "index_key_is_user_key": 0, "index_value_is_delta_encoded": 0,
>>>>> >>>>>> "filter_size": 197, "raw_key_size": 1088,
>>>>> "raw_average_key_size": 16,
>>>>> >>>>>> "raw_value_size": 3246252, "raw_average_value_size": 47739,
>>>>> >>>>>> "num_data_blocks": 36, "num_entries": 68, "num_deletions": 32,
>>>>> >>>>>> "num_merge_operands": 0, "num_range_deletions": 0,
>>>>> "format_version":
>>>>> >> 0,
>>>>> >>>>>> "fixed_key_len": 0, "filter_policy":
>>>>> "rocksdb.BuiltinBloomFilter",
>>>>> >>>>>> "column_family_name": "L", "column_family_id": 10, "comparator":
>>>>> >>>>>> "leveldb.BytewiseComparator", "merge_operator": "nullptr",
>>>>> >>>>>> "prefix_extractor_name": "nullptr", "property_collectors": "[]",
>>>>> >>>>>> "compression": "NoCompression", "compression_options":
>>>>> >> "window_bits=-14;
>>>>> >>>>>> level=32767; strategy=0; max_dict_bytes=0;
>>>>> zstd_max_train_bytes=0;
>>>>> >>>>>> enabled=0; ", "creation_time": 1633511730, "oldest_key_time":
>>>>> >>>>> 1633511730,
>>>>> >>>>>> "file_creation_time": 1633511850}}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: [flush_job.cc:401] [L] [JOB 5709]
>>>>> Level-0
>>>>> >> flush
>>>>> >>>>>> table #9589: 3249934 bytes OK
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>>>>> >> 2021/10/06-09:17:30.559362)
>>>>> >>>>>> [memtable_list.cc:447] [L] Level-0 commit table #9589 started
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>>>>> >> 2021/10/06-09:17:30.559583)
>>>>> >>>>>> [memtable_list.cc:503] [L] Level-0 commit table #9589: memtable
>>>>> #1
>>>>> >> done
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>>>>> >> 2021/10/06-09:17:30.559586)
>>>>> >>>>>> [memtable_list.cc:503] [L] Level-0 commit table #9589: memtable
>>>>> #2
>>>>> >> done
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>>>>> >> 2021/10/06-09:17:30.559601)
>>>>> >>>>>> EVENT_LOG_v1 {"time_micros": 1633511850559593, "job": 5709,
>>>>> "event":
>>>>> >>>>>> "flush_finished", "output_compression": "NoCompression",
>>>>> "lsm_state":
>>>>> >>>>> [8,
>>>>> >>>>>> 1, 0, 0, 0, 0, 0], "immutable_memtables": 0}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>>>>> >>>>>> 7f3905c02700  4 rocksdb: (Original Log Time
>>>>> >> 2021/10/06-09:17:30.559638)
>>>>> >>>>>> [db_impl/db_impl_compaction_flush.cc:205] [L] Level summary:
>>>>> files[8 1
>>>>> >>>>> 0 0
>>>>> >>>>>> 0 0 0] max score 1.00
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1676]
>>>>> [L] [JOB
>>>>> >>>>> 5710]
>>>>> >>>>>> Compacting 8@0 + 1@1 files to L1, score 1.00
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1680] [L]
>>>>> >>>>> Compaction
>>>>> >>>>>> start summary: Base version 3090 Base level 0, inputs:
>>>>> [9589(3173KB)
>>>>> >>>>>> 9586(4793KB) 9583(1876KB) 9580(194KB) 9576(6417KB) 9573(1078KB)
>>>>> >>>>> 9570(405KB)
>>>>> >>>>>> 9567(29KB)], [9564(1115KB)]
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850559956,
>>>>> >>>>>> "job": 5710, "event": "compaction_started", "compaction_reason":
>>>>> >>>>>> "LevelL0FilesNum", "files_L0": [9589, 9586, 9583, 9580, 9576,
>>>>> 9573,
>>>>> >>>>> 9570,
>>>>> >>>>>> 9567], "files_L1": [9564], "score": 1, "input_data_size":
>>>>> 19542092}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1349]
>>>>> [L] [JOB
>>>>> >>>>> 5710]
>>>>> >>>>>> Generated table #9590: 36 keys, 3249524 bytes
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850582987,
>>>>> >>>>>> "cf_name": "L", "job": 5710, "event": "table_file_creation",
>>>>> >>>>> "file_number":
>>>>> >>>>>> 9590, "file_size": 3249524, "table_properties": {"data_size":
>>>>> 3247449,
>>>>> >>>>>> "index_size": 1031, "index_partitions": 0,
>>>>> "top_level_index_size": 0,
>>>>> >>>>>> "index_key_is_user_key": 0, "index_value_is_delta_encoded": 0,
>>>>> >>>>>> "filter_size": 197, "raw_key_size": 576,
>>>>> "raw_average_key_size": 16,
>>>>> >>>>>> "raw_value_size": 3246252, "raw_average_value_size": 90173,
>>>>> >>>>>> "num_data_blocks": 36, "num_entries": 36, "num_deletions": 0,
>>>>> >>>>>> "num_merge_operands": 0, "num_range_deletions": 0,
>>>>> "format_version":
>>>>> >> 0,
>>>>> >>>>>> "fixed_key_len": 0, "filter_policy":
>>>>> "rocksdb.BuiltinBloomFilter",
>>>>> >>>>>> "column_family_name": "L", "column_family_id": 10, "comparator":
>>>>> >>>>>> "leveldb.BytewiseComparator", "merge_operator": "nullptr",
>>>>> >>>>>> "prefix_extractor_name": "nullptr", "property_collectors": "[]",
>>>>> >>>>>> "compression": "NoCompression", "compression_options":
>>>>> >> "window_bits=-14;
>>>>> >>>>>> level=32767; strategy=0; max_dict_bytes=0;
>>>>> zstd_max_train_bytes=0;
>>>>> >>>>>> enabled=0; ", "creation_time": 1633471854, "oldest_key_time": 0,
>>>>> >>>>>> "file_creation_time": 1633511850}}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1415]
>>>>> [L] [JOB
>>>>> >>>>> 5710]
>>>>> >>>>>> Compacted 8@0 + 1@1 files to L1 => 3249524 bytes
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: (Original Log Time
>>>>> >> 2021/10/06-09:17:30.583469)
>>>>> >>>>>> [compaction/compaction_job.cc:760] [L] compacted to: files[0 1
>>>>> 0 0 0 0
>>>>> >>>>> 0]
>>>>> >>>>>> max score 0.01, MB/sec: 846.1 rd, 140.7 wr, level 1, files
>>>>> in(8, 1)
>>>>> >>>>> out(1)
>>>>> >>>>>> MB in(17.5, 1.1) out(3.1), read-write-amplify(1.2)
>>>>> write-amplify(0.2)
>>>>> >>>>> OK,
>>>>> >>>>>> records in: 376, records dropped: 340 output_compression:
>>>>> >> NoCompression
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: (Original Log Time
>>>>> >> 2021/10/06-09:17:30.583498)
>>>>> >>>>>> EVENT_LOG_v1 {"time_micros": 1633511850583485, "job": 5710,
>>>>> "event":
>>>>> >>>>>> "compaction_finished", "compaction_time_micros": 23098,
>>>>> >>>>>> "compaction_time_cpu_micros": 20039, "output_level": 1,
>>>>> >>>>> "num_output_files":
>>>>> >>>>>> 1, "total_output_size": 3249524, "num_input_records": 376,
>>>>> >>>>>> "num_output_records": 36, "num_subcompactions": 1,
>>>>> >> "output_compression":
>>>>> >>>>>> "NoCompression", "num_single_delete_mismatches": 0,
>>>>> >>>>>> "num_single_delete_fallthrough": 0, "lsm_state": [0, 1, 0, 0,
>>>>> 0, 0,
>>>>> >> 0]}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850583615,
>>>>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number":
>>>>> 9589}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850583648,
>>>>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number":
>>>>> 9586}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850583675,
>>>>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number":
>>>>> 9583}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850583709,
>>>>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number":
>>>>> 9580}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850583739,
>>>>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number":
>>>>> 9576}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850583769,
>>>>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number":
>>>>> 9573}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850583804,
>>>>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number":
>>>>> 9570}
>>>>> >>>>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
>>>>> >>>>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
>>>>> >> 1633511850583835,
>>>>> >>>>>> "job": 5710, "event": "table_file_deletion", "file_number":
>>>>> 9567}
>>>>> >>>>>> _______________________________________________
>>>>> >>>>>> ceph-users mailing list -- ceph-users@xxxxxxx
>>>>> >>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>>>> >>>> --
>>>>>
>>>>
>>>>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux