Re: bluefs _allocate unable to allocate

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



hmm... so it looks like RocksDB still doesn't perform WAL cleanup during regular operation but applies it on OSD startup....

Does single OSD startup (after if's experiencing "unable to allocate) takes 20 mins as well?

Could you please share OSD log containing both that long startup and following (e.g. 1+ hour) regular operation?

Preferable for OSD.2  (or whatever one which has been using default settings from the deployment).


Thanks,

Igor


On 10/9/2021 12:18 AM, José H. Freidhof wrote:
Hi Igor,

"And was osd.2 redeployed AFTER settings had been reset to defaults ?"
A: YES

"Anything particular about current cluster use cases?"
A: we are using it temporary as a iscsi target for a vmware esxi cluster with 6 hosts. We created two 10tb iscsi images/luns for vmware, because the other datastore are at 90%. We plan in the future, after ceph is working right, and stable to install openstack and kvm and we want to convert all vms into rbd images. Like i told you is a three osd nodes cluster with 32 cores and 256gb ram and two 10g bond network cards on a 10g network

"E.g. is it a sort of regular usage (with load flukes and peak) or may be some permanently running stress load testing. The latter might tend to hold the resources and e.g. prevent from internal house keeping... A: Its a SAN for vmware and there are running 43 VMs at the moment... at the daytime is more stress on the disks because the people are working and in the afternoon the iops goes down because the users are at home
noting speculative...

There is something else that i noticed... if i reboot one osd with 20osds then it takes 20min to come up... if i tail the logs of the osd i can see a lot of " recovery log mode 2" on all osd after the 20min the osd comes one after one up and the waldb are small and no error in the logs about bluefs _allocate unable to allocate...

it seems that the problem is rocking up after a longer time (12h)


Am Fr., 8. Okt. 2021 um 15:24 Uhr schrieb Igor Fedotov <igor.fedotov@xxxxxxxx <mailto:igor.fedotov@xxxxxxxx>>:

    And was osd.2 redeployed AFTER settings had been reset to defaults ?

    Anything particular about current cluster use cases?

    E.g. is it a sort of regular usage (with load flukes and peak) or
    may be some permanently running stress load testing. The latter
    might tend to hold the resources and e.g. prevent from internal
    house keeping...

    Igor


    On 10/8/2021 12:16 AM, José H. Freidhof wrote:

    Hi Igor,

    yes the same problem is on osd.2

    we have 3 OSD Nodes... Each Node has 20 Bluestore OSDs ... in
    total we have 60 OSDs
    i checked right now one node... and 15 of 20 OSDs have this
    problem and error in the log.

    the settings that you have complained some emails ago .. i have
    reverted them to default.

    ceph.conf file:

    [global]
            fsid = 462c44b4-eed6-11eb-8b2c-a1ad45f88a97
            mon_host = [v2:10.50.50.21:3300/0,v1:10.50.50.21:6789/0
    <http://10.50.50.21:3300/0,v1:10.50.50.21:6789/0>]
    [v2:10.50.50.22:3300/0,v1:10.50.50.22:6789/0
    <http://10.50.50.22:3300/0,v1:10.50.50.22:6789/0>]
    [v2:10.50.50.20:3300/0,v1:10.50.50.20:6789/0
    <http://10.50.50.20:3300/0,v1:10.50.50.20:6789/0>]
            log file = /var/log/ceph/$cluster-$type-$id.log
            max open files = 131072
            mon compact on trim = False
            osd deep scrub interval = 137438953472
            osd max scrubs = 16
            osd objectstore = bluestore
            osd op threads = 2
            osd scrub load threshold = 0.01
            osd scrub max interval = 137438953472
            osd scrub min interval = 137438953472
            perf = True
            rbd readahead disable after bytes = 0
            rbd readahead max bytes = 4194304
            throttler perf counter = False

    [client]
            rbd cache = False


    [mon]
            mon health preluminous compat = True
            mon osd down out interval = 300

    [osd]
            bluestore cache autotune = 0
            bluestore cache kv ratio = 0.2
            bluestore cache meta ratio = 0.8
            bluestore extent map shard max size = 200
            bluestore extent map shard min size = 50
            bluestore extent map shard target size = 100
            bluestore rocksdb options =
    compression=kNoCompression,max_write_buffer_number=32,min_write_buffer_number_to_merge=2,recycle_log_file_num=32,compaction_style=kCompactionStyleLevel,write_buffer_size=67108864,target_file_size_base=67108864,max_background_compactions=31,level0_file_num_compaction_trigger=8,level0_slowdown_writes_trigger=32,level0_stop_writes_trigger=64,max_bytes_for_level_base=536870912,compaction_threads=32,max_bytes_for_level_multiplier=8,flusher_threads=8,compaction_readahead_size=2MB
            osd map share max epochs = 100
            osd max backfills = 5
            osd op num shards = 8
            osd op num threads per shard = 2
            osd min pg log entries = 10
            osd max pg log entries = 10
            osd pg log dups tracked = 10
            osd pg log trim min = 10



    root@cd133-ceph-osdh-01:~# ceph config dump
    WHO                                               MASK          
               LEVEL     OPTION                       VALUE          
                   RO
    global                      basic     container_image
    docker.io/ceph/ceph@sha256:829ebf54704f2d827de00913b171e5da741aad9b53c1f35ad59251524790eceb
    <http://docker.io/ceph/ceph@sha256:829ebf54704f2d827de00913b171e5da741aad9b53c1f35ad59251524790eceb>
                       *
    global                      advanced  leveldb_max_open_files    
                      131072
    global                      advanced  mon_compact_on_trim        
                 false
    global                      dev       ms_crc_data                
         false
    global                      advanced  osd_deep_scrub_interval    
                     1209600.000000
    global                      advanced  osd_max_scrubs            
              16
    global                      advanced  osd_scrub_load_threshold  
                      0.010000
    global                      advanced  osd_scrub_max_interval    
                      1209600.000000
    global                      advanced  osd_scrub_min_interval    
                      86400.000000
    global                      advanced  perf                      
    true
    global                      advanced
     rbd_readahead_disable_after_bytes            0
    global                      advanced  rbd_readahead_max_bytes    
                     4194304
    global                      advanced  throttler_perf_counter    
                      false
      mon                       advanced
     auth_allow_insecure_global_id_reclaim        false
      mon                       advanced  cluster_network
    10.50.50.0/24 <http://10.50.50.0/24>                            
                   *
      mon                       advanced  mon_osd_down_out_interval  
                     300
      mon                       advanced  public_network
    10.50.50.0/24 <http://10.50.50.0/24>                            
                   *
      mgr                       advanced  mgr/cephadm/container_init
                      True                                          
          *
      mgr                       advanced
     mgr/cephadm/device_enhanced_scan             true              
                                      *
      mgr                       advanced
     mgr/cephadm/migration_current                2                  
                                   *
      mgr                       advanced
     mgr/cephadm/warn_on_stray_daemons            false              
                                       *
      mgr                       advanced
     mgr/cephadm/warn_on_stray_hosts              false              
                                       *
      mgr                       advanced
     mgr/dashboard/10.50.50.21/server_addr
    <http://10.50.50.21/server_addr>           *
                                          *
      mgr                       advanced
     mgr/dashboard/camdatadash/ssl_server_port    8443              
                                      *
      mgr                       advanced
     mgr/dashboard/cd133-ceph-mon-01/server_addr                    
                                *
      mgr                       advanced
     mgr/dashboard/dasboard/server_port           80                
                                    *
      mgr                       advanced
     mgr/dashboard/dashboard/server_addr  10.251.133.161     *
      mgr                       advanced
     mgr/dashboard/dashboard/ssl_server_port      8443              
                                      *
      mgr                       advanced  mgr/dashboard/server_addr  
                     0.0.0.0                                        
             *
      mgr                       advanced  mgr/dashboard/server_port  
                     8080                                            
        *
      mgr                       advanced  mgr/dashboard/ssl          
                 false                          *
      mgr                       advanced
     mgr/dashboard/ssl_server_port                8443              
                                      *
      mgr                       advanced
     mgr/orchestrator/orchestrator                cephadm
      mgr                       advanced  mgr/prometheus/server_addr
                      0.0.0.0                                        
             *
      mgr                       advanced  mgr/telemetry/channel_ident
                     true                                            
        *
      mgr                       advanced  mgr/telemetry/enabled      
                     true                           *
      mgr                       advanced
     mgr/telemetry/last_opt_revision              3                  
                                   *
      osd                       dev bluestore_cache_autotune        
                true
      osd                       dev bluestore_cache_kv_ratio        
                0.200000
      osd                       dev bluestore_cache_meta_ratio      
                0.800000
      osd                       dev       bluestore_cache_size      
                    2147483648
      osd                       dev bluestore_cache_size_hdd 2147483648
      osd                       dev
    bluestore_extent_map_shard_max_size          200
      osd                       dev
    bluestore_extent_map_shard_min_size          50
      osd                       dev
    bluestore_extent_map_shard_target_size       100
      osd                       advanced  bluestore_rocksdb_options
     compression=kNoCompression,max_write_buffer_number=64,min_write_buffer_number_to_merge=32,recycle_log_file_num=64,compaction_style=kCompactionStyleLevel,write_buffer_size=4MB,target_file_size_base=4MB,max_background_compactions=64,level0_file_num_compaction_trigger=64,level0_slowdown_writes_trigger=128,level0_stop_writes_trigger=256,max_bytes_for_level_base=6GB,compaction_threads=32,flusher_threads=8,compaction_readahead_size=2MB
     *
      osd                       advanced  mon_osd_cache_size        
                  1024
      osd                       dev       ms_crc_data                
           false
      osd                       advanced  osd_map_share_max_epochs  
                      5
      osd                       advanced  osd_max_backfills          
                 1
      osd                       dev       osd_max_pg_log_entries    
                      10
      osd                       dev       osd_memory_cache_min      
                    3000000000
      osd host:cd133-ceph-osdh-01   basic     osd_memory_target      
                         5797322096
      osd host:cd133k-ceph-osdh-01  basic     osd_memory_target      
                         9402402385
      osd host:cd88-ceph-osdh-01    basic     osd_memory_target      
                         5797322096
      osd                       advanced  osd_memory_target_autotune
                      true
      osd                       dev       osd_min_pg_log_entries    
                      10
      osd                       advanced  osd_op_num_shards          
                 8                          *
      osd                       advanced
     osd_op_num_threads_per_shard                 2                  
                                   *
      osd                       dev osd_pg_log_dups_tracked          
               10
      osd                       dev       osd_pg_log_trim_min        
                   10
      osd                       advanced  osd_recovery_max_active    
                     3
      osd                       advanced
     osd_recovery_max_single_start                1
      osd                       advanced  osd_recovery_sleep        
                  0.000000
      client                      advanced  rbd_cache                
         false

    Am Do., 7. Okt. 2021 um 19:27 Uhr schrieb Igor Fedotov
    <igor.fedotov@xxxxxxxx <mailto:igor.fedotov@xxxxxxxx>>:

        And does redeployed osd.2 expose the same issue (or at least
        DB/WAL disbalance) again? Were settings reverted to defaults
        for it as well?


        Thanks

        Igor

        On 10/7/2021 12:46 PM, José H. Freidhof wrote:
        Good morning,

        i checked today the osd.8 and the log shows again the same error
        bluefs _allocate unable to allocate 0x100000 on bdev 0,
        allocator name bluefs-wal, allocator type hybrid, capacity
        0xb40000000, block size 0x100000, free 0xff000,
        fragmentation 0, allocated 0x0

        any idea why that could be?

        Am Mi., 6. Okt. 2021 um 22:23 Uhr schrieb José H. Freidhof
        <harald.freidhof@xxxxxxxxxxxxxx
        <mailto:harald.freidhof@xxxxxxxxxxxxxx>>:

            Hi Igor,

            today i repaired one osd node and all osd´s on the node,
            creating them new again....
            after that i waited for the rebalance/recovery process
            and the cluster was healthy after some hours..

            i notices that the osd.2 does not have any more this
            error in the log.
            but i noticed it now on the same node on osd.8... so i
            did the test that you suggested on osd.8

            it took nearly 20minutes to compact those db´s on the
            bluestore, but it helped... the problem on osd.8 is gone...

            *so the problem that i have with the alloc on the wal
            device seems to be random on different nodes and osd´s
            and looks like it comes, stay a while and disappears
            after a longer while...
            *

            here are the results that you suggested:

            root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluestore
            bluefs device info
            {
                "dev": {
                    "device": "BDEV_WAL",
                    "total": 48318377984,
            "free": 1044480,
                    "bluefs_used": 48317333504
                },
                "dev": {
                    "device": "BDEV_DB",
                    "total": 187904811008,
                    "free": 79842762752,
                    "bluefs_used": 108062048256
                },
                "dev": {
                    "device": "BDEV_SLOW",
                    "total": 6001172414464,
                    "free": 5510727389184,
                    "bluefs_used": 0,
                    "bluefs max available": 5508815847424
                }
            }
            root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluefs stats
            0 : device size 0xb3ffff000 : using 0xb3ff00000(45 GiB)
            1 : device size 0x2bbfffe000 : using 0x1931500000(101 GiB)
            2 : device size 0x57541c00000 : using 0x7235e3e000(457 GiB)
            RocksDBBlueFSVolumeSelector: wal_total:45902462976,
            db_total:178509578240, slow_total:5701113793740,
            db_avail:103884521472
            Usage matrix:
            DEV/LEV     WAL         DB  SLOW        *           *
            REAL        FILES
            LOG         304 MiB     7.9 GiB     0 B         0 B    
                0 B         9.7 MiB     1
            WAL         45 GiB      100 GiB 0 B         0 B        
            0 B 144 GiB     2319
            DB          0 B         276 MiB     0 B         0 B    
                0 B         249 MiB     47
            SLOW        0 B         0 B         0 B         0 B    
                0 B         0 B         0
            TOTALS      45 GiB      109 GiB     0 B         0 B    
                0 B         0 B         2367
            MAXIMUMS:
            LOG         304 MiB     7.9 GiB     0 B         0 B    
                0 B         20 MiB
            WAL         45 GiB      149 GiB     0 B         0 B    
                0 B         192 GiB
            DB          0 B         762 MiB     0 B         0 B    
                0 B         738 MiB
            SLOW        0 B         0 B         0 B         0 B    
                0 B         0 B
            TOTALS      45 GiB      150 GiB     0 B         0 B    
                0 B         0 B

            ---

            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.464+0000 7f4a9483a700  1 bluefs
            _allocate unable to allocate 0x400000 on bdev 0,
            allocator name bluefs-wal, allocator type hybrid,
            capacity 0xb40000000, block size 0x100000, free 0xff000,
            fragmentation 0, allocated 0x0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.472+0000 7f4a9483a700  1 bluefs
            _allocate unable to allocate 0x100000 on bdev 0,
            allocator name bluefs-wal, allocator type hybrid,
            capacity 0xb40000000, block size 0x100000, free 0xff000,
            fragmentation 0, allocated 0x0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.480+0000 7f4a9483a700  1 bluefs
            _allocate unable to allocate 0x100000 on bdev 0,
            allocator name bluefs-wal, allocator type hybrid,
            capacity 0xb40000000, block size 0x100000, free 0xff000,
            fragmentation 0, allocated 0x0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.500+0000 7f4a9483a700  1 bluefs
            _allocate unable to allocate 0x100000 on bdev 0,
            allocator name bluefs-wal, allocator type hybrid,
            capacity 0xb40000000, block size 0x100000, free 0xff000,
            fragmentation 0, allocated 0x0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.576+0000 7f4a9483a700  1 bluefs
            _allocate unable to allocate 0x100000 on bdev 0,
            allocator name bluefs-wal, allocator type hybrid,
            capacity 0xb40000000, block size 0x100000, free 0xff000,
            fragmentation 0, allocated 0x0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.624+0000 7f4a9483a700  1 bluefs
            _allocate unable to allocate 0x100000 on bdev 0,
            allocator name bluefs-wal, allocator type hybrid,
            capacity 0xb40000000, block size 0x100000, free 0xff000,
            fragmentation 0, allocated 0x0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.636+0000 7f4a9483a700  1 bluefs
            _allocate unable to allocate 0x100000 on bdev 0,
            allocator name bluefs-wal, allocator type hybrid,
            capacity 0xb40000000, block size 0x100000, free 0xff000,
            fragmentation 0, allocated 0x0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.884+0000 7f4a9483a700  1 bluefs
            _allocate unable to allocate 0x100000 on bdev 0,
            allocator name bluefs-wal, allocator type hybrid,
            capacity 0xb40000000, block size 0x100000, free 0xff000,
            fragmentation 0, allocated 0x0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.968+0000 7f4a9483a700  1 bluefs
            _allocate unable to allocate 0x100000 on bdev 0,
            allocator name bluefs-wal, allocator type hybrid,
            capacity 0xb40000000, block size 0x100000, free 0xff000,
            fragmentation 0, allocated 0x0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.992+0000 7f4a9483a700  4 rocksdb:
            [db_impl/db_impl_write.cc:1668] [L] New memtable created
            with log file: #13656. Immutable memtables: 1.
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.992+0000 7f4a9483a700  1 bluefs
            _allocate unable to allocate 0x100000 on bdev 0,
            allocator name bluefs-wal, allocator type hybrid,
            capacity 0xb40000000, block size 0x100000, free 0xff000,
            fragmentation 0, allocated 0x0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.992+0000 7f4a9483a700  1 bluefs
            _allocate unable to allocate 0x100000 on bdev 0,
            allocator name bluefs-wal, allocator type hybrid,
            capacity 0xb40000000, block size 0x100000, free 0xff000,
            fragmentation 0, allocated 0x0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb:
            (Original Log Time 2021/10/06-19:51:34.996331)
            [db_impl/db_impl_compaction_flush.cc:2198] Calling
            FlushMemTableToOutputFile with column family [L], flush
            slots available 1, compaction slots available 1, flush
            slots scheduled 1, compaction slots scheduled 0
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb:
            [flush_job.cc:321] [L] [JOB 8859] Flushing memtable with
            next log file: 13655
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb:
            [flush_job.cc:321] [L] [JOB 8859] Flushing memtable with
            next log file: 13656
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb:
            EVENT_LOG_v1 {"time_micros": 1633549894998273, "job":
            8859, "event": "flush_started", "num_memtables": 2,
            "num_entries": 3662, "num_deletes": 0,
            "total_data_size": 130482337, "memory_usage": 132976224,
            "flush_reason": "Write Buffer Full"}
            Oct 06 21:51:34 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:34.996+0000 7f4aab067700  4 rocksdb:
            [flush_job.cc:350] [L] [JOB 8859] Level-0 flush table
            #13657: started
            Oct 06 21:51:35 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:35.004+0000 7f4aab067700  4 rocksdb:
            EVENT_LOG_v1 {"time_micros": 1633549895008271,
            "cf_name": "L", "job": 8859, "event":
            "table_file_creation", "file_number": 13657,
            "file_size": 2952537, "table_properties": {"data_size":
            2951222, "index_size": 267, "index_partitions": 0,
            "top_level_index_size": 0, "index_key_is_user_key": 0,
            "index_value_is_delta_encoded": 0, "filter_size": 197,
            "raw_key_size": 1120, "raw_average_key_size": 16,
            "raw_value_size": 2950151, "raw_average_value_size":
            42145, "num_data_blocks": 9, "num_entries": 70,
            "num_deletions": 61, "num_merge_operands": 0,
            "num_range_deletions": 0, "format_version": 0,
            "fixed_key_len": 0, "filter_policy":
            "rocksdb.BuiltinBloomFilter", "column_family_name": "L",
            "column_family_id": 10, "comparator":
            "leveldb.BytewiseComparator", "merge_operator":
            "nullptr", "prefix_extractor_name": "nullptr",
            "property_collectors": "[]", "compression":
            "NoCompression", "compression_options":
            "window_bits=-14; level=32767; strategy=0;
            max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; ",
            "creation_time": 1633549854, "oldest_key_time":
            1633549854, "file_creation_time": 1633549894}}
            Oct 06 21:51:35 cd88-ceph-osdh-01 bash[6328]: debug
            2021-10-06T19:51:35.004+0000 7f4aab067700  4 rocksdb:
            [flush_job.cc:401] [L] [JOB 8859] Level-0 flush table
            #13657: 2952537 bytes OK

            ---

            root@cd88-ceph-osdh-01:~# ceph osd set noout
            root@cd88-ceph-osdh-01:~# ceph orch daemon stop osd.8
            root@cd88-ceph-osdh-01:~# ceph orch ps
            ...
            osd.7  cd133-ceph-osdh-01  running (4h)     44s ago    -
             2738M    5528M  16.2.5 6933c2a0b7dd  8a98ae61f0eb
            osd.8  cd88-ceph-osdh-01 stopped           5s ago    -
             -    5528M  <unknown>  <unknown>     <unknown>
            osd.9  cd133k-ceph-osdh-01 running (3d)      5m ago    -
               4673M    8966M  16.2.5     6933c2a0b7dd  0ff7584b1808
            ...

            ---

            root@cd88-ceph-osdh-01:~# ceph-kvstore-tool bluestore-kv
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/
            compact
            2021-10-06T21:53:50.559+0200 7f87bde3c240  0
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _open_db_and_around read-only:0 repair:0
            2021-10-06T21:53:50.559+0200 7f87bde3c240  1
            bdev(0x5644f056c800
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
            open path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1
            bdev(0x5644f056c800
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
            open size 6001172414464 (0x57541c00000, 5.5 TiB)
            block_size 4096 (4 KiB) rotational discard not supported
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _set_cache_sizes cache_size 1073741824 meta 0.45 kv 0.45
            data 0.06
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1
            bdev(0x5644f056cc00
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
            open path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1
            bdev(0x5644f056cc00
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
            open size 187904819200 (0x2bc0000000, 175 GiB)
            block_size 4096 (4 KiB) rotational discard not supported
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs
            add_block_device bdev 1 path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
            size 175 GiB
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1
            bdev(0x5644f056d000
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
            open path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1
            bdev(0x5644f056d000
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
            open size 6001172414464 (0x57541c00000, 5.5 TiB)
            block_size 4096 (4 KiB) rotational discard not supported
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs
            add_block_device bdev 2 path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
            size 5.5 TiB
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1
            bdev(0x5644f056d400
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
            open path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1
            bdev(0x5644f056d400
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
            open size 48318382080 (0xb40000000, 45 GiB) block_size
            4096 (4 KiB) non-rotational discard supported
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs
            add_block_device bdev 0 path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
            size 45 GiB
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs mount
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs
            _init_alloc new, id 0, allocator name bluefs-wal,
            allocator type hybrid, capacity 0xb40000000, block size
            0x100000
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs
            _init_alloc new, id 1, allocator name bluefs-db,
            allocator type hybrid, capacity 0x2bc0000000, block size
            0x100000
            2021-10-06T21:53:50.563+0200 7f87bde3c240  1 bluefs
            _init_alloc shared, id 2, capacity 0x57541c00000, block
            size 0x10000
            2021-10-06T21:53:50.655+0200 7f87bde3c240  1 bluefs
            mount shared_bdev_used = 0
            2021-10-06T21:53:50.655+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _prepare_db_environment set db_paths to db,178509578240
            db.slow,5701113793740
            2021-10-06T22:01:32.715+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _open_db opened rocksdb path db options
            compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2,max_total_wal_size=1073741824
            2021-10-06T22:01:32.715+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _open_super_meta old nid_max 167450
            2021-10-06T22:01:32.715+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _open_super_meta old blobid_max 30720
            2021-10-06T22:01:32.715+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _open_super_meta freelist_type bitmap
            2021-10-06T22:01:32.715+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _open_super_meta ondisk_format 4 compat_ondisk_format 3
            2021-10-06T22:01:32.715+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _open_super_meta min_alloc_size 0x1000
            2021-10-06T22:01:33.347+0200 7f87bde3c240  1 freelist init
            2021-10-06T22:01:33.347+0200 7f87bde3c240  1 freelist
            _read_cfg
            2021-10-06T22:01:33.347+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _init_alloc opening allocation metadata
            2021-10-06T22:01:41.031+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _init_alloc loaded 5.0 TiB in 37191 extents, allocator
            type hybrid, capacity 0x57541c00000, block size 0x1000,
            free 0x502f8f9a000, fragmentation 2.76445e-05
            2021-10-06T22:01:41.039+0200 7f87bde3c240  1 bluefs umount
            2021-10-06T22:01:41.043+0200 7f87bde3c240  1
            bdev(0x5644f056d400
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
            close
            2021-10-06T22:01:43.623+0200 7f87bde3c240  1
            bdev(0x5644f056cc00
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
            close
            2021-10-06T22:01:54.727+0200 7f87bde3c240  1
            bdev(0x5644f056d000
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
            close
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1
            bdev(0x5644f056d000
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
            open path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1
            bdev(0x5644f056d000
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
            open size 187904819200 (0x2bc0000000, 175 GiB)
            block_size 4096 (4 KiB) rotational discard not supported
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs
            add_block_device bdev 1 path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
            size 175 GiB
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1
            bdev(0x5644f056cc00
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
            open path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1
            bdev(0x5644f056cc00
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
            open size 6001172414464 (0x57541c00000, 5.5 TiB)
            block_size 4096 (4 KiB) rotational discard not supported
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs
            add_block_device bdev 2 path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
            size 5.5 TiB
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1
            bdev(0x5644f056d400
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
            open path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1
            bdev(0x5644f056d400
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
            open size 48318382080 (0xb40000000, 45 GiB) block_size
            4096 (4 KiB) non-rotational discard supported
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs
            add_block_device bdev 0 path
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
            size 45 GiB
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs mount
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs
            _init_alloc new, id 0, allocator name bluefs-wal,
            allocator type hybrid, capacity 0xb40000000, block size
            0x100000
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs
            _init_alloc new, id 1, allocator name bluefs-db,
            allocator type hybrid, capacity 0x2bc0000000, block size
            0x100000
            2021-10-06T22:01:54.995+0200 7f87bde3c240  1 bluefs
            _init_alloc shared, id 2, capacity 0x57541c00000, block
            size 0x10000
            2021-10-06T22:01:55.079+0200 7f87bde3c240  1 bluefs
            mount shared_bdev_used = 0
            2021-10-06T22:01:55.079+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _prepare_db_environment set db_paths to db,178509578240
            db.slow,5701113793740
            2021-10-06T22:09:36.519+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            _open_db opened rocksdb path db options
            compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2,max_total_wal_size=1073741824
            2021-10-06T22:09:54.067+0200 7f87bde3c240  1
            bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
            umount
            2021-10-06T22:09:54.079+0200 7f87bde3c240  1 bluefs umount
            2021-10-06T22:09:54.079+0200 7f87bde3c240  1
            bdev(0x5644f056d400
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
            close
            2021-10-06T22:09:56.612+0200 7f87bde3c240  1
            bdev(0x5644f056d000
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
            close
            2021-10-06T22:10:07.520+0200 7f87bde3c240  1
            bdev(0x5644f056cc00
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
            close
            2021-10-06T22:10:07.688+0200 7f87bde3c240  1 freelist
            shutdown
            2021-10-06T22:10:07.692+0200 7f87bde3c240  1
            bdev(0x5644f056c800
            /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
            close

            ---

            root@cd88-ceph-osdh-01:~# ceph orch daemon start osd.8

            ---

            root@cd88-ceph-osdh-01:/# ceph -s
              cluster:
                id: 462c44b4-eed6-11eb-8b2c-a1ad45f88a97
                health: HEALTH_OK

              services:
                mon:         3 daemons, quorum
            cd133-ceph-mon-01,cd88-ceph-mon-01,cd133k-ceph-mon-01
            (age 15h)
                mgr: cd133-ceph-mon-01.mzapob(active, since 15h),
            standbys: cd133k-ceph-mon-01.imikwh
                osd:         60 osds: 60 up (since 2m), 60 in (since 3h)
                rgw:         4 daemons active (2 hosts, 1 zones)
                tcmu-runner: 10 portals active (2 hosts)

              data:
                pools:   6 pools, 361 pgs
                objects: 2.46M objects, 8.0 TiB
                usage:   33 TiB used, 304 TiB / 338 TiB avail
                pgs:     361 active+clean

              io:
                client:   45 MiB/s rd, 50 MiB/s wr, 921 op/s rd, 674
            op/s wr



            ---

            root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluestore
            bluefs device info
            {
                "dev": {
                    "device": "BDEV_WAL",
                    "total": 48318377984,
            *"free": 41354784768,
            *        "bluefs_used": 6963593216
                },
                "dev": {
                    "device": "BDEV_DB",
                    "total": 187904811008,
                    "free": 187302928384,
                    "bluefs_used": 601882624
                },
                "dev": {
                    "device": "BDEV_SLOW",
                    "total": 6001172414464,
                    "free": 5507531620352,
                    "bluefs_used": 0,
                    "bluefs max available": 5505566572544
                }
            }

            ---

            root@cd88-ceph-osdh-01:/# ceph daemon osd.8 bluefs stats
            0 : device size 0xb3ffff000 : using 0x1a0c00000(6.5 GiB)
            1 : device size 0x2bbfffe000 : using 0x23e00000(574 MiB)
            2 : device size 0x57541c00000 : using 0x72f0803000(460 GiB)
            RocksDBBlueFSVolumeSelector: wal_total:45902462976,
            db_total:178509578240, slow_total:5701113793740,
            db_avail:103884521472
            Usage matrix:
            DEV/LEV     WAL         DB  SLOW        *           *
            REAL        FILES
            LOG         12 MiB      18 MiB      0 B         0 B    
                0 B         10 MiB      0
            WAL         6.5 GiB     0 B 0 B         0 B         0 B
            6.4 GiB 102
            DB          0 B         573 MiB     0 B         0 B    
                0 B         557 MiB     22
            SLOW        0 B         0 B         0 B         0 B    
                0 B         0 B         0
            TOTALS      6.5 GiB     591 MiB     0 B         0 B    
                0 B         0 B         125
            MAXIMUMS:
            LOG         12 MiB      18 MiB      0 B         0 B    
                0 B         17 MiB
            WAL         45 GiB      101 GiB 0 B         0 B        
            0 B 145 GiB
            DB          0 B         688 MiB     0 B         0 B    
                0 B         670 MiB
            SLOW        0 B         0 B         0 B         0 B    
                0 B         0 B
            TOTALS      45 GiB      101 GiB     0 B         0 B    
                0 B         0 B

            ----


            Here are the osd.2... the problem disapeared from alone
            very strange...

            root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluefs stats
            0 : device size 0xb3ffff000 : using 0x7bcc00000(31 GiB)
            1 : device size 0x2bbfffe000 : using 0x458c00000(17 GiB)
            2 : device size 0x57541c00000 : using 0x5cd3665000(371 GiB)
            RocksDBBlueFSVolumeSelector: wal_total:45902462976,
            db_total:178509578240, slow_total:5701113793740,
            db_avail:103884521472
            Usage matrix:
            DEV/LEV     WAL         DB  SLOW        *           *
            REAL        FILES
            LOG         920 MiB     4.0 GiB     0 B         0 B    
                0 B         10 MiB      1
            WAL         31 GiB      17 GiB      0 B         0 B    
                0 B         48 GiB      765
            DB          0 B         193 MiB     0 B         0 B    
                0 B         175 MiB     30
            SLOW        0 B         0 B         0 B         0 B    
                0 B         0 B         0
            TOTALS      32 GiB      21 GiB      0 B         0 B    
                0 B         0 B         796
            MAXIMUMS:
            LOG         920 MiB     4.0 GiB     0 B         0 B    
                0 B         17 MiB
            WAL         45 GiB      149 GiB     0 B         0 B    
                0 B         192 GiB
            DB          0 B         762 MiB     0 B         0 B    
                0 B         741 MiB
            SLOW        0 B         0 B         0 B         0 B    
                0 B         0 B
            TOTALS      45 GiB      153 GiB     0 B         0 B    
                0 B         0 B
            root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluestore
            bluefs device info
            {
                "dev": {
                    "device": "BDEV_WAL",
                    "total": 48318377984,
                    "free": 15043915776,
                    "bluefs_used": 33274462208
                },
                "dev": {
                    "device": "BDEV_DB",
                    "total": 187904811008,
                    "free": 169235963904,
                    "bluefs_used": 18668847104
                },
                "dev": {
                    "device": "BDEV_SLOW",
                    "total": 6001172414464,
                    "free": 5602453327872,
                    "bluefs_used": 0,
                    "bluefs max available": 5600865222656
                }
            }






            Am Mi., 6. Okt. 2021 um 18:11 Uhr schrieb Igor Fedotov
            <igor.fedotov@xxxxxxxx <mailto:igor.fedotov@xxxxxxxx>>:


                On 10/6/2021 4:25 PM, José H. Freidhof wrote:
                > hi,
                >
                > no risk no fun 😂 okay
                >   I have reset the settings you mentioned to standard.
                >
                > what you exactly mean with taking offline the osd?
                ceph orch daemon stop
                > osd.2? or mark down?
                "daemon stop" is enough. You  might want to set
                noout flag before that
                though...
                >
                > for the command which path i use? you mean:
                >
                > bluestore-kv /var/lib/ceph/$fsid/osd.2 compact???
                yep
                >
                >
                > Igor Fedotov <ifedotov@xxxxxxx
                <mailto:ifedotov@xxxxxxx>> schrieb am Mi., 6. Okt.
                2021, 13:33:
                >
                >> On 10/6/2021 2:16 PM, José H. Freidhof wrote:
                >>> Hi Igor,
                >>>
                >>> yes i have some osd settings set :-) here are my
                ceph config dump. those
                >>> settings are from a redhat document for
                bluestore devices
                >>> maybe it is that setting causing this problem?
                "advanced
                >>>    mon_compact_on_trim false"???
                >> OMG!!!
                >>
                >> No - mon_compact_on_trim has nothing to deal with
                bluestore.
                >>
                >> Highly likely it's bluestore_rocksdb_options
                which hurts...
                >> Documentations tend to fall behind the best
                practices.... I would
                >> strongly discourage you from using non-default
                settings unless it's
                >> absolutely clear why this is necessary.
                >>
                >> Even at the first glance the following settings
                (just a few ones I'm
                >> completely aware) are suboptimal/non-recommended:
                >>
                >> rocksdb_perf
                >>
                >> bluefs_sync_write
                >>
                >> bluefs_csum_type
                >>
                >>
                >> Not to mention bluestore_rocksdb_options which
                hasn't got much adoption
                >> so far and apparently greatly alters rocksdb
                behavior...
                >>
                >>
                >> So I would suggest to revert rocksdb options back
                to default, run the
                >> compaction and if it succeeds monitor the OSD for
                a while. Then if it
                >> works fine - apply the same for others
                >>
                >>
                >> Hope this helps,
                >>
                >> Igor
                >>
                >>
                >>
                >>> i will test it this afternoon... at the moment
                are everything semi
                >>> prodcuctive and i need to repair one osd node..
                because i think of this
                >>> reason the osds crashed on the node and the osd
                container crashes with a
                >>> dump while coming up now.
                >>> need first to replicate all between all three
                nodes and then i can take
                >>> offline the osd.2.and test your command. i will
                inform you later...
                >>>
                >>> root@cd88-ceph-osdh-01:/# ceph config dump
                >>> WHO                          MASK
                >>>    LEVEL     OPTION                            
                     VALUE
                >>>
                >>>
                >>>
                >>>
                >>>
                >>>             RO
                >>> global
                >>>    advanced leveldb_max_open_files        131072
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    advanced mon_compact_on_trim       false
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    dev       ms_crc_data                        
                        false
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    advanced osd_deep_scrub_interval      
                1209600.000000
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    advanced osd_max_scrubs        16
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    advanced osd_scrub_load_threshold        0.010000
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    advanced osd_scrub_max_interval      
                 1209600.000000
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    advanced osd_scrub_min_interval      
                 86400.000000
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    advanced  perf                              
                   true
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    advanced rbd_readahead_disable_after_bytes  
                    0
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    advanced rbd_readahead_max_bytes       4194304
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    advanced rocksdb_perf        true
                >>>
                >>>
                >>>
                >>>
                >>> global
                >>>    advanced throttler_perf_counter        false
                >>>
                >>>
                >>>
                >>>
                >>>     mon
                >>> advanced auth_allow_insecure_global_id_reclaim  
                    false
                >>>
                >>>
                >>>
                >>>
                >>>     mon
                >>> advanced cluster_network 10.50.50.0/24
                <http://10.50.50.0/24>
                >>>
                >>>
                >>>
                >>>
                >>>
                >>>             *
                >>>     mon
                >>> advanced mon_osd_down_out_interval       300
                >>>
                >>>
                >>>
                >>>
                >>>     mon
                >>> advanced  public_network 10.50.50.0/24
                <http://10.50.50.0/24>
                >>>
                >>>
                >>>
                >>>
                >>>
                >>>             *
                >>>     mgr
                >>> advanced mgr/cephadm/container_init        True
                >>>
                >>>
                >>>
                >>>
                >>>
                >>>            *
                >>>     mgr
                >>> advanced mgr/cephadm/device_enhanced_scan      
                 true
                >>>
                >>>
                >>>
                >>>
                >>>
                >>>            *
                >>>     mgr
                >>> advanced mgr/cephadm/migration_current       2
                >>>
                >>>
                >>>
                >>>
                >>>
                >>>             *
                >>>     mgr
                >>> advanced mgr/cephadm/warn_on_stray_daemons      
                false
                >>>
                >>>
                >>>
                >>>
                >>>
                >>>             *
                >>>     mgr
                >>> advanced mgr/cephadm/warn_on_stray_hosts       false
                >>>
                >>>
                >>>
                >>>
                >>>
                >>>             *
                >>>     osd
                >>> advanced bluefs_sync_write       true
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  bluestore_cache_autotune        true
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  bluestore_cache_kv_ratio        0.200000
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  bluestore_cache_meta_ratio        0.800000
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  bluestore_cache_size        2147483648
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  bluestore_cache_size_hdd        2147483648
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> advanced bluestore_csum_type       none
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  bluestore_extent_map_shard_max_size         200
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  bluestore_extent_map_shard_min_size         50
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  bluestore_extent_map_shard_target_size      100
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> advanced bluestore_rocksdb_options
                >>>
                >>
                 compression=kNoCompression,max_write_buffer_number=64,min_write_buffer_number_to_merge=32,recycle_log_file_num=64,compaction_style=kCompactionStyleLevel,write_buffer_size=4MB,target_file_size_base=4MB,max_background_compactions=64,level0_file_num_compaction_trigger=64,level0_slowdown_writes_trigger=128,level0_stop_writes_trigger=256,max_bytes_for_level_base=6GB,compaction_threads=32,flusher_threads=8,compaction_readahead_size=2MB
                >>>    *
                >>>     osd
                >>> advanced mon_osd_cache_size        1024
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev       ms_crc_data                          
                    false
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> advanced osd_map_share_max_epochs        5
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> advanced osd_max_backfills       1
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  osd_max_pg_log_entries        10
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  osd_memory_cache_min        3000000000
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>   host:cd133-ceph-osdh-01
                >>> basic  osd_memory_target         5797322383
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>   host:cd133k-ceph-osdh-01
                >>>    basic  osd_memory_target         9402402385
                >>>
                >>>
                >>>
                >>>
                >>>     osd  host:cd88-ceph-osdh-01
                >>>    basic  osd_memory_target         5797322096
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> advanced osd_memory_target_autotune        true
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  osd_min_pg_log_entries        10
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> advanced osd_op_num_shards       8
                >>>
                >>>
                >>>
                >>>
                >>>
                >>>             *
                >>>     osd
                >>> advanced osd_op_num_threads_per_shard        2
                >>>
                >>>
                >>>
                >>>
                >>>
                >>>             *
                >>>     osd
                >>> dev  osd_pg_log_dups_tracked         10
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> dev  osd_pg_log_trim_min         10
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> advanced osd_recovery_max_active       3
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> advanced osd_recovery_max_single_start       1
                >>>
                >>>
                >>>
                >>>
                >>>     osd
                >>> advanced osd_recovery_sleep        0.000000
                >>>
                >>>
                >>> Am Mi., 6. Okt. 2021 um 12:55 Uhr schrieb Igor
                Fedotov <ifedotov@xxxxxxx <mailto:ifedotov@xxxxxxx>
                >>> :
                >>>
                >>>> Jose,
                >>>>
                >>>> In fact 48GB is a way too much for WAL drive -
                usually the write ahead
                >> log
                >>>> tend to be 2-4 GBs.
                >>>>
                >>>> But in your case it's ~150GB, while DB itself
                is very small (146MB!!!):
                >>>>
                >>>> WAL         45 GiB   111 GiB     0 B         0
                B  0 B
                >>>> 154 GiB     2400
                >>>>
                >>>> DB          0 B    164 MiB     0 B         0 B  0 B
                >>>> 146 MiB     30
                >>>>
                >>>>
                >>>> which means that there are some issues with
                RocksDB's WAL processing,
                >>>> which needs some troubleshooting...
                >>>>
                >>>> Curious if other OSDs are suffering from the
                same and whether you have
                >> any
                >>>> custom settings for your OSD(s)?
                >>>>
                >>>> Additionally you might want to try the
                following command to compact this
                >>>> specific OSD manually and check if this would
                normalize the DB layout -
                >> the
                >>>> majority of data has to be at DB level not WAL.
                Please share the
                >> resulting
                >>>> layout (reported by "ceph daemon osd.2 bluefs
                stats" command) after the
                >>>> compaction is fulfiled and OSD is restarted.
                >>>>
                >>>> The compaction command to be applied on an
                offline OSD:
                >> "ceph-kvstore-tool
                >>>> bluestore-kv <path-to-osd> compact"
                >>>>
                >>>> Even if the above works great please refrain
                from applying that
                >> compaction
                >>>> to every OSD - let's see how that "compacted"
                OSD evolves.Would WAL grow
                >>>> again or not?
                >>>>
                >>>> Thanks,
                >>>>
                >>>> Igor
                >>>>
                >>>>
                >>>>
                >>>>
                >>>>
                >>>>
                >>>> On 10/6/2021 1:35 PM, José H. Freidhof wrote:
                >>>>
                >>>> Hello Igor,
                >>>>
                >>>> yes the volume is nvme wal partitions for the
                bluestore devicegroups
                >> are
                >>>> only 48gb each
                >>>>
                >>>> on each osd node are 1 nvme with 1tb splitted
                in 20 lvs with 48gb (WAL)
                >>>> on each osd node are 4 ssd with 1tb splitted in
                5 lvs with 175gb
                >> (rock.db)
                >>>> on each osd node are 20 hdd with 5.5tb with 1
                lvs (block.db)
                >>>>
                >>>> each blustore have 1 partition nvme,ssd and hdd
                like described in the
                >>>> documentation
                >>>>
                >>
                https://docs.ceph.com/en/latest/rados/configuration/bluestore-config-ref/
                <https://docs.ceph.com/en/latest/rados/configuration/bluestore-config-ref/>
                >>>> is this to small or can i adjust the max
                allocation on the wal nvme
                >> device
                >>>> in the ceph configuration?
                >>>> i know that the ssd and nvme are to small for
                those 5.5tb disk... its 1%
                >>>> only ot the rotation disk.
                >>>> i am new in ceph and still or always learning,
                but we are in a little
                >>>> hurry because our other datastores are old and
                full.
                >>>>
                >>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.2
                bluestore bluefs device
                >> info
                >>>> {
                >>>>       "dev": {
                >>>>           "device": "BDEV_WAL",
                >>>>           "total": 48318377984,
                >>>>           "free": 1044480,
                >>>>  "bluefs_used": 48317333504
                >>>>       },
                >>>>       "dev": {
                >>>>           "device": "BDEV_DB",
                >>>>           "total": 187904811008,
                >>>>           "free": 68757217280,
                >>>>  "bluefs_used": 119147593728
                >>>>       },
                >>>>       "dev": {
                >>>>           "device": "BDEV_SLOW",
                >>>>           "total": 6001172414464,
                >>>>           "free": 5624912359424,
                >>>>  "bluefs_used": 0,
                >>>>           "bluefs max available": 5624401231872
                >>>>       }
                >>>> }
                >>>> root@cd88-ceph-osdh-01:/# ceph daemon osd.2
                bluefs stats
                >>>> 0 : device size 0xb3ffff000 : using
                0xb3ff00000(45 GiB)
                >>>> 1 : device size 0x2bbfffe000 : using
                0x1bbeb00000(111 GiB)
                >>>> 2 : device size 0x57541c00000 : using
                0x579b592000(350 GiB)
                >>>> RocksDBBlueFSVolumeSelector: wal_total:45902462976,
                >> db_total:178509578240,
                >>>> slow_total:5701113793740, db_avail:103884521472
                >>>> Usage matrix:
                >>>> DEV/LEV     WAL    DB          SLOW        *  *
                >>>> REAL        FILES
                >>>> LOG         124 MiB    2.3 GiB     0 B       
                 0 B  0 B
                >>>> 7.5 MiB     1
                >>>> WAL         45 GiB   111 GiB     0 B         0
                B  0 B
                >>>> 154 GiB     2400
                >>>> DB          0 B    164 MiB     0 B         0 B  0 B
                >>>> 146 MiB     30
                >>>> SLOW        0 B    0 B         0 B         0 B  0 B
                >>   0
                >>>> B         0
                >>>> TOTALS      45 GiB   113 GiB     0 B         0
                B  0 B
                >>   0
                >>>> B         2431
                >>>> MAXIMUMS:
                >>>> LOG         124 MiB    2.3 GiB     0 B       
                 0 B  0 B
                >>   17
                >>>> MiB
                >>>> WAL         45 GiB   149 GiB     0 B         0
                B  0 B
                >>>> 192 GiB
                >>>> DB          0 B    762 MiB     0 B         0 B  0 B
                >>>> 741 MiB
                >>>> SLOW        0 B    0 B         0 B         0 B  0 B
                >>   0 B
                >>>> TOTALS      45 GiB   150 GiB     0 B         0
                B  0 B
                >>   0 B
                >>>> Am Mi., 6. Okt. 2021 um 11:45 Uhr schrieb Igor
                Fedotov <
                >> ifedotov@xxxxxxx <mailto:ifedotov@xxxxxxx>>:
                >>>>> Hey Jose,
                >>>>>
                >>>>> it looks like your WAL volume is out of space
                which looks weird given
                >>>>> its capacity = 48Gb.
                >>>>>
                >>>>> Could you please share the output of the
                following commands:
                >>>>>
                >>>>> ceph daemon osd.N bluestore bluefs device info
                >>>>>
                >>>>> ceph daemon osd.N bluefs stats
                >>>>>
                >>>>>
                >>>>> Thanks,
                >>>>>
                >>>>> Igor
                >>>>>
                >>>>>
                >>>>> On 10/6/2021 12:24 PM, José H. Freidhof wrote:
                >>>>>> Hello together
                >>>>>>
                >>>>>> we have a running ceph pacific 16.2.5 cluster
                and i found this
                >> messages
                >>>>> in
                >>>>>> the service logs of the osd daemons.
                >>>>>>
                >>>>>> we have three osd nodes .. each node has
                20osds as bluestore with
                >>>>>> nvme/ssd/hdd
                >>>>>>
                >>>>>> is this a bug or maybe i have some settings
                wrong?
                >>>>>>
                >>>>>>
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:25.821+0000
                >>>>>> 7f38eebd4700 1 bluefs _allocate unable to
                allocate 0x100000 on bdev
                >> 0,
                >>>>>> allocator name bluefs-wal, allocator type
                hybrid, capacity
                >> 0xb40000000,
                >>>>>> block size 0x100000, free 0xff000,
                fragmentation 0, allocated 0x0
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:29.857+0000
                >>>>>> 7f38eebd4700 1 bluefs _allocate unable to
                allocate 0x100000 on bdev
                >> 0,
                >>>>>> allocator name bluefs-wal, allocator type
                hybrid, capacity
                >> 0xb40000000,
                >>>>>> block size 0x100000, free 0xff000,
                fragmentation 0, allocated 0x0
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.073+0000
                >>>>>> 7f38eebd4700 1 bluefs _allocate unable to
                allocate 0x400000 on bdev
                >> 0,
                >>>>>> allocator name bluefs-wal, allocator type
                hybrid, capacity
                >> 0xb40000000,
                >>>>>> block size 0x100000, free 0xff000,
                fragmentation 0, allocated 0x0
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.405+0000
                >>>>>> 7f38eebd4700 1 bluefs _allocate unable to
                allocate 0x100000 on bdev
                >> 0,
                >>>>>> allocator name bluefs-wal, allocator type
                hybrid, capacity
                >> 0xb40000000,
                >>>>>> block size 0x100000, free 0xff000,
                fragmentation 0, allocated 0x0
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.465+0000
                >>>>>> 7f38eebd4700 1 bluefs _allocate unable to
                allocate 0x100000 on bdev
                >> 0,
                >>>>>> allocator name bluefs-wal, allocator type
                hybrid, capacity
                >> 0xb40000000,
                >>>>>> block size 0x100000, free 0xff000,
                fragmentation 0, allocated 0x0
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.529+0000
                >>>>>> 7f38eebd4700 1 bluefs _allocate unable to
                allocate 0x100000 on bdev
                >> 0,
                >>>>>> allocator name bluefs-wal, allocator type
                hybrid, capacity
                >> 0xb40000000,
                >>>>>> block size 0x100000, free 0xff000,
                fragmentation 0, allocated 0x0
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.545+0000
                >>>>>> 7f38eebd4700 4 rocksdb:
                [db_impl/db_impl_write.cc:1668] [L] New
                >>>>> memtable
                >>>>>> created with log file: #9588. Immutable
                memtables: 1.
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.545+0000
                >>>>>> 7f38eebd4700 1 bluefs _allocate unable to
                allocate 0x100000 on bdev
                >> 0,
                >>>>>> allocator name bluefs-wal, allocator type
                hybrid, capacity
                >> 0xb40000000,
                >>>>>> block size 0x100000, free 0xff000,
                fragmentation 0, allocated 0x0
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.545+0000
                >>>>>> 7f3905c02700 4 rocksdb: (Original Log Time
                >> 2021/10/06-09:17:30.547575)
                >>>>>> [db_impl/db_impl_compaction_flush.cc:2198]
                Calling
                >>>>>> FlushMemTableToOutputFile with column family
                [L], flush slots
                >> available
                >>>>> 1,
                >>>>>> compaction slots available 1, flush slots
                scheduled 1, compaction
                >> slots
                >>>>>> scheduled 0
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.545+0000
                >>>>>> 7f3905c02700 4 rocksdb: [flush_job.cc:321]
                [L] [JOB 5709] Flushing
                >>>>>> memtable with next log file: 9587
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.545+0000
                >>>>>> 7f3905c02700 4 rocksdb: [flush_job.cc:321]
                [L] [JOB 5709] Flushing
                >>>>>> memtable with next log file: 9588
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.545+0000
                >>>>>> 7f3905c02700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850547916,
                >>>>>> "job": 5709, "event": "flush_started",
                "num_memtables": 2,
                >>>>> "num_entries":
                >>>>>> 4146, "num_deletes": 0, "total_data_size":
                127203926, "memory_usage":
                >>>>>> 130479920, "flush_reason": "Write Buffer Full"}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.545+0000
                >>>>>> 7f3905c02700 4 rocksdb: [flush_job.cc:350]
                [L] [JOB 5709] Level-0
                >> flush
                >>>>>> table #9589: started
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.557+0000
                >>>>>> 7f3905c02700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850559292,
                >>>>>> "cf_name": "L", "job": 5709, "event":
                "table_file_creation",
                >>>>> "file_number":
                >>>>>> 9589, "file_size": 3249934,
                "table_properties": {"data_size": 3247855,
                >>>>>> "index_size": 1031, "index_partitions": 0,
                "top_level_index_size": 0,
                >>>>>> "index_key_is_user_key": 0,
                "index_value_is_delta_encoded": 0,
                >>>>>> "filter_size": 197, "raw_key_size": 1088,
                "raw_average_key_size": 16,
                >>>>>> "raw_value_size": 3246252,
                "raw_average_value_size": 47739,
                >>>>>> "num_data_blocks": 36, "num_entries": 68,
                "num_deletions": 32,
                >>>>>> "num_merge_operands": 0,
                "num_range_deletions": 0, "format_version":
                >> 0,
                >>>>>> "fixed_key_len": 0, "filter_policy":
                "rocksdb.BuiltinBloomFilter",
                >>>>>> "column_family_name": "L",
                "column_family_id": 10, "comparator":
                >>>>>> "leveldb.BytewiseComparator",
                "merge_operator": "nullptr",
                >>>>>> "prefix_extractor_name": "nullptr",
                "property_collectors": "[]",
                >>>>>> "compression": "NoCompression",
                "compression_options":
                >> "window_bits=-14;
                >>>>>> level=32767; strategy=0; max_dict_bytes=0;
                zstd_max_train_bytes=0;
                >>>>>> enabled=0; ", "creation_time": 1633511730,
                "oldest_key_time":
                >>>>> 1633511730,
                >>>>>> "file_creation_time": 1633511850}}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.557+0000
                >>>>>> 7f3905c02700 4 rocksdb: [flush_job.cc:401]
                [L] [JOB 5709] Level-0
                >> flush
                >>>>>> table #9589: 3249934 bytes OK
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.557+0000
                >>>>>> 7f3905c02700 4 rocksdb: (Original Log Time
                >> 2021/10/06-09:17:30.559362)
                >>>>>> [memtable_list.cc:447] [L] Level-0 commit
                table #9589 started
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.557+0000
                >>>>>> 7f3905c02700 4 rocksdb: (Original Log Time
                >> 2021/10/06-09:17:30.559583)
                >>>>>> [memtable_list.cc:503] [L] Level-0 commit
                table #9589: memtable #1
                >> done
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.557+0000
                >>>>>> 7f3905c02700 4 rocksdb: (Original Log Time
                >> 2021/10/06-09:17:30.559586)
                >>>>>> [memtable_list.cc:503] [L] Level-0 commit
                table #9589: memtable #2
                >> done
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.557+0000
                >>>>>> 7f3905c02700 4 rocksdb: (Original Log Time
                >> 2021/10/06-09:17:30.559601)
                >>>>>> EVENT_LOG_v1 {"time_micros":
                1633511850559593, "job": 5709, "event":
                >>>>>> "flush_finished", "output_compression":
                "NoCompression", "lsm_state":
                >>>>> [8,
                >>>>>> 1, 0, 0, 0, 0, 0], "immutable_memtables": 0}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.557+0000
                >>>>>> 7f3905c02700 4 rocksdb: (Original Log Time
                >> 2021/10/06-09:17:30.559638)
                >>>>>> [db_impl/db_impl_compaction_flush.cc:205] [L]
                Level summary: files[8 1
                >>>>> 0 0
                >>>>>> 0 0 0] max score 1.00
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.557+0000
                >>>>>> 7f38fb3ed700 4 rocksdb:
                [compaction/compaction_job.cc:1676] [L] [JOB
                >>>>> 5710]
                >>>>>> Compacting 8@0 + 1@1 files to L1, score 1.00
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.557+0000
                >>>>>> 7f38fb3ed700 4 rocksdb:
                [compaction/compaction_job.cc:1680] [L]
                >>>>> Compaction
                >>>>>> start summary: Base version 3090 Base level
                0, inputs: [9589(3173KB)
                >>>>>> 9586(4793KB) 9583(1876KB) 9580(194KB)
                9576(6417KB) 9573(1078KB)
                >>>>> 9570(405KB)
                >>>>>> 9567(29KB)], [9564(1115KB)]
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.557+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850559956,
                >>>>>> "job": 5710, "event": "compaction_started",
                "compaction_reason":
                >>>>>> "LevelL0FilesNum", "files_L0": [9589, 9586,
                9583, 9580, 9576, 9573,
                >>>>> 9570,
                >>>>>> 9567], "files_L1": [9564], "score": 1,
                "input_data_size": 19542092}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb:
                [compaction/compaction_job.cc:1349] [L] [JOB
                >>>>> 5710]
                >>>>>> Generated table #9590: 36 keys, 3249524 bytes
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850582987,
                >>>>>> "cf_name": "L", "job": 5710, "event":
                "table_file_creation",
                >>>>> "file_number":
                >>>>>> 9590, "file_size": 3249524,
                "table_properties": {"data_size": 3247449,
                >>>>>> "index_size": 1031, "index_partitions": 0,
                "top_level_index_size": 0,
                >>>>>> "index_key_is_user_key": 0,
                "index_value_is_delta_encoded": 0,
                >>>>>> "filter_size": 197, "raw_key_size": 576,
                "raw_average_key_size": 16,
                >>>>>> "raw_value_size": 3246252,
                "raw_average_value_size": 90173,
                >>>>>> "num_data_blocks": 36, "num_entries": 36,
                "num_deletions": 0,
                >>>>>> "num_merge_operands": 0,
                "num_range_deletions": 0, "format_version":
                >> 0,
                >>>>>> "fixed_key_len": 0, "filter_policy":
                "rocksdb.BuiltinBloomFilter",
                >>>>>> "column_family_name": "L",
                "column_family_id": 10, "comparator":
                >>>>>> "leveldb.BytewiseComparator",
                "merge_operator": "nullptr",
                >>>>>> "prefix_extractor_name": "nullptr",
                "property_collectors": "[]",
                >>>>>> "compression": "NoCompression",
                "compression_options":
                >> "window_bits=-14;
                >>>>>> level=32767; strategy=0; max_dict_bytes=0;
                zstd_max_train_bytes=0;
                >>>>>> enabled=0; ", "creation_time": 1633471854,
                "oldest_key_time": 0,
                >>>>>> "file_creation_time": 1633511850}}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb:
                [compaction/compaction_job.cc:1415] [L] [JOB
                >>>>> 5710]
                >>>>>> Compacted 8@0 + 1@1 files to L1 => 3249524 bytes
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: (Original Log Time
                >> 2021/10/06-09:17:30.583469)
                >>>>>> [compaction/compaction_job.cc:760] [L]
                compacted to: files[0 1 0 0 0 0
                >>>>> 0]
                >>>>>> max score 0.01, MB/sec: 846.1 rd, 140.7 wr,
                level 1, files in(8, 1)
                >>>>> out(1)
                >>>>>> MB in(17.5, 1.1) out(3.1),
                read-write-amplify(1.2) write-amplify(0.2)
                >>>>> OK,
                >>>>>> records in: 376, records dropped: 340
                output_compression:
                >> NoCompression
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: (Original Log Time
                >> 2021/10/06-09:17:30.583498)
                >>>>>> EVENT_LOG_v1 {"time_micros":
                1633511850583485, "job": 5710, "event":
                >>>>>> "compaction_finished",
                "compaction_time_micros": 23098,
                >>>>>> "compaction_time_cpu_micros": 20039,
                "output_level": 1,
                >>>>> "num_output_files":
                >>>>>> 1, "total_output_size": 3249524,
                "num_input_records": 376,
                >>>>>> "num_output_records": 36,
                "num_subcompactions": 1,
                >> "output_compression":
                >>>>>> "NoCompression",
                "num_single_delete_mismatches": 0,
                >>>>>> "num_single_delete_fallthrough": 0,
                "lsm_state": [0, 1, 0, 0, 0, 0,
                >> 0]}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850583615,
                >>>>>> "job": 5710, "event": "table_file_deletion",
                "file_number": 9589}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850583648,
                >>>>>> "job": 5710, "event": "table_file_deletion",
                "file_number": 9586}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850583675,
                >>>>>> "job": 5710, "event": "table_file_deletion",
                "file_number": 9583}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850583709,
                >>>>>> "job": 5710, "event": "table_file_deletion",
                "file_number": 9580}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850583739,
                >>>>>> "job": 5710, "event": "table_file_deletion",
                "file_number": 9576}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850583769,
                >>>>>> "job": 5710, "event": "table_file_deletion",
                "file_number": 9573}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850583804,
                >>>>>> "job": 5710, "event": "table_file_deletion",
                "file_number": 9570}
                >>>>>> cd88-ceph-osdh-01 bash[6283]: debug
                2021-10-06T09:17:30.581+0000
                >>>>>> 7f38fb3ed700 4 rocksdb: EVENT_LOG_v1
                {"time_micros":
                >> 1633511850583835,
                >>>>>> "job": 5710, "event": "table_file_deletion",
                "file_number": 9567}
                >>>>>> _______________________________________________
                >>>>>> ceph-users mailing list -- ceph-users@xxxxxxx
                <mailto:ceph-users@xxxxxxx>
                >>>>>> To unsubscribe send an email to
                ceph-users-leave@xxxxxxx
                <mailto:ceph-users-leave@xxxxxxx>
                >>>> --




--
    Mit freundlichen Grüßen,

     -

    José H. Freidhof

    Reyerhütterstrasse 130b
    41065 Mönchengladbach
    eMail: harald.freidhof@xxxxxxxxx <mailto:harald.freidhof@xxxxxxxxx>
    mobil: +49 (0) 1523 – 717 7801



--

Mit freundlichen Grüßen,

 -

José H. Freidhof

Reyerhütterstrasse 130b
41065 Mönchengladbach
eMail: harald.freidhof@xxxxxxxxx <mailto:harald.freidhof@xxxxxxxxx>
mobil: +49 (0) 1523 – 717 7801

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux