Re: bluefs _allocate unable to allocate

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Jose,

I can't say I'm a big guru in tuning RocksDB, IMO default settings should be good enough. So I wouldn't recommend none of them - apparently one should try and choose...

But I'd strongly encourage you to set defaults for now and make sure whether this fixes huge WAL issue or not... It might also be the case that all the "performance gain" from these updated settings is caused by a postponed (or even disabled) data processing - DB just doesn't sort data out properly. So let's make it working properly first!


Igor


On 10/12/2021 1:56 PM, José H. Freidhof wrote:
Hi Igor

the reason why i tested differekt rocksdb options is that was having a really bad write performance with the default settings (30-60mb/s) on the cluster...

actually i have 200mb/s read and 180mb/s write performance

now i dont now which of the booth settings are the good ones

Another question:
which of the booth can you recommend?

https://gist.github.com/likid0/1b52631ff5d0d649a22a3f30106ccea7 <https://gist.github.com/likid0/1b52631ff5d0d649a22a3f30106ccea7> bluestore rocksdb options = compression=kNoCompression,max_write_buffer_number=32,min_write_buffer_number_to_merge=2,recycle_log_file_num=32,compaction_style=kCompactionStyleLevel,write_buffer_size=67108864,target_file_size_base=67108864,max_background_compactions=31,level0_file_num_compaction_trigger=8,level0_slowdown_writes_trigger=32,level0_stop_writes_trigger=64,max_bytes_for_level_base=536870912,compaction_threads=32,max_bytes_for_level_multiplier=8,flusher_threads=8,compaction_readahead_size=2MB

https://yourcmc.ru/wiki/Ceph_performance <https://yourcmc.ru/wiki/Ceph_performance> bluestore_rocksdb_options = compression=kNoCompression,max_write_buffer_number=64,min_write_buffer_number_to_merge=32,recycle_log_file_num=64,compaction_style=kCompactionStyleLevel,write_buffer_size=4MB,target_file_size_base=4MB,max_background_compactions=64,level0_file_num_compaction_trigger=64,level0_slowdown_writes_trigger=128,level0_stop_writes_trigger=256,max_bytes_for_level_base=6GB,compaction_threads=32,flusher_threads=8,compaction_readahead_size=2MB



Am Di., 12. Okt. 2021 um 12:35 Uhr schrieb José H. Freidhof <harald.freidhof@xxxxxxxxxxxxxx <mailto:harald.freidhof@xxxxxxxxxxxxxx>>:

    Hi Igor,

    Thx for checking the logs.. but what the hell is going on here? :-)
    Yes its true i tested the and created the osd´s with three
    different rockdb options.
    I can not understand why the osd dont have the same rockdb option,
    because i have created ALL OSDs new after set and test those settings.

    Maybe i do something wrong with the re-deployment of the osds?
    What i do:
    ceph osd out osd.x
    ceph osd down osd.x
    systemctl stop ceph-osd@x
    ceph osd rm osd.x
    ceph osd crush rm osd.x
    ceph auth del osd.x
    ceph-volume lvm zap --destroy /dev/ceph-block-0/block-0 (lvm hdd
    partition)
    ceph-volume lvm zap --destroy /dev/ceph-db-0/db-0 (lvm ssd partition)
    ceph-volume lvm zap --destroy /dev/ceph-wal-0/wal-db-0 (lvm nvme
    partition)
    ...

    Later i recreate the osds with:
    cephadm shell -m /var/lib/ceph
    ceph auth export client.bootstrap-osd
    vi /var/lib/ceph/bootstrap-osd/ceph.keyring
    ceph-volume lvm prepare --no-systemd --bluestore --data
    ceph-block-4/block-4 --block.wal ceph-wal-0/waldb-4 --block.db
    ceph-db-0/db-4
    cp -r /var/lib/ceph/osd /mnt/ceph/
    Exit the shell in the container.
    cephadm --image ceph/ceph:v16.2.5 adopt --style legacy --name osd.X
    systemctl start
    ceph-462c44b4-eed6-11eb-8b2c-a1ad45f88a97@xxxxxx.service


    Igor one question:
    is there actually an easier way to recreate the osd? maybe over
    the dashboard?
    can you recommend something?

    i have no problem to create the osd on the nodes again, but i need
    to be sure that no old setting stays on the osd.



    Am Di., 12. Okt. 2021 um 12:03 Uhr schrieb Igor Fedotov
    <igor.fedotov@xxxxxxxx <mailto:igor.fedotov@xxxxxxxx>>:

        Hey Jose,

        your rocksdb settings are still different from the default ones.

        These are options you shared originally:

        compression=kNoCompression,max_write_buffer_number=64,min_write_buffer_number_to_merge=32,recycle_log_file_num=64,compaction_style=kCompactionStyleLevel,write_buffer_size=4MB,target_file_size_base=4MB,max_background_compactions=64,level0_file_num_compaction_trigger=64,level0_slowdown_writes_trigger=128,level0_stop_writes_trigger=256,max_bytes_for_level_base=6GB,compaction_threads=32,flusher_threads=8,compaction_readahead_size=2MB

        These are ones I could find in  osd.5 startup log, note e.g.
        max_write_buffer_number:

        Oct 12 09:09:30 cd88-ceph-osdh-01 bash[1572206]: debug
        2021-10-12T07:09:30.686+0000 7f16d24a0080  1
        bluestore(/var/lib/ceph/osd/ceph-5) _open_db opened rocksdb
        path db options
        compression=kNoCompression,max_write_buffer_number=32,min_write_buffer_number_to_merge=2,recycle_log_file_num=32,compaction_style=kCompactionStyleLevel,write_buffer_size=67108864,target_file_size_base=67108864,max_background_compactions=31,level0_file_num_compaction_trigger=8,level0_slowdown_writes_trigger=32,level0_stop_writes_trigger=64,max_bytes_for_level_base=536870912,compaction_threads=32,max_bytes_for_level_multiplier=8,flusher_threads=8,compaction_readahead_size=2MB

        And here are the ones I'd expect as defaults - again please
        note max_write_buffer_number:

        compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2,max_total_wal_size=1073741824


        And here is the source code for v16.2.5 where the expected
        default line comes from:

        https://github.com/ceph/ceph/blob/0883bdea7337b95e4b611c768c0279868462204a/src/common/options.cc#L4644
        <https://github.com/ceph/ceph/blob/0883bdea7337b95e4b611c768c0279868462204a/src/common/options.cc#L4644>


        Not that I'm absolutely sure this is the actual root cause but
        I'd suggest to revert back to the baseline prior to proceeding
        with the troubleshooting...

        So please adjust properly and restart OSDs!!! Hopefully it
        wouldn't need a redeployment...


        As for https://tracker.ceph.com/issues/50656
        <https://tracker.ceph.com/issues/50656> - it's irrelevant to
        your case. It was unexpected ENOSPC result from an allocator
        which still had enough free space. But in your case bluefs
        allocator doesn't have free space at all as the latter is
        totally wasted by tons of WAL files.


        Thanks,

        Igor



        On 10/12/2021 10:51 AM, José H. Freidhof wrote:
        Hello Igor

        "Does single OSD startup (after if's experiencing "unable to
        allocate) takes 20 mins as well?"
        A: YES

        Here the example log of the startup and recovery of a
        problematic osd.
        https://paste.ubuntu.com/p/2WVJbg7cBy/
        <https://paste.ubuntu.com/p/2WVJbg7cBy/>

        Here the example log of a problematic osd
        https://paste.ubuntu.com/p/qbB6y7663f/
        <https://paste.ubuntu.com/p/qbB6y7663f/>

        I found this post about a similar error and a bug in
        16.2.4... we are running 16.2.5...maybe the bug is not really
        fixed???
        https://tracker.ceph.com/issues/50656
        <https://tracker.ceph.com/issues/50656>
        https://forum.proxmox.com/threads/ceph-16-2-pacific-cluster-crash.92367/
        <https://forum.proxmox.com/threads/ceph-16-2-pacific-cluster-crash.92367/>



        Am Mo., 11. Okt. 2021 um 11:53 Uhr schrieb Igor Fedotov
        <igor.fedotov@xxxxxxxx <mailto:igor.fedotov@xxxxxxxx>>:

            hmm... so it looks like RocksDB still doesn't perform WAL
            cleanup during regular operation but applies it on OSD
            startup....

            Does single OSD startup (after if's experiencing "unable
            to allocate) takes 20 mins as well?

            Could you please share OSD log containing both that long
            startup and following (e.g. 1+ hour) regular operation?

            Preferable for OSD.2  (or whatever one which has been
            using default settings from the deployment).


            Thanks,

            Igor


            On 10/9/2021 12:18 AM, José H. Freidhof wrote:
            Hi Igor,

            "And was osd.2 redeployed AFTER settings had been reset
            to defaults ?"
            A: YES

            "Anything particular about current cluster use cases?"
            A: we are using it temporary as a iscsi target for a
            vmware esxi cluster with 6 hosts. We created two 10tb
            iscsi images/luns for vmware, because the other
            datastore are at 90%.
            We plan in the future, after ceph is working right, and
            stable to install openstack and kvm and we want to
            convert all vms into rbd images.
            Like i told you is a three osd nodes cluster with 32
            cores and 256gb ram and two 10g bond network cards on a
            10g network

            "E.g. is it a sort of regular usage (with load flukes
            and peak) or may be some permanently running stress load
            testing. The latter might tend to hold the resources and
            e.g. prevent from internal house keeping...
            A: Its a SAN for vmware and there are running 43 VMs at
            the moment... at the daytime is more stress on the disks
            because the people are working and in the afternoon the
            iops goes down because the users are at home
            noting speculative...

            There is something else that i noticed... if i reboot
            one osd with 20osds then it takes 20min to come up... if
            i tail the logs of the osd i can see a lot of " recovery
            log mode 2" on all osd
            after the 20min the osd comes one after one up and the
            waldb are small and no error in the logs about bluefs
            _allocate unable to allocate...

            it seems that the problem is rocking up after a longer
            time (12h)


            Am Fr., 8. Okt. 2021 um 15:24 Uhr schrieb Igor Fedotov
            <igor.fedotov@xxxxxxxx <mailto:igor.fedotov@xxxxxxxx>>:

                And was osd.2 redeployed AFTER settings had been
                reset to defaults ?

                Anything particular about current cluster use cases?

                E.g. is it a sort of regular usage (with load flukes
                and peak) or may be some permanently running stress
                load testing. The latter might tend to hold the
                resources and e.g. prevent from internal house
                keeping...

                Igor


                On 10/8/2021 12:16 AM, José H. Freidhof wrote:

                Hi Igor,

                yes the same problem is on osd.2

                we have 3 OSD Nodes... Each Node has 20 Bluestore
                OSDs ... in total we have 60 OSDs
                i checked right now one node... and 15 of 20 OSDs
                have this problem and error in the log.

                the settings that you have complained some emails
                ago .. i have reverted them to default.

                ceph.conf file:

                [global]
                        fsid = 462c44b4-eed6-11eb-8b2c-a1ad45f88a97
                        mon_host =
                [v2:10.50.50.21:3300/0,v1:10.50.50.21:6789/0
                <http://10.50.50.21:3300/0,v1:10.50.50.21:6789/0>]
                [v2:10.50.50.22:3300/0,v1:10.50.50.22:6789/0
                <http://10.50.50.22:3300/0,v1:10.50.50.22:6789/0>]
                [v2:10.50.50.20:3300/0,v1:10.50.50.20:6789/0
                <http://10.50.50.20:3300/0,v1:10.50.50.20:6789/0>]
                        log file = /var/log/ceph/$cluster-$type-$id.log
                        max open files = 131072
                        mon compact on trim = False
                        osd deep scrub interval = 137438953472
                        osd max scrubs = 16
                        osd objectstore = bluestore
                        osd op threads = 2
                        osd scrub load threshold = 0.01
                        osd scrub max interval = 137438953472
                        osd scrub min interval = 137438953472
                        perf = True
                        rbd readahead disable after bytes = 0
                        rbd readahead max bytes = 4194304
                        throttler perf counter = False

                [client]
                        rbd cache = False


                [mon]
                        mon health preluminous compat = True
                        mon osd down out interval = 300

                [osd]
                        bluestore cache autotune = 0
                        bluestore cache kv ratio = 0.2
                        bluestore cache meta ratio = 0.8
                        bluestore extent map shard max size = 200
                        bluestore extent map shard min size = 50
                        bluestore extent map shard target size = 100
                        bluestore rocksdb options =
                compression=kNoCompression,max_write_buffer_number=32,min_write_buffer_number_to_merge=2,recycle_log_file_num=32,compaction_style=kCompactionStyleLevel,write_buffer_size=67108864,target_file_size_base=67108864,max_background_compactions=31,level0_file_num_compaction_trigger=8,level0_slowdown_writes_trigger=32,level0_stop_writes_trigger=64,max_bytes_for_level_base=536870912,compaction_threads=32,max_bytes_for_level_multiplier=8,flusher_threads=8,compaction_readahead_size=2MB
                        osd map share max epochs = 100
                        osd max backfills = 5
                        osd op num shards = 8
                        osd op num threads per shard = 2
                        osd min pg log entries = 10
                        osd max pg log entries = 10
                        osd pg log dups tracked = 10
                        osd pg log trim min = 10



                root@cd133-ceph-osdh-01:~# ceph config dump
                WHO                     MASK                LEVEL
                OPTION               VALUE  RO
                global                basic container_image
                docker.io/ceph/ceph@sha256:829ebf54704f2d827de00913b171e5da741aad9b53c1f35ad59251524790eceb
                <http://docker.io/ceph/ceph@sha256:829ebf54704f2d827de00913b171e5da741aad9b53c1f35ad59251524790eceb>
                             *
                global                advanced
                 leveldb_max_open_files                 131072
                global                advanced  mon_compact_on_trim
                               false
                global                dev ms_crc_data              
                 false
                global                advanced
                 osd_deep_scrub_interval                1209600.000000
                global                advanced  osd_max_scrubs    
                            16
                global                advanced
                 osd_scrub_load_threshold                 0.010000
                global                advanced
                 osd_scrub_max_interval                 1209600.000000
                global                advanced
                 osd_scrub_min_interval                 86400.000000
                global                advanced  perf           true
                global                advanced
                 rbd_readahead_disable_after_bytes            0
                global                advanced
                 rbd_readahead_max_bytes                4194304
                global                advanced
                 throttler_perf_counter                 false
                  mon                 advanced
                 auth_allow_insecure_global_id_reclaim        false
                  mon                 advanced  cluster_network
                10.50.50.0/24 <http://10.50.50.0/24>  *
                  mon                 advanced
                 mon_osd_down_out_interval                300
                  mon                 advanced  public_network
                10.50.50.0/24 <http://10.50.50.0/24>  *
                  mgr                 advanced
                 mgr/cephadm/container_init                 True *
                  mgr                 advanced
                 mgr/cephadm/device_enhanced_scan             true
                                          *
                  mgr                 advanced
                 mgr/cephadm/migration_current                2    
                                         *
                  mgr                 advanced
                 mgr/cephadm/warn_on_stray_daemons            false
                                         *
                  mgr                 advanced
                 mgr/cephadm/warn_on_stray_hosts              false
                                           *
                  mgr                 advanced
                 mgr/dashboard/10.50.50.21/server_addr
                <http://10.50.50.21/server_addr>                  
                    *
                              *
                  mgr                 advanced
                 mgr/dashboard/camdatadash/ssl_server_port    8443
                                  *
                  mgr                 advanced
                 mgr/dashboard/cd133-ceph-mon-01/server_addr      
                          *
                  mgr                 advanced
                 mgr/dashboard/dasboard/server_port           80  
                                      *
                  mgr                 advanced
                 mgr/dashboard/dashboard/server_addr        
                 10.251.133.161                         *
                  mgr                 advanced
                 mgr/dashboard/dashboard/ssl_server_port      8443
                                    *
                  mgr                 advanced
                 mgr/dashboard/server_addr                0.0.0.0  
                                           *
                  mgr                 advanced
                 mgr/dashboard/server_port                8080 *
                  mgr                 advanced  mgr/dashboard/ssl  
                             false                              *
                  mgr                 advanced
                 mgr/dashboard/ssl_server_port                8443 *
                  mgr                 advanced
                 mgr/orchestrator/orchestrator                cephadm
                  mgr                 advanced
                 mgr/prometheus/server_addr                 0.0.0.0  *
                  mgr                 advanced
                 mgr/telemetry/channel_ident                true *
                  mgr                 advanced
                 mgr/telemetry/enabled                true *
                  mgr                 advanced
                 mgr/telemetry/last_opt_revision              3    
                                       *
                  osd                 dev bluestore_cache_autotune
                              true
                  osd                 dev bluestore_cache_kv_ratio
                              0.200000
                  osd                 dev
                bluestore_cache_meta_ratio               0.800000
                  osd                 dev bluestore_cache_size    
                          2147483648
                  osd                 dev bluestore_cache_size_hdd
                              2147483648
                  osd                 dev
                bluestore_extent_map_shard_max_size          200
                  osd                 dev
                bluestore_extent_map_shard_min_size          50
                  osd                 dev
                bluestore_extent_map_shard_target_size       100
                  osd                 advanced
                 bluestore_rocksdb_options
                 compression=kNoCompression,max_write_buffer_number=64,min_write_buffer_number_to_merge=32,recycle_log_file_num=64,compaction_style=kCompactionStyleLevel,write_buffer_size=4MB,target_file_size_base=4MB,max_background_compactions=64,level0_file_num_compaction_trigger=64,level0_slowdown_writes_trigger=128,level0_stop_writes_trigger=256,max_bytes_for_level_base=6GB,compaction_threads=32,flusher_threads=8,compaction_readahead_size=2MB
                 *
                  osd                 advanced  mon_osd_cache_size
                                1024
                  osd                 dev ms_crc_data              
                 false
                  osd                 advanced
                 osd_map_share_max_epochs                 5
                  osd                 advanced  osd_max_backfills  
                             1
                  osd                 dev osd_max_pg_log_entries  
                            10
                  osd                 dev osd_memory_cache_min    
                          3000000000
                  osd host:cd133-ceph-osdh-01 basic    
                osd_memory_target  5797322096
                  osd host:cd133k-ceph-osdh-01  basic    
                osd_memory_target  9402402385
                  osd host:cd88-ceph-osdh-01  basic    
                osd_memory_target  5797322096
                  osd                 advanced
                 osd_memory_target_autotune                 true
                  osd                 dev osd_min_pg_log_entries  
                            10
                  osd                 advanced  osd_op_num_shards  
                             8                              *
                  osd                 advanced
                 osd_op_num_threads_per_shard                 2  *
                  osd                 dev osd_pg_log_dups_tracked  
                             10
                  osd                 dev osd_pg_log_trim_min      
                         10
                  osd                 advanced
                 osd_recovery_max_active                3
                  osd                 advanced
                 osd_recovery_max_single_start                1
                  osd                 advanced  osd_recovery_sleep
                                0.000000
                  client                advanced  rbd_cache        
                       false

                Am Do., 7. Okt. 2021 um 19:27 Uhr schrieb Igor
                Fedotov <igor.fedotov@xxxxxxxx
                <mailto:igor.fedotov@xxxxxxxx>>:

                    And does redeployed osd.2 expose the same issue
                    (or at least DB/WAL disbalance) again? Were
                    settings reverted to defaults for it as well?


                    Thanks

                    Igor

                    On 10/7/2021 12:46 PM, José H. Freidhof wrote:
                    Good morning,

                    i checked today the osd.8 and the log shows
                    again the same error
                    bluefs _allocate unable to allocate 0x100000
                    on bdev 0, allocator name bluefs-wal,
                    allocator type hybrid, capacity 0xb40000000,
                    block size 0x100000, free 0xff000,
                    fragmentation 0, allocated 0x0

                    any idea why that could be?

                    Am Mi., 6. Okt. 2021 um 22:23 Uhr schrieb José
                    H. Freidhof <harald.freidhof@xxxxxxxxxxxxxx
                    <mailto:harald.freidhof@xxxxxxxxxxxxxx>>:

                        Hi Igor,

                        today i repaired one osd node and all
                        osd´s on the node, creating them new again....
                        after that i waited for the
                        rebalance/recovery process and the cluster
                        was healthy after some hours..

                        i notices that the osd.2 does not have any
                        more this error in the log.
                        but i noticed it now on the same node on
                        osd.8... so i did the test that you
                        suggested on osd.8

                        it took nearly 20minutes to compact those
                        db´s on the bluestore, but it helped...
                        the problem on osd.8 is gone...

                        *so the problem that i have with the alloc
                        on the wal device seems to be random on
                        different nodes and osd´s and looks like
                        it comes, stay a while and disappears
                        after a longer while...
                        *

                        here are the results that you suggested:

                        root@cd88-ceph-osdh-01:/# ceph daemon
                        osd.8 bluestore bluefs device info
                        {
                            "dev": {
                        "device": "BDEV_WAL",
                        "total": 48318377984,
                        "free": 1044480,
                        "bluefs_used": 48317333504
                            },
                            "dev": {
                        "device": "BDEV_DB",
                        "total": 187904811008,
                        "free": 79842762752,
                        "bluefs_used": 108062048256
                            },
                            "dev": {
                        "device": "BDEV_SLOW",
                        "total": 6001172414464,
                        "free": 5510727389184,
                        "bluefs_used": 0,
                        "bluefs max available": 5508815847424
                            }
                        }
                        root@cd88-ceph-osdh-01:/# ceph daemon
                        osd.8 bluefs stats
                        0 : device size 0xb3ffff000 : using
                        0xb3ff00000(45 GiB)
                        1 : device size 0x2bbfffe000 : using
                        0x1931500000(101 GiB)
                        2 : device size 0x57541c00000 : using
                        0x7235e3e000(457 GiB)
                        RocksDBBlueFSVolumeSelector:
                        wal_total:45902462976,
                        db_total:178509578240,
                        slow_total:5701113793740,
                        db_avail:103884521472
                        Usage matrix:
                        DEV/LEV WAL         DB          SLOW      
                         *     *   REAL  FILES
                        LOG         304 MiB     7.9 GiB     0 B  
                            0 B     0 B   9.7 MiB 1
                        WAL         45 GiB      100 GiB     0 B  
                            0 B     0 B   144 GiB 2319
                        DB          0 B         276 MiB     0 B  
                            0 B     0 B   249 MiB 47
                        SLOW        0 B         0 B         0 B  
                            0 B     0 B   0 B 0
                        TOTALS      45 GiB      109 GiB     0 B  
                            0 B     0 B   0 B 2367
                        MAXIMUMS:
                        LOG 304 MiB 7.9 GiB     0 B         0 B  
                              0 B       20 MiB
                        WAL         45 GiB      149 GiB     0 B  
                            0 B     0 B   192 GiB
                        DB          0 B         762 MiB     0 B  
                            0 B     0 B   738 MiB
                        SLOW        0 B         0 B         0 B  
                            0 B     0 B   0 B
                        TOTALS      45 GiB      150 GiB     0 B  
                            0 B     0 B   0 B

                        ---

                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.464+0000 7f4a9483a700
                         1 bluefs _allocate unable to allocate
                        0x400000 on bdev 0, allocator name
                        bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000,
                        free 0xff000, fragmentation 0, allocated 0x0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.472+0000 7f4a9483a700
                         1 bluefs _allocate unable to allocate
                        0x100000 on bdev 0, allocator name
                        bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000,
                        free 0xff000, fragmentation 0, allocated 0x0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.480+0000 7f4a9483a700
                         1 bluefs _allocate unable to allocate
                        0x100000 on bdev 0, allocator name
                        bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000,
                        free 0xff000, fragmentation 0, allocated 0x0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.500+0000 7f4a9483a700
                         1 bluefs _allocate unable to allocate
                        0x100000 on bdev 0, allocator name
                        bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000,
                        free 0xff000, fragmentation 0, allocated 0x0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.576+0000 7f4a9483a700
                         1 bluefs _allocate unable to allocate
                        0x100000 on bdev 0, allocator name
                        bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000,
                        free 0xff000, fragmentation 0, allocated 0x0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.624+0000 7f4a9483a700
                         1 bluefs _allocate unable to allocate
                        0x100000 on bdev 0, allocator name
                        bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000,
                        free 0xff000, fragmentation 0, allocated 0x0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.636+0000 7f4a9483a700
                         1 bluefs _allocate unable to allocate
                        0x100000 on bdev 0, allocator name
                        bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000,
                        free 0xff000, fragmentation 0, allocated 0x0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.884+0000 7f4a9483a700
                         1 bluefs _allocate unable to allocate
                        0x100000 on bdev 0, allocator name
                        bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000,
                        free 0xff000, fragmentation 0, allocated 0x0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.968+0000 7f4a9483a700
                         1 bluefs _allocate unable to allocate
                        0x100000 on bdev 0, allocator name
                        bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000,
                        free 0xff000, fragmentation 0, allocated 0x0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.992+0000 7f4a9483a700
                         4 rocksdb:
                        [db_impl/db_impl_write.cc:1668] [L] New
                        memtable created with log file: #13656.
                        Immutable memtables: 1.
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.992+0000 7f4a9483a700
                         1 bluefs _allocate unable to allocate
                        0x100000 on bdev 0, allocator name
                        bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000,
                        free 0xff000, fragmentation 0, allocated 0x0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.992+0000 7f4a9483a700
                         1 bluefs _allocate unable to allocate
                        0x100000 on bdev 0, allocator name
                        bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000,
                        free 0xff000, fragmentation 0, allocated 0x0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.996+0000 7f4aab067700
                         4 rocksdb: (Original Log Time
                        2021/10/06-19:51:34.996331)
                        [db_impl/db_impl_compaction_flush.cc:2198]
                        Calling FlushMemTableToOutputFile with
                        column family [L], flush slots available
                        1, compaction slots available 1, flush
                        slots scheduled 1, compaction slots
                        scheduled 0
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.996+0000 7f4aab067700
                         4 rocksdb: [flush_job.cc:321] [L] [JOB
                        8859] Flushing memtable with next log
                        file: 13655
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.996+0000 7f4aab067700
                         4 rocksdb: [flush_job.cc:321] [L] [JOB
                        8859] Flushing memtable with next log
                        file: 13656
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.996+0000 7f4aab067700
                         4 rocksdb: EVENT_LOG_v1 {"time_micros":
                        1633549894998273, "job": 8859, "event":
                        "flush_started", "num_memtables": 2,
                        "num_entries": 3662, "num_deletes": 0,
                        "total_data_size": 130482337,
                        "memory_usage": 132976224, "flush_reason":
                        "Write Buffer Full"}
                        Oct 06 21:51:34 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:34.996+0000 7f4aab067700
                         4 rocksdb: [flush_job.cc:350] [L] [JOB
                        8859] Level-0 flush table #13657: started
                        Oct 06 21:51:35 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:35.004+0000 7f4aab067700
                         4 rocksdb: EVENT_LOG_v1 {"time_micros":
                        1633549895008271, "cf_name": "L", "job":
                        8859, "event": "table_file_creation",
                        "file_number": 13657, "file_size":
                        2952537, "table_properties": {"data_size":
                        2951222, "index_size": 267,
                        "index_partitions": 0,
                        "top_level_index_size": 0,
                        "index_key_is_user_key": 0,
                        "index_value_is_delta_encoded": 0,
                        "filter_size": 197, "raw_key_size": 1120,
                        "raw_average_key_size": 16,
                        "raw_value_size": 2950151,
                        "raw_average_value_size": 42145,
                        "num_data_blocks": 9, "num_entries": 70,
                        "num_deletions": 61, "num_merge_operands":
                        0, "num_range_deletions": 0,
                        "format_version": 0, "fixed_key_len": 0,
                        "filter_policy":
                        "rocksdb.BuiltinBloomFilter",
                        "column_family_name": "L",
                        "column_family_id": 10, "comparator":
                        "leveldb.BytewiseComparator",
                        "merge_operator": "nullptr",
                        "prefix_extractor_name": "nullptr",
                        "property_collectors": "[]",
                        "compression": "NoCompression",
                        "compression_options": "window_bits=-14;
                        level=32767; strategy=0; max_dict_bytes=0;
                        zstd_max_train_bytes=0; enabled=0; ",
                        "creation_time": 1633549854,
                        "oldest_key_time": 1633549854,
                        "file_creation_time": 1633549894}}
                        Oct 06 21:51:35 cd88-ceph-osdh-01
                        bash[6328]: debug
                        2021-10-06T19:51:35.004+0000 7f4aab067700
                         4 rocksdb: [flush_job.cc:401] [L] [JOB
                        8859] Level-0 flush table #13657: 2952537
                        bytes OK

                        ---

                        root@cd88-ceph-osdh-01:~# ceph osd set noout
                        root@cd88-ceph-osdh-01:~# ceph orch daemon
                        stop osd.8
                        root@cd88-ceph-osdh-01:~# ceph orch ps
                        ...
                        osd.7  cd133-ceph-osdh-01  running (4h)  
                          44s ago  -    2738M  5528M  16.2.5
                        6933c2a0b7dd  8a98ae61f0eb
                        osd.8  cd88-ceph-osdh-01   stopped      
                        5s ago  -        -  5528M  <unknown>
                         <unknown>     <unknown>
                        osd.9  cd133k-ceph-osdh-01 running (3d)  
                         5m ago    -    4673M  8966M  16.2.5
                        6933c2a0b7dd  0ff7584b1808
                        ...

                        ---

                        root@cd88-ceph-osdh-01:~#
                        ceph-kvstore-tool bluestore-kv
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/
                        compact
                        2021-10-06T21:53:50.559+0200 7f87bde3c240
                         0
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _open_db_and_around read-only:0 repair:0
                        2021-10-06T21:53:50.559+0200 7f87bde3c240
                         1 bdev(0x5644f056c800
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
                        open path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bdev(0x5644f056c800
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
                        open size 6001172414464 (0x57541c00000,
                        5.5 TiB) block_size 4096 (4 KiB)
                        rotational discard not supported
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _set_cache_sizes cache_size 1073741824
                        meta 0.45 kv 0.45 data 0.06
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bdev(0x5644f056cc00
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
                        open path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bdev(0x5644f056cc00
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
                        open size 187904819200 (0x2bc0000000, 175
                        GiB) block_size 4096 (4 KiB) rotational
                        discard not supported
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bluefs add_block_device bdev 1 path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
                        size 175 GiB
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bdev(0x5644f056d000
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
                        open path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bdev(0x5644f056d000
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
                        open size 6001172414464 (0x57541c00000,
                        5.5 TiB) block_size 4096 (4 KiB)
                        rotational discard not supported
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bluefs add_block_device bdev 2 path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
                        size 5.5 TiB
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bdev(0x5644f056d400
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
                        open path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bdev(0x5644f056d400
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
                        open size 48318382080 (0xb40000000, 45
                        GiB) block_size 4096 (4 KiB)
                        non-rotational discard supported
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bluefs add_block_device bdev 0 path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
                        size 45 GiB
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bluefs mount
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bluefs _init_alloc new, id 0, allocator
                        name bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bluefs _init_alloc new, id 1, allocator
                        name bluefs-db, allocator type hybrid,
                        capacity 0x2bc0000000, block size 0x100000
                        2021-10-06T21:53:50.563+0200 7f87bde3c240
                         1 bluefs _init_alloc shared, id 2,
                        capacity 0x57541c00000, block size 0x10000
                        2021-10-06T21:53:50.655+0200 7f87bde3c240
                         1 bluefs mount shared_bdev_used = 0
                        2021-10-06T21:53:50.655+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _prepare_db_environment set db_paths to
                        db,178509578240 db.slow,5701113793740
                        2021-10-06T22:01:32.715+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _open_db opened rocksdb path db options
                        compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2,max_total_wal_size=1073741824
                        2021-10-06T22:01:32.715+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _open_super_meta old nid_max 167450
                        2021-10-06T22:01:32.715+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _open_super_meta old blobid_max 30720
                        2021-10-06T22:01:32.715+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _open_super_meta freelist_type bitmap
                        2021-10-06T22:01:32.715+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _open_super_meta ondisk_format 4
                        compat_ondisk_format 3
                        2021-10-06T22:01:32.715+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _open_super_meta min_alloc_size 0x1000
                        2021-10-06T22:01:33.347+0200 7f87bde3c240
                         1 freelist init
                        2021-10-06T22:01:33.347+0200 7f87bde3c240
                         1 freelist _read_cfg
                        2021-10-06T22:01:33.347+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _init_alloc opening allocation metadata
                        2021-10-06T22:01:41.031+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _init_alloc loaded 5.0 TiB in 37191
                        extents, allocator type hybrid, capacity
                        0x57541c00000, block size 0x1000, free
                        0x502f8f9a000, fragmentation 2.76445e-05
                        2021-10-06T22:01:41.039+0200 7f87bde3c240
                         1 bluefs umount
                        2021-10-06T22:01:41.043+0200 7f87bde3c240
                         1 bdev(0x5644f056d400
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
                        close
                        2021-10-06T22:01:43.623+0200 7f87bde3c240
                         1 bdev(0x5644f056cc00
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
                        close
                        2021-10-06T22:01:54.727+0200 7f87bde3c240
                         1 bdev(0x5644f056d000
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
                        close
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bdev(0x5644f056d000
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
                        open path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bdev(0x5644f056d000
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
                        open size 187904819200 (0x2bc0000000, 175
                        GiB) block_size 4096 (4 KiB) rotational
                        discard not supported
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bluefs add_block_device bdev 1 path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db
                        size 175 GiB
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bdev(0x5644f056cc00
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
                        open path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bdev(0x5644f056cc00
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
                        open size 6001172414464 (0x57541c00000,
                        5.5 TiB) block_size 4096 (4 KiB)
                        rotational discard not supported
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bluefs add_block_device bdev 2 path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block
                        size 5.5 TiB
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bdev(0x5644f056d400
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
                        open path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bdev(0x5644f056d400
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
                        open size 48318382080 (0xb40000000, 45
                        GiB) block_size 4096 (4 KiB)
                        non-rotational discard supported
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bluefs add_block_device bdev 0 path
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal
                        size 45 GiB
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bluefs mount
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bluefs _init_alloc new, id 0, allocator
                        name bluefs-wal, allocator type hybrid,
                        capacity 0xb40000000, block size 0x100000
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bluefs _init_alloc new, id 1, allocator
                        name bluefs-db, allocator type hybrid,
                        capacity 0x2bc0000000, block size 0x100000
                        2021-10-06T22:01:54.995+0200 7f87bde3c240
                         1 bluefs _init_alloc shared, id 2,
                        capacity 0x57541c00000, block size 0x10000
                        2021-10-06T22:01:55.079+0200 7f87bde3c240
                         1 bluefs mount shared_bdev_used = 0
                        2021-10-06T22:01:55.079+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _prepare_db_environment set db_paths to
                        db,178509578240 db.slow,5701113793740
                        2021-10-06T22:09:36.519+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        _open_db opened rocksdb path db options
                        compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2,max_total_wal_size=1073741824
                        2021-10-06T22:09:54.067+0200 7f87bde3c240
                         1
                        bluestore(/var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8/)
                        umount
                        2021-10-06T22:09:54.079+0200 7f87bde3c240
                         1 bluefs umount
                        2021-10-06T22:09:54.079+0200 7f87bde3c240
                         1 bdev(0x5644f056d400
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.wal)
                        close
                        2021-10-06T22:09:56.612+0200 7f87bde3c240
                         1 bdev(0x5644f056d000
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block.db)
                        close
                        2021-10-06T22:10:07.520+0200 7f87bde3c240
                         1 bdev(0x5644f056cc00
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
                        close
                        2021-10-06T22:10:07.688+0200 7f87bde3c240
                         1 freelist shutdown
                        2021-10-06T22:10:07.692+0200 7f87bde3c240
                         1 bdev(0x5644f056c800
                        /var/lib/ceph/462c44b4-eed6-11eb-8b2c-a1ad45f88a97/osd.8//block)
                        close

                        ---

                        root@cd88-ceph-osdh-01:~# ceph orch daemon
                        start osd.8

                        ---

                        root@cd88-ceph-osdh-01:/# ceph -s
                          cluster:
                            id: 462c44b4-eed6-11eb-8b2c-a1ad45f88a97
                            health: HEALTH_OK

                          services:
                            mon:   3 daemons, quorum
                        cd133-ceph-mon-01,cd88-ceph-mon-01,cd133k-ceph-mon-01
                        (age 15h)
                            mgr: cd133-ceph-mon-01.mzapob(active,
                        since 15h), standbys:
                        cd133k-ceph-mon-01.imikwh
                            osd:   60 osds: 60 up (since 2m), 60
                        in (since 3h)
                            rgw:   4 daemons active (2 hosts, 1 zones)
                        tcmu-runner: 10 portals active (2 hosts)

                          data:
                            pools:   6 pools, 361 pgs
                            objects: 2.46M objects, 8.0 TiB
                            usage: 33 TiB used, 304 TiB / 338 TiB
                        avail
                            pgs: 361 active+clean

                          io:
                            client: 45 MiB/s rd, 50 MiB/s wr, 921
                        op/s rd, 674 op/s wr



                        ---

                        root@cd88-ceph-osdh-01:/# ceph daemon
                        osd.8 bluestore bluefs device info
                        {
                            "dev": {
                        "device": "BDEV_WAL",
                        "total": 48318377984,
                        *      "free": 41354784768,
                        * "bluefs_used": 6963593216
                            },
                            "dev": {
                        "device": "BDEV_DB",
                        "total": 187904811008,
                        "free": 187302928384,
                        "bluefs_used": 601882624
                            },
                            "dev": {
                        "device": "BDEV_SLOW",
                        "total": 6001172414464,
                        "free": 5507531620352,
                        "bluefs_used": 0,
                        "bluefs max available": 5505566572544
                            }
                        }

                        ---

                        root@cd88-ceph-osdh-01:/# ceph daemon
                        osd.8 bluefs stats
                        0 : device size 0xb3ffff000 : using
                        0x1a0c00000(6.5 GiB)
                        1 : device size 0x2bbfffe000 : using
                        0x23e00000(574 MiB)
                        2 : device size 0x57541c00000 : using
                        0x72f0803000(460 GiB)
                        RocksDBBlueFSVolumeSelector:
                        wal_total:45902462976,
                        db_total:178509578240,
                        slow_total:5701113793740,
                        db_avail:103884521472
                        Usage matrix:
                        DEV/LEV WAL         DB          SLOW      
                         *     *   REAL  FILES
                        LOG         12 MiB      18 MiB    0 B 0 B
                                0 B         10 MiB      0
                        WAL 6.5 GiB     0 B         0 B         0
                        B       0 B     6.4 GiB 102
                        DB          0 B         573 MiB     0 B  
                            0 B     0 B   557 MiB 22
                        SLOW        0 B         0 B         0 B  
                            0 B     0 B   0 B 0
                        TOTALS  6.5 GiB 591 MiB     0 B         0
                        B         0 B       0 B     125
                        MAXIMUMS:
                        LOG         12 MiB      18 MiB    0 B 0 B
                                0 B         17 MiB
                        WAL         45 GiB      101 GiB     0 B  
                            0 B     0 B   145 GiB
                        DB          0 B         688 MiB     0 B  
                            0 B     0 B   670 MiB
                        SLOW        0 B         0 B         0 B  
                            0 B     0 B   0 B
                        TOTALS      45 GiB      101 GiB     0 B  
                            0 B     0 B   0 B

                        ----


                        Here are the osd.2... the problem
                        disapeared from alone
                        very strange...

                        root@cd88-ceph-osdh-01:/# ceph daemon
                        osd.2 bluefs stats
                        0 : device size 0xb3ffff000 : using
                        0x7bcc00000(31 GiB)
                        1 : device size 0x2bbfffe000 : using
                        0x458c00000(17 GiB)
                        2 : device size 0x57541c00000 : using
                        0x5cd3665000(371 GiB)
                        RocksDBBlueFSVolumeSelector:
                        wal_total:45902462976,
                        db_total:178509578240,
                        slow_total:5701113793740,
                        db_avail:103884521472
                        Usage matrix:
                        DEV/LEV WAL         DB          SLOW      
                         *     *   REAL  FILES
                        LOG 920 MiB 4.0 GiB     0 B         0 B  
                              0 B       10 MiB    1
                        WAL         31 GiB      17 GiB      0 B  
                            0 B     0 B   48 GiB  765
                        DB          0 B         193 MiB     0 B  
                            0 B     0 B   175 MiB 30
                        SLOW        0 B         0 B         0 B  
                            0 B     0 B   0 B 0
                        TOTALS      32 GiB      21 GiB      0 B  
                            0 B     0 B   0 B 796
                        MAXIMUMS:
                        LOG 920 MiB 4.0 GiB     0 B         0 B  
                              0 B       17 MiB
                        WAL         45 GiB      149 GiB     0 B  
                            0 B     0 B   192 GiB
                        DB          0 B         762 MiB     0 B  
                            0 B     0 B   741 MiB
                        SLOW        0 B         0 B         0 B  
                            0 B     0 B   0 B
                        TOTALS      45 GiB      153 GiB     0 B  
                            0 B     0 B   0 B
                        root@cd88-ceph-osdh-01:/# ceph daemon
                        osd.2 bluestore bluefs device info
                        {
                            "dev": {
                        "device": "BDEV_WAL",
                        "total": 48318377984,
                        "free": 15043915776,
                        "bluefs_used": 33274462208
                            },
                            "dev": {
                        "device": "BDEV_DB",
                        "total": 187904811008,
                        "free": 169235963904,
                        "bluefs_used": 18668847104
                            },
                            "dev": {
                        "device": "BDEV_SLOW",
                        "total": 6001172414464,
                        "free": 5602453327872,
                        "bluefs_used": 0,
                        "bluefs max available": 5600865222656
                            }
                        }






                        Am Mi., 6. Okt. 2021 um 18:11 Uhr schrieb
                        Igor Fedotov <igor.fedotov@xxxxxxxx
                        <mailto:igor.fedotov@xxxxxxxx>>:


                            On 10/6/2021 4:25 PM, José H. Freidhof
                            wrote:
                            > hi,
                            >
                            > no risk no fun 😂 okay
                            >   I have reset the settings you
                            mentioned to standard.
                            >
                            > what you exactly mean with taking
                            offline the osd? ceph orch daemon stop
                            > osd.2? or mark down?
                            "daemon stop" is enough. You  might
                            want to set noout flag before that
                            though...
                            >
                            > for the command which path i use?
                            you mean:
                            >
                            > bluestore-kv
                            /var/lib/ceph/$fsid/osd.2 compact???
                            yep
                            >
                            >
                            > Igor Fedotov <ifedotov@xxxxxxx
                            <mailto:ifedotov@xxxxxxx>> schrieb am
                            Mi., 6. Okt. 2021, 13:33:
                            >
                            >> On 10/6/2021 2:16 PM, José H.
                            Freidhof wrote:
                            >>> Hi Igor,
                            >>>
                            >>> yes i have some osd settings set
                            :-) here are my ceph config dump. those
                            >>> settings are from a redhat
                            document for bluestore devices
                            >>> maybe it is that setting causing
                            this problem? "advanced
                            >>> mon_compact_on_trim   false"???
                            >> OMG!!!
                            >>
                            >> No - mon_compact_on_trim has
                            nothing to deal with bluestore.
                            >>
                            >> Highly likely it's
                            bluestore_rocksdb_options which hurts...
                            >> Documentations tend to fall behind
                            the best practices.... I would
                            >> strongly discourage you from using
                            non-default settings unless it's
                            >> absolutely clear why this is necessary.
                            >>
                            >> Even at the first glance the
                            following settings (just a few ones I'm
                            >> completely aware) are
                            suboptimal/non-recommended:
                            >>
                            >> rocksdb_perf
                            >>
                            >> bluefs_sync_write
                            >>
                            >> bluefs_csum_type
                            >>
                            >>
                            >> Not to mention
                            bluestore_rocksdb_options which hasn't
                            got much adoption
                            >> so far and apparently greatly
                            alters rocksdb behavior...
                            >>
                            >>
                            >> So I would suggest to revert
                            rocksdb options back to default, run the
                            >> compaction and if it succeeds
                            monitor the OSD for a while. Then if it
                            >> works fine - apply the same for others
                            >>
                            >>
                            >> Hope this helps,
                            >>
                            >> Igor
                            >>
                            >>
                            >>
                            >>> i will test it this afternoon...
                            at the moment are everything semi
                            >>> prodcuctive and i need to repair
                            one osd node.. because i think of this
                            >>> reason the osds crashed on the
                            node and the osd container crashes with a
                            >>> dump while coming up now.
                            >>> need first to replicate all
                            between all three nodes and then i can
                            take
                            >>> offline the osd.2.and test your
                            command. i will inform you later...
                            >>>
                            >>> root@cd88-ceph-osdh-01:/# ceph
                            config dump
                            >>> WHO        MASK
                            >>>   LEVEL  OPTION    VALUE
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>         RO
                            >>> global
                            >>>   advanced leveldb_max_open_files
                                   131072
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   advanced mon_compact_on_trim false
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   dev  ms_crc_data     false
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   advanced osd_deep_scrub_interval
                            1209600.000000
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   advanced osd_max_scrubs  16
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   advanced
                            osd_scrub_load_threshold      0.010000
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   advanced osd_scrub_max_interval
                             1209600.000000
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   advanced osd_scrub_min_interval
                             86400.000000
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   advanced perf    true
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   advanced
                            rbd_readahead_disable_after_bytes    
                                  0
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   advanced rbd_readahead_max_bytes
                                  4194304
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   advanced rocksdb_perf    true
                            >>>
                            >>>
                            >>>
                            >>>
                            >>> global
                            >>>   advanced throttler_perf_counter
                                   false
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    mon
                            >>> advanced
                            auth_allow_insecure_global_id_reclaim
                                  false
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    mon
                            >>> advanced cluster_network
                            10.50.50.0/24 <http://10.50.50.0/24>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>         *
                            >>>    mon
                            >>> advanced mon_osd_down_out_interval
                                300
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    mon
                            >>> advanced public_network
                            10.50.50.0/24 <http://10.50.50.0/24>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>         *
                            >>>    mgr
                            >>> advanced
                            mgr/cephadm/container_init    True
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>        *
                            >>>    mgr
                            >>> advanced
                            mgr/cephadm/device_enhanced_scan  true
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>        *
                            >>>    mgr
                            >>> advanced
                            mgr/cephadm/migration_current 2
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>         *
                            >>>    mgr
                            >>> advanced
                            mgr/cephadm/warn_on_stray_daemons false
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>         *
                            >>>    mgr
                            >>> advanced
                            mgr/cephadm/warn_on_stray_hosts false
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>         *
                            >>>    osd
                            >>> advanced bluefs_sync_write true
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev  bluestore_cache_autotune    
                             true
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev  bluestore_cache_kv_ratio    
                             0.200000
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev  bluestore_cache_meta_ratio  
                             0.800000
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev  bluestore_cache_size  2147483648
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev  bluestore_cache_size_hdd
                             2147483648
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> advanced bluestore_csum_type      
                                none
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev
                             bluestore_extent_map_shard_max_size  
                                  200
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev
                             bluestore_extent_map_shard_min_size  
                                  50
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev
                             bluestore_extent_map_shard_target_size
                                 100
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> advanced bluestore_rocksdb_options
                            >>>
                            >>
                             compression=kNoCompression,max_write_buffer_number=64,min_write_buffer_number_to_merge=32,recycle_log_file_num=64,compaction_style=kCompactionStyleLevel,write_buffer_size=4MB,target_file_size_base=4MB,max_background_compactions=64,level0_file_num_compaction_trigger=64,level0_slowdown_writes_trigger=128,level0_stop_writes_trigger=256,max_bytes_for_level_base=6GB,compaction_threads=32,flusher_threads=8,compaction_readahead_size=2MB
                            >>>   *
                            >>>    osd
                            >>> advanced mon_osd_cache_size  1024
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev  ms_crc_data     false
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> advanced osd_map_share_max_epochs
                                 5
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> advanced osd_max_backfills        
                                1
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev  osd_max_pg_log_entries        10
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev  osd_memory_cache_min  3000000000
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>  host:cd133-ceph-osdh-01
                            >>> basic  osd_memory_target 5797322383
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>  host:cd133k-ceph-osdh-01
                            >>>   basic  osd_memory_target 9402402385
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd  host:cd88-ceph-osdh-01
                            >>>   basic  osd_memory_target 5797322096
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> advanced
                            osd_memory_target_autotune    true
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev  osd_min_pg_log_entries        10
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> advanced osd_op_num_shards        
                                8
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>         *
                            >>>    osd
                            >>> advanced
                            osd_op_num_threads_per_shard  2
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>         *
                            >>>    osd
                            >>> dev  osd_pg_log_dups_tracked       10
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> dev  osd_pg_log_trim_min           10
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> advanced osd_recovery_max_active  
                                3
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> advanced
                            osd_recovery_max_single_start 1
                            >>>
                            >>>
                            >>>
                            >>>
                            >>>    osd
                            >>> advanced osd_recovery_sleep  0.000000
                            >>>
                            >>>
                            >>> Am Mi., 6. Okt. 2021 um 12:55 Uhr
                            schrieb Igor Fedotov <ifedotov@xxxxxxx
                            <mailto:ifedotov@xxxxxxx>
                            >>> :
                            >>>
                            >>>> Jose,
                            >>>>
                            >>>> In fact 48GB is a way too much
                            for WAL drive - usually the write ahead
                            >> log
                            >>>> tend to be 2-4 GBs.
                            >>>>
                            >>>> But in your case it's ~150GB,
                            while DB itself is very small (146MB!!!):
                            >>>>
                            >>>> WAL         45 GiB      111 GiB 
                               0 B         0 B      0 B
                            >>>> 154 GiB     2400
                            >>>>
                            >>>> DB          0 B         164 MiB 
                               0 B         0 B      0 B
                            >>>> 146 MiB     30
                            >>>>
                            >>>>
                            >>>> which means that there are some
                            issues with RocksDB's WAL processing,
                            >>>> which needs some troubleshooting...
                            >>>>
                            >>>> Curious if other OSDs are
                            suffering from the same and whether
                            you have
                            >> any
                            >>>> custom settings for your OSD(s)?
                            >>>>
                            >>>> Additionally you might want to
                            try the following command to compact this
                            >>>> specific OSD manually and check
                            if this would normalize the DB layout -
                            >> the
                            >>>> majority of data has to be at DB
                            level not WAL. Please share the
                            >> resulting
                            >>>> layout (reported by "ceph daemon
                            osd.2 bluefs stats" command) after the
                            >>>> compaction is fulfiled and OSD is
                            restarted.
                            >>>>
                            >>>> The compaction command to be
                            applied on an offline OSD:
                            >> "ceph-kvstore-tool
                            >>>> bluestore-kv <path-to-osd> compact"
                            >>>>
                            >>>> Even if the above works great
                            please refrain from applying that
                            >> compaction
                            >>>> to every OSD - let's see how that
                            "compacted" OSD evolves.Would WAL grow
                            >>>> again or not?
                            >>>>
                            >>>> Thanks,
                            >>>>
                            >>>> Igor
                            >>>>
                            >>>>
                            >>>>
                            >>>>
                            >>>>
                            >>>>
                            >>>> On 10/6/2021 1:35 PM, José H.
                            Freidhof wrote:
                            >>>>
                            >>>> Hello Igor,
                            >>>>
                            >>>> yes the volume is  nvme wal
                            partitions for the bluestore devicegroups
                            >> are
                            >>>> only 48gb each
                            >>>>
                            >>>> on each osd node are 1 nvme with
                            1tb splitted in 20 lvs with 48gb (WAL)
                            >>>> on each osd node are 4 ssd with
                            1tb splitted in 5 lvs with 175gb
                            >> (rock.db)
                            >>>> on each osd node are 20 hdd with
                            5.5tb with 1 lvs (block.db)
                            >>>>
                            >>>> each blustore have 1 partition
                            nvme,ssd and hdd like described in the
                            >>>> documentation
                            >>>>
                            >>
                            https://docs.ceph.com/en/latest/rados/configuration/bluestore-config-ref/
                            <https://docs.ceph.com/en/latest/rados/configuration/bluestore-config-ref/>
                            >>>> is this to small or can i adjust
                            the max allocation on the wal nvme
                            >> device
                            >>>> in the ceph configuration?
                            >>>> i know that the ssd and nvme are
                            to small for those 5.5tb disk... its 1%
                            >>>> only ot the rotation disk.
                            >>>> i am new in ceph and still or
                            always learning, but we are in a little
                            >>>> hurry because our other
                            datastores are old and full.
                            >>>>
                            >>>> root@cd88-ceph-osdh-01:/# ceph
                            daemon osd.2 bluestore bluefs device
                            >> info
                            >>>> {
                            >>>>       "dev": {
                            >>>>           "device": "BDEV_WAL",
                            >>>>           "total": 48318377984,
                            >>>>           "free": 1044480,
                            >>>>           "bluefs_used": 48317333504
                            >>>>       },
                            >>>>       "dev": {
                            >>>>           "device": "BDEV_DB",
                            >>>>           "total": 187904811008,
                            >>>>           "free": 68757217280,
                            >>>>           "bluefs_used": 119147593728
                            >>>>       },
                            >>>>       "dev": {
                            >>>>           "device": "BDEV_SLOW",
                            >>>>           "total": 6001172414464,
                            >>>>           "free": 5624912359424,
                            >>>>           "bluefs_used": 0,
                            >>>>           "bluefs max available":
                            5624401231872
                            >>>>       }
                            >>>> }
                            >>>> root@cd88-ceph-osdh-01:/# ceph
                            daemon osd.2 bluefs stats
                            >>>> 0 : device size 0xb3ffff000 :
                            using 0xb3ff00000(45 GiB)
                            >>>> 1 : device size 0x2bbfffe000 :
                            using 0x1bbeb00000(111 GiB)
                            >>>> 2 : device size 0x57541c00000 :
                            using 0x579b592000(350 GiB)
                            >>>> RocksDBBlueFSVolumeSelector:
                            wal_total:45902462976,
                            >> db_total:178509578240,
                            >>>> slow_total:5701113793740,
                            db_avail:103884521472
                            >>>> Usage matrix:
                            >>>> DEV/LEV     WAL         DB       
                              SLOW        *      *
                            >>>> REAL        FILES
                            >>>> LOG         124 MiB     2.3 GiB 
                               0 B         0 B      0 B
                            >>>> 7.5 MiB     1
                            >>>> WAL         45 GiB      111 GiB 
                               0 B         0 B      0 B
                            >>>> 154 GiB     2400
                            >>>> DB          0 B         164 MiB 
                               0 B         0 B      0 B
                            >>>> 146 MiB     30
                            >>>> SLOW        0 B         0 B     
                               0 B         0 B      0 B
                            >>   0
                            >>>> B         0
                            >>>> TOTALS      45 GiB      113 GiB 
                               0 B         0 B      0 B
                            >>   0
                            >>>> B         2431
                            >>>> MAXIMUMS:
                            >>>> LOG         124 MiB     2.3 GiB 
                               0 B         0 B      0 B
                            >>   17
                            >>>> MiB
                            >>>> WAL         45 GiB      149 GiB 
                               0 B         0 B      0 B
                            >>>> 192 GiB
                            >>>> DB          0 B         762 MiB 
                               0 B         0 B      0 B
                            >>>> 741 MiB
                            >>>> SLOW        0 B         0 B     
                               0 B         0 B      0 B
                            >>   0 B
                            >>>> TOTALS      45 GiB      150 GiB 
                               0 B         0 B      0 B
                            >>   0 B
                            >>>> Am Mi., 6. Okt. 2021 um 11:45 Uhr
                            schrieb Igor Fedotov <
                            >> ifedotov@xxxxxxx
                            <mailto:ifedotov@xxxxxxx>>:
                            >>>>> Hey Jose,
                            >>>>>
                            >>>>> it looks like your WAL volume is
                            out of space which looks weird given
                            >>>>> its capacity = 48Gb.
                            >>>>>
                            >>>>> Could you please share the
                            output of the following commands:
                            >>>>>
                            >>>>> ceph daemon osd.N bluestore
                            bluefs device info
                            >>>>>
                            >>>>> ceph daemon osd.N bluefs stats
                            >>>>>
                            >>>>>
                            >>>>> Thanks,
                            >>>>>
                            >>>>> Igor
                            >>>>>
                            >>>>>
                            >>>>> On 10/6/2021 12:24 PM, José H.
                            Freidhof wrote:
                            >>>>>> Hello together
                            >>>>>>
                            >>>>>> we have a running ceph pacific
                            16.2.5 cluster and i found this
                            >> messages
                            >>>>> in
                            >>>>>> the service logs of the osd
                            daemons.
                            >>>>>>
                            >>>>>> we have three osd nodes .. each
                            node has 20osds as bluestore with
                            >>>>>> nvme/ssd/hdd
                            >>>>>>
                            >>>>>> is this a bug or maybe i have
                            some settings wrong?
                            >>>>>>
                            >>>>>>
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:25.821+0000
                            >>>>>> 7f38eebd4700  1 bluefs
                            _allocate unable to allocate 0x100000
                            on bdev
                            >> 0,
                            >>>>>> allocator name bluefs-wal,
                            allocator type hybrid, capacity
                            >> 0xb40000000,
                            >>>>>> block size 0x100000, free
                            0xff000, fragmentation 0, allocated 0x0
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:29.857+0000
                            >>>>>> 7f38eebd4700  1 bluefs
                            _allocate unable to allocate 0x100000
                            on bdev
                            >> 0,
                            >>>>>> allocator name bluefs-wal,
                            allocator type hybrid, capacity
                            >> 0xb40000000,
                            >>>>>> block size 0x100000, free
                            0xff000, fragmentation 0, allocated 0x0
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.073+0000
                            >>>>>> 7f38eebd4700  1 bluefs
                            _allocate unable to allocate 0x400000
                            on bdev
                            >> 0,
                            >>>>>> allocator name bluefs-wal,
                            allocator type hybrid, capacity
                            >> 0xb40000000,
                            >>>>>> block size 0x100000, free
                            0xff000, fragmentation 0, allocated 0x0
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.405+0000
                            >>>>>> 7f38eebd4700  1 bluefs
                            _allocate unable to allocate 0x100000
                            on bdev
                            >> 0,
                            >>>>>> allocator name bluefs-wal,
                            allocator type hybrid, capacity
                            >> 0xb40000000,
                            >>>>>> block size 0x100000, free
                            0xff000, fragmentation 0, allocated 0x0
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.465+0000
                            >>>>>> 7f38eebd4700  1 bluefs
                            _allocate unable to allocate 0x100000
                            on bdev
                            >> 0,
                            >>>>>> allocator name bluefs-wal,
                            allocator type hybrid, capacity
                            >> 0xb40000000,
                            >>>>>> block size 0x100000, free
                            0xff000, fragmentation 0, allocated 0x0
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.529+0000
                            >>>>>> 7f38eebd4700  1 bluefs
                            _allocate unable to allocate 0x100000
                            on bdev
                            >> 0,
                            >>>>>> allocator name bluefs-wal,
                            allocator type hybrid, capacity
                            >> 0xb40000000,
                            >>>>>> block size 0x100000, free
                            0xff000, fragmentation 0, allocated 0x0
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.545+0000
                            >>>>>> 7f38eebd4700  4 rocksdb:
                            [db_impl/db_impl_write.cc:1668] [L] New
                            >>>>> memtable
                            >>>>>> created with log file: #9588.
                            Immutable memtables: 1.
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.545+0000
                            >>>>>> 7f38eebd4700  1 bluefs
                            _allocate unable to allocate 0x100000
                            on bdev
                            >> 0,
                            >>>>>> allocator name bluefs-wal,
                            allocator type hybrid, capacity
                            >> 0xb40000000,
                            >>>>>> block size 0x100000, free
                            0xff000, fragmentation 0, allocated 0x0
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.545+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            (Original Log Time
                            >> 2021/10/06-09:17:30.547575)
                            >>>>>>
                            [db_impl/db_impl_compaction_flush.cc:2198]
                            Calling
                            >>>>>> FlushMemTableToOutputFile with
                            column family [L], flush slots
                            >> available
                            >>>>> 1,
                            >>>>>> compaction slots available 1,
                            flush slots scheduled 1, compaction
                            >> slots
                            >>>>>> scheduled 0
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.545+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            [flush_job.cc:321] [L] [JOB 5709] Flushing
                            >>>>>> memtable with next log file: 9587
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.545+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            [flush_job.cc:321] [L] [JOB 5709] Flushing
                            >>>>>> memtable with next log file: 9588
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.545+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850547916,
                            >>>>>> "job": 5709, "event":
                            "flush_started", "num_memtables": 2,
                            >>>>> "num_entries":
                            >>>>>> 4146, "num_deletes": 0,
                            "total_data_size": 127203926,
                            "memory_usage":
                            >>>>>> 130479920, "flush_reason":
                            "Write Buffer Full"}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.545+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            [flush_job.cc:350] [L] [JOB 5709] Level-0
                            >> flush
                            >>>>>> table #9589: started
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.557+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850559292,
                            >>>>>> "cf_name": "L", "job": 5709,
                            "event": "table_file_creation",
                            >>>>> "file_number":
                            >>>>>> 9589, "file_size": 3249934,
                            "table_properties": {"data_size": 3247855,
                            >>>>>> "index_size": 1031,
                            "index_partitions": 0,
                            "top_level_index_size": 0,
                            >>>>>> "index_key_is_user_key": 0,
                            "index_value_is_delta_encoded": 0,
                            >>>>>> "filter_size": 197,
                            "raw_key_size": 1088,
                            "raw_average_key_size": 16,
                            >>>>>> "raw_value_size": 3246252,
                            "raw_average_value_size": 47739,
                            >>>>>> "num_data_blocks": 36,
                            "num_entries": 68, "num_deletions": 32,
                            >>>>>> "num_merge_operands": 0,
                            "num_range_deletions": 0,
                            "format_version":
                            >> 0,
                            >>>>>> "fixed_key_len": 0,
                            "filter_policy":
                            "rocksdb.BuiltinBloomFilter",
                            >>>>>> "column_family_name": "L",
                            "column_family_id": 10, "comparator":
                            >>>>>> "leveldb.BytewiseComparator",
                            "merge_operator": "nullptr",
                            >>>>>> "prefix_extractor_name":
                            "nullptr", "property_collectors": "[]",
                            >>>>>> "compression": "NoCompression",
                            "compression_options":
                            >> "window_bits=-14;
                            >>>>>> level=32767; strategy=0;
                            max_dict_bytes=0; zstd_max_train_bytes=0;
                            >>>>>> enabled=0; ", "creation_time":
                            1633511730, "oldest_key_time":
                            >>>>> 1633511730,
                            >>>>>> "file_creation_time": 1633511850}}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.557+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            [flush_job.cc:401] [L] [JOB 5709] Level-0
                            >> flush
                            >>>>>> table #9589: 3249934 bytes OK
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.557+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            (Original Log Time
                            >> 2021/10/06-09:17:30.559362)
                            >>>>>> [memtable_list.cc:447] [L]
                            Level-0 commit table #9589 started
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.557+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            (Original Log Time
                            >> 2021/10/06-09:17:30.559583)
                            >>>>>> [memtable_list.cc:503] [L]
                            Level-0 commit table #9589: memtable #1
                            >> done
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.557+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            (Original Log Time
                            >> 2021/10/06-09:17:30.559586)
                            >>>>>> [memtable_list.cc:503] [L]
                            Level-0 commit table #9589: memtable #2
                            >> done
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.557+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            (Original Log Time
                            >> 2021/10/06-09:17:30.559601)
                            >>>>>> EVENT_LOG_v1 {"time_micros":
                            1633511850559593, "job": 5709, "event":
                            >>>>>> "flush_finished",
                            "output_compression": "NoCompression",
                            "lsm_state":
                            >>>>> [8,
                            >>>>>> 1, 0, 0, 0, 0, 0],
                            "immutable_memtables": 0}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.557+0000
                            >>>>>> 7f3905c02700  4 rocksdb:
                            (Original Log Time
                            >> 2021/10/06-09:17:30.559638)
                            >>>>>>
                            [db_impl/db_impl_compaction_flush.cc:205]
                            [L] Level summary: files[8 1
                            >>>>> 0 0
                            >>>>>> 0 0 0] max score 1.00
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.557+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            [compaction/compaction_job.cc:1676]
                            [L] [JOB
                            >>>>> 5710]
                            >>>>>> Compacting 8@0 + 1@1 files to
                            L1, score 1.00
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.557+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            [compaction/compaction_job.cc:1680] [L]
                            >>>>> Compaction
                            >>>>>> start summary: Base version
                            3090 Base level 0, inputs: [9589(3173KB)
                            >>>>>> 9586(4793KB) 9583(1876KB)
                            9580(194KB) 9576(6417KB) 9573(1078KB)
                            >>>>> 9570(405KB)
                            >>>>>> 9567(29KB)], [9564(1115KB)]
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.557+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850559956,
                            >>>>>> "job": 5710, "event":
                            "compaction_started", "compaction_reason":
                            >>>>>> "LevelL0FilesNum", "files_L0":
                            [9589, 9586, 9583, 9580, 9576, 9573,
                            >>>>> 9570,
                            >>>>>> 9567], "files_L1": [9564],
                            "score": 1, "input_data_size": 19542092}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            [compaction/compaction_job.cc:1349]
                            [L] [JOB
                            >>>>> 5710]
                            >>>>>> Generated table #9590: 36 keys,
                            3249524 bytes
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850582987,
                            >>>>>> "cf_name": "L", "job": 5710,
                            "event": "table_file_creation",
                            >>>>> "file_number":
                            >>>>>> 9590, "file_size": 3249524,
                            "table_properties": {"data_size": 3247449,
                            >>>>>> "index_size": 1031,
                            "index_partitions": 0,
                            "top_level_index_size": 0,
                            >>>>>> "index_key_is_user_key": 0,
                            "index_value_is_delta_encoded": 0,
                            >>>>>> "filter_size": 197,
                            "raw_key_size": 576,
                            "raw_average_key_size": 16,
                            >>>>>> "raw_value_size": 3246252,
                            "raw_average_value_size": 90173,
                            >>>>>> "num_data_blocks": 36,
                            "num_entries": 36, "num_deletions": 0,
                            >>>>>> "num_merge_operands": 0,
                            "num_range_deletions": 0,
                            "format_version":
                            >> 0,
                            >>>>>> "fixed_key_len": 0,
                            "filter_policy":
                            "rocksdb.BuiltinBloomFilter",
                            >>>>>> "column_family_name": "L",
                            "column_family_id": 10, "comparator":
                            >>>>>> "leveldb.BytewiseComparator",
                            "merge_operator": "nullptr",
                            >>>>>> "prefix_extractor_name":
                            "nullptr", "property_collectors": "[]",
                            >>>>>> "compression": "NoCompression",
                            "compression_options":
                            >> "window_bits=-14;
                            >>>>>> level=32767; strategy=0;
                            max_dict_bytes=0; zstd_max_train_bytes=0;
                            >>>>>> enabled=0; ", "creation_time":
                            1633471854, "oldest_key_time": 0,
                            >>>>>> "file_creation_time": 1633511850}}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            [compaction/compaction_job.cc:1415]
                            [L] [JOB
                            >>>>> 5710]
                            >>>>>> Compacted 8@0 + 1@1 files to L1
                            => 3249524 bytes
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            (Original Log Time
                            >> 2021/10/06-09:17:30.583469)
                            >>>>>>
                            [compaction/compaction_job.cc:760] [L]
                            compacted to: files[0 1 0 0 0 0
                            >>>>> 0]
                            >>>>>> max score 0.01, MB/sec: 846.1
                            rd, 140.7 wr, level 1, files in(8, 1)
                            >>>>> out(1)
                            >>>>>> MB in(17.5, 1.1) out(3.1),
                            read-write-amplify(1.2) write-amplify(0.2)
                            >>>>> OK,
                            >>>>>> records in: 376, records
                            dropped: 340 output_compression:
                            >> NoCompression
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            (Original Log Time
                            >> 2021/10/06-09:17:30.583498)
                            >>>>>> EVENT_LOG_v1 {"time_micros":
                            1633511850583485, "job": 5710, "event":
                            >>>>>> "compaction_finished",
                            "compaction_time_micros": 23098,
                            >>>>>> "compaction_time_cpu_micros":
                            20039, "output_level": 1,
                            >>>>> "num_output_files":
                            >>>>>> 1, "total_output_size":
                            3249524, "num_input_records": 376,
                            >>>>>> "num_output_records": 36,
                            "num_subcompactions": 1,
                            >> "output_compression":
                            >>>>>> "NoCompression",
                            "num_single_delete_mismatches": 0,
                            >>>>>>
                            "num_single_delete_fallthrough": 0,
                            "lsm_state": [0, 1, 0, 0, 0, 0,
                            >> 0]}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850583615,
                            >>>>>> "job": 5710, "event":
                            "table_file_deletion", "file_number":
                            9589}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850583648,
                            >>>>>> "job": 5710, "event":
                            "table_file_deletion", "file_number":
                            9586}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850583675,
                            >>>>>> "job": 5710, "event":
                            "table_file_deletion", "file_number":
                            9583}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850583709,
                            >>>>>> "job": 5710, "event":
                            "table_file_deletion", "file_number":
                            9580}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850583739,
                            >>>>>> "job": 5710, "event":
                            "table_file_deletion", "file_number":
                            9576}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850583769,
                            >>>>>> "job": 5710, "event":
                            "table_file_deletion", "file_number":
                            9573}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850583804,
                            >>>>>> "job": 5710, "event":
                            "table_file_deletion", "file_number":
                            9570}
                            >>>>>> cd88-ceph-osdh-01 bash[6283]:
                            debug 2021-10-06T09:17:30.581+0000
                            >>>>>> 7f38fb3ed700  4 rocksdb:
                            EVENT_LOG_v1 {"time_micros":
                            >> 1633511850583835,
                            >>>>>> "job": 5710, "event":
                            "table_file_deletion", "file_number":
                            9567}
                            >>>>>>
                            _______________________________________________
                            >>>>>> ceph-users mailing list --
                            ceph-users@xxxxxxx
                            <mailto:ceph-users@xxxxxxx>
                            >>>>>> To unsubscribe send an email to
                            ceph-users-leave@xxxxxxx
                            <mailto:ceph-users-leave@xxxxxxx>
                            >>>> --






--

Mit freundlichen Grüßen,

 -

José H. Freidhof

Reyerhütterstrasse 130b
41065 Mönchengladbach
eMail: harald.freidhof@xxxxxxxxx <mailto:harald.freidhof@xxxxxxxxx>
mobil: +49 (0) 1523 – 717 7801

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux