Re: bluefs _allocate unable to allocate

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



hi,

no risk no fun 😂 okay
 I have reset the settings you mentioned to standard.

what you exactly mean with taking offline the osd? ceph orch daemon stop
osd.2? or mark down?

for the command which path i use? you mean:

bluestore-kv /var/lib/ceph/$fsid/osd.2 compact???


Igor Fedotov <ifedotov@xxxxxxx> schrieb am Mi., 6. Okt. 2021, 13:33:

> On 10/6/2021 2:16 PM, José H. Freidhof wrote:
> > Hi Igor,
> >
> > yes i have some osd settings set :-) here are my ceph config dump. those
> > settings are from a redhat document for bluestore devices
> > maybe it is that setting causing this problem? "advanced
> >   mon_compact_on_trim    false"???
>
> OMG!!!
>
> No - mon_compact_on_trim has nothing to deal with bluestore.
>
> Highly likely it's bluestore_rocksdb_options which hurts...
> Documentations tend to fall behind the best practices.... I would
> strongly discourage you from using non-default settings unless it's
> absolutely clear why this is necessary.
>
> Even at the first glance the following settings (just a few ones I'm
> completely aware) are suboptimal/non-recommended:
>
> rocksdb_perf
>
> bluefs_sync_write
>
> bluefs_csum_type
>
>
> Not to mention bluestore_rocksdb_options which hasn't got much adoption
> so far and apparently greatly alters rocksdb behavior...
>
>
> So I would suggest to revert rocksdb options back to default, run the
> compaction and if it succeeds monitor the OSD for a while. Then if it
> works fine - apply the same for others
>
>
> Hope this helps,
>
> Igor
>
>
>
> >
> > i will test it this afternoon... at the moment are everything semi
> > prodcuctive and i need to repair one osd node.. because i think of this
> > reason the osds crashed on the node and the osd container crashes with a
> > dump while coming up now.
> > need first to replicate all between all three nodes and then i can take
> > offline the osd.2.and test your command. i will inform you later...
> >
> > root@cd88-ceph-osdh-01:/# ceph config dump
> > WHO                                               MASK
> >   LEVEL     OPTION                                       VALUE
> >
> >
> >
> >
> >
> >                                     RO
> > global
> >   advanced  leveldb_max_open_files                       131072
> >
> >
> >
> >
> > global
> >   advanced  mon_compact_on_trim                          false
> >
> >
> >
> >
> > global
> >   dev       ms_crc_data                                  false
> >
> >
> >
> >
> > global
> >   advanced  osd_deep_scrub_interval                      1209600.000000
> >
> >
> >
> >
> > global
> >   advanced  osd_max_scrubs                               16
> >
> >
> >
> >
> > global
> >   advanced  osd_scrub_load_threshold                     0.010000
> >
> >
> >
> >
> > global
> >   advanced  osd_scrub_max_interval                       1209600.000000
> >
> >
> >
> >
> > global
> >   advanced  osd_scrub_min_interval                       86400.000000
> >
> >
> >
> >
> > global
> >   advanced  perf                                         true
> >
> >
> >
> >
> > global
> >   advanced  rbd_readahead_disable_after_bytes            0
> >
> >
> >
> >
> > global
> >   advanced  rbd_readahead_max_bytes                      4194304
> >
> >
> >
> >
> > global
> >   advanced  rocksdb_perf                                 true
> >
> >
> >
> >
> > global
> >   advanced  throttler_perf_counter                       false
> >
> >
> >
> >
> >    mon
> > advanced  auth_allow_insecure_global_id_reclaim        false
> >
> >
> >
> >
> >    mon
> > advanced  cluster_network                              10.50.50.0/24
> >
> >
> >
> >
> >
> >                                     *
> >    mon
> > advanced  mon_osd_down_out_interval                    300
> >
> >
> >
> >
> >    mon
> > advanced  public_network                               10.50.50.0/24
> >
> >
> >
> >
> >
> >                                     *
> >    mgr
> > advanced  mgr/cephadm/container_init                   True
> >
> >
> >
> >
> >
> >                                    *
> >    mgr
> > advanced  mgr/cephadm/device_enhanced_scan             true
> >
> >
> >
> >
> >
> >                                    *
> >    mgr
> > advanced  mgr/cephadm/migration_current                2
> >
> >
> >
> >
> >
> >                                     *
> >    mgr
> > advanced  mgr/cephadm/warn_on_stray_daemons            false
> >
> >
> >
> >
> >
> >                                     *
> >    mgr
> > advanced  mgr/cephadm/warn_on_stray_hosts              false
> >
> >
> >
> >
> >
> >                                     *
> >    osd
> > advanced  bluefs_sync_write                            true
> >
> >
> >
> >
> >    osd
> > dev       bluestore_cache_autotune                     true
> >
> >
> >
> >
> >    osd
> > dev       bluestore_cache_kv_ratio                     0.200000
> >
> >
> >
> >
> >    osd
> > dev       bluestore_cache_meta_ratio                   0.800000
> >
> >
> >
> >
> >    osd
> > dev       bluestore_cache_size                         2147483648
> >
> >
> >
> >
> >    osd
> > dev       bluestore_cache_size_hdd                     2147483648
> >
> >
> >
> >
> >    osd
> > advanced  bluestore_csum_type                          none
> >
> >
> >
> >
> >    osd
> > dev       bluestore_extent_map_shard_max_size          200
> >
> >
> >
> >
> >    osd
> > dev       bluestore_extent_map_shard_min_size          50
> >
> >
> >
> >
> >    osd
> > dev       bluestore_extent_map_shard_target_size       100
> >
> >
> >
> >
> >    osd
> > advanced  bluestore_rocksdb_options
> >
>  compression=kNoCompression,max_write_buffer_number=64,min_write_buffer_number_to_merge=32,recycle_log_file_num=64,compaction_style=kCompactionStyleLevel,write_buffer_size=4MB,target_file_size_base=4MB,max_background_compactions=64,level0_file_num_compaction_trigger=64,level0_slowdown_writes_trigger=128,level0_stop_writes_trigger=256,max_bytes_for_level_base=6GB,compaction_threads=32,flusher_threads=8,compaction_readahead_size=2MB
> >   *
> >    osd
> > advanced  mon_osd_cache_size                           1024
> >
> >
> >
> >
> >    osd
> > dev       ms_crc_data                                  false
> >
> >
> >
> >
> >    osd
> > advanced  osd_map_share_max_epochs                     5
> >
> >
> >
> >
> >    osd
> > advanced  osd_max_backfills                            1
> >
> >
> >
> >
> >    osd
> > dev       osd_max_pg_log_entries                       10
> >
> >
> >
> >
> >    osd
> > dev       osd_memory_cache_min                         3000000000
> >
> >
> >
> >
> >    osd
>  host:cd133-ceph-osdh-01
> > basic     osd_memory_target                            5797322383
> >
> >
> >
> >
> >    osd
>  host:cd133k-ceph-osdh-01
> >   basic     osd_memory_target                            9402402385
> >
> >
> >
> >
> >    osd                                             host:cd88-ceph-osdh-01
> >   basic     osd_memory_target                            5797322096
> >
> >
> >
> >
> >    osd
> > advanced  osd_memory_target_autotune                   true
> >
> >
> >
> >
> >    osd
> > dev       osd_min_pg_log_entries                       10
> >
> >
> >
> >
> >    osd
> > advanced  osd_op_num_shards                            8
> >
> >
> >
> >
> >
> >                                     *
> >    osd
> > advanced  osd_op_num_threads_per_shard                 2
> >
> >
> >
> >
> >
> >                                     *
> >    osd
> > dev       osd_pg_log_dups_tracked                      10
> >
> >
> >
> >
> >    osd
> > dev       osd_pg_log_trim_min                          10
> >
> >
> >
> >
> >    osd
> > advanced  osd_recovery_max_active                      3
> >
> >
> >
> >
> >    osd
> > advanced  osd_recovery_max_single_start                1
> >
> >
> >
> >
> >    osd
> > advanced  osd_recovery_sleep                           0.000000
> >
> >
> > Am Mi., 6. Okt. 2021 um 12:55 Uhr schrieb Igor Fedotov <ifedotov@xxxxxxx
> >:
> >
> >> Jose,
> >>
> >> In fact 48GB is a way too much for WAL drive - usually the write ahead
> log
> >> tend to be 2-4 GBs.
> >>
> >> But in your case it's ~150GB, while DB itself is very small (146MB!!!):
> >>
> >> WAL         45 GiB      111 GiB     0 B         0 B         0 B
> >> 154 GiB     2400
> >>
> >> DB          0 B         164 MiB     0 B         0 B         0 B
> >> 146 MiB     30
> >>
> >>
> >> which means that there are some issues with RocksDB's WAL processing,
> >> which needs some troubleshooting...
> >>
> >> Curious if other OSDs are suffering from the same and whether you have
> any
> >> custom settings for your OSD(s)?
> >>
> >> Additionally you might want to try the following command to compact this
> >> specific OSD manually and check if this would normalize the DB layout -
> the
> >> majority of data has to be at DB level not WAL. Please share the
> resulting
> >> layout (reported by "ceph daemon osd.2 bluefs stats" command) after the
> >> compaction is fulfiled and OSD is restarted.
> >>
> >> The compaction command to be applied on an offline OSD:
> "ceph-kvstore-tool
> >> bluestore-kv <path-to-osd> compact"
> >>
> >> Even if the above works great please refrain from applying that
> compaction
> >> to every OSD - let's see how that "compacted" OSD evolves.Would WAL grow
> >> again or not?
> >>
> >> Thanks,
> >>
> >> Igor
> >>
> >>
> >>
> >>
> >>
> >>
> >> On 10/6/2021 1:35 PM, José H. Freidhof wrote:
> >>
> >> Hello Igor,
> >>
> >> yes the volume is  nvme wal partitions for the bluestore devicegroups
> are
> >> only 48gb each
> >>
> >> on each osd node are 1 nvme with 1tb splitted in 20 lvs with 48gb (WAL)
> >> on each osd node are 4 ssd with 1tb splitted in 5 lvs with 175gb
> (rock.db)
> >> on each osd node are 20 hdd with 5.5tb with 1 lvs (block.db)
> >>
> >> each blustore have 1 partition nvme,ssd and hdd like described in the
> >> documentation
> >>
> https://docs.ceph.com/en/latest/rados/configuration/bluestore-config-ref/
> >>
> >> is this to small or can i adjust the max allocation on the wal nvme
> device
> >> in the ceph configuration?
> >> i know that the ssd and nvme are to small for those 5.5tb disk... its 1%
> >> only ot the rotation disk.
> >> i am new in ceph and still or always learning, but we are in a little
> >> hurry because our other datastores are old and full.
> >>
> >> root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluestore bluefs device
> info
> >> {
> >>      "dev": {
> >>          "device": "BDEV_WAL",
> >>          "total": 48318377984,
> >>          "free": 1044480,
> >>          "bluefs_used": 48317333504
> >>      },
> >>      "dev": {
> >>          "device": "BDEV_DB",
> >>          "total": 187904811008,
> >>          "free": 68757217280,
> >>          "bluefs_used": 119147593728
> >>      },
> >>      "dev": {
> >>          "device": "BDEV_SLOW",
> >>          "total": 6001172414464,
> >>          "free": 5624912359424,
> >>          "bluefs_used": 0,
> >>          "bluefs max available": 5624401231872
> >>      }
> >> }
> >> root@cd88-ceph-osdh-01:/# ceph daemon osd.2 bluefs stats
> >> 0 : device size 0xb3ffff000 : using 0xb3ff00000(45 GiB)
> >> 1 : device size 0x2bbfffe000 : using 0x1bbeb00000(111 GiB)
> >> 2 : device size 0x57541c00000 : using 0x579b592000(350 GiB)
> >> RocksDBBlueFSVolumeSelector: wal_total:45902462976,
> db_total:178509578240,
> >> slow_total:5701113793740, db_avail:103884521472
> >> Usage matrix:
> >> DEV/LEV     WAL         DB          SLOW        *           *
> >> REAL        FILES
> >> LOG         124 MiB     2.3 GiB     0 B         0 B         0 B
> >> 7.5 MiB     1
> >> WAL         45 GiB      111 GiB     0 B         0 B         0 B
> >> 154 GiB     2400
> >> DB          0 B         164 MiB     0 B         0 B         0 B
> >> 146 MiB     30
> >> SLOW        0 B         0 B         0 B         0 B         0 B
>  0
> >> B         0
> >> TOTALS      45 GiB      113 GiB     0 B         0 B         0 B
>  0
> >> B         2431
> >> MAXIMUMS:
> >> LOG         124 MiB     2.3 GiB     0 B         0 B         0 B
>  17
> >> MiB
> >> WAL         45 GiB      149 GiB     0 B         0 B         0 B
> >> 192 GiB
> >> DB          0 B         762 MiB     0 B         0 B         0 B
> >> 741 MiB
> >> SLOW        0 B         0 B         0 B         0 B         0 B
>  0 B
> >> TOTALS      45 GiB      150 GiB     0 B         0 B         0 B
>  0 B
> >>
> >> Am Mi., 6. Okt. 2021 um 11:45 Uhr schrieb Igor Fedotov <
> ifedotov@xxxxxxx>:
> >>
> >>> Hey Jose,
> >>>
> >>> it looks like your WAL volume is out of space which looks weird given
> >>> its capacity = 48Gb.
> >>>
> >>> Could you please share the output of the following commands:
> >>>
> >>> ceph daemon osd.N bluestore bluefs device info
> >>>
> >>> ceph daemon osd.N bluefs stats
> >>>
> >>>
> >>> Thanks,
> >>>
> >>> Igor
> >>>
> >>>
> >>> On 10/6/2021 12:24 PM, José H. Freidhof wrote:
> >>>> Hello together
> >>>>
> >>>> we have a running ceph pacific 16.2.5 cluster and i found this
> messages
> >>> in
> >>>> the service logs of the osd daemons.
> >>>>
> >>>> we have three osd nodes .. each node has 20osds as bluestore with
> >>>> nvme/ssd/hdd
> >>>>
> >>>> is this a bug or maybe i have some settings wrong?
> >>>>
> >>>>
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:25.821+0000
> >>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on bdev
> 0,
> >>>> allocator name bluefs-wal, allocator type hybrid, capacity
> 0xb40000000,
> >>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:29.857+0000
> >>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on bdev
> 0,
> >>>> allocator name bluefs-wal, allocator type hybrid, capacity
> 0xb40000000,
> >>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.073+0000
> >>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x400000 on bdev
> 0,
> >>>> allocator name bluefs-wal, allocator type hybrid, capacity
> 0xb40000000,
> >>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.405+0000
> >>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on bdev
> 0,
> >>>> allocator name bluefs-wal, allocator type hybrid, capacity
> 0xb40000000,
> >>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.465+0000
> >>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on bdev
> 0,
> >>>> allocator name bluefs-wal, allocator type hybrid, capacity
> 0xb40000000,
> >>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.529+0000
> >>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on bdev
> 0,
> >>>> allocator name bluefs-wal, allocator type hybrid, capacity
> 0xb40000000,
> >>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
> >>>> 7f38eebd4700  4 rocksdb: [db_impl/db_impl_write.cc:1668] [L] New
> >>> memtable
> >>>> created with log file: #9588. Immutable memtables: 1.
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
> >>>> 7f38eebd4700  1 bluefs _allocate unable to allocate 0x100000 on bdev
> 0,
> >>>> allocator name bluefs-wal, allocator type hybrid, capacity
> 0xb40000000,
> >>>> block size 0x100000, free 0xff000, fragmentation 0, allocated 0x0
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
> >>>> 7f3905c02700  4 rocksdb: (Original Log Time
> 2021/10/06-09:17:30.547575)
> >>>> [db_impl/db_impl_compaction_flush.cc:2198] Calling
> >>>> FlushMemTableToOutputFile with column family [L], flush slots
> available
> >>> 1,
> >>>> compaction slots available 1, flush slots scheduled 1, compaction
> slots
> >>>> scheduled 0
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
> >>>> 7f3905c02700  4 rocksdb: [flush_job.cc:321] [L] [JOB 5709] Flushing
> >>>> memtable with next log file: 9587
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
> >>>> 7f3905c02700  4 rocksdb: [flush_job.cc:321] [L] [JOB 5709] Flushing
> >>>> memtable with next log file: 9588
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
> >>>> 7f3905c02700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850547916,
> >>>> "job": 5709, "event": "flush_started", "num_memtables": 2,
> >>> "num_entries":
> >>>> 4146, "num_deletes": 0, "total_data_size": 127203926, "memory_usage":
> >>>> 130479920, "flush_reason": "Write Buffer Full"}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.545+0000
> >>>> 7f3905c02700  4 rocksdb: [flush_job.cc:350] [L] [JOB 5709] Level-0
> flush
> >>>> table #9589: started
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
> >>>> 7f3905c02700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850559292,
> >>>> "cf_name": "L", "job": 5709, "event": "table_file_creation",
> >>> "file_number":
> >>>> 9589, "file_size": 3249934, "table_properties": {"data_size": 3247855,
> >>>> "index_size": 1031, "index_partitions": 0, "top_level_index_size": 0,
> >>>> "index_key_is_user_key": 0, "index_value_is_delta_encoded": 0,
> >>>> "filter_size": 197, "raw_key_size": 1088, "raw_average_key_size": 16,
> >>>> "raw_value_size": 3246252, "raw_average_value_size": 47739,
> >>>> "num_data_blocks": 36, "num_entries": 68, "num_deletions": 32,
> >>>> "num_merge_operands": 0, "num_range_deletions": 0, "format_version":
> 0,
> >>>> "fixed_key_len": 0, "filter_policy": "rocksdb.BuiltinBloomFilter",
> >>>> "column_family_name": "L", "column_family_id": 10, "comparator":
> >>>> "leveldb.BytewiseComparator", "merge_operator": "nullptr",
> >>>> "prefix_extractor_name": "nullptr", "property_collectors": "[]",
> >>>> "compression": "NoCompression", "compression_options":
> "window_bits=-14;
> >>>> level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0;
> >>>> enabled=0; ", "creation_time": 1633511730, "oldest_key_time":
> >>> 1633511730,
> >>>> "file_creation_time": 1633511850}}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
> >>>> 7f3905c02700  4 rocksdb: [flush_job.cc:401] [L] [JOB 5709] Level-0
> flush
> >>>> table #9589: 3249934 bytes OK
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
> >>>> 7f3905c02700  4 rocksdb: (Original Log Time
> 2021/10/06-09:17:30.559362)
> >>>> [memtable_list.cc:447] [L] Level-0 commit table #9589 started
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
> >>>> 7f3905c02700  4 rocksdb: (Original Log Time
> 2021/10/06-09:17:30.559583)
> >>>> [memtable_list.cc:503] [L] Level-0 commit table #9589: memtable #1
> done
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
> >>>> 7f3905c02700  4 rocksdb: (Original Log Time
> 2021/10/06-09:17:30.559586)
> >>>> [memtable_list.cc:503] [L] Level-0 commit table #9589: memtable #2
> done
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
> >>>> 7f3905c02700  4 rocksdb: (Original Log Time
> 2021/10/06-09:17:30.559601)
> >>>> EVENT_LOG_v1 {"time_micros": 1633511850559593, "job": 5709, "event":
> >>>> "flush_finished", "output_compression": "NoCompression", "lsm_state":
> >>> [8,
> >>>> 1, 0, 0, 0, 0, 0], "immutable_memtables": 0}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
> >>>> 7f3905c02700  4 rocksdb: (Original Log Time
> 2021/10/06-09:17:30.559638)
> >>>> [db_impl/db_impl_compaction_flush.cc:205] [L] Level summary: files[8 1
> >>> 0 0
> >>>> 0 0 0] max score 1.00
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
> >>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1676] [L] [JOB
> >>> 5710]
> >>>> Compacting 8@0 + 1@1 files to L1, score 1.00
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
> >>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1680] [L]
> >>> Compaction
> >>>> start summary: Base version 3090 Base level 0, inputs: [9589(3173KB)
> >>>> 9586(4793KB) 9583(1876KB) 9580(194KB) 9576(6417KB) 9573(1078KB)
> >>> 9570(405KB)
> >>>> 9567(29KB)], [9564(1115KB)]
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.557+0000
> >>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850559956,
> >>>> "job": 5710, "event": "compaction_started", "compaction_reason":
> >>>> "LevelL0FilesNum", "files_L0": [9589, 9586, 9583, 9580, 9576, 9573,
> >>> 9570,
> >>>> 9567], "files_L1": [9564], "score": 1, "input_data_size": 19542092}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1349] [L] [JOB
> >>> 5710]
> >>>> Generated table #9590: 36 keys, 3249524 bytes
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850582987,
> >>>> "cf_name": "L", "job": 5710, "event": "table_file_creation",
> >>> "file_number":
> >>>> 9590, "file_size": 3249524, "table_properties": {"data_size": 3247449,
> >>>> "index_size": 1031, "index_partitions": 0, "top_level_index_size": 0,
> >>>> "index_key_is_user_key": 0, "index_value_is_delta_encoded": 0,
> >>>> "filter_size": 197, "raw_key_size": 576, "raw_average_key_size": 16,
> >>>> "raw_value_size": 3246252, "raw_average_value_size": 90173,
> >>>> "num_data_blocks": 36, "num_entries": 36, "num_deletions": 0,
> >>>> "num_merge_operands": 0, "num_range_deletions": 0, "format_version":
> 0,
> >>>> "fixed_key_len": 0, "filter_policy": "rocksdb.BuiltinBloomFilter",
> >>>> "column_family_name": "L", "column_family_id": 10, "comparator":
> >>>> "leveldb.BytewiseComparator", "merge_operator": "nullptr",
> >>>> "prefix_extractor_name": "nullptr", "property_collectors": "[]",
> >>>> "compression": "NoCompression", "compression_options":
> "window_bits=-14;
> >>>> level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0;
> >>>> enabled=0; ", "creation_time": 1633471854, "oldest_key_time": 0,
> >>>> "file_creation_time": 1633511850}}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: [compaction/compaction_job.cc:1415] [L] [JOB
> >>> 5710]
> >>>> Compacted 8@0 + 1@1 files to L1 => 3249524 bytes
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: (Original Log Time
> 2021/10/06-09:17:30.583469)
> >>>> [compaction/compaction_job.cc:760] [L] compacted to: files[0 1 0 0 0 0
> >>> 0]
> >>>> max score 0.01, MB/sec: 846.1 rd, 140.7 wr, level 1, files in(8, 1)
> >>> out(1)
> >>>> MB in(17.5, 1.1) out(3.1), read-write-amplify(1.2) write-amplify(0.2)
> >>> OK,
> >>>> records in: 376, records dropped: 340 output_compression:
> NoCompression
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: (Original Log Time
> 2021/10/06-09:17:30.583498)
> >>>> EVENT_LOG_v1 {"time_micros": 1633511850583485, "job": 5710, "event":
> >>>> "compaction_finished", "compaction_time_micros": 23098,
> >>>> "compaction_time_cpu_micros": 20039, "output_level": 1,
> >>> "num_output_files":
> >>>> 1, "total_output_size": 3249524, "num_input_records": 376,
> >>>> "num_output_records": 36, "num_subcompactions": 1,
> "output_compression":
> >>>> "NoCompression", "num_single_delete_mismatches": 0,
> >>>> "num_single_delete_fallthrough": 0, "lsm_state": [0, 1, 0, 0, 0, 0,
> 0]}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850583615,
> >>>> "job": 5710, "event": "table_file_deletion", "file_number": 9589}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850583648,
> >>>> "job": 5710, "event": "table_file_deletion", "file_number": 9586}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850583675,
> >>>> "job": 5710, "event": "table_file_deletion", "file_number": 9583}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850583709,
> >>>> "job": 5710, "event": "table_file_deletion", "file_number": 9580}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850583739,
> >>>> "job": 5710, "event": "table_file_deletion", "file_number": 9576}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850583769,
> >>>> "job": 5710, "event": "table_file_deletion", "file_number": 9573}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850583804,
> >>>> "job": 5710, "event": "table_file_deletion", "file_number": 9570}
> >>>> cd88-ceph-osdh-01 bash[6283]: debug 2021-10-06T09:17:30.581+0000
> >>>> 7f38fb3ed700  4 rocksdb: EVENT_LOG_v1 {"time_micros":
> 1633511850583835,
> >>>> "job": 5710, "event": "table_file_deletion", "file_number": 9567}
> >>>> _______________________________________________
> >>>> ceph-users mailing list -- ceph-users@xxxxxxx
> >>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >>
> >> --
> >>
> >> Mit freundlichen Grüßen,
> >>
> >>   -
> >>
> >> José H. Freidhof
> >>
> >> Reyerhütterstrasse 130b
> >> 41065 Mönchengladbach
> >> eMail: harald.freidhof@xxxxxxxxx
> >> mobil: +49 (0) 1523 – 717 7801
> >>
> >>
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux