Hi Jeremy,
good to know you managed to bring your OSDs up.
Have you been able to reweight them to 0 and migrate data out of these
"broken" OSDs?
If so I suggest to redeploy them - the corruption is still in the DB and
it might pop-up one day.
If not please do that first - you might still hit into issues along the
process since DB is still corrupted. Disabling compaction just bypassed
the reads from broken files - but this might happen again during
different operations.
On 12/22/2020 7:17 AM, Jeremy Austin wrote:
Igor,
You're a bloomin' genius, as they say.
Disabling auto compaction allowed OSDs 11 and 12 to spin up/out. The 7
down PGs recovered; there were a few unfound items previously which I
went ahead and deleted, given that this is EC, revert not being an option.
HEALTH OK :)
I'm now intending to re-enable auto compaction. Should I also fsck the
rest of the OSDs, or is the typical scrub/deep scrub cycle sufficient?
(No PGs are behind on scrubbing, whereas they were during the degraded
period.)
Please do not re-enable the compaction before you're ready to kill
"broken" OSDs...
There is no much sense in fsck-ing other OSDs - this is unlikely to
detect inconsistencies at PG level if any.
Hence it's [deep]-scrub which you might want to apply manually once
reweight/rebalance is done. PGs located at "broken" OSDs are of the top
priority to scrub.
Time will tell if I actually lost data, I guess.
On Mon, Dec 21, 2020 at 8:37 AM Igor Fedotov <ifedotov@xxxxxxx
<mailto:ifedotov@xxxxxxx>> wrote:
Hi Jeremy,
you might want to try RocksDB's disable_auto_compactions option
for that.
To adjust rocksdb's options one should edit/insert
bluestore_rocksdb_options in ceph.conf.
E.g.
bluestore_rocksdb_options =
"disable_auto_compactions=true,compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2,max_total_wal_size=1073741824"
Please note bluestore specific defaults for rocksdb settings are
re-provided to make sure they aren't reset to rocksdb's ones.
Hope this helps.
Thanks,
Igor
On 12/21/2020 2:56 AM, Jeremy Austin wrote:
On Sun, Dec 20, 2020 at 2:25 PM Jeremy Austin <jhaustin@xxxxxxxxx
<mailto:jhaustin@xxxxxxxxx>> wrote:
Will attempt to disable compaction and report.
Not sure I'm doing this right. In [osd] section of ceph.conf, I added
periodic_compaction_seconds=0
and attempted to start the OSDs in question. Same error as
before. Am I setting compaction options correctly?
--
Jeremy Austin
jhaustin@xxxxxxxxx <mailto:jhaustin@xxxxxxxxx>
--
Jeremy Austin
jhaustin@xxxxxxxxx <mailto:jhaustin@xxxxxxxxx>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx