Re: Bluefs WAL : bluefs _allocate failed to allocate on bdev 0

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

replying to my own message.

After I restarted the OSD it seems some of the wal partition got purged.
However there are still ~790MB used. As far as I think, it should get
completely emptied. At least this is what happens when I restart another
OSD, there its associated wal gets copletely flushed.
Is it somhow possible to reinitialize the wal for that OSD in question?

Thanks
  Dietmar


On 01/26/2018 05:11 PM, Dietmar Rieder wrote:
> Hi all,
> 
> I've a question regarding bluestore wal.db:
> 
> 
> We are running a 10 OSD node + 3 MON/MDS node cluster (luminous 12.2.2).
> Each OSD node has 22xHDD (8TB) OSDs, 2xSSD (1.6TB) OSDs and 2xNVME (800
> GB) for bluestore wal and db.
> 
> We have separated wal and db partitions
> wal partitions are 1GB
> db partitions are 64GB
> 
> The cluster is providing cephfs from one HDD (EC 6+3) and one SSD
> (3xrep) pool.
> Since the cluster is "new" we have not much data ~30TB (HDD EC) and
> ~140GB (SSD rep) stored on it yet.
> 
> I just noticed that the wal db usage for the SSD OSDs is all more or
> less equal ~518MB. The wal db usage for the HDD OSDs is as well quite
> balanced at 284-306MB, however there is one OSD whose wal db usage is ~ 1GB
> 
> 
>    "bluefs": {
>         "gift_bytes": 0,
>         "reclaim_bytes": 0,
>         "db_total_bytes": 68719468544,
>         "db_used_bytes": 1114636288,
>         "wal_total_bytes": 1073737728,
>         "wal_used_bytes": 1072693248,
>         "slow_total_bytes": 320057901056,
>         "slow_used_bytes": 0,
>         "num_files": 16,
>         "log_bytes": 862326784,
>         "log_compactions": 0,
>         "logged_bytes": 850575360,
>         "files_written_wal": 2,
>         "files_written_sst": 9,
>         "bytes_written_wal": 744469265,
>         "bytes_written_sst": 568855830
>     },
> 
> 
> and I got the following log entries:
> 
> 2018-01-26 16:31:05.484284 7f65ea28a700  1 bluefs _allocate failed to
> allocate 0x400000 on bdev 0, free 0xff000; fallback to bdev 1
> 
> Is there any reason for this difference ~300MB vs 1GB?
> I have in mind that 1GB of wal should be enough, and old logs should be
> purged to free space. (can this be triggered manually?)
> 
> Could this be related to the fact that the HDD OSD in question was
> failing some week ago and we replaced it with with a new HDD?
> 
> Do we have to expect problems/performace reductions, with the falling
> back to bdev 1?
> 
> Thanks for any clarifying comment
>    Dietmar
> 
> 
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 


-- 
_________________________________________
D i e t m a r  R i e d e r, Mag.Dr.
Innsbruck Medical University
Biocenter - Division for Bioinformatics
Email: dietmar.rieder@xxxxxxxxxxx
Web:   http://www.icbi.at


Attachment: signature.asc
Description: OpenPGP digital signature

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux