inline...
On Fri, Mar 22, 2019 at 1:08 PM Konstantin Shalygin <k0ste@xxxxxxxx> wrote:
On 3/22/19 11:57 PM, Mazzystr wrote:
> I am also seeing BlueFS spill since updating to Nautilus. I also see
> high slow_used_bytes and slow_total_bytes metrics. It sure looks to
> me that the only solution is to zap and rebuilt the osd. I had to
> manually check 36 osds some of them traditional processes and some
> containerized. The lack of tooling here is underwhelming... As soon
> as I rebuilt the osd the "BlueFS spill..." warning went away.
>
> I use 50Gb db partitions on an nvme with 3 or 6 Tb spinning disks. I
> don't understand the spillove
Wow, it's something new. What is your upgrade path?
I keep current with community. All osds have all been rebuilt as of luminous.
Also, you record cluster metrics, like via prometheus? To see diff
between upgrades.
Unfortunately not. I've only had prometheus running for about two weeks aaaand I had it turned off for a couple days for some unknown reason... :/
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com