Aw: Re: [PATCH] mark rbd requiring stable pages

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I have done some testing through the weekend on the same testmachine
i used the week before, but wasn't successfull to reproduce any corruption
this time.
 
I was rethinking my tests and why the results where different this time and
what your might also be different in your own test-environment.  This
difference might be the recovering state of my ceph installation and its effect
on the available memory and maybe other stats of the machine.
 
Back then when the corruptions occured on my systems i would have several
problems because of the closed connections caused by the bad crcs, with OSDs
either being restarted by myself or getting marked out and in of the cluster,
because the other OSDs didn't get the necessary ping from the OSD for too long.
 
The result of this was a noticeable increase of RAM-use. After some peerings
and recoverys the machines would use ca. 500-800 MB swap, when it would
have used none before.
 
This might also explain why my own attempts to reproduce the corruptions on
purpose were never successfull. This tests usually happened pretty much shortly
after a fresh start of the cluster.
 
Same situation on my test machine.  Before i run into a corruption on this machine 2 
weeks ago, it had already executed some recoverys and backfills on the OSDs that
are hosted on this machine. This resulted in ca. 900 MB swap being used.
 
This weekend the machine had the same amount of OSDs started, but there was not
much of recovering necessary, so there was no swap used and enogh Ram
available for ca. 2 GB of vfs-caches.
 
Whatever it might be, the recovering/peering process might be the important part
that brings the machine in the necessary state for the corruptions to occur.  This
memory increase should be easy to reproduce by sending a "ceph osd down x" to some
of the OSDs  This has always the described result in my installation.
 
I will check if this actually works on my side also. But that will take some time.
 
 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux