Re: [PATCHv2] KVM: optimize apic interrupt delivery

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, Nov 28, 2013 at 10:55:06AM +0200, Gleb Natapov wrote:
> On Wed, Nov 27, 2013 at 09:06:36AM -0800, Paul E. McKenney wrote:
> > On Wed, Nov 27, 2013 at 10:00:09AM +0200, Gleb Natapov wrote:
> > > On Tue, Nov 26, 2013 at 11:35:06AM -0800, Paul E. McKenney wrote:
> > > > On Tue, Nov 26, 2013 at 06:24:13PM +0200, Michael S. Tsirkin wrote:
> > > > > On Wed, Sep 12, 2012 at 08:13:54AM -0700, Paul E. McKenney wrote:
> > > > > > On Wed, Sep 12, 2012 at 03:44:26PM +0300, Gleb Natapov wrote:
> > > > > > > On Wed, Sep 12, 2012 at 03:36:57PM +0300, Avi Kivity wrote:
> > > > > > > > On 09/12/2012 03:34 PM, Gleb Natapov wrote:
> > > > > > > > > On Wed, Sep 12, 2012 at 10:45:22AM +0300, Avi Kivity wrote:
> > > > > > > > >> On 09/12/2012 04:03 AM, Paul E. McKenney wrote:
> > > > > > > > >> >> > > Paul, I'd like to check something with you here:
> > > > > > > > >> >> > > this function can be triggered by userspace,
> > > > > > > > >> >> > > any number of times; we allocate
> > > > > > > > >> >> > > a 2K chunk of memory that is later freed by
> > > > > > > > >> >> > > kfree_rcu.
> > > > > > > > >> >> > > 
> > > > > > > > >> >> > > Is there a risk of DOS if RCU is delayed while
> > > > > > > > >> >> > > lots of memory is queued up in this way?
> > > > > > > > >> >> > > If yes is this a generic problem with kfree_rcu
> > > > > > > > >> >> > > that should be addressed in core kernel?
> > > > > > > > >> >> > 
> > > > > > > > >> >> > There is indeed a risk.
> > > > > > > > >> >> 
> > > > > > > > >> >> In our case it's a 2K object. Is it a practical risk?
> > > > > > > > >> > 
> > > > > > > > >> > How many kfree_rcu()s per second can a given user cause to happen?
> > > > > > > > >> 
> > > > > > > > >> Not much more than a few hundred thousand per second per process (normal
> > > > > > > > >> operation is zero).
> > > > > > > > >> 
> > > > > > > > > I managed to do 21466 per second.
> > > > > > > > 
> > > > > > > > Strange, why so slow?
> > > > > > > > 
> > > > > > > Because ftrace buffer overflows :) With bigger buffer I get 169940.
> > > > > > 
> > > > > > Ah, good, should not be a problem.  In contrast, if you ran kfree_rcu() in
> > > > > > a tight loop, you could probably do in excess of 100M per CPU per second.
> > > > > > Now -that- might be a problem.
> > > > > > 
> > > > > > Well, it -might- be a problem if you somehow figured out how to allocate
> > > > > > memory that quickly in a steady-state manner.  ;-)
> > > > > > 
> > > > > > > > >> Good idea.  Michael, is should be easy to modify kvm-unit-tests to write
> > > > > > > > >> to the APIC ID register in a loop.
> > > > > > > > >> 
> > > > > > > > > I did. Memory consumption does not grow on otherwise idle host.
> > > > > > 
> > > > > > Very good -- the checks in __call_rcu(), which is common code invoked by
> > > > > > kfree_rcu(), seem to be doing their job, then.  These do keep a per-CPU
> > > > > > counter, which can be adjusted via rcutree.blimit, which defaults
> > > > > > to taking evasive action if more than 10K callbacks are waiting on a
> > > > > > given CPU.
> > > > > > 
> > > > > > My concern was that you might be overrunning that limit in way less
> > > > > > than a grace period (as in about a hundred microseconds.  My concern
> > > > > > was of course unfounded -- you take several grace periods in push 10K
> > > > > > callbacks through.
> > > > > > 
> > > > > > 							Thanx, Paul
> > > > > 
> > > > > Gleb noted that Documentation/RCU/checklist.txt has this text:
> > > > > 
> > > > >         An especially important property of the synchronize_rcu()
> > > > >         primitive is that it automatically self-limits: if grace periods
> > > > >         are delayed for whatever reason, then the synchronize_rcu()
> > > > >         primitive will correspondingly delay updates.  In contrast,
> > > > >         code using call_rcu() should explicitly limit update rate in
> > > > >         cases where grace periods are delayed, as failing to do so can
> > > > >         result in excessive realtime latencies or even OOM conditions.
> > > > > 
> > > > > If call_rcu is self-limiting maybe this should be documented ...
> > > > 
> > > > It would be more accurate to say that takes has some measures to limit
> > > > the damage -- you can overwhelm these measures if you try hard enough.
> > > > 
> > > The question is: Is it safe to have a call_rcu() without any additional rate limiting
> > > on user triggerable code path?
> > 
> > That would be a good way to allow users to run your system out of memory,
> > especially on systems with limited memory.  (If you have several GB of
> > free space, you might be OK.)
> > 
> Thanks! Got it.

Does the following help?

							Thanx, Paul

------------------------------------------------------------------------

rcu: Document call_rcu() safety mechanisms and limitations

The call_rcu() family of primitives will take action to accelerate
grace periods when the number of callbacks pending on a given CPU
becomes excessive.  Although this safety mechanism can be useful,
it is no substitute for users of call_rcu() having rate-limit controls
in place.  This commit adds this nuance to the documentation.

Reported-by: "Michael S. Tsirkin" <mst@xxxxxxxxxx>
Reported-by: Gleb Natapov <gleb@xxxxxxxxxx>
Signed-off-by: Paul E. McKenney <paulmck@xxxxxxxxxxxxxxxxxx>

diff --git a/Documentation/RCU/checklist.txt b/Documentation/RCU/checklist.txt
index 91266193b8f4..5733e31836b5 100644
--- a/Documentation/RCU/checklist.txt
+++ b/Documentation/RCU/checklist.txt
@@ -256,10 +256,11 @@ over a rather long period of time, but improvements are always welcome!
 		variations on this theme.
 
 	b.	Limiting update rate.  For example, if updates occur only
-		once per hour, then no explicit rate limiting is required,
-		unless your system is already badly broken.  The dcache
-		subsystem takes this approach -- updates are guarded
-		by a global lock, limiting their rate.
+		once per hour, then no explicit rate limiting is
+		required, unless your system is already badly broken.
+		Older versions of the dcache subsystem takes this
+		approach -- updates were guarded by a global lock,
+		limiting their rate.
 
 	c.	Trusted update -- if updates can only be done manually by
 		superuser or some other trusted user, then it might not
@@ -268,7 +269,8 @@ over a rather long period of time, but improvements are always welcome!
 		the machine.
 
 	d.	Use call_rcu_bh() rather than call_rcu(), in order to take
-		advantage of call_rcu_bh()'s faster grace periods.
+		advantage of call_rcu_bh()'s faster grace periods.  (This
+		is only a partial solution, though.)
 
 	e.	Periodically invoke synchronize_rcu(), permitting a limited
 		number of updates per grace period.
@@ -276,6 +278,13 @@ over a rather long period of time, but improvements are always welcome!
 	The same cautions apply to call_rcu_bh(), call_rcu_sched(),
 	call_srcu(), and kfree_rcu().
 
+	Note that although these primitives do take action to avoid memory
+	exhaustion when any given CPU has too many callbacks, a determined
+	user could still exhaust memory.  This is especially the case
+	if a system with a large number of CPUs has been configured to
+	offload all of its RCU callbacks onto a single CPU, or if the
+	system has relatively little free memory.
+
 9.	All RCU list-traversal primitives, which include
 	rcu_dereference(), list_for_each_entry_rcu(), and
 	list_for_each_safe_rcu(), must be either within an RCU read-side

--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [KVM ARM]     [KVM ia64]     [KVM ppc]     [Virtualization Tools]     [Spice Development]     [Libvirt]     [Libvirt Users]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite Questions]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux