Re: needed lru_add_drain_all() change

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



(2012/06/29 8:42), Minchan Kim wrote:
On 06/28/2012 04:43 PM, Kamezawa Hiroyuki wrote:

(2012/06/27 6:37), Andrew Morton wrote:
https://bugzilla.kernel.org/show_bug.cgi?id=43811

lru_add_drain_all() uses schedule_on_each_cpu().  But
schedule_on_each_cpu() hangs if a realtime thread is spinning, pinned
to a CPU.  There's no intention to change the scheduler behaviour, so I
think we should remove schedule_on_each_cpu() from the kernel.

The biggest user of schedule_on_each_cpu() is lru_add_drain_all().

Does anyone have any thoughts on how we can do this?  The obvious
approach is to declare these:

static DEFINE_PER_CPU(struct pagevec[NR_LRU_LISTS], lru_add_pvecs);
static DEFINE_PER_CPU(struct pagevec, lru_rotate_pvecs);
static DEFINE_PER_CPU(struct pagevec, lru_deactivate_pvecs);

to be irq-safe and use on_each_cpu().  lru_rotate_pvecs is already
irq-safe and converting lru_add_pvecs and lru_deactivate_pvecs looks
pretty simple.

Thoughts?


How about this kind of RCU synchronization ?
==
/*
  * Double buffered pagevec for quick drain.
  * The usual per-cpu-pvec user need to take rcu_read_lock() before
accessing.
  * External drainer of pvecs will relpace pvec vector and call
synchroize_rcu(),
  * and drain all pages on unused pvecs in turn.
  */
static DEFINE_PER_CPU(struct pagevec[NR_LRU_LISTS * 2], lru_pvecs);

atomic_t pvec_idx; /* must be placed onto some aligned address...*/


struct pagevec *my_pagevec(enum lru)
{
     return  pvec = &__get_cpu_var(lru_pvecs[lru << atomic_read(pvec_idx)]);
}

/*
  * percpu pagevec access should be surrounded by these calls.
  */
static inline void pagevec_start_access()
{
     rcu_read_lock();
}

static inline void pagevec_end_access()
{
     rcu_read_unlock();
}


/*
  * changing pagevec array vec 0 <-> 1
  */
static void lru_pvec_update()
{
     if (atomic_read(&pvec_idx))
         atomic_set(&pvec_idx, 0);
     else
         atomic_set(&pvec_idx, 1);
}

/*
  * drain all LRUS on per-cpu pagevecs.
  */
DEFINE_MUTEX(lru_add_drain_all_mutex);
static void lru_add_drain_all()
{
     mutex_lock(&lru_add_drain_mutex);
     lru_pvec_update();
     synchronize_rcu();  /* waits for all accessors to pvec quits. */


I don't know RCU internal but conceptually, I understood synchronize_rcu need
context switching of all CPU. If it's partly true, it could be a problem, too.


Hmm, from Documenatation/RCU/stallwarn.txt
==

o       For !CONFIG_PREEMPT kernels, a CPU looping anywhere in the kernel
        without invoking schedule().

o       A CPU-bound real-time task in a CONFIG_PREEMPT kernel, which might
        happen to preempt a low-priority task in the middle of an RCU
        read-side critical section.   This is especially damaging if
        that low-priority task is not permitted to run on any other CPU,
        in which case the next RCU grace period can never complete, which
        will eventually cause the system to run out of memory and hang.
        While the system is in the process of running itself out of
        memory, you might see stall-warning messages.

o       A CPU-bound real-time task in a CONFIG_PREEMPT_RT kernel that
        is running at a higher priority than the RCU softirq threads.
        This will prevent RCU callbacks from ever being invoked,
        and in a CONFIG_TREE_PREEMPT_RCU kernel will further prevent
        RCU grace periods from ever completing.  Either way, the
        system will eventually run out of memory and hang.  In the
        CONFIG_TREE_PREEMPT_RCU case, you might see stall-warning
        messages.
==
you're right. (RCU stall warning seems to be shown per 60secs at default.)

I'm wondering to do sync without RCU...
==
pvec_start_access(struct pagevec *pvec)
{
	atomic_inc(&pvec->using);
}

pvec_end_access(struct pagevec *pvec)
{
	atomic_dec(&pvec->using);
}

synchronize_pvec()
{
	for_each_cpu(cpu)
		wait for pvec->using to be 0.
}

static void lru_add_drain_all()
{
	mutex_lock();
	lru_pvec_update(); //switch pvec
	synchronize_pvec(); // wait for all user exits
	for_each_cpu()
		drain pages in pvec
	mutex_unlock()
}
==

"disable_irq() + intterupt()" will be easier.

What is the cost of IRQ-disable v.s. atomic_inc() for local variable...

Regards,
-Kame











--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@xxxxxxxxx.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@xxxxxxxxx";> email@xxxxxxxxx </a>


[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [ECOS]     [Asterisk Internet PBX]     [Linux API]