On 09/22/2016 04:08 PM, Waiman Long wrote:
On 09/22/2016 11:11 AM, Davidlohr Bueso wrote:
On Thu, 22 Sep 2016, Thomas Gleixner wrote:
On Thu, 22 Sep 2016, Davidlohr Bueso wrote:
On Thu, 22 Sep 2016, Thomas Gleixner wrote:
> Also what's the reason that we can't do probabilistic spinning for
> FUTEX_WAIT and have to add yet another specialized variant of
futexes?
Where would this leave the respective FUTEX_WAKE? A nop? Probably
have to
differentiate the fact that the queue was empty, but there was a
spinning,
instead of straightforward returning 0.
Sorry, but I really can't parse this answer.
Can you folks please communicate with proper and coherent explanations
instead of throwing a few gnawed off bones in my direction?
I actually think that FUTEX_WAIT is the better/nicer approach. But my
immediate
question above was how to handle the FUTEX_WAKE counter-part. If we
want to
maintain current FIFO ordering for wakeups, now with WAIT spinners
this will
create lock stealing scenarios (including if we even guard against
starvation).
Or we could reduce the scope of spinners, due to the restrictions,
similar to
the top-waiter only being able to spin for rtmutexes. This of course
will hurt
the effectiveness of spinning in FUTEX_WAIT in the first place.
Actually, there can be a lot of lock stealing going on with the
wait-wake futexes. If the critical section is short enough, many of
the lock waiters can be waiting in the hash bucket spinlock queue and
not sleeping yet while the futex value changes. As a result, they will
exit the futex syscall and back to user space with EAGAIN where one of
them may get the lock. So we can't assume that they will get the lock
in the FIFO order anyway.
BTW, my initial attempt for the new futex was to use the same workflow
as the PI futexes, but use mutex which has optimistic spinning instead
of rt_mutex. That version can double the throughput compared with PI
futexes but still far short of what can be achieved with wait-wake
futex. Looking at the performance figures from the patch:
wait-wake futex PI futex TO futex
--------------- -------- --------
max time 3.49s 50.91s 2.65s
min time 3.24s 50.84s 0.07s
average time 3.41s 50.90s 1.84s
sys time 7m22.4s 55.73s 2m32.9s
lock count 3,090,294 9,999,813 698,318
unlock count 3,268,896 9,999,814 134
The problem with a PI futexes like version is that almost all the
lock/unlock operations were done in the kernel which added overhead and
latency. Now looking at the numbers for the TO futexes, less than 1/10
of the lock operations were done in the kernel, the number of unlock was
insignificant. Locking was done mostly by lock stealing. This is where
most of the performance benefit comes from, not optimistic spinning.
This is also the reason that a lock handoff mechanism is implemented to
prevent lock starvation which is likely to happen without one.
Cheers,
Longman
--
To unsubscribe from this list: send the line "unsubscribe linux-doc" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html