Le mardi 01 juin 2010 à 19:52 +0300, Avi Kivity a écrit : > What I'd like to see eventually is a short-term-unfair, long-term-fair > spinlock. Might make sense for bare metal as well. But it won't be > easy to write. > This thread rings a bell here :) Yes, ticket spinlocks are sometime slower, especially in workloads where a spinlock needs to be taken several times to handle one unit of work, and many cpus competing. We currently have kind of a similar problem in network stack, and we have a patch to speedup xmit path by an order of magnitude, letting one cpu (the consumer cpu) to get unfair access to the (ticket) spinlock. (It can compete with no more than one other cpu) Boost from ~50.000 to ~600.000 pps on a dual quad core machine (E5450 @3.00GHz) on a particular workload (many cpus want to xmit their packets) ( patch : http://patchwork.ozlabs.org/patch/53163/ ) It could be possible to write such a generic beast, with a cascade or regular ticket spinlocks ? One ticket spinlock at first stage (only if some conditions are met, aka slow path), then an 'primary' spinlock at second stage. // generic implementation // (x86 could use 16bit fields for users_in & user_out) struct cascade_lock { atomic_t users_in; int users_out; spinlock_t primlock; spinlock_t slowpathlock; // could be outside of this structure, shared by many 'cascade_locks' }; /* * In kvm case, you might call hypervisor when slowpathlock is about to be taken ? * When a cascade lock is unlocked, and relocked right after, this cpu has unfair * priority and could get the lock before cpus blocked in slowpathlock (especially if * an hypervisor call was done) * * In network xmit path, the dequeue thread would use highprio_user=true mode * In network xmit path, the 'contended' enqueueing thread would set a negative threshold, * to force a 'lowprio_user' mode. */ void cascade_lock(struct cascade_lock *l, bool highprio_user, int threshold) { bool slowpath = false; atomic_inc(&l->users_in); // no real need for atomic_inc_return() if (atomic_read(&l->users_in) - l->users_out > threshold && !highprio_user)) { spin_lock(&l->slowpathlock); slowpath = true; } spin_lock(&l->primlock); if (slowpath) spin_unlock(&l->slowpathlock); } void cascade_unlock(struct cascade_lock *l) { l->users_out++; spin_unlock(&l->primlock); } -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html