Santosh Shilimkar <santosh.shilimkar@xxxxxx> writes: > On Thursday 04 April 2013 02:55 AM, Kevin Hilman wrote: >> Santosh Shilimkar <santosh.shilimkar@xxxxxx> writes: >> >>> The OMAP5 idle driver can re-use OMAP4 CPUidle driver implementation thanks >>> to compatible MPUSS design. >>> >>> Though unlike OMAP4, on OMAP5 devices, MPUSS CSWR (Close Switch >>> Retention) power states can be achieved with respective power states >>> on CPU0 and CPU1 power domain. This mode was broken on OMAP4 devices >>> because of hardware limitation. >> >> I'm a bit confused by the description here. >> >> I gather from the code that this means that CPU0 and CPU1 can hit CSWR >> independently, correct? >> > They can be programmed independently without any ordering(like > CPU0 last etc), but the actual transition to the low power CSWR > happens together. Till that the first CPU hit WFI remains in WFI > state waiting for other CPU to hit WFI and then both transition > together. > Completely managed inside hardware. OK, elaborating this in the changelog would be helpful. Use the "will I understand this changelog in a year" rule to see if the changelog is detailed enough. Or better, "will Kevin understand this changelog in a year." (hint: the answer is usually no.) ;) >>> Also there is no CPU low power entry order requirement like >>> master CPU etc for CSWR C-state, which is icing on the cake. >> >> Even on secure devices? >> > Yes. On secure devices too. Actually since we don't loose context, > secure entry/exit doesn't come into picture. > >>> Code makes use of voting scheme for cluster low power state to support >>> MPUSS CSWR C-state. >> >> The voting scheme and associated locking should be better described >> here, or commented in the code itself. >> > You are right. It deserves some description. > >>> Supported OMAP5 CPUidle C-states: >>> C1 - CPU0 ON(WFI) + CPU1 ON(WFI) + MPUSS ON >>> C2 - CPU0 CSWR + CPU1 CSWR + MPUSS CSWR >>> C3 - CPU0 OFF + CPU1 OFF + MPUSS OSWR >>> >>> Acked-by: Nishanth Menon <nm@xxxxxx> >>> Signed-off-by: Santosh Shilimkar <santosh.shilimkar@xxxxxx> >> >> [...] >> >>> +static int omap_enter_idle_smp(struct cpuidle_device *dev, >>> + struct cpuidle_driver *drv, >>> + int index) >>> +{ >>> + struct idle_statedata *cx = state_ptr + index; >>> + int cpu_id = smp_processor_id(); >>> + unsigned long flag; >>> + >>> + clockevents_notify(CLOCK_EVT_NOTIFY_BROADCAST_ENTER, &cpu_id); >> >> I think the CPUidle core handles the broadcast notification now. >> > Not in mainline yet. And those patches came after my patches and > I don't wanted un-necessary merge dependency, I avoided it. Its trivial > though to drop if from here once the idle next is merged. OK. I believe that stuff is already queued, no? Maybe ahave this as an add-on separate patch that can be used for your loal testing, but does not go upstream. I would only include this if you're sure the other series is not going upstream. >>> + raw_spin_lock_irqsave(&mpu_lock, flag); >>> + cx->mpu_state_vote++; >> >> How about using an atomic_t and atomic_inc()/atomic_dec() instead, which >> will avoid the need for a spinlock. >> > Spin lock is not just for the vote variable. I had atomics opps in first > version I gave it to product team. But they found a race condition in > where MPU power state was getting overwritten by other CPU. > >> Even with that, it still seems potentially racy. Do you need a memory >> barrier here to be sure that any changes from another CPU are visible >> here, and vice versa? >> > With locks, you don't need barriers since the updated copy is guaranteed. It's guaranteed because the spinlock implementation uses barriers. > Can you please elaborate on potential race ? I have given pretty hard > thought and didn't see any race which can be exposed with locks in place. I was referring to using atomic ops. With atomic ops, you'd need an explicit barrier (which you're getting inside the spinlock implementation) That being said, I have not thought through all the corner cases as you have, so I'll defer to your choice (as long as it's well documented.) If you decide to stick with spinlocks, be sure to describe all of what the spinlock is protecting, and why. Thanks, Kevin -- To unsubscribe from this list: send the line "unsubscribe linux-omap" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html