Re: [Ltt-dev] [PATCH 5/10] local_t : MIPS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Sorry, I meant MIPS.

* Mathieu Desnoyers (mathieu.desnoyers@xxxxxxxxxx) wrote:
> ia64 architecture local_t extension.
> 
> Signed-off-by: Mathieu Desnoyers <mathieu.desnoyers@xxxxxxxxxx>
> 
> --- a/include/asm-mips/system.h
> +++ b/include/asm-mips/system.h
> @@ -253,6 +253,58 @@ static inline unsigned long __cmpxchg_u3
>  	return retval;
>  }
>  
> +static inline unsigned long __cmpxchg_u32_local(volatile int * m,
> +	unsigned long old, unsigned long new)
> +{
> +	__u32 retval;
> +
> +	if (cpu_has_llsc && R10000_LLSC_WAR) {
> +		__asm__ __volatile__(
> +		"	.set	push					\n"
> +		"	.set	noat					\n"
> +		"	.set	mips3					\n"
> +		"1:	ll	%0, %2			# __cmpxchg_u32	\n"
> +		"	bne	%0, %z3, 2f				\n"
> +		"	.set	mips0					\n"
> +		"	move	$1, %z4					\n"
> +		"	.set	mips3					\n"
> +		"	sc	$1, %1					\n"
> +		"	beqzl	$1, 1b					\n"
> +		"2:							\n"
> +		"	.set	pop					\n"
> +		: "=&r" (retval), "=R" (*m)
> +		: "R" (*m), "Jr" (old), "Jr" (new)
> +		: "memory");
> +	} else if (cpu_has_llsc) {
> +		__asm__ __volatile__(
> +		"	.set	push					\n"
> +		"	.set	noat					\n"
> +		"	.set	mips3					\n"
> +		"1:	ll	%0, %2			# __cmpxchg_u32	\n"
> +		"	bne	%0, %z3, 2f				\n"
> +		"	.set	mips0					\n"
> +		"	move	$1, %z4					\n"
> +		"	.set	mips3					\n"
> +		"	sc	$1, %1					\n"
> +		"	beqz	$1, 1b					\n"
> +		"2:							\n"
> +		"	.set	pop					\n"
> +		: "=&r" (retval), "=R" (*m)
> +		: "R" (*m), "Jr" (old), "Jr" (new)
> +		: "memory");
> +	} else {
> +		unsigned long flags;
> +
> +		local_irq_save(flags);
> +		retval = *m;
> +		if (retval == old)
> +			*m = new;
> +		local_irq_restore(flags);	/* implies memory barrier  */
> +	}
> +
> +	return retval;
> +}
> +
>  #ifdef CONFIG_64BIT
>  static inline unsigned long __cmpxchg_u64(volatile int * m, unsigned long old,
>  	unsigned long new)
> @@ -303,10 +355,62 @@ static inline unsigned long __cmpxchg_u6
>  
>  	return retval;
>  }
> +
> +static inline unsigned long __cmpxchg_u64_local(volatile int * m,
> +	unsigned long old, unsigned long new)
> +{
> +	__u64 retval;
> +
> +	if (cpu_has_llsc && R10000_LLSC_WAR) {
> +		__asm__ __volatile__(
> +		"	.set	push					\n"
> +		"	.set	noat					\n"
> +		"	.set	mips3					\n"
> +		"1:	lld	%0, %2			# __cmpxchg_u64	\n"
> +		"	bne	%0, %z3, 2f				\n"
> +		"	move	$1, %z4					\n"
> +		"	scd	$1, %1					\n"
> +		"	beqzl	$1, 1b					\n"
> +		"2:							\n"
> +		"	.set	pop					\n"
> +		: "=&r" (retval), "=R" (*m)
> +		: "R" (*m), "Jr" (old), "Jr" (new)
> +		: "memory");
> +	} else if (cpu_has_llsc) {
> +		__asm__ __volatile__(
> +		"	.set	push					\n"
> +		"	.set	noat					\n"
> +		"	.set	mips3					\n"
> +		"1:	lld	%0, %2			# __cmpxchg_u64	\n"
> +		"	bne	%0, %z3, 2f				\n"
> +		"	move	$1, %z4					\n"
> +		"	scd	$1, %1					\n"
> +		"	beqz	$1, 1b					\n"
> +		"2:							\n"
> +		"	.set	pop					\n"
> +		: "=&r" (retval), "=R" (*m)
> +		: "R" (*m), "Jr" (old), "Jr" (new)
> +		: "memory");
> +	} else {
> +		unsigned long flags;
> +
> +		local_irq_save(flags);
> +		retval = *m;
> +		if (retval == old)
> +			*m = new;
> +		local_irq_restore(flags);	/* implies memory barrier  */
> +	}
> +
> +	return retval;
> +}
> +
>  #else
>  extern unsigned long __cmpxchg_u64_unsupported_on_32bit_kernels(
>  	volatile int * m, unsigned long old, unsigned long new);
>  #define __cmpxchg_u64 __cmpxchg_u64_unsupported_on_32bit_kernels
> +extern unsigned long __cmpxchg_u64_local_unsupported_on_32bit_kernels(
> +	volatile int * m, unsigned long old, unsigned long new);
> +#define __cmpxchg_u64_local __cmpxchg_u64_local_unsupported_on_32bit_kernels
>  #endif
>  
>  /* This function doesn't exist, so you'll get a linker error
> @@ -326,7 +430,26 @@ static inline unsigned long __cmpxchg(vo
>  	return old;
>  }
>  
> -#define cmpxchg(ptr,old,new) ((__typeof__(*(ptr)))__cmpxchg((ptr), (unsigned long)(old), (unsigned long)(new),sizeof(*(ptr))))
> +static inline unsigned long __cmpxchg_local(volatile void * ptr,
> +	unsigned long old, unsigned long new, int size)
> +{
> +	switch (size) {
> +	case 4:
> +		return __cmpxchg_u32_local(ptr, old, new);
> +	case 8:
> +		return __cmpxchg_u64_local(ptr, old, new);
> +	}
> +	__cmpxchg_called_with_bad_pointer();
> +	return old;
> +}
> +
> +#define cmpxchg(ptr,old,new) \
> +	((__typeof__(*(ptr)))__cmpxchg((ptr), \
> +		(unsigned long)(old), (unsigned long)(new),sizeof(*(ptr))))
> +
> +#define cmpxchg_local(ptr,old,new) \
> +	((__typeof__(*(ptr)))__cmpxchg_local((ptr), \
> +		(unsigned long)(old), (unsigned long)(new),sizeof(*(ptr))))
>  
>  extern void set_handler (unsigned long offset, void *addr, unsigned long len);
>  extern void set_uncached_handler (unsigned long offset, void *addr, unsigned long len);
> --- a/include/asm-mips/local.h
> +++ b/include/asm-mips/local.h
> @@ -1,60 +1,527 @@
> -#ifndef _ASM_LOCAL_H
> -#define _ASM_LOCAL_H
> +#ifndef _ARCH_POWERPC_LOCAL_H
> +#define _ARCH_POWERPC_LOCAL_H
>  
>  #include <linux/percpu.h>
>  #include <asm/atomic.h>
>  
> -#ifdef CONFIG_32BIT
> +typedef struct
> +{
> +	local_long_t a;
> +} local_t;
>  
> -typedef atomic_t local_t;
> +#define LOCAL_INIT(i)	{ local_LONG_INIT(i) }
>  
> -#define LOCAL_INIT(i)	ATOMIC_INIT(i)
> -#define local_read(v)	atomic_read(v)
> -#define local_set(v,i)	atomic_set(v,i)
> +#define local_read(l)	local_long_read(&(l)->a)
> +#define local_set(l,i)	local_long_set(&(l)->a, (i))
>  
> -#define local_inc(v)	atomic_inc(v)
> -#define local_dec(v)	atomic_dec(v)
> -#define local_add(i, v)	atomic_add(i, v)
> -#define local_sub(i, v)	atomic_sub(i, v)
> +#define local_add(i,l)	local_long_add((i),(&(l)->a))
> +#define local_sub(i,l)	local_long_sub((i),(&(l)->a))
> +#define local_inc(l)	local_long_inc(&(l)->a)
> +#define local_dec(l)	local_long_dec(&(l)->a)
>  
> -#endif
>  
> -#ifdef CONFIG_64BIT
> +#ifndef CONFIG_64BITS
>  
> -typedef atomic64_t local_t;
> +/*
> + * Same as above, but return the result value
> + */
> +static __inline__ int local_add_return(int i, local_t * l)
> +{
> +	unsigned long result;
> +
> +	if (cpu_has_llsc && R10000_LLSC_WAR) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	ll	%1, %2		# local_add_return	\n"
> +		"	addu	%0, %1, %3				\n"
> +		"	sc	%0, %2					\n"
> +		"	beqzl	%0, 1b					\n"
> +		"	addu	%0, %1, %3				\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else if (cpu_has_llsc) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	ll	%1, %2		# local_add_return	\n"
> +		"	addu	%0, %1, %3				\n"
> +		"	sc	%0, %2					\n"
> +		"	beqz	%0, 1b					\n"
> +		"	addu	%0, %1, %3				\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else {
> +		unsigned long flags;
> +
> +		local_irq_save(flags);
> +		result = &(l->a.counter);
> +		result += i;
> +		&(l->a.counter) = result;
> +		local_irq_restore(flags);
> +	}
> +
> +	return result;
> +}
> +
> +static __inline__ int local_sub_return(int i, local_t * l)
> +{
> +	unsigned long result;
> +
> +	if (cpu_has_llsc && R10000_LLSC_WAR) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	ll	%1, %2		# local_sub_return	\n"
> +		"	subu	%0, %1, %3				\n"
> +		"	sc	%0, %2					\n"
> +		"	beqzl	%0, 1b					\n"
> +		"	subu	%0, %1, %3				\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else if (cpu_has_llsc) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	ll	%1, %2		# local_sub_return	\n"
> +		"	subu	%0, %1, %3				\n"
> +		"	sc	%0, %2					\n"
> +		"	beqz	%0, 1b					\n"
> +		"	subu	%0, %1, %3				\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else {
> +		unsigned long flags;
> +
> +		local_irq_save(flags);
> +		result = &(l->a.counter);
> +		result -= i;
> +		&(l->a.counter) = result;
> +		local_irq_restore(flags);
> +	}
> +
> +	return result;
> +}
> +
> +/*
> + * local_sub_if_positive - conditionally subtract integer from atomic variable
> + * @i: integer value to subtract
> + * @l: pointer of type local_t
> + *
> + * Atomically test @l and subtract @i if @l is greater or equal than @i.
> + * The function returns the old value of @l minus @i.
> + */
> +static __inline__ int local_sub_if_positive(int i, local_t * l)
> +{
> +	unsigned long result;
> +
> +	if (cpu_has_llsc && R10000_LLSC_WAR) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	ll	%1, %2		# local_sub_if_positive\n"
> +		"	subu	%0, %1, %3				\n"
> +		"	bltz	%0, 1f					\n"
> +		"	sc	%0, %2					\n"
> +		"	.set	noreorder				\n"
> +		"	beqzl	%0, 1b					\n"
> +		"	 subu	%0, %1, %3				\n"
> +		"	.set	reorder					\n"
> +		"1:							\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else if (cpu_has_llsc) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	ll	%1, %2		# local_sub_if_positive\n"
> +		"	subu	%0, %1, %3				\n"
> +		"	bltz	%0, 1f					\n"
> +		"	sc	%0, %2					\n"
> +		"	.set	noreorder				\n"
> +		"	beqz	%0, 1b					\n"
> +		"	 subu	%0, %1, %3				\n"
> +		"	.set	reorder					\n"
> +		"1:							\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else {
> +		unsigned long flags;
> +
> +		local_irq_save(flags);
> +		result = &(l->a.counter);
> +		result -= i;
> +		if (result >= 0)
> +			&(l->a.counter) = result;
> +		local_irq_restore(flags);
> +	}
> +
> +	return result;
> +}
> +
> +#define local_cmpxchg(l, o, n) \
> +	((long)cmpxchg(&((l)->a.counter), (o), (n)))
> +#define local_xchg(l, new) (xchg(&((l)->a.counter), new))
> +
> +/**
> + * local_add_unless - add unless the number is a given value
> + * @l: pointer of type local_t
> + * @a: the amount to add to l...
> + * @u: ...unless l is equal to u.
> + *
> + * Atomically adds @a to @l, so long as it was not @u.
> + * Returns non-zero if @l was not @u, and zero otherwise.
> + */
> +#define local_add_unless(l, a, u)				\
> +({								\
> +	long c, old;						\
> +	c = local_read(l);					\
> +	while (c != (u) && (old = local_cmpxchg((l), c, c + (a))) != c) \
> +		c = old;					\
> +	c != (u);						\
> +})
> +#define local_inc_not_zero(l) local_add_unless((l), 1, 0)
> +
> +#define local_dec_return(l) local_sub_return(1,(l))
> +#define local_inc_return(l) local_add_return(1,(l))
> +
> +/*
> + * local_sub_and_test - subtract value from variable and test result
> + * @i: integer value to subtract
> + * @l: pointer of type local_t
> + *
> + * Atomically subtracts @i from @l and returns
> + * true if the result is zero, or false for all
> + * other cases.
> + */
> +#define local_sub_and_test(i,l) (local_sub_return((i), (l)) == 0)
> +
> +/*
> + * local_inc_and_test - increment and test
> + * @l: pointer of type local_t
> + *
> + * Atomically increments @l by 1
> + * and returns true if the result is zero, or false for all
> + * other cases.
> + */
> +#define local_inc_and_test(l) (local_inc_return(l) == 0)
> +
> +/*
> + * local_dec_and_test - decrement by 1 and test
> + * @l: pointer of type local_t
> + *
> + * Atomically decrements @l by 1 and
> + * returns true if the result is 0, or false for all other
> + * cases.
> + */
> +#define local_dec_and_test(l) (local_sub_return(1, (l)) == 0)
> +
> +/*
> + * local_dec_if_positive - decrement by 1 if old value positive
> + * @l: pointer of type local_t
> + */
> +#define local_dec_if_positive(l)	local_sub_if_positive(1, l)
> +
> +/*
> + * local_add_negative - add and test if negative
> + * @l: pointer of type local_t
> + * @i: integer value to add
> + *
> + * Atomically adds @i to @l and returns true
> + * if the result is negative, or false when
> + * result is greater than or equal to zero.
> + */
> +#define local_add_negative(i,l) (local_add_return(i, (l)) < 0)
> +
> +#else /* CONFIG_64BITS */
>  
> -#define LOCAL_INIT(i)	ATOMIC64_INIT(i)
> -#define local_read(v)	atomic64_read(v)
> -#define local_set(v,i)	atomic64_set(v,i)
> +/*
> + * Same as above, but return the result value
> + */
> +static __inline__ long local_add_return(long i, local_t * l)
> +{
> +	unsigned long result;
> +
> +	if (cpu_has_llsc && R10000_LLSC_WAR) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	lld	%1, %2		# local_add_return	\n"
> +		"	addu	%0, %1, %3				\n"
> +		"	scd	%0, %2					\n"
> +		"	beqzl	%0, 1b					\n"
> +		"	addu	%0, %1, %3				\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else if (cpu_has_llsc) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	lld	%1, %2		# local_add_return	\n"
> +		"	addu	%0, %1, %3				\n"
> +		"	scd	%0, %2					\n"
> +		"	beqz	%0, 1b					\n"
> +		"	addu	%0, %1, %3				\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else {
> +		unsigned long flags;
> +
> +		local_irq_save(flags);
> +		result = &(l->a.counter);
> +		result += i;
> +		&(l->a.counter) = result;
> +		local_irq_restore(flags);
> +	}
>  
> -#define local_inc(v)	atomic64_inc(v)
> -#define local_dec(v)	atomic64_dec(v)
> -#define local_add(i, v)	atomic64_add(i, v)
> -#define local_sub(i, v)	atomic64_sub(i, v)
> +	return result;
> +}
>  
> -#endif
> +static __inline__ long local_sub_return(long i, local_t * l)
> +{
> +	unsigned long result;
>  
> -#define __local_inc(v)		((v)->counter++)
> -#define __local_dec(v)		((v)->counter--)
> -#define __local_add(i,v)	((v)->counter+=(i))
> -#define __local_sub(i,v)	((v)->counter-=(i))
> +	if (cpu_has_llsc && R10000_LLSC_WAR) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	lld	%1, %2		# local_sub_return	\n"
> +		"	subu	%0, %1, %3				\n"
> +		"	scd	%0, %2					\n"
> +		"	beqzl	%0, 1b					\n"
> +		"	subu	%0, %1, %3				\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else if (cpu_has_llsc) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	lld	%1, %2		# local_sub_return	\n"
> +		"	subu	%0, %1, %3				\n"
> +		"	scd	%0, %2					\n"
> +		"	beqz	%0, 1b					\n"
> +		"	subu	%0, %1, %3				\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else {
> +		unsigned long flags;
> +
> +		local_irq_save(flags);
> +		result = &(l->a.counter);
> +		result -= i;
> +		&(l->a.counter) = result;
> +		local_irq_restore(flags);
> +	}
> +
> +	return result;
> +}
>  
>  /*
> - * Use these for per-cpu local_t variables: on some archs they are
> + * local_sub_if_positive - conditionally subtract integer from atomic variable
> + * @i: integer value to subtract
> + * @l: pointer of type local_t
> + *
> + * Atomically test @l and subtract @i if @l is greater or equal than @i.
> + * The function returns the old value of @l minus @i.
> + */
> +static __inline__ long local_sub_if_positive(long i, local_t * l)
> +{
> +	unsigned long result;
> +
> +	if (cpu_has_llsc && R10000_LLSC_WAR) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	lld	%1, %2		# local_sub_if_positive\n"
> +		"	dsubu	%0, %1, %3				\n"
> +		"	bltz	%0, 1f					\n"
> +		"	scd	%0, %2					\n"
> +		"	.set	noreorder				\n"
> +		"	beqzl	%0, 1b					\n"
> +		"	 dsubu	%0, %1, %3				\n"
> +		"	.set	reorder					\n"
> +		"1:							\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else if (cpu_has_llsc) {
> +		unsigned long temp;
> +
> +		__asm__ __volatile__(
> +		"	.set	mips3					\n"
> +		"1:	lld	%1, %2		# local_sub_if_positive\n"
> +		"	dsubu	%0, %1, %3				\n"
> +		"	bltz	%0, 1f					\n"
> +		"	scd	%0, %2					\n"
> +		"	.set	noreorder				\n"
> +		"	beqz	%0, 1b					\n"
> +		"	 dsubu	%0, %1, %3				\n"
> +		"	.set	reorder					\n"
> +		"1:							\n"
> +		"	.set	mips0					\n"
> +		: "=&r" (result), "=&r" (temp), "=m" (&(l->a.counter))
> +		: "Ir" (i), "m" (&(l->a.counter))
> +		: "memory");
> +	} else {
> +		unsigned long flags;
> +
> +		local_irq_save(flags);
> +		result = &(l->a.counter);
> +		result -= i;
> +		if (result >= 0)
> +			&(l->a.counter) = result;
> +		local_irq_restore(flags);
> +	}
> +
> +	return result;
> +}
> +
> +
> +#define local_cmpxchg(l, o, n) \
> +	((long)cmpxchg(&((l)->a.counter), (o), (n)))
> +#define local_xchg(l, new) (xchg(&((l)->a.counter), new))
> +
> +/**
> + * local_add_unless - add unless the number is a given value
> + * @l: pointer of type local_t
> + * @a: the amount to add to l...
> + * @u: ...unless l is equal to u.
> + *
> + * Atomically adds @a to @l, so long as it was not @u.
> + * Returns non-zero if @l was not @u, and zero otherwise.
> + */
> +#define local_add_unless(l, a, u)				\
> +({								\
> +	long c, old;						\
> +	c = local_read(l);					\
> +	while (c != (u) && (old = local_cmpxchg((l), c, c + (a))) != c) \
> +		c = old;					\
> +	c != (u);						\
> +})
> +#define local_inc_not_zero(l) local_add_unless((l), 1, 0)
> +
> +#define local_dec_return(l) local_sub_return(1,(l))
> +#define local_inc_return(l) local_add_return(1,(l))
> +
> +/*
> + * local_sub_and_test - subtract value from variable and test result
> + * @i: integer value to subtract
> + * @l: pointer of type local_t
> + *
> + * Atomically subtracts @i from @l and returns
> + * true if the result is zero, or false for all
> + * other cases.
> + */
> +#define local_sub_and_test(i,l) (local_sub_return((i), (l)) == 0)
> +
> +/*
> + * local_inc_and_test - increment and test
> + * @l: pointer of type local_t
> + *
> + * Atomically increments @l by 1
> + * and returns true if the result is zero, or false for all
> + * other cases.
> + */
> +#define local_inc_and_test(l) (local_inc_return(l) == 0)
> +
> +/*
> + * local_dec_and_test - decrement by 1 and test
> + * @l: pointer of type local_t
> + *
> + * Atomically decrements @l by 1 and
> + * returns true if the result is 0, or false for all other
> + * cases.
> + */
> +#define local_dec_and_test(l) (local_sub_return(1, (l)) == 0)
> +
> +/*
> + * local_dec_if_positive - decrement by 1 if old value positive
> + * @l: pointer of type local_t
> + */
> +#define local_dec_if_positive(l)	local_sub_if_positive(1, l)
> +
> +/*
> + * local_add_negative - add and test if negative
> + * @l: pointer of type local_t
> + * @i: integer value to add
> + *
> + * Atomically adds @i to @l and returns true
> + * if the result is negative, or false when
> + * result is greater than or equal to zero.
> + */
> +#define local_add_negative(i,l) (local_add_return(i, (l)) < 0)
> +
> +#endif /* !CONFIG_64BITS */
> +
> +
> +/* Use these for per-cpu local_t variables: on some archs they are
>   * much more efficient than these naive implementations.  Note they take
>   * a variable, not an address.
> + *
> + * This could be done better if we moved the per cpu data directly
> + * after GS.
>   */
> -#define cpu_local_read(v)	local_read(&__get_cpu_var(v))
> -#define cpu_local_set(v, i)	local_set(&__get_cpu_var(v), (i))
>  
> -#define cpu_local_inc(v)	local_inc(&__get_cpu_var(v))
> -#define cpu_local_dec(v)	local_dec(&__get_cpu_var(v))
> -#define cpu_local_add(i, v)	local_add((i), &__get_cpu_var(v))
> -#define cpu_local_sub(i, v)	local_sub((i), &__get_cpu_var(v))
> +#define __local_inc(l)		((l)->a.counter++)
> +#define __local_dec(l)		((l)->a.counter++)
> +#define __local_add(i,l)	((l)->a.counter+=(i))
> +#define __local_sub(i,l)	((l)->a.counter-=(i))
> +
> +/* Need to disable preemption for the cpu local counters otherwise we could
> +   still access a variable of a previous CPU in a non atomic way. */
> +#define cpu_local_wrap_v(l)	 	\
> +	({ local_t res__;		\
> +	   preempt_disable(); 		\
> +	   res__ = (l);			\
> +	   preempt_enable();		\
> +	   res__; })
> +#define cpu_local_wrap(l)		\
> +	({ preempt_disable();		\
> +	   l;				\
> +	   preempt_enable(); })		\
> +
> +#define cpu_local_read(l)    cpu_local_wrap_v(local_read(&__get_cpu_var(l)))
> +#define cpu_local_set(l, i)  cpu_local_wrap(local_set(&__get_cpu_var(l), (i)))
> +#define cpu_local_inc(l)     cpu_local_wrap(local_inc(&__get_cpu_var(l)))
> +#define cpu_local_dec(l)     cpu_local_wrap(local_dec(&__get_cpu_var(l)))
> +#define cpu_local_add(i, l)  cpu_local_wrap(local_add((i), &__get_cpu_var(l)))
> +#define cpu_local_sub(i, l)  cpu_local_wrap(local_sub((i), &__get_cpu_var(l)))
>  
> -#define __cpu_local_inc(v)	__local_inc(&__get_cpu_var(v))
> -#define __cpu_local_dec(v)	__local_dec(&__get_cpu_var(v))
> -#define __cpu_local_add(i, v)	__local_add((i), &__get_cpu_var(v))
> -#define __cpu_local_sub(i, v)	__local_sub((i), &__get_cpu_var(v))
> +#define __cpu_local_inc(l)	cpu_local_inc(l)
> +#define __cpu_local_dec(l)	cpu_local_dec(l)
> +#define __cpu_local_add(i, l)	cpu_local_add((i), (l))
> +#define __cpu_local_sub(i, l)	cpu_local_sub((i), (l))
>  
> -#endif /* _ASM_LOCAL_H */
> +#endif /* _ARCH_POWERPC_LOCAL_H */
> 

-- 
OpenPGP public key:              http://krystal.dyndns.org:8080/key/compudj.gpg
Key fingerprint:     8CD5 52C3 8E3C 4140 715F  BA06 3F25 A8FE 3BAE 9A68 


[Index of Archives]     [Linux MIPS Home]     [LKML Archive]     [Linux ARM Kernel]     [Linux ARM]     [Linux]     [Git]     [Yosemite News]     [Linux SCSI]     [Linux Hams]

  Powered by Linux