Change the index to unsigned long in all bitops for [m32r] Signed-off-by: Justin Chen <justin.chen@xxxxxx> Reviewed-by: Bjorn Helgaas <bjorn.helgaas@xxxxxx> --- include/asm-m32r/bitops.h | 14 ++++++++------ 1 file changed, 8 insertions(+), 6 deletions(-) diff -Nru a/include/asm-m32r/bitops.h b/include/asm-m32r/bitops.h --- a/include/asm-m32r/bitops.h 2009-02-13 15:31:30.000000000 -0800 +++ b/include/asm-m32r/bitops.h 2009-02-15 18:19:39.793041022 -0800 @@ -39,7 +39,7 @@ * Note that @nr may be almost arbitrarily large; this function is not * restricted to acting on a single-word quantity. */ -static __inline__ void set_bit(int nr, volatile void * addr) +static __inline__ void set_bit(unsigned long nr, volatile void *addr) { __u32 mask; volatile __u32 *a = addr; @@ -75,7 +75,7 @@ * you should call smp_mb__before_clear_bit() and/or smp_mb__after_clear_bit() * in order to ensure changes are visible on other processors. */ -static __inline__ void clear_bit(int nr, volatile void * addr) +static __inline__ void clear_bit(unsigned long nr, volatile void *addr) { __u32 mask; volatile __u32 *a = addr; @@ -114,7 +114,7 @@ * Note that @nr may be almost arbitrarily large; this function is not * restricted to acting on a single-word quantity. */ -static __inline__ void change_bit(int nr, volatile void * addr) +static __inline__ void change_bit(unsigned long nr, volatile void *addr) { __u32 mask; volatile __u32 *a = addr; @@ -148,7 +148,7 @@ * This operation is atomic and cannot be reordered. * It also implies a memory barrier. */ -static __inline__ int test_and_set_bit(int nr, volatile void * addr) +static __inline__ int test_and_set_bit(unsigned long nr, volatile void *addr) { __u32 mask, oldbit; volatile __u32 *a = addr; @@ -183,7 +183,8 @@ * This operation is atomic and cannot be reordered. * It also implies a memory barrier. */ -static __inline__ int test_and_clear_bit(int nr, volatile void * addr) +static __inline__ int +test_and_clear_bit(unsigned long nr, volatile void *addr) { __u32 mask, oldbit; volatile __u32 *a = addr; @@ -220,7 +221,8 @@ * This operation is atomic and cannot be reordered. * It also implies a memory barrier. */ -static __inline__ int test_and_change_bit(int nr, volatile void * addr) +static __inline__ int +test_and_change_bit(unsigned long nr, volatile void *addr) { __u32 mask, oldbit; volatile __u32 *a = addr; -- To unsubscribe from this list: send the line "unsubscribe linux-arch" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html