|
@@ -0,0 +1,165 @@
|
|
|
|
+--- /dev/null 2008-09-18 06:50:54.356228028 -0700
|
|
|
|
++++ uClibc-0.9.29/libc/sysdeps/linux/xtensa/bits/atomic.h 2008-10-04 11:40:21.000000000 -0700
|
|
|
|
+@@ -0,0 +1,162 @@
|
|
|
|
++/*
|
|
|
|
++ * Copyright (C) 2008 Tensilica, Inc.
|
|
|
|
++ * Contributed by Joe Taylor <joe@tensilica.com> 2008
|
|
|
|
++ *
|
|
|
|
++ * This file is subject to the terms and conditions of the GNU Lesser General
|
|
|
|
++ * Public License. See the file "COPYING.LIB" in the main directory of this
|
|
|
|
++ * archive for more details.
|
|
|
|
++ */
|
|
|
|
++
|
|
|
|
++#ifndef _XTENSA_BITS_ATOMIC_H
|
|
|
|
++#define _XTENSA_BITS_ATOMIC_H 1
|
|
|
|
++
|
|
|
|
++/* Xtensa has only a 32-bit form of a store-conditional instruction,
|
|
|
|
++ so just stub out the rest. */
|
|
|
|
++
|
|
|
|
++/* Atomically store NEWVAL in *MEM if *MEM is equal to OLDVAL.
|
|
|
|
++ Return the old *MEM value. */
|
|
|
|
++
|
|
|
|
++#define __arch_compare_and_exchange_val_32_acq(mem, newval, oldval) \
|
|
|
|
++ ({__typeof__(*(mem)) __tmp, __value; \
|
|
|
|
++ __asm__ __volatile__( \
|
|
|
|
++ "1: l32i %1, %2, 0 \n" \
|
|
|
|
++ " bne %1, %4, 2f \n" \
|
|
|
|
++ " wsr %1, SCOMPARE1 \n" \
|
|
|
|
++ " mov %0, %1 \n" \
|
|
|
|
++ " mov %1, %3 \n" \
|
|
|
|
++ " s32c1i %1, %2, 0 \n" \
|
|
|
|
++ " bne %0, %1, 1b \n" \
|
|
|
|
++ "2: \n" \
|
|
|
|
++ : "=&a" (__value), "=&a" (__tmp) \
|
|
|
|
++ : "a" (mem), "a" (newval), "a" (oldval) \
|
|
|
|
++ : "memory" ); \
|
|
|
|
++ __tmp; \
|
|
|
|
++ })
|
|
|
|
++
|
|
|
|
++/* Atomically store NEWVAL in *MEM if *MEM is equal to OLDVAL.
|
|
|
|
++ Return zero if *MEM was changed or non-zero if no exchange happened. */
|
|
|
|
++
|
|
|
|
++#define __arch_compare_and_exchange_bool_32_acq(mem, newval, oldval) \
|
|
|
|
++ ({__typeof__(*(mem)) __tmp, __value; \
|
|
|
|
++ __asm__ __volatile__( \
|
|
|
|
++ "1: l32i %0, %2, 0 \n" \
|
|
|
|
++ " sub %1, %4, %0 \n" \
|
|
|
|
++ " bnez %1, 2f \n" \
|
|
|
|
++ " wsr %0, SCOMPARE1 \n" \
|
|
|
|
++ " mov %1, %3 \n" \
|
|
|
|
++ " s32c1i %1, %2, 0 \n" \
|
|
|
|
++ " bne %0, %1, 1b \n" \
|
|
|
|
++ " movi %1, 0 \n" \
|
|
|
|
++ "2: \n" \
|
|
|
|
++ : "=&a" (__value), "=&a" (__tmp) \
|
|
|
|
++ : "a" (mem), "a" (newval), "a" (oldval) \
|
|
|
|
++ : "memory" ); \
|
|
|
|
++ __tmp != 0; \
|
|
|
|
++ })
|
|
|
|
++
|
|
|
|
++/* Store NEWVALUE in *MEM and return the old value. */
|
|
|
|
++
|
|
|
|
++#define __arch_exchange_32_acq(mem, newval) \
|
|
|
|
++ ({__typeof__(*(mem)) __tmp, __value; \
|
|
|
|
++ __asm__ __volatile__( \
|
|
|
|
++ "1: l32i %0, %2, 0 \n" \
|
|
|
|
++ " wsr %0, SCOMPARE1 \n" \
|
|
|
|
++ " mov %1, %3 \n" \
|
|
|
|
++ " s32c1i %1, %2, 0 \n" \
|
|
|
|
++ " bne %0, %1, 1b \n" \
|
|
|
|
++ : "=&a" (__value), "=&a" (__tmp) \
|
|
|
|
++ : "a" (mem), "a" (newval) \
|
|
|
|
++ : "memory" ); \
|
|
|
|
++ __tmp; \
|
|
|
|
++ })
|
|
|
|
++
|
|
|
|
++/* Add VALUE to *MEM and return the old value of *MEM. */
|
|
|
|
++
|
|
|
|
++#define __arch_atomic_exchange_and_add_32(mem, value) \
|
|
|
|
++ ({__typeof__(*(mem)) __tmp, __value; \
|
|
|
|
++ __asm__ __volatile__( \
|
|
|
|
++ "1: l32i %0, %2, 0 \n" \
|
|
|
|
++ " wsr %0, SCOMPARE1 \n" \
|
|
|
|
++ " add %1, %0, %3 \n" \
|
|
|
|
++ " s32c1i %1, %2, 0 \n" \
|
|
|
|
++ " bne %0, %1, 1b \n" \
|
|
|
|
++ : "=&a" (__value), "=&a" (__tmp) \
|
|
|
|
++ : "a" (mem), "a" (value) \
|
|
|
|
++ : "memory" ); \
|
|
|
|
++ __tmp; \
|
|
|
|
++ })
|
|
|
|
++
|
|
|
|
++/* Subtract VALUE from *MEM and return the old value of *MEM. */
|
|
|
|
++
|
|
|
|
++#define __arch_atomic_exchange_and_sub_32(mem, value) \
|
|
|
|
++ ({__typeof__(*(mem)) __tmp, __value; \
|
|
|
|
++ __asm__ __volatile__( \
|
|
|
|
++ "1: l32i %0, %2, 0 \n" \
|
|
|
|
++ " wsr %0, SCOMPARE1 \n" \
|
|
|
|
++ " sub %1, %0, %3 \n" \
|
|
|
|
++ " s32c1i %1, %2, 0 \n" \
|
|
|
|
++ " bne %0, %1, 1b \n" \
|
|
|
|
++ : "=&a" (__value), "=&a" (__tmp) \
|
|
|
|
++ : "a" (mem), "a" (value) \
|
|
|
|
++ : "memory" ); \
|
|
|
|
++ __tmp; \
|
|
|
|
++ })
|
|
|
|
++
|
|
|
|
++/* Decrement *MEM if it is > 0, and return the old value. */
|
|
|
|
++
|
|
|
|
++#define __arch_atomic_decrement_if_positive_32(mem) \
|
|
|
|
++ ({__typeof__(*(mem)) __tmp, __value; \
|
|
|
|
++ __asm__ __volatile__( \
|
|
|
|
++ "1: l32i %0, %2, 0 \n" \
|
|
|
|
++ " blti %0, 1, 2f \n" \
|
|
|
|
++ " wsr %0, SCOMPARE1 \n" \
|
|
|
|
++ " addi %1, %0, -1 \n" \
|
|
|
|
++ " s32c1i %1, %2, 0 \n" \
|
|
|
|
++ " bne %0, %1, 1b \n" \
|
|
|
|
++ "2: \n" \
|
|
|
|
++ : "=&a" (__value), "=&a" (__tmp) \
|
|
|
|
++ : "a" (mem) \
|
|
|
|
++ : "memory" ); \
|
|
|
|
++ __tmp; \
|
|
|
|
++ })
|
|
|
|
++
|
|
|
|
++
|
|
|
|
++/* These are the preferred public interfaces: */
|
|
|
|
++
|
|
|
|
++#define atomic_compare_and_exchange_val_acq(mem, newval, oldval) \
|
|
|
|
++ ({ \
|
|
|
|
++ if (sizeof (*mem) != 4) \
|
|
|
|
++ abort(); \
|
|
|
|
++ __arch_compare_and_exchange_val_32_acq(mem, newval, oldval); \
|
|
|
|
++ })
|
|
|
|
++
|
|
|
|
++#define atomic_exchange_acq(mem, newval) \
|
|
|
|
++ ({ \
|
|
|
|
++ if (sizeof(*(mem)) != 4) \
|
|
|
|
++ abort(); \
|
|
|
|
++ __arch_exchange_32_acq(mem, newval); \
|
|
|
|
++ })
|
|
|
|
++
|
|
|
|
++#define atomic_exchange_and_add(mem, newval) \
|
|
|
|
++ ({ \
|
|
|
|
++ if (sizeof(*(mem)) != 4) \
|
|
|
|
++ abort(); \
|
|
|
|
++ __arch_atomic_exchange_and_add_32(mem, newval); \
|
|
|
|
++ })
|
|
|
|
++
|
|
|
|
++#define atomic_exchange_and_sub(mem, newval) \
|
|
|
|
++ ({ \
|
|
|
|
++ if (sizeof(*(mem)) != 4) \
|
|
|
|
++ abort(); \
|
|
|
|
++ __arch_atomic_exchange_and_sub_32(mem, newval); \
|
|
|
|
++ })
|
|
|
|
++
|
|
|
|
++#define atomic_decrement_if_positive(mem) \
|
|
|
|
++ ({ \
|
|
|
|
++ if (sizeof(*(mem)) != 4) \
|
|
|
|
++ abort(); \
|
|
|
|
++ __arch_atomic_decrement_if_positive_32(mem); \
|
|
|
|
++ })
|
|
|
|
++
|
|
|
|
++#endif /* _XTENSA_BITS_ATOMIC_H */
|
|
|
|
++
|