aboutsummaryrefslogtreecommitdiff
path: root/sys/powerpc/include/atomic.h
diff options
context:
space:
mode:
Diffstat (limited to 'sys/powerpc/include/atomic.h')
-rw-r--r--sys/powerpc/include/atomic.h1183
1 files changed, 1183 insertions, 0 deletions
diff --git a/sys/powerpc/include/atomic.h b/sys/powerpc/include/atomic.h
new file mode 100644
index 000000000000..b2d7549e5bd0
--- /dev/null
+++ b/sys/powerpc/include/atomic.h
@@ -0,0 +1,1183 @@
+/*-
+ * SPDX-License-Identifier: BSD-2-Clause
+ *
+ * Copyright (c) 2008 Marcel Moolenaar
+ * Copyright (c) 2001 Benno Rice
+ * Copyright (c) 2001 David E. O'Brien
+ * Copyright (c) 1998 Doug Rabson
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
+ * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
+ * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
+ * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
+ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ */
+
+#ifndef _MACHINE_ATOMIC_H_
+#define _MACHINE_ATOMIC_H_
+
+#include <sys/atomic_common.h>
+
+#ifndef __powerpc64__
+#include <sys/_atomic64e.h>
+#endif
+
+/*
+ * The __ATOMIC_REL/ACQ() macros provide memory barriers only in conjunction
+ * with the atomic lXarx/stXcx. sequences below. They are not exposed outside
+ * of this file. See also Appendix B.2 of Book II of the architecture manual.
+ *
+ * Note that not all Book-E processors accept the light-weight sync variant.
+ * In particular, early models of E500 cores are known to wedge. Bank on all
+ * 64-bit capable CPUs to accept lwsync properly and pressimize 32-bit CPUs
+ * to use the heavier-weight sync.
+ */
+
+#ifdef __powerpc64__
+#define mb() __asm __volatile("sync" : : : "memory")
+#define rmb() __asm __volatile("lwsync" : : : "memory")
+#define wmb() __asm __volatile("lwsync" : : : "memory")
+#define __ATOMIC_REL() __asm __volatile("lwsync" : : : "memory")
+#define __ATOMIC_ACQ() __asm __volatile("isync" : : : "memory")
+#else
+#define mb() __asm __volatile("sync" : : : "memory")
+#define rmb() __asm __volatile("sync" : : : "memory")
+#define wmb() __asm __volatile("sync" : : : "memory")
+#define __ATOMIC_REL() __asm __volatile("sync" : : : "memory")
+#define __ATOMIC_ACQ() __asm __volatile("isync" : : : "memory")
+#endif
+
+static __inline void
+powerpc_lwsync(void)
+{
+
+#ifdef __powerpc64__
+ __asm __volatile("lwsync" : : : "memory");
+#else
+ __asm __volatile("sync" : : : "memory");
+#endif
+}
+
+/*
+ * atomic_add(p, v)
+ * { *p += v; }
+ */
+
+#define __atomic_add_int(p, v, t) \
+ __asm __volatile( \
+ "1: lwarx %0, 0, %2\n" \
+ " add %0, %3, %0\n" \
+ " stwcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_add_int */
+
+#ifdef __powerpc64__
+#define __atomic_add_long(p, v, t) \
+ __asm __volatile( \
+ "1: ldarx %0, 0, %2\n" \
+ " add %0, %3, %0\n" \
+ " stdcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_add_long */
+#else
+#define __atomic_add_long(p, v, t) \
+ __asm __volatile( \
+ "1: lwarx %0, 0, %2\n" \
+ " add %0, %3, %0\n" \
+ " stwcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_add_long */
+#endif
+
+#define _ATOMIC_ADD(type) \
+ static __inline void \
+ atomic_add_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __atomic_add_##type(p, v, t); \
+ } \
+ \
+ static __inline void \
+ atomic_add_acq_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __atomic_add_##type(p, v, t); \
+ __ATOMIC_ACQ(); \
+ } \
+ \
+ static __inline void \
+ atomic_add_rel_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __ATOMIC_REL(); \
+ __atomic_add_##type(p, v, t); \
+ } \
+ /* _ATOMIC_ADD */
+
+_ATOMIC_ADD(int)
+_ATOMIC_ADD(long)
+
+#define atomic_add_32 atomic_add_int
+#define atomic_add_acq_32 atomic_add_acq_int
+#define atomic_add_rel_32 atomic_add_rel_int
+
+#ifdef __powerpc64__
+#define atomic_add_64 atomic_add_long
+#define atomic_add_acq_64 atomic_add_acq_long
+#define atomic_add_rel_64 atomic_add_rel_long
+
+#define atomic_add_ptr atomic_add_long
+#define atomic_add_acq_ptr atomic_add_acq_long
+#define atomic_add_rel_ptr atomic_add_rel_long
+#else
+#define atomic_add_ptr atomic_add_int
+#define atomic_add_acq_ptr atomic_add_acq_int
+#define atomic_add_rel_ptr atomic_add_rel_int
+#endif
+#undef _ATOMIC_ADD
+#undef __atomic_add_long
+#undef __atomic_add_int
+
+/*
+ * atomic_clear(p, v)
+ * { *p &= ~v; }
+ */
+
+#define __atomic_clear_int(p, v, t) \
+ __asm __volatile( \
+ "1: lwarx %0, 0, %2\n" \
+ " andc %0, %0, %3\n" \
+ " stwcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_clear_int */
+
+#ifdef __powerpc64__
+#define __atomic_clear_long(p, v, t) \
+ __asm __volatile( \
+ "1: ldarx %0, 0, %2\n" \
+ " andc %0, %0, %3\n" \
+ " stdcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_clear_long */
+#else
+#define __atomic_clear_long(p, v, t) \
+ __asm __volatile( \
+ "1: lwarx %0, 0, %2\n" \
+ " andc %0, %0, %3\n" \
+ " stwcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_clear_long */
+#endif
+
+#define _ATOMIC_CLEAR(type) \
+ static __inline void \
+ atomic_clear_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __atomic_clear_##type(p, v, t); \
+ } \
+ \
+ static __inline void \
+ atomic_clear_acq_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __atomic_clear_##type(p, v, t); \
+ __ATOMIC_ACQ(); \
+ } \
+ \
+ static __inline void \
+ atomic_clear_rel_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __ATOMIC_REL(); \
+ __atomic_clear_##type(p, v, t); \
+ } \
+ /* _ATOMIC_CLEAR */
+
+_ATOMIC_CLEAR(int)
+_ATOMIC_CLEAR(long)
+
+#define atomic_clear_32 atomic_clear_int
+#define atomic_clear_acq_32 atomic_clear_acq_int
+#define atomic_clear_rel_32 atomic_clear_rel_int
+
+#ifdef __powerpc64__
+#define atomic_clear_64 atomic_clear_long
+#define atomic_clear_acq_64 atomic_clear_acq_long
+#define atomic_clear_rel_64 atomic_clear_rel_long
+
+#define atomic_clear_ptr atomic_clear_long
+#define atomic_clear_acq_ptr atomic_clear_acq_long
+#define atomic_clear_rel_ptr atomic_clear_rel_long
+#else
+#define atomic_clear_ptr atomic_clear_int
+#define atomic_clear_acq_ptr atomic_clear_acq_int
+#define atomic_clear_rel_ptr atomic_clear_rel_int
+#endif
+#undef _ATOMIC_CLEAR
+#undef __atomic_clear_long
+#undef __atomic_clear_int
+
+/*
+ * atomic_cmpset(p, o, n)
+ */
+/* TODO -- see below */
+
+/*
+ * atomic_load_acq(p)
+ */
+/* TODO -- see below */
+
+/*
+ * atomic_readandclear(p)
+ */
+/* TODO -- see below */
+
+/*
+ * atomic_set(p, v)
+ * { *p |= v; }
+ */
+
+#define __atomic_set_int(p, v, t) \
+ __asm __volatile( \
+ "1: lwarx %0, 0, %2\n" \
+ " or %0, %3, %0\n" \
+ " stwcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_set_int */
+
+#ifdef __powerpc64__
+#define __atomic_set_long(p, v, t) \
+ __asm __volatile( \
+ "1: ldarx %0, 0, %2\n" \
+ " or %0, %3, %0\n" \
+ " stdcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_set_long */
+#else
+#define __atomic_set_long(p, v, t) \
+ __asm __volatile( \
+ "1: lwarx %0, 0, %2\n" \
+ " or %0, %3, %0\n" \
+ " stwcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_set_long */
+#endif
+
+#define _ATOMIC_SET(type) \
+ static __inline void \
+ atomic_set_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __atomic_set_##type(p, v, t); \
+ } \
+ \
+ static __inline void \
+ atomic_set_acq_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __atomic_set_##type(p, v, t); \
+ __ATOMIC_ACQ(); \
+ } \
+ \
+ static __inline void \
+ atomic_set_rel_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __ATOMIC_REL(); \
+ __atomic_set_##type(p, v, t); \
+ } \
+ /* _ATOMIC_SET */
+
+_ATOMIC_SET(int)
+_ATOMIC_SET(long)
+
+#define atomic_set_32 atomic_set_int
+#define atomic_set_acq_32 atomic_set_acq_int
+#define atomic_set_rel_32 atomic_set_rel_int
+
+#ifdef __powerpc64__
+#define atomic_set_64 atomic_set_long
+#define atomic_set_acq_64 atomic_set_acq_long
+#define atomic_set_rel_64 atomic_set_rel_long
+
+#define atomic_set_ptr atomic_set_long
+#define atomic_set_acq_ptr atomic_set_acq_long
+#define atomic_set_rel_ptr atomic_set_rel_long
+#else
+#define atomic_set_ptr atomic_set_int
+#define atomic_set_acq_ptr atomic_set_acq_int
+#define atomic_set_rel_ptr atomic_set_rel_int
+#endif
+#undef _ATOMIC_SET
+#undef __atomic_set_long
+#undef __atomic_set_int
+
+/*
+ * atomic_subtract(p, v)
+ * { *p -= v; }
+ */
+
+#define __atomic_subtract_int(p, v, t) \
+ __asm __volatile( \
+ "1: lwarx %0, 0, %2\n" \
+ " subf %0, %3, %0\n" \
+ " stwcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_subtract_int */
+
+#ifdef __powerpc64__
+#define __atomic_subtract_long(p, v, t) \
+ __asm __volatile( \
+ "1: ldarx %0, 0, %2\n" \
+ " subf %0, %3, %0\n" \
+ " stdcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_subtract_long */
+#else
+#define __atomic_subtract_long(p, v, t) \
+ __asm __volatile( \
+ "1: lwarx %0, 0, %2\n" \
+ " subf %0, %3, %0\n" \
+ " stwcx. %0, 0, %2\n" \
+ " bne- 1b\n" \
+ : "=&r" (t), "=m" (*p) \
+ : "r" (p), "r" (v), "m" (*p) \
+ : "cr0", "memory") \
+ /* __atomic_subtract_long */
+#endif
+
+#define _ATOMIC_SUBTRACT(type) \
+ static __inline void \
+ atomic_subtract_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __atomic_subtract_##type(p, v, t); \
+ } \
+ \
+ static __inline void \
+ atomic_subtract_acq_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __atomic_subtract_##type(p, v, t); \
+ __ATOMIC_ACQ(); \
+ } \
+ \
+ static __inline void \
+ atomic_subtract_rel_##type(volatile u_##type *p, u_##type v) { \
+ u_##type t; \
+ __ATOMIC_REL(); \
+ __atomic_subtract_##type(p, v, t); \
+ } \
+ /* _ATOMIC_SUBTRACT */
+
+_ATOMIC_SUBTRACT(int)
+_ATOMIC_SUBTRACT(long)
+
+#define atomic_subtract_32 atomic_subtract_int
+#define atomic_subtract_acq_32 atomic_subtract_acq_int
+#define atomic_subtract_rel_32 atomic_subtract_rel_int
+
+#ifdef __powerpc64__
+#define atomic_subtract_64 atomic_subtract_long
+#define atomic_subtract_acq_64 atomic_subract_acq_long
+#define atomic_subtract_rel_64 atomic_subtract_rel_long
+
+#define atomic_subtract_ptr atomic_subtract_long
+#define atomic_subtract_acq_ptr atomic_subtract_acq_long
+#define atomic_subtract_rel_ptr atomic_subtract_rel_long
+#else
+#define atomic_subtract_ptr atomic_subtract_int
+#define atomic_subtract_acq_ptr atomic_subtract_acq_int
+#define atomic_subtract_rel_ptr atomic_subtract_rel_int
+#endif
+#undef _ATOMIC_SUBTRACT
+#undef __atomic_subtract_long
+#undef __atomic_subtract_int
+
+/*
+ * atomic_store_rel(p, v)
+ */
+/* TODO -- see below */
+
+/*
+ * Old/original implementations that still need revisiting.
+ */
+
+static __inline u_int
+atomic_readandclear_int(volatile u_int *addr)
+{
+ u_int result,temp;
+
+ __asm __volatile (
+ "\tsync\n" /* drain writes */
+ "1:\tlwarx %0, 0, %3\n\t" /* load old value */
+ "li %1, 0\n\t" /* load new value */
+ "stwcx. %1, 0, %3\n\t" /* attempt to store */
+ "bne- 1b\n\t" /* spin if failed */
+ : "=&r"(result), "=&r"(temp), "=m" (*addr)
+ : "r" (addr), "m" (*addr)
+ : "cr0", "memory");
+
+ return (result);
+}
+
+#ifdef __powerpc64__
+static __inline u_long
+atomic_readandclear_long(volatile u_long *addr)
+{
+ u_long result,temp;
+
+ __asm __volatile (
+ "\tsync\n" /* drain writes */
+ "1:\tldarx %0, 0, %3\n\t" /* load old value */
+ "li %1, 0\n\t" /* load new value */
+ "stdcx. %1, 0, %3\n\t" /* attempt to store */
+ "bne- 1b\n\t" /* spin if failed */
+ : "=&r"(result), "=&r"(temp), "=m" (*addr)
+ : "r" (addr), "m" (*addr)
+ : "cr0", "memory");
+
+ return (result);
+}
+#endif
+
+#define atomic_readandclear_32 atomic_readandclear_int
+
+#ifdef __powerpc64__
+#define atomic_readandclear_64 atomic_readandclear_long
+
+#define atomic_readandclear_ptr atomic_readandclear_long
+#else
+static __inline u_long
+atomic_readandclear_long(volatile u_long *addr)
+{
+
+ return ((u_long)atomic_readandclear_int((volatile u_int *)addr));
+}
+
+#define atomic_readandclear_ptr atomic_readandclear_int
+#endif
+
+/*
+ * We assume that a = b will do atomic loads and stores.
+ */
+#define ATOMIC_STORE_LOAD(TYPE) \
+static __inline u_##TYPE \
+atomic_load_acq_##TYPE(const volatile u_##TYPE *p) \
+{ \
+ u_##TYPE v; \
+ \
+ v = *p; \
+ powerpc_lwsync(); \
+ return (v); \
+} \
+ \
+static __inline void \
+atomic_store_rel_##TYPE(volatile u_##TYPE *p, u_##TYPE v) \
+{ \
+ \
+ powerpc_lwsync(); \
+ *p = v; \
+}
+
+ATOMIC_STORE_LOAD(int)
+
+#define atomic_load_acq_32 atomic_load_acq_int
+#define atomic_store_rel_32 atomic_store_rel_int
+
+#ifdef __powerpc64__
+ATOMIC_STORE_LOAD(long)
+
+#define atomic_load_acq_64 atomic_load_acq_long
+#define atomic_store_rel_64 atomic_store_rel_long
+
+#define atomic_load_acq_ptr atomic_load_acq_long
+#define atomic_store_rel_ptr atomic_store_rel_long
+#else
+static __inline u_long
+atomic_load_acq_long(const volatile u_long *addr)
+{
+
+ return ((u_long)atomic_load_acq_int((const volatile u_int *)addr));
+}
+
+static __inline void
+atomic_store_rel_long(volatile u_long *addr, u_long val)
+{
+
+ atomic_store_rel_int((volatile u_int *)addr, (u_int)val);
+}
+
+#define atomic_load_acq_ptr atomic_load_acq_int
+#define atomic_store_rel_ptr atomic_store_rel_int
+#endif
+#undef ATOMIC_STORE_LOAD
+
+/*
+ * Atomically compare the value stored at *p with cmpval and if the
+ * two values are equal, update the value of *p with newval. Returns
+ * zero if the compare failed, nonzero otherwise.
+ */
+#ifdef ISA_206_ATOMICS
+static __inline int
+atomic_cmpset_char(volatile u_char *p, u_char cmpval, u_char newval)
+{
+ int ret;
+
+ __asm __volatile (
+ "1:\tlbarx %0, 0, %2\n\t" /* load old value */
+ "cmplw %3, %0\n\t" /* compare */
+ "bne- 2f\n\t" /* exit if not equal */
+ "stbcx. %4, 0, %2\n\t" /* attempt to store */
+ "bne- 1b\n\t" /* spin if failed */
+ "li %0, 1\n\t" /* success - retval = 1 */
+ "b 3f\n\t" /* we've succeeded */
+ "2:\n\t"
+ "stbcx. %0, 0, %2\n\t" /* clear reservation (74xx) */
+ "li %0, 0\n\t" /* failure - retval = 0 */
+ "3:\n\t"
+ : "=&r" (ret), "=m" (*p)
+ : "r" (p), "r" (cmpval), "r" (newval), "m" (*p)
+ : "cr0", "memory");
+
+ return (ret);
+}
+
+static __inline int
+atomic_cmpset_short(volatile u_short *p, u_short cmpval, u_short newval)
+{
+ int ret;
+
+ __asm __volatile (
+ "1:\tlharx %0, 0, %2\n\t" /* load old value */
+ "cmplw %3, %0\n\t" /* compare */
+ "bne- 2f\n\t" /* exit if not equal */
+ "sthcx. %4, 0, %2\n\t" /* attempt to store */
+ "bne- 1b\n\t" /* spin if failed */
+ "li %0, 1\n\t" /* success - retval = 1 */
+ "b 3f\n\t" /* we've succeeded */
+ "2:\n\t"
+ "sthcx. %0, 0, %2\n\t" /* clear reservation (74xx) */
+ "li %0, 0\n\t" /* failure - retval = 0 */
+ "3:\n\t"
+ : "=&r" (ret), "=m" (*p)
+ : "r" (p), "r" (cmpval), "r" (newval), "m" (*p)
+ : "cr0", "memory");
+
+ return (ret);
+}
+#else
+static __inline int
+atomic_cmpset_masked(uint32_t *p, uint32_t cmpval, uint32_t newval,
+ uint32_t mask)
+{
+ int ret;
+ uint32_t tmp;
+
+ __asm __volatile (
+ "1:\tlwarx %2, 0, %3\n\t" /* load old value */
+ "and %0, %2, %7\n\t"
+ "cmplw %4, %0\n\t" /* compare */
+ "bne- 2f\n\t" /* exit if not equal */
+ "andc %2, %2, %7\n\t"
+ "or %2, %2, %5\n\t"
+ "stwcx. %2, 0, %3\n\t" /* attempt to store */
+ "bne- 1b\n\t" /* spin if failed */
+ "li %0, 1\n\t" /* success - retval = 1 */
+ "b 3f\n\t" /* we've succeeded */
+ "2:\n\t"
+ "stwcx. %2, 0, %3\n\t" /* clear reservation (74xx) */
+ "li %0, 0\n\t" /* failure - retval = 0 */
+ "3:\n\t"
+ : "=&r" (ret), "=m" (*p), "+&r" (tmp)
+ : "r" (p), "r" (cmpval), "r" (newval), "m" (*p),
+ "r" (mask)
+ : "cr0", "memory");
+
+ return (ret);
+}
+
+#define _atomic_cmpset_masked_word(a,o,v,m) atomic_cmpset_masked(a, o, v, m)
+#endif
+
+static __inline int
+atomic_cmpset_int(volatile u_int* p, u_int cmpval, u_int newval)
+{
+ int ret;
+
+ __asm __volatile (
+ "1:\tlwarx %0, 0, %2\n\t" /* load old value */
+ "cmplw %3, %0\n\t" /* compare */
+ "bne- 2f\n\t" /* exit if not equal */
+ "stwcx. %4, 0, %2\n\t" /* attempt to store */
+ "bne- 1b\n\t" /* spin if failed */
+ "li %0, 1\n\t" /* success - retval = 1 */
+ "b 3f\n\t" /* we've succeeded */
+ "2:\n\t"
+ "stwcx. %0, 0, %2\n\t" /* clear reservation (74xx) */
+ "li %0, 0\n\t" /* failure - retval = 0 */
+ "3:\n\t"
+ : "=&r" (ret), "=m" (*p)
+ : "r" (p), "r" (cmpval), "r" (newval), "m" (*p)
+ : "cr0", "memory");
+
+ return (ret);
+}
+static __inline int
+atomic_cmpset_long(volatile u_long* p, u_long cmpval, u_long newval)
+{
+ int ret;
+
+ __asm __volatile (
+ #ifdef __powerpc64__
+ "1:\tldarx %0, 0, %2\n\t" /* load old value */
+ "cmpld %3, %0\n\t" /* compare */
+ "bne- 2f\n\t" /* exit if not equal */
+ "stdcx. %4, 0, %2\n\t" /* attempt to store */
+ #else
+ "1:\tlwarx %0, 0, %2\n\t" /* load old value */
+ "cmplw %3, %0\n\t" /* compare */
+ "bne- 2f\n\t" /* exit if not equal */
+ "stwcx. %4, 0, %2\n\t" /* attempt to store */
+ #endif
+ "bne- 1b\n\t" /* spin if failed */
+ "li %0, 1\n\t" /* success - retval = 1 */
+ "b 3f\n\t" /* we've succeeded */
+ "2:\n\t"
+ #ifdef __powerpc64__
+ "stdcx. %0, 0, %2\n\t" /* clear reservation (74xx) */
+ #else
+ "stwcx. %0, 0, %2\n\t" /* clear reservation (74xx) */
+ #endif
+ "li %0, 0\n\t" /* failure - retval = 0 */
+ "3:\n\t"
+ : "=&r" (ret), "=m" (*p)
+ : "r" (p), "r" (cmpval), "r" (newval), "m" (*p)
+ : "cr0", "memory");
+
+ return (ret);
+}
+
+#define ATOMIC_CMPSET_ACQ_REL(type) \
+ static __inline int \
+ atomic_cmpset_acq_##type(volatile u_##type *p, \
+ u_##type cmpval, u_##type newval)\
+ {\
+ u_##type retval; \
+ retval = atomic_cmpset_##type(p, cmpval, newval);\
+ __ATOMIC_ACQ();\
+ return (retval);\
+ }\
+ static __inline int \
+ atomic_cmpset_rel_##type(volatile u_##type *p, \
+ u_##type cmpval, u_##type newval)\
+ {\
+ __ATOMIC_REL();\
+ return (atomic_cmpset_##type(p, cmpval, newval));\
+ }\
+ struct hack
+
+ATOMIC_CMPSET_ACQ_REL(int);
+ATOMIC_CMPSET_ACQ_REL(long);
+
+#ifdef ISA_206_ATOMICS
+#define atomic_cmpset_8 atomic_cmpset_char
+#endif
+#define atomic_cmpset_acq_8 atomic_cmpset_acq_char
+#define atomic_cmpset_rel_8 atomic_cmpset_rel_char
+
+#ifdef ISA_206_ATOMICS
+#define atomic_cmpset_16 atomic_cmpset_short
+#endif
+#define atomic_cmpset_acq_16 atomic_cmpset_acq_short
+#define atomic_cmpset_rel_16 atomic_cmpset_rel_short
+
+#define atomic_cmpset_32 atomic_cmpset_int
+#define atomic_cmpset_acq_32 atomic_cmpset_acq_int
+#define atomic_cmpset_rel_32 atomic_cmpset_rel_int
+
+#ifdef __powerpc64__
+#define atomic_cmpset_64 atomic_cmpset_long
+#define atomic_cmpset_acq_64 atomic_cmpset_acq_long
+#define atomic_cmpset_rel_64 atomic_cmpset_rel_long
+
+#define atomic_cmpset_ptr atomic_cmpset_long
+#define atomic_cmpset_acq_ptr atomic_cmpset_acq_long
+#define atomic_cmpset_rel_ptr atomic_cmpset_rel_long
+#else
+#define atomic_cmpset_ptr atomic_cmpset_int
+#define atomic_cmpset_acq_ptr atomic_cmpset_acq_int
+#define atomic_cmpset_rel_ptr atomic_cmpset_rel_int
+#endif
+
+/*
+ * Atomically compare the value stored at *p with *cmpval and if the
+ * two values are equal, update the value of *p with newval. Returns
+ * zero if the compare failed and sets *cmpval to the read value from *p,
+ * nonzero otherwise.
+ */
+#ifdef ISA_206_ATOMICS
+static __inline int
+atomic_fcmpset_char(volatile u_char *p, u_char *cmpval, u_char newval)
+{
+ int ret;
+
+ __asm __volatile (
+ "lbarx %0, 0, %3\n\t" /* load old value */
+ "cmplw %4, %0\n\t" /* compare */
+ "bne- 1f\n\t" /* exit if not equal */
+ "stbcx. %5, 0, %3\n\t" /* attempt to store */
+ "bne- 1f\n\t" /* exit if failed */
+ "li %0, 1\n\t" /* success - retval = 1 */
+ "b 2f\n\t" /* we've succeeded */
+ "1:\n\t"
+ "stbcx. %0, 0, %3\n\t" /* clear reservation (74xx) */
+ "stbx %0, 0, %7\n\t"
+ "li %0, 0\n\t" /* failure - retval = 0 */
+ "2:\n\t"
+ : "=&r" (ret), "=m" (*p), "=m" (*cmpval)
+ : "r" (p), "r" (*cmpval), "r" (newval), "m" (*p), "r"(cmpval)
+ : "cr0", "memory");
+
+ return (ret);
+}
+
+static __inline int
+atomic_fcmpset_short(volatile u_short *p, u_short *cmpval, u_short newval)
+{
+ int ret;
+
+ __asm __volatile (
+ "lharx %0, 0, %3\n\t" /* load old value */
+ "cmplw %4, %0\n\t" /* compare */
+ "bne- 1f\n\t" /* exit if not equal */
+ "sthcx. %5, 0, %3\n\t" /* attempt to store */
+ "bne- 1f\n\t" /* exit if failed */
+ "li %0, 1\n\t" /* success - retval = 1 */
+ "b 2f\n\t" /* we've succeeded */
+ "1:\n\t"
+ "sthcx. %0, 0, %3\n\t" /* clear reservation (74xx) */
+ "sthx %0, 0, %7\n\t"
+ "li %0, 0\n\t" /* failure - retval = 0 */
+ "2:\n\t"
+ : "=&r" (ret), "=m" (*p), "=m" (*cmpval)
+ : "r" (p), "r" (*cmpval), "r" (newval), "m" (*p), "r"(cmpval)
+ : "cr0", "memory");
+
+ return (ret);
+}
+#endif /* ISA_206_ATOMICS */
+
+static __inline int
+atomic_fcmpset_int(volatile u_int *p, u_int *cmpval, u_int newval)
+{
+ int ret;
+
+ __asm __volatile (
+ "lwarx %0, 0, %3\n\t" /* load old value */
+ "cmplw %4, %0\n\t" /* compare */
+ "bne- 1f\n\t" /* exit if not equal */
+ "stwcx. %5, 0, %3\n\t" /* attempt to store */
+ "bne- 1f\n\t" /* exit if failed */
+ "li %0, 1\n\t" /* success - retval = 1 */
+ "b 2f\n\t" /* we've succeeded */
+ "1:\n\t"
+ "stwcx. %0, 0, %3\n\t" /* clear reservation (74xx) */
+ "stwx %0, 0, %7\n\t"
+ "li %0, 0\n\t" /* failure - retval = 0 */
+ "2:\n\t"
+ : "=&r" (ret), "=m" (*p), "=m" (*cmpval)
+ : "r" (p), "r" (*cmpval), "r" (newval), "m" (*p), "r"(cmpval)
+ : "cr0", "memory");
+
+ return (ret);
+}
+static __inline int
+atomic_fcmpset_long(volatile u_long *p, u_long *cmpval, u_long newval)
+{
+ int ret;
+
+ __asm __volatile (
+ #ifdef __powerpc64__
+ "ldarx %0, 0, %3\n\t" /* load old value */
+ "cmpld %4, %0\n\t" /* compare */
+ "bne- 1f\n\t" /* exit if not equal */
+ "stdcx. %5, 0, %3\n\t" /* attempt to store */
+ #else
+ "lwarx %0, 0, %3\n\t" /* load old value */
+ "cmplw %4, %0\n\t" /* compare */
+ "bne- 1f\n\t" /* exit if not equal */
+ "stwcx. %5, 0, %3\n\t" /* attempt to store */
+ #endif
+ "bne- 1f\n\t" /* exit if failed */
+ "li %0, 1\n\t" /* success - retval = 1 */
+ "b 2f\n\t" /* we've succeeded */
+ "1:\n\t"
+ #ifdef __powerpc64__
+ "stdcx. %0, 0, %3\n\t" /* clear reservation (74xx) */
+ "stdx %0, 0, %7\n\t"
+ #else
+ "stwcx. %0, 0, %3\n\t" /* clear reservation (74xx) */
+ "stwx %0, 0, %7\n\t"
+ #endif
+ "li %0, 0\n\t" /* failure - retval = 0 */
+ "2:\n\t"
+ : "=&r" (ret), "=m" (*p), "=m" (*cmpval)
+ : "r" (p), "r" (*cmpval), "r" (newval), "m" (*p), "r"(cmpval)
+ : "cr0", "memory");
+
+ return (ret);
+}
+
+#define ATOMIC_FCMPSET_ACQ_REL(type) \
+ static __inline int \
+ atomic_fcmpset_acq_##type(volatile u_##type *p, \
+ u_##type *cmpval, u_##type newval)\
+ {\
+ u_##type retval; \
+ retval = atomic_fcmpset_##type(p, cmpval, newval);\
+ __ATOMIC_ACQ();\
+ return (retval);\
+ }\
+ static __inline int \
+ atomic_fcmpset_rel_##type(volatile u_##type *p, \
+ u_##type *cmpval, u_##type newval)\
+ {\
+ __ATOMIC_REL();\
+ return (atomic_fcmpset_##type(p, cmpval, newval));\
+ }\
+ struct hack
+
+ATOMIC_FCMPSET_ACQ_REL(int);
+ATOMIC_FCMPSET_ACQ_REL(long);
+
+#ifdef ISA_206_ATOMICS
+#define atomic_fcmpset_8 atomic_fcmpset_char
+#endif
+#define atomic_fcmpset_acq_8 atomic_fcmpset_acq_char
+#define atomic_fcmpset_rel_8 atomic_fcmpset_rel_char
+
+#ifdef ISA_206_ATOMICS
+#define atomic_fcmpset_16 atomic_fcmpset_short
+#endif
+#define atomic_fcmpset_acq_16 atomic_fcmpset_acq_short
+#define atomic_fcmpset_rel_16 atomic_fcmpset_rel_short
+
+#define atomic_fcmpset_32 atomic_fcmpset_int
+#define atomic_fcmpset_acq_32 atomic_fcmpset_acq_int
+#define atomic_fcmpset_rel_32 atomic_fcmpset_rel_int
+
+#ifdef __powerpc64__
+#define atomic_fcmpset_64 atomic_fcmpset_long
+#define atomic_fcmpset_acq_64 atomic_fcmpset_acq_long
+#define atomic_fcmpset_rel_64 atomic_fcmpset_rel_long
+
+#define atomic_fcmpset_ptr atomic_fcmpset_long
+#define atomic_fcmpset_acq_ptr atomic_fcmpset_acq_long
+#define atomic_fcmpset_rel_ptr atomic_fcmpset_rel_long
+#else
+#define atomic_fcmpset_ptr atomic_fcmpset_int
+#define atomic_fcmpset_acq_ptr atomic_fcmpset_acq_int
+#define atomic_fcmpset_rel_ptr atomic_fcmpset_rel_int
+#endif
+
+static __inline u_int
+atomic_fetchadd_int(volatile u_int *p, u_int v)
+{
+ u_int value;
+
+ do {
+ value = *p;
+ } while (!atomic_cmpset_int(p, value, value + v));
+ return (value);
+}
+
+static __inline u_long
+atomic_fetchadd_long(volatile u_long *p, u_long v)
+{
+ u_long value;
+
+ do {
+ value = *p;
+ } while (!atomic_cmpset_long(p, value, value + v));
+ return (value);
+}
+
+static __inline u_int
+atomic_swap_32(volatile u_int *p, u_int v)
+{
+ u_int prev;
+
+ __asm __volatile(
+ "1: lwarx %0,0,%2\n"
+ " stwcx. %3,0,%2\n"
+ " bne- 1b\n"
+ : "=&r" (prev), "+m" (*(volatile u_int *)p)
+ : "r" (p), "r" (v)
+ : "cr0", "memory");
+
+ return (prev);
+}
+
+#ifdef __powerpc64__
+static __inline u_long
+atomic_swap_64(volatile u_long *p, u_long v)
+{
+ u_long prev;
+
+ __asm __volatile(
+ "1: ldarx %0,0,%2\n"
+ " stdcx. %3,0,%2\n"
+ " bne- 1b\n"
+ : "=&r" (prev), "+m" (*(volatile u_long *)p)
+ : "r" (p), "r" (v)
+ : "cr0", "memory");
+
+ return (prev);
+}
+#endif
+
+#define atomic_fetchadd_32 atomic_fetchadd_int
+#define atomic_swap_int atomic_swap_32
+
+#ifdef __powerpc64__
+#define atomic_fetchadd_64 atomic_fetchadd_long
+#define atomic_swap_long atomic_swap_64
+#define atomic_swap_ptr atomic_swap_64
+#else
+#define atomic_swap_long(p,v) atomic_swap_32((volatile u_int *)(p), v)
+#define atomic_swap_ptr(p,v) atomic_swap_32((volatile u_int *)(p), v)
+#endif
+
+static __inline int
+atomic_testandset_int(volatile u_int *p, u_int v)
+{
+ u_int m = (1u << (v & 0x1f));
+ u_int res;
+ u_int tmp;
+
+ __asm __volatile(
+ "1: lwarx %0,0,%3\n"
+ " and %1,%0,%4\n"
+ " or %0,%0,%4\n"
+ " stwcx. %0,0,%3\n"
+ " bne- 1b\n"
+ : "=&r"(tmp), "=&r"(res), "+m"(*p)
+ : "r"(p), "r"(m)
+ : "cr0", "memory");
+
+ return (res != 0);
+}
+
+static __inline int
+atomic_testandclear_int(volatile u_int *p, u_int v)
+{
+ u_int m = (1u << (v & 0x1f));
+ u_int res;
+ u_int tmp;
+
+ __asm __volatile(
+ "1: lwarx %0,0,%3\n"
+ " and %1,%0,%4\n"
+ " andc %0,%0,%4\n"
+ " stwcx. %0,0,%3\n"
+ " bne- 1b\n"
+ : "=&r"(tmp), "=&r"(res), "+m"(*p)
+ : "r"(p), "r"(m)
+ : "cr0", "memory");
+
+ return (res != 0);
+}
+
+#ifdef __powerpc64__
+static __inline int
+atomic_testandset_long(volatile u_long *p, u_int v)
+{
+ u_long m = (1ul << (v & 0x3f));
+ u_long res;
+ u_long tmp;
+
+ __asm __volatile(
+ "1: ldarx %0,0,%3\n"
+ " and %1,%0,%4\n"
+ " or %0,%0,%4\n"
+ " stdcx. %0,0,%3\n"
+ " bne- 1b\n"
+ : "=&r"(tmp), "=&r"(res), "+m"(*(volatile u_long *)p)
+ : "r"(p), "r"(m)
+ : "cr0", "memory");
+
+ return (res != 0);
+}
+
+static __inline int
+atomic_testandclear_long(volatile u_long *p, u_int v)
+{
+ u_long m = (1ul << (v & 0x3f));
+ u_long res;
+ u_long tmp;
+
+ __asm __volatile(
+ "1: ldarx %0,0,%3\n"
+ " and %1,%0,%4\n"
+ " andc %0,%0,%4\n"
+ " stdcx. %0,0,%3\n"
+ " bne- 1b\n"
+ : "=&r"(tmp), "=&r"(res), "+m"(*p)
+ : "r"(p), "r"(m)
+ : "cr0", "memory");
+
+ return (res != 0);
+}
+#else
+static __inline int
+atomic_testandset_long(volatile u_long *p, u_int v)
+{
+ return (atomic_testandset_int((volatile u_int *)p, v));
+}
+
+static __inline int
+atomic_testandclear_long(volatile u_long *p, u_int v)
+{
+ return (atomic_testandclear_int((volatile u_int *)p, v));
+}
+#endif
+
+#define atomic_testandclear_32 atomic_testandclear_int
+#define atomic_testandset_32 atomic_testandset_int
+
+static __inline int
+atomic_testandset_acq_long(volatile u_long *p, u_int v)
+{
+ u_int a = atomic_testandset_long(p, v);
+ __ATOMIC_ACQ();
+ return (a);
+}
+
+#ifdef __powerpc64__
+#define atomic_testandclear_ptr atomic_testandclear_long
+#define atomic_testandset_ptr atomic_testandset_long
+#else
+#define atomic_testandclear_ptr(p,v) \
+ atomic_testandclear_32((volatile u_int *)(p), v)
+#define atomic_testandset_ptr(p,v) \
+ atomic_testandset_32((volatile u_int *)(p), v)
+#endif
+
+static __inline void
+atomic_thread_fence_acq(void)
+{
+
+ powerpc_lwsync();
+}
+
+static __inline void
+atomic_thread_fence_rel(void)
+{
+
+ powerpc_lwsync();
+}
+
+static __inline void
+atomic_thread_fence_acq_rel(void)
+{
+
+ powerpc_lwsync();
+}
+
+static __inline void
+atomic_thread_fence_seq_cst(void)
+{
+
+ __asm __volatile("sync" : : : "memory");
+}
+
+#ifndef ISA_206_ATOMICS
+#include <sys/_atomic_subword.h>
+#define atomic_cmpset_char atomic_cmpset_8
+#define atomic_cmpset_short atomic_cmpset_16
+#define atomic_fcmpset_char atomic_fcmpset_8
+#define atomic_fcmpset_short atomic_fcmpset_16
+#define atomic_set_short atomic_set_16
+#define atomic_clear_short atomic_clear_16
+#else
+
+static __inline void
+atomic_set_short(volatile u_short *p, u_short bit)
+{
+ u_short v;
+
+ v = atomic_load_short(p);
+ for (;;) {
+ if (atomic_fcmpset_16(p, &v, v | bit))
+ break;
+ }
+}
+
+static __inline void
+atomic_clear_short(volatile u_short *p, u_short bit)
+{
+ u_short v;
+
+ v = atomic_load_short(p);
+ for (;;) {
+ if (atomic_fcmpset_16(p, &v, v & ~bit))
+ break;
+ }
+}
+
+#define atomic_set_16 atomic_set_short
+#define atomic_clear_16 atomic_clear_short
+
+#endif /* ISA_206_ATOMICS */
+
+/* These need sys/_atomic_subword.h on non-ISA-2.06-atomic platforms. */
+ATOMIC_CMPSET_ACQ_REL(char);
+ATOMIC_CMPSET_ACQ_REL(short);
+
+ATOMIC_FCMPSET_ACQ_REL(char);
+ATOMIC_FCMPSET_ACQ_REL(short);
+
+#undef __ATOMIC_REL
+#undef __ATOMIC_ACQ
+
+#endif /* ! _MACHINE_ATOMIC_H_ */