| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627 | /* Atomic operations.  PowerPC Common version.   Copyright (C) 2003, 2004 Free Software Foundation, Inc.   This file is part of the GNU C Library.   Contributed by Paul Mackerras <paulus@au.ibm.com>, 2003.   The GNU C Library is free software; you can redistribute it and/or   modify it under the terms of the GNU Lesser General Public   License as published by the Free Software Foundation; either   version 2.1 of the License, or (at your option) any later version.   The GNU C Library is distributed in the hope that it will be useful,   but WITHOUT ANY WARRANTY; without even the implied warranty of   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU   Lesser General Public License for more details.   You should have received a copy of the GNU Lesser General Public   License along with the GNU C Library; if not, see   <http://www.gnu.org/licenses/>.  */#include <bits/wordsize.h>#if __WORDSIZE == 64/* Atomic operations.  PowerPC64 version.   Copyright (C) 2003, 2004 Free Software Foundation, Inc.   This file is part of the GNU C Library.   Contributed by Paul Mackerras <paulus@au.ibm.com>, 2003.   The GNU C Library is free software; you can redistribute it and/or   modify it under the terms of the GNU Lesser General Public   License as published by the Free Software Foundation; either   version 2.1 of the License, or (at your option) any later version.   The GNU C Library is distributed in the hope that it will be useful,   but WITHOUT ANY WARRANTY; without even the implied warranty of   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU   Lesser General Public License for more details.   You should have received a copy of the GNU Lesser General Public   License along with the GNU C Library; if not, see   <http://www.gnu.org/licenses/>.  *//* The 32-bit exchange_bool is different on powerpc64 because the subf   does signed 64-bit arthmatic while the lwarx is 32-bit unsigned   (a load word and zero (high 32) form) load.   In powerpc64 register values are 64-bit by default,  including oldval.   The value in old val unknown sign extension, lwarx loads the 32-bit   value as unsigned.  So we explicitly clear the high 32 bits in oldval.  */# define __arch_compare_and_exchange_bool_32_acq(mem, newval, oldval) \({									      \  unsigned int __tmp, __tmp2;						      \  __asm__ __volatile__ ("   clrldi  %1,%1,32\n"				      \		    "1:	lwarx	%0,0,%2\n"				      \		    "	subf.	%0,%1,%0\n"				      \		    "	bne	2f\n"					      \		    "	stwcx.	%4,0,%2\n"				      \		    "	bne-	1b\n"					      \		    "2:	" __ARCH_ACQ_INSTR				      \		    : "=&r" (__tmp), "=r" (__tmp2)			      \		    : "b" (mem), "1" (oldval), "r" (newval)		      \		    : "cr0", "memory");					      \  __tmp != 0;								      \})# define __arch_compare_and_exchange_bool_32_rel(mem, newval, oldval) \({									      \  unsigned int __tmp, __tmp2;						      \  __asm__ __volatile__ (__ARCH_REL_INSTR "\n"				      \		    "   clrldi  %1,%1,32\n"				      \		    "1:	lwarx	%0,0,%2\n"				      \		    "	subf.	%0,%1,%0\n"				      \		    "	bne	2f\n"					      \		    "	stwcx.	%4,0,%2\n"				      \		    "	bne-	1b\n"					      \		    "2:	"						      \		    : "=&r" (__tmp), "=r" (__tmp2)			      \		    : "b" (mem), "1" (oldval), "r" (newval)		      \		    : "cr0", "memory");					      \  __tmp != 0;								      \})/* * Only powerpc64 processors support Load doubleword and reserve index (ldarx) * and Store doubleword conditional indexed (stdcx) instructions.  So here * we define the 64-bit forms. */# define __arch_compare_and_exchange_bool_64_acq(mem, newval, oldval) \({									      \  unsigned long	__tmp;							      \  __asm__ __volatile__ (							      \		    "1:	ldarx	%0,0,%1\n"				      \		    "	subf.	%0,%2,%0\n"				      \		    "	bne	2f\n"					      \		    "	stdcx.	%3,0,%1\n"				      \		    "	bne-	1b\n"					      \		    "2:	" __ARCH_ACQ_INSTR				      \		    : "=&r" (__tmp)					      \		    : "b" (mem), "r" (oldval), "r" (newval)		      \		    : "cr0", "memory");					      \  __tmp != 0;								      \})# define __arch_compare_and_exchange_bool_64_rel(mem, newval, oldval) \({									      \  unsigned long	__tmp;							      \  __asm__ __volatile__ (__ARCH_REL_INSTR "\n"				      \		    "1:	ldarx	%0,0,%1\n"				      \		    "	subf.	%0,%2,%0\n"				      \		    "	bne	2f\n"					      \		    "	stdcx.	%3,0,%1\n"				      \		    "	bne-	1b\n"					      \		    "2:	"						      \		    : "=&r" (__tmp)					      \		    : "b" (mem), "r" (oldval), "r" (newval)		      \		    : "cr0", "memory");					      \  __tmp != 0;								      \})#define __arch_compare_and_exchange_val_64_acq(mem, newval, oldval) \  ({									      \      __typeof (*(mem)) __tmp;						      \      __typeof (mem)  __memp = (mem);					      \      __asm__ __volatile__ (						      \		        "1:	ldarx	%0,0,%1\n"			      \		        "	cmpd	%0,%2\n"			      \		        "	bne	2f\n"				      \		        "	stdcx.	%3,0,%1\n"			      \		        "	bne-	1b\n"				      \		        "2:	" __ARCH_ACQ_INSTR			      \		        : "=&r" (__tmp)					      \		        : "b" (__memp), "r" (oldval), "r" (newval)	      \		        : "cr0", "memory");				      \      __tmp;								      \  })#define __arch_compare_and_exchange_val_64_rel(mem, newval, oldval) \  ({									      \      __typeof (*(mem)) __tmp;						      \      __typeof (mem)  __memp = (mem);					      \      __asm__ __volatile__ (__ARCH_REL_INSTR "\n"				      \		        "1:	ldarx	%0,0,%1\n"			      \		        "	cmpd	%0,%2\n"			      \		        "	bne	2f\n"				      \		        "	stdcx.	%3,0,%1\n"			      \		        "	bne-	1b\n"				      \		        "2:	"					      \		        : "=&r" (__tmp)					      \		        : "b" (__memp), "r" (oldval), "r" (newval)	      \		        : "cr0", "memory");				      \      __tmp;								      \  })# define __arch_atomic_exchange_64_acq(mem, value) \    ({									      \      __typeof (*mem) __val;						      \      __asm__ __volatile__ (__ARCH_REL_INSTR "\n"				      \			"1:	ldarx	%0,0,%2\n"			      \			"	stdcx.	%3,0,%2\n"			      \			"	bne-	1b\n"				      \		  " " __ARCH_ACQ_INSTR					      \			: "=&r" (__val), "=m" (*mem)			      \			: "b" (mem), "r" (value), "m" (*mem)		      \			: "cr0", "memory");				      \      __val;								      \    })# define __arch_atomic_exchange_64_rel(mem, value) \    ({									      \      __typeof (*mem) __val;						      \      __asm__ __volatile__ (__ARCH_REL_INSTR "\n"				      \			"1:	ldarx	%0,0,%2\n"			      \			"	stdcx.	%3,0,%2\n"			      \			"	bne-	1b"				      \			: "=&r" (__val), "=m" (*mem)			      \			: "b" (mem), "r" (value), "m" (*mem)		      \			: "cr0", "memory");				      \      __val;								      \    })# define __arch_atomic_exchange_and_add_64(mem, value) \    ({									      \      __typeof (*mem) __val, __tmp;					      \      __asm__ __volatile__ ("1:	ldarx	%0,0,%3\n"			      \			"	add	%1,%0,%4\n"			      \			"	stdcx.	%1,0,%3\n"			      \			"	bne-	1b"				      \			: "=&b" (__val), "=&r" (__tmp), "=m" (*mem)	      \			: "b" (mem), "r" (value), "m" (*mem)		      \			: "cr0", "memory");				      \      __val;								      \    })# define __arch_atomic_increment_val_64(mem) \    ({									      \      __typeof (*(mem)) __val;						      \      __asm__ __volatile__ ("1:	ldarx	%0,0,%2\n"			      \			"	addi	%0,%0,1\n"			      \			"	stdcx.	%0,0,%2\n"			      \			"	bne-	1b"				      \			: "=&b" (__val), "=m" (*mem)			      \			: "b" (mem), "m" (*mem)				      \			: "cr0", "memory");				      \      __val;								      \    })# define __arch_atomic_decrement_val_64(mem) \    ({									      \      __typeof (*(mem)) __val;						      \      __asm__ __volatile__ ("1:	ldarx	%0,0,%2\n"			      \			"	subi	%0,%0,1\n"			      \			"	stdcx.	%0,0,%2\n"			      \			"	bne-	1b"				      \			: "=&b" (__val), "=m" (*mem)			      \			: "b" (mem), "m" (*mem)				      \			: "cr0", "memory");				      \      __val;								      \    })# define __arch_atomic_decrement_if_positive_64(mem) \  ({ int __val, __tmp;							      \     __asm__ __volatile__ ("1:	ldarx	%0,0,%3\n"			      \		       "	cmpdi	0,%0,0\n"			      \		       "	addi	%1,%0,-1\n"			      \		       "	ble	2f\n"				      \		       "	stdcx.	%1,0,%3\n"			      \		       "	bne-	1b\n"				      \		       "2:	" __ARCH_ACQ_INSTR			      \		       : "=&b" (__val), "=&r" (__tmp), "=m" (*mem)	      \		       : "b" (mem), "m" (*mem)				      \		       : "cr0", "memory");				      \     __val;								      \  })/* * All powerpc64 processors support the new "light weight"  sync (lwsync). */# define atomic_read_barrier()	__asm__ ("lwsync" ::: "memory")/* * "light weight" sync can also be used for the release barrier. */# ifndef UP#  define __ARCH_REL_INSTR	"lwsync"# endif#else/* Atomic operations.  PowerPC32 version.   Copyright (C) 2003, 2004 Free Software Foundation, Inc.   This file is part of the GNU C Library.   Contributed by Paul Mackerras <paulus@au.ibm.com>, 2003.   The GNU C Library is free software; you can redistribute it and/or   modify it under the terms of the GNU Lesser General Public   License as published by the Free Software Foundation; either   version 2.1 of the License, or (at your option) any later version.   The GNU C Library is distributed in the hope that it will be useful,   but WITHOUT ANY WARRANTY; without even the implied warranty of   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU   Lesser General Public License for more details.   You should have received a copy of the GNU Lesser General Public   License along with the GNU C Library; if not, see   <http://www.gnu.org/licenses/>.  *//* * The 32-bit exchange_bool is different on powerpc64 because the subf * does signed 64-bit arthmatic while the lwarx is 32-bit unsigned * (a load word and zero (high 32) form).  So powerpc64 has a slightly * different version in sysdeps/powerpc/powerpc64/bits/atomic.h. */# define __arch_compare_and_exchange_bool_32_acq(mem, newval, oldval)         \({									      \  unsigned int __tmp;							      \  __asm__ __volatile__ (							      \		    "1:	lwarx	%0,0,%1\n"				      \		    "	subf.	%0,%2,%0\n"				      \		    "	bne	2f\n"					      \		    "	stwcx.	%3,0,%1\n"				      \		    "	bne-	1b\n"					      \		    "2:	" __ARCH_ACQ_INSTR				      \		    : "=&r" (__tmp)					      \		    : "b" (mem), "r" (oldval), "r" (newval)		      \		    : "cr0", "memory");					      \  __tmp != 0;								      \})# define __arch_compare_and_exchange_bool_32_rel(mem, newval, oldval)	      \({									      \  unsigned int __tmp;							      \  __asm__ __volatile__ (__ARCH_REL_INSTR "\n"				      \		    "1:	lwarx	%0,0,%1\n"				      \		    "	subf.	%0,%2,%0\n"				      \		    "	bne	2f\n"					      \		    "	stwcx.	%3,0,%1\n"				      \		    "	bne-	1b\n"					      \		    "2:	"						      \		    : "=&r" (__tmp)					      \		    : "b" (mem), "r" (oldval), "r" (newval)		      \		    : "cr0", "memory");					      \  __tmp != 0;								      \})/* Powerpc32 processors don't implement the 64-bit (doubleword) forms of   load and reserve (ldarx) and store conditional (stdcx.) instructions.   So for powerpc32 we stub out the 64-bit forms.  */# define __arch_compare_and_exchange_bool_64_acq(mem, newval, oldval) \  (abort (), 0)# define __arch_compare_and_exchange_val_64_acq(mem, newval, oldval) \  (abort (), (__typeof (*mem)) 0)# define __arch_compare_and_exchange_bool_64_rel(mem, newval, oldval) \  (abort (), 0)# define __arch_compare_and_exchange_val_64_rel(mem, newval, oldval) \  (abort (), (__typeof (*mem)) 0)# define __arch_atomic_exchange_64_acq(mem, value) \    ({ abort (); (*mem) = (value); })# define __arch_atomic_exchange_64_rel(mem, value) \    ({ abort (); (*mem) = (value); })# define __arch_atomic_exchange_and_add_64(mem, value) \    ({ abort (); (*mem) = (value); })# define __arch_atomic_increment_val_64(mem) \    ({ abort (); (*mem)++; })# define __arch_atomic_decrement_val_64(mem) \    ({ abort (); (*mem)--; })# define __arch_atomic_decrement_if_positive_64(mem) \    ({ abort (); (*mem)--; })#ifdef _ARCH_PWR4/* * Newer powerpc64 processors support the new "light weight" sync (lwsync) * So if the build is using -mcpu=[power4,power5,power5+,970] we can * safely use lwsync. */# define atomic_read_barrier()	__asm__ ("lwsync" ::: "memory")/* * "light weight" sync can also be used for the release barrier. */# ifndef UP#  define __ARCH_REL_INSTR	"lwsync"# endif#else/* * Older powerpc32 processors don't support the new "light weight" * sync (lwsync).  So the only safe option is to use normal sync * for all powerpc32 applications. */# define atomic_read_barrier()	__asm__ ("sync" ::: "memory")#endif#endif#include <stdint.h>typedef int32_t atomic32_t;typedef uint32_t uatomic32_t;typedef int_fast32_t atomic_fast32_t;typedef uint_fast32_t uatomic_fast32_t;typedef int64_t atomic64_t;typedef uint64_t uatomic64_t;typedef int_fast64_t atomic_fast64_t;typedef uint_fast64_t uatomic_fast64_t;typedef intptr_t atomicptr_t;typedef uintptr_t uatomicptr_t;typedef intmax_t atomic_max_t;typedef uintmax_t uatomic_max_t;/* * Powerpc does not have byte and halfword forms of load and reserve and * store conditional. So for powerpc we stub out the 8- and 16-bit forms. */#define __arch_compare_and_exchange_bool_8_acq(mem, newval, oldval) \  (abort (), 0)#define __arch_compare_and_exchange_bool_16_acq(mem, newval, oldval) \  (abort (), 0)#define __arch_compare_and_exchange_bool_8_rel(mem, newval, oldval) \  (abort (), 0)#define __arch_compare_and_exchange_bool_16_rel(mem, newval, oldval) \  (abort (), 0)#ifdef UP# define __ARCH_ACQ_INSTR	""# define __ARCH_REL_INSTR	""#else# define __ARCH_ACQ_INSTR	"isync"# ifndef __ARCH_REL_INSTR#  define __ARCH_REL_INSTR	"sync"# endif#endif#ifndef MUTEX_HINT_ACQ# define MUTEX_HINT_ACQ#endif#ifndef MUTEX_HINT_REL# define MUTEX_HINT_REL#endif#define atomic_full_barrier()	__asm__ ("sync" ::: "memory")#define atomic_write_barrier()	__asm__ ("eieio" ::: "memory")#define __arch_compare_and_exchange_val_32_acq(mem, newval, oldval)	      \  ({									      \      __typeof (*(mem)) __tmp;						      \      __typeof (mem)  __memp = (mem);					      \      __asm__ __volatile__ (						      \		        "1:	lwarx	%0,0,%1\n"			      \		        "	cmpw	%0,%2\n"			      \		        "	bne	2f\n"				      \		        "	stwcx.	%3,0,%1\n"			      \		        "	bne-	1b\n"				      \		        "2:	" __ARCH_ACQ_INSTR			      \		        : "=&r" (__tmp)					      \		        : "b" (__memp), "r" (oldval), "r" (newval)	      \		        : "cr0", "memory");				      \      __tmp;								      \  })#define __arch_compare_and_exchange_val_32_rel(mem, newval, oldval)	      \  ({									      \      __typeof (*(mem)) __tmp;						      \      __typeof (mem)  __memp = (mem);					      \      __asm__ __volatile__ (__ARCH_REL_INSTR "\n"				      \		        "1:	lwarx	%0,0,%1\n"			      \		        "	cmpw	%0,%2\n"			      \		        "	bne	2f\n"				      \		        "	stwcx.	%3,0,%1\n"			      \		        "	bne-	1b\n"				      \		        "2:	"					      \		        : "=&r" (__tmp)					      \		        : "b" (__memp), "r" (oldval), "r" (newval)	      \		        : "cr0", "memory");				      \      __tmp;								      \  })#define __arch_atomic_exchange_32_acq(mem, value)			      \  ({									      \    __typeof (*mem) __val;						      \    __asm__ __volatile__ (							      \		      "1:	lwarx	%0,0,%2\n"			      \		      "		stwcx.	%3,0,%2\n"			      \		      "		bne-	1b\n"				      \		      "   " __ARCH_ACQ_INSTR				      \		      : "=&r" (__val), "=m" (*mem)			      \		      : "b" (mem), "r" (value), "m" (*mem)		      \		      : "cr0", "memory");				      \    __val;								      \  })#define __arch_atomic_exchange_32_rel(mem, value) \  ({									      \    __typeof (*mem) __val;						      \    __asm__ __volatile__ (__ARCH_REL_INSTR "\n"				      \		      "1:	lwarx	%0,0,%2\n"			      \		      "		stwcx.	%3,0,%2\n"			      \		      "		bne-	1b"				      \		      : "=&r" (__val), "=m" (*mem)			      \		      : "b" (mem), "r" (value), "m" (*mem)		      \		      : "cr0", "memory");				      \    __val;								      \  })#define __arch_atomic_exchange_and_add_32(mem, value) \  ({									      \    __typeof (*mem) __val, __tmp;					      \    __asm__ __volatile__ ("1:	lwarx	%0,0,%3\n"			      \		      "		add	%1,%0,%4\n"			      \		      "		stwcx.	%1,0,%3\n"			      \		      "		bne-	1b"				      \		      : "=&b" (__val), "=&r" (__tmp), "=m" (*mem)	      \		      : "b" (mem), "r" (value), "m" (*mem)		      \		      : "cr0", "memory");				      \    __val;								      \  })#define __arch_atomic_increment_val_32(mem) \  ({									      \    __typeof (*(mem)) __val;						      \    __asm__ __volatile__ ("1:	lwarx	%0,0,%2\n"			      \		      "		addi	%0,%0,1\n"			      \		      "		stwcx.	%0,0,%2\n"			      \		      "		bne-	1b"				      \		      : "=&b" (__val), "=m" (*mem)			      \		      : "b" (mem), "m" (*mem)				      \		      : "cr0", "memory");				      \    __val;								      \  })#define __arch_atomic_decrement_val_32(mem) \  ({									      \    __typeof (*(mem)) __val;						      \    __asm__ __volatile__ ("1:	lwarx	%0,0,%2\n"			      \		      "		subi	%0,%0,1\n"			      \		      "		stwcx.	%0,0,%2\n"			      \		      "		bne-	1b"				      \		      : "=&b" (__val), "=m" (*mem)			      \		      : "b" (mem), "m" (*mem)				      \		      : "cr0", "memory");				      \    __val;								      \  })#define __arch_atomic_decrement_if_positive_32(mem) \  ({ int __val, __tmp;							      \     __asm__ __volatile__ ("1:	lwarx	%0,0,%3\n"			      \		       "	cmpwi	0,%0,0\n"			      \		       "	addi	%1,%0,-1\n"			      \		       "	ble	2f\n"				      \		       "	stwcx.	%1,0,%3\n"			      \		       "	bne-	1b\n"				      \		       "2:	" __ARCH_ACQ_INSTR			      \		       : "=&b" (__val), "=&r" (__tmp), "=m" (*mem)	      \		       : "b" (mem), "m" (*mem)				      \		       : "cr0", "memory");				      \     __val;								      \  })#define atomic_compare_and_exchange_val_acq(mem, newval, oldval) \  ({									      \    __typeof (*(mem)) __result;						      \    if (sizeof (*mem) == 4)						      \      __result = __arch_compare_and_exchange_val_32_acq(mem, newval, oldval); \    else if (sizeof (*mem) == 8)					      \      __result = __arch_compare_and_exchange_val_64_acq(mem, newval, oldval); \    else								      \       abort ();							      \    __result;								      \  })#define atomic_compare_and_exchange_val_rel(mem, newval, oldval) \  ({									      \    __typeof (*(mem)) __result;						      \    if (sizeof (*mem) == 4)						      \      __result = __arch_compare_and_exchange_val_32_rel(mem, newval, oldval); \    else if (sizeof (*mem) == 8)					      \      __result = __arch_compare_and_exchange_val_64_rel(mem, newval, oldval); \    else								      \       abort ();							      \    __result;								      \  })#define atomic_exchange_acq(mem, value) \  ({									      \    __typeof (*(mem)) __result;						      \    if (sizeof (*mem) == 4)						      \      __result = __arch_atomic_exchange_32_acq (mem, value);		      \    else if (sizeof (*mem) == 8)					      \      __result = __arch_atomic_exchange_64_acq (mem, value);		      \    else								      \       abort ();							      \    __result;								      \  })#define atomic_exchange_rel(mem, value) \  ({									      \    __typeof (*(mem)) __result;						      \    if (sizeof (*mem) == 4)						      \      __result = __arch_atomic_exchange_32_rel (mem, value);		      \    else if (sizeof (*mem) == 8)					      \      __result = __arch_atomic_exchange_64_rel (mem, value);		      \    else								      \       abort ();							      \    __result;								      \  })#define atomic_exchange_and_add(mem, value) \  ({									      \    __typeof (*(mem)) __result;						      \    if (sizeof (*mem) == 4)						      \      __result = __arch_atomic_exchange_and_add_32 (mem, value);	      \    else if (sizeof (*mem) == 8)					      \      __result = __arch_atomic_exchange_and_add_64 (mem, value);	      \    else								      \       abort ();							      \    __result;								      \  })#define atomic_increment_val(mem) \  ({									      \    __typeof (*(mem)) __result;						      \    if (sizeof (*(mem)) == 4)						      \      __result = __arch_atomic_increment_val_32 (mem);			      \    else if (sizeof (*(mem)) == 8)					      \      __result = __arch_atomic_increment_val_64 (mem);			      \    else								      \       abort ();							      \    __result;								      \  })#define atomic_increment(mem) ({ atomic_increment_val (mem); (void) 0; })#define atomic_decrement_val(mem) \  ({									      \    __typeof (*(mem)) __result;						      \    if (sizeof (*(mem)) == 4)						      \      __result = __arch_atomic_decrement_val_32 (mem);			      \    else if (sizeof (*(mem)) == 8)					      \      __result = __arch_atomic_decrement_val_64 (mem);			      \    else								      \       abort ();							      \    __result;								      \  })#define atomic_decrement(mem) ({ atomic_decrement_val (mem); (void) 0; })/* Decrement *MEM if it is > 0, and return the old value.  */#define atomic_decrement_if_positive(mem) \  ({ __typeof (*(mem)) __result;					      \    if (sizeof (*mem) == 4)						      \      __result = __arch_atomic_decrement_if_positive_32 (mem);		      \    else if (sizeof (*mem) == 8)					      \      __result = __arch_atomic_decrement_if_positive_64 (mem);		      \    else								      \       abort ();							      \    __result;								      \  })
 |