123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281 |
- /*
- * Copyright (c) 2003-2011 Hewlett-Packard Development Company, L.P.
- *
- * Permission is hereby granted, free of charge, to any person obtaining a copy
- * of this software and associated documentation files (the "Software"), to deal
- * in the Software without restriction, including without limitation the rights
- * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
- * copies of the Software, and to permit persons to whom the Software is
- * furnished to do so, subject to the following conditions:
- *
- * The above copyright notice and this permission notice shall be included in
- * all copies or substantial portions of the Software.
- *
- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
- * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
- * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
- * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
- * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
- * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
- * SOFTWARE.
- */
- #include "../all_atomic_load_store.h"
- #include "../all_acquire_release_volatile.h"
- #include "../test_and_set_t_is_char.h"
- #ifdef _ILP32
- /* 32-bit HP/UX code. */
- /* This requires pointer "swizzling". Pointers need to be expanded */
- /* to 64 bits using the addp4 instruction before use. This makes it */
- /* hard to share code, but we try anyway. */
- # define AO_LEN "4"
- /* We assume that addr always appears in argument position 1 in asm */
- /* code. If it is clobbered due to swizzling, we also need it in */
- /* second position. Any later arguments are referenced symbolically, */
- /* so that we don't have to worry about their position. This requires*/
- /* gcc 3.1, but you shouldn't be using anything older than that on */
- /* IA64 anyway. */
- /* The AO_MASK macro is a workaround for the fact that HP/UX gcc */
- /* appears to otherwise store 64-bit pointers in ar.ccv, i.e. it */
- /* doesn't appear to clear high bits in a pointer value we pass into */
- /* assembly code, even if it is supposedly of type AO_t. */
- # define AO_IN_ADDR "1"(addr)
- # define AO_OUT_ADDR , "=r"(addr)
- # define AO_SWIZZLE "addp4 %1=0,%1;;\n"
- # define AO_MASK(ptr) __asm__ __volatile__("zxt4 %1=%1": "=r"(ptr) : "0"(ptr))
- #else
- # define AO_LEN "8"
- # define AO_IN_ADDR "r"(addr)
- # define AO_OUT_ADDR
- # define AO_SWIZZLE
- # define AO_MASK(ptr) /* empty */
- #endif /* !_ILP32 */
- AO_INLINE void
- AO_nop_full(void)
- {
- __asm__ __volatile__("mf" : : : "memory");
- }
- #define AO_HAVE_nop_full
- #ifndef AO_PREFER_GENERALIZED
- AO_INLINE AO_t
- AO_fetch_and_add1_acquire (volatile AO_t *addr)
- {
- AO_t result;
- __asm__ __volatile__ (AO_SWIZZLE
- "fetchadd" AO_LEN ".acq %0=[%1],1":
- "=r" (result) AO_OUT_ADDR: AO_IN_ADDR :"memory");
- return result;
- }
- #define AO_HAVE_fetch_and_add1_acquire
- AO_INLINE AO_t
- AO_fetch_and_add1_release (volatile AO_t *addr)
- {
- AO_t result;
- __asm__ __volatile__ (AO_SWIZZLE
- "fetchadd" AO_LEN ".rel %0=[%1],1":
- "=r" (result) AO_OUT_ADDR: AO_IN_ADDR :"memory");
- return result;
- }
- #define AO_HAVE_fetch_and_add1_release
- AO_INLINE AO_t
- AO_fetch_and_sub1_acquire (volatile AO_t *addr)
- {
- AO_t result;
- __asm__ __volatile__ (AO_SWIZZLE
- "fetchadd" AO_LEN ".acq %0=[%1],-1":
- "=r" (result) AO_OUT_ADDR: AO_IN_ADDR :"memory");
- return result;
- }
- #define AO_HAVE_fetch_and_sub1_acquire
- AO_INLINE AO_t
- AO_fetch_and_sub1_release (volatile AO_t *addr)
- {
- AO_t result;
- __asm__ __volatile__ (AO_SWIZZLE
- "fetchadd" AO_LEN ".rel %0=[%1],-1":
- "=r" (result) AO_OUT_ADDR: AO_IN_ADDR :"memory");
- return result;
- }
- #define AO_HAVE_fetch_and_sub1_release
- #endif /* !AO_PREFER_GENERALIZED */
- AO_INLINE AO_t
- AO_fetch_compare_and_swap_acquire(volatile AO_t *addr, AO_t old, AO_t new_val)
- {
- AO_t fetched_val;
- AO_MASK(old);
- __asm__ __volatile__(AO_SWIZZLE
- "mov ar.ccv=%[old] ;; cmpxchg" AO_LEN
- ".acq %0=[%1],%[new_val],ar.ccv"
- : "=r"(fetched_val) AO_OUT_ADDR
- : AO_IN_ADDR, [new_val]"r"(new_val), [old]"r"(old)
- : "memory");
- return fetched_val;
- }
- #define AO_HAVE_fetch_compare_and_swap_acquire
- AO_INLINE AO_t
- AO_fetch_compare_and_swap_release(volatile AO_t *addr, AO_t old, AO_t new_val)
- {
- AO_t fetched_val;
- AO_MASK(old);
- __asm__ __volatile__(AO_SWIZZLE
- "mov ar.ccv=%[old] ;; cmpxchg" AO_LEN
- ".rel %0=[%1],%[new_val],ar.ccv"
- : "=r"(fetched_val) AO_OUT_ADDR
- : AO_IN_ADDR, [new_val]"r"(new_val), [old]"r"(old)
- : "memory");
- return fetched_val;
- }
- #define AO_HAVE_fetch_compare_and_swap_release
- AO_INLINE unsigned char
- AO_char_fetch_compare_and_swap_acquire(volatile unsigned char *addr,
- unsigned char old, unsigned char new_val)
- {
- unsigned char fetched_val;
- __asm__ __volatile__(AO_SWIZZLE
- "mov ar.ccv=%[old] ;; cmpxchg1.acq %0=[%1],%[new_val],ar.ccv"
- : "=r"(fetched_val) AO_OUT_ADDR
- : AO_IN_ADDR, [new_val]"r"(new_val), [old]"r"((AO_t)old)
- : "memory");
- return fetched_val;
- }
- #define AO_HAVE_char_fetch_compare_and_swap_acquire
- AO_INLINE unsigned char
- AO_char_fetch_compare_and_swap_release(volatile unsigned char *addr,
- unsigned char old, unsigned char new_val)
- {
- unsigned char fetched_val;
- __asm__ __volatile__(AO_SWIZZLE
- "mov ar.ccv=%[old] ;; cmpxchg1.rel %0=[%1],%[new_val],ar.ccv"
- : "=r"(fetched_val) AO_OUT_ADDR
- : AO_IN_ADDR, [new_val]"r"(new_val), [old]"r"((AO_t)old)
- : "memory");
- return fetched_val;
- }
- #define AO_HAVE_char_fetch_compare_and_swap_release
- AO_INLINE unsigned short
- AO_short_fetch_compare_and_swap_acquire(volatile unsigned short *addr,
- unsigned short old, unsigned short new_val)
- {
- unsigned short fetched_val;
- __asm__ __volatile__(AO_SWIZZLE
- "mov ar.ccv=%[old] ;; cmpxchg2.acq %0=[%1],%[new_val],ar.ccv"
- : "=r"(fetched_val) AO_OUT_ADDR
- : AO_IN_ADDR, [new_val]"r"(new_val), [old]"r"((AO_t)old)
- : "memory");
- return fetched_val;
- }
- #define AO_HAVE_short_fetch_compare_and_swap_acquire
- AO_INLINE unsigned short
- AO_short_fetch_compare_and_swap_release(volatile unsigned short *addr,
- unsigned short old, unsigned short new_val)
- {
- unsigned short fetched_val;
- __asm__ __volatile__(AO_SWIZZLE
- "mov ar.ccv=%[old] ;; cmpxchg2.rel %0=[%1],%[new_val],ar.ccv"
- : "=r"(fetched_val) AO_OUT_ADDR
- : AO_IN_ADDR, [new_val]"r"(new_val), [old]"r"((AO_t)old)
- : "memory");
- return fetched_val;
- }
- #define AO_HAVE_short_fetch_compare_and_swap_release
- #ifdef _ILP32
- # define AO_T_IS_INT
- /* TODO: Add compare_double_and_swap_double for the _ILP32 case. */
- #else
- # ifndef AO_PREFER_GENERALIZED
- AO_INLINE unsigned int
- AO_int_fetch_and_add1_acquire(volatile unsigned int *addr)
- {
- unsigned int result;
- __asm__ __volatile__("fetchadd4.acq %0=[%1],1"
- : "=r" (result) : AO_IN_ADDR
- : "memory");
- return result;
- }
- # define AO_HAVE_int_fetch_and_add1_acquire
- AO_INLINE unsigned int
- AO_int_fetch_and_add1_release(volatile unsigned int *addr)
- {
- unsigned int result;
- __asm__ __volatile__("fetchadd4.rel %0=[%1],1"
- : "=r" (result) : AO_IN_ADDR
- : "memory");
- return result;
- }
- # define AO_HAVE_int_fetch_and_add1_release
- AO_INLINE unsigned int
- AO_int_fetch_and_sub1_acquire(volatile unsigned int *addr)
- {
- unsigned int result;
- __asm__ __volatile__("fetchadd4.acq %0=[%1],-1"
- : "=r" (result) : AO_IN_ADDR
- : "memory");
- return result;
- }
- # define AO_HAVE_int_fetch_and_sub1_acquire
- AO_INLINE unsigned int
- AO_int_fetch_and_sub1_release(volatile unsigned int *addr)
- {
- unsigned int result;
- __asm__ __volatile__("fetchadd4.rel %0=[%1],-1"
- : "=r" (result) : AO_IN_ADDR
- : "memory");
- return result;
- }
- # define AO_HAVE_int_fetch_and_sub1_release
- # endif /* !AO_PREFER_GENERALIZED */
- AO_INLINE unsigned int
- AO_int_fetch_compare_and_swap_acquire(volatile unsigned int *addr,
- unsigned int old, unsigned int new_val)
- {
- unsigned int fetched_val;
- __asm__ __volatile__("mov ar.ccv=%3 ;; cmpxchg4.acq %0=[%1],%2,ar.ccv"
- : "=r"(fetched_val)
- : AO_IN_ADDR, "r"(new_val), "r"((AO_t)old)
- : "memory");
- return fetched_val;
- }
- # define AO_HAVE_int_fetch_compare_and_swap_acquire
- AO_INLINE unsigned int
- AO_int_fetch_compare_and_swap_release(volatile unsigned int *addr,
- unsigned int old, unsigned int new_val)
- {
- unsigned int fetched_val;
- __asm__ __volatile__("mov ar.ccv=%3 ;; cmpxchg4.rel %0=[%1],%2,ar.ccv"
- : "=r"(fetched_val)
- : AO_IN_ADDR, "r"(new_val), "r"((AO_t)old)
- : "memory");
- return fetched_val;
- }
- # define AO_HAVE_int_fetch_compare_and_swap_release
- #endif /* !_ILP32 */
- /* TODO: Add compare_and_swap_double as soon as there is widely */
- /* available hardware that implements it. */
|