123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113 |
- /*
- Copyright 2005-2013 Intel Corporation. All Rights Reserved.
- This file is part of Threading Building Blocks.
- Threading Building Blocks is free software; you can redistribute it
- and/or modify it under the terms of the GNU General Public License
- version 2 as published by the Free Software Foundation.
- Threading Building Blocks is distributed in the hope that it will be
- useful, but WITHOUT ANY WARRANTY; without even the implied warranty
- of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- GNU General Public License for more details.
- You should have received a copy of the GNU General Public License
- along with Threading Building Blocks; if not, write to the Free Software
- Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
- As a special exception, you may use this file as part of a free software
- library without restriction. Specifically, if other files instantiate
- templates or use macros or inline functions from this file, or you compile
- this file and link it with other files to produce an executable, this
- file does not by itself cause the resulting executable to be covered by
- the GNU General Public License. This exception does not however
- invalidate any other reasons why the executable file might be covered by
- the GNU General Public License.
- */
- #if !defined(__TBB_machine_H) || defined(__TBB_machine_windows_intel64_H)
- #error Do not #include this internal file directly; use public TBB headers instead.
- #endif
- #define __TBB_machine_windows_intel64_H
- #define __TBB_WORDSIZE 8
- #define __TBB_BIG_ENDIAN 0
- #include <intrin.h>
- #include "msvc_ia32_common.h"
- //TODO: Use _InterlockedXXX16 intrinsics for 2 byte operations
- #if !__INTEL_COMPILER
- #pragma intrinsic(_InterlockedOr64)
- #pragma intrinsic(_InterlockedAnd64)
- #pragma intrinsic(_InterlockedCompareExchange)
- #pragma intrinsic(_InterlockedCompareExchange64)
- #pragma intrinsic(_InterlockedExchangeAdd)
- #pragma intrinsic(_InterlockedExchangeAdd64)
- #pragma intrinsic(_InterlockedExchange)
- #pragma intrinsic(_InterlockedExchange64)
- #endif /* !(__INTEL_COMPILER) */
- #if __INTEL_COMPILER && (__INTEL_COMPILER < 1100)
- #define __TBB_compiler_fence() __asm { __asm nop }
- #define __TBB_full_memory_fence() __asm { __asm mfence }
- #elif _MSC_VER >= 1300 || __INTEL_COMPILER
- #pragma intrinsic(_ReadWriteBarrier)
- #pragma intrinsic(_mm_mfence)
- #define __TBB_compiler_fence() _ReadWriteBarrier()
- #define __TBB_full_memory_fence() _mm_mfence()
- #endif
- #define __TBB_control_consistency_helper() __TBB_compiler_fence()
- #define __TBB_acquire_consistency_helper() __TBB_compiler_fence()
- #define __TBB_release_consistency_helper() __TBB_compiler_fence()
- // ATTENTION: if you ever change argument types in machine-specific primitives,
- // please take care of atomic_word<> specializations in tbb/atomic.h
- extern "C" {
- __int8 __TBB_EXPORTED_FUNC __TBB_machine_cmpswp1 (volatile void *ptr, __int8 value, __int8 comparand );
- __int8 __TBB_EXPORTED_FUNC __TBB_machine_fetchadd1 (volatile void *ptr, __int8 addend );
- __int8 __TBB_EXPORTED_FUNC __TBB_machine_fetchstore1 (volatile void *ptr, __int8 value );
- __int16 __TBB_EXPORTED_FUNC __TBB_machine_cmpswp2 (volatile void *ptr, __int16 value, __int16 comparand );
- __int16 __TBB_EXPORTED_FUNC __TBB_machine_fetchadd2 (volatile void *ptr, __int16 addend );
- __int16 __TBB_EXPORTED_FUNC __TBB_machine_fetchstore2 (volatile void *ptr, __int16 value );
- }
- inline long __TBB_machine_cmpswp4 (volatile void *ptr, __int32 value, __int32 comparand ) {
- return _InterlockedCompareExchange( (long*)ptr, value, comparand );
- }
- inline long __TBB_machine_fetchadd4 (volatile void *ptr, __int32 addend ) {
- return _InterlockedExchangeAdd( (long*)ptr, addend );
- }
- inline long __TBB_machine_fetchstore4 (volatile void *ptr, __int32 value ) {
- return _InterlockedExchange( (long*)ptr, value );
- }
- inline __int64 __TBB_machine_cmpswp8 (volatile void *ptr, __int64 value, __int64 comparand ) {
- return _InterlockedCompareExchange64( (__int64*)ptr, value, comparand );
- }
- inline __int64 __TBB_machine_fetchadd8 (volatile void *ptr, __int64 addend ) {
- return _InterlockedExchangeAdd64( (__int64*)ptr, addend );
- }
- inline __int64 __TBB_machine_fetchstore8 (volatile void *ptr, __int64 value ) {
- return _InterlockedExchange64( (__int64*)ptr, value );
- }
- #define __TBB_USE_FETCHSTORE_AS_FULL_FENCED_STORE 1
- #define __TBB_USE_GENERIC_HALF_FENCED_LOAD_STORE 1
- #define __TBB_USE_GENERIC_RELAXED_LOAD_STORE 1
- #define __TBB_USE_GENERIC_SEQUENTIAL_CONSISTENCY_LOAD_STORE 1
- inline void __TBB_machine_OR( volatile void *operand, intptr_t addend ) {
- _InterlockedOr64((__int64*)operand, addend);
- }
- inline void __TBB_machine_AND( volatile void *operand, intptr_t addend ) {
- _InterlockedAnd64((__int64*)operand, addend);
- }
- #define __TBB_AtomicOR(P,V) __TBB_machine_OR(P,V)
- #define __TBB_AtomicAND(P,V) __TBB_machine_AND(P,V)
|