mirror of
https://github.com/autc04/Retro68.git
synced 2024-11-19 18:46:30 +00:00
371 lines
12 KiB
C++
371 lines
12 KiB
C++
// -*- C++ -*-
|
|
|
|
// Copyright (C) 2007, 2008, 2009, 2010, 2012 Free Software Foundation, Inc.
|
|
//
|
|
// This file is part of the GNU ISO C++ Library. This library is free
|
|
// software; you can redistribute it and/or modify it under the terms
|
|
// of the GNU General Public License as published by the Free Software
|
|
// Foundation; either version 3, or (at your option) any later
|
|
// version.
|
|
|
|
// This library is distributed in the hope that it will be useful, but
|
|
// WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
// General Public License for more details.
|
|
|
|
// Under Section 7 of GPL version 3, you are granted additional
|
|
// permissions described in the GCC Runtime Library Exception, version
|
|
// 3.1, as published by the Free Software Foundation.
|
|
|
|
// You should have received a copy of the GNU General Public License and
|
|
// a copy of the GCC Runtime Library Exception along with this program;
|
|
// see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
|
|
// <http://www.gnu.org/licenses/>.
|
|
|
|
/** @file parallel/compatibility.h
|
|
* @brief Compatibility layer, mostly concerned with atomic operations.
|
|
* This file is a GNU parallel extension to the Standard C++ Library.
|
|
*/
|
|
|
|
// Written by Felix Putze.
|
|
|
|
#ifndef _GLIBCXX_PARALLEL_COMPATIBILITY_H
|
|
#define _GLIBCXX_PARALLEL_COMPATIBILITY_H 1
|
|
|
|
#include <parallel/types.h>
|
|
#include <parallel/base.h>
|
|
|
|
#if defined(__SUNPRO_CC) && defined(__sparc)
|
|
#include <sys/atomic.h>
|
|
#endif
|
|
|
|
#if !defined(_WIN32) || defined (__CYGWIN__)
|
|
#include <sched.h>
|
|
#endif
|
|
|
|
#if defined(_MSC_VER)
|
|
#include <Windows.h>
|
|
#include <intrin.h>
|
|
#undef max
|
|
#undef min
|
|
#endif
|
|
|
|
#ifdef __MINGW32__
|
|
// Including <windows.h> will drag in all the windows32 names. Since
|
|
// that can cause user code portability problems, we just declare the
|
|
// one needed function here.
|
|
extern "C"
|
|
__attribute((dllimport)) void __attribute__((stdcall)) Sleep (unsigned long);
|
|
#endif
|
|
|
|
namespace __gnu_parallel
|
|
{
|
|
#if defined(__ICC)
|
|
template<typename _MustBeInt = int>
|
|
int32_t __faa32(int32_t* __x, int32_t __inc)
|
|
{
|
|
asm volatile("lock xadd %0,%1"
|
|
: "=__r" (__inc), "=__m" (*__x)
|
|
: "0" (__inc)
|
|
: "memory");
|
|
return __inc;
|
|
}
|
|
#if defined(__x86_64)
|
|
template<typename _MustBeInt = int>
|
|
int64_t __faa64(int64_t* __x, int64_t __inc)
|
|
{
|
|
asm volatile("lock xadd %0,%1"
|
|
: "=__r" (__inc), "=__m" (*__x)
|
|
: "0" (__inc)
|
|
: "memory");
|
|
return __inc;
|
|
}
|
|
#endif
|
|
#endif
|
|
|
|
// atomic functions only work on integers
|
|
|
|
/** @brief Add a value to a variable, atomically.
|
|
*
|
|
* Implementation is heavily platform-dependent.
|
|
* @param __ptr Pointer to a 32-bit signed integer.
|
|
* @param __addend Value to add.
|
|
*/
|
|
inline int32_t
|
|
__fetch_and_add_32(volatile int32_t* __ptr, int32_t __addend)
|
|
{
|
|
#if defined(__ICC) //x86 version
|
|
return _InterlockedExchangeAdd((void*)__ptr, __addend);
|
|
#elif defined(__ECC) //IA-64 version
|
|
return _InterlockedExchangeAdd((void*)__ptr, __addend);
|
|
#elif defined(__ICL) || defined(_MSC_VER)
|
|
return _InterlockedExchangeAdd(reinterpret_cast<volatile long*>(__ptr),
|
|
__addend);
|
|
#elif defined(__GNUC__)
|
|
return __atomic_fetch_add(__ptr, __addend, __ATOMIC_ACQ_REL);
|
|
#elif defined(__SUNPRO_CC) && defined(__sparc)
|
|
volatile int32_t __before, __after;
|
|
do
|
|
{
|
|
__before = *__ptr;
|
|
__after = __before + __addend;
|
|
} while (atomic_cas_32((volatile unsigned int*)__ptr, __before,
|
|
__after) != __before);
|
|
return __before;
|
|
#else //fallback, slow
|
|
#pragma message("slow __fetch_and_add_32")
|
|
int32_t __res;
|
|
#pragma omp critical
|
|
{
|
|
__res = *__ptr;
|
|
*(__ptr) += __addend;
|
|
}
|
|
return __res;
|
|
#endif
|
|
}
|
|
|
|
/** @brief Add a value to a variable, atomically.
|
|
*
|
|
* Implementation is heavily platform-dependent.
|
|
* @param __ptr Pointer to a 64-bit signed integer.
|
|
* @param __addend Value to add.
|
|
*/
|
|
inline int64_t
|
|
__fetch_and_add_64(volatile int64_t* __ptr, int64_t __addend)
|
|
{
|
|
#if defined(__ICC) && defined(__x86_64) //x86 version
|
|
return __faa64<int>((int64_t*)__ptr, __addend);
|
|
#elif defined(__ECC) //IA-64 version
|
|
return _InterlockedExchangeAdd64((void*)__ptr, __addend);
|
|
#elif defined(__ICL) || defined(_MSC_VER)
|
|
#ifndef _WIN64
|
|
_GLIBCXX_PARALLEL_ASSERT(false); //not available in this case
|
|
return 0;
|
|
#else
|
|
return _InterlockedExchangeAdd64(__ptr, __addend);
|
|
#endif
|
|
#elif defined(__GNUC__) && defined(__x86_64)
|
|
return __atomic_fetch_add(__ptr, __addend, __ATOMIC_ACQ_REL);
|
|
#elif defined(__GNUC__) && defined(__i386) && \
|
|
(defined(__i686) || defined(__pentium4) || defined(__athlon) \
|
|
|| defined(__k8) || defined(__core2))
|
|
return __atomic_fetch_add(__ptr, __addend, __ATOMIC_ACQ_REL);
|
|
#elif defined(__SUNPRO_CC) && defined(__sparc)
|
|
volatile int64_t __before, __after;
|
|
do
|
|
{
|
|
__before = *__ptr;
|
|
__after = __before + __addend;
|
|
} while (atomic_cas_64((volatile unsigned long long*)__ptr, __before,
|
|
__after) != __before);
|
|
return __before;
|
|
#else //fallback, slow
|
|
#if defined(__GNUC__) && defined(__i386)
|
|
// XXX doesn'__t work with -march=native
|
|
//#warning "please compile with -march=i686 or better"
|
|
#endif
|
|
#pragma message("slow __fetch_and_add_64")
|
|
int64_t __res;
|
|
#pragma omp critical
|
|
{
|
|
__res = *__ptr;
|
|
*(__ptr) += __addend;
|
|
}
|
|
return __res;
|
|
#endif
|
|
}
|
|
|
|
/** @brief Add a value to a variable, atomically.
|
|
*
|
|
* Implementation is heavily platform-dependent.
|
|
* @param __ptr Pointer to a signed integer.
|
|
* @param __addend Value to add.
|
|
*/
|
|
template<typename _Tp>
|
|
inline _Tp
|
|
__fetch_and_add(volatile _Tp* __ptr, _Tp __addend)
|
|
{
|
|
if (sizeof(_Tp) == sizeof(int32_t))
|
|
return
|
|
(_Tp)__fetch_and_add_32((volatile int32_t*) __ptr, (int32_t)__addend);
|
|
else if (sizeof(_Tp) == sizeof(int64_t))
|
|
return
|
|
(_Tp)__fetch_and_add_64((volatile int64_t*) __ptr, (int64_t)__addend);
|
|
else
|
|
_GLIBCXX_PARALLEL_ASSERT(false);
|
|
}
|
|
|
|
|
|
#if defined(__ICC)
|
|
|
|
template<typename _MustBeInt = int>
|
|
inline int32_t
|
|
__cas32(volatile int32_t* __ptr, int32_t __old, int32_t __nw)
|
|
{
|
|
int32_t __before;
|
|
__asm__ __volatile__("lock; cmpxchgl %1,%2"
|
|
: "=a"(__before)
|
|
: "q"(__nw), "__m"(*(volatile long long*)(__ptr)),
|
|
"0"(__old)
|
|
: "memory");
|
|
return __before;
|
|
}
|
|
|
|
#if defined(__x86_64)
|
|
template<typename _MustBeInt = int>
|
|
inline int64_t
|
|
__cas64(volatile int64_t *__ptr, int64_t __old, int64_t __nw)
|
|
{
|
|
int64_t __before;
|
|
__asm__ __volatile__("lock; cmpxchgq %1,%2"
|
|
: "=a"(__before)
|
|
: "q"(__nw), "__m"(*(volatile long long*)(__ptr)),
|
|
"0"(__old)
|
|
: "memory");
|
|
return __before;
|
|
}
|
|
#endif
|
|
|
|
#endif
|
|
|
|
/** @brief Compare @c *__ptr and @c __comparand. If equal, let @c
|
|
* *__ptr=__replacement and return @c true, return @c false otherwise.
|
|
*
|
|
* Implementation is heavily platform-dependent.
|
|
* @param __ptr Pointer to 32-bit signed integer.
|
|
* @param __comparand Compare value.
|
|
* @param __replacement Replacement value.
|
|
*/
|
|
inline bool
|
|
__compare_and_swap_32(volatile int32_t* __ptr, int32_t __comparand,
|
|
int32_t __replacement)
|
|
{
|
|
#if defined(__ICC) //x86 version
|
|
return _InterlockedCompareExchange((void*)__ptr, __replacement,
|
|
__comparand) == __comparand;
|
|
#elif defined(__ECC) //IA-64 version
|
|
return _InterlockedCompareExchange((void*)__ptr, __replacement,
|
|
__comparand) == __comparand;
|
|
#elif defined(__ICL) || defined(_MSC_VER)
|
|
return _InterlockedCompareExchange(
|
|
reinterpret_cast<volatile long*>(__ptr),
|
|
__replacement, __comparand)
|
|
== __comparand;
|
|
#elif defined(__GNUC__)
|
|
return __atomic_compare_exchange_n(__ptr, &__comparand, __replacement,
|
|
false, __ATOMIC_ACQ_REL,
|
|
__ATOMIC_RELAXED);
|
|
#elif defined(__SUNPRO_CC) && defined(__sparc)
|
|
return atomic_cas_32((volatile unsigned int*)__ptr, __comparand,
|
|
__replacement) == __comparand;
|
|
#else
|
|
#pragma message("slow __compare_and_swap_32")
|
|
bool __res = false;
|
|
#pragma omp critical
|
|
{
|
|
if (*__ptr == __comparand)
|
|
{
|
|
*__ptr = __replacement;
|
|
__res = true;
|
|
}
|
|
}
|
|
return __res;
|
|
#endif
|
|
}
|
|
|
|
/** @brief Compare @c *__ptr and @c __comparand. If equal, let @c
|
|
* *__ptr=__replacement and return @c true, return @c false otherwise.
|
|
*
|
|
* Implementation is heavily platform-dependent.
|
|
* @param __ptr Pointer to 64-bit signed integer.
|
|
* @param __comparand Compare value.
|
|
* @param __replacement Replacement value.
|
|
*/
|
|
inline bool
|
|
__compare_and_swap_64(volatile int64_t* __ptr, int64_t __comparand,
|
|
int64_t __replacement)
|
|
{
|
|
#if defined(__ICC) && defined(__x86_64) //x86 version
|
|
return __cas64<int>(__ptr, __comparand, __replacement) == __comparand;
|
|
#elif defined(__ECC) //IA-64 version
|
|
return _InterlockedCompareExchange64((void*)__ptr, __replacement,
|
|
__comparand) == __comparand;
|
|
#elif defined(__ICL) || defined(_MSC_VER)
|
|
#ifndef _WIN64
|
|
_GLIBCXX_PARALLEL_ASSERT(false); //not available in this case
|
|
return 0;
|
|
#else
|
|
return _InterlockedCompareExchange64(__ptr, __replacement,
|
|
__comparand) == __comparand;
|
|
#endif
|
|
|
|
#elif defined(__GNUC__) && defined(__x86_64)
|
|
return __atomic_compare_exchange_n(__ptr, &__comparand, __replacement,
|
|
false, __ATOMIC_ACQ_REL,
|
|
__ATOMIC_RELAXED);
|
|
#elif defined(__GNUC__) && defined(__i386) && \
|
|
(defined(__i686) || defined(__pentium4) || defined(__athlon) \
|
|
|| defined(__k8) || defined(__core2))
|
|
return __atomic_compare_exchange_n(__ptr, &__comparand, __replacement,
|
|
false, __ATOMIC_ACQ_REL,
|
|
__ATOMIC_RELAXED);
|
|
#elif defined(__SUNPRO_CC) && defined(__sparc)
|
|
return atomic_cas_64((volatile unsigned long long*)__ptr,
|
|
__comparand, __replacement) == __comparand;
|
|
#else
|
|
#if defined(__GNUC__) && defined(__i386)
|
|
// XXX -march=native
|
|
//#warning "please compile with -march=i686 or better"
|
|
#endif
|
|
#pragma message("slow __compare_and_swap_64")
|
|
bool __res = false;
|
|
#pragma omp critical
|
|
{
|
|
if (*__ptr == __comparand)
|
|
{
|
|
*__ptr = __replacement;
|
|
__res = true;
|
|
}
|
|
}
|
|
return __res;
|
|
#endif
|
|
}
|
|
|
|
/** @brief Compare @c *__ptr and @c __comparand. If equal, let @c
|
|
* *__ptr=__replacement and return @c true, return @c false otherwise.
|
|
*
|
|
* Implementation is heavily platform-dependent.
|
|
* @param __ptr Pointer to signed integer.
|
|
* @param __comparand Compare value.
|
|
* @param __replacement Replacement value. */
|
|
template<typename _Tp>
|
|
inline bool
|
|
__compare_and_swap(volatile _Tp* __ptr, _Tp __comparand, _Tp __replacement)
|
|
{
|
|
if (sizeof(_Tp) == sizeof(int32_t))
|
|
return __compare_and_swap_32((volatile int32_t*) __ptr,
|
|
(int32_t)__comparand,
|
|
(int32_t)__replacement);
|
|
else if (sizeof(_Tp) == sizeof(int64_t))
|
|
return __compare_and_swap_64((volatile int64_t*) __ptr,
|
|
(int64_t)__comparand,
|
|
(int64_t)__replacement);
|
|
else
|
|
_GLIBCXX_PARALLEL_ASSERT(false);
|
|
}
|
|
|
|
/** @brief Yield the control to another thread, without waiting for
|
|
the end to the time slice. */
|
|
inline void
|
|
__yield()
|
|
{
|
|
#if defined (_WIN32) && !defined (__CYGWIN__)
|
|
Sleep(0);
|
|
#else
|
|
sched_yield();
|
|
#endif
|
|
}
|
|
} // end namespace
|
|
|
|
#endif /* _GLIBCXX_PARALLEL_COMPATIBILITY_H */
|