re PR libstdc++/54754 ([parallel mode] 'make check-parallel' only works on x86-64)
PR libstdc++/54754 * include/parallel/compatibility.h: Use atomic built-ins when they are lock-free. From-SVN: r192240
This commit is contained in:
parent
eef26c05fb
commit
8024199176
2 changed files with 50 additions and 128 deletions
|
@ -1,3 +1,9 @@
|
||||||
|
2012-10-09 Jonathan Wakely <jwakely.gcc@gmail.com>
|
||||||
|
|
||||||
|
PR libstdc++/54754
|
||||||
|
* include/parallel/compatibility.h: Use atomic built-ins when they are
|
||||||
|
lock-free.
|
||||||
|
|
||||||
2012-10-09 Uros Bizjak <ubizjak@gmail.com>
|
2012-10-09 Uros Bizjak <ubizjak@gmail.com>
|
||||||
|
|
||||||
* testsuite/util/testsuite_abi.cc (check_version): Add CXXABI_1.3.7.
|
* testsuite/util/testsuite_abi.cc (check_version): Add CXXABI_1.3.7.
|
||||||
|
|
|
@ -51,154 +51,70 @@ __attribute((dllimport)) void __attribute__((stdcall)) Sleep (unsigned long);
|
||||||
|
|
||||||
namespace __gnu_parallel
|
namespace __gnu_parallel
|
||||||
{
|
{
|
||||||
// These atomic functions only work on integers
|
template<typename _Tp>
|
||||||
|
inline _Tp
|
||||||
/** @brief Add a value to a variable, atomically.
|
__add_omp(volatile _Tp* __ptr, _Tp __addend)
|
||||||
*
|
|
||||||
* Implementation is heavily platform-dependent.
|
|
||||||
* @param __ptr Pointer to a 32-bit signed integer.
|
|
||||||
* @param __addend Value to add.
|
|
||||||
*/
|
|
||||||
inline int32_t
|
|
||||||
__fetch_and_add_32(volatile int32_t* __ptr, int32_t __addend)
|
|
||||||
{
|
|
||||||
return __atomic_fetch_add(__ptr, __addend, __ATOMIC_ACQ_REL);
|
|
||||||
}
|
|
||||||
|
|
||||||
/** @brief Add a value to a variable, atomically.
|
|
||||||
*
|
|
||||||
* Implementation is heavily platform-dependent.
|
|
||||||
* @param __ptr Pointer to a 64-bit signed integer.
|
|
||||||
* @param __addend Value to add.
|
|
||||||
*/
|
|
||||||
inline int64_t
|
|
||||||
__fetch_and_add_64(volatile int64_t* __ptr, int64_t __addend)
|
|
||||||
{
|
|
||||||
#if defined(__x86_64)
|
|
||||||
return __atomic_fetch_add(__ptr, __addend, __ATOMIC_ACQ_REL);
|
|
||||||
#elif defined(__i386) && \
|
|
||||||
(defined(__i686) || defined(__pentium4) || defined(__athlon) \
|
|
||||||
|| defined(__k8) || defined(__core2))
|
|
||||||
return __atomic_fetch_add(__ptr, __addend, __ATOMIC_ACQ_REL);
|
|
||||||
#else //fallback, slow
|
|
||||||
#if defined(__i386)
|
|
||||||
// XXX doesn'__t work with -march=native
|
|
||||||
//#warning "please compile with -march=i686 or better"
|
|
||||||
#endif
|
|
||||||
#pragma message("slow __fetch_and_add_64")
|
|
||||||
int64_t __res;
|
|
||||||
#pragma omp critical
|
|
||||||
{
|
{
|
||||||
__res = *__ptr;
|
int64_t __res;
|
||||||
*(__ptr) += __addend;
|
#pragma omp critical
|
||||||
|
{
|
||||||
|
__res = *__ptr;
|
||||||
|
*(__ptr) += __addend;
|
||||||
|
}
|
||||||
|
return __res;
|
||||||
}
|
}
|
||||||
return __res;
|
|
||||||
#endif
|
|
||||||
}
|
|
||||||
|
|
||||||
/** @brief Add a value to a variable, atomically.
|
/** @brief Add a value to a variable, atomically.
|
||||||
*
|
*
|
||||||
* Implementation is heavily platform-dependent.
|
|
||||||
* @param __ptr Pointer to a signed integer.
|
* @param __ptr Pointer to a signed integer.
|
||||||
* @param __addend Value to add.
|
* @param __addend Value to add.
|
||||||
*/
|
*/
|
||||||
template<typename _Tp>
|
template<typename _Tp>
|
||||||
inline _Tp
|
inline _Tp
|
||||||
__fetch_and_add(volatile _Tp* __ptr, _Tp __addend)
|
__fetch_and_add(volatile _Tp* __ptr, _Tp __addend)
|
||||||
{
|
|
||||||
if (sizeof(_Tp) == sizeof(int32_t))
|
|
||||||
return
|
|
||||||
(_Tp)__fetch_and_add_32((volatile int32_t*) __ptr, (int32_t)__addend);
|
|
||||||
else if (sizeof(_Tp) == sizeof(int64_t))
|
|
||||||
return
|
|
||||||
(_Tp)__fetch_and_add_64((volatile int64_t*) __ptr, (int64_t)__addend);
|
|
||||||
else
|
|
||||||
_GLIBCXX_PARALLEL_ASSERT(false);
|
|
||||||
}
|
|
||||||
|
|
||||||
/** @brief Compare @c *__ptr and @c __comparand. If equal, let @c
|
|
||||||
* *__ptr=__replacement and return @c true, return @c false otherwise.
|
|
||||||
*
|
|
||||||
* Implementation is heavily platform-dependent.
|
|
||||||
* @param __ptr Pointer to 32-bit signed integer.
|
|
||||||
* @param __comparand Compare value.
|
|
||||||
* @param __replacement Replacement value.
|
|
||||||
*/
|
|
||||||
inline bool
|
|
||||||
__compare_and_swap_32(volatile int32_t* __ptr, int32_t __comparand,
|
|
||||||
int32_t __replacement)
|
|
||||||
{
|
|
||||||
return __atomic_compare_exchange_n(__ptr, &__comparand, __replacement,
|
|
||||||
false, __ATOMIC_ACQ_REL,
|
|
||||||
__ATOMIC_RELAXED);
|
|
||||||
}
|
|
||||||
|
|
||||||
/** @brief Compare @c *__ptr and @c __comparand. If equal, let @c
|
|
||||||
* *__ptr=__replacement and return @c true, return @c false otherwise.
|
|
||||||
*
|
|
||||||
* Implementation is heavily platform-dependent.
|
|
||||||
* @param __ptr Pointer to 64-bit signed integer.
|
|
||||||
* @param __comparand Compare value.
|
|
||||||
* @param __replacement Replacement value.
|
|
||||||
*/
|
|
||||||
inline bool
|
|
||||||
__compare_and_swap_64(volatile int64_t* __ptr, int64_t __comparand,
|
|
||||||
int64_t __replacement)
|
|
||||||
{
|
|
||||||
#if defined(__x86_64)
|
|
||||||
return __atomic_compare_exchange_n(__ptr, &__comparand, __replacement,
|
|
||||||
false, __ATOMIC_ACQ_REL,
|
|
||||||
__ATOMIC_RELAXED);
|
|
||||||
#elif defined(__i386) && \
|
|
||||||
(defined(__i686) || defined(__pentium4) || defined(__athlon) \
|
|
||||||
|| defined(__k8) || defined(__core2))
|
|
||||||
return __atomic_compare_exchange_n(__ptr, &__comparand, __replacement,
|
|
||||||
false, __ATOMIC_ACQ_REL,
|
|
||||||
__ATOMIC_RELAXED);
|
|
||||||
#else
|
|
||||||
#if defined(__i386)
|
|
||||||
// XXX -march=native
|
|
||||||
//#warning "please compile with -march=i686 or better"
|
|
||||||
#endif
|
|
||||||
#pragma message("slow __compare_and_swap_64")
|
|
||||||
bool __res = false;
|
|
||||||
#pragma omp critical
|
|
||||||
{
|
{
|
||||||
if (*__ptr == __comparand)
|
if (__atomic_always_lock_free(sizeof(_Tp), __ptr))
|
||||||
{
|
return __atomic_fetch_add(__ptr, __addend, __ATOMIC_ACQ_REL);
|
||||||
*__ptr = __replacement;
|
return __add_omp(__ptr, __addend);
|
||||||
__res = true;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return __res;
|
|
||||||
#endif
|
|
||||||
}
|
|
||||||
|
|
||||||
/** @brief Compare @c *__ptr and @c __comparand. If equal, let @c
|
template<typename _Tp>
|
||||||
|
inline bool
|
||||||
|
__cas_omp(volatile _Tp* __ptr, _Tp __comparand, _Tp __replacement)
|
||||||
|
{
|
||||||
|
bool __res = false;
|
||||||
|
#pragma omp critical
|
||||||
|
{
|
||||||
|
if (*__ptr == __comparand)
|
||||||
|
{
|
||||||
|
*__ptr = __replacement;
|
||||||
|
__res = true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return __res;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** @brief Compare-and-swap
|
||||||
|
*
|
||||||
|
* Compare @c *__ptr and @c __comparand. If equal, let @c
|
||||||
* *__ptr=__replacement and return @c true, return @c false otherwise.
|
* *__ptr=__replacement and return @c true, return @c false otherwise.
|
||||||
*
|
*
|
||||||
* Implementation is heavily platform-dependent.
|
|
||||||
* @param __ptr Pointer to signed integer.
|
* @param __ptr Pointer to signed integer.
|
||||||
* @param __comparand Compare value.
|
* @param __comparand Compare value.
|
||||||
* @param __replacement Replacement value.
|
* @param __replacement Replacement value.
|
||||||
*/
|
*/
|
||||||
template<typename _Tp>
|
template<typename _Tp>
|
||||||
inline bool
|
inline bool
|
||||||
__compare_and_swap(volatile _Tp* __ptr, _Tp __comparand, _Tp __replacement)
|
__compare_and_swap(volatile _Tp* __ptr, _Tp __comparand, _Tp __replacement)
|
||||||
{
|
{
|
||||||
if (sizeof(_Tp) == sizeof(int32_t))
|
if (__atomic_always_lock_free(sizeof(_Tp), __ptr))
|
||||||
return __compare_and_swap_32((volatile int32_t*) __ptr,
|
return __atomic_compare_exchange_n(__ptr, &__comparand, __replacement,
|
||||||
(int32_t)__comparand,
|
false, __ATOMIC_ACQ_REL,
|
||||||
(int32_t)__replacement);
|
__ATOMIC_RELAXED);
|
||||||
else if (sizeof(_Tp) == sizeof(int64_t))
|
return __cas_omp(__ptr, __comparand, __replacement);
|
||||||
return __compare_and_swap_64((volatile int64_t*) __ptr,
|
}
|
||||||
(int64_t)__comparand,
|
|
||||||
(int64_t)__replacement);
|
|
||||||
else
|
|
||||||
_GLIBCXX_PARALLEL_ASSERT(false);
|
|
||||||
}
|
|
||||||
|
|
||||||
/** @brief Yield the control to another thread, without waiting for
|
/** @brief Yield control to another thread, without waiting for
|
||||||
* the end of the time slice.
|
* the end of the time slice.
|
||||||
*/
|
*/
|
||||||
inline void
|
inline void
|
||||||
|
|
Loading…
Add table
Reference in a new issue