forked from OSchip/llvm-project
1559 lines
57 KiB
C
1559 lines
57 KiB
C
/*===---- __clang_cuda_device_functions.h - CUDA runtime support -----------===
|
|
*
|
|
* Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
* See https://llvm.org/LICENSE.txt for license information.
|
|
* SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
*
|
|
*===-----------------------------------------------------------------------===
|
|
*/
|
|
|
|
#ifndef __CLANG_CUDA_DEVICE_FUNCTIONS_H__
|
|
#define __CLANG_CUDA_DEVICE_FUNCTIONS_H__
|
|
|
|
#ifndef __OPENMP_NVPTX__
|
|
#if CUDA_VERSION < 9000
|
|
#error This file is intended to be used with CUDA-9+ only.
|
|
#endif
|
|
#endif
|
|
|
|
// __DEVICE__ is a helper macro with common set of attributes for the wrappers
|
|
// we implement in this file. We need static in order to avoid emitting unused
|
|
// functions and __forceinline__ helps inlining these wrappers at -O1.
|
|
#pragma push_macro("__DEVICE__")
|
|
#ifdef __OPENMP_NVPTX__
|
|
#define __DEVICE__ static __attribute__((always_inline, nothrow))
|
|
#else
|
|
#define __DEVICE__ static __device__ __forceinline__
|
|
#endif
|
|
|
|
__DEVICE__ int __all(int __a) { return __nvvm_vote_all(__a); }
|
|
__DEVICE__ int __any(int __a) { return __nvvm_vote_any(__a); }
|
|
__DEVICE__ unsigned int __ballot(int __a) { return __nvvm_vote_ballot(__a); }
|
|
__DEVICE__ unsigned int __brev(unsigned int __a) { return __nv_brev(__a); }
|
|
__DEVICE__ unsigned long long __brevll(unsigned long long __a) {
|
|
return __nv_brevll(__a);
|
|
}
|
|
#if defined(__cplusplus)
|
|
__DEVICE__ void __brkpt() { __asm__ __volatile__("brkpt;"); }
|
|
__DEVICE__ void __brkpt(int __a) { __brkpt(); }
|
|
#else
|
|
__DEVICE__ void __attribute__((overloadable)) __brkpt(void) {
|
|
__asm__ __volatile__("brkpt;");
|
|
}
|
|
__DEVICE__ void __attribute__((overloadable)) __brkpt(int __a) { __brkpt(); }
|
|
#endif
|
|
__DEVICE__ unsigned int __byte_perm(unsigned int __a, unsigned int __b,
|
|
unsigned int __c) {
|
|
return __nv_byte_perm(__a, __b, __c);
|
|
}
|
|
__DEVICE__ int __clz(int __a) { return __nv_clz(__a); }
|
|
__DEVICE__ int __clzll(long long __a) { return __nv_clzll(__a); }
|
|
__DEVICE__ float __cosf(float __a) { return __nv_fast_cosf(__a); }
|
|
__DEVICE__ double __dAtomicAdd(double *__p, double __v) {
|
|
return __nvvm_atom_add_gen_d(__p, __v);
|
|
}
|
|
__DEVICE__ double __dAtomicAdd_block(double *__p, double __v) {
|
|
return __nvvm_atom_cta_add_gen_d(__p, __v);
|
|
}
|
|
__DEVICE__ double __dAtomicAdd_system(double *__p, double __v) {
|
|
return __nvvm_atom_sys_add_gen_d(__p, __v);
|
|
}
|
|
__DEVICE__ double __dadd_rd(double __a, double __b) {
|
|
return __nv_dadd_rd(__a, __b);
|
|
}
|
|
__DEVICE__ double __dadd_rn(double __a, double __b) {
|
|
return __nv_dadd_rn(__a, __b);
|
|
}
|
|
__DEVICE__ double __dadd_ru(double __a, double __b) {
|
|
return __nv_dadd_ru(__a, __b);
|
|
}
|
|
__DEVICE__ double __dadd_rz(double __a, double __b) {
|
|
return __nv_dadd_rz(__a, __b);
|
|
}
|
|
__DEVICE__ double __ddiv_rd(double __a, double __b) {
|
|
return __nv_ddiv_rd(__a, __b);
|
|
}
|
|
__DEVICE__ double __ddiv_rn(double __a, double __b) {
|
|
return __nv_ddiv_rn(__a, __b);
|
|
}
|
|
__DEVICE__ double __ddiv_ru(double __a, double __b) {
|
|
return __nv_ddiv_ru(__a, __b);
|
|
}
|
|
__DEVICE__ double __ddiv_rz(double __a, double __b) {
|
|
return __nv_ddiv_rz(__a, __b);
|
|
}
|
|
__DEVICE__ double __dmul_rd(double __a, double __b) {
|
|
return __nv_dmul_rd(__a, __b);
|
|
}
|
|
__DEVICE__ double __dmul_rn(double __a, double __b) {
|
|
return __nv_dmul_rn(__a, __b);
|
|
}
|
|
__DEVICE__ double __dmul_ru(double __a, double __b) {
|
|
return __nv_dmul_ru(__a, __b);
|
|
}
|
|
__DEVICE__ double __dmul_rz(double __a, double __b) {
|
|
return __nv_dmul_rz(__a, __b);
|
|
}
|
|
__DEVICE__ float __double2float_rd(double __a) {
|
|
return __nv_double2float_rd(__a);
|
|
}
|
|
__DEVICE__ float __double2float_rn(double __a) {
|
|
return __nv_double2float_rn(__a);
|
|
}
|
|
__DEVICE__ float __double2float_ru(double __a) {
|
|
return __nv_double2float_ru(__a);
|
|
}
|
|
__DEVICE__ float __double2float_rz(double __a) {
|
|
return __nv_double2float_rz(__a);
|
|
}
|
|
__DEVICE__ int __double2hiint(double __a) { return __nv_double2hiint(__a); }
|
|
__DEVICE__ int __double2int_rd(double __a) { return __nv_double2int_rd(__a); }
|
|
__DEVICE__ int __double2int_rn(double __a) { return __nv_double2int_rn(__a); }
|
|
__DEVICE__ int __double2int_ru(double __a) { return __nv_double2int_ru(__a); }
|
|
__DEVICE__ int __double2int_rz(double __a) { return __nv_double2int_rz(__a); }
|
|
__DEVICE__ long long __double2ll_rd(double __a) {
|
|
return __nv_double2ll_rd(__a);
|
|
}
|
|
__DEVICE__ long long __double2ll_rn(double __a) {
|
|
return __nv_double2ll_rn(__a);
|
|
}
|
|
__DEVICE__ long long __double2ll_ru(double __a) {
|
|
return __nv_double2ll_ru(__a);
|
|
}
|
|
__DEVICE__ long long __double2ll_rz(double __a) {
|
|
return __nv_double2ll_rz(__a);
|
|
}
|
|
__DEVICE__ int __double2loint(double __a) { return __nv_double2loint(__a); }
|
|
__DEVICE__ unsigned int __double2uint_rd(double __a) {
|
|
return __nv_double2uint_rd(__a);
|
|
}
|
|
__DEVICE__ unsigned int __double2uint_rn(double __a) {
|
|
return __nv_double2uint_rn(__a);
|
|
}
|
|
__DEVICE__ unsigned int __double2uint_ru(double __a) {
|
|
return __nv_double2uint_ru(__a);
|
|
}
|
|
__DEVICE__ unsigned int __double2uint_rz(double __a) {
|
|
return __nv_double2uint_rz(__a);
|
|
}
|
|
__DEVICE__ unsigned long long __double2ull_rd(double __a) {
|
|
return __nv_double2ull_rd(__a);
|
|
}
|
|
__DEVICE__ unsigned long long __double2ull_rn(double __a) {
|
|
return __nv_double2ull_rn(__a);
|
|
}
|
|
__DEVICE__ unsigned long long __double2ull_ru(double __a) {
|
|
return __nv_double2ull_ru(__a);
|
|
}
|
|
__DEVICE__ unsigned long long __double2ull_rz(double __a) {
|
|
return __nv_double2ull_rz(__a);
|
|
}
|
|
__DEVICE__ long long __double_as_longlong(double __a) {
|
|
return __nv_double_as_longlong(__a);
|
|
}
|
|
__DEVICE__ double __drcp_rd(double __a) { return __nv_drcp_rd(__a); }
|
|
__DEVICE__ double __drcp_rn(double __a) { return __nv_drcp_rn(__a); }
|
|
__DEVICE__ double __drcp_ru(double __a) { return __nv_drcp_ru(__a); }
|
|
__DEVICE__ double __drcp_rz(double __a) { return __nv_drcp_rz(__a); }
|
|
__DEVICE__ double __dsqrt_rd(double __a) { return __nv_dsqrt_rd(__a); }
|
|
__DEVICE__ double __dsqrt_rn(double __a) { return __nv_dsqrt_rn(__a); }
|
|
__DEVICE__ double __dsqrt_ru(double __a) { return __nv_dsqrt_ru(__a); }
|
|
__DEVICE__ double __dsqrt_rz(double __a) { return __nv_dsqrt_rz(__a); }
|
|
__DEVICE__ double __dsub_rd(double __a, double __b) {
|
|
return __nv_dsub_rd(__a, __b);
|
|
}
|
|
__DEVICE__ double __dsub_rn(double __a, double __b) {
|
|
return __nv_dsub_rn(__a, __b);
|
|
}
|
|
__DEVICE__ double __dsub_ru(double __a, double __b) {
|
|
return __nv_dsub_ru(__a, __b);
|
|
}
|
|
__DEVICE__ double __dsub_rz(double __a, double __b) {
|
|
return __nv_dsub_rz(__a, __b);
|
|
}
|
|
__DEVICE__ float __exp10f(float __a) { return __nv_fast_exp10f(__a); }
|
|
__DEVICE__ float __expf(float __a) { return __nv_fast_expf(__a); }
|
|
__DEVICE__ float __fAtomicAdd(float *__p, float __v) {
|
|
return __nvvm_atom_add_gen_f(__p, __v);
|
|
}
|
|
__DEVICE__ float __fAtomicAdd_block(float *__p, float __v) {
|
|
return __nvvm_atom_cta_add_gen_f(__p, __v);
|
|
}
|
|
__DEVICE__ float __fAtomicAdd_system(float *__p, float __v) {
|
|
return __nvvm_atom_sys_add_gen_f(__p, __v);
|
|
}
|
|
__DEVICE__ float __fAtomicExch(float *__p, float __v) {
|
|
return __nv_int_as_float(
|
|
__nvvm_atom_xchg_gen_i((int *)__p, __nv_float_as_int(__v)));
|
|
}
|
|
__DEVICE__ float __fAtomicExch_block(float *__p, float __v) {
|
|
return __nv_int_as_float(
|
|
__nvvm_atom_cta_xchg_gen_i((int *)__p, __nv_float_as_int(__v)));
|
|
}
|
|
__DEVICE__ float __fAtomicExch_system(float *__p, float __v) {
|
|
return __nv_int_as_float(
|
|
__nvvm_atom_sys_xchg_gen_i((int *)__p, __nv_float_as_int(__v)));
|
|
}
|
|
__DEVICE__ float __fadd_rd(float __a, float __b) {
|
|
return __nv_fadd_rd(__a, __b);
|
|
}
|
|
__DEVICE__ float __fadd_rn(float __a, float __b) {
|
|
return __nv_fadd_rn(__a, __b);
|
|
}
|
|
__DEVICE__ float __fadd_ru(float __a, float __b) {
|
|
return __nv_fadd_ru(__a, __b);
|
|
}
|
|
__DEVICE__ float __fadd_rz(float __a, float __b) {
|
|
return __nv_fadd_rz(__a, __b);
|
|
}
|
|
__DEVICE__ float __fdiv_rd(float __a, float __b) {
|
|
return __nv_fdiv_rd(__a, __b);
|
|
}
|
|
__DEVICE__ float __fdiv_rn(float __a, float __b) {
|
|
return __nv_fdiv_rn(__a, __b);
|
|
}
|
|
__DEVICE__ float __fdiv_ru(float __a, float __b) {
|
|
return __nv_fdiv_ru(__a, __b);
|
|
}
|
|
__DEVICE__ float __fdiv_rz(float __a, float __b) {
|
|
return __nv_fdiv_rz(__a, __b);
|
|
}
|
|
__DEVICE__ float __fdividef(float __a, float __b) {
|
|
return __nv_fast_fdividef(__a, __b);
|
|
}
|
|
__DEVICE__ int __ffs(int __a) { return __nv_ffs(__a); }
|
|
__DEVICE__ int __ffsll(long long __a) { return __nv_ffsll(__a); }
|
|
__DEVICE__ int __finite(double __a) { return __nv_isfinited(__a); }
|
|
__DEVICE__ int __finitef(float __a) { return __nv_finitef(__a); }
|
|
#ifdef _MSC_VER
|
|
__DEVICE__ int __finitel(long double __a);
|
|
#endif
|
|
__DEVICE__ int __float2int_rd(float __a) { return __nv_float2int_rd(__a); }
|
|
__DEVICE__ int __float2int_rn(float __a) { return __nv_float2int_rn(__a); }
|
|
__DEVICE__ int __float2int_ru(float __a) { return __nv_float2int_ru(__a); }
|
|
__DEVICE__ int __float2int_rz(float __a) { return __nv_float2int_rz(__a); }
|
|
__DEVICE__ long long __float2ll_rd(float __a) { return __nv_float2ll_rd(__a); }
|
|
__DEVICE__ long long __float2ll_rn(float __a) { return __nv_float2ll_rn(__a); }
|
|
__DEVICE__ long long __float2ll_ru(float __a) { return __nv_float2ll_ru(__a); }
|
|
__DEVICE__ long long __float2ll_rz(float __a) { return __nv_float2ll_rz(__a); }
|
|
__DEVICE__ unsigned int __float2uint_rd(float __a) {
|
|
return __nv_float2uint_rd(__a);
|
|
}
|
|
__DEVICE__ unsigned int __float2uint_rn(float __a) {
|
|
return __nv_float2uint_rn(__a);
|
|
}
|
|
__DEVICE__ unsigned int __float2uint_ru(float __a) {
|
|
return __nv_float2uint_ru(__a);
|
|
}
|
|
__DEVICE__ unsigned int __float2uint_rz(float __a) {
|
|
return __nv_float2uint_rz(__a);
|
|
}
|
|
__DEVICE__ unsigned long long __float2ull_rd(float __a) {
|
|
return __nv_float2ull_rd(__a);
|
|
}
|
|
__DEVICE__ unsigned long long __float2ull_rn(float __a) {
|
|
return __nv_float2ull_rn(__a);
|
|
}
|
|
__DEVICE__ unsigned long long __float2ull_ru(float __a) {
|
|
return __nv_float2ull_ru(__a);
|
|
}
|
|
__DEVICE__ unsigned long long __float2ull_rz(float __a) {
|
|
return __nv_float2ull_rz(__a);
|
|
}
|
|
__DEVICE__ int __float_as_int(float __a) { return __nv_float_as_int(__a); }
|
|
__DEVICE__ unsigned int __float_as_uint(float __a) {
|
|
return __nv_float_as_uint(__a);
|
|
}
|
|
__DEVICE__ double __fma_rd(double __a, double __b, double __c) {
|
|
return __nv_fma_rd(__a, __b, __c);
|
|
}
|
|
__DEVICE__ double __fma_rn(double __a, double __b, double __c) {
|
|
return __nv_fma_rn(__a, __b, __c);
|
|
}
|
|
__DEVICE__ double __fma_ru(double __a, double __b, double __c) {
|
|
return __nv_fma_ru(__a, __b, __c);
|
|
}
|
|
__DEVICE__ double __fma_rz(double __a, double __b, double __c) {
|
|
return __nv_fma_rz(__a, __b, __c);
|
|
}
|
|
__DEVICE__ float __fmaf_ieee_rd(float __a, float __b, float __c) {
|
|
return __nv_fmaf_ieee_rd(__a, __b, __c);
|
|
}
|
|
__DEVICE__ float __fmaf_ieee_rn(float __a, float __b, float __c) {
|
|
return __nv_fmaf_ieee_rn(__a, __b, __c);
|
|
}
|
|
__DEVICE__ float __fmaf_ieee_ru(float __a, float __b, float __c) {
|
|
return __nv_fmaf_ieee_ru(__a, __b, __c);
|
|
}
|
|
__DEVICE__ float __fmaf_ieee_rz(float __a, float __b, float __c) {
|
|
return __nv_fmaf_ieee_rz(__a, __b, __c);
|
|
}
|
|
__DEVICE__ float __fmaf_rd(float __a, float __b, float __c) {
|
|
return __nv_fmaf_rd(__a, __b, __c);
|
|
}
|
|
__DEVICE__ float __fmaf_rn(float __a, float __b, float __c) {
|
|
return __nv_fmaf_rn(__a, __b, __c);
|
|
}
|
|
__DEVICE__ float __fmaf_ru(float __a, float __b, float __c) {
|
|
return __nv_fmaf_ru(__a, __b, __c);
|
|
}
|
|
__DEVICE__ float __fmaf_rz(float __a, float __b, float __c) {
|
|
return __nv_fmaf_rz(__a, __b, __c);
|
|
}
|
|
__DEVICE__ float __fmul_rd(float __a, float __b) {
|
|
return __nv_fmul_rd(__a, __b);
|
|
}
|
|
__DEVICE__ float __fmul_rn(float __a, float __b) {
|
|
return __nv_fmul_rn(__a, __b);
|
|
}
|
|
__DEVICE__ float __fmul_ru(float __a, float __b) {
|
|
return __nv_fmul_ru(__a, __b);
|
|
}
|
|
__DEVICE__ float __fmul_rz(float __a, float __b) {
|
|
return __nv_fmul_rz(__a, __b);
|
|
}
|
|
__DEVICE__ float __frcp_rd(float __a) { return __nv_frcp_rd(__a); }
|
|
__DEVICE__ float __frcp_rn(float __a) { return __nv_frcp_rn(__a); }
|
|
__DEVICE__ float __frcp_ru(float __a) { return __nv_frcp_ru(__a); }
|
|
__DEVICE__ float __frcp_rz(float __a) { return __nv_frcp_rz(__a); }
|
|
__DEVICE__ float __frsqrt_rn(float __a) { return __nv_frsqrt_rn(__a); }
|
|
__DEVICE__ float __fsqrt_rd(float __a) { return __nv_fsqrt_rd(__a); }
|
|
__DEVICE__ float __fsqrt_rn(float __a) { return __nv_fsqrt_rn(__a); }
|
|
__DEVICE__ float __fsqrt_ru(float __a) { return __nv_fsqrt_ru(__a); }
|
|
__DEVICE__ float __fsqrt_rz(float __a) { return __nv_fsqrt_rz(__a); }
|
|
__DEVICE__ float __fsub_rd(float __a, float __b) {
|
|
return __nv_fsub_rd(__a, __b);
|
|
}
|
|
__DEVICE__ float __fsub_rn(float __a, float __b) {
|
|
return __nv_fsub_rn(__a, __b);
|
|
}
|
|
__DEVICE__ float __fsub_ru(float __a, float __b) {
|
|
return __nv_fsub_ru(__a, __b);
|
|
}
|
|
__DEVICE__ float __fsub_rz(float __a, float __b) {
|
|
return __nv_fsub_rz(__a, __b);
|
|
}
|
|
__DEVICE__ int __hadd(int __a, int __b) { return __nv_hadd(__a, __b); }
|
|
__DEVICE__ double __hiloint2double(int __a, int __b) {
|
|
return __nv_hiloint2double(__a, __b);
|
|
}
|
|
__DEVICE__ int __iAtomicAdd(int *__p, int __v) {
|
|
return __nvvm_atom_add_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicAdd_block(int *__p, int __v) {
|
|
return __nvvm_atom_cta_add_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicAdd_system(int *__p, int __v) {
|
|
return __nvvm_atom_sys_add_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicAnd(int *__p, int __v) {
|
|
return __nvvm_atom_and_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicAnd_block(int *__p, int __v) {
|
|
return __nvvm_atom_cta_and_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicAnd_system(int *__p, int __v) {
|
|
return __nvvm_atom_sys_and_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicCAS(int *__p, int __cmp, int __v) {
|
|
return __nvvm_atom_cas_gen_i(__p, __cmp, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicCAS_block(int *__p, int __cmp, int __v) {
|
|
return __nvvm_atom_cta_cas_gen_i(__p, __cmp, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicCAS_system(int *__p, int __cmp, int __v) {
|
|
return __nvvm_atom_sys_cas_gen_i(__p, __cmp, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicExch(int *__p, int __v) {
|
|
return __nvvm_atom_xchg_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicExch_block(int *__p, int __v) {
|
|
return __nvvm_atom_cta_xchg_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicExch_system(int *__p, int __v) {
|
|
return __nvvm_atom_sys_xchg_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicMax(int *__p, int __v) {
|
|
return __nvvm_atom_max_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicMax_block(int *__p, int __v) {
|
|
return __nvvm_atom_cta_max_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicMax_system(int *__p, int __v) {
|
|
return __nvvm_atom_sys_max_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicMin(int *__p, int __v) {
|
|
return __nvvm_atom_min_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicMin_block(int *__p, int __v) {
|
|
return __nvvm_atom_cta_min_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicMin_system(int *__p, int __v) {
|
|
return __nvvm_atom_sys_min_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicOr(int *__p, int __v) {
|
|
return __nvvm_atom_or_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicOr_block(int *__p, int __v) {
|
|
return __nvvm_atom_cta_or_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicOr_system(int *__p, int __v) {
|
|
return __nvvm_atom_sys_or_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicXor(int *__p, int __v) {
|
|
return __nvvm_atom_xor_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicXor_block(int *__p, int __v) {
|
|
return __nvvm_atom_cta_xor_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ int __iAtomicXor_system(int *__p, int __v) {
|
|
return __nvvm_atom_sys_xor_gen_i(__p, __v);
|
|
}
|
|
__DEVICE__ long long __illAtomicMax(long long *__p, long long __v) {
|
|
return __nvvm_atom_max_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __illAtomicMax_block(long long *__p, long long __v) {
|
|
return __nvvm_atom_cta_max_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __illAtomicMax_system(long long *__p, long long __v) {
|
|
return __nvvm_atom_sys_max_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __illAtomicMin(long long *__p, long long __v) {
|
|
return __nvvm_atom_min_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __illAtomicMin_block(long long *__p, long long __v) {
|
|
return __nvvm_atom_cta_min_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __illAtomicMin_system(long long *__p, long long __v) {
|
|
return __nvvm_atom_sys_min_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ double __int2double_rn(int __a) { return __nv_int2double_rn(__a); }
|
|
__DEVICE__ float __int2float_rd(int __a) { return __nv_int2float_rd(__a); }
|
|
__DEVICE__ float __int2float_rn(int __a) { return __nv_int2float_rn(__a); }
|
|
__DEVICE__ float __int2float_ru(int __a) { return __nv_int2float_ru(__a); }
|
|
__DEVICE__ float __int2float_rz(int __a) { return __nv_int2float_rz(__a); }
|
|
__DEVICE__ float __int_as_float(int __a) { return __nv_int_as_float(__a); }
|
|
__DEVICE__ int __isfinited(double __a) { return __nv_isfinited(__a); }
|
|
__DEVICE__ int __isinf(double __a) { return __nv_isinfd(__a); }
|
|
__DEVICE__ int __isinff(float __a) { return __nv_isinff(__a); }
|
|
#ifdef _MSC_VER
|
|
__DEVICE__ int __isinfl(long double __a);
|
|
#endif
|
|
__DEVICE__ int __isnan(double __a) { return __nv_isnand(__a); }
|
|
__DEVICE__ int __isnanf(float __a) { return __nv_isnanf(__a); }
|
|
#ifdef _MSC_VER
|
|
__DEVICE__ int __isnanl(long double __a);
|
|
#endif
|
|
__DEVICE__ double __ll2double_rd(long long __a) {
|
|
return __nv_ll2double_rd(__a);
|
|
}
|
|
__DEVICE__ double __ll2double_rn(long long __a) {
|
|
return __nv_ll2double_rn(__a);
|
|
}
|
|
__DEVICE__ double __ll2double_ru(long long __a) {
|
|
return __nv_ll2double_ru(__a);
|
|
}
|
|
__DEVICE__ double __ll2double_rz(long long __a) {
|
|
return __nv_ll2double_rz(__a);
|
|
}
|
|
__DEVICE__ float __ll2float_rd(long long __a) { return __nv_ll2float_rd(__a); }
|
|
__DEVICE__ float __ll2float_rn(long long __a) { return __nv_ll2float_rn(__a); }
|
|
__DEVICE__ float __ll2float_ru(long long __a) { return __nv_ll2float_ru(__a); }
|
|
__DEVICE__ float __ll2float_rz(long long __a) { return __nv_ll2float_rz(__a); }
|
|
__DEVICE__ long long __llAtomicAnd(long long *__p, long long __v) {
|
|
return __nvvm_atom_and_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __llAtomicAnd_block(long long *__p, long long __v) {
|
|
return __nvvm_atom_cta_and_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __llAtomicAnd_system(long long *__p, long long __v) {
|
|
return __nvvm_atom_sys_and_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __llAtomicOr(long long *__p, long long __v) {
|
|
return __nvvm_atom_or_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __llAtomicOr_block(long long *__p, long long __v) {
|
|
return __nvvm_atom_cta_or_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __llAtomicOr_system(long long *__p, long long __v) {
|
|
return __nvvm_atom_sys_or_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __llAtomicXor(long long *__p, long long __v) {
|
|
return __nvvm_atom_xor_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __llAtomicXor_block(long long *__p, long long __v) {
|
|
return __nvvm_atom_cta_xor_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ long long __llAtomicXor_system(long long *__p, long long __v) {
|
|
return __nvvm_atom_sys_xor_gen_ll(__p, __v);
|
|
}
|
|
__DEVICE__ float __log10f(float __a) { return __nv_fast_log10f(__a); }
|
|
__DEVICE__ float __log2f(float __a) { return __nv_fast_log2f(__a); }
|
|
__DEVICE__ float __logf(float __a) { return __nv_fast_logf(__a); }
|
|
__DEVICE__ double __longlong_as_double(long long __a) {
|
|
return __nv_longlong_as_double(__a);
|
|
}
|
|
__DEVICE__ int __mul24(int __a, int __b) { return __nv_mul24(__a, __b); }
|
|
__DEVICE__ long long __mul64hi(long long __a, long long __b) {
|
|
return __nv_mul64hi(__a, __b);
|
|
}
|
|
__DEVICE__ int __mulhi(int __a, int __b) { return __nv_mulhi(__a, __b); }
|
|
__DEVICE__ unsigned int __pm0(void) { return __nvvm_read_ptx_sreg_pm0(); }
|
|
__DEVICE__ unsigned int __pm1(void) { return __nvvm_read_ptx_sreg_pm1(); }
|
|
__DEVICE__ unsigned int __pm2(void) { return __nvvm_read_ptx_sreg_pm2(); }
|
|
__DEVICE__ unsigned int __pm3(void) { return __nvvm_read_ptx_sreg_pm3(); }
|
|
__DEVICE__ int __popc(int __a) { return __nv_popc(__a); }
|
|
__DEVICE__ int __popcll(long long __a) { return __nv_popcll(__a); }
|
|
__DEVICE__ float __powf(float __a, float __b) {
|
|
return __nv_fast_powf(__a, __b);
|
|
}
|
|
|
|
// Parameter must have a known integer value.
|
|
#define __prof_trigger(__a) __asm__ __volatile__("pmevent \t%0;" ::"i"(__a))
|
|
__DEVICE__ int __rhadd(int __a, int __b) { return __nv_rhadd(__a, __b); }
|
|
__DEVICE__ unsigned int __sad(int __a, int __b, unsigned int __c) {
|
|
return __nv_sad(__a, __b, __c);
|
|
}
|
|
__DEVICE__ float __saturatef(float __a) { return __nv_saturatef(__a); }
|
|
__DEVICE__ int __signbitd(double __a) { return __nv_signbitd(__a); }
|
|
__DEVICE__ int __signbitf(float __a) { return __nv_signbitf(__a); }
|
|
__DEVICE__ void __sincosf(float __a, float *__s, float *__c) {
|
|
return __nv_fast_sincosf(__a, __s, __c);
|
|
}
|
|
__DEVICE__ float __sinf(float __a) { return __nv_fast_sinf(__a); }
|
|
__DEVICE__ int __syncthreads_and(int __a) { return __nvvm_bar0_and(__a); }
|
|
__DEVICE__ int __syncthreads_count(int __a) { return __nvvm_bar0_popc(__a); }
|
|
__DEVICE__ int __syncthreads_or(int __a) { return __nvvm_bar0_or(__a); }
|
|
__DEVICE__ float __tanf(float __a) { return __nv_fast_tanf(__a); }
|
|
__DEVICE__ void __threadfence(void) { __nvvm_membar_gl(); }
|
|
__DEVICE__ void __threadfence_block(void) { __nvvm_membar_cta(); };
|
|
__DEVICE__ void __threadfence_system(void) { __nvvm_membar_sys(); };
|
|
__DEVICE__ void __trap(void) { __asm__ __volatile__("trap;"); }
|
|
__DEVICE__ unsigned int __uAtomicAdd(unsigned int *__p, unsigned int __v) {
|
|
return __nvvm_atom_add_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicAdd_block(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_cta_add_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicAdd_system(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_sys_add_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicAnd(unsigned int *__p, unsigned int __v) {
|
|
return __nvvm_atom_and_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicAnd_block(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_cta_and_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicAnd_system(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_sys_and_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicCAS(unsigned int *__p, unsigned int __cmp,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_cas_gen_i((int *)__p, __cmp, __v);
|
|
}
|
|
__DEVICE__ unsigned int
|
|
__uAtomicCAS_block(unsigned int *__p, unsigned int __cmp, unsigned int __v) {
|
|
return __nvvm_atom_cta_cas_gen_i((int *)__p, __cmp, __v);
|
|
}
|
|
__DEVICE__ unsigned int
|
|
__uAtomicCAS_system(unsigned int *__p, unsigned int __cmp, unsigned int __v) {
|
|
return __nvvm_atom_sys_cas_gen_i((int *)__p, __cmp, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicDec(unsigned int *__p, unsigned int __v) {
|
|
return __nvvm_atom_dec_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicDec_block(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_cta_dec_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicDec_system(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_sys_dec_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicExch(unsigned int *__p, unsigned int __v) {
|
|
return __nvvm_atom_xchg_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicExch_block(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_cta_xchg_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicExch_system(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_sys_xchg_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicInc(unsigned int *__p, unsigned int __v) {
|
|
return __nvvm_atom_inc_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicInc_block(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_cta_inc_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicInc_system(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_sys_inc_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicMax(unsigned int *__p, unsigned int __v) {
|
|
return __nvvm_atom_max_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicMax_block(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_cta_max_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicMax_system(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_sys_max_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicMin(unsigned int *__p, unsigned int __v) {
|
|
return __nvvm_atom_min_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicMin_block(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_cta_min_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicMin_system(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_sys_min_gen_ui(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicOr(unsigned int *__p, unsigned int __v) {
|
|
return __nvvm_atom_or_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicOr_block(unsigned int *__p, unsigned int __v) {
|
|
return __nvvm_atom_cta_or_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicOr_system(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_sys_or_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicXor(unsigned int *__p, unsigned int __v) {
|
|
return __nvvm_atom_xor_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicXor_block(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_cta_xor_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uAtomicXor_system(unsigned int *__p,
|
|
unsigned int __v) {
|
|
return __nvvm_atom_sys_xor_gen_i((int *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __uhadd(unsigned int __a, unsigned int __b) {
|
|
return __nv_uhadd(__a, __b);
|
|
}
|
|
__DEVICE__ double __uint2double_rn(unsigned int __a) {
|
|
return __nv_uint2double_rn(__a);
|
|
}
|
|
__DEVICE__ float __uint2float_rd(unsigned int __a) {
|
|
return __nv_uint2float_rd(__a);
|
|
}
|
|
__DEVICE__ float __uint2float_rn(unsigned int __a) {
|
|
return __nv_uint2float_rn(__a);
|
|
}
|
|
__DEVICE__ float __uint2float_ru(unsigned int __a) {
|
|
return __nv_uint2float_ru(__a);
|
|
}
|
|
__DEVICE__ float __uint2float_rz(unsigned int __a) {
|
|
return __nv_uint2float_rz(__a);
|
|
}
|
|
__DEVICE__ float __uint_as_float(unsigned int __a) {
|
|
return __nv_uint_as_float(__a);
|
|
} //
|
|
__DEVICE__ double __ull2double_rd(unsigned long long __a) {
|
|
return __nv_ull2double_rd(__a);
|
|
}
|
|
__DEVICE__ double __ull2double_rn(unsigned long long __a) {
|
|
return __nv_ull2double_rn(__a);
|
|
}
|
|
__DEVICE__ double __ull2double_ru(unsigned long long __a) {
|
|
return __nv_ull2double_ru(__a);
|
|
}
|
|
__DEVICE__ double __ull2double_rz(unsigned long long __a) {
|
|
return __nv_ull2double_rz(__a);
|
|
}
|
|
__DEVICE__ float __ull2float_rd(unsigned long long __a) {
|
|
return __nv_ull2float_rd(__a);
|
|
}
|
|
__DEVICE__ float __ull2float_rn(unsigned long long __a) {
|
|
return __nv_ull2float_rn(__a);
|
|
}
|
|
__DEVICE__ float __ull2float_ru(unsigned long long __a) {
|
|
return __nv_ull2float_ru(__a);
|
|
}
|
|
__DEVICE__ float __ull2float_rz(unsigned long long __a) {
|
|
return __nv_ull2float_rz(__a);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicAdd(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_add_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicAdd_block(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_cta_add_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicAdd_system(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_sys_add_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicAnd(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_and_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicAnd_block(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_cta_and_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicAnd_system(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_sys_and_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicCAS(unsigned long long *__p,
|
|
unsigned long long __cmp,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_cas_gen_ll((long long *)__p, __cmp, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicCAS_block(unsigned long long *__p,
|
|
unsigned long long __cmp,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_cta_cas_gen_ll((long long *)__p, __cmp, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicCAS_system(unsigned long long *__p,
|
|
unsigned long long __cmp,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_sys_cas_gen_ll((long long *)__p, __cmp, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicExch(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_xchg_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicExch_block(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_cta_xchg_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicExch_system(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_sys_xchg_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicMax(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_max_gen_ull(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicMax_block(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_cta_max_gen_ull(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicMax_system(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_sys_max_gen_ull(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicMin(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_min_gen_ull(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicMin_block(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_cta_min_gen_ull(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicMin_system(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_sys_min_gen_ull(__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicOr(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_or_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicOr_block(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_cta_or_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicOr_system(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_sys_or_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicXor(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_xor_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicXor_block(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_cta_xor_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned long long __ullAtomicXor_system(unsigned long long *__p,
|
|
unsigned long long __v) {
|
|
return __nvvm_atom_sys_xor_gen_ll((long long *)__p, __v);
|
|
}
|
|
__DEVICE__ unsigned int __umul24(unsigned int __a, unsigned int __b) {
|
|
return __nv_umul24(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned long long __umul64hi(unsigned long long __a,
|
|
unsigned long long __b) {
|
|
return __nv_umul64hi(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __umulhi(unsigned int __a, unsigned int __b) {
|
|
return __nv_umulhi(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __urhadd(unsigned int __a, unsigned int __b) {
|
|
return __nv_urhadd(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __usad(unsigned int __a, unsigned int __b,
|
|
unsigned int __c) {
|
|
return __nv_usad(__a, __b, __c);
|
|
}
|
|
|
|
#if CUDA_VERSION >= 9000 && CUDA_VERSION < 9020
|
|
__DEVICE__ unsigned int __vabs2(unsigned int __a) { return __nv_vabs2(__a); }
|
|
__DEVICE__ unsigned int __vabs4(unsigned int __a) { return __nv_vabs4(__a); }
|
|
__DEVICE__ unsigned int __vabsdiffs2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vabsdiffs2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vabsdiffs4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vabsdiffs4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vabsdiffu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vabsdiffu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vabsdiffu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vabsdiffu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vabsss2(unsigned int __a) {
|
|
return __nv_vabsss2(__a);
|
|
}
|
|
__DEVICE__ unsigned int __vabsss4(unsigned int __a) {
|
|
return __nv_vabsss4(__a);
|
|
}
|
|
__DEVICE__ unsigned int __vadd2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vadd2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vadd4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vadd4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vaddss2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vaddss2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vaddss4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vaddss4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vaddus2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vaddus2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vaddus4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vaddus4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vavgs2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vavgs2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vavgs4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vavgs4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vavgu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vavgu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vavgu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vavgu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpeq2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpeq2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpeq4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpeq4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpges2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpges2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpges4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpges4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgeu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpgeu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgeu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpgeu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgts2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpgts2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgts4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpgts4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgtu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpgtu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgtu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpgtu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmples2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmples2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmples4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmples4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpleu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpleu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpleu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpleu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmplts2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmplts2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmplts4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmplts4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpltu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpltu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpltu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpltu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpne2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpne2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vcmpne4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vcmpne4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vhaddu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vhaddu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vhaddu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vhaddu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vmaxs2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vmaxs2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vmaxs4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vmaxs4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vmaxu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vmaxu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vmaxu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vmaxu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vmins2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vmins2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vmins4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vmins4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vminu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vminu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vminu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vminu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vneg2(unsigned int __a) { return __nv_vneg2(__a); }
|
|
__DEVICE__ unsigned int __vneg4(unsigned int __a) { return __nv_vneg4(__a); }
|
|
__DEVICE__ unsigned int __vnegss2(unsigned int __a) {
|
|
return __nv_vnegss2(__a);
|
|
}
|
|
__DEVICE__ unsigned int __vnegss4(unsigned int __a) {
|
|
return __nv_vnegss4(__a);
|
|
}
|
|
__DEVICE__ unsigned int __vsads2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsads2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsads4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsads4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsadu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsadu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsadu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsadu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vseteq2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vseteq2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vseteq4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vseteq4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetges2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetges2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetges4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetges4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgeu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetgeu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgeu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetgeu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgts2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetgts2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgts4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetgts4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgtu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetgtu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgtu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetgtu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetles2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetles2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetles4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetles4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetleu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetleu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetleu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetleu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetlts2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetlts2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetlts4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetlts4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetltu2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetltu2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetltu4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetltu4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetne2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetne2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsetne4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsetne4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsub2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsub2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsub4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsub4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsubss2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsubss2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsubss4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsubss4(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsubus2(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsubus2(__a, __b);
|
|
}
|
|
__DEVICE__ unsigned int __vsubus4(unsigned int __a, unsigned int __b) {
|
|
return __nv_vsubus4(__a, __b);
|
|
}
|
|
#else // CUDA_VERSION >= 9020
|
|
// CUDA no longer provides inline assembly (or bitcode) implementation of these
|
|
// functions, so we have to reimplment them. The implementation is naive and is
|
|
// not optimized for performance.
|
|
|
|
// Helper function to convert N-bit boolean subfields into all-0 or all-1.
|
|
// E.g. __bool2mask(0x01000100,8) -> 0xff00ff00
|
|
// __bool2mask(0x00010000,16) -> 0xffff0000
|
|
__DEVICE__ unsigned int __bool2mask(unsigned int __a, int shift) {
|
|
return (__a << shift) - __a;
|
|
}
|
|
__DEVICE__ unsigned int __vabs2(unsigned int __a) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff2.s32.s32.s32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(0), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vabs4(unsigned int __a) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff4.s32.s32.s32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(0), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vabsdiffs2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff2.s32.s32.s32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
|
|
__DEVICE__ unsigned int __vabsdiffs4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff4.s32.s32.s32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vabsdiffu2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff2.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vabsdiffu4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff4.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vabsss2(unsigned int __a) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff2.s32.s32.s32.sat %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(0), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vabsss4(unsigned int __a) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff4.s32.s32.s32.sat %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(0), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vadd2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vadd2.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vadd4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vadd4.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vaddss2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vadd2.s32.s32.s32.sat %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vaddss4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vadd4.s32.s32.s32.sat %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vaddus2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vadd2.u32.u32.u32.sat %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vaddus4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vadd4.u32.u32.u32.sat %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vavgs2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vavrg2.s32.s32.s32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vavgs4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vavrg4.s32.s32.s32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vavgu2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vavrg2.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vavgu4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vavrg4.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vseteq2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset2.u32.u32.eq %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpeq2(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vseteq2(__a, __b), 16);
|
|
}
|
|
__DEVICE__ unsigned int __vseteq4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset4.u32.u32.eq %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpeq4(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vseteq4(__a, __b), 8);
|
|
}
|
|
__DEVICE__ unsigned int __vsetges2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset2.s32.s32.ge %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpges2(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetges2(__a, __b), 16);
|
|
}
|
|
__DEVICE__ unsigned int __vsetges4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset4.s32.s32.ge %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpges4(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetges4(__a, __b), 8);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgeu2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset2.u32.u32.ge %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgeu2(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetgeu2(__a, __b), 16);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgeu4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset4.u32.u32.ge %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgeu4(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetgeu4(__a, __b), 8);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgts2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset2.s32.s32.gt %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgts2(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetgts2(__a, __b), 16);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgts4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset4.s32.s32.gt %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgts4(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetgts4(__a, __b), 8);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgtu2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset2.u32.u32.gt %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgtu2(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetgtu2(__a, __b), 16);
|
|
}
|
|
__DEVICE__ unsigned int __vsetgtu4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset4.u32.u32.gt %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpgtu4(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetgtu4(__a, __b), 8);
|
|
}
|
|
__DEVICE__ unsigned int __vsetles2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset2.s32.s32.le %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmples2(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetles2(__a, __b), 16);
|
|
}
|
|
__DEVICE__ unsigned int __vsetles4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset4.s32.s32.le %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmples4(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetles4(__a, __b), 8);
|
|
}
|
|
__DEVICE__ unsigned int __vsetleu2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset2.u32.u32.le %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpleu2(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetleu2(__a, __b), 16);
|
|
}
|
|
__DEVICE__ unsigned int __vsetleu4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset4.u32.u32.le %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpleu4(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetleu4(__a, __b), 8);
|
|
}
|
|
__DEVICE__ unsigned int __vsetlts2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset2.s32.s32.lt %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmplts2(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetlts2(__a, __b), 16);
|
|
}
|
|
__DEVICE__ unsigned int __vsetlts4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset4.s32.s32.lt %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmplts4(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetlts4(__a, __b), 8);
|
|
}
|
|
__DEVICE__ unsigned int __vsetltu2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset2.u32.u32.lt %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpltu2(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetltu2(__a, __b), 16);
|
|
}
|
|
__DEVICE__ unsigned int __vsetltu4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset4.u32.u32.lt %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpltu4(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetltu4(__a, __b), 8);
|
|
}
|
|
__DEVICE__ unsigned int __vsetne2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset2.u32.u32.ne %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpne2(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetne2(__a, __b), 16);
|
|
}
|
|
__DEVICE__ unsigned int __vsetne4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vset4.u32.u32.ne %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vcmpne4(unsigned int __a, unsigned int __b) {
|
|
return __bool2mask(__vsetne4(__a, __b), 8);
|
|
}
|
|
|
|
// Based on ITEM 23 in AIM-239: http://dspace.mit.edu/handle/1721.1/6086
|
|
// (a & b) + (a | b) = a + b = (a ^ b) + 2 * (a & b) =>
|
|
// (a + b) / 2 = ((a ^ b) >> 1) + (a & b)
|
|
// To operate on multiple sub-elements we need to make sure to mask out bits
|
|
// that crossed over into adjacent elements during the shift.
|
|
__DEVICE__ unsigned int __vhaddu2(unsigned int __a, unsigned int __b) {
|
|
return (((__a ^ __b) >> 1) & ~0x80008000u) + (__a & __b);
|
|
}
|
|
__DEVICE__ unsigned int __vhaddu4(unsigned int __a, unsigned int __b) {
|
|
return (((__a ^ __b) >> 1) & ~0x80808080u) + (__a & __b);
|
|
}
|
|
|
|
__DEVICE__ unsigned int __vmaxs2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
if ((__a & 0x8000) && (__b & 0x8000)) {
|
|
// Work around a bug in ptxas which produces invalid result if low element
|
|
// is negative.
|
|
unsigned mask = __vcmpgts2(__a, __b);
|
|
r = (__a & mask) | (__b & ~mask);
|
|
} else {
|
|
__asm__("vmax2.s32.s32.s32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
}
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vmaxs4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vmax4.s32.s32.s32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vmaxu2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vmax2.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vmaxu4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vmax4.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vmins2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vmin2.s32.s32.s32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vmins4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vmin4.s32.s32.s32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vminu2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vmin2.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vminu4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vmin4.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vsads2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff2.s32.s32.s32.add %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vsads4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff4.s32.s32.s32.add %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vsadu2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff2.u32.u32.u32.add %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vsadu4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vabsdiff4.u32.u32.u32.add %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
|
|
__DEVICE__ unsigned int __vsub2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vsub2.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vneg2(unsigned int __a) { return __vsub2(0, __a); }
|
|
|
|
__DEVICE__ unsigned int __vsub4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vsub4.u32.u32.u32 %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vneg4(unsigned int __a) { return __vsub4(0, __a); }
|
|
__DEVICE__ unsigned int __vsubss2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vsub2.s32.s32.s32.sat %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vnegss2(unsigned int __a) {
|
|
return __vsubss2(0, __a);
|
|
}
|
|
__DEVICE__ unsigned int __vsubss4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vsub4.s32.s32.s32.sat %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vnegss4(unsigned int __a) {
|
|
return __vsubss4(0, __a);
|
|
}
|
|
__DEVICE__ unsigned int __vsubus2(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vsub2.u32.u32.u32.sat %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
__DEVICE__ unsigned int __vsubus4(unsigned int __a, unsigned int __b) {
|
|
unsigned int r;
|
|
__asm__("vsub4.u32.u32.u32.sat %0,%1,%2,%3;"
|
|
: "=r"(r)
|
|
: "r"(__a), "r"(__b), "r"(0));
|
|
return r;
|
|
}
|
|
#endif // CUDA_VERSION >= 9020
|
|
|
|
// For OpenMP we require the user to include <time.h> as we need to know what
|
|
// clock_t is on the system.
|
|
#ifndef __OPENMP_NVPTX__
|
|
__DEVICE__ /* clock_t= */ int clock() { return __nvvm_read_ptx_sreg_clock(); }
|
|
#endif
|
|
__DEVICE__ long long clock64() { return __nvvm_read_ptx_sreg_clock64(); }
|
|
|
|
// These functions shouldn't be declared when including this header
|
|
// for math function resolution purposes.
|
|
#ifndef __OPENMP_NVPTX__
|
|
__DEVICE__ void *memcpy(void *__a, const void *__b, size_t __c) {
|
|
return __builtin_memcpy(__a, __b, __c);
|
|
}
|
|
__DEVICE__ void *memset(void *__a, int __b, size_t __c) {
|
|
return __builtin_memset(__a, __b, __c);
|
|
}
|
|
#endif
|
|
|
|
#pragma pop_macro("__DEVICE__")
|
|
#endif // __CLANG_CUDA_DEVICE_FUNCTIONS_H__
|