[X86] Add builtins for VALIGNQ/VALIGND to enable proper target feature checking.

We still emit shufflevector instructions we just do it from CGBuiltin.cpp now. This ensures the intrinsics that use this are only available on CPUs that support the feature.

I also added range checking to the immediate, but only checked it is 8 bits or smaller. We should maybe be stricter since we never use all 8 bits, but gcc doesn't seem to do that.

llvm-svn: 334237
This commit is contained in:
Craig Topper 2018-06-07 21:27:41 +00:00
parent 188a619e56
commit e56819eb69
5 changed files with 44 additions and 54 deletions

View File

@ -909,6 +909,12 @@ TARGET_BUILTIN(__builtin_ia32_storeupd512_mask, "vd*V8dUc", "n", "avx512f")
TARGET_BUILTIN(__builtin_ia32_storeapd512_mask, "vV8d*V8dUc", "n", "avx512f") TARGET_BUILTIN(__builtin_ia32_storeapd512_mask, "vV8d*V8dUc", "n", "avx512f")
TARGET_BUILTIN(__builtin_ia32_storeups512_mask, "vf*V16fUs", "n", "avx512f") TARGET_BUILTIN(__builtin_ia32_storeups512_mask, "vf*V16fUs", "n", "avx512f")
TARGET_BUILTIN(__builtin_ia32_storeaps512_mask, "vV16f*V16fUs", "n", "avx512f") TARGET_BUILTIN(__builtin_ia32_storeaps512_mask, "vV16f*V16fUs", "n", "avx512f")
TARGET_BUILTIN(__builtin_ia32_alignq512, "V8LLiV8LLiV8LLiIi", "nc", "avx512f")
TARGET_BUILTIN(__builtin_ia32_alignd512, "V16iV16iV16iIi", "nc", "avx512f")
TARGET_BUILTIN(__builtin_ia32_alignd128, "V4iV4iV4iIi", "nc", "avx512vl")
TARGET_BUILTIN(__builtin_ia32_alignd256, "V8iV8iV8iIi", "nc", "avx512vl")
TARGET_BUILTIN(__builtin_ia32_alignq128, "V2LLiV2LLiV2LLiIi", "nc", "avx512vl")
TARGET_BUILTIN(__builtin_ia32_alignq256, "V4LLiV4LLiV4LLiIi", "nc", "avx512vl")
TARGET_BUILTIN(__builtin_ia32_vpdpbusd128, "V4iV4iV4iV4i", "nc", "avx512vl,avx512vnni") TARGET_BUILTIN(__builtin_ia32_vpdpbusd128, "V4iV4iV4iV4i", "nc", "avx512vl,avx512vnni")
TARGET_BUILTIN(__builtin_ia32_vpdpbusd256, "V8iV8iV8iV8i", "nc", "avx512vl,avx512vnni") TARGET_BUILTIN(__builtin_ia32_vpdpbusd256, "V8iV8iV8iV8i", "nc", "avx512vl,avx512vnni")

View File

@ -9222,6 +9222,26 @@ Value *CodeGenFunction::EmitX86BuiltinExpr(unsigned BuiltinID,
makeArrayRef(Indices, NumElts), makeArrayRef(Indices, NumElts),
"palignr"); "palignr");
} }
case X86::BI__builtin_ia32_alignd128:
case X86::BI__builtin_ia32_alignd256:
case X86::BI__builtin_ia32_alignd512:
case X86::BI__builtin_ia32_alignq128:
case X86::BI__builtin_ia32_alignq256:
case X86::BI__builtin_ia32_alignq512: {
unsigned NumElts = Ops[0]->getType()->getVectorNumElements();
unsigned ShiftVal = cast<llvm::ConstantInt>(Ops[2])->getZExtValue();
// Mask the shift amount to width of two vectors.
ShiftVal &= (2 * NumElts) - 1;
uint32_t Indices[16];
for (unsigned i = 0; i != NumElts; ++i)
Indices[i] = i + ShiftVal;
return Builder.CreateShuffleVector(Ops[1], Ops[0],
makeArrayRef(Indices, NumElts),
"valign");
}
case X86::BI__builtin_ia32_vperm2f128_pd256: case X86::BI__builtin_ia32_vperm2f128_pd256:
case X86::BI__builtin_ia32_vperm2f128_ps256: case X86::BI__builtin_ia32_vperm2f128_ps256:

View File

@ -3465,16 +3465,8 @@ _mm512_maskz_permutex2var_epi64(__mmask8 __U, __m512i __A, __m512i __I,
} }
#define _mm512_alignr_epi64(A, B, I) \ #define _mm512_alignr_epi64(A, B, I) \
(__m512i)__builtin_shufflevector((__v8di)(__m512i)(B), \ (__m512i)__builtin_ia32_alignq512((__v8di)(__m512i)(A), \
(__v8di)(__m512i)(A), \ (__v8di)(__m512i)(B), (int)(I))
((int)(I) & 0x7) + 0, \
((int)(I) & 0x7) + 1, \
((int)(I) & 0x7) + 2, \
((int)(I) & 0x7) + 3, \
((int)(I) & 0x7) + 4, \
((int)(I) & 0x7) + 5, \
((int)(I) & 0x7) + 6, \
((int)(I) & 0x7) + 7)
#define _mm512_mask_alignr_epi64(W, U, A, B, imm) \ #define _mm512_mask_alignr_epi64(W, U, A, B, imm) \
(__m512i)__builtin_ia32_selectq_512((__mmask8)(U), \ (__m512i)__builtin_ia32_selectq_512((__mmask8)(U), \
@ -3487,24 +3479,8 @@ _mm512_maskz_permutex2var_epi64(__mmask8 __U, __m512i __A, __m512i __I,
(__v8di)_mm512_setzero_si512()) (__v8di)_mm512_setzero_si512())
#define _mm512_alignr_epi32(A, B, I) \ #define _mm512_alignr_epi32(A, B, I) \
(__m512i)__builtin_shufflevector((__v16si)(__m512i)(B), \ (__m512i)__builtin_ia32_alignd512((__v16si)(__m512i)(A), \
(__v16si)(__m512i)(A), \ (__v16si)(__m512i)(B), (int)(I))
((int)(I) & 0xf) + 0, \
((int)(I) & 0xf) + 1, \
((int)(I) & 0xf) + 2, \
((int)(I) & 0xf) + 3, \
((int)(I) & 0xf) + 4, \
((int)(I) & 0xf) + 5, \
((int)(I) & 0xf) + 6, \
((int)(I) & 0xf) + 7, \
((int)(I) & 0xf) + 8, \
((int)(I) & 0xf) + 9, \
((int)(I) & 0xf) + 10, \
((int)(I) & 0xf) + 11, \
((int)(I) & 0xf) + 12, \
((int)(I) & 0xf) + 13, \
((int)(I) & 0xf) + 14, \
((int)(I) & 0xf) + 15)
#define _mm512_mask_alignr_epi32(W, U, A, B, imm) \ #define _mm512_mask_alignr_epi32(W, U, A, B, imm) \
(__m512i)__builtin_ia32_selectd_512((__mmask16)(U), \ (__m512i)__builtin_ia32_selectd_512((__mmask16)(U), \

View File

@ -8082,12 +8082,8 @@ _mm256_maskz_permutexvar_epi32(__mmask8 __M, __m256i __X, __m256i __Y)
} }
#define _mm_alignr_epi32(A, B, imm) \ #define _mm_alignr_epi32(A, B, imm) \
(__m128i)__builtin_shufflevector((__v4si)(__m128i)(B), \ (__m128i)__builtin_ia32_alignd128((__v4si)(__m128i)(A), \
(__v4si)(__m128i)(A), \ (__v4si)(__m128i)(B), (int)(imm))
((int)(imm) & 0x3) + 0, \
((int)(imm) & 0x3) + 1, \
((int)(imm) & 0x3) + 2, \
((int)(imm) & 0x3) + 3)
#define _mm_mask_alignr_epi32(W, U, A, B, imm) \ #define _mm_mask_alignr_epi32(W, U, A, B, imm) \
(__m128i)__builtin_ia32_selectd_128((__mmask8)(U), \ (__m128i)__builtin_ia32_selectd_128((__mmask8)(U), \
@ -8100,16 +8096,8 @@ _mm256_maskz_permutexvar_epi32(__mmask8 __M, __m256i __X, __m256i __Y)
(__v4si)_mm_setzero_si128()) (__v4si)_mm_setzero_si128())
#define _mm256_alignr_epi32(A, B, imm) \ #define _mm256_alignr_epi32(A, B, imm) \
(__m256i)__builtin_shufflevector((__v8si)(__m256i)(B), \ (__m256i)__builtin_ia32_alignd256((__v8si)(__m256i)(A), \
(__v8si)(__m256i)(A), \ (__v8si)(__m256i)(B), (int)(imm))
((int)(imm) & 0x7) + 0, \
((int)(imm) & 0x7) + 1, \
((int)(imm) & 0x7) + 2, \
((int)(imm) & 0x7) + 3, \
((int)(imm) & 0x7) + 4, \
((int)(imm) & 0x7) + 5, \
((int)(imm) & 0x7) + 6, \
((int)(imm) & 0x7) + 7)
#define _mm256_mask_alignr_epi32(W, U, A, B, imm) \ #define _mm256_mask_alignr_epi32(W, U, A, B, imm) \
(__m256i)__builtin_ia32_selectd_256((__mmask8)(U), \ (__m256i)__builtin_ia32_selectd_256((__mmask8)(U), \
@ -8122,10 +8110,8 @@ _mm256_maskz_permutexvar_epi32(__mmask8 __M, __m256i __X, __m256i __Y)
(__v8si)_mm256_setzero_si256()) (__v8si)_mm256_setzero_si256())
#define _mm_alignr_epi64(A, B, imm) \ #define _mm_alignr_epi64(A, B, imm) \
(__m128i)__builtin_shufflevector((__v2di)(__m128i)(B), \ (__m128i)__builtin_ia32_alignq128((__v2di)(__m128i)(A), \
(__v2di)(__m128i)(A), \ (__v2di)(__m128i)(B), (int)(imm))
((int)(imm) & 0x1) + 0, \
((int)(imm) & 0x1) + 1)
#define _mm_mask_alignr_epi64(W, U, A, B, imm) \ #define _mm_mask_alignr_epi64(W, U, A, B, imm) \
(__m128i)__builtin_ia32_selectq_128((__mmask8)(U), \ (__m128i)__builtin_ia32_selectq_128((__mmask8)(U), \
@ -8138,12 +8124,8 @@ _mm256_maskz_permutexvar_epi32(__mmask8 __M, __m256i __X, __m256i __Y)
(__v2di)_mm_setzero_si128()) (__v2di)_mm_setzero_si128())
#define _mm256_alignr_epi64(A, B, imm) \ #define _mm256_alignr_epi64(A, B, imm) \
(__m256i)__builtin_shufflevector((__v4di)(__m256i)(B), \ (__m256i)__builtin_ia32_alignq256((__v4di)(__m256i)(A), \
(__v4di)(__m256i)(A), \ (__v4di)(__m256i)(B), (int)(imm))
((int)(imm) & 0x3) + 0, \
((int)(imm) & 0x3) + 1, \
((int)(imm) & 0x3) + 2, \
((int)(imm) & 0x3) + 3)
#define _mm256_mask_alignr_epi64(W, U, A, B, imm) \ #define _mm256_mask_alignr_epi64(W, U, A, B, imm) \
(__m256i)__builtin_ia32_selectq_256((__mmask8)(U), \ (__m256i)__builtin_ia32_selectq_256((__mmask8)(U), \

View File

@ -2712,6 +2712,12 @@ bool Sema::CheckX86BuiltinFunctionCall(unsigned BuiltinID, CallExpr *TheCall) {
case X86::BI__builtin_ia32_palignr128: case X86::BI__builtin_ia32_palignr128:
case X86::BI__builtin_ia32_palignr256: case X86::BI__builtin_ia32_palignr256:
case X86::BI__builtin_ia32_palignr512: case X86::BI__builtin_ia32_palignr512:
case X86::BI__builtin_ia32_alignq512:
case X86::BI__builtin_ia32_alignd512:
case X86::BI__builtin_ia32_alignd128:
case X86::BI__builtin_ia32_alignd256:
case X86::BI__builtin_ia32_alignq128:
case X86::BI__builtin_ia32_alignq256:
case X86::BI__builtin_ia32_vcomisd: case X86::BI__builtin_ia32_vcomisd:
case X86::BI__builtin_ia32_vcomiss: case X86::BI__builtin_ia32_vcomiss:
case X86::BI__builtin_ia32_dbpsadbw128_mask: case X86::BI__builtin_ia32_dbpsadbw128_mask: