forked from OSchip/llvm-project
[X86] Remove mask argument from more builtins that are handled completely in CGBuiltin.cpp. Just wrap a select builtin around them in the header file instead.
llvm-svn: 333061
This commit is contained in:
parent
3c90fcebd4
commit
f2043b08b4
|
@ -945,16 +945,16 @@ TARGET_BUILTIN(__builtin_ia32_cvtudq2ps512_mask, "V16fV16iV16fUsIi", "nc", "avx5
|
|||
TARGET_BUILTIN(__builtin_ia32_cvtpd2ps512_mask, "V8fV8dV8fUcIi", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_vcvtps2ph512_mask, "V16sV16fIiV16sUs", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_vcvtph2ps512_mask, "V16fV16sV16fUsIi", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsd512_mask, "V16iV16iV16iUs", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsq512_mask, "V8LLiV8LLiV8LLiUc", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsd512_mask, "V16iV16iV16iV16iUs", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsq512_mask, "V8LLiV8LLiV8LLiV8LLiUc", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxud512_mask, "V16iV16iV16iV16iUs", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxuq512_mask, "V8LLiV8LLiV8LLiV8LLiUc", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsd512_mask, "V16iV16iV16iV16iUs", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsq512_mask, "V8LLiV8LLiV8LLiV8LLiUc", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminud512_mask, "V16iV16iV16iV16iUs", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminuq512_mask, "V8LLiV8LLiV8LLiV8LLiUc", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsd512, "V16iV16i", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsq512, "V8LLiV8LLi", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsd512, "V16iV16iV16i", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsq512, "V8LLiV8LLiV8LLi", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxud512, "V16iV16iV16i", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxuq512, "V8LLiV8LLiV8LLi", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsd512, "V16iV16iV16i", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsq512, "V8LLiV8LLiV8LLi", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminud512, "V16iV16iV16i", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminuq512, "V8LLiV8LLiV8LLi", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmuldq512, "V8LLiV16iV16i", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmuludq512, "V8LLiV16iV16i", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_loaddqusi512_mask, "V16iiC*V16iUs", "n", "avx512f")
|
||||
|
@ -1068,8 +1068,8 @@ TARGET_BUILTIN(__builtin_ia32_ucmpd512_mask, "UsV16iV16iIiUs", "nc", "avx512f")
|
|||
TARGET_BUILTIN(__builtin_ia32_ucmpq512_mask, "UcV8LLiV8LLiIiUc", "nc", "avx512f")
|
||||
TARGET_BUILTIN(__builtin_ia32_ucmpw512_mask, "UiV32sV32sIiUi", "nc", "avx512bw")
|
||||
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsb512_mask, "V64cV64cV64cULLi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsw512_mask, "V32sV32sV32sUi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsb512, "V64cV64c", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsw512, "V32sV32s", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_packssdw512, "V32sV16iV16i", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_packsswb512, "V64cV32sV32s", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_packusdw512, "V32sV16iV16i", "nc", "avx512bw")
|
||||
|
@ -1078,14 +1078,14 @@ TARGET_BUILTIN(__builtin_ia32_paddsb512_mask, "V64cV64cV64cV64cULLi", "nc", "avx
|
|||
TARGET_BUILTIN(__builtin_ia32_paddsw512_mask, "V32sV32sV32sV32sUi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_paddusb512_mask, "V64cV64cV64cV64cULLi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_paddusw512_mask, "V32sV32sV32sV32sUi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsb512_mask, "V64cV64cV64cV64cULLi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsw512_mask, "V32sV32sV32sV32sUi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxub512_mask, "V64cV64cV64cV64cULLi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxuw512_mask, "V32sV32sV32sV32sUi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsb512_mask, "V64cV64cV64cV64cULLi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsw512_mask, "V32sV32sV32sV32sUi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminub512_mask, "V64cV64cV64cV64cULLi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminuw512_mask, "V32sV32sV32sV32sUi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsb512, "V64cV64cV64c", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsw512, "V32sV32sV32s", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxub512, "V64cV64cV64c", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxuw512, "V32sV32sV32s", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsb512, "V64cV64cV64c", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsw512, "V32sV32sV32s", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminub512, "V64cV64cV64c", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminuw512, "V32sV32sV32s", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_pshufb512, "V64cV64cV64c", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_psubsb512_mask, "V64cV64cV64cV64cULLi", "nc", "avx512bw")
|
||||
TARGET_BUILTIN(__builtin_ia32_psubsw512_mask, "V32sV32sV32sV32sUi", "nc", "avx512bw")
|
||||
|
@ -1230,16 +1230,16 @@ TARGET_BUILTIN(__builtin_ia32_getexppd128_mask, "V2dV2dV2dUc", "nc", "avx512vl")
|
|||
TARGET_BUILTIN(__builtin_ia32_getexppd256_mask, "V4dV4dV4dUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_getexpps128_mask, "V4fV4fV4fUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_getexpps256_mask, "V8fV8fV8fUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsq128_mask, "V2LLiV2LLiV2LLiUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsq256_mask, "V4LLiV4LLiV4LLiUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsq128_mask, "V2LLiV2LLiV2LLiV2LLiUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsq256_mask, "V4LLiV4LLiV4LLiV4LLiUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxuq128_mask, "V2LLiV2LLiV2LLiV2LLiUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxuq256_mask, "V4LLiV4LLiV4LLiV4LLiUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsq128_mask, "V2LLiV2LLiV2LLiV2LLiUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsq256_mask, "V4LLiV4LLiV4LLiV4LLiUc", "n", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminuq128_mask, "V2LLiV2LLiV2LLiV2LLiUc", "n", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminuq256_mask, "V4LLiV4LLiV4LLiV4LLiUc", "n", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsq128, "V2LLiV2LLi", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pabsq256, "V4LLiV4LLi", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsq128, "V2LLiV2LLiV2LLi", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxsq256, "V4LLiV4LLiV4LLi", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxuq128, "V2LLiV2LLiV2LLi", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pmaxuq256, "V4LLiV4LLiV4LLi", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsq128, "V2LLiV2LLiV2LLi", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminsq256, "V4LLiV4LLiV4LLi", "n", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminuq128, "V2LLiV2LLiV2LLi", "n", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_pminuq256, "V4LLiV4LLiV4LLi", "n", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_rndscalepd_128_mask, "V2dV2dIiV2dUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_rndscalepd_256_mask, "V4dV4dIiV4dUc", "nc", "avx512vl")
|
||||
TARGET_BUILTIN(__builtin_ia32_rndscaleps_128_mask, "V4fV4fIiV4fUc", "nc", "avx512vl")
|
||||
|
|
|
@ -8399,9 +8399,7 @@ static Value *EmitX86Abs(CodeGenFunction &CGF, ArrayRef<Value *> Ops) {
|
|||
Value *Sub = CGF.Builder.CreateSub(Zero, Ops[0]);
|
||||
Value *Cmp = CGF.Builder.CreateICmp(ICmpInst::ICMP_SGT, Ops[0], Zero);
|
||||
Value *Res = CGF.Builder.CreateSelect(Cmp, Ops[0], Sub);
|
||||
if (Ops.size() == 1)
|
||||
return Res;
|
||||
return EmitX86Select(CGF, Ops[2], Res, Ops[1]);
|
||||
return Res;
|
||||
}
|
||||
|
||||
static Value *EmitX86MinMax(CodeGenFunction &CGF, ICmpInst::Predicate Pred,
|
||||
|
@ -8409,11 +8407,8 @@ static Value *EmitX86MinMax(CodeGenFunction &CGF, ICmpInst::Predicate Pred,
|
|||
Value *Cmp = CGF.Builder.CreateICmp(Pred, Ops[0], Ops[1]);
|
||||
Value *Res = CGF.Builder.CreateSelect(Cmp, Ops[0], Ops[1]);
|
||||
|
||||
if (Ops.size() == 2)
|
||||
return Res;
|
||||
|
||||
assert(Ops.size() == 4);
|
||||
return EmitX86Select(CGF, Ops[3], Res, Ops[2]);
|
||||
assert(Ops.size() == 2);
|
||||
return Res;
|
||||
}
|
||||
|
||||
static Value *EmitX86Muldq(CodeGenFunction &CGF, bool IsSigned,
|
||||
|
@ -9108,65 +9103,65 @@ Value *CodeGenFunction::EmitX86BuiltinExpr(unsigned BuiltinID,
|
|||
case X86::BI__builtin_ia32_pabsb256:
|
||||
case X86::BI__builtin_ia32_pabsw256:
|
||||
case X86::BI__builtin_ia32_pabsd256:
|
||||
case X86::BI__builtin_ia32_pabsq128_mask:
|
||||
case X86::BI__builtin_ia32_pabsq256_mask:
|
||||
case X86::BI__builtin_ia32_pabsb512_mask:
|
||||
case X86::BI__builtin_ia32_pabsw512_mask:
|
||||
case X86::BI__builtin_ia32_pabsd512_mask:
|
||||
case X86::BI__builtin_ia32_pabsq512_mask:
|
||||
case X86::BI__builtin_ia32_pabsq128:
|
||||
case X86::BI__builtin_ia32_pabsq256:
|
||||
case X86::BI__builtin_ia32_pabsb512:
|
||||
case X86::BI__builtin_ia32_pabsw512:
|
||||
case X86::BI__builtin_ia32_pabsd512:
|
||||
case X86::BI__builtin_ia32_pabsq512:
|
||||
return EmitX86Abs(*this, Ops);
|
||||
|
||||
case X86::BI__builtin_ia32_pmaxsb128:
|
||||
case X86::BI__builtin_ia32_pmaxsw128:
|
||||
case X86::BI__builtin_ia32_pmaxsd128:
|
||||
case X86::BI__builtin_ia32_pmaxsq128_mask:
|
||||
case X86::BI__builtin_ia32_pmaxsq128:
|
||||
case X86::BI__builtin_ia32_pmaxsb256:
|
||||
case X86::BI__builtin_ia32_pmaxsw256:
|
||||
case X86::BI__builtin_ia32_pmaxsd256:
|
||||
case X86::BI__builtin_ia32_pmaxsq256_mask:
|
||||
case X86::BI__builtin_ia32_pmaxsb512_mask:
|
||||
case X86::BI__builtin_ia32_pmaxsw512_mask:
|
||||
case X86::BI__builtin_ia32_pmaxsd512_mask:
|
||||
case X86::BI__builtin_ia32_pmaxsq512_mask:
|
||||
case X86::BI__builtin_ia32_pmaxsq256:
|
||||
case X86::BI__builtin_ia32_pmaxsb512:
|
||||
case X86::BI__builtin_ia32_pmaxsw512:
|
||||
case X86::BI__builtin_ia32_pmaxsd512:
|
||||
case X86::BI__builtin_ia32_pmaxsq512:
|
||||
return EmitX86MinMax(*this, ICmpInst::ICMP_SGT, Ops);
|
||||
case X86::BI__builtin_ia32_pmaxub128:
|
||||
case X86::BI__builtin_ia32_pmaxuw128:
|
||||
case X86::BI__builtin_ia32_pmaxud128:
|
||||
case X86::BI__builtin_ia32_pmaxuq128_mask:
|
||||
case X86::BI__builtin_ia32_pmaxuq128:
|
||||
case X86::BI__builtin_ia32_pmaxub256:
|
||||
case X86::BI__builtin_ia32_pmaxuw256:
|
||||
case X86::BI__builtin_ia32_pmaxud256:
|
||||
case X86::BI__builtin_ia32_pmaxuq256_mask:
|
||||
case X86::BI__builtin_ia32_pmaxub512_mask:
|
||||
case X86::BI__builtin_ia32_pmaxuw512_mask:
|
||||
case X86::BI__builtin_ia32_pmaxud512_mask:
|
||||
case X86::BI__builtin_ia32_pmaxuq512_mask:
|
||||
case X86::BI__builtin_ia32_pmaxuq256:
|
||||
case X86::BI__builtin_ia32_pmaxub512:
|
||||
case X86::BI__builtin_ia32_pmaxuw512:
|
||||
case X86::BI__builtin_ia32_pmaxud512:
|
||||
case X86::BI__builtin_ia32_pmaxuq512:
|
||||
return EmitX86MinMax(*this, ICmpInst::ICMP_UGT, Ops);
|
||||
case X86::BI__builtin_ia32_pminsb128:
|
||||
case X86::BI__builtin_ia32_pminsw128:
|
||||
case X86::BI__builtin_ia32_pminsd128:
|
||||
case X86::BI__builtin_ia32_pminsq128_mask:
|
||||
case X86::BI__builtin_ia32_pminsq128:
|
||||
case X86::BI__builtin_ia32_pminsb256:
|
||||
case X86::BI__builtin_ia32_pminsw256:
|
||||
case X86::BI__builtin_ia32_pminsd256:
|
||||
case X86::BI__builtin_ia32_pminsq256_mask:
|
||||
case X86::BI__builtin_ia32_pminsb512_mask:
|
||||
case X86::BI__builtin_ia32_pminsw512_mask:
|
||||
case X86::BI__builtin_ia32_pminsd512_mask:
|
||||
case X86::BI__builtin_ia32_pminsq512_mask:
|
||||
case X86::BI__builtin_ia32_pminsq256:
|
||||
case X86::BI__builtin_ia32_pminsb512:
|
||||
case X86::BI__builtin_ia32_pminsw512:
|
||||
case X86::BI__builtin_ia32_pminsd512:
|
||||
case X86::BI__builtin_ia32_pminsq512:
|
||||
return EmitX86MinMax(*this, ICmpInst::ICMP_SLT, Ops);
|
||||
case X86::BI__builtin_ia32_pminub128:
|
||||
case X86::BI__builtin_ia32_pminuw128:
|
||||
case X86::BI__builtin_ia32_pminud128:
|
||||
case X86::BI__builtin_ia32_pminuq128_mask:
|
||||
case X86::BI__builtin_ia32_pminuq128:
|
||||
case X86::BI__builtin_ia32_pminub256:
|
||||
case X86::BI__builtin_ia32_pminuw256:
|
||||
case X86::BI__builtin_ia32_pminud256:
|
||||
case X86::BI__builtin_ia32_pminuq256_mask:
|
||||
case X86::BI__builtin_ia32_pminub512_mask:
|
||||
case X86::BI__builtin_ia32_pminuw512_mask:
|
||||
case X86::BI__builtin_ia32_pminud512_mask:
|
||||
case X86::BI__builtin_ia32_pminuq512_mask:
|
||||
case X86::BI__builtin_ia32_pminuq256:
|
||||
case X86::BI__builtin_ia32_pminub512:
|
||||
case X86::BI__builtin_ia32_pminuw512:
|
||||
case X86::BI__builtin_ia32_pminud512:
|
||||
case X86::BI__builtin_ia32_pminuq512:
|
||||
return EmitX86MinMax(*this, ICmpInst::ICMP_ULT, Ops);
|
||||
|
||||
case X86::BI__builtin_ia32_pmuludq128:
|
||||
|
|
|
@ -310,49 +310,45 @@ _mm512_mask_blend_epi16 (__mmask32 __U, __m512i __A, __m512i __W)
|
|||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_abs_epi8 (__m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsb512_mask ((__v64qi) __A,
|
||||
(__v64qi) _mm512_setzero_qi(),
|
||||
(__mmask64) -1);
|
||||
return (__m512i)__builtin_ia32_pabsb512((__v64qi)__A);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_abs_epi8 (__m512i __W, __mmask64 __U, __m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsb512_mask ((__v64qi) __A,
|
||||
(__v64qi) __W,
|
||||
(__mmask64) __U);
|
||||
return (__m512i)__builtin_ia32_selectb_512((__mmask64)__U,
|
||||
(__v64qi)_mm512_abs_epi8(__A),
|
||||
(__v64qi)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_abs_epi8 (__mmask64 __U, __m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsb512_mask ((__v64qi) __A,
|
||||
(__v64qi) _mm512_setzero_qi(),
|
||||
(__mmask64) __U);
|
||||
return (__m512i)__builtin_ia32_selectb_512((__mmask64)__U,
|
||||
(__v64qi)_mm512_abs_epi8(__A),
|
||||
(__v64qi)_mm512_setzero_qi());
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_abs_epi16 (__m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsw512_mask ((__v32hi) __A,
|
||||
(__v32hi) _mm512_setzero_hi(),
|
||||
(__mmask32) -1);
|
||||
return (__m512i)__builtin_ia32_pabsw512((__v32hi)__A);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_abs_epi16 (__m512i __W, __mmask32 __U, __m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __W,
|
||||
(__mmask32) __U);
|
||||
return (__m512i)__builtin_ia32_selectw_512((__mmask32)__U,
|
||||
(__v32hi)_mm512_abs_epi16(__A),
|
||||
(__v32hi)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_abs_epi16 (__mmask32 __U, __m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsw512_mask ((__v32hi) __A,
|
||||
(__v32hi) _mm512_setzero_hi(),
|
||||
(__mmask32) __U);
|
||||
return (__m512i)__builtin_ia32_selectw_512((__mmask32)__U,
|
||||
(__v32hi)_mm512_abs_epi16(__A),
|
||||
(__v32hi)_mm512_setzero_hi());
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
|
@ -612,225 +608,178 @@ _mm512_maskz_avg_epu16 (__mmask32 __U, __m512i __A, __m512i __B)
|
|||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_max_epi8 (__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsb512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) _mm512_setzero_qi(),
|
||||
(__mmask64) -1);
|
||||
return (__m512i)__builtin_ia32_pmaxsb512((__v64qi) __A, (__v64qi) __B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_max_epi8 (__mmask64 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsb512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) _mm512_setzero_qi(),
|
||||
(__mmask64) __M);
|
||||
return (__m512i)__builtin_ia32_selectb_512((__mmask64)__M,
|
||||
(__v64qi)_mm512_max_epi8(__A, __B),
|
||||
(__v64qi)_mm512_setzero_qi());
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_max_epi8 (__m512i __W, __mmask64 __M, __m512i __A,
|
||||
__m512i __B)
|
||||
_mm512_mask_max_epi8 (__m512i __W, __mmask64 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsb512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) __W,
|
||||
(__mmask64) __M);
|
||||
return (__m512i)__builtin_ia32_selectb_512((__mmask64)__M,
|
||||
(__v64qi)_mm512_max_epi8(__A, __B),
|
||||
(__v64qi)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_max_epi16 (__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) _mm512_setzero_hi(),
|
||||
(__mmask32) -1);
|
||||
return (__m512i)__builtin_ia32_pmaxsw512((__v32hi) __A, (__v32hi) __B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_max_epi16 (__mmask32 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) _mm512_setzero_hi(),
|
||||
(__mmask32) __M);
|
||||
return (__m512i)__builtin_ia32_selectw_512((__mmask32)__M,
|
||||
(__v32hi)_mm512_max_epi16(__A, __B),
|
||||
(__v32hi)_mm512_setzero_hi());
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_max_epi16 (__m512i __W, __mmask32 __M, __m512i __A,
|
||||
__m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) __W,
|
||||
(__mmask32) __M);
|
||||
return (__m512i)__builtin_ia32_selectw_512((__mmask32)__M,
|
||||
(__v32hi)_mm512_max_epi16(__A, __B),
|
||||
(__v32hi)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_max_epu8 (__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxub512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) _mm512_setzero_qi(),
|
||||
(__mmask64) -1);
|
||||
return (__m512i)__builtin_ia32_pmaxub512((__v64qi)__A, (__v64qi)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_max_epu8 (__mmask64 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxub512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) _mm512_setzero_qi(),
|
||||
(__mmask64) __M);
|
||||
return (__m512i)__builtin_ia32_selectb_512((__mmask64)__M,
|
||||
(__v64qi)_mm512_max_epu8(__A, __B),
|
||||
(__v64qi)_mm512_setzero_qi());
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_max_epu8 (__m512i __W, __mmask64 __M, __m512i __A,
|
||||
__m512i __B)
|
||||
_mm512_mask_max_epu8 (__m512i __W, __mmask64 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxub512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) __W,
|
||||
(__mmask64) __M);
|
||||
return (__m512i)__builtin_ia32_selectb_512((__mmask64)__M,
|
||||
(__v64qi)_mm512_max_epu8(__A, __B),
|
||||
(__v64qi)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_max_epu16 (__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxuw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) _mm512_setzero_hi(),
|
||||
(__mmask32) -1);
|
||||
return (__m512i)__builtin_ia32_pmaxuw512((__v32hi)__A, (__v32hi)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_max_epu16 (__mmask32 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxuw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) _mm512_setzero_hi(),
|
||||
(__mmask32) __M);
|
||||
return (__m512i)__builtin_ia32_selectw_512((__mmask32)__M,
|
||||
(__v32hi)_mm512_max_epu16(__A, __B),
|
||||
(__v32hi)_mm512_setzero_hi());
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_max_epu16 (__m512i __W, __mmask32 __M, __m512i __A,
|
||||
__m512i __B)
|
||||
_mm512_mask_max_epu16 (__m512i __W, __mmask32 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxuw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) __W,
|
||||
(__mmask32) __M);
|
||||
return (__m512i)__builtin_ia32_selectw_512((__mmask32)__M,
|
||||
(__v32hi)_mm512_max_epu16(__A, __B),
|
||||
(__v32hi)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_min_epi8 (__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsb512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) _mm512_setzero_qi(),
|
||||
(__mmask64) -1);
|
||||
return (__m512i)__builtin_ia32_pminsb512((__v64qi) __A, (__v64qi) __B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_min_epi8 (__mmask64 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsb512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) _mm512_setzero_qi(),
|
||||
(__mmask64) __M);
|
||||
return (__m512i)__builtin_ia32_selectb_512((__mmask64)__M,
|
||||
(__v64qi)_mm512_min_epi8(__A, __B),
|
||||
(__v64qi)_mm512_setzero_qi());
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_min_epi8 (__m512i __W, __mmask64 __M, __m512i __A,
|
||||
__m512i __B)
|
||||
_mm512_mask_min_epi8 (__m512i __W, __mmask64 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsb512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) __W,
|
||||
(__mmask64) __M);
|
||||
return (__m512i)__builtin_ia32_selectb_512((__mmask64)__M,
|
||||
(__v64qi)_mm512_min_epi8(__A, __B),
|
||||
(__v64qi)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_min_epi16 (__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) _mm512_setzero_hi(),
|
||||
(__mmask32) -1);
|
||||
return (__m512i)__builtin_ia32_pminsw512((__v32hi) __A, (__v32hi) __B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_min_epi16 (__mmask32 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) _mm512_setzero_hi(),
|
||||
(__mmask32) __M);
|
||||
return (__m512i)__builtin_ia32_selectw_512((__mmask32)__M,
|
||||
(__v32hi)_mm512_min_epi16(__A, __B),
|
||||
(__v32hi)_mm512_setzero_hi());
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_min_epi16 (__m512i __W, __mmask32 __M, __m512i __A,
|
||||
__m512i __B)
|
||||
_mm512_mask_min_epi16 (__m512i __W, __mmask32 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) __W,
|
||||
(__mmask32) __M);
|
||||
return (__m512i)__builtin_ia32_selectw_512((__mmask32)__M,
|
||||
(__v32hi)_mm512_min_epi16(__A, __B),
|
||||
(__v32hi)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_min_epu8 (__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminub512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) _mm512_setzero_qi(),
|
||||
(__mmask64) -1);
|
||||
return (__m512i)__builtin_ia32_pminub512((__v64qi)__A, (__v64qi)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_min_epu8 (__mmask64 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminub512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) _mm512_setzero_qi(),
|
||||
(__mmask64) __M);
|
||||
return (__m512i)__builtin_ia32_selectb_512((__mmask64)__M,
|
||||
(__v64qi)_mm512_min_epu8(__A, __B),
|
||||
(__v64qi)_mm512_setzero_qi());
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_min_epu8 (__m512i __W, __mmask64 __M, __m512i __A,
|
||||
__m512i __B)
|
||||
_mm512_mask_min_epu8 (__m512i __W, __mmask64 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminub512_mask ((__v64qi) __A,
|
||||
(__v64qi) __B,
|
||||
(__v64qi) __W,
|
||||
(__mmask64) __M);
|
||||
return (__m512i)__builtin_ia32_selectb_512((__mmask64)__M,
|
||||
(__v64qi)_mm512_min_epu8(__A, __B),
|
||||
(__v64qi)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_min_epu16 (__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminuw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) _mm512_setzero_hi(),
|
||||
(__mmask32) -1);
|
||||
return (__m512i)__builtin_ia32_pminuw512((__v32hi)__A, (__v32hi)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_min_epu16 (__mmask32 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminuw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) _mm512_setzero_hi(),
|
||||
(__mmask32) __M);
|
||||
return (__m512i)__builtin_ia32_selectw_512((__mmask32)__M,
|
||||
(__v32hi)_mm512_min_epu16(__A, __B),
|
||||
(__v32hi)_mm512_setzero_hi());
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_min_epu16 (__m512i __W, __mmask32 __M, __m512i __A,
|
||||
__m512i __B)
|
||||
_mm512_mask_min_epu16 (__m512i __W, __mmask32 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminuw512_mask ((__v32hi) __A,
|
||||
(__v32hi) __B,
|
||||
(__v32hi) __W,
|
||||
(__mmask32) __M);
|
||||
return (__m512i)__builtin_ia32_selectw_512((__mmask32)__M,
|
||||
(__v32hi)_mm512_min_epu16(__A, __B),
|
||||
(__v32hi)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
|
|
|
@ -1121,113 +1121,89 @@ static __inline __m512i
|
|||
__DEFAULT_FN_ATTRS
|
||||
_mm512_max_epi32(__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsd512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask16) -1);
|
||||
return (__m512i)__builtin_ia32_pmaxsd512((__v16si)__A, (__v16si)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_max_epi32 (__m512i __W, __mmask16 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsd512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si) __W, __M);
|
||||
return (__m512i)__builtin_ia32_selectd_512((__mmask16)__M,
|
||||
(__v16si)_mm512_max_epi32(__A, __B),
|
||||
(__v16si)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_max_epi32 (__mmask16 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsd512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si)
|
||||
_mm512_setzero_si512 (),
|
||||
__M);
|
||||
return (__m512i)__builtin_ia32_selectd_512((__mmask16)__M,
|
||||
(__v16si)_mm512_max_epi32(__A, __B),
|
||||
(__v16si)_mm512_setzero_si512());
|
||||
}
|
||||
|
||||
static __inline __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_max_epu32(__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxud512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask16) -1);
|
||||
return (__m512i)__builtin_ia32_pmaxud512((__v16si)__A, (__v16si)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_max_epu32 (__m512i __W, __mmask16 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxud512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si) __W, __M);
|
||||
return (__m512i)__builtin_ia32_selectd_512((__mmask16)__M,
|
||||
(__v16si)_mm512_max_epu32(__A, __B),
|
||||
(__v16si)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_max_epu32 (__mmask16 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxud512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si)
|
||||
_mm512_setzero_si512 (),
|
||||
__M);
|
||||
return (__m512i)__builtin_ia32_selectd_512((__mmask16)__M,
|
||||
(__v16si)_mm512_max_epu32(__A, __B),
|
||||
(__v16si)_mm512_setzero_si512());
|
||||
}
|
||||
|
||||
static __inline __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_max_epi64(__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask8) -1);
|
||||
return (__m512i)__builtin_ia32_pmaxsq512((__v8di)__A, (__v8di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_max_epi64 (__m512i __W, __mmask8 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di) __W, __M);
|
||||
return (__m512i)__builtin_ia32_selectq_512((__mmask8)__M,
|
||||
(__v8di)_mm512_max_epi64(__A, __B),
|
||||
(__v8di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_max_epi64 (__mmask8 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxsq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di)
|
||||
_mm512_setzero_si512 (),
|
||||
__M);
|
||||
return (__m512i)__builtin_ia32_selectq_512((__mmask8)__M,
|
||||
(__v8di)_mm512_max_epi64(__A, __B),
|
||||
(__v8di)_mm512_setzero_si512());
|
||||
}
|
||||
|
||||
static __inline __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_max_epu64(__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxuq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask8) -1);
|
||||
return (__m512i)__builtin_ia32_pmaxuq512((__v8di)__A, (__v8di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_max_epu64 (__m512i __W, __mmask8 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxuq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di) __W, __M);
|
||||
return (__m512i)__builtin_ia32_selectq_512((__mmask8)__M,
|
||||
(__v8di)_mm512_max_epu64(__A, __B),
|
||||
(__v8di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_max_epu64 (__mmask8 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pmaxuq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di)
|
||||
_mm512_setzero_si512 (),
|
||||
__M);
|
||||
return (__m512i)__builtin_ia32_selectq_512((__mmask8)__M,
|
||||
(__v8di)_mm512_max_epu64(__A, __B),
|
||||
(__v8di)_mm512_setzero_si512());
|
||||
}
|
||||
|
||||
#define _mm512_mask_min_round_pd(W, U, A, B, R) __extension__ ({ \
|
||||
|
@ -1406,113 +1382,89 @@ static __inline __m512i
|
|||
__DEFAULT_FN_ATTRS
|
||||
_mm512_min_epi32(__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsd512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask16) -1);
|
||||
return (__m512i)__builtin_ia32_pminsd512((__v16si)__A, (__v16si)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_min_epi32 (__m512i __W, __mmask16 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsd512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si) __W, __M);
|
||||
return (__m512i)__builtin_ia32_selectd_512((__mmask16)__M,
|
||||
(__v16si)_mm512_min_epi32(__A, __B),
|
||||
(__v16si)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_min_epi32 (__mmask16 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsd512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si)
|
||||
_mm512_setzero_si512 (),
|
||||
__M);
|
||||
return (__m512i)__builtin_ia32_selectd_512((__mmask16)__M,
|
||||
(__v16si)_mm512_min_epi32(__A, __B),
|
||||
(__v16si)_mm512_setzero_si512());
|
||||
}
|
||||
|
||||
static __inline __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_min_epu32(__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminud512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask16) -1);
|
||||
return (__m512i)__builtin_ia32_pminud512((__v16si)__A, (__v16si)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_min_epu32 (__m512i __W, __mmask16 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminud512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si) __W, __M);
|
||||
return (__m512i)__builtin_ia32_selectd_512((__mmask16)__M,
|
||||
(__v16si)_mm512_min_epu32(__A, __B),
|
||||
(__v16si)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_min_epu32 (__mmask16 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminud512_mask ((__v16si) __A,
|
||||
(__v16si) __B,
|
||||
(__v16si)
|
||||
_mm512_setzero_si512 (),
|
||||
__M);
|
||||
return (__m512i)__builtin_ia32_selectd_512((__mmask16)__M,
|
||||
(__v16si)_mm512_min_epu32(__A, __B),
|
||||
(__v16si)_mm512_setzero_si512());
|
||||
}
|
||||
|
||||
static __inline __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_min_epi64(__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask8) -1);
|
||||
return (__m512i)__builtin_ia32_pminsq512((__v8di)__A, (__v8di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_min_epi64 (__m512i __W, __mmask8 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di) __W, __M);
|
||||
return (__m512i)__builtin_ia32_selectq_512((__mmask8)__M,
|
||||
(__v8di)_mm512_min_epi64(__A, __B),
|
||||
(__v8di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_min_epi64 (__mmask8 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminsq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di)
|
||||
_mm512_setzero_si512 (),
|
||||
__M);
|
||||
return (__m512i)__builtin_ia32_selectq_512((__mmask8)__M,
|
||||
(__v8di)_mm512_min_epi64(__A, __B),
|
||||
(__v8di)_mm512_setzero_si512());
|
||||
}
|
||||
|
||||
static __inline __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_min_epu64(__m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminuq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask8) -1);
|
||||
return (__m512i)__builtin_ia32_pminuq512((__v8di)__A, (__v8di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_min_epu64 (__m512i __W, __mmask8 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminuq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di) __W, __M);
|
||||
return (__m512i)__builtin_ia32_selectq_512((__mmask8)__M,
|
||||
(__v8di)_mm512_min_epu64(__A, __B),
|
||||
(__v8di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_min_epu64 (__mmask8 __M, __m512i __A, __m512i __B)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pminuq512_mask ((__v8di) __A,
|
||||
(__v8di) __B,
|
||||
(__v8di)
|
||||
_mm512_setzero_si512 (),
|
||||
__M);
|
||||
return (__m512i)__builtin_ia32_selectq_512((__mmask8)__M,
|
||||
(__v8di)_mm512_min_epu64(__A, __B),
|
||||
(__v8di)_mm512_setzero_si512());
|
||||
}
|
||||
|
||||
static __inline __m512i __DEFAULT_FN_ATTRS
|
||||
|
@ -1968,53 +1920,45 @@ _mm512_mask_ceil_pd (__m512d __W, __mmask8 __U, __m512d __A)
|
|||
static __inline __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_abs_epi64(__m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsq512_mask ((__v8di) __A,
|
||||
(__v8di)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask8) -1);
|
||||
return (__m512i)__builtin_ia32_pabsq512((__v8di)__A);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_abs_epi64 (__m512i __W, __mmask8 __U, __m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsq512_mask ((__v8di) __A,
|
||||
(__v8di) __W,
|
||||
(__mmask8) __U);
|
||||
return (__m512i)__builtin_ia32_selectq_512((__mmask8)__U,
|
||||
(__v8di)_mm512_abs_epi64(__A),
|
||||
(__v8di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_abs_epi64 (__mmask8 __U, __m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsq512_mask ((__v8di) __A,
|
||||
(__v8di)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask8) __U);
|
||||
return (__m512i)__builtin_ia32_selectq_512((__mmask8)__U,
|
||||
(__v8di)_mm512_abs_epi64(__A),
|
||||
(__v8di)_mm512_setzero_si512());
|
||||
}
|
||||
|
||||
static __inline __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_abs_epi32(__m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsd512_mask ((__v16si) __A,
|
||||
(__v16si)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask16) -1);
|
||||
return (__m512i)__builtin_ia32_pabsd512((__v16si) __A);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_mask_abs_epi32 (__m512i __W, __mmask16 __U, __m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsd512_mask ((__v16si) __A,
|
||||
(__v16si) __W,
|
||||
(__mmask16) __U);
|
||||
return (__m512i)__builtin_ia32_selectd_512((__mmask8)__U,
|
||||
(__v16si)_mm512_abs_epi32(__A),
|
||||
(__v16si)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m512i __DEFAULT_FN_ATTRS
|
||||
_mm512_maskz_abs_epi32 (__mmask16 __U, __m512i __A)
|
||||
{
|
||||
return (__m512i) __builtin_ia32_pabsd512_mask ((__v16si) __A,
|
||||
(__v16si)
|
||||
_mm512_setzero_si512 (),
|
||||
(__mmask16) __U);
|
||||
return (__m512i)__builtin_ia32_selectd_512((__mmask8)__U,
|
||||
(__v16si)_mm512_abs_epi32(__A),
|
||||
(__v16si)_mm512_setzero_si512());
|
||||
}
|
||||
|
||||
static __inline__ __m128 __DEFAULT_FN_ATTRS
|
||||
|
|
|
@ -2844,48 +2844,40 @@ _mm256_maskz_abs_epi32(__mmask8 __U, __m256i __A) {
|
|||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_abs_epi64 (__m128i __A) {
|
||||
return (__m128i) __builtin_ia32_pabsq128_mask ((__v2di) __A,
|
||||
(__v2di)
|
||||
_mm_setzero_si128 (),
|
||||
(__mmask8) -1);
|
||||
return (__m128i)__builtin_ia32_pabsq128((__v2di)__A);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_mask_abs_epi64 (__m128i __W, __mmask8 __U, __m128i __A) {
|
||||
return (__m128i) __builtin_ia32_pabsq128_mask ((__v2di) __A,
|
||||
(__v2di) __W,
|
||||
(__mmask8) __U);
|
||||
return (__m128i)__builtin_ia32_selectq_128((__mmask8)__U,
|
||||
(__v2di)_mm_abs_epi64(__A),
|
||||
(__v2di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_maskz_abs_epi64 (__mmask8 __U, __m128i __A) {
|
||||
return (__m128i) __builtin_ia32_pabsq128_mask ((__v2di) __A,
|
||||
(__v2di)
|
||||
_mm_setzero_si128 (),
|
||||
(__mmask8) __U);
|
||||
return (__m128i)__builtin_ia32_selectq_128((__mmask8)__U,
|
||||
(__v2di)_mm_abs_epi64(__A),
|
||||
(__v2di)_mm_setzero_si128());
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_abs_epi64 (__m256i __A) {
|
||||
return (__m256i) __builtin_ia32_pabsq256_mask ((__v4di) __A,
|
||||
(__v4di)
|
||||
_mm256_setzero_si256 (),
|
||||
(__mmask8) -1);
|
||||
return (__m256i)__builtin_ia32_pabsq256 ((__v4di)__A);
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_mask_abs_epi64 (__m256i __W, __mmask8 __U, __m256i __A) {
|
||||
return (__m256i) __builtin_ia32_pabsq256_mask ((__v4di) __A,
|
||||
(__v4di) __W,
|
||||
(__mmask8) __U);
|
||||
return (__m256i)__builtin_ia32_selectq_256((__mmask8)__U,
|
||||
(__v4di)_mm256_abs_epi64(__A),
|
||||
(__v4di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_maskz_abs_epi64 (__mmask8 __U, __m256i __A) {
|
||||
return (__m256i) __builtin_ia32_pabsq256_mask ((__v4di) __A,
|
||||
(__v4di)
|
||||
_mm256_setzero_si256 (),
|
||||
(__mmask8) __U);
|
||||
return (__m256i)__builtin_ia32_selectq_256((__mmask8)__U,
|
||||
(__v4di)_mm256_abs_epi64(__A),
|
||||
(__v4di)_mm256_setzero_si256());
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
|
@ -2916,56 +2908,42 @@ _mm256_mask_max_epi32(__m256i __W, __mmask8 __M, __m256i __A, __m256i __B) {
|
|||
(__v8si)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_maskz_max_epi64 (__mmask8 __M, __m128i __A, __m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pmaxsq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di)
|
||||
_mm_setzero_si128 (),
|
||||
__M);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_mask_max_epi64 (__m128i __W, __mmask8 __M, __m128i __A,
|
||||
__m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pmaxsq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di) __W, __M);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_max_epi64 (__m128i __A, __m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pmaxsq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di)
|
||||
_mm_setzero_si128 (),
|
||||
(__mmask8) -1);
|
||||
return (__m128i)__builtin_ia32_pmaxsq128((__v2di)__A, (__v2di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_maskz_max_epi64 (__mmask8 __M, __m256i __A, __m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pmaxsq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di)
|
||||
_mm256_setzero_si256 (),
|
||||
__M);
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_maskz_max_epi64 (__mmask8 __M, __m128i __A, __m128i __B) {
|
||||
return (__m128i)__builtin_ia32_selectq_128((__mmask8)__M,
|
||||
(__v2di)_mm_max_epi64(__A, __B),
|
||||
(__v2di)_mm_setzero_si128());
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_mask_max_epi64 (__m256i __W, __mmask8 __M, __m256i __A,
|
||||
__m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pmaxsq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di) __W, __M);
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_mask_max_epi64 (__m128i __W, __mmask8 __M, __m128i __A, __m128i __B) {
|
||||
return (__m128i)__builtin_ia32_selectq_128((__mmask8)__M,
|
||||
(__v2di)_mm_max_epi64(__A, __B),
|
||||
(__v2di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_max_epi64 (__m256i __A, __m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pmaxsq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di)
|
||||
_mm256_setzero_si256 (),
|
||||
(__mmask8) -1);
|
||||
return (__m256i)__builtin_ia32_pmaxsq256((__v4di)__A, (__v4di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_maskz_max_epi64 (__mmask8 __M, __m256i __A, __m256i __B) {
|
||||
return (__m256i)__builtin_ia32_selectq_256((__mmask8)__M,
|
||||
(__v4di)_mm256_max_epi64(__A, __B),
|
||||
(__v4di)_mm256_setzero_si256());
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_mask_max_epi64 (__m256i __W, __mmask8 __M, __m256i __A, __m256i __B) {
|
||||
return (__m256i)__builtin_ia32_selectq_256((__mmask8)__M,
|
||||
(__v4di)_mm256_max_epi64(__A, __B),
|
||||
(__v4di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
|
@ -2996,56 +2974,42 @@ _mm256_mask_max_epu32(__m256i __W, __mmask8 __M, __m256i __A, __m256i __B) {
|
|||
(__v8si)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_maskz_max_epu64 (__mmask8 __M, __m128i __A, __m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pmaxuq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di)
|
||||
_mm_setzero_si128 (),
|
||||
__M);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_max_epu64 (__m128i __A, __m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pmaxuq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di)
|
||||
_mm_setzero_si128 (),
|
||||
(__mmask8) -1);
|
||||
return (__m128i)__builtin_ia32_pmaxuq128((__v2di)__A, (__v2di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_mask_max_epu64 (__m128i __W, __mmask8 __M, __m128i __A,
|
||||
__m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pmaxuq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di) __W, __M);
|
||||
_mm_maskz_max_epu64 (__mmask8 __M, __m128i __A, __m128i __B) {
|
||||
return (__m128i)__builtin_ia32_selectq_128((__mmask8)__M,
|
||||
(__v2di)_mm_max_epu64(__A, __B),
|
||||
(__v2di)_mm_setzero_si128());
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_maskz_max_epu64 (__mmask8 __M, __m256i __A, __m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pmaxuq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di)
|
||||
_mm256_setzero_si256 (),
|
||||
__M);
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_mask_max_epu64 (__m128i __W, __mmask8 __M, __m128i __A, __m128i __B) {
|
||||
return (__m128i)__builtin_ia32_selectq_128((__mmask8)__M,
|
||||
(__v2di)_mm_max_epu64(__A, __B),
|
||||
(__v2di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_max_epu64 (__m256i __A, __m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pmaxuq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di)
|
||||
_mm256_setzero_si256 (),
|
||||
(__mmask8) -1);
|
||||
return (__m256i)__builtin_ia32_pmaxuq256((__v4di)__A, (__v4di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_mask_max_epu64 (__m256i __W, __mmask8 __M, __m256i __A,
|
||||
__m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pmaxuq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di) __W, __M);
|
||||
_mm256_maskz_max_epu64 (__mmask8 __M, __m256i __A, __m256i __B) {
|
||||
return (__m256i)__builtin_ia32_selectq_256((__mmask8)__M,
|
||||
(__v4di)_mm256_max_epu64(__A, __B),
|
||||
(__v4di)_mm256_setzero_si256());
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_mask_max_epu64 (__m256i __W, __mmask8 __M, __m256i __A, __m256i __B) {
|
||||
return (__m256i)__builtin_ia32_selectq_256((__mmask8)__M,
|
||||
(__v4di)_mm256_max_epu64(__A, __B),
|
||||
(__v4di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
|
@ -3078,54 +3042,40 @@ _mm256_mask_min_epi32(__m256i __W, __mmask8 __M, __m256i __A, __m256i __B) {
|
|||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_min_epi64 (__m128i __A, __m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pminsq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di)
|
||||
_mm_setzero_si128 (),
|
||||
(__mmask8) -1);
|
||||
return (__m128i)__builtin_ia32_pminsq128((__v2di)__A, (__v2di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_mask_min_epi64 (__m128i __W, __mmask8 __M, __m128i __A,
|
||||
__m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pminsq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di) __W, __M);
|
||||
_mm_mask_min_epi64 (__m128i __W, __mmask8 __M, __m128i __A, __m128i __B) {
|
||||
return (__m128i)__builtin_ia32_selectq_128((__mmask8)__M,
|
||||
(__v2di)_mm_min_epi64(__A, __B),
|
||||
(__v2di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_maskz_min_epi64 (__mmask8 __M, __m128i __A, __m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pminsq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di)
|
||||
_mm_setzero_si128 (),
|
||||
__M);
|
||||
return (__m128i)__builtin_ia32_selectq_128((__mmask8)__M,
|
||||
(__v2di)_mm_min_epi64(__A, __B),
|
||||
(__v2di)_mm_setzero_si128());
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_min_epi64 (__m256i __A, __m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pminsq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di)
|
||||
_mm256_setzero_si256 (),
|
||||
(__mmask8) -1);
|
||||
return (__m256i)__builtin_ia32_pminsq256((__v4di)__A, (__v4di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_mask_min_epi64 (__m256i __W, __mmask8 __M, __m256i __A,
|
||||
__m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pminsq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di) __W, __M);
|
||||
_mm256_mask_min_epi64 (__m256i __W, __mmask8 __M, __m256i __A, __m256i __B) {
|
||||
return (__m256i)__builtin_ia32_selectq_256((__mmask8)__M,
|
||||
(__v4di)_mm256_min_epi64(__A, __B),
|
||||
(__v4di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_maskz_min_epi64 (__mmask8 __M, __m256i __A, __m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pminsq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di)
|
||||
_mm256_setzero_si256 (),
|
||||
__M);
|
||||
return (__m256i)__builtin_ia32_selectq_256((__mmask8)__M,
|
||||
(__v4di)_mm256_min_epi64(__A, __B),
|
||||
(__v4di)_mm256_setzero_si256());
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
|
@ -3158,54 +3108,40 @@ _mm256_mask_min_epu32(__m256i __W, __mmask8 __M, __m256i __A, __m256i __B) {
|
|||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_min_epu64 (__m128i __A, __m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pminuq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di)
|
||||
_mm_setzero_si128 (),
|
||||
(__mmask8) -1);
|
||||
return (__m128i)__builtin_ia32_pminuq128((__v2di)__A, (__v2di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_mask_min_epu64 (__m128i __W, __mmask8 __M, __m128i __A,
|
||||
__m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pminuq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di) __W, __M);
|
||||
_mm_mask_min_epu64 (__m128i __W, __mmask8 __M, __m128i __A, __m128i __B) {
|
||||
return (__m128i)__builtin_ia32_selectq_128((__mmask8)__M,
|
||||
(__v2di)_mm_min_epu64(__A, __B),
|
||||
(__v2di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m128i __DEFAULT_FN_ATTRS
|
||||
_mm_maskz_min_epu64 (__mmask8 __M, __m128i __A, __m128i __B) {
|
||||
return (__m128i) __builtin_ia32_pminuq128_mask ((__v2di) __A,
|
||||
(__v2di) __B,
|
||||
(__v2di)
|
||||
_mm_setzero_si128 (),
|
||||
__M);
|
||||
return (__m128i)__builtin_ia32_selectq_128((__mmask8)__M,
|
||||
(__v2di)_mm_min_epu64(__A, __B),
|
||||
(__v2di)_mm_setzero_si128());
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_min_epu64 (__m256i __A, __m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pminuq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di)
|
||||
_mm256_setzero_si256 (),
|
||||
(__mmask8) -1);
|
||||
return (__m256i)__builtin_ia32_pminuq256((__v4di)__A, (__v4di)__B);
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_mask_min_epu64 (__m256i __W, __mmask8 __M, __m256i __A,
|
||||
__m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pminuq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di) __W, __M);
|
||||
_mm256_mask_min_epu64 (__m256i __W, __mmask8 __M, __m256i __A, __m256i __B) {
|
||||
return (__m256i)__builtin_ia32_selectq_256((__mmask8)__M,
|
||||
(__v4di)_mm256_min_epu64(__A, __B),
|
||||
(__v4di)__W);
|
||||
}
|
||||
|
||||
static __inline__ __m256i __DEFAULT_FN_ATTRS
|
||||
_mm256_maskz_min_epu64 (__mmask8 __M, __m256i __A, __m256i __B) {
|
||||
return (__m256i) __builtin_ia32_pminuq256_mask ((__v4di) __A,
|
||||
(__v4di) __B,
|
||||
(__v4di)
|
||||
_mm256_setzero_si256 (),
|
||||
__M);
|
||||
return (__m256i)__builtin_ia32_selectq_256((__mmask8)__M,
|
||||
(__v4di)_mm256_min_epu64(__A, __B),
|
||||
(__v4di)_mm256_setzero_si256());
|
||||
}
|
||||
|
||||
#define _mm_roundscale_pd(A, imm) __extension__ ({ \
|
||||
|
|
|
@ -3,13 +3,10 @@
|
|||
#include <immintrin.h>
|
||||
|
||||
// CHECK-LABEL: define i64 @test_mm512_reduce_max_epi64(<8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I8_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I9_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I10_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__V_ADDR_I:%.*]] = alloca <8 x i64>, align 64
|
||||
|
@ -27,8 +24,6 @@
|
|||
// CHECK: store <8 x i64> [[SHUFFLE1_I]], <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP5:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP6:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP7:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP8:%.*]] = icmp sgt <8 x i64> [[TMP5]], [[TMP6]]
|
||||
// CHECK: [[TMP9:%.*]] = select <8 x i1> [[TMP8]], <8 x i64> [[TMP5]], <8 x i64> [[TMP6]]
|
||||
// CHECK: store <8 x i64> [[TMP9]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -42,8 +37,6 @@
|
|||
// CHECK: store <8 x i64> [[SHUFFLE3_I]], <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP14:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I12_I]], align 64
|
||||
// CHECK: [[TMP15:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP16:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP17:%.*]] = icmp sgt <8 x i64> [[TMP14]], [[TMP15]]
|
||||
// CHECK: [[TMP18:%.*]] = select <8 x i1> [[TMP17]], <8 x i64> [[TMP14]], <8 x i64> [[TMP15]]
|
||||
// CHECK: store <8 x i64> [[TMP18]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -57,8 +50,6 @@
|
|||
// CHECK: store <8 x i64> [[SHUFFLE6_I]], <8 x i64>* [[__B_ADDR_I10_I]], align 64
|
||||
// CHECK: [[TMP23:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I9_I]], align 64
|
||||
// CHECK: [[TMP24:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I10_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I8_I]], align 64
|
||||
// CHECK: [[TMP25:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I8_I]], align 64
|
||||
// CHECK: [[TMP26:%.*]] = icmp sgt <8 x i64> [[TMP23]], [[TMP24]]
|
||||
// CHECK: [[TMP27:%.*]] = select <8 x i1> [[TMP26]], <8 x i64> [[TMP23]], <8 x i64> [[TMP24]]
|
||||
// CHECK: store <8 x i64> [[TMP27]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -70,13 +61,10 @@ long long test_mm512_reduce_max_epi64(__m512i __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i64 @test_mm512_reduce_max_epu64(<8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I8_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I9_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I10_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__V_ADDR_I:%.*]] = alloca <8 x i64>, align 64
|
||||
|
@ -94,8 +82,6 @@ long long test_mm512_reduce_max_epi64(__m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE1_I]], <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP5:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP6:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP7:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP8:%.*]] = icmp ugt <8 x i64> [[TMP5]], [[TMP6]]
|
||||
// CHECK: [[TMP9:%.*]] = select <8 x i1> [[TMP8]], <8 x i64> [[TMP5]], <8 x i64> [[TMP6]]
|
||||
// CHECK: store <8 x i64> [[TMP9]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -109,8 +95,6 @@ long long test_mm512_reduce_max_epi64(__m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE3_I]], <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP14:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I12_I]], align 64
|
||||
// CHECK: [[TMP15:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP16:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP17:%.*]] = icmp ugt <8 x i64> [[TMP14]], [[TMP15]]
|
||||
// CHECK: [[TMP18:%.*]] = select <8 x i1> [[TMP17]], <8 x i64> [[TMP14]], <8 x i64> [[TMP15]]
|
||||
// CHECK: store <8 x i64> [[TMP18]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -124,8 +108,6 @@ long long test_mm512_reduce_max_epi64(__m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE6_I]], <8 x i64>* [[__B_ADDR_I10_I]], align 64
|
||||
// CHECK: [[TMP23:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I9_I]], align 64
|
||||
// CHECK: [[TMP24:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I10_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I8_I]], align 64
|
||||
// CHECK: [[TMP25:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I8_I]], align 64
|
||||
// CHECK: [[TMP26:%.*]] = icmp ugt <8 x i64> [[TMP23]], [[TMP24]]
|
||||
// CHECK: [[TMP27:%.*]] = select <8 x i1> [[TMP26]], <8 x i64> [[TMP23]], <8 x i64> [[TMP24]]
|
||||
// CHECK: store <8 x i64> [[TMP27]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -201,13 +183,10 @@ double test_mm512_reduce_max_pd(__m512d __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i64 @test_mm512_reduce_min_epi64(<8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I8_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I9_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I10_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__V_ADDR_I:%.*]] = alloca <8 x i64>, align 64
|
||||
|
@ -225,8 +204,6 @@ double test_mm512_reduce_max_pd(__m512d __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE1_I]], <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP5:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP6:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP7:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP8:%.*]] = icmp slt <8 x i64> [[TMP5]], [[TMP6]]
|
||||
// CHECK: [[TMP9:%.*]] = select <8 x i1> [[TMP8]], <8 x i64> [[TMP5]], <8 x i64> [[TMP6]]
|
||||
// CHECK: store <8 x i64> [[TMP9]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -240,8 +217,6 @@ double test_mm512_reduce_max_pd(__m512d __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE3_I]], <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP14:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I12_I]], align 64
|
||||
// CHECK: [[TMP15:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP16:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP17:%.*]] = icmp slt <8 x i64> [[TMP14]], [[TMP15]]
|
||||
// CHECK: [[TMP18:%.*]] = select <8 x i1> [[TMP17]], <8 x i64> [[TMP14]], <8 x i64> [[TMP15]]
|
||||
// CHECK: store <8 x i64> [[TMP18]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -255,8 +230,6 @@ double test_mm512_reduce_max_pd(__m512d __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE6_I]], <8 x i64>* [[__B_ADDR_I10_I]], align 64
|
||||
// CHECK: [[TMP23:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I9_I]], align 64
|
||||
// CHECK: [[TMP24:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I10_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I8_I]], align 64
|
||||
// CHECK: [[TMP25:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I8_I]], align 64
|
||||
// CHECK: [[TMP26:%.*]] = icmp slt <8 x i64> [[TMP23]], [[TMP24]]
|
||||
// CHECK: [[TMP27:%.*]] = select <8 x i1> [[TMP26]], <8 x i64> [[TMP23]], <8 x i64> [[TMP24]]
|
||||
// CHECK: store <8 x i64> [[TMP27]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -268,13 +241,10 @@ long long test_mm512_reduce_min_epi64(__m512i __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i64 @test_mm512_reduce_min_epu64(<8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I8_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I9_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I10_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__V_ADDR_I:%.*]] = alloca <8 x i64>, align 64
|
||||
|
@ -292,8 +262,6 @@ long long test_mm512_reduce_min_epi64(__m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE1_I]], <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP5:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP6:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP7:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP8:%.*]] = icmp ult <8 x i64> [[TMP5]], [[TMP6]]
|
||||
// CHECK: [[TMP9:%.*]] = select <8 x i1> [[TMP8]], <8 x i64> [[TMP5]], <8 x i64> [[TMP6]]
|
||||
// CHECK: store <8 x i64> [[TMP9]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -307,8 +275,6 @@ long long test_mm512_reduce_min_epi64(__m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE3_I]], <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP14:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I12_I]], align 64
|
||||
// CHECK: [[TMP15:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP16:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP17:%.*]] = icmp ult <8 x i64> [[TMP14]], [[TMP15]]
|
||||
// CHECK: [[TMP18:%.*]] = select <8 x i1> [[TMP17]], <8 x i64> [[TMP14]], <8 x i64> [[TMP15]]
|
||||
// CHECK: store <8 x i64> [[TMP18]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -322,8 +288,6 @@ long long test_mm512_reduce_min_epi64(__m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE6_I]], <8 x i64>* [[__B_ADDR_I10_I]], align 64
|
||||
// CHECK: [[TMP23:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I9_I]], align 64
|
||||
// CHECK: [[TMP24:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I10_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I8_I]], align 64
|
||||
// CHECK: [[TMP25:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I8_I]], align 64
|
||||
// CHECK: [[TMP26:%.*]] = icmp ult <8 x i64> [[TMP23]], [[TMP24]]
|
||||
// CHECK: [[TMP27:%.*]] = select <8 x i1> [[TMP26]], <8 x i64> [[TMP23]], <8 x i64> [[TMP24]]
|
||||
// CHECK: store <8 x i64> [[TMP27]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -399,13 +363,10 @@ double test_mm512_reduce_min_pd(__m512d __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i64 @test_mm512_mask_reduce_max_epi64(i8 zeroext %__M, <8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I9_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I10_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__D_ADDR_I_I:%.*]] = alloca i64, align 8
|
||||
|
@ -454,8 +415,6 @@ double test_mm512_reduce_min_pd(__m512d __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE1_I]], <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: [[TMP19:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP20:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP21:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP22:%.*]] = icmp sgt <8 x i64> [[TMP19]], [[TMP20]]
|
||||
// CHECK: [[TMP23:%.*]] = select <8 x i1> [[TMP22]], <8 x i64> [[TMP19]], <8 x i64> [[TMP20]]
|
||||
// CHECK: store <8 x i64> [[TMP23]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -469,8 +428,6 @@ double test_mm512_reduce_min_pd(__m512d __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE4_I]], <8 x i64>* [[__B_ADDR_I11_I]], align 64
|
||||
// CHECK: [[TMP28:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I10_I]], align 64
|
||||
// CHECK: [[TMP29:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I11_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I9_I]], align 64
|
||||
// CHECK: [[TMP30:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I9_I]], align 64
|
||||
// CHECK: [[TMP31:%.*]] = icmp sgt <8 x i64> [[TMP28]], [[TMP29]]
|
||||
// CHECK: [[TMP32:%.*]] = select <8 x i1> [[TMP31]], <8 x i64> [[TMP28]], <8 x i64> [[TMP29]]
|
||||
// CHECK: store <8 x i64> [[TMP32]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -484,8 +441,6 @@ double test_mm512_reduce_min_pd(__m512d __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE7_I]], <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP37:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP38:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP39:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP40:%.*]] = icmp sgt <8 x i64> [[TMP37]], [[TMP38]]
|
||||
// CHECK: [[TMP41:%.*]] = select <8 x i1> [[TMP40]], <8 x i64> [[TMP37]], <8 x i64> [[TMP38]]
|
||||
// CHECK: store <8 x i64> [[TMP41]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -497,13 +452,10 @@ long long test_mm512_mask_reduce_max_epi64(__mmask8 __M, __m512i __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i64 @test_mm512_mask_reduce_max_epu64(i8 zeroext %__M, <8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I9_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I10_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__D_ADDR_I_I:%.*]] = alloca i64, align 8
|
||||
|
@ -552,8 +504,6 @@ long long test_mm512_mask_reduce_max_epi64(__mmask8 __M, __m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE1_I]], <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: [[TMP19:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP20:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP21:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP22:%.*]] = icmp ugt <8 x i64> [[TMP19]], [[TMP20]]
|
||||
// CHECK: [[TMP23:%.*]] = select <8 x i1> [[TMP22]], <8 x i64> [[TMP19]], <8 x i64> [[TMP20]]
|
||||
// CHECK: store <8 x i64> [[TMP23]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -567,8 +517,6 @@ long long test_mm512_mask_reduce_max_epi64(__mmask8 __M, __m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE4_I]], <8 x i64>* [[__B_ADDR_I11_I]], align 64
|
||||
// CHECK: [[TMP28:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I10_I]], align 64
|
||||
// CHECK: [[TMP29:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I11_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I9_I]], align 64
|
||||
// CHECK: [[TMP30:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I9_I]], align 64
|
||||
// CHECK: [[TMP31:%.*]] = icmp ugt <8 x i64> [[TMP28]], [[TMP29]]
|
||||
// CHECK: [[TMP32:%.*]] = select <8 x i1> [[TMP31]], <8 x i64> [[TMP28]], <8 x i64> [[TMP29]]
|
||||
// CHECK: store <8 x i64> [[TMP32]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -582,8 +530,6 @@ long long test_mm512_mask_reduce_max_epi64(__mmask8 __M, __m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE7_I]], <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP37:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP38:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP39:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP40:%.*]] = icmp ugt <8 x i64> [[TMP37]], [[TMP38]]
|
||||
// CHECK: [[TMP41:%.*]] = select <8 x i1> [[TMP40]], <8 x i64> [[TMP37]], <8 x i64> [[TMP38]]
|
||||
// CHECK: store <8 x i64> [[TMP41]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -692,13 +638,10 @@ long long test_mm512_mask_reduce_max_pd(__mmask8 __M, __m512d __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i64 @test_mm512_mask_reduce_min_epi64(i8 zeroext %__M, <8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I9_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I10_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__D_ADDR_I_I:%.*]] = alloca i64, align 8
|
||||
|
@ -747,8 +690,6 @@ long long test_mm512_mask_reduce_max_pd(__mmask8 __M, __m512d __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE1_I]], <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: [[TMP19:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP20:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP21:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP22:%.*]] = icmp slt <8 x i64> [[TMP19]], [[TMP20]]
|
||||
// CHECK: [[TMP23:%.*]] = select <8 x i1> [[TMP22]], <8 x i64> [[TMP19]], <8 x i64> [[TMP20]]
|
||||
// CHECK: store <8 x i64> [[TMP23]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -762,8 +703,6 @@ long long test_mm512_mask_reduce_max_pd(__mmask8 __M, __m512d __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE4_I]], <8 x i64>* [[__B_ADDR_I11_I]], align 64
|
||||
// CHECK: [[TMP28:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I10_I]], align 64
|
||||
// CHECK: [[TMP29:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I11_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I9_I]], align 64
|
||||
// CHECK: [[TMP30:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I9_I]], align 64
|
||||
// CHECK: [[TMP31:%.*]] = icmp slt <8 x i64> [[TMP28]], [[TMP29]]
|
||||
// CHECK: [[TMP32:%.*]] = select <8 x i1> [[TMP31]], <8 x i64> [[TMP28]], <8 x i64> [[TMP29]]
|
||||
// CHECK: store <8 x i64> [[TMP32]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -777,8 +716,6 @@ long long test_mm512_mask_reduce_max_pd(__mmask8 __M, __m512d __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE7_I]], <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP37:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP38:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP39:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP40:%.*]] = icmp slt <8 x i64> [[TMP37]], [[TMP38]]
|
||||
// CHECK: [[TMP41:%.*]] = select <8 x i1> [[TMP40]], <8 x i64> [[TMP37]], <8 x i64> [[TMP38]]
|
||||
// CHECK: store <8 x i64> [[TMP41]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -790,13 +727,10 @@ long long test_mm512_mask_reduce_min_epi64(__mmask8 __M, __m512i __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i64 @test_mm512_mask_reduce_min_epu64(i8 zeroext %__M, <8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I9_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I10_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__D_ADDR_I_I:%.*]] = alloca i64, align 8
|
||||
|
@ -845,8 +779,6 @@ long long test_mm512_mask_reduce_min_epi64(__mmask8 __M, __m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE1_I]], <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: [[TMP19:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP20:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP21:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP22:%.*]] = icmp ult <8 x i64> [[TMP19]], [[TMP20]]
|
||||
// CHECK: [[TMP23:%.*]] = select <8 x i1> [[TMP22]], <8 x i64> [[TMP19]], <8 x i64> [[TMP20]]
|
||||
// CHECK: store <8 x i64> [[TMP23]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -860,8 +792,6 @@ long long test_mm512_mask_reduce_min_epi64(__mmask8 __M, __m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE4_I]], <8 x i64>* [[__B_ADDR_I11_I]], align 64
|
||||
// CHECK: [[TMP28:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I10_I]], align 64
|
||||
// CHECK: [[TMP29:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I11_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I9_I]], align 64
|
||||
// CHECK: [[TMP30:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I9_I]], align 64
|
||||
// CHECK: [[TMP31:%.*]] = icmp ult <8 x i64> [[TMP28]], [[TMP29]]
|
||||
// CHECK: [[TMP32:%.*]] = select <8 x i1> [[TMP31]], <8 x i64> [[TMP28]], <8 x i64> [[TMP29]]
|
||||
// CHECK: store <8 x i64> [[TMP32]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -875,8 +805,6 @@ long long test_mm512_mask_reduce_min_epi64(__mmask8 __M, __m512i __W){
|
|||
// CHECK: store <8 x i64> [[SHUFFLE7_I]], <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP37:%.*]] = load <8 x i64>, <8 x i64>* [[__A_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP38:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP39:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP40:%.*]] = icmp ult <8 x i64> [[TMP37]], [[TMP38]]
|
||||
// CHECK: [[TMP41:%.*]] = select <8 x i1> [[TMP40]], <8 x i64> [[TMP37]], <8 x i64> [[TMP38]]
|
||||
// CHECK: store <8 x i64> [[TMP41]], <8 x i64>* [[__V_ADDR_I]], align 64
|
||||
|
@ -983,16 +911,12 @@ double test_mm512_mask_reduce_min_pd(__mmask8 __M, __m512d __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i32 @test_mm512_reduce_max_epi32(<8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I17_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I18_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I19_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I15_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I16_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[A_ADDR_I:%.*]] = alloca <8 x i64>, align 64
|
||||
|
@ -1018,9 +942,6 @@ double test_mm512_mask_reduce_min_pd(__mmask8 __M, __m512d __W){
|
|||
// CHECK: [[TMP12:%.*]] = bitcast <8 x i64> [[TMP11]] to <16 x i32>
|
||||
// CHECK: [[TMP13:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP14:%.*]] = bitcast <8 x i64> [[TMP13]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP15:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP16:%.*]] = bitcast <8 x i64> [[TMP15]] to <16 x i32>
|
||||
// CHECK: [[TMP17:%.*]] = icmp sgt <16 x i32> [[TMP12]], [[TMP14]]
|
||||
// CHECK: [[TMP18:%.*]] = select <16 x i1> [[TMP17]], <16 x i32> [[TMP12]], <16 x i32> [[TMP14]]
|
||||
// CHECK: [[TMP19:%.*]] = bitcast <16 x i32> [[TMP18]] to <8 x i64>
|
||||
|
@ -1043,9 +964,6 @@ double test_mm512_mask_reduce_min_pd(__mmask8 __M, __m512d __W){
|
|||
// CHECK: [[TMP31:%.*]] = bitcast <8 x i64> [[TMP30]] to <16 x i32>
|
||||
// CHECK: [[TMP32:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I19_I]], align 64
|
||||
// CHECK: [[TMP33:%.*]] = bitcast <8 x i64> [[TMP32]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I17_I]], align 64
|
||||
// CHECK: [[TMP34:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I17_I]], align 64
|
||||
// CHECK: [[TMP35:%.*]] = bitcast <8 x i64> [[TMP34]] to <16 x i32>
|
||||
// CHECK: [[TMP36:%.*]] = icmp sgt <16 x i32> [[TMP31]], [[TMP33]]
|
||||
// CHECK: [[TMP37:%.*]] = select <16 x i1> [[TMP36]], <16 x i32> [[TMP31]], <16 x i32> [[TMP33]]
|
||||
// CHECK: [[TMP38:%.*]] = bitcast <16 x i32> [[TMP37]] to <8 x i64>
|
||||
|
@ -1068,9 +986,6 @@ double test_mm512_mask_reduce_min_pd(__mmask8 __M, __m512d __W){
|
|||
// CHECK: [[TMP50:%.*]] = bitcast <8 x i64> [[TMP49]] to <16 x i32>
|
||||
// CHECK: [[TMP51:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I16_I]], align 64
|
||||
// CHECK: [[TMP52:%.*]] = bitcast <8 x i64> [[TMP51]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I14_I]], align 64
|
||||
// CHECK: [[TMP53:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I14_I]], align 64
|
||||
// CHECK: [[TMP54:%.*]] = bitcast <8 x i64> [[TMP53]] to <16 x i32>
|
||||
// CHECK: [[TMP55:%.*]] = icmp sgt <16 x i32> [[TMP50]], [[TMP52]]
|
||||
// CHECK: [[TMP56:%.*]] = select <16 x i1> [[TMP55]], <16 x i32> [[TMP50]], <16 x i32> [[TMP52]]
|
||||
// CHECK: [[TMP57:%.*]] = bitcast <16 x i32> [[TMP56]] to <8 x i64>
|
||||
|
@ -1093,9 +1008,6 @@ double test_mm512_mask_reduce_min_pd(__mmask8 __M, __m512d __W){
|
|||
// CHECK: [[TMP69:%.*]] = bitcast <8 x i64> [[TMP68]] to <16 x i32>
|
||||
// CHECK: [[TMP70:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP71:%.*]] = bitcast <8 x i64> [[TMP70]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP72:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP73:%.*]] = bitcast <8 x i64> [[TMP72]] to <16 x i32>
|
||||
// CHECK: [[TMP74:%.*]] = icmp sgt <16 x i32> [[TMP69]], [[TMP71]]
|
||||
// CHECK: [[TMP75:%.*]] = select <16 x i1> [[TMP74]], <16 x i32> [[TMP69]], <16 x i32> [[TMP71]]
|
||||
// CHECK: [[TMP76:%.*]] = bitcast <16 x i32> [[TMP75]] to <8 x i64>
|
||||
|
@ -1109,16 +1021,12 @@ int test_mm512_reduce_max_epi32(__m512i __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i32 @test_mm512_reduce_max_epu32(<8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I17_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I18_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I19_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I15_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I16_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[A_ADDR_I:%.*]] = alloca <8 x i64>, align 64
|
||||
|
@ -1144,9 +1052,6 @@ int test_mm512_reduce_max_epi32(__m512i __W){
|
|||
// CHECK: [[TMP12:%.*]] = bitcast <8 x i64> [[TMP11]] to <16 x i32>
|
||||
// CHECK: [[TMP13:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP14:%.*]] = bitcast <8 x i64> [[TMP13]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP15:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP16:%.*]] = bitcast <8 x i64> [[TMP15]] to <16 x i32>
|
||||
// CHECK: [[TMP17:%.*]] = icmp ugt <16 x i32> [[TMP12]], [[TMP14]]
|
||||
// CHECK: [[TMP18:%.*]] = select <16 x i1> [[TMP17]], <16 x i32> [[TMP12]], <16 x i32> [[TMP14]]
|
||||
// CHECK: [[TMP19:%.*]] = bitcast <16 x i32> [[TMP18]] to <8 x i64>
|
||||
|
@ -1169,9 +1074,6 @@ int test_mm512_reduce_max_epi32(__m512i __W){
|
|||
// CHECK: [[TMP31:%.*]] = bitcast <8 x i64> [[TMP30]] to <16 x i32>
|
||||
// CHECK: [[TMP32:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I19_I]], align 64
|
||||
// CHECK: [[TMP33:%.*]] = bitcast <8 x i64> [[TMP32]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I17_I]], align 64
|
||||
// CHECK: [[TMP34:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I17_I]], align 64
|
||||
// CHECK: [[TMP35:%.*]] = bitcast <8 x i64> [[TMP34]] to <16 x i32>
|
||||
// CHECK: [[TMP36:%.*]] = icmp ugt <16 x i32> [[TMP31]], [[TMP33]]
|
||||
// CHECK: [[TMP37:%.*]] = select <16 x i1> [[TMP36]], <16 x i32> [[TMP31]], <16 x i32> [[TMP33]]
|
||||
// CHECK: [[TMP38:%.*]] = bitcast <16 x i32> [[TMP37]] to <8 x i64>
|
||||
|
@ -1194,9 +1096,6 @@ int test_mm512_reduce_max_epi32(__m512i __W){
|
|||
// CHECK: [[TMP50:%.*]] = bitcast <8 x i64> [[TMP49]] to <16 x i32>
|
||||
// CHECK: [[TMP51:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I16_I]], align 64
|
||||
// CHECK: [[TMP52:%.*]] = bitcast <8 x i64> [[TMP51]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I14_I]], align 64
|
||||
// CHECK: [[TMP53:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I14_I]], align 64
|
||||
// CHECK: [[TMP54:%.*]] = bitcast <8 x i64> [[TMP53]] to <16 x i32>
|
||||
// CHECK: [[TMP55:%.*]] = icmp ugt <16 x i32> [[TMP50]], [[TMP52]]
|
||||
// CHECK: [[TMP56:%.*]] = select <16 x i1> [[TMP55]], <16 x i32> [[TMP50]], <16 x i32> [[TMP52]]
|
||||
// CHECK: [[TMP57:%.*]] = bitcast <16 x i32> [[TMP56]] to <8 x i64>
|
||||
|
@ -1219,9 +1118,6 @@ int test_mm512_reduce_max_epi32(__m512i __W){
|
|||
// CHECK: [[TMP69:%.*]] = bitcast <8 x i64> [[TMP68]] to <16 x i32>
|
||||
// CHECK: [[TMP70:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP71:%.*]] = bitcast <8 x i64> [[TMP70]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP72:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP73:%.*]] = bitcast <8 x i64> [[TMP72]] to <16 x i32>
|
||||
// CHECK: [[TMP74:%.*]] = icmp ugt <16 x i32> [[TMP69]], [[TMP71]]
|
||||
// CHECK: [[TMP75:%.*]] = select <16 x i1> [[TMP74]], <16 x i32> [[TMP69]], <16 x i32> [[TMP71]]
|
||||
// CHECK: [[TMP76:%.*]] = bitcast <16 x i32> [[TMP75]] to <8 x i64>
|
||||
|
@ -1316,16 +1212,12 @@ float test_mm512_reduce_max_ps(__m512 __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i32 @test_mm512_reduce_min_epi32(<8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I17_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I18_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I19_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I15_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I16_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[A_ADDR_I:%.*]] = alloca <8 x i64>, align 64
|
||||
|
@ -1351,9 +1243,6 @@ float test_mm512_reduce_max_ps(__m512 __W){
|
|||
// CHECK: [[TMP12:%.*]] = bitcast <8 x i64> [[TMP11]] to <16 x i32>
|
||||
// CHECK: [[TMP13:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP14:%.*]] = bitcast <8 x i64> [[TMP13]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP15:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP16:%.*]] = bitcast <8 x i64> [[TMP15]] to <16 x i32>
|
||||
// CHECK: [[TMP17:%.*]] = icmp slt <16 x i32> [[TMP12]], [[TMP14]]
|
||||
// CHECK: [[TMP18:%.*]] = select <16 x i1> [[TMP17]], <16 x i32> [[TMP12]], <16 x i32> [[TMP14]]
|
||||
// CHECK: [[TMP19:%.*]] = bitcast <16 x i32> [[TMP18]] to <8 x i64>
|
||||
|
@ -1376,9 +1265,6 @@ float test_mm512_reduce_max_ps(__m512 __W){
|
|||
// CHECK: [[TMP31:%.*]] = bitcast <8 x i64> [[TMP30]] to <16 x i32>
|
||||
// CHECK: [[TMP32:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I19_I]], align 64
|
||||
// CHECK: [[TMP33:%.*]] = bitcast <8 x i64> [[TMP32]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I17_I]], align 64
|
||||
// CHECK: [[TMP34:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I17_I]], align 64
|
||||
// CHECK: [[TMP35:%.*]] = bitcast <8 x i64> [[TMP34]] to <16 x i32>
|
||||
// CHECK: [[TMP36:%.*]] = icmp slt <16 x i32> [[TMP31]], [[TMP33]]
|
||||
// CHECK: [[TMP37:%.*]] = select <16 x i1> [[TMP36]], <16 x i32> [[TMP31]], <16 x i32> [[TMP33]]
|
||||
// CHECK: [[TMP38:%.*]] = bitcast <16 x i32> [[TMP37]] to <8 x i64>
|
||||
|
@ -1401,9 +1287,6 @@ float test_mm512_reduce_max_ps(__m512 __W){
|
|||
// CHECK: [[TMP50:%.*]] = bitcast <8 x i64> [[TMP49]] to <16 x i32>
|
||||
// CHECK: [[TMP51:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I16_I]], align 64
|
||||
// CHECK: [[TMP52:%.*]] = bitcast <8 x i64> [[TMP51]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I14_I]], align 64
|
||||
// CHECK: [[TMP53:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I14_I]], align 64
|
||||
// CHECK: [[TMP54:%.*]] = bitcast <8 x i64> [[TMP53]] to <16 x i32>
|
||||
// CHECK: [[TMP55:%.*]] = icmp slt <16 x i32> [[TMP50]], [[TMP52]]
|
||||
// CHECK: [[TMP56:%.*]] = select <16 x i1> [[TMP55]], <16 x i32> [[TMP50]], <16 x i32> [[TMP52]]
|
||||
// CHECK: [[TMP57:%.*]] = bitcast <16 x i32> [[TMP56]] to <8 x i64>
|
||||
|
@ -1426,9 +1309,6 @@ float test_mm512_reduce_max_ps(__m512 __W){
|
|||
// CHECK: [[TMP69:%.*]] = bitcast <8 x i64> [[TMP68]] to <16 x i32>
|
||||
// CHECK: [[TMP70:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP71:%.*]] = bitcast <8 x i64> [[TMP70]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP72:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP73:%.*]] = bitcast <8 x i64> [[TMP72]] to <16 x i32>
|
||||
// CHECK: [[TMP74:%.*]] = icmp slt <16 x i32> [[TMP69]], [[TMP71]]
|
||||
// CHECK: [[TMP75:%.*]] = select <16 x i1> [[TMP74]], <16 x i32> [[TMP69]], <16 x i32> [[TMP71]]
|
||||
// CHECK: [[TMP76:%.*]] = bitcast <16 x i32> [[TMP75]] to <8 x i64>
|
||||
|
@ -1442,16 +1322,12 @@ int test_mm512_reduce_min_epi32(__m512i __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i32 @test_mm512_reduce_min_epu32(<8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I17_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I18_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I19_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I15_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I16_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I11_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[A_ADDR_I:%.*]] = alloca <8 x i64>, align 64
|
||||
|
@ -1477,9 +1353,6 @@ int test_mm512_reduce_min_epi32(__m512i __W){
|
|||
// CHECK: [[TMP12:%.*]] = bitcast <8 x i64> [[TMP11]] to <16 x i32>
|
||||
// CHECK: [[TMP13:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP14:%.*]] = bitcast <8 x i64> [[TMP13]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP15:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP16:%.*]] = bitcast <8 x i64> [[TMP15]] to <16 x i32>
|
||||
// CHECK: [[TMP17:%.*]] = icmp ult <16 x i32> [[TMP12]], [[TMP14]]
|
||||
// CHECK: [[TMP18:%.*]] = select <16 x i1> [[TMP17]], <16 x i32> [[TMP12]], <16 x i32> [[TMP14]]
|
||||
// CHECK: [[TMP19:%.*]] = bitcast <16 x i32> [[TMP18]] to <8 x i64>
|
||||
|
@ -1502,9 +1375,6 @@ int test_mm512_reduce_min_epi32(__m512i __W){
|
|||
// CHECK: [[TMP31:%.*]] = bitcast <8 x i64> [[TMP30]] to <16 x i32>
|
||||
// CHECK: [[TMP32:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I19_I]], align 64
|
||||
// CHECK: [[TMP33:%.*]] = bitcast <8 x i64> [[TMP32]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I17_I]], align 64
|
||||
// CHECK: [[TMP34:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I17_I]], align 64
|
||||
// CHECK: [[TMP35:%.*]] = bitcast <8 x i64> [[TMP34]] to <16 x i32>
|
||||
// CHECK: [[TMP36:%.*]] = icmp ult <16 x i32> [[TMP31]], [[TMP33]]
|
||||
// CHECK: [[TMP37:%.*]] = select <16 x i1> [[TMP36]], <16 x i32> [[TMP31]], <16 x i32> [[TMP33]]
|
||||
// CHECK: [[TMP38:%.*]] = bitcast <16 x i32> [[TMP37]] to <8 x i64>
|
||||
|
@ -1527,9 +1397,6 @@ int test_mm512_reduce_min_epi32(__m512i __W){
|
|||
// CHECK: [[TMP50:%.*]] = bitcast <8 x i64> [[TMP49]] to <16 x i32>
|
||||
// CHECK: [[TMP51:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I16_I]], align 64
|
||||
// CHECK: [[TMP52:%.*]] = bitcast <8 x i64> [[TMP51]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I14_I]], align 64
|
||||
// CHECK: [[TMP53:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I14_I]], align 64
|
||||
// CHECK: [[TMP54:%.*]] = bitcast <8 x i64> [[TMP53]] to <16 x i32>
|
||||
// CHECK: [[TMP55:%.*]] = icmp ult <16 x i32> [[TMP50]], [[TMP52]]
|
||||
// CHECK: [[TMP56:%.*]] = select <16 x i1> [[TMP55]], <16 x i32> [[TMP50]], <16 x i32> [[TMP52]]
|
||||
// CHECK: [[TMP57:%.*]] = bitcast <16 x i32> [[TMP56]] to <8 x i64>
|
||||
|
@ -1552,9 +1419,6 @@ int test_mm512_reduce_min_epi32(__m512i __W){
|
|||
// CHECK: [[TMP69:%.*]] = bitcast <8 x i64> [[TMP68]] to <16 x i32>
|
||||
// CHECK: [[TMP70:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I13_I]], align 64
|
||||
// CHECK: [[TMP71:%.*]] = bitcast <8 x i64> [[TMP70]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP72:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I11_I]], align 64
|
||||
// CHECK: [[TMP73:%.*]] = bitcast <8 x i64> [[TMP72]] to <16 x i32>
|
||||
// CHECK: [[TMP74:%.*]] = icmp ult <16 x i32> [[TMP69]], [[TMP71]]
|
||||
// CHECK: [[TMP75:%.*]] = select <16 x i1> [[TMP74]], <16 x i32> [[TMP69]], <16 x i32> [[TMP71]]
|
||||
// CHECK: [[TMP76:%.*]] = bitcast <16 x i32> [[TMP75]] to <8 x i64>
|
||||
|
@ -1649,16 +1513,12 @@ float test_mm512_reduce_min_ps(__m512 __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i32 @test_mm512_mask_reduce_max_epi32(i16 zeroext %__M, <8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I18_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I19_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I20_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I15_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I16_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I17_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__S_ADDR_I_I:%.*]] = alloca i32, align 4
|
||||
|
@ -1734,9 +1594,6 @@ float test_mm512_reduce_min_ps(__m512 __W){
|
|||
// CHECK: [[TMP37:%.*]] = bitcast <8 x i64> [[TMP36]] to <16 x i32>
|
||||
// CHECK: [[TMP38:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I20_I]], align 64
|
||||
// CHECK: [[TMP39:%.*]] = bitcast <8 x i64> [[TMP38]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I18_I]], align 64
|
||||
// CHECK: [[TMP40:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I18_I]], align 64
|
||||
// CHECK: [[TMP41:%.*]] = bitcast <8 x i64> [[TMP40]] to <16 x i32>
|
||||
// CHECK: [[TMP42:%.*]] = icmp sgt <16 x i32> [[TMP37]], [[TMP39]]
|
||||
// CHECK: [[TMP43:%.*]] = select <16 x i1> [[TMP42]], <16 x i32> [[TMP37]], <16 x i32> [[TMP39]]
|
||||
// CHECK: [[TMP44:%.*]] = bitcast <16 x i32> [[TMP43]] to <8 x i64>
|
||||
|
@ -1759,9 +1616,6 @@ float test_mm512_reduce_min_ps(__m512 __W){
|
|||
// CHECK: [[TMP56:%.*]] = bitcast <8 x i64> [[TMP55]] to <16 x i32>
|
||||
// CHECK: [[TMP57:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I17_I]], align 64
|
||||
// CHECK: [[TMP58:%.*]] = bitcast <8 x i64> [[TMP57]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I15_I]], align 64
|
||||
// CHECK: [[TMP59:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I15_I]], align 64
|
||||
// CHECK: [[TMP60:%.*]] = bitcast <8 x i64> [[TMP59]] to <16 x i32>
|
||||
// CHECK: [[TMP61:%.*]] = icmp sgt <16 x i32> [[TMP56]], [[TMP58]]
|
||||
// CHECK: [[TMP62:%.*]] = select <16 x i1> [[TMP61]], <16 x i32> [[TMP56]], <16 x i32> [[TMP58]]
|
||||
// CHECK: [[TMP63:%.*]] = bitcast <16 x i32> [[TMP62]] to <8 x i64>
|
||||
|
@ -1784,9 +1638,6 @@ float test_mm512_reduce_min_ps(__m512 __W){
|
|||
// CHECK: [[TMP75:%.*]] = bitcast <8 x i64> [[TMP74]] to <16 x i32>
|
||||
// CHECK: [[TMP76:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: [[TMP77:%.*]] = bitcast <8 x i64> [[TMP76]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP78:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP79:%.*]] = bitcast <8 x i64> [[TMP78]] to <16 x i32>
|
||||
// CHECK: [[TMP80:%.*]] = icmp sgt <16 x i32> [[TMP75]], [[TMP77]]
|
||||
// CHECK: [[TMP81:%.*]] = select <16 x i1> [[TMP80]], <16 x i32> [[TMP75]], <16 x i32> [[TMP77]]
|
||||
// CHECK: [[TMP82:%.*]] = bitcast <16 x i32> [[TMP81]] to <8 x i64>
|
||||
|
@ -1809,9 +1660,6 @@ float test_mm512_reduce_min_ps(__m512 __W){
|
|||
// CHECK: [[TMP94:%.*]] = bitcast <8 x i64> [[TMP93]] to <16 x i32>
|
||||
// CHECK: [[TMP95:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP96:%.*]] = bitcast <8 x i64> [[TMP95]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP97:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP98:%.*]] = bitcast <8 x i64> [[TMP97]] to <16 x i32>
|
||||
// CHECK: [[TMP99:%.*]] = icmp sgt <16 x i32> [[TMP94]], [[TMP96]]
|
||||
// CHECK: [[TMP100:%.*]] = select <16 x i1> [[TMP99]], <16 x i32> [[TMP94]], <16 x i32> [[TMP96]]
|
||||
// CHECK: [[TMP101:%.*]] = bitcast <16 x i32> [[TMP100]] to <8 x i64>
|
||||
|
@ -1825,16 +1673,12 @@ int test_mm512_mask_reduce_max_epi32(__mmask16 __M, __m512i __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i32 @test_mm512_mask_reduce_max_epu32(i16 zeroext %__M, <8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I18_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I19_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I20_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I15_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I16_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I17_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__S_ADDR_I_I:%.*]] = alloca i32, align 4
|
||||
|
@ -1910,9 +1754,6 @@ int test_mm512_mask_reduce_max_epi32(__mmask16 __M, __m512i __W){
|
|||
// CHECK: [[TMP37:%.*]] = bitcast <8 x i64> [[TMP36]] to <16 x i32>
|
||||
// CHECK: [[TMP38:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I20_I]], align 64
|
||||
// CHECK: [[TMP39:%.*]] = bitcast <8 x i64> [[TMP38]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I18_I]], align 64
|
||||
// CHECK: [[TMP40:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I18_I]], align 64
|
||||
// CHECK: [[TMP41:%.*]] = bitcast <8 x i64> [[TMP40]] to <16 x i32>
|
||||
// CHECK: [[TMP42:%.*]] = icmp ugt <16 x i32> [[TMP37]], [[TMP39]]
|
||||
// CHECK: [[TMP43:%.*]] = select <16 x i1> [[TMP42]], <16 x i32> [[TMP37]], <16 x i32> [[TMP39]]
|
||||
// CHECK: [[TMP44:%.*]] = bitcast <16 x i32> [[TMP43]] to <8 x i64>
|
||||
|
@ -1935,9 +1776,6 @@ int test_mm512_mask_reduce_max_epi32(__mmask16 __M, __m512i __W){
|
|||
// CHECK: [[TMP56:%.*]] = bitcast <8 x i64> [[TMP55]] to <16 x i32>
|
||||
// CHECK: [[TMP57:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I17_I]], align 64
|
||||
// CHECK: [[TMP58:%.*]] = bitcast <8 x i64> [[TMP57]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I15_I]], align 64
|
||||
// CHECK: [[TMP59:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I15_I]], align 64
|
||||
// CHECK: [[TMP60:%.*]] = bitcast <8 x i64> [[TMP59]] to <16 x i32>
|
||||
// CHECK: [[TMP61:%.*]] = icmp ugt <16 x i32> [[TMP56]], [[TMP58]]
|
||||
// CHECK: [[TMP62:%.*]] = select <16 x i1> [[TMP61]], <16 x i32> [[TMP56]], <16 x i32> [[TMP58]]
|
||||
// CHECK: [[TMP63:%.*]] = bitcast <16 x i32> [[TMP62]] to <8 x i64>
|
||||
|
@ -1960,9 +1798,6 @@ int test_mm512_mask_reduce_max_epi32(__mmask16 __M, __m512i __W){
|
|||
// CHECK: [[TMP75:%.*]] = bitcast <8 x i64> [[TMP74]] to <16 x i32>
|
||||
// CHECK: [[TMP76:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: [[TMP77:%.*]] = bitcast <8 x i64> [[TMP76]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP78:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP79:%.*]] = bitcast <8 x i64> [[TMP78]] to <16 x i32>
|
||||
// CHECK: [[TMP80:%.*]] = icmp ugt <16 x i32> [[TMP75]], [[TMP77]]
|
||||
// CHECK: [[TMP81:%.*]] = select <16 x i1> [[TMP80]], <16 x i32> [[TMP75]], <16 x i32> [[TMP77]]
|
||||
// CHECK: [[TMP82:%.*]] = bitcast <16 x i32> [[TMP81]] to <8 x i64>
|
||||
|
@ -1985,9 +1820,6 @@ int test_mm512_mask_reduce_max_epi32(__mmask16 __M, __m512i __W){
|
|||
// CHECK: [[TMP94:%.*]] = bitcast <8 x i64> [[TMP93]] to <16 x i32>
|
||||
// CHECK: [[TMP95:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP96:%.*]] = bitcast <8 x i64> [[TMP95]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP97:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP98:%.*]] = bitcast <8 x i64> [[TMP97]] to <16 x i32>
|
||||
// CHECK: [[TMP99:%.*]] = icmp ugt <16 x i32> [[TMP94]], [[TMP96]]
|
||||
// CHECK: [[TMP100:%.*]] = select <16 x i1> [[TMP99]], <16 x i32> [[TMP94]], <16 x i32> [[TMP96]]
|
||||
// CHECK: [[TMP101:%.*]] = bitcast <16 x i32> [[TMP100]] to <8 x i64>
|
||||
|
@ -2130,16 +1962,12 @@ float test_mm512_mask_reduce_max_ps(__mmask16 __M, __m512 __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i32 @test_mm512_mask_reduce_min_epi32(i16 zeroext %__M, <8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I18_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I19_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I20_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I15_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I16_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I17_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__S_ADDR_I_I:%.*]] = alloca i32, align 4
|
||||
|
@ -2215,9 +2043,6 @@ float test_mm512_mask_reduce_max_ps(__mmask16 __M, __m512 __W){
|
|||
// CHECK: [[TMP37:%.*]] = bitcast <8 x i64> [[TMP36]] to <16 x i32>
|
||||
// CHECK: [[TMP38:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I20_I]], align 64
|
||||
// CHECK: [[TMP39:%.*]] = bitcast <8 x i64> [[TMP38]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I18_I]], align 64
|
||||
// CHECK: [[TMP40:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I18_I]], align 64
|
||||
// CHECK: [[TMP41:%.*]] = bitcast <8 x i64> [[TMP40]] to <16 x i32>
|
||||
// CHECK: [[TMP42:%.*]] = icmp slt <16 x i32> [[TMP37]], [[TMP39]]
|
||||
// CHECK: [[TMP43:%.*]] = select <16 x i1> [[TMP42]], <16 x i32> [[TMP37]], <16 x i32> [[TMP39]]
|
||||
// CHECK: [[TMP44:%.*]] = bitcast <16 x i32> [[TMP43]] to <8 x i64>
|
||||
|
@ -2240,9 +2065,6 @@ float test_mm512_mask_reduce_max_ps(__mmask16 __M, __m512 __W){
|
|||
// CHECK: [[TMP56:%.*]] = bitcast <8 x i64> [[TMP55]] to <16 x i32>
|
||||
// CHECK: [[TMP57:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I17_I]], align 64
|
||||
// CHECK: [[TMP58:%.*]] = bitcast <8 x i64> [[TMP57]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I15_I]], align 64
|
||||
// CHECK: [[TMP59:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I15_I]], align 64
|
||||
// CHECK: [[TMP60:%.*]] = bitcast <8 x i64> [[TMP59]] to <16 x i32>
|
||||
// CHECK: [[TMP61:%.*]] = icmp slt <16 x i32> [[TMP56]], [[TMP58]]
|
||||
// CHECK: [[TMP62:%.*]] = select <16 x i1> [[TMP61]], <16 x i32> [[TMP56]], <16 x i32> [[TMP58]]
|
||||
// CHECK: [[TMP63:%.*]] = bitcast <16 x i32> [[TMP62]] to <8 x i64>
|
||||
|
@ -2265,9 +2087,6 @@ float test_mm512_mask_reduce_max_ps(__mmask16 __M, __m512 __W){
|
|||
// CHECK: [[TMP75:%.*]] = bitcast <8 x i64> [[TMP74]] to <16 x i32>
|
||||
// CHECK: [[TMP76:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: [[TMP77:%.*]] = bitcast <8 x i64> [[TMP76]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP78:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP79:%.*]] = bitcast <8 x i64> [[TMP78]] to <16 x i32>
|
||||
// CHECK: [[TMP80:%.*]] = icmp slt <16 x i32> [[TMP75]], [[TMP77]]
|
||||
// CHECK: [[TMP81:%.*]] = select <16 x i1> [[TMP80]], <16 x i32> [[TMP75]], <16 x i32> [[TMP77]]
|
||||
// CHECK: [[TMP82:%.*]] = bitcast <16 x i32> [[TMP81]] to <8 x i64>
|
||||
|
@ -2290,9 +2109,6 @@ float test_mm512_mask_reduce_max_ps(__mmask16 __M, __m512 __W){
|
|||
// CHECK: [[TMP94:%.*]] = bitcast <8 x i64> [[TMP93]] to <16 x i32>
|
||||
// CHECK: [[TMP95:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP96:%.*]] = bitcast <8 x i64> [[TMP95]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP97:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP98:%.*]] = bitcast <8 x i64> [[TMP97]] to <16 x i32>
|
||||
// CHECK: [[TMP99:%.*]] = icmp slt <16 x i32> [[TMP94]], [[TMP96]]
|
||||
// CHECK: [[TMP100:%.*]] = select <16 x i1> [[TMP99]], <16 x i32> [[TMP94]], <16 x i32> [[TMP96]]
|
||||
// CHECK: [[TMP101:%.*]] = bitcast <16 x i32> [[TMP100]] to <8 x i64>
|
||||
|
@ -2306,16 +2122,12 @@ int test_mm512_mask_reduce_min_epi32(__mmask16 __M, __m512i __W){
|
|||
}
|
||||
|
||||
// CHECK-LABEL: define i32 @test_mm512_mask_reduce_min_epu32(i16 zeroext %__M, <8 x i64> %__W) #0 {
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I18_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I19_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I20_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I15_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I16_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I17_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I12_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I13_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I14_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[_COMPOUNDLITERAL_I_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__A_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__B_ADDR_I_I:%.*]] = alloca <8 x i64>, align 64
|
||||
// CHECK: [[__S_ADDR_I_I:%.*]] = alloca i32, align 4
|
||||
|
@ -2391,9 +2203,6 @@ int test_mm512_mask_reduce_min_epi32(__mmask16 __M, __m512i __W){
|
|||
// CHECK: [[TMP37:%.*]] = bitcast <8 x i64> [[TMP36]] to <16 x i32>
|
||||
// CHECK: [[TMP38:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I20_I]], align 64
|
||||
// CHECK: [[TMP39:%.*]] = bitcast <8 x i64> [[TMP38]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I18_I]], align 64
|
||||
// CHECK: [[TMP40:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I18_I]], align 64
|
||||
// CHECK: [[TMP41:%.*]] = bitcast <8 x i64> [[TMP40]] to <16 x i32>
|
||||
// CHECK: [[TMP42:%.*]] = icmp ult <16 x i32> [[TMP37]], [[TMP39]]
|
||||
// CHECK: [[TMP43:%.*]] = select <16 x i1> [[TMP42]], <16 x i32> [[TMP37]], <16 x i32> [[TMP39]]
|
||||
// CHECK: [[TMP44:%.*]] = bitcast <16 x i32> [[TMP43]] to <8 x i64>
|
||||
|
@ -2416,9 +2225,6 @@ int test_mm512_mask_reduce_min_epi32(__mmask16 __M, __m512i __W){
|
|||
// CHECK: [[TMP56:%.*]] = bitcast <8 x i64> [[TMP55]] to <16 x i32>
|
||||
// CHECK: [[TMP57:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I17_I]], align 64
|
||||
// CHECK: [[TMP58:%.*]] = bitcast <8 x i64> [[TMP57]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I15_I]], align 64
|
||||
// CHECK: [[TMP59:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I15_I]], align 64
|
||||
// CHECK: [[TMP60:%.*]] = bitcast <8 x i64> [[TMP59]] to <16 x i32>
|
||||
// CHECK: [[TMP61:%.*]] = icmp ult <16 x i32> [[TMP56]], [[TMP58]]
|
||||
// CHECK: [[TMP62:%.*]] = select <16 x i1> [[TMP61]], <16 x i32> [[TMP56]], <16 x i32> [[TMP58]]
|
||||
// CHECK: [[TMP63:%.*]] = bitcast <16 x i32> [[TMP62]] to <8 x i64>
|
||||
|
@ -2441,9 +2247,6 @@ int test_mm512_mask_reduce_min_epi32(__mmask16 __M, __m512i __W){
|
|||
// CHECK: [[TMP75:%.*]] = bitcast <8 x i64> [[TMP74]] to <16 x i32>
|
||||
// CHECK: [[TMP76:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I14_I]], align 64
|
||||
// CHECK: [[TMP77:%.*]] = bitcast <8 x i64> [[TMP76]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP78:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I12_I]], align 64
|
||||
// CHECK: [[TMP79:%.*]] = bitcast <8 x i64> [[TMP78]] to <16 x i32>
|
||||
// CHECK: [[TMP80:%.*]] = icmp ult <16 x i32> [[TMP75]], [[TMP77]]
|
||||
// CHECK: [[TMP81:%.*]] = select <16 x i1> [[TMP80]], <16 x i32> [[TMP75]], <16 x i32> [[TMP77]]
|
||||
// CHECK: [[TMP82:%.*]] = bitcast <16 x i32> [[TMP81]] to <8 x i64>
|
||||
|
@ -2466,9 +2269,6 @@ int test_mm512_mask_reduce_min_epi32(__mmask16 __M, __m512i __W){
|
|||
// CHECK: [[TMP94:%.*]] = bitcast <8 x i64> [[TMP93]] to <16 x i32>
|
||||
// CHECK: [[TMP95:%.*]] = load <8 x i64>, <8 x i64>* [[__B_ADDR_I_I]], align 64
|
||||
// CHECK: [[TMP96:%.*]] = bitcast <8 x i64> [[TMP95]] to <16 x i32>
|
||||
// CHECK: store <8 x i64> zeroinitializer, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP97:%.*]] = load <8 x i64>, <8 x i64>* [[_COMPOUNDLITERAL_I_I_I]], align 64
|
||||
// CHECK: [[TMP98:%.*]] = bitcast <8 x i64> [[TMP97]] to <16 x i32>
|
||||
// CHECK: [[TMP99:%.*]] = icmp ult <16 x i32> [[TMP94]], [[TMP96]]
|
||||
// CHECK: [[TMP100:%.*]] = select <16 x i1> [[TMP99]], <16 x i32> [[TMP94]], <16 x i32> [[TMP96]]
|
||||
// CHECK: [[TMP101:%.*]] = bitcast <16 x i32> [[TMP100]] to <8 x i64>
|
||||
|
|
Loading…
Reference in New Issue