forked from OSchip/llvm-project
[DAGCombiner] Teach createBuildVecShuffle to handle cases where input vectors are less than half of the output vector size.
This will be needed by a future commit to support sign/zero extending from v8i8 to v8i64 which requires a sign/zero_extend_vector_inreg to be created which requires v8i8 to be concatenated upto v64i8 and goes through this code. llvm-svn: 284204
This commit is contained in:
parent
2bd52b5d91
commit
40feb7f157
|
@ -12983,11 +12983,15 @@ SDValue DAGCombiner::createBuildVecShuffle(SDLoc DL, SDNode *N,
|
|||
// We can't generate a shuffle node with mismatched input and output types.
|
||||
// Try to make the types match the type of the output.
|
||||
if (InVT1 != VT || InVT2 != VT) {
|
||||
if (InVT1.getSizeInBits() * 2 == VT.getSizeInBits() && InVT1 == InVT2) {
|
||||
// If both input vectors are exactly half the size of the output, concat
|
||||
// them. If we have only one (non-zero) input, concat it with undef.
|
||||
VecIn1 = DAG.getNode(ISD::CONCAT_VECTORS, DL, VT, VecIn1,
|
||||
VecIn2.getNode() ? VecIn2 : DAG.getUNDEF(InVT1));
|
||||
if ((VT.getSizeInBits() % InVT1.getSizeInBits() == 0) && InVT1 == InVT2) {
|
||||
// If the output vector length is a multiple of both input lengths,
|
||||
// we can concatenate them and pad the rest with undefs.
|
||||
unsigned NumConcats = VT.getSizeInBits() / InVT1.getSizeInBits();
|
||||
assert(NumConcats >= 2 && "Concat needs at least two inputs!");
|
||||
SmallVector<SDValue, 2> ConcatOps(NumConcats, DAG.getUNDEF(InVT1));
|
||||
ConcatOps[0] = VecIn1;
|
||||
ConcatOps[1] = VecIn2 ? VecIn2 : DAG.getUNDEF(InVT1);
|
||||
VecIn1 = DAG.getNode(ISD::CONCAT_VECTORS, DL, VT, ConcatOps);
|
||||
VecIn2 = SDValue();
|
||||
} else if (InVT1.getSizeInBits() == VT.getSizeInBits() * 2) {
|
||||
if (!TLI.isExtractSubvectorCheap(VT, NumElems))
|
||||
|
|
|
@ -14,11 +14,11 @@ define <16 x i32> @test2(<16 x i32> %x) {
|
|||
define <16 x float> @test3(<4 x float> %a) {
|
||||
; CHECK-LABEL: test3:
|
||||
; CHECK: ## BB#0:
|
||||
; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
||||
; CHECK-NEXT: vpshufb {{.*#+}} xmm0 = zero,zero,zero,zero,xmm0[8,9,10,11,0,1,2,3],zero,zero,zero,zero
|
||||
; CHECK-NEXT: vinserti128 $1, %xmm0, %ymm1, %ymm0
|
||||
; CHECK-NEXT: vpxor %ymm1, %ymm1, %ymm1
|
||||
; CHECK-NEXT: vinsertf64x4 $1, %ymm1, %zmm0, %zmm0
|
||||
; CHECK-NEXT: ## kill: %XMM0<def> %XMM0<kill> %ZMM0<def>
|
||||
; CHECK-NEXT: vmovdqa32 {{.*#+}} zmm2 = [0,1,2,3,4,18,16,7,8,9,10,11,12,13,14,15]
|
||||
; CHECK-NEXT: vpxord %zmm1, %zmm1, %zmm1
|
||||
; CHECK-NEXT: vpermt2ps %zmm0, %zmm2, %zmm1
|
||||
; CHECK-NEXT: vmovdqa64 %zmm1, %zmm0
|
||||
; CHECK-NEXT: retq
|
||||
%b = extractelement <4 x float> %a, i32 2
|
||||
%c = insertelement <16 x float> <float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float undef, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00>, float %b, i32 5
|
||||
|
|
|
@ -117,7 +117,7 @@ define void @legal_vzmovl_2i32_8i32(<2 x i32>* %in, <8 x i32>* %out) {
|
|||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
||||
; X32-NEXT: vpmovzxdq {{.*#+}} xmm0 = mem[0],zero,mem[1],zero
|
||||
; X32-NEXT: vmovsd {{.*#+}} xmm0 = mem[0],zero
|
||||
; X32-NEXT: vxorps %ymm1, %ymm1, %ymm1
|
||||
; X32-NEXT: vblendps {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3,4,5,6,7]
|
||||
; X32-NEXT: vmovaps %ymm0, (%eax)
|
||||
|
@ -126,7 +126,7 @@ define void @legal_vzmovl_2i32_8i32(<2 x i32>* %in, <8 x i32>* %out) {
|
|||
;
|
||||
; X64-LABEL: legal_vzmovl_2i32_8i32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: vpmovzxdq {{.*#+}} xmm0 = mem[0],zero,mem[1],zero
|
||||
; X64-NEXT: vmovq {{.*#+}} xmm0 = mem[0],zero
|
||||
; X64-NEXT: vxorps %ymm1, %ymm1, %ymm1
|
||||
; X64-NEXT: vblendps {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3,4,5,6,7]
|
||||
; X64-NEXT: vmovaps %ymm0, (%rsi)
|
||||
|
|
Loading…
Reference in New Issue