forked from OSchip/llvm-project
74 lines
2.5 KiB
YAML
74 lines
2.5 KiB
YAML
# NOTE: Assertions have been autogenerated by utils/update_mir_test_checks.py
|
|
# RUN: llc -O0 -march=aarch64 -run-pass=legalizer -global-isel-abort=1 %s -o - | FileCheck %s
|
|
|
|
---
|
|
name: fadd_v2s32
|
|
tracksRegLiveness: true
|
|
body: |
|
|
bb.1:
|
|
liveins: $d0
|
|
|
|
; CHECK-LABEL: name: fadd_v2s32
|
|
; CHECK: liveins: $d0
|
|
; CHECK: [[COPY:%[0-9]+]]:_(<2 x s32>) = COPY $d0
|
|
; CHECK: [[VECREDUCE_FADD:%[0-9]+]]:_(s32) = G_VECREDUCE_FADD [[COPY]](<2 x s32>)
|
|
; CHECK: $w0 = COPY [[VECREDUCE_FADD]](s32)
|
|
; CHECK: RET_ReallyLR implicit $w0
|
|
%0:_(<2 x s32>) = COPY $d0
|
|
%1:_(s32) = G_VECREDUCE_FADD %0(<2 x s32>)
|
|
$w0 = COPY %1(s32)
|
|
RET_ReallyLR implicit $w0
|
|
|
|
...
|
|
---
|
|
name: fadd_v2s64
|
|
tracksRegLiveness: true
|
|
body: |
|
|
bb.1:
|
|
liveins: $q0
|
|
|
|
; CHECK-LABEL: name: fadd_v2s64
|
|
; CHECK: liveins: $q0
|
|
; CHECK: [[COPY:%[0-9]+]]:_(<2 x s64>) = COPY $q0
|
|
; CHECK: [[VECREDUCE_FADD:%[0-9]+]]:_(s64) = G_VECREDUCE_FADD [[COPY]](<2 x s64>)
|
|
; CHECK: $x0 = COPY [[VECREDUCE_FADD]](s64)
|
|
; CHECK: RET_ReallyLR implicit $x0
|
|
%0:_(<2 x s64>) = COPY $q0
|
|
%2:_(s64) = G_VECREDUCE_FADD %0(<2 x s64>)
|
|
$x0 = COPY %2(s64)
|
|
RET_ReallyLR implicit $x0
|
|
|
|
...
|
|
---
|
|
name: fadd_v8s64
|
|
alignment: 4
|
|
tracksRegLiveness: true
|
|
body: |
|
|
bb.1:
|
|
liveins: $q0, $q1, $q2, $q3
|
|
; This is a power-of-2 legalization, so use a tree reduction.
|
|
; CHECK-LABEL: name: fadd_v8s64
|
|
; CHECK: liveins: $q0, $q1, $q2, $q3
|
|
; CHECK: [[COPY:%[0-9]+]]:_(<2 x s64>) = COPY $q0
|
|
; CHECK: [[COPY1:%[0-9]+]]:_(<2 x s64>) = COPY $q1
|
|
; CHECK: [[COPY2:%[0-9]+]]:_(<2 x s64>) = COPY $q2
|
|
; CHECK: [[COPY3:%[0-9]+]]:_(<2 x s64>) = COPY $q3
|
|
; CHECK: [[FADD:%[0-9]+]]:_(<2 x s64>) = G_FADD [[COPY]], [[COPY1]]
|
|
; CHECK: [[FADD1:%[0-9]+]]:_(<2 x s64>) = G_FADD [[COPY2]], [[COPY3]]
|
|
; CHECK: [[FADD2:%[0-9]+]]:_(<2 x s64>) = G_FADD [[FADD]], [[FADD1]]
|
|
; CHECK: [[VECREDUCE_FADD:%[0-9]+]]:_(s64) = G_VECREDUCE_FADD [[FADD2]](<2 x s64>)
|
|
; CHECK: $x0 = COPY [[VECREDUCE_FADD]](s64)
|
|
; CHECK: RET_ReallyLR implicit $x0
|
|
%0:_(<2 x s64>) = COPY $q0
|
|
%1:_(<2 x s64>) = COPY $q1
|
|
%2:_(<2 x s64>) = COPY $q2
|
|
%3:_(<2 x s64>) = COPY $q3
|
|
%4:_(<4 x s64>) = G_CONCAT_VECTORS %0(<2 x s64>), %1(<2 x s64>)
|
|
%5:_(<4 x s64>) = G_CONCAT_VECTORS %2(<2 x s64>), %3(<2 x s64>)
|
|
%6:_(<8 x s64>) = G_CONCAT_VECTORS %4(<4 x s64>), %5(<4 x s64>)
|
|
%7:_(s64) = G_VECREDUCE_FADD %6(<8 x s64>)
|
|
$x0 = COPY %7(s64)
|
|
RET_ReallyLR implicit $x0
|
|
|
|
...
|