forked from OSchip/llvm-project
311 lines
6.3 KiB
LLVM
311 lines
6.3 KiB
LLVM
; RUN: llc < %s -march=sparcv9 -mattr=+popc -disable-sparc-delay-filler -disable-sparc-leaf-proc | FileCheck %s
|
|
; RUN: llc < %s -march=sparcv9 -mattr=+popc | FileCheck %s -check-prefix=OPT
|
|
|
|
; CHECK-LABEL: ret2:
|
|
; CHECK: or %g0, %i1, %i0
|
|
|
|
; OPT-LABEL: ret2:
|
|
; OPT: retl
|
|
; OPT: or %g0, %o1, %o0
|
|
define i64 @ret2(i64 %a, i64 %b) {
|
|
ret i64 %b
|
|
}
|
|
|
|
; CHECK: shl_imm
|
|
; CHECK: sllx %i0, 7, %i0
|
|
|
|
; OPT-LABEL: shl_imm:
|
|
; OPT: retl
|
|
; OPT: sllx %o0, 7, %o0
|
|
define i64 @shl_imm(i64 %a) {
|
|
%x = shl i64 %a, 7
|
|
ret i64 %x
|
|
}
|
|
|
|
; CHECK: sra_reg
|
|
; CHECK: srax %i0, %i1, %i0
|
|
|
|
; OPT-LABEL: sra_reg:
|
|
; OPT: retl
|
|
; OPT: srax %o0, %o1, %o0
|
|
define i64 @sra_reg(i64 %a, i64 %b) {
|
|
%x = ashr i64 %a, %b
|
|
ret i64 %x
|
|
}
|
|
|
|
; Immediate materialization. Many of these patterns could actually be merged
|
|
; into the restore instruction:
|
|
;
|
|
; restore %g0, %g0, %o0
|
|
;
|
|
; CHECK: ret_imm0
|
|
; CHECK: or %g0, 0, %i0
|
|
|
|
; OPT: ret_imm0
|
|
; OPT: retl
|
|
; OPT: or %g0, 0, %o0
|
|
define i64 @ret_imm0() {
|
|
ret i64 0
|
|
}
|
|
|
|
; CHECK: ret_simm13
|
|
; CHECK: or %g0, -4096, %i0
|
|
|
|
; OPT: ret_simm13
|
|
; OPT: retl
|
|
; OPT: or %g0, -4096, %o0
|
|
define i64 @ret_simm13() {
|
|
ret i64 -4096
|
|
}
|
|
|
|
; CHECK: ret_sethi
|
|
; CHECK: sethi 4, %i0
|
|
; CHECK-NOT: or
|
|
; CHECK: restore
|
|
|
|
; OPT: ret_sethi
|
|
; OPT: retl
|
|
; OPT: sethi 4, %o0
|
|
define i64 @ret_sethi() {
|
|
ret i64 4096
|
|
}
|
|
|
|
; CHECK: ret_sethi_or
|
|
; CHECK: sethi 4, [[R:%[goli][0-7]]]
|
|
; CHECK: or [[R]], 1, %i0
|
|
|
|
; OPT: ret_sethi_or
|
|
; OPT: sethi 4, [[R:%[go][0-7]]]
|
|
; OPT: retl
|
|
; OPT: or [[R]], 1, %o0
|
|
|
|
define i64 @ret_sethi_or() {
|
|
ret i64 4097
|
|
}
|
|
|
|
; CHECK: ret_nimm33
|
|
; CHECK: sethi 4, [[R:%[goli][0-7]]]
|
|
; CHECK: xor [[R]], -4, %i0
|
|
|
|
; OPT: ret_nimm33
|
|
; OPT: sethi 4, [[R:%[go][0-7]]]
|
|
; OPT: retl
|
|
; OPT: xor [[R]], -4, %o0
|
|
|
|
define i64 @ret_nimm33() {
|
|
ret i64 -4100
|
|
}
|
|
|
|
; CHECK: ret_bigimm
|
|
; CHECK: sethi
|
|
; CHECK: sethi
|
|
define i64 @ret_bigimm() {
|
|
ret i64 6800754272627607872
|
|
}
|
|
|
|
; CHECK: ret_bigimm2
|
|
; CHECK: sethi 1048576
|
|
define i64 @ret_bigimm2() {
|
|
ret i64 4611686018427387904 ; 0x4000000000000000
|
|
}
|
|
|
|
; CHECK: reg_reg_alu
|
|
; CHECK: add %i0, %i1, [[R0:%[goli][0-7]]]
|
|
; CHECK: sub [[R0]], %i2, [[R1:%[goli][0-7]]]
|
|
; CHECK: andn [[R1]], %i0, %i0
|
|
define i64 @reg_reg_alu(i64 %x, i64 %y, i64 %z) {
|
|
%a = add i64 %x, %y
|
|
%b = sub i64 %a, %z
|
|
%c = xor i64 %x, -1
|
|
%d = and i64 %b, %c
|
|
ret i64 %d
|
|
}
|
|
|
|
; CHECK: reg_imm_alu
|
|
; CHECK: add %i0, -5, [[R0:%[goli][0-7]]]
|
|
; CHECK: xor [[R0]], 2, %i0
|
|
define i64 @reg_imm_alu(i64 %x, i64 %y, i64 %z) {
|
|
%a = add i64 %x, -5
|
|
%b = xor i64 %a, 2
|
|
ret i64 %b
|
|
}
|
|
|
|
; CHECK: loads
|
|
; CHECK: ldx [%i0]
|
|
; CHECK: stx %
|
|
; CHECK: ld [%i1]
|
|
; CHECK: st %
|
|
; CHECK: ldsw [%i2]
|
|
; CHECK: stx %
|
|
; CHECK: ldsh [%i3]
|
|
; CHECK: sth %
|
|
define i64 @loads(i64* %p, i32* %q, i32* %r, i16* %s) {
|
|
%a = load i64* %p
|
|
%ai = add i64 1, %a
|
|
store i64 %ai, i64* %p
|
|
%b = load i32* %q
|
|
%b2 = zext i32 %b to i64
|
|
%bi = trunc i64 %ai to i32
|
|
store i32 %bi, i32* %q
|
|
%c = load i32* %r
|
|
%c2 = sext i32 %c to i64
|
|
store i64 %ai, i64* %p
|
|
%d = load i16* %s
|
|
%d2 = sext i16 %d to i64
|
|
%di = trunc i64 %ai to i16
|
|
store i16 %di, i16* %s
|
|
|
|
%x1 = add i64 %a, %b2
|
|
%x2 = add i64 %c2, %d2
|
|
%x3 = add i64 %x1, %x2
|
|
ret i64 %x3
|
|
}
|
|
|
|
; CHECK: load_bool
|
|
; CHECK: ldub [%i0], %i0
|
|
define i64 @load_bool(i1* %p) {
|
|
%a = load i1* %p
|
|
%b = zext i1 %a to i64
|
|
ret i64 %b
|
|
}
|
|
|
|
; CHECK: stores
|
|
; CHECK: ldx [%i0+8], [[R:%[goli][0-7]]]
|
|
; CHECK: stx [[R]], [%i0+16]
|
|
; CHECK: st [[R]], [%i1+-8]
|
|
; CHECK: sth [[R]], [%i2+40]
|
|
; CHECK: stb [[R]], [%i3+-20]
|
|
define void @stores(i64* %p, i32* %q, i16* %r, i8* %s) {
|
|
%p1 = getelementptr i64* %p, i64 1
|
|
%p2 = getelementptr i64* %p, i64 2
|
|
%pv = load i64* %p1
|
|
store i64 %pv, i64* %p2
|
|
|
|
%q2 = getelementptr i32* %q, i32 -2
|
|
%qv = trunc i64 %pv to i32
|
|
store i32 %qv, i32* %q2
|
|
|
|
%r2 = getelementptr i16* %r, i16 20
|
|
%rv = trunc i64 %pv to i16
|
|
store i16 %rv, i16* %r2
|
|
|
|
%s2 = getelementptr i8* %s, i8 -20
|
|
%sv = trunc i64 %pv to i8
|
|
store i8 %sv, i8* %s2
|
|
|
|
ret void
|
|
}
|
|
|
|
; CHECK: promote_shifts
|
|
; CHECK: ldub [%i0], [[R:%[goli][0-7]]]
|
|
; CHECK: sll [[R]], [[R]], %i0
|
|
define i8 @promote_shifts(i8* %p) {
|
|
%L24 = load i8* %p
|
|
%L32 = load i8* %p
|
|
%B36 = shl i8 %L24, %L32
|
|
ret i8 %B36
|
|
}
|
|
|
|
; CHECK: multiply
|
|
; CHECK: mulx %i0, %i1, %i0
|
|
define i64 @multiply(i64 %a, i64 %b) {
|
|
%r = mul i64 %a, %b
|
|
ret i64 %r
|
|
}
|
|
|
|
; CHECK: signed_divide
|
|
; CHECK: sdivx %i0, %i1, %i0
|
|
define i64 @signed_divide(i64 %a, i64 %b) {
|
|
%r = sdiv i64 %a, %b
|
|
ret i64 %r
|
|
}
|
|
|
|
; CHECK: unsigned_divide
|
|
; CHECK: udivx %i0, %i1, %i0
|
|
define i64 @unsigned_divide(i64 %a, i64 %b) {
|
|
%r = udiv i64 %a, %b
|
|
ret i64 %r
|
|
}
|
|
|
|
define void @access_fi() {
|
|
entry:
|
|
%b = alloca [32 x i8], align 1
|
|
%arraydecay = getelementptr inbounds [32 x i8]* %b, i64 0, i64 0
|
|
call void @g(i8* %arraydecay) #2
|
|
ret void
|
|
}
|
|
|
|
declare void @g(i8*)
|
|
|
|
; CHECK: expand_setcc
|
|
; CHECK: cmp %i0, 1
|
|
; CHECK: movl %xcc, 1,
|
|
define i32 @expand_setcc(i64 %a) {
|
|
%cond = icmp sle i64 %a, 0
|
|
%cast2 = zext i1 %cond to i32
|
|
%RV = sub i32 1, %cast2
|
|
ret i32 %RV
|
|
}
|
|
|
|
; CHECK: spill_i64
|
|
; CHECK: stx
|
|
; CHECK: ldx
|
|
define i64 @spill_i64(i64 %x) {
|
|
call void asm sideeffect "", "~{i0},~{i1},~{i2},~{i3},~{i4},~{i5},~{o0},~{o1},~{o2},~{o3},~{o4},~{o5},~{o7},~{l0},~{l1},~{l2},~{l3},~{l4},~{l5},~{l6},~{l7},~{g1},~{g2},~{g3},~{g4},~{g5},~{g6},~{g7}"()
|
|
ret i64 %x
|
|
}
|
|
|
|
; CHECK: bitcast_i64_f64
|
|
; CHECK: std
|
|
; CHECK: ldx
|
|
define i64 @bitcast_i64_f64(double %x) {
|
|
%y = bitcast double %x to i64
|
|
ret i64 %y
|
|
}
|
|
|
|
; CHECK: bitcast_f64_i64
|
|
; CHECK: stx
|
|
; CHECK: ldd
|
|
define double @bitcast_f64_i64(i64 %x) {
|
|
%y = bitcast i64 %x to double
|
|
ret double %y
|
|
}
|
|
|
|
; CHECK-LABEL: store_zero:
|
|
; CHECK: stx %g0, [%i0]
|
|
; CHECK: stx %g0, [%i1+8]
|
|
|
|
; OPT-LABEL: store_zero:
|
|
; OPT: stx %g0, [%o0]
|
|
; OPT: stx %g0, [%o1+8]
|
|
define i64 @store_zero(i64* nocapture %a, i64* nocapture %b) {
|
|
entry:
|
|
store i64 0, i64* %a, align 8
|
|
%0 = getelementptr inbounds i64* %b, i32 1
|
|
store i64 0, i64* %0, align 8
|
|
ret i64 0
|
|
}
|
|
|
|
; CHECK-LABEL: bit_ops
|
|
; CHECK: popc
|
|
|
|
; OPT-LABEL: bit_ops
|
|
; OPT: popc
|
|
|
|
define i64 @bit_ops(i64 %arg) {
|
|
entry:
|
|
%0 = tail call i64 @llvm.ctpop.i64(i64 %arg)
|
|
%1 = tail call i64 @llvm.ctlz.i64(i64 %arg, i1 true)
|
|
%2 = tail call i64 @llvm.cttz.i64(i64 %arg, i1 true)
|
|
%3 = tail call i64 @llvm.bswap.i64(i64 %arg)
|
|
%4 = add i64 %0, %1
|
|
%5 = add i64 %2, %3
|
|
%6 = add i64 %4, %5
|
|
ret i64 %6
|
|
}
|
|
|
|
declare i64 @llvm.ctpop.i64(i64) nounwind readnone
|
|
declare i64 @llvm.ctlz.i64(i64, i1) nounwind readnone
|
|
declare i64 @llvm.cttz.i64(i64, i1) nounwind readnone
|
|
declare i64 @llvm.bswap.i64(i64) nounwind readnone
|