2015-02-28 02:32:11 +08:00
|
|
|
; RUN: llc < %s -mattr=-avx -fast-isel -mcpu=core2 -O0 -regalloc=fast -asm-verbose=0 -fast-isel-abort=1 | FileCheck %s
|
[CodeGen] Pass SDAG an ORE, and replace FastISel stats with remarks.
In the long-term, we want to replace statistics with something
finer-grained that lets us gather per-function data.
Remarks are that replacement.
Create an ORE instance in SelectionDAGISel, and pass it to
SelectionDAG.
SelectionDAG was used so that we can emit remarks from all
SelectionDAG-related code, including TargetLowering and DAGCombiner.
This isn't used in the current patch but Adam tells me he's interested
for the fp-contract combines.
Use the ORE instance to emit FastISel failures as remarks (instead of
the mix of dbgs() dumps and statistics that we currently have).
Eventually, we want to have an API that tells us whether remarks are
enabled (http://llvm.org/PR32352) so that we don't emit expensive
remarks (in this case, dumping IR) when it's not needed. For now, use
'isEnabled' as a crude replacement.
This does mean that the replacement for '-fast-isel-verbose' is now
'-pass-remarks-missed=isel'. Additionally, clang users also need to
enable remark diagnostics, using '-Rpass-missed=isel'.
This also removes '-fast-isel-verbose2': there are no static statistics
that we want to only enable in asserts builds, so we can always use
the remarks regardless of the build type.
Differential Revision: https://reviews.llvm.org/D31405
llvm-svn: 299093
2017-03-31 01:49:58 +08:00
|
|
|
; RUN: llc < %s -mattr=-avx -fast-isel -mcpu=core2 -O0 -regalloc=fast -asm-verbose=0 -pass-remarks-missed=isel 2>&1 >/dev/null | FileCheck %s --check-prefix=STDERR --allow-empty
|
2015-02-28 02:32:11 +08:00
|
|
|
; RUN: llc < %s -mattr=+avx -fast-isel -mcpu=core2 -O0 -regalloc=fast -asm-verbose=0 -fast-isel-abort=1 | FileCheck %s --check-prefix=AVX
|
2011-04-17 09:16:47 +08:00
|
|
|
|
|
|
|
target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64"
|
|
|
|
target triple = "x86_64-apple-darwin10.0.0"
|
|
|
|
|
|
|
|
; Make sure that fast-isel folds the immediate into the binop even though it
|
|
|
|
; is non-canonical.
|
|
|
|
define i32 @test1(i32 %i) nounwind ssp {
|
|
|
|
%and = and i32 8, %i
|
|
|
|
ret i32 %and
|
|
|
|
}
|
|
|
|
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test1:
|
2011-04-17 09:16:47 +08:00
|
|
|
; CHECK: andl $8,
|
2011-04-17 14:35:44 +08:00
|
|
|
|
|
|
|
|
2011-04-23 05:59:37 +08:00
|
|
|
; rdar://9289512 - The load should fold into the compare.
|
|
|
|
define void @test2(i64 %x) nounwind ssp {
|
|
|
|
entry:
|
|
|
|
%x.addr = alloca i64, align 8
|
|
|
|
store i64 %x, i64* %x.addr, align 8
|
2015-02-28 05:17:42 +08:00
|
|
|
%tmp = load i64, i64* %x.addr, align 8
|
2011-04-23 05:59:37 +08:00
|
|
|
%cmp = icmp sgt i64 %tmp, 42
|
|
|
|
br i1 %cmp, label %if.then, label %if.end
|
|
|
|
|
|
|
|
if.then: ; preds = %entry
|
|
|
|
br label %if.end
|
|
|
|
|
|
|
|
if.end: ; preds = %if.then, %entry
|
|
|
|
ret void
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test2:
|
2011-04-23 05:59:37 +08:00
|
|
|
; CHECK: movq %rdi, -8(%rsp)
|
|
|
|
; CHECK: cmpq $42, -8(%rsp)
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2011-04-18 01:12:08 +08:00
|
|
|
@G = external global i32
|
|
|
|
define i64 @test3() nounwind {
|
|
|
|
%A = ptrtoint i32* @G to i64
|
|
|
|
ret i64 %A
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test3:
|
2011-04-18 01:12:08 +08:00
|
|
|
; CHECK: movq _G@GOTPCREL(%rip), %rax
|
|
|
|
; CHECK-NEXT: ret
|
2011-04-18 01:47:38 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
; rdar://9289558
|
|
|
|
@rtx_length = external global [153 x i8]
|
|
|
|
|
|
|
|
define i32 @test4(i64 %idxprom9) nounwind {
|
[opaque pointer type] Add textual IR support for explicit type parameter to getelementptr instruction
One of several parallel first steps to remove the target type of pointers,
replacing them with a single opaque pointer type.
This adds an explicit type parameter to the gep instruction so that when the
first parameter becomes an opaque pointer type, the type to gep through is
still available to the instructions.
* This doesn't modify gep operators, only instructions (operators will be
handled separately)
* Textual IR changes only. Bitcode (including upgrade) and changing the
in-memory representation will be in separate changes.
* geps of vectors are transformed as:
getelementptr <4 x float*> %x, ...
->getelementptr float, <4 x float*> %x, ...
Then, once the opaque pointer type is introduced, this will ultimately look
like:
getelementptr float, <4 x ptr> %x
with the unambiguous interpretation that it is a vector of pointers to float.
* address spaces remain on the pointer, not the type:
getelementptr float addrspace(1)* %x
->getelementptr float, float addrspace(1)* %x
Then, eventually:
getelementptr float, ptr addrspace(1) %x
Importantly, the massive amount of test case churn has been automated by
same crappy python code. I had to manually update a few test cases that
wouldn't fit the script's model (r228970,r229196,r229197,r229198). The
python script just massages stdin and writes the result to stdout, I
then wrapped that in a shell script to handle replacing files, then
using the usual find+xargs to migrate all the files.
update.py:
import fileinput
import sys
import re
ibrep = re.compile(r"(^.*?[^%\w]getelementptr inbounds )(((?:<\d* x )?)(.*?)(| addrspace\(\d\)) *\*(|>)(?:$| *(?:%|@|null|undef|blockaddress|getelementptr|addrspacecast|bitcast|inttoptr|\[\[[a-zA-Z]|\{\{).*$))")
normrep = re.compile( r"(^.*?[^%\w]getelementptr )(((?:<\d* x )?)(.*?)(| addrspace\(\d\)) *\*(|>)(?:$| *(?:%|@|null|undef|blockaddress|getelementptr|addrspacecast|bitcast|inttoptr|\[\[[a-zA-Z]|\{\{).*$))")
def conv(match, line):
if not match:
return line
line = match.groups()[0]
if len(match.groups()[5]) == 0:
line += match.groups()[2]
line += match.groups()[3]
line += ", "
line += match.groups()[1]
line += "\n"
return line
for line in sys.stdin:
if line.find("getelementptr ") == line.find("getelementptr inbounds"):
if line.find("getelementptr inbounds") != line.find("getelementptr inbounds ("):
line = conv(re.match(ibrep, line), line)
elif line.find("getelementptr ") != line.find("getelementptr ("):
line = conv(re.match(normrep, line), line)
sys.stdout.write(line)
apply.sh:
for name in "$@"
do
python3 `dirname "$0"`/update.py < "$name" > "$name.tmp" && mv "$name.tmp" "$name"
rm -f "$name.tmp"
done
The actual commands:
From llvm/src:
find test/ -name *.ll | xargs ./apply.sh
From llvm/src/tools/clang:
find test/ -name *.mm -o -name *.m -o -name *.cpp -o -name *.c | xargs -I '{}' ../../apply.sh "{}"
From llvm/src/tools/polly:
find test/ -name *.ll | xargs ./apply.sh
After that, check-all (with llvm, clang, clang-tools-extra, lld,
compiler-rt, and polly all checked out).
The extra 'rm' in the apply.sh script is due to a few files in clang's test
suite using interesting unicode stuff that my python script was throwing
exceptions on. None of those files needed to be migrated, so it seemed
sufficient to ignore those cases.
Reviewers: rafael, dexonsmith, grosser
Differential Revision: http://reviews.llvm.org/D7636
llvm-svn: 230786
2015-02-28 03:29:02 +08:00
|
|
|
%arrayidx10 = getelementptr inbounds [153 x i8], [153 x i8]* @rtx_length, i32 0, i64 %idxprom9
|
2015-02-28 05:17:42 +08:00
|
|
|
%tmp11 = load i8, i8* %arrayidx10, align 1
|
2011-04-18 01:47:38 +08:00
|
|
|
%conv = zext i8 %tmp11 to i32
|
|
|
|
ret i32 %conv
|
|
|
|
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test4:
|
2011-04-18 01:47:38 +08:00
|
|
|
; CHECK: movq _rtx_length@GOTPCREL(%rip), %rax
|
|
|
|
; CHECK-NEXT: movzbl (%rax,%rdi), %eax
|
|
|
|
; CHECK-NEXT: ret
|
|
|
|
}
|
2011-04-18 04:23:29 +08:00
|
|
|
|
|
|
|
|
|
|
|
; PR3242 - Out of range shifts should not be folded by fastisel.
|
|
|
|
define void @test5(i32 %x, i32* %p) nounwind {
|
|
|
|
%y = ashr i32 %x, 50000
|
|
|
|
store i32 %y, i32* %p
|
|
|
|
ret void
|
|
|
|
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test5:
|
2011-04-18 04:23:29 +08:00
|
|
|
; CHECK: movl $50000, %ecx
|
|
|
|
; CHECK: sarl %cl, %edi
|
|
|
|
; CHECK: ret
|
|
|
|
}
|
|
|
|
|
|
|
|
; rdar://9289501 - fast isel should fold trivial multiplies to shifts.
|
|
|
|
define i64 @test6(i64 %x) nounwind ssp {
|
|
|
|
entry:
|
|
|
|
%mul = mul nsw i64 %x, 8
|
|
|
|
ret i64 %mul
|
|
|
|
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test6:
|
2011-11-17 02:44:48 +08:00
|
|
|
; CHECK: shlq $3, %rdi
|
2011-04-18 04:23:29 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
define i32 @test7(i32 %x) nounwind ssp {
|
|
|
|
entry:
|
|
|
|
%mul = mul nsw i32 %x, 8
|
|
|
|
ret i32 %mul
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test7:
|
2011-11-17 02:44:48 +08:00
|
|
|
; CHECK: shll $3, %edi
|
2011-04-18 04:23:29 +08:00
|
|
|
}
|
|
|
|
|
2011-04-18 14:22:33 +08:00
|
|
|
|
|
|
|
; rdar://9289507 - folding of immediates into 64-bit operations.
|
|
|
|
define i64 @test8(i64 %x) nounwind ssp {
|
|
|
|
entry:
|
|
|
|
%add = add nsw i64 %x, 7
|
|
|
|
ret i64 %add
|
|
|
|
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test8:
|
2011-04-18 14:22:33 +08:00
|
|
|
; CHECK: addq $7, %rdi
|
|
|
|
}
|
|
|
|
|
|
|
|
define i64 @test9(i64 %x) nounwind ssp {
|
|
|
|
entry:
|
|
|
|
%add = mul nsw i64 %x, 7
|
|
|
|
ret i64 %add
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test9:
|
2011-04-18 14:22:33 +08:00
|
|
|
; CHECK: imulq $7, %rdi, %rax
|
|
|
|
}
|
2011-04-18 14:55:51 +08:00
|
|
|
|
|
|
|
; rdar://9297011 - Don't reject udiv by a power of 2.
|
|
|
|
define i32 @test10(i32 %X) nounwind {
|
|
|
|
%Y = udiv i32 %X, 8
|
|
|
|
ret i32 %Y
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test10:
|
2011-04-18 14:55:51 +08:00
|
|
|
; CHECK: shrl $3,
|
|
|
|
}
|
2011-04-18 15:00:40 +08:00
|
|
|
|
|
|
|
define i32 @test11(i32 %X) nounwind {
|
|
|
|
%Y = sdiv exact i32 %X, 8
|
|
|
|
ret i32 %Y
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test11:
|
2011-04-18 15:00:40 +08:00
|
|
|
; CHECK: sarl $3,
|
|
|
|
}
|
|
|
|
|
2011-04-19 12:22:17 +08:00
|
|
|
|
|
|
|
; rdar://9297006 - Trunc to bool.
|
|
|
|
define void @test12(i8 %tmp) nounwind ssp noredzone {
|
|
|
|
entry:
|
|
|
|
%tobool = trunc i8 %tmp to i1
|
|
|
|
br i1 %tobool, label %if.then, label %if.end
|
|
|
|
|
|
|
|
if.then: ; preds = %entry
|
|
|
|
call void @test12(i8 0) noredzone
|
|
|
|
br label %if.end
|
|
|
|
|
|
|
|
if.end: ; preds = %if.then, %entry
|
|
|
|
ret void
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test12:
|
2011-04-19 12:22:17 +08:00
|
|
|
; CHECK: testb $1,
|
2011-04-19 12:26:32 +08:00
|
|
|
; CHECK-NEXT: je L
|
2014-08-20 03:44:10 +08:00
|
|
|
; CHECK-NEXT: xorl %edi, %edi
|
2011-04-19 12:42:38 +08:00
|
|
|
; CHECK-NEXT: callq
|
|
|
|
}
|
|
|
|
|
|
|
|
declare void @test13f(i1 %X)
|
|
|
|
|
|
|
|
define void @test13() nounwind {
|
|
|
|
call void @test13f(i1 0)
|
|
|
|
ret void
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test13:
|
2014-08-20 03:44:10 +08:00
|
|
|
; CHECK: xorl %edi, %edi
|
2011-04-19 12:42:38 +08:00
|
|
|
; CHECK-NEXT: callq
|
2011-04-19 12:22:17 +08:00
|
|
|
}
|
|
|
|
|
2011-04-19 13:09:50 +08:00
|
|
|
|
|
|
|
|
|
|
|
; rdar://9297003 - fast isel bails out on all functions taking bools
|
|
|
|
define void @test14(i8 %tmp) nounwind ssp noredzone {
|
|
|
|
entry:
|
|
|
|
%tobool = trunc i8 %tmp to i1
|
|
|
|
call void @test13f(i1 zeroext %tobool) noredzone
|
|
|
|
ret void
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test14:
|
2011-04-19 13:09:50 +08:00
|
|
|
; CHECK: andb $1,
|
|
|
|
; CHECK: callq
|
|
|
|
}
|
|
|
|
|
Remove alignment argument from memcpy/memmove/memset in favour of alignment attributes (Step 1)
Summary:
This is a resurrection of work first proposed and discussed in Aug 2015:
http://lists.llvm.org/pipermail/llvm-dev/2015-August/089384.html
and initially landed (but then backed out) in Nov 2015:
http://lists.llvm.org/pipermail/llvm-commits/Week-of-Mon-20151109/312083.html
The @llvm.memcpy/memmove/memset intrinsics currently have an explicit argument
which is required to be a constant integer. It represents the alignment of the
dest (and source), and so must be the minimum of the actual alignment of the
two.
This change is the first in a series that allows source and dest to each
have their own alignments by using the alignment attribute on their arguments.
In this change we:
1) Remove the alignment argument.
2) Add alignment attributes to the source & dest arguments. We, temporarily,
require that the alignments for source & dest be equal.
For example, code which used to read:
call void @llvm.memcpy.p0i8.p0i8.i32(i8* %dest, i8* %src, i32 100, i32 4, i1 false)
will now read
call void @llvm.memcpy.p0i8.p0i8.i32(i8* align 4 %dest, i8* align 4 %src, i32 100, i1 false)
Downstream users may have to update their lit tests that check for
@llvm.memcpy/memmove/memset call/declaration patterns. The following extended sed script
may help with updating the majority of your tests, but it does not catch all possible
patterns so some manual checking and updating will be required.
s~declare void @llvm\.mem(set|cpy|move)\.p([^(]*)\((.*), i32, i1\)~declare void @llvm.mem\1.p\2(\3, i1)~g
s~call void @llvm\.memset\.p([^(]*)i8\(i8([^*]*)\* (.*), i8 (.*), i8 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.memset.p\1i8(i8\2* \3, i8 \4, i8 \5, i1 \6)~g
s~call void @llvm\.memset\.p([^(]*)i16\(i8([^*]*)\* (.*), i8 (.*), i16 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.memset.p\1i16(i8\2* \3, i8 \4, i16 \5, i1 \6)~g
s~call void @llvm\.memset\.p([^(]*)i32\(i8([^*]*)\* (.*), i8 (.*), i32 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.memset.p\1i32(i8\2* \3, i8 \4, i32 \5, i1 \6)~g
s~call void @llvm\.memset\.p([^(]*)i64\(i8([^*]*)\* (.*), i8 (.*), i64 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.memset.p\1i64(i8\2* \3, i8 \4, i64 \5, i1 \6)~g
s~call void @llvm\.memset\.p([^(]*)i128\(i8([^*]*)\* (.*), i8 (.*), i128 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.memset.p\1i128(i8\2* \3, i8 \4, i128 \5, i1 \6)~g
s~call void @llvm\.memset\.p([^(]*)i8\(i8([^*]*)\* (.*), i8 (.*), i8 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.memset.p\1i8(i8\2* align \6 \3, i8 \4, i8 \5, i1 \7)~g
s~call void @llvm\.memset\.p([^(]*)i16\(i8([^*]*)\* (.*), i8 (.*), i16 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.memset.p\1i16(i8\2* align \6 \3, i8 \4, i16 \5, i1 \7)~g
s~call void @llvm\.memset\.p([^(]*)i32\(i8([^*]*)\* (.*), i8 (.*), i32 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.memset.p\1i32(i8\2* align \6 \3, i8 \4, i32 \5, i1 \7)~g
s~call void @llvm\.memset\.p([^(]*)i64\(i8([^*]*)\* (.*), i8 (.*), i64 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.memset.p\1i64(i8\2* align \6 \3, i8 \4, i64 \5, i1 \7)~g
s~call void @llvm\.memset\.p([^(]*)i128\(i8([^*]*)\* (.*), i8 (.*), i128 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.memset.p\1i128(i8\2* align \6 \3, i8 \4, i128 \5, i1 \7)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i8\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i8 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.mem\1.p\2i8(i8\3* \4, i8\5* \6, i8 \7, i1 \8)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i16\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i16 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.mem\1.p\2i16(i8\3* \4, i8\5* \6, i16 \7, i1 \8)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i32\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i32 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.mem\1.p\2i32(i8\3* \4, i8\5* \6, i32 \7, i1 \8)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i64\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i64 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.mem\1.p\2i64(i8\3* \4, i8\5* \6, i64 \7, i1 \8)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i128\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i128 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.mem\1.p\2i128(i8\3* \4, i8\5* \6, i128 \7, i1 \8)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i8\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i8 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.mem\1.p\2i8(i8\3* align \8 \4, i8\5* align \8 \6, i8 \7, i1 \9)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i16\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i16 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.mem\1.p\2i16(i8\3* align \8 \4, i8\5* align \8 \6, i16 \7, i1 \9)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i32\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i32 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.mem\1.p\2i32(i8\3* align \8 \4, i8\5* align \8 \6, i32 \7, i1 \9)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i64\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i64 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.mem\1.p\2i64(i8\3* align \8 \4, i8\5* align \8 \6, i64 \7, i1 \9)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i128\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i128 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.mem\1.p\2i128(i8\3* align \8 \4, i8\5* align \8 \6, i128 \7, i1 \9)~g
The remaining changes in the series will:
Step 2) Expand the IRBuilder API to allow creation of memcpy/memmove with differing
source and dest alignments.
Step 3) Update Clang to use the new IRBuilder API.
Step 4) Update Polly to use the new IRBuilder API.
Step 5) Update LLVM passes that create memcpy/memmove calls to use the new IRBuilder API,
and those that use use MemIntrinsicInst::[get|set]Alignment() to use
getDestAlignment() and getSourceAlignment() instead.
Step 6) Remove the single-alignment IRBuilder API for memcpy/memmove, and the
MemIntrinsicInst::[get|set]Alignment() methods.
Reviewers: pete, hfinkel, lhames, reames, bollu
Reviewed By: reames
Subscribers: niosHD, reames, jholewinski, qcolombet, jfb, sanjoy, arsenm, dschuff, dylanmckay, mehdi_amini, sdardis, nemanjai, david2050, nhaehnle, javed.absar, sbc100, jgravelle-google, eraman, aheejin, kbarton, JDevlieghere, asb, rbar, johnrusso, simoncook, jordy.potman.lists, apazos, sabuasal, llvm-commits
Differential Revision: https://reviews.llvm.org/D41675
llvm-svn: 322965
2018-01-20 01:13:12 +08:00
|
|
|
declare void @llvm.memcpy.p0i8.p0i8.i64(i8*, i8*, i64, i1)
|
2011-04-19 13:52:03 +08:00
|
|
|
|
|
|
|
; rdar://9289488 - fast-isel shouldn't bail out on llvm.memcpy
|
|
|
|
define void @test15(i8* %a, i8* %b) nounwind {
|
Remove alignment argument from memcpy/memmove/memset in favour of alignment attributes (Step 1)
Summary:
This is a resurrection of work first proposed and discussed in Aug 2015:
http://lists.llvm.org/pipermail/llvm-dev/2015-August/089384.html
and initially landed (but then backed out) in Nov 2015:
http://lists.llvm.org/pipermail/llvm-commits/Week-of-Mon-20151109/312083.html
The @llvm.memcpy/memmove/memset intrinsics currently have an explicit argument
which is required to be a constant integer. It represents the alignment of the
dest (and source), and so must be the minimum of the actual alignment of the
two.
This change is the first in a series that allows source and dest to each
have their own alignments by using the alignment attribute on their arguments.
In this change we:
1) Remove the alignment argument.
2) Add alignment attributes to the source & dest arguments. We, temporarily,
require that the alignments for source & dest be equal.
For example, code which used to read:
call void @llvm.memcpy.p0i8.p0i8.i32(i8* %dest, i8* %src, i32 100, i32 4, i1 false)
will now read
call void @llvm.memcpy.p0i8.p0i8.i32(i8* align 4 %dest, i8* align 4 %src, i32 100, i1 false)
Downstream users may have to update their lit tests that check for
@llvm.memcpy/memmove/memset call/declaration patterns. The following extended sed script
may help with updating the majority of your tests, but it does not catch all possible
patterns so some manual checking and updating will be required.
s~declare void @llvm\.mem(set|cpy|move)\.p([^(]*)\((.*), i32, i1\)~declare void @llvm.mem\1.p\2(\3, i1)~g
s~call void @llvm\.memset\.p([^(]*)i8\(i8([^*]*)\* (.*), i8 (.*), i8 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.memset.p\1i8(i8\2* \3, i8 \4, i8 \5, i1 \6)~g
s~call void @llvm\.memset\.p([^(]*)i16\(i8([^*]*)\* (.*), i8 (.*), i16 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.memset.p\1i16(i8\2* \3, i8 \4, i16 \5, i1 \6)~g
s~call void @llvm\.memset\.p([^(]*)i32\(i8([^*]*)\* (.*), i8 (.*), i32 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.memset.p\1i32(i8\2* \3, i8 \4, i32 \5, i1 \6)~g
s~call void @llvm\.memset\.p([^(]*)i64\(i8([^*]*)\* (.*), i8 (.*), i64 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.memset.p\1i64(i8\2* \3, i8 \4, i64 \5, i1 \6)~g
s~call void @llvm\.memset\.p([^(]*)i128\(i8([^*]*)\* (.*), i8 (.*), i128 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.memset.p\1i128(i8\2* \3, i8 \4, i128 \5, i1 \6)~g
s~call void @llvm\.memset\.p([^(]*)i8\(i8([^*]*)\* (.*), i8 (.*), i8 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.memset.p\1i8(i8\2* align \6 \3, i8 \4, i8 \5, i1 \7)~g
s~call void @llvm\.memset\.p([^(]*)i16\(i8([^*]*)\* (.*), i8 (.*), i16 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.memset.p\1i16(i8\2* align \6 \3, i8 \4, i16 \5, i1 \7)~g
s~call void @llvm\.memset\.p([^(]*)i32\(i8([^*]*)\* (.*), i8 (.*), i32 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.memset.p\1i32(i8\2* align \6 \3, i8 \4, i32 \5, i1 \7)~g
s~call void @llvm\.memset\.p([^(]*)i64\(i8([^*]*)\* (.*), i8 (.*), i64 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.memset.p\1i64(i8\2* align \6 \3, i8 \4, i64 \5, i1 \7)~g
s~call void @llvm\.memset\.p([^(]*)i128\(i8([^*]*)\* (.*), i8 (.*), i128 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.memset.p\1i128(i8\2* align \6 \3, i8 \4, i128 \5, i1 \7)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i8\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i8 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.mem\1.p\2i8(i8\3* \4, i8\5* \6, i8 \7, i1 \8)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i16\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i16 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.mem\1.p\2i16(i8\3* \4, i8\5* \6, i16 \7, i1 \8)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i32\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i32 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.mem\1.p\2i32(i8\3* \4, i8\5* \6, i32 \7, i1 \8)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i64\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i64 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.mem\1.p\2i64(i8\3* \4, i8\5* \6, i64 \7, i1 \8)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i128\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i128 (.*), i32 [01], i1 ([^)]*)\)~call void @llvm.mem\1.p\2i128(i8\3* \4, i8\5* \6, i128 \7, i1 \8)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i8\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i8 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.mem\1.p\2i8(i8\3* align \8 \4, i8\5* align \8 \6, i8 \7, i1 \9)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i16\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i16 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.mem\1.p\2i16(i8\3* align \8 \4, i8\5* align \8 \6, i16 \7, i1 \9)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i32\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i32 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.mem\1.p\2i32(i8\3* align \8 \4, i8\5* align \8 \6, i32 \7, i1 \9)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i64\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i64 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.mem\1.p\2i64(i8\3* align \8 \4, i8\5* align \8 \6, i64 \7, i1 \9)~g
s~call void @llvm\.mem(cpy|move)\.p([^(]*)i128\(i8([^*]*)\* (.*), i8([^*]*)\* (.*), i128 (.*), i32 ([0-9]*), i1 ([^)]*)\)~call void @llvm.mem\1.p\2i128(i8\3* align \8 \4, i8\5* align \8 \6, i128 \7, i1 \9)~g
The remaining changes in the series will:
Step 2) Expand the IRBuilder API to allow creation of memcpy/memmove with differing
source and dest alignments.
Step 3) Update Clang to use the new IRBuilder API.
Step 4) Update Polly to use the new IRBuilder API.
Step 5) Update LLVM passes that create memcpy/memmove calls to use the new IRBuilder API,
and those that use use MemIntrinsicInst::[get|set]Alignment() to use
getDestAlignment() and getSourceAlignment() instead.
Step 6) Remove the single-alignment IRBuilder API for memcpy/memmove, and the
MemIntrinsicInst::[get|set]Alignment() methods.
Reviewers: pete, hfinkel, lhames, reames, bollu
Reviewed By: reames
Subscribers: niosHD, reames, jholewinski, qcolombet, jfb, sanjoy, arsenm, dschuff, dylanmckay, mehdi_amini, sdardis, nemanjai, david2050, nhaehnle, javed.absar, sbc100, jgravelle-google, eraman, aheejin, kbarton, JDevlieghere, asb, rbar, johnrusso, simoncook, jordy.potman.lists, apazos, sabuasal, llvm-commits
Differential Revision: https://reviews.llvm.org/D41675
llvm-svn: 322965
2018-01-20 01:13:12 +08:00
|
|
|
call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 4 %a, i8* align 4 %b, i64 4, i1 false)
|
2011-04-19 13:52:03 +08:00
|
|
|
ret void
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test15:
|
2011-04-19 13:52:03 +08:00
|
|
|
; CHECK-NEXT: movl (%rsi), %eax
|
|
|
|
; CHECK-NEXT: movl %eax, (%rdi)
|
|
|
|
; CHECK-NEXT: ret
|
|
|
|
}
|
2011-04-20 01:22:22 +08:00
|
|
|
|
|
|
|
; Handling for varargs calls
|
|
|
|
declare void @test16callee(...) nounwind
|
|
|
|
define void @test16() nounwind {
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test16:
|
2011-04-20 01:22:22 +08:00
|
|
|
; CHECK: movl $1, %edi
|
|
|
|
; CHECK: movb $0, %al
|
|
|
|
; CHECK: callq _test16callee
|
[opaque pointer type] Add textual IR support for explicit type parameter to the call instruction
See r230786 and r230794 for similar changes to gep and load
respectively.
Call is a bit different because it often doesn't have a single explicit
type - usually the type is deduced from the arguments, and just the
return type is explicit. In those cases there's no need to change the
IR.
When that's not the case, the IR usually contains the pointer type of
the first operand - but since typed pointers are going away, that
representation is insufficient so I'm just stripping the "pointerness"
of the explicit type away.
This does make the IR a bit weird - it /sort of/ reads like the type of
the first operand: "call void () %x(" but %x is actually of type "void
()*" and will eventually be just of type "ptr". But this seems not too
bad and I don't think it would benefit from repeating the type
("void (), void () * %x(" and then eventually "void (), ptr %x(") as has
been done with gep and load.
This also has a side benefit: since the explicit type is no longer a
pointer, there's no ambiguity between an explicit type and a function
that returns a function pointer. Previously this case needed an explicit
type (eg: a function returning a void() function was written as
"call void () () * @x(" rather than "call void () * @x(" because of the
ambiguity between a function returning a pointer to a void() function
and a function returning void).
No ambiguity means even function pointer return types can just be
written alone, without writing the whole function's type.
This leaves /only/ the varargs case where the explicit type is required.
Given the special type syntax in call instructions, the regex-fu used
for migration was a bit more involved in its own unique way (as every
one of these is) so here it is. Use it in conjunction with the apply.sh
script and associated find/xargs commands I've provided in rr230786 to
migrate your out of tree tests. Do let me know if any of this doesn't
cover your cases & we can iterate on a more general script/regexes to
help others with out of tree tests.
About 9 test cases couldn't be automatically migrated - half of those
were functions returning function pointers, where I just had to manually
delete the function argument types now that we didn't need an explicit
function type there. The other half were typedefs of function types used
in calls - just had to manually drop the * from those.
import fileinput
import sys
import re
pat = re.compile(r'((?:=|:|^|\s)call\s(?:[^@]*?))(\s*$|\s*(?:(?:\[\[[a-zA-Z0-9_]+\]\]|[@%](?:(")?[\\\?@a-zA-Z0-9_.]*?(?(3)"|)|{{.*}}))(?:\(|$)|undef|inttoptr|bitcast|null|asm).*$)')
addrspace_end = re.compile(r"addrspace\(\d+\)\s*\*$")
func_end = re.compile("(?:void.*|\)\s*)\*$")
def conv(match, line):
if not match or re.search(addrspace_end, match.group(1)) or not re.search(func_end, match.group(1)):
return line
return line[:match.start()] + match.group(1)[:match.group(1).rfind('*')].rstrip() + match.group(2) + line[match.end():]
for line in sys.stdin:
sys.stdout.write(conv(re.search(pat, line), line))
llvm-svn: 235145
2015-04-17 07:24:18 +08:00
|
|
|
call void (...) @test16callee(i32 1)
|
2011-04-20 01:22:22 +08:00
|
|
|
br label %block2
|
|
|
|
|
|
|
|
block2:
|
2014-08-20 03:05:24 +08:00
|
|
|
; CHECK: movsd LCP{{.*}}_{{.*}}(%rip), %xmm0
|
2011-04-20 01:22:22 +08:00
|
|
|
; CHECK: movb $1, %al
|
|
|
|
; CHECK: callq _test16callee
|
Introduce 'UseSSEx' to force SSE legacy encoding
- Add 'UseSSEx' to force SSE legacy insn not being selected when AVX is
enabled.
As the penalty of inter-mixing SSE and AVX instructions, we need
prevent SSE legacy insn from being generated except explicitly
specified through some intrinsics. For patterns supported by both
SSE and AVX, so far, we force AVX insn will be tried first relying on
AddedComplexity or position in td file. It's error-prone and
introduces bugs accidentally.
'UseSSEx' is disabled when AVX is turned on. For SSE insns inherited
by AVX, we need this predicate to force VEX encoding or SSE legacy
encoding only.
For insns not inherited by AVX, we still use the previous predicates,
i.e. 'HasSSEx'. So far, these insns fall into the following
categories:
* SSE insns with MMX operands
* SSE insns with GPR/MEM operands only (xFENCE, PREFETCH, CLFLUSH,
CRC, and etc.)
* SSE4A insns.
* MMX insns.
* x87 insns added by SSE.
2 test cases are modified:
- test/CodeGen/X86/fast-isel-x86-64.ll
AVX code generation is different from SSE one. 'vcvtsi2sdq' cannot be
selected by fast-isel due to complicated pattern and fast-isel
fallback to materialize it from constant pool.
- test/CodeGen/X86/widen_load-1.ll
AVX code generation is different from SSE one after fixing SSE/AVX
inter-mixing. Exec-domain fixing prefers 'vmovapd' instead of
'vmovaps'.
llvm-svn: 162919
2012-08-31 00:54:46 +08:00
|
|
|
|
|
|
|
; AVX: vmovsd LCP{{.*}}_{{.*}}(%rip), %xmm0
|
|
|
|
; AVX: movb $1, %al
|
|
|
|
; AVX: callq _test16callee
|
[opaque pointer type] Add textual IR support for explicit type parameter to the call instruction
See r230786 and r230794 for similar changes to gep and load
respectively.
Call is a bit different because it often doesn't have a single explicit
type - usually the type is deduced from the arguments, and just the
return type is explicit. In those cases there's no need to change the
IR.
When that's not the case, the IR usually contains the pointer type of
the first operand - but since typed pointers are going away, that
representation is insufficient so I'm just stripping the "pointerness"
of the explicit type away.
This does make the IR a bit weird - it /sort of/ reads like the type of
the first operand: "call void () %x(" but %x is actually of type "void
()*" and will eventually be just of type "ptr". But this seems not too
bad and I don't think it would benefit from repeating the type
("void (), void () * %x(" and then eventually "void (), ptr %x(") as has
been done with gep and load.
This also has a side benefit: since the explicit type is no longer a
pointer, there's no ambiguity between an explicit type and a function
that returns a function pointer. Previously this case needed an explicit
type (eg: a function returning a void() function was written as
"call void () () * @x(" rather than "call void () * @x(" because of the
ambiguity between a function returning a pointer to a void() function
and a function returning void).
No ambiguity means even function pointer return types can just be
written alone, without writing the whole function's type.
This leaves /only/ the varargs case where the explicit type is required.
Given the special type syntax in call instructions, the regex-fu used
for migration was a bit more involved in its own unique way (as every
one of these is) so here it is. Use it in conjunction with the apply.sh
script and associated find/xargs commands I've provided in rr230786 to
migrate your out of tree tests. Do let me know if any of this doesn't
cover your cases & we can iterate on a more general script/regexes to
help others with out of tree tests.
About 9 test cases couldn't be automatically migrated - half of those
were functions returning function pointers, where I just had to manually
delete the function argument types now that we didn't need an explicit
function type there. The other half were typedefs of function types used
in calls - just had to manually drop the * from those.
import fileinput
import sys
import re
pat = re.compile(r'((?:=|:|^|\s)call\s(?:[^@]*?))(\s*$|\s*(?:(?:\[\[[a-zA-Z0-9_]+\]\]|[@%](?:(")?[\\\?@a-zA-Z0-9_.]*?(?(3)"|)|{{.*}}))(?:\(|$)|undef|inttoptr|bitcast|null|asm).*$)')
addrspace_end = re.compile(r"addrspace\(\d+\)\s*\*$")
func_end = re.compile("(?:void.*|\)\s*)\*$")
def conv(match, line):
if not match or re.search(addrspace_end, match.group(1)) or not re.search(func_end, match.group(1)):
return line
return line[:match.start()] + match.group(1)[:match.group(1).rfind('*')].rstrip() + match.group(2) + line[match.end():]
for line in sys.stdin:
sys.stdout.write(conv(re.search(pat, line), line))
llvm-svn: 235145
2015-04-17 07:24:18 +08:00
|
|
|
call void (...) @test16callee(double 1.000000e+00)
|
2011-04-20 01:22:22 +08:00
|
|
|
ret void
|
|
|
|
}
|
2011-04-23 05:59:37 +08:00
|
|
|
|
|
|
|
|
|
|
|
declare void @foo() unnamed_addr ssp align 2
|
|
|
|
|
|
|
|
; Verify that we don't fold the load into the compare here. That would move it
|
|
|
|
; w.r.t. the call.
|
|
|
|
define i32 @test17(i32 *%P) ssp nounwind {
|
|
|
|
entry:
|
2015-02-28 05:17:42 +08:00
|
|
|
%tmp = load i32, i32* %P
|
2011-04-23 05:59:37 +08:00
|
|
|
%cmp = icmp ne i32 %tmp, 5
|
|
|
|
call void @foo()
|
|
|
|
br i1 %cmp, label %if.then, label %if.else
|
|
|
|
|
|
|
|
if.then: ; preds = %entry
|
|
|
|
ret i32 1
|
|
|
|
|
|
|
|
if.else: ; preds = %entry
|
|
|
|
ret i32 2
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test17:
|
2011-04-23 05:59:37 +08:00
|
|
|
; CHECK: movl (%rdi), %eax
|
|
|
|
; CHECK: callq _foo
|
|
|
|
; CHECK: cmpl $5, %eax
|
|
|
|
; CHECK-NEXT: je
|
|
|
|
}
|
|
|
|
|
2011-11-30 06:27:25 +08:00
|
|
|
; Check that 0.0 is materialized using xorps
|
2011-04-28 06:41:55 +08:00
|
|
|
define void @test18(float* %p1) {
|
|
|
|
store float 0.0, float* %p1
|
|
|
|
ret void
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test18:
|
2011-11-30 06:27:25 +08:00
|
|
|
; CHECK: xorps
|
2011-04-28 06:41:55 +08:00
|
|
|
}
|
2011-11-30 06:27:25 +08:00
|
|
|
|
|
|
|
; Without any type hints, doubles use the smaller xorps instead of xorpd.
|
2011-04-28 06:41:55 +08:00
|
|
|
define void @test19(double* %p1) {
|
|
|
|
store double 0.0, double* %p1
|
|
|
|
ret void
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test19:
|
2011-11-30 06:27:25 +08:00
|
|
|
; CHECK: xorps
|
2011-04-28 06:41:55 +08:00
|
|
|
}
|
2011-04-28 07:58:52 +08:00
|
|
|
|
2011-04-29 04:19:12 +08:00
|
|
|
; Check that we fast-isel sret
|
|
|
|
%struct.a = type { i64, i64, i64 }
|
|
|
|
define void @test20() nounwind ssp {
|
|
|
|
entry:
|
|
|
|
%tmp = alloca %struct.a, align 8
|
|
|
|
call void @test20sret(%struct.a* sret %tmp)
|
|
|
|
ret void
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test20:
|
2016-09-26 14:42:07 +08:00
|
|
|
; CHECK: movq %rsp, %rdi
|
2011-04-29 04:19:12 +08:00
|
|
|
; CHECK: callq _test20sret
|
|
|
|
}
|
|
|
|
declare void @test20sret(%struct.a* sret)
|
|
|
|
|
2011-11-30 06:27:25 +08:00
|
|
|
; Check that -0.0 is not materialized using xor
|
2011-04-28 08:42:03 +08:00
|
|
|
define void @test21(double* %p1) {
|
|
|
|
store double -0.0, double* %p1
|
|
|
|
ret void
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test21:
|
2011-11-30 06:27:25 +08:00
|
|
|
; CHECK-NOT: xor
|
2011-04-28 08:42:03 +08:00
|
|
|
; CHECK: movsd LCPI
|
2011-08-19 06:06:10 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
; Check that immediate arguments to a function
|
|
|
|
; do not cause massive spilling and are used
|
|
|
|
; as immediates just before the call.
|
|
|
|
define void @test22() nounwind {
|
|
|
|
entry:
|
|
|
|
call void @foo22(i32 0)
|
|
|
|
call void @foo22(i32 1)
|
|
|
|
call void @foo22(i32 2)
|
|
|
|
call void @foo22(i32 3)
|
|
|
|
ret void
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test22:
|
2014-08-20 03:44:10 +08:00
|
|
|
; CHECK: xorl %edi, %edi
|
2011-08-19 06:06:10 +08:00
|
|
|
; CHECK: callq _foo22
|
|
|
|
; CHECK: movl $1, %edi
|
|
|
|
; CHECK: callq _foo22
|
|
|
|
; CHECK: movl $2, %edi
|
|
|
|
; CHECK: callq _foo22
|
|
|
|
; CHECK: movl $3, %edi
|
|
|
|
; CHECK: callq _foo22
|
|
|
|
}
|
|
|
|
|
|
|
|
declare void @foo22(i32)
|
2012-10-03 06:45:06 +08:00
|
|
|
|
|
|
|
; PR13563
|
|
|
|
define void @test23(i8* noalias sret %result) {
|
|
|
|
%a = alloca i8
|
|
|
|
%b = call i8* @foo23()
|
|
|
|
ret void
|
2013-07-14 04:38:47 +08:00
|
|
|
; CHECK-LABEL: test23:
|
2012-10-03 06:45:06 +08:00
|
|
|
; CHECK: call
|
|
|
|
; CHECK: movq %rdi, %rax
|
|
|
|
; CHECK: ret
|
|
|
|
}
|
|
|
|
|
|
|
|
declare i8* @foo23()
|
2014-11-06 03:27:21 +08:00
|
|
|
|
|
|
|
declare void @takesi32ptr(i32* %arg)
|
|
|
|
|
|
|
|
; CHECK-LABEL: allocamaterialize
|
|
|
|
define void @allocamaterialize() {
|
|
|
|
%a = alloca i32
|
|
|
|
; CHECK: leaq {{.*}}, %rdi
|
|
|
|
call void @takesi32ptr(i32* %a)
|
|
|
|
ret void
|
|
|
|
}
|
2016-07-16 04:18:37 +08:00
|
|
|
|
|
|
|
; STDERR-NOT: FastISel missed terminator: ret void
|
|
|
|
; CHECK-LABEL: win64ccfun
|
2017-07-18 04:05:19 +08:00
|
|
|
define win64cc void @win64ccfun(i32 %i) {
|
2016-07-16 04:18:37 +08:00
|
|
|
; CHECK: ret
|
|
|
|
ret void
|
|
|
|
}
|