forked from OSchip/llvm-project
1030 lines
32 KiB
LLVM
1030 lines
32 KiB
LLVM
; RUN: opt < %s -msan-check-access-address=0 -S -passes='module(msan-module),function(msan)' 2>&1 | FileCheck -allow-deprecated-dag-overlap %s
|
|
; RUN: opt < %s --passes='module(msan-module),function(msan)' -msan-check-access-address=0 -S | FileCheck -allow-deprecated-dag-overlap %s
|
|
; RUN: opt < %s -msan-check-access-address=0 -msan-track-origins=1 -S -passes='module(msan-module),function(msan)' 2>&1 | \
|
|
; RUN: FileCheck -allow-deprecated-dag-overlap -check-prefixes=CHECK,CHECK-ORIGINS %s
|
|
; RUN: opt < %s -passes='module(msan-module),function(msan)' -msan-check-access-address=0 -msan-track-origins=1 -S | \
|
|
; RUN: FileCheck -allow-deprecated-dag-overlap -check-prefixes=CHECK,CHECK-ORIGINS %s
|
|
; RUN: opt < %s -passes='module(msan-module),function(msan)' -msan-instrumentation-with-call-threshold=0 -msan-track-origins=1 -S | \
|
|
; RUN: FileCheck -allow-deprecated-dag-overlap -check-prefixes=CHECK-CALLS %s
|
|
|
|
target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64-S128"
|
|
target triple = "x86_64-unknown-linux-gnu"
|
|
|
|
; CHECK: @llvm.used = appending global [1 x i8*] [i8* bitcast (void ()* @msan.module_ctor to i8*)]
|
|
; CHECK: @llvm.global_ctors {{.*}} { i32 0, void ()* @msan.module_ctor, i8* null }
|
|
|
|
; Check the presence and the linkage type of __msan_track_origins and
|
|
; other interface symbols.
|
|
; CHECK-NOT: @__msan_track_origins
|
|
; CHECK-ORIGINS: @__msan_track_origins = weak_odr constant i32 1
|
|
; CHECK-NOT: @__msan_keep_going = weak_odr constant i32 0
|
|
; CHECK: @__msan_retval_tls = external thread_local(initialexec) global [{{.*}}]
|
|
; CHECK: @__msan_retval_origin_tls = external thread_local(initialexec) global i32
|
|
; CHECK: @__msan_param_tls = external thread_local(initialexec) global [{{.*}}]
|
|
; CHECK: @__msan_param_origin_tls = external thread_local(initialexec) global [{{.*}}]
|
|
; CHECK: @__msan_va_arg_tls = external thread_local(initialexec) global [{{.*}}]
|
|
; CHECK: @__msan_va_arg_overflow_size_tls = external thread_local(initialexec) global i64
|
|
|
|
|
|
; Check instrumentation of stores
|
|
|
|
define void @Store(i32* nocapture %p, i32 %x) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
store i32 %x, i32* %p, align 4
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @Store
|
|
; CHECK: load {{.*}} @__msan_param_tls
|
|
; CHECK-ORIGINS: load {{.*}} @__msan_param_origin_tls
|
|
; CHECK: store
|
|
; CHECK-ORIGINS: icmp
|
|
; CHECK-ORIGINS: br i1
|
|
; CHECK-ORIGINS: {{^[0-9]+}}:
|
|
; CHECK-ORIGINS: store
|
|
; CHECK-ORIGINS: br label
|
|
; CHECK-ORIGINS: {{^[0-9]+}}:
|
|
; CHECK: store
|
|
; CHECK: ret void
|
|
|
|
|
|
; Check instrumentation of aligned stores
|
|
; Shadow store has the same alignment as the original store; origin store
|
|
; does not specify explicit alignment.
|
|
|
|
define void @AlignedStore(i32* nocapture %p, i32 %x) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
store i32 %x, i32* %p, align 32
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @AlignedStore
|
|
; CHECK: load {{.*}} @__msan_param_tls
|
|
; CHECK-ORIGINS: load {{.*}} @__msan_param_origin_tls
|
|
; CHECK: store {{.*}} align 32
|
|
; CHECK-ORIGINS: icmp
|
|
; CHECK-ORIGINS: br i1
|
|
; CHECK-ORIGINS: {{^[0-9]+}}:
|
|
; CHECK-ORIGINS: store {{.*}} align 32
|
|
; CHECK-ORIGINS: br label
|
|
; CHECK-ORIGINS: {{^[0-9]+}}:
|
|
; CHECK: store {{.*}} align 32
|
|
; CHECK: ret void
|
|
|
|
|
|
; load followed by cmp: check that we load the shadow and call __msan_warning_with_origin.
|
|
define void @LoadAndCmp(i32* nocapture %a) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
%0 = load i32, i32* %a, align 4
|
|
%tobool = icmp eq i32 %0, 0
|
|
br i1 %tobool, label %if.end, label %if.then
|
|
|
|
if.then: ; preds = %entry
|
|
tail call void (...) @foo() nounwind
|
|
br label %if.end
|
|
|
|
if.end: ; preds = %entry, %if.then
|
|
ret void
|
|
}
|
|
|
|
declare void @foo(...)
|
|
|
|
; CHECK-LABEL: @LoadAndCmp
|
|
; CHECK: %0 = load i32,
|
|
; CHECK: = load
|
|
; CHECK-ORIGINS: %[[ORIGIN:.*]] = load
|
|
; CHECK: call void @__msan_warning_with_origin_noreturn(i32
|
|
; CHECK-ORIGINS-SAME %[[ORIGIN]])
|
|
; CHECK-CONT:
|
|
; CHECK-NEXT: unreachable
|
|
; CHECK: br i1 %tobool
|
|
; CHECK: ret void
|
|
|
|
; Check that we store the shadow for the retval.
|
|
define i32 @ReturnInt() nounwind uwtable readnone sanitize_memory {
|
|
entry:
|
|
ret i32 123
|
|
}
|
|
|
|
; CHECK-LABEL: @ReturnInt
|
|
; CHECK: store i32 0,{{.*}}__msan_retval_tls
|
|
; CHECK: ret i32
|
|
|
|
; Check that we get the shadow for the retval.
|
|
define void @CopyRetVal(i32* nocapture %a) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
%call = tail call i32 @ReturnInt() nounwind
|
|
store i32 %call, i32* %a, align 4
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @CopyRetVal
|
|
; CHECK: load{{.*}}__msan_retval_tls
|
|
; CHECK: store
|
|
; CHECK: store
|
|
; CHECK: ret void
|
|
|
|
|
|
; Check that we generate PHIs for shadow.
|
|
define void @FuncWithPhi(i32* nocapture %a, i32* %b, i32* nocapture %c) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
%tobool = icmp eq i32* %b, null
|
|
br i1 %tobool, label %if.else, label %if.then
|
|
|
|
if.then: ; preds = %entry
|
|
%0 = load i32, i32* %b, align 4
|
|
br label %if.end
|
|
|
|
if.else: ; preds = %entry
|
|
%1 = load i32, i32* %c, align 4
|
|
br label %if.end
|
|
|
|
if.end: ; preds = %if.else, %if.then
|
|
%t.0 = phi i32 [ %0, %if.then ], [ %1, %if.else ]
|
|
store i32 %t.0, i32* %a, align 4
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @FuncWithPhi
|
|
; CHECK: = phi
|
|
; CHECK-NEXT: = phi
|
|
; CHECK: store
|
|
; CHECK: store
|
|
; CHECK: ret void
|
|
|
|
; Compute shadow for "x << 10"
|
|
define void @ShlConst(i32* nocapture %x) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
%0 = load i32, i32* %x, align 4
|
|
%1 = shl i32 %0, 10
|
|
store i32 %1, i32* %x, align 4
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @ShlConst
|
|
; CHECK: = load
|
|
; CHECK: = load
|
|
; CHECK: shl
|
|
; CHECK: shl
|
|
; CHECK: store
|
|
; CHECK: store
|
|
; CHECK: ret void
|
|
|
|
; Compute shadow for "10 << x": it should have 'sext i1'.
|
|
define void @ShlNonConst(i32* nocapture %x) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
%0 = load i32, i32* %x, align 4
|
|
%1 = shl i32 10, %0
|
|
store i32 %1, i32* %x, align 4
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @ShlNonConst
|
|
; CHECK: = load
|
|
; CHECK: = load
|
|
; CHECK: = sext i1
|
|
; CHECK: store
|
|
; CHECK: store
|
|
; CHECK: ret void
|
|
|
|
; SExt
|
|
define void @SExt(i32* nocapture %a, i16* nocapture %b) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
%0 = load i16, i16* %b, align 2
|
|
%1 = sext i16 %0 to i32
|
|
store i32 %1, i32* %a, align 4
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @SExt
|
|
; CHECK: = load
|
|
; CHECK: = load
|
|
; CHECK: = sext
|
|
; CHECK: = sext
|
|
; CHECK: store
|
|
; CHECK: store
|
|
; CHECK: ret void
|
|
|
|
|
|
; memset
|
|
define void @MemSet(i8* nocapture %x) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
call void @llvm.memset.p0i8.i64(i8* %x, i8 42, i64 10, i1 false)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.memset.p0i8.i64(i8* nocapture, i8, i64, i1) nounwind
|
|
|
|
; CHECK-LABEL: @MemSet
|
|
; CHECK: call i8* @__msan_memset
|
|
; CHECK: ret void
|
|
|
|
|
|
; memcpy
|
|
define void @MemCpy(i8* nocapture %x, i8* nocapture %y) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
call void @llvm.memcpy.p0i8.p0i8.i64(i8* %x, i8* %y, i64 10, i1 false)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.memcpy.p0i8.p0i8.i64(i8* nocapture, i8* nocapture, i64, i1) nounwind
|
|
|
|
; CHECK-LABEL: @MemCpy
|
|
; CHECK: call i8* @__msan_memcpy
|
|
; CHECK: ret void
|
|
|
|
|
|
; memmove is lowered to a call
|
|
define void @MemMove(i8* nocapture %x, i8* nocapture %y) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
call void @llvm.memmove.p0i8.p0i8.i64(i8* %x, i8* %y, i64 10, i1 false)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.memmove.p0i8.p0i8.i64(i8* nocapture, i8* nocapture, i64, i1) nounwind
|
|
|
|
; CHECK-LABEL: @MemMove
|
|
; CHECK: call i8* @__msan_memmove
|
|
; CHECK: ret void
|
|
|
|
;; ------------
|
|
;; Placeholder tests that will fail once element atomic @llvm.mem[cpy|move|set] intrinsics have
|
|
;; been added to the MemIntrinsic class hierarchy. These will act as a reminder to
|
|
;; verify that MSAN handles these intrinsics properly once they have been
|
|
;; added to that class hierarchy.
|
|
declare void @llvm.memset.element.unordered.atomic.p0i8.i64(i8* nocapture writeonly, i8, i64, i32) nounwind
|
|
declare void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* nocapture writeonly, i8* nocapture readonly, i64, i32) nounwind
|
|
declare void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* nocapture writeonly, i8* nocapture readonly, i64, i32) nounwind
|
|
|
|
define void @atomic_memcpy(i8* nocapture %x, i8* nocapture %y) nounwind {
|
|
; CHECK-LABEL: atomic_memcpy
|
|
; CHECK-NEXT: call void @llvm.donothing
|
|
; CHECK-NEXT: call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %x, i8* align 2 %y, i64 16, i32 1)
|
|
; CHECK-NEXT: ret void
|
|
call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %x, i8* align 2 %y, i64 16, i32 1)
|
|
ret void
|
|
}
|
|
|
|
define void @atomic_memmove(i8* nocapture %x, i8* nocapture %y) nounwind {
|
|
; CHECK-LABEL: atomic_memmove
|
|
; CHECK-NEXT: call void @llvm.donothing
|
|
; CHECK-NEXT: call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %x, i8* align 2 %y, i64 16, i32 1)
|
|
; CHECK-NEXT: ret void
|
|
call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %x, i8* align 2 %y, i64 16, i32 1)
|
|
ret void
|
|
}
|
|
|
|
define void @atomic_memset(i8* nocapture %x) nounwind {
|
|
; CHECK-LABEL: atomic_memset
|
|
; CHECK-NEXT: call void @llvm.donothing
|
|
; CHECK-NEXT: call void @llvm.memset.element.unordered.atomic.p0i8.i64(i8* align 1 %x, i8 88, i64 16, i32 1)
|
|
; CHECK-NEXT: ret void
|
|
call void @llvm.memset.element.unordered.atomic.p0i8.i64(i8* align 1 %x, i8 88, i64 16, i32 1)
|
|
ret void
|
|
}
|
|
|
|
;; ------------
|
|
|
|
|
|
; Check that we propagate shadow for "select"
|
|
|
|
define i32 @Select(i32 %a, i32 %b, i1 %c) nounwind uwtable readnone sanitize_memory {
|
|
entry:
|
|
%cond = select i1 %c, i32 %a, i32 %b
|
|
ret i32 %cond
|
|
}
|
|
|
|
; CHECK-LABEL: @Select
|
|
; CHECK: select i1
|
|
; CHECK-DAG: or i32
|
|
; CHECK-DAG: xor i32
|
|
; CHECK: or i32
|
|
; CHECK-DAG: select i1
|
|
; CHECK-ORIGINS-DAG: select
|
|
; CHECK-ORIGINS-DAG: select
|
|
; CHECK-DAG: select i1
|
|
; CHECK: store i32{{.*}}@__msan_retval_tls
|
|
; CHECK-ORIGINS: store i32{{.*}}@__msan_retval_origin_tls
|
|
; CHECK: ret i32
|
|
|
|
|
|
; Check that we propagate origin for "select" with vector condition.
|
|
; Select condition is flattened to i1, which is then used to select one of the
|
|
; argument origins.
|
|
|
|
define <8 x i16> @SelectVector(<8 x i16> %a, <8 x i16> %b, <8 x i1> %c) nounwind uwtable readnone sanitize_memory {
|
|
entry:
|
|
%cond = select <8 x i1> %c, <8 x i16> %a, <8 x i16> %b
|
|
ret <8 x i16> %cond
|
|
}
|
|
|
|
; CHECK-LABEL: @SelectVector
|
|
; CHECK: select <8 x i1>
|
|
; CHECK-DAG: or <8 x i16>
|
|
; CHECK-DAG: xor <8 x i16>
|
|
; CHECK: or <8 x i16>
|
|
; CHECK-DAG: select <8 x i1>
|
|
; CHECK-ORIGINS-DAG: select
|
|
; CHECK-ORIGINS-DAG: select
|
|
; CHECK-DAG: select <8 x i1>
|
|
; CHECK: store <8 x i16>{{.*}}@__msan_retval_tls
|
|
; CHECK-ORIGINS: store i32{{.*}}@__msan_retval_origin_tls
|
|
; CHECK: ret <8 x i16>
|
|
|
|
|
|
; Check that we propagate origin for "select" with scalar condition and vector
|
|
; arguments. Select condition shadow is sign-extended to the vector type and
|
|
; mixed into the result shadow.
|
|
|
|
define <8 x i16> @SelectVector2(<8 x i16> %a, <8 x i16> %b, i1 %c) nounwind uwtable readnone sanitize_memory {
|
|
entry:
|
|
%cond = select i1 %c, <8 x i16> %a, <8 x i16> %b
|
|
ret <8 x i16> %cond
|
|
}
|
|
|
|
; CHECK-LABEL: @SelectVector2
|
|
; CHECK: select i1
|
|
; CHECK-DAG: or <8 x i16>
|
|
; CHECK-DAG: xor <8 x i16>
|
|
; CHECK: or <8 x i16>
|
|
; CHECK-DAG: select i1
|
|
; CHECK-ORIGINS-DAG: select i1
|
|
; CHECK-ORIGINS-DAG: select i1
|
|
; CHECK-DAG: select i1
|
|
; CHECK: ret <8 x i16>
|
|
|
|
|
|
define { i64, i64 } @SelectStruct(i1 zeroext %x, { i64, i64 } %a, { i64, i64 } %b) readnone sanitize_memory {
|
|
entry:
|
|
%c = select i1 %x, { i64, i64 } %a, { i64, i64 } %b
|
|
ret { i64, i64 } %c
|
|
}
|
|
|
|
; CHECK-LABEL: @SelectStruct
|
|
; CHECK: select i1 {{.*}}, { i64, i64 }
|
|
; CHECK-NEXT: select i1 {{.*}}, { i64, i64 } { i64 -1, i64 -1 }, { i64, i64 }
|
|
; CHECK-ORIGINS: select i1
|
|
; CHECK-ORIGINS: select i1
|
|
; CHECK-NEXT: select i1 {{.*}}, { i64, i64 }
|
|
; CHECK: ret { i64, i64 }
|
|
|
|
|
|
define { i64*, double } @SelectStruct2(i1 zeroext %x, { i64*, double } %a, { i64*, double } %b) readnone sanitize_memory {
|
|
entry:
|
|
%c = select i1 %x, { i64*, double } %a, { i64*, double } %b
|
|
ret { i64*, double } %c
|
|
}
|
|
|
|
; CHECK-LABEL: @SelectStruct2
|
|
; CHECK: select i1 {{.*}}, { i64, i64 }
|
|
; CHECK-NEXT: select i1 {{.*}}, { i64, i64 } { i64 -1, i64 -1 }, { i64, i64 }
|
|
; CHECK-ORIGINS: select i1
|
|
; CHECK-ORIGINS: select i1
|
|
; CHECK-NEXT: select i1 {{.*}}, { i64*, double }
|
|
; CHECK: ret { i64*, double }
|
|
|
|
|
|
define i8* @IntToPtr(i64 %x) nounwind uwtable readnone sanitize_memory {
|
|
entry:
|
|
%0 = inttoptr i64 %x to i8*
|
|
ret i8* %0
|
|
}
|
|
|
|
; CHECK-LABEL: @IntToPtr
|
|
; CHECK: load i64, i64*{{.*}}__msan_param_tls
|
|
; CHECK-ORIGINS-NEXT: load i32, i32*{{.*}}__msan_param_origin_tls
|
|
; CHECK-NEXT: call void @llvm.donothing
|
|
; CHECK-NEXT: inttoptr
|
|
; CHECK-NEXT: store i64{{.*}}__msan_retval_tls
|
|
; CHECK: ret i8*
|
|
|
|
|
|
define i8* @IntToPtr_ZExt(i16 %x) nounwind uwtable readnone sanitize_memory {
|
|
entry:
|
|
%0 = inttoptr i16 %x to i8*
|
|
ret i8* %0
|
|
}
|
|
|
|
; CHECK-LABEL: @IntToPtr_ZExt
|
|
; CHECK: load i16, i16*{{.*}}__msan_param_tls
|
|
; CHECK: zext
|
|
; CHECK-NEXT: inttoptr
|
|
; CHECK-NEXT: store i64{{.*}}__msan_retval_tls
|
|
; CHECK: ret i8*
|
|
|
|
|
|
; Check that we insert exactly one check on udiv
|
|
; (2nd arg shadow is checked, 1st arg shadow is propagated)
|
|
|
|
define i32 @Div(i32 %a, i32 %b) nounwind uwtable readnone sanitize_memory {
|
|
entry:
|
|
%div = udiv i32 %a, %b
|
|
ret i32 %div
|
|
}
|
|
|
|
; CHECK-LABEL: @Div
|
|
; CHECK: icmp
|
|
; CHECK: call void @__msan_warning_with_origin
|
|
; CHECK-NOT: icmp
|
|
; CHECK: udiv
|
|
; CHECK-NOT: icmp
|
|
; CHECK: ret i32
|
|
|
|
; Check that fdiv, unlike udiv, simply propagates shadow.
|
|
|
|
define float @FDiv(float %a, float %b) nounwind uwtable readnone sanitize_memory {
|
|
entry:
|
|
%c = fdiv float %a, %b
|
|
ret float %c
|
|
}
|
|
|
|
; CHECK-LABEL: @FDiv
|
|
; CHECK: %[[SA:.*]] = load i32,{{.*}}@__msan_param_tls
|
|
; CHECK: %[[SB:.*]] = load i32,{{.*}}@__msan_param_tls
|
|
; CHECK: %[[SC:.*]] = or i32 %[[SA]], %[[SB]]
|
|
; CHECK: = fdiv float
|
|
; CHECK: store i32 %[[SC]], i32* {{.*}}@__msan_retval_tls
|
|
; CHECK: ret float
|
|
|
|
; Check that fneg simply propagates shadow.
|
|
|
|
define float @FNeg(float %a) nounwind uwtable readnone sanitize_memory {
|
|
entry:
|
|
%c = fneg float %a
|
|
ret float %c
|
|
}
|
|
|
|
; CHECK-LABEL: @FNeg
|
|
; CHECK: %[[SA:.*]] = load i32,{{.*}}@__msan_param_tls
|
|
; CHECK-ORIGINS: %[[SB:.*]] = load i32,{{.*}}@__msan_param_origin_tls
|
|
; CHECK: = fneg float
|
|
; CHECK: store i32 %[[SA]], i32* {{.*}}@__msan_retval_tls
|
|
; CHECK-ORIGINS: store i32{{.*}}@__msan_retval_origin_tls
|
|
; CHECK: ret float
|
|
|
|
; Check that we propagate shadow for x<0, x>=0, etc (i.e. sign bit tests)
|
|
|
|
define zeroext i1 @ICmpSLTZero(i32 %x) nounwind uwtable readnone sanitize_memory {
|
|
%1 = icmp slt i32 %x, 0
|
|
ret i1 %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ICmpSLTZero
|
|
; CHECK: icmp slt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp slt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret i1
|
|
|
|
define zeroext i1 @ICmpSGEZero(i32 %x) nounwind uwtable readnone sanitize_memory {
|
|
%1 = icmp sge i32 %x, 0
|
|
ret i1 %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ICmpSGEZero
|
|
; CHECK: icmp slt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp sge
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret i1
|
|
|
|
define zeroext i1 @ICmpSGTZero(i32 %x) nounwind uwtable readnone sanitize_memory {
|
|
%1 = icmp sgt i32 0, %x
|
|
ret i1 %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ICmpSGTZero
|
|
; CHECK: icmp slt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp sgt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret i1
|
|
|
|
define zeroext i1 @ICmpSLEZero(i32 %x) nounwind uwtable readnone sanitize_memory {
|
|
%1 = icmp sle i32 0, %x
|
|
ret i1 %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ICmpSLEZero
|
|
; CHECK: icmp slt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp sle
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret i1
|
|
|
|
|
|
; Check that we propagate shadow for x<=-1, x>-1, etc (i.e. sign bit tests)
|
|
|
|
define zeroext i1 @ICmpSLTAllOnes(i32 %x) nounwind uwtable readnone sanitize_memory {
|
|
%1 = icmp slt i32 -1, %x
|
|
ret i1 %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ICmpSLTAllOnes
|
|
; CHECK: icmp slt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp slt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret i1
|
|
|
|
define zeroext i1 @ICmpSGEAllOnes(i32 %x) nounwind uwtable readnone sanitize_memory {
|
|
%1 = icmp sge i32 -1, %x
|
|
ret i1 %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ICmpSGEAllOnes
|
|
; CHECK: icmp slt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp sge
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret i1
|
|
|
|
define zeroext i1 @ICmpSGTAllOnes(i32 %x) nounwind uwtable readnone sanitize_memory {
|
|
%1 = icmp sgt i32 %x, -1
|
|
ret i1 %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ICmpSGTAllOnes
|
|
; CHECK: icmp slt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp sgt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret i1
|
|
|
|
define zeroext i1 @ICmpSLEAllOnes(i32 %x) nounwind uwtable readnone sanitize_memory {
|
|
%1 = icmp sle i32 %x, -1
|
|
ret i1 %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ICmpSLEAllOnes
|
|
; CHECK: icmp slt
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp sle
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret i1
|
|
|
|
|
|
; Check that we propagate shadow for x<0, x>=0, etc (i.e. sign bit tests)
|
|
; of the vector arguments.
|
|
|
|
define <2 x i1> @ICmpSLT_vector_Zero(<2 x i32*> %x) nounwind uwtable readnone sanitize_memory {
|
|
%1 = icmp slt <2 x i32*> %x, zeroinitializer
|
|
ret <2 x i1> %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ICmpSLT_vector_Zero
|
|
; CHECK: icmp slt <2 x i64>
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp slt <2 x i32*>
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret <2 x i1>
|
|
|
|
; Check that we propagate shadow for x<=-1, x>0, etc (i.e. sign bit tests)
|
|
; of the vector arguments.
|
|
|
|
define <2 x i1> @ICmpSLT_vector_AllOnes(<2 x i32> %x) nounwind uwtable readnone sanitize_memory {
|
|
%1 = icmp slt <2 x i32> <i32 -1, i32 -1>, %x
|
|
ret <2 x i1> %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ICmpSLT_vector_AllOnes
|
|
; CHECK: icmp slt <2 x i32>
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp slt <2 x i32>
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret <2 x i1>
|
|
|
|
|
|
; Check that we propagate shadow for unsigned relational comparisons with
|
|
; constants
|
|
|
|
define zeroext i1 @ICmpUGTConst(i32 %x) nounwind uwtable readnone sanitize_memory {
|
|
entry:
|
|
%cmp = icmp ugt i32 %x, 7
|
|
ret i1 %cmp
|
|
}
|
|
|
|
; CHECK-LABEL: @ICmpUGTConst
|
|
; CHECK: icmp ugt i32
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp ugt i32
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: icmp ugt i32
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret i1
|
|
|
|
|
|
; Check that loads of shadow have the same alignment as the original loads.
|
|
; Check that loads of origin have the alignment of max(4, original alignment).
|
|
|
|
define i32 @ShadowLoadAlignmentLarge() nounwind uwtable sanitize_memory {
|
|
%y = alloca i32, align 64
|
|
%1 = load volatile i32, i32* %y, align 64
|
|
ret i32 %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ShadowLoadAlignmentLarge
|
|
; CHECK: load volatile i32, i32* {{.*}} align 64
|
|
; CHECK: load i32, i32* {{.*}} align 64
|
|
; CHECK: ret i32
|
|
|
|
define i32 @ShadowLoadAlignmentSmall() nounwind uwtable sanitize_memory {
|
|
%y = alloca i32, align 2
|
|
%1 = load volatile i32, i32* %y, align 2
|
|
ret i32 %1
|
|
}
|
|
|
|
; CHECK-LABEL: @ShadowLoadAlignmentSmall
|
|
; CHECK: load volatile i32, i32* {{.*}} align 2
|
|
; CHECK: load i32, i32* {{.*}} align 2
|
|
; CHECK-ORIGINS: load i32, i32* {{.*}} align 4
|
|
; CHECK: ret i32
|
|
|
|
|
|
; Test vector manipulation instructions.
|
|
; Check that the same bit manipulation is applied to the shadow values.
|
|
; Check that there is a zero test of the shadow of %idx argument, where present.
|
|
|
|
define i32 @ExtractElement(<4 x i32> %vec, i32 %idx) sanitize_memory {
|
|
%x = extractelement <4 x i32> %vec, i32 %idx
|
|
ret i32 %x
|
|
}
|
|
|
|
; CHECK-LABEL: @ExtractElement
|
|
; CHECK: extractelement
|
|
; CHECK: call void @__msan_warning_with_origin
|
|
; CHECK: extractelement
|
|
; CHECK: ret i32
|
|
|
|
define <4 x i32> @InsertElement(<4 x i32> %vec, i32 %idx, i32 %x) sanitize_memory {
|
|
%vec1 = insertelement <4 x i32> %vec, i32 %x, i32 %idx
|
|
ret <4 x i32> %vec1
|
|
}
|
|
|
|
; CHECK-LABEL: @InsertElement
|
|
; CHECK: insertelement
|
|
; CHECK: call void @__msan_warning_with_origin
|
|
; CHECK: insertelement
|
|
; CHECK: ret <4 x i32>
|
|
|
|
define <4 x i32> @ShuffleVector(<4 x i32> %vec, <4 x i32> %vec1) sanitize_memory {
|
|
%vec2 = shufflevector <4 x i32> %vec, <4 x i32> %vec1,
|
|
<4 x i32> <i32 0, i32 4, i32 1, i32 5>
|
|
ret <4 x i32> %vec2
|
|
}
|
|
|
|
; CHECK-LABEL: @ShuffleVector
|
|
; CHECK: shufflevector
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: shufflevector
|
|
; CHECK: ret <4 x i32>
|
|
|
|
|
|
; Test bswap intrinsic instrumentation
|
|
define i32 @BSwap(i32 %x) nounwind uwtable readnone sanitize_memory {
|
|
%y = tail call i32 @llvm.bswap.i32(i32 %x)
|
|
ret i32 %y
|
|
}
|
|
|
|
declare i32 @llvm.bswap.i32(i32) nounwind readnone
|
|
|
|
; CHECK-LABEL: @BSwap
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: @llvm.bswap.i32
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: @llvm.bswap.i32
|
|
; CHECK-NOT: call void @__msan_warning_with_origin
|
|
; CHECK: ret i32
|
|
|
|
; Test handling of vectors of pointers.
|
|
; Check that shadow of such vector is a vector of integers.
|
|
|
|
define <8 x i8*> @VectorOfPointers(<8 x i8*>* %p) nounwind uwtable sanitize_memory {
|
|
%x = load <8 x i8*>, <8 x i8*>* %p
|
|
ret <8 x i8*> %x
|
|
}
|
|
|
|
; CHECK-LABEL: @VectorOfPointers
|
|
; CHECK: load <8 x i8*>, <8 x i8*>*
|
|
; CHECK: load <8 x i64>, <8 x i64>*
|
|
; CHECK: store <8 x i64> {{.*}} @__msan_retval_tls
|
|
; CHECK: ret <8 x i8*>
|
|
|
|
; Test handling of va_copy.
|
|
|
|
declare void @llvm.va_copy(i8*, i8*) nounwind
|
|
|
|
define void @VACopy(i8* %p1, i8* %p2) nounwind uwtable sanitize_memory {
|
|
call void @llvm.va_copy(i8* %p1, i8* %p2) nounwind
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @VACopy
|
|
; CHECK: call void @llvm.memset.p0i8.i64({{.*}}, i8 0, i64 24, i1 false)
|
|
; CHECK: ret void
|
|
|
|
|
|
; Test that va_start instrumentation does not use va_arg_tls*.
|
|
; It should work with a local stack copy instead.
|
|
|
|
%struct.__va_list_tag = type { i32, i32, i8*, i8* }
|
|
declare void @llvm.va_start(i8*) nounwind
|
|
|
|
; Function Attrs: nounwind uwtable
|
|
define void @VAStart(i32 %x, ...) sanitize_memory {
|
|
entry:
|
|
%x.addr = alloca i32, align 4
|
|
%va = alloca [1 x %struct.__va_list_tag], align 16
|
|
store i32 %x, i32* %x.addr, align 4
|
|
%arraydecay = getelementptr inbounds [1 x %struct.__va_list_tag], [1 x %struct.__va_list_tag]* %va, i32 0, i32 0
|
|
%arraydecay1 = bitcast %struct.__va_list_tag* %arraydecay to i8*
|
|
call void @llvm.va_start(i8* %arraydecay1)
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @VAStart
|
|
; CHECK: call void @llvm.va_start
|
|
; CHECK-NOT: @__msan_va_arg_tls
|
|
; CHECK-NOT: @__msan_va_arg_overflow_size_tls
|
|
; CHECK: ret void
|
|
|
|
|
|
; Test handling of volatile stores.
|
|
; Check that MemorySanitizer does not add a check of the value being stored.
|
|
|
|
define void @VolatileStore(i32* nocapture %p, i32 %x) nounwind uwtable sanitize_memory {
|
|
entry:
|
|
store volatile i32 %x, i32* %p, align 4
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @VolatileStore
|
|
; CHECK-NOT: @__msan_warning_with_origin
|
|
; CHECK: ret void
|
|
|
|
|
|
; Test that checks are omitted and returned value is always initialized if
|
|
; sanitize_memory attribute is missing.
|
|
|
|
define i32 @NoSanitizeMemory(i32 %x) uwtable {
|
|
entry:
|
|
%tobool = icmp eq i32 %x, 0
|
|
br i1 %tobool, label %if.end, label %if.then
|
|
|
|
if.then: ; preds = %entry
|
|
tail call void @bar()
|
|
br label %if.end
|
|
|
|
if.end: ; preds = %entry, %if.then
|
|
ret i32 %x
|
|
}
|
|
|
|
declare void @bar()
|
|
|
|
; CHECK-LABEL: @NoSanitizeMemory
|
|
; CHECK-NOT: @__msan_warning_with_origin
|
|
; CHECK: store i32 0, {{.*}} @__msan_retval_tls
|
|
; CHECK-NOT: @__msan_warning_with_origin
|
|
; CHECK: ret i32
|
|
|
|
|
|
; Test that stack allocations are unpoisoned in functions missing
|
|
; sanitize_memory attribute
|
|
|
|
define i32 @NoSanitizeMemoryAlloca() {
|
|
entry:
|
|
%p = alloca i32, align 4
|
|
%x = call i32 @NoSanitizeMemoryAllocaHelper(i32* %p)
|
|
ret i32 %x
|
|
}
|
|
|
|
declare i32 @NoSanitizeMemoryAllocaHelper(i32* %p)
|
|
|
|
; CHECK-LABEL: @NoSanitizeMemoryAlloca
|
|
; CHECK: call void @llvm.memset.p0i8.i64(i8* align 4 {{.*}}, i8 0, i64 4, i1 false)
|
|
; CHECK: call i32 @NoSanitizeMemoryAllocaHelper(i32*
|
|
; CHECK: ret i32
|
|
|
|
|
|
; Test that undef is unpoisoned in functions missing
|
|
; sanitize_memory attribute
|
|
|
|
define i32 @NoSanitizeMemoryUndef() {
|
|
entry:
|
|
%x = call i32 @NoSanitizeMemoryUndefHelper(i32 undef)
|
|
ret i32 %x
|
|
}
|
|
|
|
declare i32 @NoSanitizeMemoryUndefHelper(i32 %x)
|
|
|
|
; CHECK-LABEL: @NoSanitizeMemoryUndef
|
|
; CHECK: store i32 0, i32* {{.*}} @__msan_param_tls
|
|
; CHECK: call i32 @NoSanitizeMemoryUndefHelper(i32 undef)
|
|
; CHECK: ret i32
|
|
|
|
|
|
; Test PHINode instrumentation in ignorelisted functions
|
|
|
|
define i32 @NoSanitizeMemoryPHI(i32 %x) {
|
|
entry:
|
|
%tobool = icmp ne i32 %x, 0
|
|
br i1 %tobool, label %cond.true, label %cond.false
|
|
|
|
cond.true: ; preds = %entry
|
|
br label %cond.end
|
|
|
|
cond.false: ; preds = %entry
|
|
br label %cond.end
|
|
|
|
cond.end: ; preds = %cond.false, %cond.true
|
|
%cond = phi i32 [ undef, %cond.true ], [ undef, %cond.false ]
|
|
ret i32 %cond
|
|
}
|
|
|
|
; CHECK: [[A:%.*]] = phi i32 [ undef, %cond.true ], [ undef, %cond.false ]
|
|
; CHECK: store i32 0, i32* bitcast {{.*}} @__msan_retval_tls
|
|
; CHECK: ret i32 [[A]]
|
|
|
|
|
|
; Test that there are no __msan_param_origin_tls stores when
|
|
; argument shadow is a compile-time zero constant (which is always the case
|
|
; in functions missing sanitize_memory attribute).
|
|
|
|
define i32 @NoSanitizeMemoryParamTLS(i32* nocapture readonly %x) {
|
|
entry:
|
|
%0 = load i32, i32* %x, align 4
|
|
%call = tail call i32 @NoSanitizeMemoryParamTLSHelper(i32 %0)
|
|
ret i32 %call
|
|
}
|
|
|
|
declare i32 @NoSanitizeMemoryParamTLSHelper(i32 %x)
|
|
|
|
; CHECK-LABEL: define i32 @NoSanitizeMemoryParamTLS(
|
|
; CHECK-NOT: __msan_param_origin_tls
|
|
; CHECK: ret i32
|
|
|
|
|
|
; Test argument shadow alignment
|
|
|
|
define <2 x i64> @ArgumentShadowAlignment(i64 %a, <2 x i64> %b) sanitize_memory {
|
|
entry:
|
|
ret <2 x i64> %b
|
|
}
|
|
|
|
; CHECK-LABEL: @ArgumentShadowAlignment
|
|
; CHECK: load <2 x i64>, <2 x i64>* {{.*}} @__msan_param_tls {{.*}}, align 8
|
|
; CHECK: store <2 x i64> {{.*}} @__msan_retval_tls {{.*}}, align 8
|
|
; CHECK: ret <2 x i64>
|
|
|
|
|
|
; Test origin propagation for insertvalue
|
|
|
|
define { i64, i32 } @make_pair_64_32(i64 %x, i32 %y) sanitize_memory {
|
|
entry:
|
|
%a = insertvalue { i64, i32 } undef, i64 %x, 0
|
|
%b = insertvalue { i64, i32 } %a, i32 %y, 1
|
|
ret { i64, i32 } %b
|
|
}
|
|
|
|
; CHECK-ORIGINS: @make_pair_64_32
|
|
; First element shadow
|
|
; CHECK-ORIGINS: insertvalue { i64, i32 } { i64 -1, i32 -1 }, i64 {{.*}}, 0
|
|
; First element origin
|
|
; CHECK-ORIGINS: icmp ne i64
|
|
; CHECK-ORIGINS: select i1
|
|
; First element app value
|
|
; CHECK-ORIGINS: insertvalue { i64, i32 } undef, i64 {{.*}}, 0
|
|
; Second element shadow
|
|
; CHECK-ORIGINS: insertvalue { i64, i32 } {{.*}}, i32 {{.*}}, 1
|
|
; Second element origin
|
|
; CHECK-ORIGINS: icmp ne i32
|
|
; CHECK-ORIGINS: select i1
|
|
; Second element app value
|
|
; CHECK-ORIGINS: insertvalue { i64, i32 } {{.*}}, i32 {{.*}}, 1
|
|
; CHECK-ORIGINS: ret { i64, i32 }
|
|
|
|
|
|
; Test shadow propagation for aggregates passed through ellipsis.
|
|
|
|
%struct.StructByVal = type { i32, i32, i32, i32 }
|
|
|
|
declare void @VAArgStructFn(i32 %guard, ...)
|
|
|
|
define void @VAArgStruct(%struct.StructByVal* nocapture %s) sanitize_memory {
|
|
entry:
|
|
%agg.tmp2 = alloca %struct.StructByVal, align 8
|
|
%0 = bitcast %struct.StructByVal* %s to i8*
|
|
%agg.tmp.sroa.0.0..sroa_cast = bitcast %struct.StructByVal* %s to i64*
|
|
%agg.tmp.sroa.0.0.copyload = load i64, i64* %agg.tmp.sroa.0.0..sroa_cast, align 4
|
|
%agg.tmp.sroa.2.0..sroa_idx = getelementptr inbounds %struct.StructByVal, %struct.StructByVal* %s, i64 0, i32 2
|
|
%agg.tmp.sroa.2.0..sroa_cast = bitcast i32* %agg.tmp.sroa.2.0..sroa_idx to i64*
|
|
%agg.tmp.sroa.2.0.copyload = load i64, i64* %agg.tmp.sroa.2.0..sroa_cast, align 4
|
|
%1 = bitcast %struct.StructByVal* %agg.tmp2 to i8*
|
|
call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 4 %1, i8* align 4 %0, i64 16, i1 false)
|
|
call void (i32, ...) @VAArgStructFn(i32 undef, i64 %agg.tmp.sroa.0.0.copyload, i64 %agg.tmp.sroa.2.0.copyload, i64 %agg.tmp.sroa.0.0.copyload, i64 %agg.tmp.sroa.2.0.copyload, %struct.StructByVal* byval(%struct.StructByVal) align 8 %agg.tmp2)
|
|
ret void
|
|
}
|
|
|
|
; "undef" and the first 2 structs go to general purpose registers;
|
|
; the third struct goes to the overflow area byval
|
|
|
|
; CHECK-LABEL: @VAArgStruct
|
|
; undef not stored to __msan_va_arg_tls - it's a fixed argument
|
|
; first struct through general purpose registers
|
|
; CHECK: store i64 {{.*}}, i64* {{.*}}@__msan_va_arg_tls{{.*}}, i64 8){{.*}}, align 8
|
|
; CHECK: store i64 {{.*}}, i64* {{.*}}@__msan_va_arg_tls{{.*}}, i64 16){{.*}}, align 8
|
|
; second struct through general purpose registers
|
|
; CHECK: store i64 {{.*}}, i64* {{.*}}@__msan_va_arg_tls{{.*}}, i64 24){{.*}}, align 8
|
|
; CHECK: store i64 {{.*}}, i64* {{.*}}@__msan_va_arg_tls{{.*}}, i64 32){{.*}}, align 8
|
|
; third struct through the overflow area byval
|
|
; CHECK: ptrtoint %struct.StructByVal* {{.*}} to i64
|
|
; CHECK: call void @llvm.memcpy.p0i8.p0i8.i64{{.*}}@__msan_va_arg_tls {{.*}}, i64 176
|
|
; CHECK: store i64 16, i64* @__msan_va_arg_overflow_size_tls
|
|
; CHECK: call void (i32, ...) @VAArgStructFn
|
|
; CHECK: ret void
|
|
|
|
; Same code compiled without SSE (see attributes below).
|
|
; The register save area is only 48 bytes instead of 176.
|
|
define void @VAArgStructNoSSE(%struct.StructByVal* nocapture %s) sanitize_memory #0 {
|
|
entry:
|
|
%agg.tmp2 = alloca %struct.StructByVal, align 8
|
|
%0 = bitcast %struct.StructByVal* %s to i8*
|
|
%agg.tmp.sroa.0.0..sroa_cast = bitcast %struct.StructByVal* %s to i64*
|
|
%agg.tmp.sroa.0.0.copyload = load i64, i64* %agg.tmp.sroa.0.0..sroa_cast, align 4
|
|
%agg.tmp.sroa.2.0..sroa_idx = getelementptr inbounds %struct.StructByVal, %struct.StructByVal* %s, i64 0, i32 2
|
|
%agg.tmp.sroa.2.0..sroa_cast = bitcast i32* %agg.tmp.sroa.2.0..sroa_idx to i64*
|
|
%agg.tmp.sroa.2.0.copyload = load i64, i64* %agg.tmp.sroa.2.0..sroa_cast, align 4
|
|
%1 = bitcast %struct.StructByVal* %agg.tmp2 to i8*
|
|
call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 4 %1, i8* align 4 %0, i64 16, i1 false)
|
|
call void (i32, ...) @VAArgStructFn(i32 undef, i64 %agg.tmp.sroa.0.0.copyload, i64 %agg.tmp.sroa.2.0.copyload, i64 %agg.tmp.sroa.0.0.copyload, i64 %agg.tmp.sroa.2.0.copyload, %struct.StructByVal* byval(%struct.StructByVal) align 8 %agg.tmp2)
|
|
ret void
|
|
}
|
|
|
|
attributes #0 = { "target-features"="+fxsr,+x87,-sse" }
|
|
|
|
; CHECK: call void @llvm.memcpy.p0i8.p0i8.i64{{.*}}@__msan_va_arg_tls {{.*}}, i64 48
|
|
|
|
declare i32 @InnerTailCall(i32 %a)
|
|
|
|
define void @MismatchedReturnTypeTailCall(i32 %a) sanitize_memory {
|
|
%b = tail call i32 @InnerTailCall(i32 %a)
|
|
ret void
|
|
}
|
|
|
|
; We used to strip off the 'tail' modifier, but now that we unpoison return slot
|
|
; shadow before the call, we don't need to anymore.
|
|
|
|
; CHECK-LABEL: define void @MismatchedReturnTypeTailCall
|
|
; CHECK: tail call i32 @InnerTailCall
|
|
; CHECK: ret void
|
|
|
|
|
|
declare i32 @MustTailCall(i32 %a)
|
|
|
|
define i32 @CallMustTailCall(i32 %a) sanitize_memory {
|
|
%b = musttail call i32 @MustTailCall(i32 %a)
|
|
ret i32 %b
|
|
}
|
|
|
|
; For "musttail" calls we can not insert any shadow manipulating code between
|
|
; call and the return instruction. And we don't need to, because everything is
|
|
; taken care of in the callee.
|
|
|
|
; CHECK-LABEL: define i32 @CallMustTailCall
|
|
; CHECK: musttail call i32 @MustTailCall
|
|
; No instrumentation between call and ret.
|
|
; CHECK-NEXT: ret i32
|
|
|
|
declare i32* @MismatchingMustTailCall(i32 %a)
|
|
|
|
define i8* @MismatchingCallMustTailCall(i32 %a) sanitize_memory {
|
|
%b = musttail call i32* @MismatchingMustTailCall(i32 %a)
|
|
%c = bitcast i32* %b to i8*
|
|
ret i8* %c
|
|
}
|
|
|
|
; For "musttail" calls we can not insert any shadow manipulating code between
|
|
; call and the return instruction. And we don't need to, because everything is
|
|
; taken care of in the callee.
|
|
|
|
; CHECK-LABEL: define i8* @MismatchingCallMustTailCall
|
|
; CHECK: musttail call i32* @MismatchingMustTailCall
|
|
; No instrumentation between call and ret.
|
|
; CHECK-NEXT: bitcast i32* {{.*}} to i8*
|
|
; CHECK-NEXT: ret i8*
|
|
|
|
|
|
; CHECK-LABEL: define internal void @msan.module_ctor() #[[#ATTR:]] {
|
|
; CHECK: call void @__msan_init()
|
|
|
|
; CHECK-CALLS: declare void @__msan_maybe_warning_1(i8 zeroext, i32 zeroext)
|
|
; CHECK-CALLS: declare void @__msan_maybe_store_origin_1(i8 zeroext, i8*, i32 zeroext)
|
|
; CHECK-CALLS: declare void @__msan_maybe_warning_2(i16 zeroext, i32 zeroext)
|
|
; CHECK-CALLS: declare void @__msan_maybe_store_origin_2(i16 zeroext, i8*, i32 zeroext)
|
|
; CHECK-CALLS: declare void @__msan_maybe_warning_4(i32 zeroext, i32 zeroext)
|
|
; CHECK-CALLS: declare void @__msan_maybe_store_origin_4(i32 zeroext, i8*, i32 zeroext)
|
|
; CHECK-CALLS: declare void @__msan_maybe_warning_8(i64 zeroext, i32 zeroext)
|
|
; CHECK-CALLS: declare void @__msan_maybe_store_origin_8(i64 zeroext, i8*, i32 zeroext)
|
|
|
|
; CHECK: attributes #[[#ATTR]] = { nounwind }
|