llvm-project/llvm/test/CodeGen/X86/memset-zero.ll

827 lines
23 KiB
LLVM

; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s -mtriple=i686-unknown-linux | FileCheck %s --check-prefix=X86
; RUN: llc < %s -mtriple=x86_64-unknown-linux -mcpu=core2 | FileCheck %s --check-prefix=CORE2
; RUN: llc < %s -mtriple=x86_64-unknown-linux -mcpu=nehalem | FileCheck %s --check-prefix=NEHALEM
; RUN: llc < %s -mtriple=x86_64-unknown-linux -mcpu=sandybridge | FileCheck %s --check-prefix=SANDYBRIDGE
; RUN: llc < %s -mtriple=x86_64-unknown-linux -mcpu=skylake | FileCheck %s --check-prefix=SKYLAKE
; RUN: llc < %s -mtriple=x86_64-unknown-linux -mcpu=knl | FileCheck %s --check-prefix=KNL
declare void @llvm.memset.p0i8.i64(i8* nocapture, i8, i64, i1) nounwind
define void @memset_0(i8* %a) nounwind {
; X86-LABEL: memset_0:
; X86: # %bb.0: # %entry
; X86-NEXT: retl
;
; CORE2-LABEL: memset_0:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_0:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_0:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_0:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_0:
; KNL: # %bb.0: # %entry
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 0, i1 false)
ret void
}
define void @memset_4(i8* %a) nounwind {
; X86-LABEL: memset_4:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_4:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movl $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_4:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: movl $0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_4:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: movl $0, (%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_4:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: movl $0, (%rdi)
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_4:
; KNL: # %bb.0: # %entry
; KNL-NEXT: movl $0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 4, i1 false)
ret void
}
define void @memset_5(i8* %a) nounwind {
; X86-LABEL: memset_5:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movb $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_5:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movb $0, 4(%rdi)
; CORE2-NEXT: movl $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_5:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: movb $0, 4(%rdi)
; NEHALEM-NEXT: movl $0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_5:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: movb $0, 4(%rdi)
; SANDYBRIDGE-NEXT: movl $0, (%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_5:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: movb $0, 4(%rdi)
; SKYLAKE-NEXT: movl $0, (%rdi)
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_5:
; KNL: # %bb.0: # %entry
; KNL-NEXT: movb $0, 4(%rdi)
; KNL-NEXT: movl $0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 5, i1 false)
ret void
}
define void @memset_7(i8* %a) nounwind {
; X86-LABEL: memset_7:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 3(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_7:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movl $0, 3(%rdi)
; CORE2-NEXT: movl $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_7:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: movl $0, 3(%rdi)
; NEHALEM-NEXT: movl $0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_7:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: movl $0, 3(%rdi)
; SANDYBRIDGE-NEXT: movl $0, (%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_7:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: movl $0, 3(%rdi)
; SKYLAKE-NEXT: movl $0, (%rdi)
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_7:
; KNL: # %bb.0: # %entry
; KNL-NEXT: movl $0, 3(%rdi)
; KNL-NEXT: movl $0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 7, i1 false)
ret void
}
define void @memset_8(i8* %a) nounwind {
; X86-LABEL: memset_8:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_8:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movq $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_8:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: movq $0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_8:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: movq $0, (%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_8:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: movq $0, (%rdi)
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_8:
; KNL: # %bb.0: # %entry
; KNL-NEXT: movq $0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 8, i1 false)
ret void
}
define void @memset_11(i8* %a) nounwind {
; X86-LABEL: memset_11:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 7(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_11:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movl $0, 7(%rdi)
; CORE2-NEXT: movq $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_11:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: movl $0, 7(%rdi)
; NEHALEM-NEXT: movq $0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_11:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: movl $0, 7(%rdi)
; SANDYBRIDGE-NEXT: movq $0, (%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_11:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: movl $0, 7(%rdi)
; SKYLAKE-NEXT: movq $0, (%rdi)
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_11:
; KNL: # %bb.0: # %entry
; KNL-NEXT: movl $0, 7(%rdi)
; KNL-NEXT: movq $0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 11, i1 false)
ret void
}
define void @memset_13(i8* %a) nounwind {
; X86-LABEL: memset_13:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movb $0, 12(%eax)
; X86-NEXT: movl $0, 8(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_13:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movq $0, 5(%rdi)
; CORE2-NEXT: movq $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_13:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: movq $0, 5(%rdi)
; NEHALEM-NEXT: movq $0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_13:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: movq $0, 5(%rdi)
; SANDYBRIDGE-NEXT: movq $0, (%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_13:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: movq $0, 5(%rdi)
; SKYLAKE-NEXT: movq $0, (%rdi)
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_13:
; KNL: # %bb.0: # %entry
; KNL-NEXT: movq $0, 5(%rdi)
; KNL-NEXT: movq $0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 13, i1 false)
ret void
}
define void @memset_15(i8* %a) nounwind {
; X86-LABEL: memset_15:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 11(%eax)
; X86-NEXT: movl $0, 8(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_15:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movq $0, 7(%rdi)
; CORE2-NEXT: movq $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_15:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: movq $0, 7(%rdi)
; NEHALEM-NEXT: movq $0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_15:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: movq $0, 7(%rdi)
; SANDYBRIDGE-NEXT: movq $0, (%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_15:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: movq $0, 7(%rdi)
; SKYLAKE-NEXT: movq $0, (%rdi)
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_15:
; KNL: # %bb.0: # %entry
; KNL-NEXT: movq $0, 7(%rdi)
; KNL-NEXT: movq $0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 15, i1 false)
ret void
}
define void @memset_16(i8* %a) nounwind {
; X86-LABEL: memset_16:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 12(%eax)
; X86-NEXT: movl $0, 8(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_16:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movq $0, 8(%rdi)
; CORE2-NEXT: movq $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_16:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: xorps %xmm0, %xmm0
; NEHALEM-NEXT: movups %xmm0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_16:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SANDYBRIDGE-NEXT: vmovups %xmm0, (%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_16:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SKYLAKE-NEXT: vmovups %xmm0, (%rdi)
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_16:
; KNL: # %bb.0: # %entry
; KNL-NEXT: vxorps %xmm0, %xmm0, %xmm0
; KNL-NEXT: vmovups %xmm0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 16, i1 false)
ret void
}
define void @memset_17(i8* %a) nounwind {
; X86-LABEL: memset_17:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movb $0, 16(%eax)
; X86-NEXT: movl $0, 12(%eax)
; X86-NEXT: movl $0, 8(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_17:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movb $0, 16(%rdi)
; CORE2-NEXT: movq $0, 8(%rdi)
; CORE2-NEXT: movq $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_17:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: xorps %xmm0, %xmm0
; NEHALEM-NEXT: movups %xmm0, (%rdi)
; NEHALEM-NEXT: movb $0, 16(%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_17:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SANDYBRIDGE-NEXT: vmovups %xmm0, (%rdi)
; SANDYBRIDGE-NEXT: movb $0, 16(%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_17:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SKYLAKE-NEXT: vmovups %xmm0, (%rdi)
; SKYLAKE-NEXT: movb $0, 16(%rdi)
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_17:
; KNL: # %bb.0: # %entry
; KNL-NEXT: vxorps %xmm0, %xmm0, %xmm0
; KNL-NEXT: vmovups %xmm0, (%rdi)
; KNL-NEXT: movb $0, 16(%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 17, i1 false)
ret void
}
define void @memset_19(i8* %a) nounwind {
; X86-LABEL: memset_19:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 15(%eax)
; X86-NEXT: movl $0, 12(%eax)
; X86-NEXT: movl $0, 8(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_19:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movl $0, 15(%rdi)
; CORE2-NEXT: movq $0, 8(%rdi)
; CORE2-NEXT: movq $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_19:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: xorps %xmm0, %xmm0
; NEHALEM-NEXT: movups %xmm0, (%rdi)
; NEHALEM-NEXT: movl $0, 15(%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_19:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SANDYBRIDGE-NEXT: vmovups %xmm0, (%rdi)
; SANDYBRIDGE-NEXT: movl $0, 15(%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_19:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SKYLAKE-NEXT: vmovups %xmm0, (%rdi)
; SKYLAKE-NEXT: movl $0, 15(%rdi)
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_19:
; KNL: # %bb.0: # %entry
; KNL-NEXT: vxorps %xmm0, %xmm0, %xmm0
; KNL-NEXT: vmovups %xmm0, (%rdi)
; KNL-NEXT: movl $0, 15(%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 19, i1 false)
ret void
}
define void @memset_31(i8* %a) nounwind {
; X86-LABEL: memset_31:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 27(%eax)
; X86-NEXT: movl $0, 24(%eax)
; X86-NEXT: movl $0, 20(%eax)
; X86-NEXT: movl $0, 16(%eax)
; X86-NEXT: movl $0, 12(%eax)
; X86-NEXT: movl $0, 8(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_31:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movq $0, 23(%rdi)
; CORE2-NEXT: movq $0, 16(%rdi)
; CORE2-NEXT: movq $0, 8(%rdi)
; CORE2-NEXT: movq $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_31:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: xorps %xmm0, %xmm0
; NEHALEM-NEXT: movups %xmm0, 15(%rdi)
; NEHALEM-NEXT: movups %xmm0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_31:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SANDYBRIDGE-NEXT: vmovups %xmm0, 15(%rdi)
; SANDYBRIDGE-NEXT: vmovups %xmm0, (%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_31:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SKYLAKE-NEXT: vmovups %xmm0, 15(%rdi)
; SKYLAKE-NEXT: vmovups %xmm0, (%rdi)
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_31:
; KNL: # %bb.0: # %entry
; KNL-NEXT: vxorps %xmm0, %xmm0, %xmm0
; KNL-NEXT: vmovups %xmm0, 15(%rdi)
; KNL-NEXT: vmovups %xmm0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 31, i1 false)
ret void
}
define void @memset_32(i8* %a) nounwind {
; X86-LABEL: memset_32:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 28(%eax)
; X86-NEXT: movl $0, 24(%eax)
; X86-NEXT: movl $0, 20(%eax)
; X86-NEXT: movl $0, 16(%eax)
; X86-NEXT: movl $0, 12(%eax)
; X86-NEXT: movl $0, 8(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_32:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movq $0, 24(%rdi)
; CORE2-NEXT: movq $0, 16(%rdi)
; CORE2-NEXT: movq $0, 8(%rdi)
; CORE2-NEXT: movq $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_32:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: xorps %xmm0, %xmm0
; NEHALEM-NEXT: movups %xmm0, 16(%rdi)
; NEHALEM-NEXT: movups %xmm0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_32:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SANDYBRIDGE-NEXT: vmovups %xmm0, 16(%rdi)
; SANDYBRIDGE-NEXT: vmovups %xmm0, (%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_32:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SKYLAKE-NEXT: vmovups %ymm0, (%rdi)
; SKYLAKE-NEXT: vzeroupper
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_32:
; KNL: # %bb.0: # %entry
; KNL-NEXT: vxorps %xmm0, %xmm0, %xmm0
; KNL-NEXT: vmovups %ymm0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 32, i1 false)
ret void
}
define void @memset_32_align32(i8* %a) nounwind {
; X86-LABEL: memset_32_align32:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 28(%eax)
; X86-NEXT: movl $0, 24(%eax)
; X86-NEXT: movl $0, 20(%eax)
; X86-NEXT: movl $0, 16(%eax)
; X86-NEXT: movl $0, 12(%eax)
; X86-NEXT: movl $0, 8(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_32_align32:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: xorps %xmm0, %xmm0
; CORE2-NEXT: movaps %xmm0, 16(%rdi)
; CORE2-NEXT: movaps %xmm0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_32_align32:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: xorps %xmm0, %xmm0
; NEHALEM-NEXT: movaps %xmm0, 16(%rdi)
; NEHALEM-NEXT: movaps %xmm0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_32_align32:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SANDYBRIDGE-NEXT: vmovaps %ymm0, (%rdi)
; SANDYBRIDGE-NEXT: vzeroupper
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_32_align32:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SKYLAKE-NEXT: vmovaps %ymm0, (%rdi)
; SKYLAKE-NEXT: vzeroupper
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_32_align32:
; KNL: # %bb.0: # %entry
; KNL-NEXT: vxorps %xmm0, %xmm0, %xmm0
; KNL-NEXT: vmovaps %ymm0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* align 32 %a, i8 0, i64 32, i1 false)
ret void
}
define void @memset_35(i8* %a) nounwind {
; X86-LABEL: memset_35:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 31(%eax)
; X86-NEXT: movl $0, 28(%eax)
; X86-NEXT: movl $0, 24(%eax)
; X86-NEXT: movl $0, 20(%eax)
; X86-NEXT: movl $0, 16(%eax)
; X86-NEXT: movl $0, 12(%eax)
; X86-NEXT: movl $0, 8(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_35:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movl $0, 31(%rdi)
; CORE2-NEXT: movq $0, 24(%rdi)
; CORE2-NEXT: movq $0, 16(%rdi)
; CORE2-NEXT: movq $0, 8(%rdi)
; CORE2-NEXT: movq $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_35:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: xorps %xmm0, %xmm0
; NEHALEM-NEXT: movups %xmm0, 16(%rdi)
; NEHALEM-NEXT: movups %xmm0, (%rdi)
; NEHALEM-NEXT: movl $0, 31(%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_35:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SANDYBRIDGE-NEXT: vmovups %xmm0, 16(%rdi)
; SANDYBRIDGE-NEXT: vmovups %xmm0, (%rdi)
; SANDYBRIDGE-NEXT: movl $0, 31(%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_35:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SKYLAKE-NEXT: vmovups %ymm0, (%rdi)
; SKYLAKE-NEXT: movl $0, 31(%rdi)
; SKYLAKE-NEXT: vzeroupper
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_35:
; KNL: # %bb.0: # %entry
; KNL-NEXT: vxorps %xmm0, %xmm0, %xmm0
; KNL-NEXT: vmovups %ymm0, (%rdi)
; KNL-NEXT: movl $0, 31(%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 35, i1 false)
ret void
}
define void @memset_64(i8* %a) nounwind {
; X86-LABEL: memset_64:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 60(%eax)
; X86-NEXT: movl $0, 56(%eax)
; X86-NEXT: movl $0, 52(%eax)
; X86-NEXT: movl $0, 48(%eax)
; X86-NEXT: movl $0, 44(%eax)
; X86-NEXT: movl $0, 40(%eax)
; X86-NEXT: movl $0, 36(%eax)
; X86-NEXT: movl $0, 32(%eax)
; X86-NEXT: movl $0, 28(%eax)
; X86-NEXT: movl $0, 24(%eax)
; X86-NEXT: movl $0, 20(%eax)
; X86-NEXT: movl $0, 16(%eax)
; X86-NEXT: movl $0, 12(%eax)
; X86-NEXT: movl $0, 8(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_64:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: movq $0, 56(%rdi)
; CORE2-NEXT: movq $0, 48(%rdi)
; CORE2-NEXT: movq $0, 40(%rdi)
; CORE2-NEXT: movq $0, 32(%rdi)
; CORE2-NEXT: movq $0, 24(%rdi)
; CORE2-NEXT: movq $0, 16(%rdi)
; CORE2-NEXT: movq $0, 8(%rdi)
; CORE2-NEXT: movq $0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_64:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: xorps %xmm0, %xmm0
; NEHALEM-NEXT: movups %xmm0, 48(%rdi)
; NEHALEM-NEXT: movups %xmm0, 32(%rdi)
; NEHALEM-NEXT: movups %xmm0, 16(%rdi)
; NEHALEM-NEXT: movups %xmm0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_64:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SANDYBRIDGE-NEXT: vmovups %xmm0, 16(%rdi)
; SANDYBRIDGE-NEXT: vmovups %xmm0, (%rdi)
; SANDYBRIDGE-NEXT: vmovups %xmm0, 48(%rdi)
; SANDYBRIDGE-NEXT: vmovups %xmm0, 32(%rdi)
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_64:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SKYLAKE-NEXT: vmovups %ymm0, 32(%rdi)
; SKYLAKE-NEXT: vmovups %ymm0, (%rdi)
; SKYLAKE-NEXT: vzeroupper
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_64:
; KNL: # %bb.0: # %entry
; KNL-NEXT: vxorps %xmm0, %xmm0, %xmm0
; KNL-NEXT: vmovups %zmm0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* %a, i8 0, i64 64, i1 false)
ret void
}
define void @memset_64_align64(i8* %a) nounwind {
; X86-LABEL: memset_64_align64:
; X86: # %bb.0: # %entry
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl $0, 60(%eax)
; X86-NEXT: movl $0, 56(%eax)
; X86-NEXT: movl $0, 52(%eax)
; X86-NEXT: movl $0, 48(%eax)
; X86-NEXT: movl $0, 44(%eax)
; X86-NEXT: movl $0, 40(%eax)
; X86-NEXT: movl $0, 36(%eax)
; X86-NEXT: movl $0, 32(%eax)
; X86-NEXT: movl $0, 28(%eax)
; X86-NEXT: movl $0, 24(%eax)
; X86-NEXT: movl $0, 20(%eax)
; X86-NEXT: movl $0, 16(%eax)
; X86-NEXT: movl $0, 12(%eax)
; X86-NEXT: movl $0, 8(%eax)
; X86-NEXT: movl $0, 4(%eax)
; X86-NEXT: movl $0, (%eax)
; X86-NEXT: retl
;
; CORE2-LABEL: memset_64_align64:
; CORE2: # %bb.0: # %entry
; CORE2-NEXT: xorps %xmm0, %xmm0
; CORE2-NEXT: movaps %xmm0, 48(%rdi)
; CORE2-NEXT: movaps %xmm0, 32(%rdi)
; CORE2-NEXT: movaps %xmm0, 16(%rdi)
; CORE2-NEXT: movaps %xmm0, (%rdi)
; CORE2-NEXT: retq
;
; NEHALEM-LABEL: memset_64_align64:
; NEHALEM: # %bb.0: # %entry
; NEHALEM-NEXT: xorps %xmm0, %xmm0
; NEHALEM-NEXT: movaps %xmm0, 48(%rdi)
; NEHALEM-NEXT: movaps %xmm0, 32(%rdi)
; NEHALEM-NEXT: movaps %xmm0, 16(%rdi)
; NEHALEM-NEXT: movaps %xmm0, (%rdi)
; NEHALEM-NEXT: retq
;
; SANDYBRIDGE-LABEL: memset_64_align64:
; SANDYBRIDGE: # %bb.0: # %entry
; SANDYBRIDGE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SANDYBRIDGE-NEXT: vmovaps %ymm0, 32(%rdi)
; SANDYBRIDGE-NEXT: vmovaps %ymm0, (%rdi)
; SANDYBRIDGE-NEXT: vzeroupper
; SANDYBRIDGE-NEXT: retq
;
; SKYLAKE-LABEL: memset_64_align64:
; SKYLAKE: # %bb.0: # %entry
; SKYLAKE-NEXT: vxorps %xmm0, %xmm0, %xmm0
; SKYLAKE-NEXT: vmovaps %ymm0, 32(%rdi)
; SKYLAKE-NEXT: vmovaps %ymm0, (%rdi)
; SKYLAKE-NEXT: vzeroupper
; SKYLAKE-NEXT: retq
;
; KNL-LABEL: memset_64_align64:
; KNL: # %bb.0: # %entry
; KNL-NEXT: vxorps %xmm0, %xmm0, %xmm0
; KNL-NEXT: vmovaps %zmm0, (%rdi)
; KNL-NEXT: retq
entry:
call void @llvm.memset.p0i8.i64(i8* align 64 %a, i8 0, i64 64, i1 false)
ret void
}