forked from OSchip/llvm-project
[SLPVectorizer][X86] Added vXi8/vXi16 sitofp/uitofp tests
Dropped useless 2i32-2f32 test llvm-svn: 277281
This commit is contained in:
parent
5e0d6b509a
commit
9e201eac32
|
@ -123,6 +123,118 @@ define void @sitofp_4i32_4f64() #0 {
|
|||
ret void
|
||||
}
|
||||
|
||||
define void @sitofp_2i16_2f64() #0 {
|
||||
; CHECK-LABEL: @sitofp_2i16_2f64(
|
||||
; CHECK-NEXT: [[LD0:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: [[LD1:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: [[CVT0:%.*]] = sitofp i16 [[LD0]] to double
|
||||
; CHECK-NEXT: [[CVT1:%.*]] = sitofp i16 [[LD1]] to double
|
||||
; CHECK-NEXT: store double [[CVT0]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: store double [[CVT1]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
%cvt0 = sitofp i16 %ld0 to double
|
||||
%cvt1 = sitofp i16 %ld1 to double
|
||||
store double %cvt0, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
store double %cvt1, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @sitofp_4i16_4f64() #0 {
|
||||
; SSE-LABEL: @sitofp_4i16_4f64(
|
||||
; SSE-NEXT: [[LD0:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
; SSE-NEXT: [[LD1:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
; SSE-NEXT: [[LD2:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 2), align 8
|
||||
; SSE-NEXT: [[LD3:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 3), align 8
|
||||
; SSE-NEXT: [[CVT0:%.*]] = sitofp i16 [[LD0]] to double
|
||||
; SSE-NEXT: [[CVT1:%.*]] = sitofp i16 [[LD1]] to double
|
||||
; SSE-NEXT: [[CVT2:%.*]] = sitofp i16 [[LD2]] to double
|
||||
; SSE-NEXT: [[CVT3:%.*]] = sitofp i16 [[LD3]] to double
|
||||
; SSE-NEXT: store double [[CVT0]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
; SSE-NEXT: store double [[CVT1]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
; SSE-NEXT: store double [[CVT2]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 2), align 8
|
||||
; SSE-NEXT: store double [[CVT3]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 3), align 8
|
||||
; SSE-NEXT: ret void
|
||||
;
|
||||
; AVX-LABEL: @sitofp_4i16_4f64(
|
||||
; AVX-NEXT: [[TMP1:%.*]] = load <4 x i16>, <4 x i16>* bitcast ([16 x i16]* @src16 to <4 x i16>*), align 8
|
||||
; AVX-NEXT: [[TMP2:%.*]] = sitofp <4 x i16> [[TMP1]] to <4 x double>
|
||||
; AVX-NEXT: store <4 x double> [[TMP2]], <4 x double>* bitcast ([4 x double]* @dst64 to <4 x double>*), align 8
|
||||
; AVX-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 2), align 8
|
||||
%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 3), align 8
|
||||
%cvt0 = sitofp i16 %ld0 to double
|
||||
%cvt1 = sitofp i16 %ld1 to double
|
||||
%cvt2 = sitofp i16 %ld2 to double
|
||||
%cvt3 = sitofp i16 %ld3 to double
|
||||
store double %cvt0, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
store double %cvt1, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
store double %cvt2, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 2), align 8
|
||||
store double %cvt3, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 3), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @sitofp_2i8_2f64() #0 {
|
||||
; CHECK-LABEL: @sitofp_2i8_2f64(
|
||||
; CHECK-NEXT: [[LD0:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: [[LD1:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: [[CVT0:%.*]] = sitofp i8 [[LD0]] to double
|
||||
; CHECK-NEXT: [[CVT1:%.*]] = sitofp i8 [[LD1]] to double
|
||||
; CHECK-NEXT: store double [[CVT0]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: store double [[CVT1]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
%ld1 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
%cvt0 = sitofp i8 %ld0 to double
|
||||
%cvt1 = sitofp i8 %ld1 to double
|
||||
store double %cvt0, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
store double %cvt1, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @sitofp_4i8_4f64() #0 {
|
||||
; SSE-LABEL: @sitofp_4i8_4f64(
|
||||
; SSE-NEXT: [[LD0:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
; SSE-NEXT: [[LD1:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
; SSE-NEXT: [[LD2:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 2), align 8
|
||||
; SSE-NEXT: [[LD3:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 3), align 8
|
||||
; SSE-NEXT: [[CVT0:%.*]] = sitofp i8 [[LD0]] to double
|
||||
; SSE-NEXT: [[CVT1:%.*]] = sitofp i8 [[LD1]] to double
|
||||
; SSE-NEXT: [[CVT2:%.*]] = sitofp i8 [[LD2]] to double
|
||||
; SSE-NEXT: [[CVT3:%.*]] = sitofp i8 [[LD3]] to double
|
||||
; SSE-NEXT: store double [[CVT0]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
; SSE-NEXT: store double [[CVT1]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
; SSE-NEXT: store double [[CVT2]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 2), align 8
|
||||
; SSE-NEXT: store double [[CVT3]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 3), align 8
|
||||
; SSE-NEXT: ret void
|
||||
;
|
||||
; AVX-LABEL: @sitofp_4i8_4f64(
|
||||
; AVX-NEXT: [[TMP1:%.*]] = load <4 x i8>, <4 x i8>* bitcast ([32 x i8]* @src8 to <4 x i8>*), align 8
|
||||
; AVX-NEXT: [[TMP2:%.*]] = sitofp <4 x i8> [[TMP1]] to <4 x double>
|
||||
; AVX-NEXT: store <4 x double> [[TMP2]], <4 x double>* bitcast ([4 x double]* @dst64 to <4 x double>*), align 8
|
||||
; AVX-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
%ld1 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
%ld2 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 2), align 8
|
||||
%ld3 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 3), align 8
|
||||
%cvt0 = sitofp i8 %ld0 to double
|
||||
%cvt1 = sitofp i8 %ld1 to double
|
||||
%cvt2 = sitofp i8 %ld2 to double
|
||||
%cvt3 = sitofp i8 %ld3 to double
|
||||
store double %cvt0, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
store double %cvt1, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
store double %cvt2, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 2), align 8
|
||||
store double %cvt3, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 3), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
;
|
||||
; SITOFP to vXf32
|
||||
;
|
||||
|
@ -177,25 +289,6 @@ define void @sitofp_4i64_4f32() #0 {
|
|||
ret void
|
||||
}
|
||||
|
||||
define void @sitofp_2i32_2f32() #0 {
|
||||
; CHECK-LABEL: @sitofp_2i32_2f32(
|
||||
; CHECK-NEXT: [[LD0:%.*]] = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: [[LD1:%.*]] = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: [[CVT0:%.*]] = sitofp i32 [[LD0]] to float
|
||||
; CHECK-NEXT: [[CVT1:%.*]] = sitofp i32 [[LD1]] to float
|
||||
; CHECK-NEXT: store float [[CVT0]], float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: store float [[CVT1]], float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 0), align 8
|
||||
%ld1 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 1), align 8
|
||||
%cvt0 = sitofp i32 %ld0 to float
|
||||
%cvt1 = sitofp i32 %ld1 to float
|
||||
store float %cvt0, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
store float %cvt1, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @sitofp_4i32_4f32() #0 {
|
||||
; CHECK-LABEL: @sitofp_4i32_4f32(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x i32>, <4 x i32>* bitcast ([8 x i32]* @src32 to <4 x i32>*), align 8
|
||||
|
@ -261,4 +354,134 @@ define void @sitofp_8i32_8f32() #0 {
|
|||
ret void
|
||||
}
|
||||
|
||||
define void @sitofp_4i16_4f32() #0 {
|
||||
; CHECK-LABEL: @sitofp_4i16_4f32(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x i16>, <4 x i16>* bitcast ([16 x i16]* @src16 to <4 x i16>*), align 8
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = sitofp <4 x i16> [[TMP1]] to <4 x float>
|
||||
; CHECK-NEXT: store <4 x float> [[TMP2]], <4 x float>* bitcast ([8 x float]* @dst32 to <4 x float>*), align 8
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 2), align 8
|
||||
%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 3), align 8
|
||||
%cvt0 = sitofp i16 %ld0 to float
|
||||
%cvt1 = sitofp i16 %ld1 to float
|
||||
%cvt2 = sitofp i16 %ld2 to float
|
||||
%cvt3 = sitofp i16 %ld3 to float
|
||||
store float %cvt0, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
store float %cvt1, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
store float %cvt2, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 2), align 8
|
||||
store float %cvt3, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 3), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @sitofp_8i16_8f32() #0 {
|
||||
; SSE-LABEL: @sitofp_8i16_8f32(
|
||||
; SSE-NEXT: [[TMP1:%.*]] = load <4 x i16>, <4 x i16>* bitcast ([16 x i16]* @src16 to <4 x i16>*), align 8
|
||||
; SSE-NEXT: [[TMP2:%.*]] = load <4 x i16>, <4 x i16>* bitcast (i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 4) to <4 x i16>*), align 8
|
||||
; SSE-NEXT: [[TMP3:%.*]] = sitofp <4 x i16> [[TMP1]] to <4 x float>
|
||||
; SSE-NEXT: [[TMP4:%.*]] = sitofp <4 x i16> [[TMP2]] to <4 x float>
|
||||
; SSE-NEXT: store <4 x float> [[TMP3]], <4 x float>* bitcast ([8 x float]* @dst32 to <4 x float>*), align 8
|
||||
; SSE-NEXT: store <4 x float> [[TMP4]], <4 x float>* bitcast (float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 4) to <4 x float>*), align 8
|
||||
; SSE-NEXT: ret void
|
||||
;
|
||||
; AVX-LABEL: @sitofp_8i16_8f32(
|
||||
; AVX-NEXT: [[TMP1:%.*]] = load <8 x i16>, <8 x i16>* bitcast ([16 x i16]* @src16 to <8 x i16>*), align 8
|
||||
; AVX-NEXT: [[TMP2:%.*]] = sitofp <8 x i16> [[TMP1]] to <8 x float>
|
||||
; AVX-NEXT: store <8 x float> [[TMP2]], <8 x float>* bitcast ([8 x float]* @dst32 to <8 x float>*), align 8
|
||||
; AVX-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 2), align 8
|
||||
%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 3), align 8
|
||||
%ld4 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 4), align 8
|
||||
%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 5), align 8
|
||||
%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 6), align 8
|
||||
%ld7 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 7), align 8
|
||||
%cvt0 = sitofp i16 %ld0 to float
|
||||
%cvt1 = sitofp i16 %ld1 to float
|
||||
%cvt2 = sitofp i16 %ld2 to float
|
||||
%cvt3 = sitofp i16 %ld3 to float
|
||||
%cvt4 = sitofp i16 %ld4 to float
|
||||
%cvt5 = sitofp i16 %ld5 to float
|
||||
%cvt6 = sitofp i16 %ld6 to float
|
||||
%cvt7 = sitofp i16 %ld7 to float
|
||||
store float %cvt0, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
store float %cvt1, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
store float %cvt2, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 2), align 8
|
||||
store float %cvt3, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 3), align 8
|
||||
store float %cvt4, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 4), align 8
|
||||
store float %cvt5, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 5), align 8
|
||||
store float %cvt6, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 6), align 8
|
||||
store float %cvt7, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 7), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @sitofp_4i8_4f32() #0 {
|
||||
; CHECK-LABEL: @sitofp_4i8_4f32(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x i8>, <4 x i8>* bitcast ([32 x i8]* @src8 to <4 x i8>*), align 8
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = sitofp <4 x i8> [[TMP1]] to <4 x float>
|
||||
; CHECK-NEXT: store <4 x float> [[TMP2]], <4 x float>* bitcast ([8 x float]* @dst32 to <4 x float>*), align 8
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
%ld1 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
%ld2 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 2), align 8
|
||||
%ld3 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 3), align 8
|
||||
%cvt0 = sitofp i8 %ld0 to float
|
||||
%cvt1 = sitofp i8 %ld1 to float
|
||||
%cvt2 = sitofp i8 %ld2 to float
|
||||
%cvt3 = sitofp i8 %ld3 to float
|
||||
store float %cvt0, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
store float %cvt1, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
store float %cvt2, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 2), align 8
|
||||
store float %cvt3, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 3), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @sitofp_8i8_8f32() #0 {
|
||||
; SSE-LABEL: @sitofp_8i8_8f32(
|
||||
; SSE-NEXT: [[TMP1:%.*]] = load <4 x i8>, <4 x i8>* bitcast ([32 x i8]* @src8 to <4 x i8>*), align 8
|
||||
; SSE-NEXT: [[TMP2:%.*]] = load <4 x i8>, <4 x i8>* bitcast (i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 4) to <4 x i8>*), align 8
|
||||
; SSE-NEXT: [[TMP3:%.*]] = sitofp <4 x i8> [[TMP1]] to <4 x float>
|
||||
; SSE-NEXT: [[TMP4:%.*]] = sitofp <4 x i8> [[TMP2]] to <4 x float>
|
||||
; SSE-NEXT: store <4 x float> [[TMP3]], <4 x float>* bitcast ([8 x float]* @dst32 to <4 x float>*), align 8
|
||||
; SSE-NEXT: store <4 x float> [[TMP4]], <4 x float>* bitcast (float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 4) to <4 x float>*), align 8
|
||||
; SSE-NEXT: ret void
|
||||
;
|
||||
; AVX-LABEL: @sitofp_8i8_8f32(
|
||||
; AVX-NEXT: [[TMP1:%.*]] = load <8 x i8>, <8 x i8>* bitcast ([32 x i8]* @src8 to <8 x i8>*), align 8
|
||||
; AVX-NEXT: [[TMP2:%.*]] = sitofp <8 x i8> [[TMP1]] to <8 x float>
|
||||
; AVX-NEXT: store <8 x float> [[TMP2]], <8 x float>* bitcast ([8 x float]* @dst32 to <8 x float>*), align 8
|
||||
; AVX-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
%ld1 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
%ld2 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 2), align 8
|
||||
%ld3 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 3), align 8
|
||||
%ld4 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 4), align 8
|
||||
%ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 5), align 8
|
||||
%ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 6), align 8
|
||||
%ld7 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 7), align 8
|
||||
%cvt0 = sitofp i8 %ld0 to float
|
||||
%cvt1 = sitofp i8 %ld1 to float
|
||||
%cvt2 = sitofp i8 %ld2 to float
|
||||
%cvt3 = sitofp i8 %ld3 to float
|
||||
%cvt4 = sitofp i8 %ld4 to float
|
||||
%cvt5 = sitofp i8 %ld5 to float
|
||||
%cvt6 = sitofp i8 %ld6 to float
|
||||
%cvt7 = sitofp i8 %ld7 to float
|
||||
store float %cvt0, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
store float %cvt1, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
store float %cvt2, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 2), align 8
|
||||
store float %cvt3, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 3), align 8
|
||||
store float %cvt4, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 4), align 8
|
||||
store float %cvt5, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 5), align 8
|
||||
store float %cvt6, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 6), align 8
|
||||
store float %cvt7, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 7), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
attributes #0 = { nounwind }
|
||||
|
|
|
@ -123,6 +123,118 @@ define void @uitofp_4i32_4f64() #0 {
|
|||
ret void
|
||||
}
|
||||
|
||||
define void @uitofp_2i16_2f64() #0 {
|
||||
; CHECK-LABEL: @uitofp_2i16_2f64(
|
||||
; CHECK-NEXT: [[LD0:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: [[LD1:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: [[CVT0:%.*]] = uitofp i16 [[LD0]] to double
|
||||
; CHECK-NEXT: [[CVT1:%.*]] = uitofp i16 [[LD1]] to double
|
||||
; CHECK-NEXT: store double [[CVT0]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: store double [[CVT1]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
%cvt0 = uitofp i16 %ld0 to double
|
||||
%cvt1 = uitofp i16 %ld1 to double
|
||||
store double %cvt0, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
store double %cvt1, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @uitofp_4i16_4f64() #0 {
|
||||
; SSE-LABEL: @uitofp_4i16_4f64(
|
||||
; SSE-NEXT: [[LD0:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
; SSE-NEXT: [[LD1:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
; SSE-NEXT: [[LD2:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 2), align 8
|
||||
; SSE-NEXT: [[LD3:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 3), align 8
|
||||
; SSE-NEXT: [[CVT0:%.*]] = uitofp i16 [[LD0]] to double
|
||||
; SSE-NEXT: [[CVT1:%.*]] = uitofp i16 [[LD1]] to double
|
||||
; SSE-NEXT: [[CVT2:%.*]] = uitofp i16 [[LD2]] to double
|
||||
; SSE-NEXT: [[CVT3:%.*]] = uitofp i16 [[LD3]] to double
|
||||
; SSE-NEXT: store double [[CVT0]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
; SSE-NEXT: store double [[CVT1]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
; SSE-NEXT: store double [[CVT2]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 2), align 8
|
||||
; SSE-NEXT: store double [[CVT3]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 3), align 8
|
||||
; SSE-NEXT: ret void
|
||||
;
|
||||
; AVX-LABEL: @uitofp_4i16_4f64(
|
||||
; AVX-NEXT: [[TMP1:%.*]] = load <4 x i16>, <4 x i16>* bitcast ([16 x i16]* @src16 to <4 x i16>*), align 8
|
||||
; AVX-NEXT: [[TMP2:%.*]] = uitofp <4 x i16> [[TMP1]] to <4 x double>
|
||||
; AVX-NEXT: store <4 x double> [[TMP2]], <4 x double>* bitcast ([4 x double]* @dst64 to <4 x double>*), align 8
|
||||
; AVX-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 2), align 8
|
||||
%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 3), align 8
|
||||
%cvt0 = uitofp i16 %ld0 to double
|
||||
%cvt1 = uitofp i16 %ld1 to double
|
||||
%cvt2 = uitofp i16 %ld2 to double
|
||||
%cvt3 = uitofp i16 %ld3 to double
|
||||
store double %cvt0, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
store double %cvt1, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
store double %cvt2, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 2), align 8
|
||||
store double %cvt3, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 3), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @uitofp_2i8_2f64() #0 {
|
||||
; CHECK-LABEL: @uitofp_2i8_2f64(
|
||||
; CHECK-NEXT: [[LD0:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: [[LD1:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: [[CVT0:%.*]] = uitofp i8 [[LD0]] to double
|
||||
; CHECK-NEXT: [[CVT1:%.*]] = uitofp i8 [[LD1]] to double
|
||||
; CHECK-NEXT: store double [[CVT0]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: store double [[CVT1]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
%ld1 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
%cvt0 = uitofp i8 %ld0 to double
|
||||
%cvt1 = uitofp i8 %ld1 to double
|
||||
store double %cvt0, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
store double %cvt1, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @uitofp_4i8_4f64() #0 {
|
||||
; SSE-LABEL: @uitofp_4i8_4f64(
|
||||
; SSE-NEXT: [[LD0:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
; SSE-NEXT: [[LD1:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
; SSE-NEXT: [[LD2:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 2), align 8
|
||||
; SSE-NEXT: [[LD3:%.*]] = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 3), align 8
|
||||
; SSE-NEXT: [[CVT0:%.*]] = uitofp i8 [[LD0]] to double
|
||||
; SSE-NEXT: [[CVT1:%.*]] = uitofp i8 [[LD1]] to double
|
||||
; SSE-NEXT: [[CVT2:%.*]] = uitofp i8 [[LD2]] to double
|
||||
; SSE-NEXT: [[CVT3:%.*]] = uitofp i8 [[LD3]] to double
|
||||
; SSE-NEXT: store double [[CVT0]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
; SSE-NEXT: store double [[CVT1]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
; SSE-NEXT: store double [[CVT2]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 2), align 8
|
||||
; SSE-NEXT: store double [[CVT3]], double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 3), align 8
|
||||
; SSE-NEXT: ret void
|
||||
;
|
||||
; AVX-LABEL: @uitofp_4i8_4f64(
|
||||
; AVX-NEXT: [[TMP1:%.*]] = load <4 x i8>, <4 x i8>* bitcast ([32 x i8]* @src8 to <4 x i8>*), align 8
|
||||
; AVX-NEXT: [[TMP2:%.*]] = uitofp <4 x i8> [[TMP1]] to <4 x double>
|
||||
; AVX-NEXT: store <4 x double> [[TMP2]], <4 x double>* bitcast ([4 x double]* @dst64 to <4 x double>*), align 8
|
||||
; AVX-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
%ld1 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
%ld2 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 2), align 8
|
||||
%ld3 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 3), align 8
|
||||
%cvt0 = uitofp i8 %ld0 to double
|
||||
%cvt1 = uitofp i8 %ld1 to double
|
||||
%cvt2 = uitofp i8 %ld2 to double
|
||||
%cvt3 = uitofp i8 %ld3 to double
|
||||
store double %cvt0, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 0), align 8
|
||||
store double %cvt1, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 1), align 8
|
||||
store double %cvt2, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 2), align 8
|
||||
store double %cvt3, double* getelementptr inbounds ([4 x double], [4 x double]* @dst64, i32 0, i64 3), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
;
|
||||
; UITOFP to vXf32
|
||||
;
|
||||
|
@ -177,25 +289,6 @@ define void @uitofp_4i64_4f32() #0 {
|
|||
ret void
|
||||
}
|
||||
|
||||
define void @uitofp_2i32_2f32() #0 {
|
||||
; CHECK-LABEL: @uitofp_2i32_2f32(
|
||||
; CHECK-NEXT: [[LD0:%.*]] = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: [[LD1:%.*]] = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: [[CVT0:%.*]] = uitofp i32 [[LD0]] to float
|
||||
; CHECK-NEXT: [[CVT1:%.*]] = uitofp i32 [[LD1]] to float
|
||||
; CHECK-NEXT: store float [[CVT0]], float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
; CHECK-NEXT: store float [[CVT1]], float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 0), align 8
|
||||
%ld1 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 1), align 8
|
||||
%cvt0 = uitofp i32 %ld0 to float
|
||||
%cvt1 = uitofp i32 %ld1 to float
|
||||
store float %cvt0, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
store float %cvt1, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @uitofp_4i32_4f32() #0 {
|
||||
; CHECK-LABEL: @uitofp_4i32_4f32(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x i32>, <4 x i32>* bitcast ([8 x i32]* @src32 to <4 x i32>*), align 8
|
||||
|
@ -261,4 +354,134 @@ define void @uitofp_8i32_8f32() #0 {
|
|||
ret void
|
||||
}
|
||||
|
||||
define void @uitofp_4i16_4f32() #0 {
|
||||
; CHECK-LABEL: @uitofp_4i16_4f32(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x i16>, <4 x i16>* bitcast ([16 x i16]* @src16 to <4 x i16>*), align 8
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = uitofp <4 x i16> [[TMP1]] to <4 x float>
|
||||
; CHECK-NEXT: store <4 x float> [[TMP2]], <4 x float>* bitcast ([8 x float]* @dst32 to <4 x float>*), align 8
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 2), align 8
|
||||
%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 3), align 8
|
||||
%cvt0 = uitofp i16 %ld0 to float
|
||||
%cvt1 = uitofp i16 %ld1 to float
|
||||
%cvt2 = uitofp i16 %ld2 to float
|
||||
%cvt3 = uitofp i16 %ld3 to float
|
||||
store float %cvt0, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
store float %cvt1, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
store float %cvt2, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 2), align 8
|
||||
store float %cvt3, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 3), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @uitofp_8i16_8f32() #0 {
|
||||
; SSE-LABEL: @uitofp_8i16_8f32(
|
||||
; SSE-NEXT: [[TMP1:%.*]] = load <4 x i16>, <4 x i16>* bitcast ([16 x i16]* @src16 to <4 x i16>*), align 8
|
||||
; SSE-NEXT: [[TMP2:%.*]] = load <4 x i16>, <4 x i16>* bitcast (i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 4) to <4 x i16>*), align 8
|
||||
; SSE-NEXT: [[TMP3:%.*]] = uitofp <4 x i16> [[TMP1]] to <4 x float>
|
||||
; SSE-NEXT: [[TMP4:%.*]] = uitofp <4 x i16> [[TMP2]] to <4 x float>
|
||||
; SSE-NEXT: store <4 x float> [[TMP3]], <4 x float>* bitcast ([8 x float]* @dst32 to <4 x float>*), align 8
|
||||
; SSE-NEXT: store <4 x float> [[TMP4]], <4 x float>* bitcast (float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 4) to <4 x float>*), align 8
|
||||
; SSE-NEXT: ret void
|
||||
;
|
||||
; AVX-LABEL: @uitofp_8i16_8f32(
|
||||
; AVX-NEXT: [[TMP1:%.*]] = load <8 x i16>, <8 x i16>* bitcast ([16 x i16]* @src16 to <8 x i16>*), align 8
|
||||
; AVX-NEXT: [[TMP2:%.*]] = uitofp <8 x i16> [[TMP1]] to <8 x float>
|
||||
; AVX-NEXT: store <8 x float> [[TMP2]], <8 x float>* bitcast ([8 x float]* @dst32 to <8 x float>*), align 8
|
||||
; AVX-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 0), align 8
|
||||
%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 1), align 8
|
||||
%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 2), align 8
|
||||
%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 3), align 8
|
||||
%ld4 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 4), align 8
|
||||
%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 5), align 8
|
||||
%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 6), align 8
|
||||
%ld7 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i32 0, i64 7), align 8
|
||||
%cvt0 = uitofp i16 %ld0 to float
|
||||
%cvt1 = uitofp i16 %ld1 to float
|
||||
%cvt2 = uitofp i16 %ld2 to float
|
||||
%cvt3 = uitofp i16 %ld3 to float
|
||||
%cvt4 = uitofp i16 %ld4 to float
|
||||
%cvt5 = uitofp i16 %ld5 to float
|
||||
%cvt6 = uitofp i16 %ld6 to float
|
||||
%cvt7 = uitofp i16 %ld7 to float
|
||||
store float %cvt0, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
store float %cvt1, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
store float %cvt2, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 2), align 8
|
||||
store float %cvt3, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 3), align 8
|
||||
store float %cvt4, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 4), align 8
|
||||
store float %cvt5, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 5), align 8
|
||||
store float %cvt6, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 6), align 8
|
||||
store float %cvt7, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 7), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @uitofp_4i8_4f32() #0 {
|
||||
; CHECK-LABEL: @uitofp_4i8_4f32(
|
||||
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x i8>, <4 x i8>* bitcast ([32 x i8]* @src8 to <4 x i8>*), align 8
|
||||
; CHECK-NEXT: [[TMP2:%.*]] = uitofp <4 x i8> [[TMP1]] to <4 x float>
|
||||
; CHECK-NEXT: store <4 x float> [[TMP2]], <4 x float>* bitcast ([8 x float]* @dst32 to <4 x float>*), align 8
|
||||
; CHECK-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
%ld1 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
%ld2 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 2), align 8
|
||||
%ld3 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 3), align 8
|
||||
%cvt0 = uitofp i8 %ld0 to float
|
||||
%cvt1 = uitofp i8 %ld1 to float
|
||||
%cvt2 = uitofp i8 %ld2 to float
|
||||
%cvt3 = uitofp i8 %ld3 to float
|
||||
store float %cvt0, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
store float %cvt1, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
store float %cvt2, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 2), align 8
|
||||
store float %cvt3, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 3), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
define void @uitofp_8i8_8f32() #0 {
|
||||
; SSE-LABEL: @uitofp_8i8_8f32(
|
||||
; SSE-NEXT: [[TMP1:%.*]] = load <4 x i8>, <4 x i8>* bitcast ([32 x i8]* @src8 to <4 x i8>*), align 8
|
||||
; SSE-NEXT: [[TMP2:%.*]] = load <4 x i8>, <4 x i8>* bitcast (i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 4) to <4 x i8>*), align 8
|
||||
; SSE-NEXT: [[TMP3:%.*]] = uitofp <4 x i8> [[TMP1]] to <4 x float>
|
||||
; SSE-NEXT: [[TMP4:%.*]] = uitofp <4 x i8> [[TMP2]] to <4 x float>
|
||||
; SSE-NEXT: store <4 x float> [[TMP3]], <4 x float>* bitcast ([8 x float]* @dst32 to <4 x float>*), align 8
|
||||
; SSE-NEXT: store <4 x float> [[TMP4]], <4 x float>* bitcast (float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 4) to <4 x float>*), align 8
|
||||
; SSE-NEXT: ret void
|
||||
;
|
||||
; AVX-LABEL: @uitofp_8i8_8f32(
|
||||
; AVX-NEXT: [[TMP1:%.*]] = load <8 x i8>, <8 x i8>* bitcast ([32 x i8]* @src8 to <8 x i8>*), align 8
|
||||
; AVX-NEXT: [[TMP2:%.*]] = uitofp <8 x i8> [[TMP1]] to <8 x float>
|
||||
; AVX-NEXT: store <8 x float> [[TMP2]], <8 x float>* bitcast ([8 x float]* @dst32 to <8 x float>*), align 8
|
||||
; AVX-NEXT: ret void
|
||||
;
|
||||
%ld0 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 0), align 8
|
||||
%ld1 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 1), align 8
|
||||
%ld2 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 2), align 8
|
||||
%ld3 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 3), align 8
|
||||
%ld4 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 4), align 8
|
||||
%ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 5), align 8
|
||||
%ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 6), align 8
|
||||
%ld7 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i32 0, i64 7), align 8
|
||||
%cvt0 = uitofp i8 %ld0 to float
|
||||
%cvt1 = uitofp i8 %ld1 to float
|
||||
%cvt2 = uitofp i8 %ld2 to float
|
||||
%cvt3 = uitofp i8 %ld3 to float
|
||||
%cvt4 = uitofp i8 %ld4 to float
|
||||
%cvt5 = uitofp i8 %ld5 to float
|
||||
%cvt6 = uitofp i8 %ld6 to float
|
||||
%cvt7 = uitofp i8 %ld7 to float
|
||||
store float %cvt0, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 0), align 8
|
||||
store float %cvt1, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 1), align 8
|
||||
store float %cvt2, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 2), align 8
|
||||
store float %cvt3, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 3), align 8
|
||||
store float %cvt4, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 4), align 8
|
||||
store float %cvt5, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 5), align 8
|
||||
store float %cvt6, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 6), align 8
|
||||
store float %cvt7, float* getelementptr inbounds ([8 x float], [8 x float]* @dst32, i32 0, i64 7), align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
attributes #0 = { nounwind }
|
||||
|
|
Loading…
Reference in New Issue