forked from OSchip/llvm-project
358 lines
14 KiB
C
358 lines
14 KiB
C
// RUN: %clang_cc1 -triple arm64-apple-ios7.0 -target-feature +neon -ffreestanding -emit-llvm -o - -O1 %s | FileCheck %s
|
|
#include <arm_neon.h>
|
|
|
|
int8x8_t test_vqshl_n_s8(int8x8_t in) {
|
|
// CHECK-LABEL: @test_vqshl_n_s8
|
|
// CHECK: call <8 x i8> @llvm.aarch64.neon.sqshl.v8i8(<8 x i8> %in, <8 x i8> <i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1>)
|
|
return vqshl_n_s8(in, 1);
|
|
}
|
|
|
|
int16x4_t test_vqshl_n_s16(int16x4_t in) {
|
|
// CHECK-LABEL: @test_vqshl_n_s16
|
|
// CHECK: call <4 x i16> @llvm.aarch64.neon.sqshl.v4i16(<4 x i16> %in, <4 x i16> <i16 1, i16 1, i16 1, i16 1>)
|
|
return vqshl_n_s16(in, 1);
|
|
}
|
|
|
|
int32x2_t test_vqshl_n_s32(int32x2_t in) {
|
|
// CHECK-LABEL: @test_vqshl_n_s32
|
|
// CHECK: call <2 x i32> @llvm.aarch64.neon.sqshl.v2i32(<2 x i32> %in, <2 x i32> <i32 1, i32 1>)
|
|
return vqshl_n_s32(in, 1);
|
|
}
|
|
|
|
int64x1_t test_vqshl_n_s64(int64x1_t in) {
|
|
// CHECK-LABEL: @test_vqshl_n_s64
|
|
// CHECK: call <1 x i64> @llvm.aarch64.neon.sqshl.v1i64(<1 x i64> %in, <1 x i64> <i64 1>)
|
|
return vqshl_n_s64(in, 1);
|
|
}
|
|
|
|
|
|
int8x16_t test_vqshlq_n_s8(int8x16_t in) {
|
|
// CHECK-LABEL: @test_vqshlq_n_s8
|
|
// CHECK: call <16 x i8> @llvm.aarch64.neon.sqshl.v16i8(<16 x i8> %in, <16 x i8> <i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1>)
|
|
return vqshlq_n_s8(in, 1);
|
|
}
|
|
|
|
int16x8_t test_vqshlq_n_s16(int16x8_t in) {
|
|
// CHECK-LABEL: @test_vqshlq_n_s16
|
|
// CHECK: call <8 x i16> @llvm.aarch64.neon.sqshl.v8i16(<8 x i16> %in, <8 x i16> <i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1>)
|
|
return vqshlq_n_s16(in, 1);
|
|
}
|
|
|
|
int32x4_t test_vqshlq_n_s32(int32x4_t in) {
|
|
// CHECK-LABEL: @test_vqshlq_n_s32
|
|
// CHECK: call <4 x i32> @llvm.aarch64.neon.sqshl.v4i32(<4 x i32> %in, <4 x i32> <i32 1, i32 1, i32 1, i32 1>)
|
|
return vqshlq_n_s32(in, 1);
|
|
}
|
|
|
|
int64x2_t test_vqshlq_n_s64(int64x2_t in) {
|
|
// CHECK-LABEL: @test_vqshlq_n_s64
|
|
// CHECK: call <2 x i64> @llvm.aarch64.neon.sqshl.v2i64(<2 x i64> %in, <2 x i64> <i64 1, i64 1>
|
|
return vqshlq_n_s64(in, 1);
|
|
}
|
|
|
|
uint8x8_t test_vqshl_n_u8(uint8x8_t in) {
|
|
// CHECK-LABEL: @test_vqshl_n_u8
|
|
// CHECK: call <8 x i8> @llvm.aarch64.neon.uqshl.v8i8(<8 x i8> %in, <8 x i8> <i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1>)
|
|
return vqshl_n_u8(in, 1);
|
|
}
|
|
|
|
uint16x4_t test_vqshl_n_u16(uint16x4_t in) {
|
|
// CHECK-LABEL: @test_vqshl_n_u16
|
|
// CHECK: call <4 x i16> @llvm.aarch64.neon.uqshl.v4i16(<4 x i16> %in, <4 x i16> <i16 1, i16 1, i16 1, i16 1>)
|
|
return vqshl_n_u16(in, 1);
|
|
}
|
|
|
|
uint32x2_t test_vqshl_n_u32(uint32x2_t in) {
|
|
// CHECK-LABEL: @test_vqshl_n_u32
|
|
// CHECK: call <2 x i32> @llvm.aarch64.neon.uqshl.v2i32(<2 x i32> %in, <2 x i32> <i32 1, i32 1>)
|
|
return vqshl_n_u32(in, 1);
|
|
}
|
|
|
|
uint64x1_t test_vqshl_n_u64(uint64x1_t in) {
|
|
// CHECK-LABEL: @test_vqshl_n_u64
|
|
// CHECK: call <1 x i64> @llvm.aarch64.neon.uqshl.v1i64(<1 x i64> %in, <1 x i64> <i64 1>)
|
|
return vqshl_n_u64(in, 1);
|
|
}
|
|
|
|
uint8x16_t test_vqshlq_n_u8(uint8x16_t in) {
|
|
// CHECK-LABEL: @test_vqshlq_n_u8
|
|
// CHECK: call <16 x i8> @llvm.aarch64.neon.uqshl.v16i8(<16 x i8> %in, <16 x i8> <i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1>)
|
|
return vqshlq_n_u8(in, 1);
|
|
}
|
|
|
|
uint16x8_t test_vqshlq_n_u16(uint16x8_t in) {
|
|
// CHECK-LABEL: @test_vqshlq_n_u16
|
|
// CHECK: call <8 x i16> @llvm.aarch64.neon.uqshl.v8i16(<8 x i16> %in, <8 x i16> <i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1>)
|
|
return vqshlq_n_u16(in, 1);
|
|
}
|
|
|
|
uint32x4_t test_vqshlq_n_u32(uint32x4_t in) {
|
|
// CHECK-LABEL: @test_vqshlq_n_u32
|
|
// CHECK: call <4 x i32> @llvm.aarch64.neon.uqshl.v4i32(<4 x i32> %in, <4 x i32> <i32 1, i32 1, i32 1, i32 1>)
|
|
return vqshlq_n_u32(in, 1);
|
|
}
|
|
|
|
uint64x2_t test_vqshlq_n_u64(uint64x2_t in) {
|
|
// CHECK-LABEL: @test_vqshlq_n_u64
|
|
// CHECK: call <2 x i64> @llvm.aarch64.neon.uqshl.v2i64(<2 x i64> %in, <2 x i64> <i64 1, i64 1>
|
|
return vqshlq_n_u64(in, 1);
|
|
}
|
|
|
|
int8x8_t test_vrshr_n_s8(int8x8_t in) {
|
|
// CHECK-LABEL: @test_vrshr_n_s8
|
|
// CHECK: call <8 x i8> @llvm.aarch64.neon.srshl.v8i8(<8 x i8> %in, <8 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>)
|
|
return vrshr_n_s8(in, 1);
|
|
}
|
|
|
|
int16x4_t test_vrshr_n_s16(int16x4_t in) {
|
|
// CHECK-LABEL: @test_vrshr_n_s16
|
|
// CHECK: call <4 x i16> @llvm.aarch64.neon.srshl.v4i16(<4 x i16> %in, <4 x i16> <i16 -1, i16 -1, i16 -1, i16 -1>)
|
|
return vrshr_n_s16(in, 1);
|
|
}
|
|
|
|
int32x2_t test_vrshr_n_s32(int32x2_t in) {
|
|
// CHECK-LABEL: @test_vrshr_n_s32
|
|
// CHECK: call <2 x i32> @llvm.aarch64.neon.srshl.v2i32(<2 x i32> %in, <2 x i32> <i32 -1, i32 -1>)
|
|
return vrshr_n_s32(in, 1);
|
|
}
|
|
|
|
int64x1_t test_vrshr_n_s64(int64x1_t in) {
|
|
// CHECK-LABEL: @test_vrshr_n_s64
|
|
// CHECK: call <1 x i64> @llvm.aarch64.neon.srshl.v1i64(<1 x i64> %in, <1 x i64> <i64 -1>)
|
|
return vrshr_n_s64(in, 1);
|
|
}
|
|
|
|
|
|
int8x16_t test_vrshrq_n_s8(int8x16_t in) {
|
|
// CHECK-LABEL: @test_vrshrq_n_s8
|
|
// CHECK: call <16 x i8> @llvm.aarch64.neon.srshl.v16i8(<16 x i8> %in, <16 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>)
|
|
return vrshrq_n_s8(in, 1);
|
|
}
|
|
|
|
int16x8_t test_vrshrq_n_s16(int16x8_t in) {
|
|
// CHECK-LABEL: @test_vrshrq_n_s16
|
|
// CHECK: call <8 x i16> @llvm.aarch64.neon.srshl.v8i16(<8 x i16> %in, <8 x i16> <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1>)
|
|
return vrshrq_n_s16(in, 1);
|
|
}
|
|
|
|
int32x4_t test_vrshrq_n_s32(int32x4_t in) {
|
|
// CHECK-LABEL: @test_vrshrq_n_s32
|
|
// CHECK: call <4 x i32> @llvm.aarch64.neon.srshl.v4i32(<4 x i32> %in, <4 x i32> <i32 -1, i32 -1, i32 -1, i32 -1>)
|
|
return vrshrq_n_s32(in, 1);
|
|
}
|
|
|
|
int64x2_t test_vrshrq_n_s64(int64x2_t in) {
|
|
// CHECK-LABEL: @test_vrshrq_n_s64
|
|
// CHECK: call <2 x i64> @llvm.aarch64.neon.srshl.v2i64(<2 x i64> %in, <2 x i64> <i64 -1, i64 -1>
|
|
return vrshrq_n_s64(in, 1);
|
|
}
|
|
|
|
uint8x8_t test_vrshr_n_u8(uint8x8_t in) {
|
|
// CHECK-LABEL: @test_vrshr_n_u8
|
|
// CHECK: call <8 x i8> @llvm.aarch64.neon.urshl.v8i8(<8 x i8> %in, <8 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>)
|
|
return vrshr_n_u8(in, 1);
|
|
}
|
|
|
|
uint16x4_t test_vrshr_n_u16(uint16x4_t in) {
|
|
// CHECK-LABEL: @test_vrshr_n_u16
|
|
// CHECK: call <4 x i16> @llvm.aarch64.neon.urshl.v4i16(<4 x i16> %in, <4 x i16> <i16 -1, i16 -1, i16 -1, i16 -1>)
|
|
return vrshr_n_u16(in, 1);
|
|
}
|
|
|
|
uint32x2_t test_vrshr_n_u32(uint32x2_t in) {
|
|
// CHECK-LABEL: @test_vrshr_n_u32
|
|
// CHECK: call <2 x i32> @llvm.aarch64.neon.urshl.v2i32(<2 x i32> %in, <2 x i32> <i32 -1, i32 -1>)
|
|
return vrshr_n_u32(in, 1);
|
|
}
|
|
|
|
uint64x1_t test_vrshr_n_u64(uint64x1_t in) {
|
|
// CHECK-LABEL: @test_vrshr_n_u64
|
|
// CHECK: call <1 x i64> @llvm.aarch64.neon.urshl.v1i64(<1 x i64> %in, <1 x i64> <i64 -1>)
|
|
return vrshr_n_u64(in, 1);
|
|
}
|
|
|
|
uint8x16_t test_vrshrq_n_u8(uint8x16_t in) {
|
|
// CHECK-LABEL: @test_vrshrq_n_u8
|
|
// CHECK: call <16 x i8> @llvm.aarch64.neon.urshl.v16i8(<16 x i8> %in, <16 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>)
|
|
return vrshrq_n_u8(in, 1);
|
|
}
|
|
|
|
uint16x8_t test_vrshrq_n_u16(uint16x8_t in) {
|
|
// CHECK-LABEL: @test_vrshrq_n_u16
|
|
// CHECK: call <8 x i16> @llvm.aarch64.neon.urshl.v8i16(<8 x i16> %in, <8 x i16> <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1>)
|
|
return vrshrq_n_u16(in, 1);
|
|
}
|
|
|
|
uint32x4_t test_vrshrq_n_u32(uint32x4_t in) {
|
|
// CHECK-LABEL: @test_vrshrq_n_u32
|
|
// CHECK: call <4 x i32> @llvm.aarch64.neon.urshl.v4i32(<4 x i32> %in, <4 x i32> <i32 -1, i32 -1, i32 -1, i32 -1>)
|
|
return vrshrq_n_u32(in, 1);
|
|
}
|
|
|
|
uint64x2_t test_vrshrq_n_u64(uint64x2_t in) {
|
|
// CHECK-LABEL: @test_vrshrq_n_u64
|
|
// CHECK: call <2 x i64> @llvm.aarch64.neon.urshl.v2i64(<2 x i64> %in, <2 x i64> <i64 -1, i64 -1>
|
|
return vrshrq_n_u64(in, 1);
|
|
}
|
|
|
|
int8x8_t test_vqshlu_n_s8(int8x8_t in) {
|
|
// CHECK-LABEL: @test_vqshlu_n_s8
|
|
// CHECK: call <8 x i8> @llvm.aarch64.neon.sqshlu.v8i8(<8 x i8> %in, <8 x i8> <i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1>)
|
|
return vqshlu_n_s8(in, 1);
|
|
}
|
|
|
|
int16x4_t test_vqshlu_n_s16(int16x4_t in) {
|
|
// CHECK-LABEL: @test_vqshlu_n_s16
|
|
// CHECK: call <4 x i16> @llvm.aarch64.neon.sqshlu.v4i16(<4 x i16> %in, <4 x i16> <i16 1, i16 1, i16 1, i16 1>)
|
|
return vqshlu_n_s16(in, 1);
|
|
}
|
|
|
|
int32x2_t test_vqshlu_n_s32(int32x2_t in) {
|
|
// CHECK-LABEL: @test_vqshlu_n_s32
|
|
// CHECK: call <2 x i32> @llvm.aarch64.neon.sqshlu.v2i32(<2 x i32> %in, <2 x i32> <i32 1, i32 1>)
|
|
return vqshlu_n_s32(in, 1);
|
|
}
|
|
|
|
int64x1_t test_vqshlu_n_s64(int64x1_t in) {
|
|
// CHECK-LABEL: @test_vqshlu_n_s64
|
|
// CHECK: call <1 x i64> @llvm.aarch64.neon.sqshlu.v1i64(<1 x i64> %in, <1 x i64> <i64 1>)
|
|
return vqshlu_n_s64(in, 1);
|
|
}
|
|
|
|
|
|
int8x16_t test_vqshluq_n_s8(int8x16_t in) {
|
|
// CHECK-LABEL: @test_vqshluq_n_s8
|
|
// CHECK: call <16 x i8> @llvm.aarch64.neon.sqshlu.v16i8(<16 x i8> %in, <16 x i8> <i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1>)
|
|
return vqshluq_n_s8(in, 1);
|
|
}
|
|
|
|
int16x8_t test_vqshluq_n_s16(int16x8_t in) {
|
|
// CHECK-LABEL: @test_vqshluq_n_s16
|
|
// CHECK: call <8 x i16> @llvm.aarch64.neon.sqshlu.v8i16(<8 x i16> %in, <8 x i16> <i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1>)
|
|
return vqshluq_n_s16(in, 1);
|
|
}
|
|
|
|
int32x4_t test_vqshluq_n_s32(int32x4_t in) {
|
|
// CHECK-LABEL: @test_vqshluq_n_s32
|
|
// CHECK: call <4 x i32> @llvm.aarch64.neon.sqshlu.v4i32(<4 x i32> %in, <4 x i32> <i32 1, i32 1, i32 1, i32 1>)
|
|
return vqshluq_n_s32(in, 1);
|
|
}
|
|
|
|
int64x2_t test_vqshluq_n_s64(int64x2_t in) {
|
|
// CHECK-LABEL: @test_vqshluq_n_s64
|
|
// CHECK: call <2 x i64> @llvm.aarch64.neon.sqshlu.v2i64(<2 x i64> %in, <2 x i64> <i64 1, i64 1>
|
|
return vqshluq_n_s64(in, 1);
|
|
}
|
|
|
|
int8x8_t test_vrsra_n_s8(int8x8_t acc, int8x8_t in) {
|
|
// CHECK-LABEL: @test_vrsra_n_s8
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <8 x i8> @llvm.aarch64.neon.srshl.v8i8(<8 x i8> %in, <8 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>)
|
|
// CHECK: add <8 x i8> [[TMP]], %acc
|
|
return vrsra_n_s8(acc, in, 1);
|
|
}
|
|
|
|
int16x4_t test_vrsra_n_s16(int16x4_t acc, int16x4_t in) {
|
|
// CHECK-LABEL: @test_vrsra_n_s16
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <4 x i16> @llvm.aarch64.neon.srshl.v4i16(<4 x i16> %in, <4 x i16> <i16 -1, i16 -1, i16 -1, i16 -1>)
|
|
// CHECK: add <4 x i16> [[TMP]], %acc
|
|
return vrsra_n_s16(acc, in, 1);
|
|
}
|
|
|
|
int32x2_t test_vrsra_n_s32(int32x2_t acc, int32x2_t in) {
|
|
// CHECK-LABEL: @test_vrsra_n_s32
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <2 x i32> @llvm.aarch64.neon.srshl.v2i32(<2 x i32> %in, <2 x i32> <i32 -1, i32 -1>)
|
|
// CHECK: add <2 x i32> [[TMP]], %acc
|
|
return vrsra_n_s32(acc, in, 1);
|
|
}
|
|
|
|
int64x1_t test_vrsra_n_s64(int64x1_t acc, int64x1_t in) {
|
|
// CHECK-LABEL: @test_vrsra_n_s64
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <1 x i64> @llvm.aarch64.neon.srshl.v1i64(<1 x i64> %in, <1 x i64> <i64 -1>)
|
|
// CHECK: add <1 x i64> [[TMP]], %acc
|
|
return vrsra_n_s64(acc, in, 1);
|
|
}
|
|
|
|
int8x16_t test_vrsraq_n_s8(int8x16_t acc, int8x16_t in) {
|
|
// CHECK-LABEL: @test_vrsraq_n_s8
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <16 x i8> @llvm.aarch64.neon.srshl.v16i8(<16 x i8> %in, <16 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>)
|
|
// CHECK: add <16 x i8> [[TMP]], %acc
|
|
return vrsraq_n_s8(acc, in, 1);
|
|
}
|
|
|
|
int16x8_t test_vrsraq_n_s16(int16x8_t acc, int16x8_t in) {
|
|
// CHECK-LABEL: @test_vrsraq_n_s16
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <8 x i16> @llvm.aarch64.neon.srshl.v8i16(<8 x i16> %in, <8 x i16> <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1>)
|
|
// CHECK: add <8 x i16> [[TMP]], %acc
|
|
return vrsraq_n_s16(acc, in, 1);
|
|
}
|
|
|
|
int32x4_t test_vrsraq_n_s32(int32x4_t acc, int32x4_t in) {
|
|
// CHECK-LABEL: @test_vrsraq_n_s32
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <4 x i32> @llvm.aarch64.neon.srshl.v4i32(<4 x i32> %in, <4 x i32> <i32 -1, i32 -1, i32 -1, i32 -1>)
|
|
// CHECK: add <4 x i32> [[TMP]], %acc
|
|
return vrsraq_n_s32(acc, in, 1);
|
|
}
|
|
|
|
int64x2_t test_vrsraq_n_s64(int64x2_t acc, int64x2_t in) {
|
|
// CHECK-LABEL: @test_vrsraq_n_s64
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <2 x i64> @llvm.aarch64.neon.srshl.v2i64(<2 x i64> %in, <2 x i64> <i64 -1, i64 -1>)
|
|
// CHECK: add <2 x i64> [[TMP]], %acc
|
|
return vrsraq_n_s64(acc, in, 1);
|
|
}
|
|
|
|
uint8x8_t test_vrsra_n_u8(uint8x8_t acc, uint8x8_t in) {
|
|
// CHECK-LABEL: @test_vrsra_n_u8
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <8 x i8> @llvm.aarch64.neon.urshl.v8i8(<8 x i8> %in, <8 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>)
|
|
// CHECK: add <8 x i8> [[TMP]], %acc
|
|
return vrsra_n_u8(acc, in, 1);
|
|
}
|
|
|
|
uint16x4_t test_vrsra_n_u16(uint16x4_t acc, uint16x4_t in) {
|
|
// CHECK-LABEL: @test_vrsra_n_u16
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <4 x i16> @llvm.aarch64.neon.urshl.v4i16(<4 x i16> %in, <4 x i16> <i16 -1, i16 -1, i16 -1, i16 -1>)
|
|
// CHECK: add <4 x i16> [[TMP]], %acc
|
|
return vrsra_n_u16(acc, in, 1);
|
|
}
|
|
|
|
uint32x2_t test_vrsra_n_u32(uint32x2_t acc, uint32x2_t in) {
|
|
// CHECK-LABEL: @test_vrsra_n_u32
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <2 x i32> @llvm.aarch64.neon.urshl.v2i32(<2 x i32> %in, <2 x i32> <i32 -1, i32 -1>)
|
|
// CHECK: add <2 x i32> [[TMP]], %acc
|
|
return vrsra_n_u32(acc, in, 1);
|
|
}
|
|
|
|
uint64x1_t test_vrsra_n_u64(uint64x1_t acc, uint64x1_t in) {
|
|
// CHECK-LABEL: @test_vrsra_n_u64
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <1 x i64> @llvm.aarch64.neon.urshl.v1i64(<1 x i64> %in, <1 x i64> <i64 -1>)
|
|
// CHECK: add <1 x i64> [[TMP]], %acc
|
|
return vrsra_n_u64(acc, in, 1);
|
|
}
|
|
|
|
uint8x16_t test_vrsraq_n_u8(uint8x16_t acc, uint8x16_t in) {
|
|
// CHECK-LABEL: @test_vrsraq_n_u8
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <16 x i8> @llvm.aarch64.neon.urshl.v16i8(<16 x i8> %in, <16 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>)
|
|
// CHECK: add <16 x i8> [[TMP]], %acc
|
|
return vrsraq_n_u8(acc, in, 1);
|
|
}
|
|
|
|
uint16x8_t test_vrsraq_n_u16(uint16x8_t acc, uint16x8_t in) {
|
|
// CHECK-LABEL: @test_vrsraq_n_u16
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <8 x i16> @llvm.aarch64.neon.urshl.v8i16(<8 x i16> %in, <8 x i16> <i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1, i16 -1>)
|
|
// CHECK: add <8 x i16> [[TMP]], %acc
|
|
return vrsraq_n_u16(acc, in, 1);
|
|
}
|
|
|
|
uint32x4_t test_vrsraq_n_u32(uint32x4_t acc, uint32x4_t in) {
|
|
// CHECK-LABEL: @test_vrsraq_n_u32
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <4 x i32> @llvm.aarch64.neon.urshl.v4i32(<4 x i32> %in, <4 x i32> <i32 -1, i32 -1, i32 -1, i32 -1>)
|
|
// CHECK: add <4 x i32> [[TMP]], %acc
|
|
return vrsraq_n_u32(acc, in, 1);
|
|
}
|
|
|
|
uint64x2_t test_vrsraq_n_u64(uint64x2_t acc, uint64x2_t in) {
|
|
// CHECK-LABEL: @test_vrsraq_n_u64
|
|
// CHECK: [[TMP:%[0-9a-zA-Z._]+]] = tail call <2 x i64> @llvm.aarch64.neon.urshl.v2i64(<2 x i64> %in, <2 x i64> <i64 -1, i64 -1>)
|
|
// CHECK: add <2 x i64> [[TMP]], %acc
|
|
return vrsraq_n_u64(acc, in, 1);
|
|
}
|