forked from OSchip/llvm-project
792 lines
27 KiB
C
792 lines
27 KiB
C
// RUN: %clang_cc1 -triple wasm32-unknown-unknown -target-feature +simd128 -target-feature +relaxed-simd -target-feature +nontrapping-fptoint -target-feature +exception-handling -target-feature +bulk-memory -target-feature +atomics -flax-vector-conversions=none -O3 -emit-llvm -o - %s | FileCheck %s -check-prefixes WEBASSEMBLY,WEBASSEMBLY32
|
|
// RUN: %clang_cc1 -triple wasm64-unknown-unknown -target-feature +simd128 -target-feature +relaxed-simd -target-feature +nontrapping-fptoint -target-feature +exception-handling -target-feature +bulk-memory -target-feature +atomics -flax-vector-conversions=none -O3 -emit-llvm -o - %s | FileCheck %s -check-prefixes WEBASSEMBLY,WEBASSEMBLY64
|
|
// RUN: not %clang_cc1 -triple wasm64-unknown-unknown -target-feature +nontrapping-fptoint -target-feature +exception-handling -target-feature +bulk-memory -target-feature +atomics -flax-vector-conversions=none -O3 -emit-llvm -o - %s 2>&1 | FileCheck %s -check-prefixes MISSING-SIMD
|
|
|
|
// SIMD convenience types
|
|
typedef signed char i8x16 __attribute((vector_size(16)));
|
|
typedef short i16x8 __attribute((vector_size(16)));
|
|
typedef int i32x4 __attribute((vector_size(16)));
|
|
typedef long long i64x2 __attribute((vector_size(16)));
|
|
typedef unsigned char u8x16 __attribute((vector_size(16)));
|
|
typedef unsigned short u16x8 __attribute((vector_size(16)));
|
|
typedef unsigned int u32x4 __attribute((vector_size(16)));
|
|
typedef unsigned long long u64x2 __attribute((vector_size(16)));
|
|
typedef float f32x4 __attribute((vector_size(16)));
|
|
typedef double f64x2 __attribute((vector_size(16)));
|
|
|
|
__SIZE_TYPE__ memory_size(void) {
|
|
return __builtin_wasm_memory_size(0);
|
|
// WEBASSEMBLY32: call {{i.*}} @llvm.wasm.memory.size.i32(i32 0)
|
|
// WEBASSEMBLY64: call {{i.*}} @llvm.wasm.memory.size.i64(i32 0)
|
|
}
|
|
|
|
__SIZE_TYPE__ memory_grow(__SIZE_TYPE__ delta) {
|
|
return __builtin_wasm_memory_grow(0, delta);
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.memory.grow.i32(i32 0, i32 %{{.*}})
|
|
// WEBASSEMBLY64: call i64 @llvm.wasm.memory.grow.i64(i32 0, i64 %{{.*}})
|
|
}
|
|
|
|
__SIZE_TYPE__ tls_size(void) {
|
|
return __builtin_wasm_tls_size();
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.tls.size.i32()
|
|
// WEBASSEMBLY64: call i64 @llvm.wasm.tls.size.i64()
|
|
}
|
|
|
|
__SIZE_TYPE__ tls_align(void) {
|
|
return __builtin_wasm_tls_align();
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.tls.align.i32()
|
|
// WEBASSEMBLY64: call i64 @llvm.wasm.tls.align.i64()
|
|
}
|
|
|
|
void *tls_base(void) {
|
|
return __builtin_wasm_tls_base();
|
|
// WEBASSEMBLY: call i8* @llvm.wasm.tls.base()
|
|
}
|
|
|
|
void throw(void *obj) {
|
|
return __builtin_wasm_throw(0, obj);
|
|
// WEBASSEMBLY32: call void @llvm.wasm.throw(i32 0, i8* %{{.*}})
|
|
// WEBASSEMBLY64: call void @llvm.wasm.throw(i32 0, i8* %{{.*}})
|
|
}
|
|
|
|
void rethrow(void) {
|
|
return __builtin_wasm_rethrow();
|
|
// WEBASSEMBLY32: call void @llvm.wasm.rethrow()
|
|
// WEBASSEMBLY64: call void @llvm.wasm.rethrow()
|
|
}
|
|
|
|
int memory_atomic_wait32(int *addr, int expected, long long timeout) {
|
|
return __builtin_wasm_memory_atomic_wait32(addr, expected, timeout);
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.memory.atomic.wait32(i32* %{{.*}}, i32 %{{.*}}, i64 %{{.*}})
|
|
// WEBASSEMBLY64: call i32 @llvm.wasm.memory.atomic.wait32(i32* %{{.*}}, i32 %{{.*}}, i64 %{{.*}})
|
|
}
|
|
|
|
int memory_atomic_wait64(long long *addr, long long expected, long long timeout) {
|
|
return __builtin_wasm_memory_atomic_wait64(addr, expected, timeout);
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.memory.atomic.wait64(i64* %{{.*}}, i64 %{{.*}}, i64 %{{.*}})
|
|
// WEBASSEMBLY64: call i32 @llvm.wasm.memory.atomic.wait64(i64* %{{.*}}, i64 %{{.*}}, i64 %{{.*}})
|
|
}
|
|
|
|
unsigned int memory_atomic_notify(int *addr, unsigned int count) {
|
|
return __builtin_wasm_memory_atomic_notify(addr, count);
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.memory.atomic.notify(i32* %{{.*}}, i32 %{{.*}})
|
|
// WEBASSEMBLY64: call i32 @llvm.wasm.memory.atomic.notify(i32* %{{.*}}, i32 %{{.*}})
|
|
}
|
|
|
|
int trunc_s_i32_f32(float f) {
|
|
return __builtin_wasm_trunc_s_i32_f32(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.signed.i32.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_u_i32_f32(float f) {
|
|
return __builtin_wasm_trunc_u_i32_f32(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.unsigned.i32.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_s_i32_f64(double f) {
|
|
return __builtin_wasm_trunc_s_i32_f64(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.signed.i32.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_u_i32_f64(double f) {
|
|
return __builtin_wasm_trunc_u_i32_f64(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.unsigned.i32.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_s_i64_f32(float f) {
|
|
return __builtin_wasm_trunc_s_i64_f32(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.signed.i64.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_u_i64_f32(float f) {
|
|
return __builtin_wasm_trunc_u_i64_f32(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.unsigned.i64.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_s_i64_f64(double f) {
|
|
return __builtin_wasm_trunc_s_i64_f64(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.signed.i64.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_u_i64_f64(double f) {
|
|
return __builtin_wasm_trunc_u_i64_f64(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.unsigned.i64.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_saturate_s_i32_f32(float f) {
|
|
return __builtin_wasm_trunc_saturate_s_i32_f32(f);
|
|
// WEBASSEMBLY: call i32 @llvm.fptosi.sat.i32.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_saturate_u_i32_f32(float f) {
|
|
return __builtin_wasm_trunc_saturate_u_i32_f32(f);
|
|
// WEBASSEMBLY: call i32 @llvm.fptoui.sat.i32.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_saturate_s_i32_f64(double f) {
|
|
return __builtin_wasm_trunc_saturate_s_i32_f64(f);
|
|
// WEBASSEMBLY: call i32 @llvm.fptosi.sat.i32.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_saturate_u_i32_f64(double f) {
|
|
return __builtin_wasm_trunc_saturate_u_i32_f64(f);
|
|
// WEBASSEMBLY: call i32 @llvm.fptoui.sat.i32.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_saturate_s_i64_f32(float f) {
|
|
return __builtin_wasm_trunc_saturate_s_i64_f32(f);
|
|
// WEBASSEMBLY: call i64 @llvm.fptosi.sat.i64.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_saturate_u_i64_f32(float f) {
|
|
return __builtin_wasm_trunc_saturate_u_i64_f32(f);
|
|
// WEBASSEMBLY: call i64 @llvm.fptoui.sat.i64.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_saturate_s_i64_f64(double f) {
|
|
return __builtin_wasm_trunc_saturate_s_i64_f64(f);
|
|
// WEBASSEMBLY: call i64 @llvm.fptosi.sat.i64.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_saturate_u_i64_f64(double f) {
|
|
return __builtin_wasm_trunc_saturate_u_i64_f64(f);
|
|
// WEBASSEMBLY: call i64 @llvm.fptoui.sat.i64.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
float min_f32(float x, float y) {
|
|
return __builtin_wasm_min_f32(x, y);
|
|
// WEBASSEMBLY: call float @llvm.minimum.f32(float %x, float %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
float max_f32(float x, float y) {
|
|
return __builtin_wasm_max_f32(x, y);
|
|
// WEBASSEMBLY: call float @llvm.maximum.f32(float %x, float %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
double min_f64(double x, double y) {
|
|
return __builtin_wasm_min_f64(x, y);
|
|
// WEBASSEMBLY: call double @llvm.minimum.f64(double %x, double %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
double max_f64(double x, double y) {
|
|
return __builtin_wasm_max_f64(x, y);
|
|
// WEBASSEMBLY: call double @llvm.maximum.f64(double %x, double %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 add_sat_s_i8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_add_sat_s_i8x16(x, y);
|
|
// MISSING-SIMD: error: '__builtin_wasm_add_sat_s_i8x16' needs target feature simd128
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.sadd.sat.v16i8(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %x, <16 x i8> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
u8x16 add_sat_u_i8x16(u8x16 x, u8x16 y) {
|
|
return __builtin_wasm_add_sat_u_i8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.uadd.sat.v16i8(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %x, <16 x i8> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i16x8 add_sat_s_i16x8(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_add_sat_s_i16x8(x, y);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.sadd.sat.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
u16x8 add_sat_u_i16x8(u16x8 x, u16x8 y) {
|
|
return __builtin_wasm_add_sat_u_i16x8(x, y);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.uadd.sat.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 sub_sat_s_i8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_sub_sat_s_i8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.sub.sat.signed.v16i8(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %x, <16 x i8> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
u8x16 sub_sat_u_i8x16(u8x16 x, u8x16 y) {
|
|
return __builtin_wasm_sub_sat_u_i8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.sub.sat.unsigned.v16i8(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %x, <16 x i8> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 abs_i8x16(i8x16 v) {
|
|
return __builtin_wasm_abs_i8x16(v);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.abs.v16i8(<16 x i8> %v, i1 false)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i16x8 abs_i16x8(i16x8 v) {
|
|
return __builtin_wasm_abs_i16x8(v);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.abs.v8i16(<8 x i16> %v, i1 false)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 abs_i32x4(i32x4 v) {
|
|
return __builtin_wasm_abs_i32x4(v);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.abs.v4i32(<4 x i32> %v, i1 false)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i64x2 abs_i64x2(i64x2 v) {
|
|
return __builtin_wasm_abs_i64x2(v);
|
|
// WEBASSEMBLY: call <2 x i64> @llvm.abs.v2i64(<2 x i64> %v, i1 false)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 min_s_i8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_min_s_i8x16(x, y);
|
|
// WEBASSEMBLY: %0 = icmp slt <16 x i8> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <16 x i1> %0, <16 x i8> %x, <16 x i8> %y
|
|
// WEBASSEMBLY-NEXT: ret <16 x i8> %1
|
|
}
|
|
|
|
u8x16 min_u_i8x16(u8x16 x, u8x16 y) {
|
|
return __builtin_wasm_min_u_i8x16(x, y);
|
|
// WEBASSEMBLY: %0 = icmp ult <16 x i8> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <16 x i1> %0, <16 x i8> %x, <16 x i8> %y
|
|
// WEBASSEMBLY-NEXT: ret <16 x i8> %1
|
|
}
|
|
|
|
i8x16 max_s_i8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_max_s_i8x16(x, y);
|
|
// WEBASSEMBLY: %0 = icmp sgt <16 x i8> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <16 x i1> %0, <16 x i8> %x, <16 x i8> %y
|
|
// WEBASSEMBLY-NEXT: ret <16 x i8> %1
|
|
}
|
|
|
|
u8x16 max_u_i8x16(u8x16 x, u8x16 y) {
|
|
return __builtin_wasm_max_u_i8x16(x, y);
|
|
// WEBASSEMBLY: %0 = icmp ugt <16 x i8> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <16 x i1> %0, <16 x i8> %x, <16 x i8> %y
|
|
// WEBASSEMBLY-NEXT: ret <16 x i8> %1
|
|
}
|
|
|
|
i16x8 min_s_i16x8(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_min_s_i16x8(x, y);
|
|
// WEBASSEMBLY: %0 = icmp slt <8 x i16> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <8 x i1> %0, <8 x i16> %x, <8 x i16> %y
|
|
// WEBASSEMBLY-NEXT: ret <8 x i16> %1
|
|
}
|
|
|
|
u16x8 min_u_i16x8(u16x8 x, u16x8 y) {
|
|
return __builtin_wasm_min_u_i16x8(x, y);
|
|
// WEBASSEMBLY: %0 = icmp ult <8 x i16> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <8 x i1> %0, <8 x i16> %x, <8 x i16> %y
|
|
// WEBASSEMBLY-NEXT: ret <8 x i16> %1
|
|
}
|
|
|
|
i16x8 max_s_i16x8(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_max_s_i16x8(x, y);
|
|
// WEBASSEMBLY: %0 = icmp sgt <8 x i16> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <8 x i1> %0, <8 x i16> %x, <8 x i16> %y
|
|
// WEBASSEMBLY-NEXT: ret <8 x i16> %1
|
|
}
|
|
|
|
u16x8 max_u_i16x8(u16x8 x, u16x8 y) {
|
|
return __builtin_wasm_max_u_i16x8(x, y);
|
|
// WEBASSEMBLY: %0 = icmp ugt <8 x i16> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <8 x i1> %0, <8 x i16> %x, <8 x i16> %y
|
|
// WEBASSEMBLY-NEXT: ret <8 x i16> %1
|
|
}
|
|
|
|
i32x4 min_s_i32x4(i32x4 x, i32x4 y) {
|
|
return __builtin_wasm_min_s_i32x4(x, y);
|
|
// WEBASSEMBLY: %0 = icmp slt <4 x i32> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <4 x i1> %0, <4 x i32> %x, <4 x i32> %y
|
|
// WEBASSEMBLY-NEXT: ret <4 x i32> %1
|
|
}
|
|
|
|
u32x4 min_u_i32x4(u32x4 x, u32x4 y) {
|
|
return __builtin_wasm_min_u_i32x4(x, y);
|
|
// WEBASSEMBLY: %0 = icmp ult <4 x i32> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <4 x i1> %0, <4 x i32> %x, <4 x i32> %y
|
|
// WEBASSEMBLY-NEXT: ret <4 x i32> %1
|
|
}
|
|
|
|
i32x4 max_s_i32x4(i32x4 x, i32x4 y) {
|
|
return __builtin_wasm_max_s_i32x4(x, y);
|
|
// WEBASSEMBLY: %0 = icmp sgt <4 x i32> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <4 x i1> %0, <4 x i32> %x, <4 x i32> %y
|
|
// WEBASSEMBLY-NEXT: ret <4 x i32> %1
|
|
}
|
|
|
|
u32x4 max_u_i32x4(u32x4 x, u32x4 y) {
|
|
return __builtin_wasm_max_u_i32x4(x, y);
|
|
// WEBASSEMBLY: %0 = icmp ugt <4 x i32> %x, %y
|
|
// WEBASSEMBLY-NEXT: %1 = select <4 x i1> %0, <4 x i32> %x, <4 x i32> %y
|
|
// WEBASSEMBLY-NEXT: ret <4 x i32> %1
|
|
}
|
|
|
|
i16x8 sub_sat_s_i16x8(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_sub_sat_s_i16x8(x, y);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.sub.sat.signed.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
u16x8 sub_sat_u_i16x8(u16x8 x, u16x8 y) {
|
|
return __builtin_wasm_sub_sat_u_i16x8(x, y);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.sub.sat.unsigned.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
u8x16 avgr_u_i8x16(u8x16 x, u8x16 y) {
|
|
return __builtin_wasm_avgr_u_i8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.avgr.unsigned.v16i8(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %x, <16 x i8> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
u16x8 avgr_u_i16x8(u16x8 x, u16x8 y) {
|
|
return __builtin_wasm_avgr_u_i16x8(x, y);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.avgr.unsigned.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i16x8 q15mulr_sat_s_i16x8(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_q15mulr_sat_s_i16x8(x, y);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.q15mulr.sat.signed(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i16x8 extadd_pairwise_i8x16_s_i16x8(i8x16 v) {
|
|
return __builtin_wasm_extadd_pairwise_i8x16_s_i16x8(v);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.extadd.pairwise.signed.v8i16(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %v)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
u16x8 extadd_pairwise_i8x16_u_i16x8(u8x16 v) {
|
|
return __builtin_wasm_extadd_pairwise_i8x16_u_i16x8(v);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.extadd.pairwise.unsigned.v8i16(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %v)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 extadd_pairwise_i16x8_s_i32x4(i16x8 v) {
|
|
return __builtin_wasm_extadd_pairwise_i16x8_s_i32x4(v);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.extadd.pairwise.signed.v4i32(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %v)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
u32x4 extadd_pairwise_i16x8_u_i32x4(u16x8 v) {
|
|
return __builtin_wasm_extadd_pairwise_i16x8_u_i32x4(v);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.extadd.pairwise.unsigned.v4i32(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %v)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 dot_i16x8_s(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_dot_s_i32x4_i16x8(x, y);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.dot(<8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 bitselect(i32x4 x, i32x4 y, i32x4 c) {
|
|
return __builtin_wasm_bitselect(x, y, c);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.bitselect.v4i32(
|
|
// WEBASSEMBLY-SAME: <4 x i32> %x, <4 x i32> %y, <4 x i32> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 popcnt(i8x16 x) {
|
|
return __builtin_wasm_popcnt_i8x16(x);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.ctpop.v16i8(<16 x i8> %x)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int any_true_v128(i8x16 x) {
|
|
return __builtin_wasm_any_true_v128(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.anytrue.v16i8(<16 x i8> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int all_true_i8x16(i8x16 x) {
|
|
return __builtin_wasm_all_true_i8x16(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v16i8(<16 x i8> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int all_true_i16x8(i16x8 x) {
|
|
return __builtin_wasm_all_true_i16x8(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v8i16(<8 x i16> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int all_true_i32x4(i32x4 x) {
|
|
return __builtin_wasm_all_true_i32x4(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v4i32(<4 x i32> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int all_true_i64x2(i64x2 x) {
|
|
return __builtin_wasm_all_true_i64x2(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v2i64(<2 x i64> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int bitmask_i8x16(i8x16 x) {
|
|
return __builtin_wasm_bitmask_i8x16(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.bitmask.v16i8(<16 x i8> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int bitmask_i16x8(i16x8 x) {
|
|
return __builtin_wasm_bitmask_i16x8(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.bitmask.v8i16(<8 x i16> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int bitmask_i32x4(i32x4 x) {
|
|
return __builtin_wasm_bitmask_i32x4(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.bitmask.v4i32(<4 x i32> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int bitmask_i64x2(i64x2 x) {
|
|
return __builtin_wasm_bitmask_i64x2(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.bitmask.v2i64(<2 x i64> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f32x4 abs_f32x4(f32x4 x) {
|
|
return __builtin_wasm_abs_f32x4(x);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.fabs.v4f32(<4 x float> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f64x2 abs_f64x2(f64x2 x) {
|
|
return __builtin_wasm_abs_f64x2(x);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.fabs.v2f64(<2 x double> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f32x4 min_f32x4(f32x4 x, f32x4 y) {
|
|
return __builtin_wasm_min_f32x4(x, y);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.minimum.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %x, <4 x float> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f32x4 max_f32x4(f32x4 x, f32x4 y) {
|
|
return __builtin_wasm_max_f32x4(x, y);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.maximum.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %x, <4 x float> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f32x4 pmin_f32x4(f32x4 x, f32x4 y) {
|
|
return __builtin_wasm_pmin_f32x4(x, y);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.wasm.pmin.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %x, <4 x float> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f32x4 pmax_f32x4(f32x4 x, f32x4 y) {
|
|
return __builtin_wasm_pmax_f32x4(x, y);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.wasm.pmax.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %x, <4 x float> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 min_f64x2(f64x2 x, f64x2 y) {
|
|
return __builtin_wasm_min_f64x2(x, y);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.minimum.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %x, <2 x double> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 max_f64x2(f64x2 x, f64x2 y) {
|
|
return __builtin_wasm_max_f64x2(x, y);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.maximum.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %x, <2 x double> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 pmin_f64x2(f64x2 x, f64x2 y) {
|
|
return __builtin_wasm_pmin_f64x2(x, y);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.wasm.pmin.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %x, <2 x double> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 pmax_f64x2(f64x2 x, f64x2 y) {
|
|
return __builtin_wasm_pmax_f64x2(x, y);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.wasm.pmax.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %x, <2 x double> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f32x4 ceil_f32x4(f32x4 x) {
|
|
return __builtin_wasm_ceil_f32x4(x);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.ceil.v4f32(<4 x float> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f32x4 floor_f32x4(f32x4 x) {
|
|
return __builtin_wasm_floor_f32x4(x);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.floor.v4f32(<4 x float> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f32x4 trunc_f32x4(f32x4 x) {
|
|
return __builtin_wasm_trunc_f32x4(x);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.trunc.v4f32(<4 x float> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f32x4 nearest_f32x4(f32x4 x) {
|
|
return __builtin_wasm_nearest_f32x4(x);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.nearbyint.v4f32(<4 x float> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f64x2 ceil_f64x2(f64x2 x) {
|
|
return __builtin_wasm_ceil_f64x2(x);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.ceil.v2f64(<2 x double> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f64x2 floor_f64x2(f64x2 x) {
|
|
return __builtin_wasm_floor_f64x2(x);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.floor.v2f64(<2 x double> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f64x2 trunc_f64x2(f64x2 x) {
|
|
return __builtin_wasm_trunc_f64x2(x);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.trunc.v2f64(<2 x double> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f64x2 nearest_f64x2(f64x2 x) {
|
|
return __builtin_wasm_nearest_f64x2(x);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.nearbyint.v2f64(<2 x double> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f32x4 sqrt_f32x4(f32x4 x) {
|
|
return __builtin_wasm_sqrt_f32x4(x);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.sqrt.v4f32(<4 x float> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f64x2 sqrt_f64x2(f64x2 x) {
|
|
return __builtin_wasm_sqrt_f64x2(x);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.sqrt.v2f64(<2 x double> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i32x4 trunc_saturate_s_i32x4_f32x4(f32x4 f) {
|
|
return __builtin_wasm_trunc_saturate_s_i32x4_f32x4(f);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.fptosi.sat.v4i32.v4f32(<4 x float> %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 trunc_saturate_u_i32x4_f32x4(f32x4 f) {
|
|
return __builtin_wasm_trunc_saturate_u_i32x4_f32x4(f);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.fptoui.sat.v4i32.v4f32(<4 x float> %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 narrow_s_i8x16_i16x8(i16x8 low, i16x8 high) {
|
|
return __builtin_wasm_narrow_s_i8x16_i16x8(low, high);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.narrow.signed.v16i8.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %low, <8 x i16> %high)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
u8x16 narrow_u_i8x16_i16x8(i16x8 low, i16x8 high) {
|
|
return __builtin_wasm_narrow_u_i8x16_i16x8(low, high);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.narrow.unsigned.v16i8.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %low, <8 x i16> %high)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i16x8 narrow_s_i16x8_i32x4(i32x4 low, i32x4 high) {
|
|
return __builtin_wasm_narrow_s_i16x8_i32x4(low, high);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.narrow.signed.v8i16.v4i32(
|
|
// WEBASSEMBLY-SAME: <4 x i32> %low, <4 x i32> %high)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
u16x8 narrow_u_i16x8_i32x4(i32x4 low, i32x4 high) {
|
|
return __builtin_wasm_narrow_u_i16x8_i32x4(low, high);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.narrow.unsigned.v8i16.v4i32(
|
|
// WEBASSEMBLY-SAME: <4 x i32> %low, <4 x i32> %high)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i32x4 trunc_sat_zero_s_f64x2_i32x4(f64x2 x) {
|
|
return __builtin_wasm_trunc_sat_zero_s_f64x2_i32x4(x);
|
|
// WEBASSEMBLY: %0 = tail call <2 x i32> @llvm.fptosi.sat.v2i32.v2f64(<2 x double> %x)
|
|
// WEBASSEMBLY: %1 = shufflevector <2 x i32> %0, <2 x i32> zeroinitializer, <4 x i32> <i32 0, i32 1, i32 2, i32 3>
|
|
// WEBASSEMBLY: ret <4 x i32> %1
|
|
}
|
|
|
|
u32x4 trunc_sat_zero_u_f64x2_i32x4(f64x2 x) {
|
|
return __builtin_wasm_trunc_sat_zero_u_f64x2_i32x4(x);
|
|
// WEBASSEMBLY: %0 = tail call <2 x i32> @llvm.fptoui.sat.v2i32.v2f64(<2 x double> %x)
|
|
// WEBASSEMBLY: %1 = shufflevector <2 x i32> %0, <2 x i32> zeroinitializer, <4 x i32> <i32 0, i32 1, i32 2, i32 3>
|
|
// WEBASSEMBLY: ret <4 x i32> %1
|
|
}
|
|
|
|
i8x16 swizzle_i8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_swizzle_i8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.swizzle(<16 x i8> %x, <16 x i8> %y)
|
|
}
|
|
|
|
i8x16 shuffle(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_shuffle_i8x16(x, y, 0, 1, 2, 3, 4, 5, 6, 7,
|
|
8, 9, 10, 11, 12, 13, 14, 15);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.shuffle(<16 x i8> %x, <16 x i8> %y,
|
|
// WEBASSEMBLY-SAME: i32 0, i32 1, i32 2, i32 3, i32 4, i32 5, i32 6, i32 7,
|
|
// WEBASSEMBLY-SAME: i32 8, i32 9, i32 10, i32 11, i32 12, i32 13, i32 14,
|
|
// WEBASSEMBLY-SAME: i32 15
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f32x4 fma_f32x4(f32x4 a, f32x4 b, f32x4 c) {
|
|
return __builtin_wasm_fma_f32x4(a, b, c);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.wasm.fma.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %a, <4 x float> %b, <4 x float> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f32x4 fms_f32x4(f32x4 a, f32x4 b, f32x4 c) {
|
|
return __builtin_wasm_fms_f32x4(a, b, c);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.wasm.fms.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %a, <4 x float> %b, <4 x float> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 fma_f64x2(f64x2 a, f64x2 b, f64x2 c) {
|
|
return __builtin_wasm_fma_f64x2(a, b, c);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.wasm.fma.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %a, <2 x double> %b, <2 x double> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 fms_f64x2(f64x2 a, f64x2 b, f64x2 c) {
|
|
return __builtin_wasm_fms_f64x2(a, b, c);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.wasm.fms.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %a, <2 x double> %b, <2 x double> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 laneselect_i8x16(i8x16 a, i8x16 b, i8x16 c) {
|
|
return __builtin_wasm_laneselect_i8x16(a, b, c);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.laneselect.v16i8(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %a, <16 x i8> %b, <16 x i8> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i16x8 laneselect_i16x8(i16x8 a, i16x8 b, i16x8 c) {
|
|
return __builtin_wasm_laneselect_i16x8(a, b, c);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.laneselect.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %a, <8 x i16> %b, <8 x i16> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 laneselect_i32x4(i32x4 a, i32x4 b, i32x4 c) {
|
|
return __builtin_wasm_laneselect_i32x4(a, b, c);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.laneselect.v4i32(
|
|
// WEBASSEMBLY-SAME: <4 x i32> %a, <4 x i32> %b, <4 x i32> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i64x2 laneselect_i64x2(i64x2 a, i64x2 b, i64x2 c) {
|
|
return __builtin_wasm_laneselect_i64x2(a, b, c);
|
|
// WEBASSEMBLY: call <2 x i64> @llvm.wasm.laneselect.v2i64(
|
|
// WEBASSEMBLY-SAME: <2 x i64> %a, <2 x i64> %b, <2 x i64> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 relaxed_swizzle_i8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_relaxed_swizzle_i8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.relaxed.swizzle(<16 x i8> %x, <16 x i8> %y)
|
|
}
|
|
|
|
f32x4 relaxed_min_f32x4(f32x4 a, f32x4 b) {
|
|
return __builtin_wasm_relaxed_min_f32x4(a, b);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.wasm.relaxed.min.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %a, <4 x float> %b)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f32x4 relaxed_max_f32x4(f32x4 a, f32x4 b) {
|
|
return __builtin_wasm_relaxed_max_f32x4(a, b);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.wasm.relaxed.max.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %a, <4 x float> %b)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 relaxed_min_f64x2(f64x2 a, f64x2 b) {
|
|
return __builtin_wasm_relaxed_min_f64x2(a, b);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.wasm.relaxed.min.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %a, <2 x double> %b)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 relaxed_max_f64x2(f64x2 a, f64x2 b) {
|
|
return __builtin_wasm_relaxed_max_f64x2(a, b);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.wasm.relaxed.max.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %a, <2 x double> %b)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 relaxed_trunc_s_i32x4_f32x4(f32x4 f) {
|
|
return __builtin_wasm_relaxed_trunc_s_i32x4_f32x4(f);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.relaxed.trunc.signed(<4 x float> %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
u32x4 relaxed_trunc_u_i32x4_f32x4(f32x4 f) {
|
|
return __builtin_wasm_relaxed_trunc_u_i32x4_f32x4(f);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.relaxed.trunc.unsigned(<4 x float> %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 relaxed_trunc_zero_s_i32x4_f64x2(f64x2 x) {
|
|
return __builtin_wasm_relaxed_trunc_zero_s_i32x4_f64x2(x);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.relaxed.trunc.zero.signed(<2 x double> %x)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
u32x4 relaxed_trunc_zero_u_i32x4_f64x2(f64x2 x) {
|
|
return __builtin_wasm_relaxed_trunc_zero_u_i32x4_f64x2(x);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.relaxed.trunc.zero.unsigned(<2 x double> %x)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|