forked from OSchip/llvm-project
101 lines
4.5 KiB
ArmAsm
101 lines
4.5 KiB
ArmAsm
@ RUN: llvm-mc -mcpu=cortex-a8 -triple thumb-unknown-unknown -show-encoding < %s | FileCheck %s
|
|
|
|
.code 16
|
|
|
|
@ CHECK: vsra.s8 d17, d16, #8 @ encoding: [0xc8,0xef,0x30,0x11]
|
|
vsra.s8 d17, d16, #8
|
|
@ CHECK: vsra.s16 d17, d16, #16 @ encoding: [0xd0,0xef,0x30,0x11]
|
|
vsra.s16 d17, d16, #16
|
|
@ CHECK: vsra.s32 d17, d16, #32 @ encoding: [0xe0,0xef,0x30,0x11]
|
|
vsra.s32 d17, d16, #32
|
|
@ CHECK: vsra.s64 d17, d16, #64 @ encoding: [0xc0,0xef,0xb0,0x11]
|
|
vsra.s64 d17, d16, #64
|
|
@ CHECK: vsra.s8 q8, q9, #8 @ encoding: [0xc8,0xef,0x72,0x01]
|
|
vsra.s8 q8, q9, #8
|
|
@ CHECK: vsra.s16 q8, q9, #16 @ encoding: [0xd0,0xef,0x72,0x01]
|
|
vsra.s16 q8, q9, #16
|
|
@ CHECK: vsra.s32 q8, q9, #32 @ encoding: [0xe0,0xef,0x72,0x01]
|
|
vsra.s32 q8, q9, #32
|
|
@ CHECK: vsra.s64 q8, q9, #64 @ encoding: [0xc0,0xef,0xf2,0x01]
|
|
vsra.s64 q8, q9, #64
|
|
@ CHECK: vsra.u8 d17, d16, #8 @ encoding: [0xc8,0xff,0x30,0x11]
|
|
vsra.u8 d17, d16, #8
|
|
@ CHECK: vsra.u16 d17, d16, #16 @ encoding: [0xd0,0xff,0x30,0x11]
|
|
vsra.u16 d17, d16, #16
|
|
@ CHECK: vsra.u32 d17, d16, #32 @ encoding: [0xe0,0xff,0x30,0x11]
|
|
vsra.u32 d17, d16, #32
|
|
@ CHECK: vsra.u64 d17, d16, #64 @ encoding: [0xc0,0xff,0xb0,0x11]
|
|
vsra.u64 d17, d16, #64
|
|
@ CHECK: vsra.u8 q8, q9, #8 @ encoding: [0xc8,0xff,0x72,0x01]
|
|
vsra.u8 q8, q9, #8
|
|
@ CHECK: vsra.u16 q8, q9, #16 @ encoding: [0xd0,0xff,0x72,0x01]
|
|
vsra.u16 q8, q9, #16
|
|
@ CHECK: vsra.u32 q8, q9, #32 @ encoding: [0xe0,0xff,0x72,0x01]
|
|
vsra.u32 q8, q9, #32
|
|
@ CHECK: vsra.u64 q8, q9, #64 @ encoding: [0xc0,0xff,0xf2,0x01]
|
|
vsra.u64 q8, q9, #64
|
|
@ CHECK: vrsra.s8 d17, d16, #8 @ encoding: [0xc8,0xef,0x30,0x13]
|
|
vrsra.s8 d17, d16, #8
|
|
@ CHECK: vrsra.s16 d17, d16, #16 @ encoding: [0xd0,0xef,0x30,0x13]
|
|
vrsra.s16 d17, d16, #16
|
|
@ CHECK: vrsra.s32 d17, d16, #32 @ encoding: [0xe0,0xef,0x30,0x13]
|
|
vrsra.s32 d17, d16, #32
|
|
@ CHECK: vrsra.s64 d17, d16, #64 @ encoding: [0xc0,0xef,0xb0,0x13]
|
|
vrsra.s64 d17, d16, #64
|
|
@ CHECK: vrsra.u8 d17, d16, #8 @ encoding: [0xc8,0xff,0x30,0x13]
|
|
vrsra.u8 d17, d16, #8
|
|
@ CHECK: vrsra.u16 d17, d16, #16 @ encoding: [0xd0,0xff,0x30,0x13]
|
|
vrsra.u16 d17, d16, #16
|
|
@ CHECK: vrsra.u32 d17, d16, #32 @ encoding: [0xe0,0xff,0x30,0x13]
|
|
vrsra.u32 d17, d16, #32
|
|
@ CHECK: vrsra.u64 d17, d16, #64 @ encoding: [0xc0,0xff,0xb0,0x13]
|
|
vrsra.u64 d17, d16, #64
|
|
@ CHECK: vrsra.s8 q8, q9, #8 @ encoding: [0xc8,0xef,0x72,0x03]
|
|
vrsra.s8 q8, q9, #8
|
|
@ CHECK: vrsra.s16 q8, q9, #16 @ encoding: [0xd0,0xef,0x72,0x03]
|
|
vrsra.s16 q8, q9, #16
|
|
@ CHECK: vrsra.s32 q8, q9, #32 @ encoding: [0xe0,0xef,0x72,0x03]
|
|
vrsra.s32 q8, q9, #32
|
|
@ CHECK: vrsra.s64 q8, q9, #64 @ encoding: [0xc0,0xef,0xf2,0x03]
|
|
vrsra.s64 q8, q9, #64
|
|
@ CHECK: vrsra.u8 q8, q9, #8 @ encoding: [0xc8,0xff,0x72,0x03]
|
|
vrsra.u8 q8, q9, #8
|
|
@ CHECK: vrsra.u16 q8, q9, #16 @ encoding: [0xd0,0xff,0x72,0x03]
|
|
vrsra.u16 q8, q9, #16
|
|
@ CHECK: vrsra.u32 q8, q9, #32 @ encoding: [0xe0,0xff,0x72,0x03]
|
|
vrsra.u32 q8, q9, #32
|
|
@ CHECK: vrsra.u64 q8, q9, #64 @ encoding: [0xc0,0xff,0xf2,0x03]
|
|
vrsra.u64 q8, q9, #64
|
|
@ CHECK: vsli.8 d17, d16, #7 @ encoding: [0xcf,0xff,0x30,0x15]
|
|
vsli.8 d17, d16, #7
|
|
@ CHECK: vsli.16 d17, d16, #15 @ encoding: [0xdf,0xff,0x30,0x15]
|
|
vsli.16 d17, d16, #15
|
|
@ CHECK: vsli.32 d17, d16, #31 @ encoding: [0xff,0xff,0x30,0x15]
|
|
vsli.32 d17, d16, #31
|
|
@ CHECK: vsli.64 d17, d16, #63 @ encoding: [0xff,0xff,0xb0,0x15]
|
|
vsli.64 d17, d16, #63
|
|
@ CHECK: vsli.8 q9, q8, #7 @ encoding: [0xcf,0xff,0x70,0x25]
|
|
vsli.8 q9, q8, #7
|
|
@ CHECK: vsli.16 q9, q8, #15 @ encoding: [0xdf,0xff,0x70,0x25]
|
|
vsli.16 q9, q8, #15
|
|
@ CHECK: vsli.32 q9, q8, #31 @ encoding: [0xff,0xff,0x70,0x25]
|
|
vsli.32 q9, q8, #31
|
|
@ CHECK: vsli.64 q9, q8, #63 @ encoding: [0xff,0xff,0xf0,0x25]
|
|
vsli.64 q9, q8, #63
|
|
@ CHECK: vsri.8 d17, d16, #8 @ encoding: [0xc8,0xff,0x30,0x14]
|
|
vsri.8 d17, d16, #8
|
|
@ CHECK: vsri.16 d17, d16, #16 @ encoding: [0xd0,0xff,0x30,0x14]
|
|
vsri.16 d17, d16, #16
|
|
@ CHECK: vsri.32 d17, d16, #32 @ encoding: [0xe0,0xff,0x30,0x14]
|
|
vsri.32 d17, d16, #32
|
|
@ CHECK: vsri.64 d17, d16, #64 @ encoding: [0xc0,0xff,0xb0,0x14]
|
|
vsri.64 d17, d16, #64
|
|
@ CHECK: vsri.8 q9, q8, #8 @ encoding: [0xc8,0xff,0x70,0x24]
|
|
vsri.8 q9, q8, #8
|
|
@ CHECK: vsri.16 q9, q8, #16 @ encoding: [0xd0,0xff,0x70,0x24]
|
|
vsri.16 q9, q8, #16
|
|
@ CHECK: vsri.32 q9, q8, #32 @ encoding: [0xe0,0xff,0x70,0x24]
|
|
vsri.32 q9, q8, #32
|
|
@ CHECK: vsri.64 q9, q8, #64 @ encoding: [0xc0,0xff,0xf0,0x24]
|
|
vsri.64 q9, q8, #64
|