| @ RUN: llvm-mc -mcpu=cortex-a8 -triple thumb-unknown-unknown -show-encoding < %s | FileCheck %s |
| |
| .code 16 |
| |
| vsra.s8 d17, d16, #8 |
| vsra.s16 d15, d14, #16 |
| vsra.s32 d13, d12, #32 |
| vsra.s64 d11, d10, #64 |
| vsra.s8 q7, q2, #8 |
| vsra.s16 q3, q6, #16 |
| vsra.s32 q9, q5, #32 |
| vsra.s64 q8, q4, #64 |
| vsra.u8 d17, d16, #8 |
| vsra.u16 d11, d14, #11 |
| vsra.u32 d12, d15, #22 |
| vsra.u64 d13, d16, #54 |
| vsra.u8 q1, q7, #8 |
| vsra.u16 q2, q7, #6 |
| vsra.u32 q3, q6, #21 |
| vsra.u64 q4, q5, #25 |
| |
| @ Two-operand syntax variant. |
| vsra.s8 d16, #8 |
| vsra.s16 d14, #16 |
| vsra.s32 d12, #32 |
| vsra.s64 d10, #64 |
| vsra.s8 q2, #8 |
| vsra.s16 q6, #16 |
| vsra.s32 q5, #32 |
| vsra.s64 q4, #64 |
| vsra.u8 d16, #8 |
| vsra.u16 d14, #11 |
| vsra.u32 d15, #22 |
| vsra.u64 d16, #54 |
| vsra.u8 q7, #8 |
| vsra.u16 q7, #6 |
| vsra.u32 q6, #21 |
| vsra.u64 q5, #25 |
| |
| @ CHECK: vsra.s8 d17, d16, #8 @ encoding: [0xc8,0xef,0x30,0x11] |
| @ CHECK: vsra.s16 d15, d14, #16 @ encoding: [0x90,0xef,0x1e,0xf1] |
| @ CHECK: vsra.s32 d13, d12, #32 @ encoding: [0xa0,0xef,0x1c,0xd1] |
| @ CHECK: vsra.s64 d11, d10, #64 @ encoding: [0x80,0xef,0x9a,0xb1] |
| @ CHECK: vsra.s8 q7, q2, #8 @ encoding: [0x88,0xef,0x54,0xe1] |
| @ CHECK: vsra.s16 q3, q6, #16 @ encoding: [0x90,0xef,0x5c,0x61] |
| @ CHECK: vsra.s32 q9, q5, #32 @ encoding: [0xe0,0xef,0x5a,0x21] |
| @ CHECK: vsra.s64 q8, q4, #64 @ encoding: [0xc0,0xef,0xd8,0x01] |
| @ CHECK: vsra.u8 d17, d16, #8 @ encoding: [0xc8,0xff,0x30,0x11] |
| @ CHECK: vsra.u16 d11, d14, #11 @ encoding: [0x95,0xff,0x1e,0xb1] |
| @ CHECK: vsra.u32 d12, d15, #22 @ encoding: [0xaa,0xff,0x1f,0xc1] |
| @ CHECK: vsra.u64 d13, d16, #54 @ encoding: [0x8a,0xff,0xb0,0xd1] |
| @ CHECK: vsra.u8 q1, q7, #8 @ encoding: [0x88,0xff,0x5e,0x21] |
| @ CHECK: vsra.u16 q2, q7, #6 @ encoding: [0x9a,0xff,0x5e,0x41] |
| @ CHECK: vsra.u32 q3, q6, #21 @ encoding: [0xab,0xff,0x5c,0x61] |
| @ CHECK: vsra.u64 q4, q5, #25 @ encoding: [0xa7,0xff,0xda,0x81] |
| |
| @ CHECK: vsra.s8 d16, d16, #8 @ encoding: [0xc8,0xef,0x30,0x01] |
| @ CHECK: vsra.s16 d14, d14, #16 @ encoding: [0x90,0xef,0x1e,0xe1] |
| @ CHECK: vsra.s32 d12, d12, #32 @ encoding: [0xa0,0xef,0x1c,0xc1] |
| @ CHECK: vsra.s64 d10, d10, #64 @ encoding: [0x80,0xef,0x9a,0xa1] |
| @ CHECK: vsra.s8 q2, q2, #8 @ encoding: [0x88,0xef,0x54,0x41] |
| @ CHECK: vsra.s16 q6, q6, #16 @ encoding: [0x90,0xef,0x5c,0xc1] |
| @ CHECK: vsra.s32 q5, q5, #32 @ encoding: [0xa0,0xef,0x5a,0xa1] |
| @ CHECK: vsra.s64 q4, q4, #64 @ encoding: [0x80,0xef,0xd8,0x81] |
| @ CHECK: vsra.u8 d16, d16, #8 @ encoding: [0xc8,0xff,0x30,0x01] |
| @ CHECK: vsra.u16 d14, d14, #11 @ encoding: [0x95,0xff,0x1e,0xe1] |
| @ CHECK: vsra.u32 d15, d15, #22 @ encoding: [0xaa,0xff,0x1f,0xf1] |
| @ CHECK: vsra.u64 d16, d16, #54 @ encoding: [0xca,0xff,0xb0,0x01] |
| @ CHECK: vsra.u8 q7, q7, #8 @ encoding: [0x88,0xff,0x5e,0xe1] |
| @ CHECK: vsra.u16 q7, q7, #6 @ encoding: [0x9a,0xff,0x5e,0xe1] |
| @ CHECK: vsra.u32 q6, q6, #21 @ encoding: [0xab,0xff,0x5c,0xc1] |
| @ CHECK: vsra.u64 q5, q5, #25 @ encoding: [0xa7,0xff,0xda,0xa1] |
| |
| |
| vrsra.s8 d5, d26, #8 |
| vrsra.s16 d6, d25, #16 |
| vrsra.s32 d7, d24, #32 |
| vrsra.s64 d14, d23, #64 |
| vrsra.u8 d15, d22, #8 |
| vrsra.u16 d16, d21, #16 |
| vrsra.u32 d17, d20, #32 |
| vrsra.u64 d18, d19, #64 |
| vrsra.s8 q1, q2, #8 |
| vrsra.s16 q2, q3, #16 |
| vrsra.s32 q3, q4, #32 |
| vrsra.s64 q4, q5, #64 |
| vrsra.u8 q5, q6, #8 |
| vrsra.u16 q6, q7, #16 |
| vrsra.u32 q7, q8, #32 |
| vrsra.u64 q8, q9, #64 |
| |
| @ Two-operand syntax variant. |
| vrsra.s8 d26, #8 |
| vrsra.s16 d25, #16 |
| vrsra.s32 d24, #32 |
| vrsra.s64 d23, #64 |
| vrsra.u8 d22, #8 |
| vrsra.u16 d21, #16 |
| vrsra.u32 d20, #32 |
| vrsra.u64 d19, #64 |
| vrsra.s8 q2, #8 |
| vrsra.s16 q3, #16 |
| vrsra.s32 q4, #32 |
| vrsra.s64 q5, #64 |
| vrsra.u8 q6, #8 |
| vrsra.u16 q7, #16 |
| vrsra.u32 q8, #32 |
| vrsra.u64 q9, #64 |
| |
| @ CHECK: vrsra.s8 d5, d26, #8 @ encoding: [0x88,0xef,0x3a,0x53] |
| @ CHECK: vrsra.s16 d6, d25, #16 @ encoding: [0x90,0xef,0x39,0x63] |
| @ CHECK: vrsra.s32 d7, d24, #32 @ encoding: [0xa0,0xef,0x38,0x73] |
| @ CHECK: vrsra.s64 d14, d23, #64 @ encoding: [0x80,0xef,0xb7,0xe3] |
| @ CHECK: vrsra.u8 d15, d22, #8 @ encoding: [0x88,0xff,0x36,0xf3] |
| @ CHECK: vrsra.u16 d16, d21, #16 @ encoding: [0xd0,0xff,0x35,0x03] |
| @ CHECK: vrsra.u32 d17, d20, #32 @ encoding: [0xe0,0xff,0x34,0x13] |
| @ CHECK: vrsra.u64 d18, d19, #64 @ encoding: [0xc0,0xff,0xb3,0x23] |
| @ CHECK: vrsra.s8 q1, q2, #8 @ encoding: [0x88,0xef,0x54,0x23] |
| @ CHECK: vrsra.s16 q2, q3, #16 @ encoding: [0x90,0xef,0x56,0x43] |
| @ CHECK: vrsra.s32 q3, q4, #32 @ encoding: [0xa0,0xef,0x58,0x63] |
| @ CHECK: vrsra.s64 q4, q5, #64 @ encoding: [0x80,0xef,0xda,0x83] |
| @ CHECK: vrsra.u8 q5, q6, #8 @ encoding: [0x88,0xff,0x5c,0xa3] |
| @ CHECK: vrsra.u16 q6, q7, #16 @ encoding: [0x90,0xff,0x5e,0xc3] |
| @ CHECK: vrsra.u32 q7, q8, #32 @ encoding: [0xa0,0xff,0x70,0xe3] |
| @ CHECK: vrsra.u64 q8, q9, #64 @ encoding: [0xc0,0xff,0xf2,0x03] |
| |
| @ CHECK: vrsra.s8 d26, d26, #8 @ encoding: [0xc8,0xef,0x3a,0xa3] |
| @ CHECK: vrsra.s16 d25, d25, #16 @ encoding: [0xd0,0xef,0x39,0x93] |
| @ CHECK: vrsra.s32 d24, d24, #32 @ encoding: [0xe0,0xef,0x38,0x83] |
| @ CHECK: vrsra.s64 d23, d23, #64 @ encoding: [0xc0,0xef,0xb7,0x73] |
| @ CHECK: vrsra.u8 d22, d22, #8 @ encoding: [0xc8,0xff,0x36,0x63] |
| @ CHECK: vrsra.u16 d21, d21, #16 @ encoding: [0xd0,0xff,0x35,0x53] |
| @ CHECK: vrsra.u32 d20, d20, #32 @ encoding: [0xe0,0xff,0x34,0x43] |
| @ CHECK: vrsra.u64 d19, d19, #64 @ encoding: [0xc0,0xff,0xb3,0x33] |
| @ CHECK: vrsra.s8 q2, q2, #8 @ encoding: [0x88,0xef,0x54,0x43] |
| @ CHECK: vrsra.s16 q3, q3, #16 @ encoding: [0x90,0xef,0x56,0x63] |
| @ CHECK: vrsra.s32 q4, q4, #32 @ encoding: [0xa0,0xef,0x58,0x83] |
| @ CHECK: vrsra.s64 q5, q5, #64 @ encoding: [0x80,0xef,0xda,0xa3] |
| @ CHECK: vrsra.u8 q6, q6, #8 @ encoding: [0x88,0xff,0x5c,0xc3] |
| @ CHECK: vrsra.u16 q7, q7, #16 @ encoding: [0x90,0xff,0x5e,0xe3] |
| @ CHECK: vrsra.u32 q8, q8, #32 @ encoding: [0xe0,0xff,0x70,0x03] |
| @ CHECK: vrsra.u64 q9, q9, #64 @ encoding: [0xc0,0xff,0xf2,0x23] |
| |
| |
| vsli.8 d11, d12, #7 |
| vsli.16 d12, d13, #15 |
| vsli.32 d13, d14, #31 |
| vsli.64 d14, d15, #63 |
| vsli.8 q1, q8, #7 |
| vsli.16 q2, q7, #15 |
| vsli.32 q3, q4, #31 |
| vsli.64 q4, q5, #63 |
| vsri.8 d28, d11, #8 |
| vsri.16 d26, d12, #16 |
| vsri.32 d24, d13, #32 |
| vsri.64 d21, d14, #64 |
| vsri.8 q1, q8, #8 |
| vsri.16 q5, q2, #16 |
| vsri.32 q7, q4, #32 |
| vsri.64 q9, q6, #64 |
| |
| @ Two-operand syntax variant. |
| vsli.8 d12, #7 |
| vsli.16 d13, #15 |
| vsli.32 d14, #31 |
| vsli.64 d15, #63 |
| vsli.8 q8, #7 |
| vsli.16 q7, #15 |
| vsli.32 q4, #31 |
| vsli.64 q5, #63 |
| vsri.8 d11, #8 |
| vsri.16 d12, #16 |
| vsri.32 d13, #32 |
| vsri.64 d14, #64 |
| vsri.8 q8, #8 |
| vsri.16 q2, #16 |
| vsri.32 q4, #32 |
| vsri.64 q6, #64 |
| |
| @ CHECK: vsli.8 d11, d12, #7 @ encoding: [0x8f,0xff,0x1c,0xb5] |
| @ CHECK: vsli.16 d12, d13, #15 @ encoding: [0x9f,0xff,0x1d,0xc5] |
| @ CHECK: vsli.32 d13, d14, #31 @ encoding: [0xbf,0xff,0x1e,0xd5] |
| @ CHECK: vsli.64 d14, d15, #63 @ encoding: [0xbf,0xff,0x9f,0xe5] |
| @ CHECK: vsli.8 q1, q8, #7 @ encoding: [0x8f,0xff,0x70,0x25] |
| @ CHECK: vsli.16 q2, q7, #15 @ encoding: [0x9f,0xff,0x5e,0x45] |
| @ CHECK: vsli.32 q3, q4, #31 @ encoding: [0xbf,0xff,0x58,0x65] |
| @ CHECK: vsli.64 q4, q5, #63 @ encoding: [0xbf,0xff,0xda,0x85] |
| @ CHECK: vsri.8 d28, d11, #8 @ encoding: [0xc8,0xff,0x1b,0xc4] |
| @ CHECK: vsri.16 d26, d12, #16 @ encoding: [0xd0,0xff,0x1c,0xa4] |
| @ CHECK: vsri.32 d24, d13, #32 @ encoding: [0xe0,0xff,0x1d,0x84] |
| @ CHECK: vsri.64 d21, d14, #64 @ encoding: [0xc0,0xff,0x9e,0x54] |
| @ CHECK: vsri.8 q1, q8, #8 @ encoding: [0x88,0xff,0x70,0x24] |
| @ CHECK: vsri.16 q5, q2, #16 @ encoding: [0x90,0xff,0x54,0xa4] |
| @ CHECK: vsri.32 q7, q4, #32 @ encoding: [0xa0,0xff,0x58,0xe4] |
| @ CHECK: vsri.64 q9, q6, #64 @ encoding: [0xc0,0xff,0xdc,0x24] |
| |
| @ CHECK: vsli.8 d12, d12, #7 @ encoding: [0x8f,0xff,0x1c,0xc5] |
| @ CHECK: vsli.16 d13, d13, #15 @ encoding: [0x9f,0xff,0x1d,0xd5] |
| @ CHECK: vsli.32 d14, d14, #31 @ encoding: [0xbf,0xff,0x1e,0xe5] |
| @ CHECK: vsli.64 d15, d15, #63 @ encoding: [0xbf,0xff,0x9f,0xf5] |
| @ CHECK: vsli.8 q8, q8, #7 @ encoding: [0xcf,0xff,0x70,0x05] |
| @ CHECK: vsli.16 q7, q7, #15 @ encoding: [0x9f,0xff,0x5e,0xe5] |
| @ CHECK: vsli.32 q4, q4, #31 @ encoding: [0xbf,0xff,0x58,0x85] |
| @ CHECK: vsli.64 q5, q5, #63 @ encoding: [0xbf,0xff,0xda,0xa5] |
| @ CHECK: vsri.8 d11, d11, #8 @ encoding: [0x88,0xff,0x1b,0xb4] |
| @ CHECK: vsri.16 d12, d12, #16 @ encoding: [0x90,0xff,0x1c,0xc4] |
| @ CHECK: vsri.32 d13, d13, #32 @ encoding: [0xa0,0xff,0x1d,0xd4] |
| @ CHECK: vsri.64 d14, d14, #64 @ encoding: [0x80,0xff,0x9e,0xe4] |
| @ CHECK: vsri.8 q8, q8, #8 @ encoding: [0xc8,0xff,0x70,0x04] |
| @ CHECK: vsri.16 q2, q2, #16 @ encoding: [0x90,0xff,0x54,0x44] |
| @ CHECK: vsri.32 q4, q4, #32 @ encoding: [0xa0,0xff,0x58,0x84] |
| @ CHECK: vsri.64 q6, q6, #64 @ encoding: [0x80,0xff,0xdc,0xc4] |