| @ RUN: llvm-mc -mcpu=cortex-a8 -triple arm-unknown-unknown -show-encoding < %s | FileCheck %s |
| |
| _foo: |
| @ CHECK: vshl.u8 d16, d17, d16 @ encoding: [0xa1,0x04,0x40,0xf3] |
| vshl.u8 d16, d17, d16 |
| @ CHECK: vshl.u16 d16, d17, d16 @ encoding: [0xa1,0x04,0x50,0xf3] |
| vshl.u16 d16, d17, d16 |
| @ CHECK: vshl.u32 d16, d17, d16 @ encoding: [0xa1,0x04,0x60,0xf3] |
| vshl.u32 d16, d17, d16 |
| @ CHECK: vshl.u64 d16, d17, d16 @ encoding: [0xa1,0x04,0x70,0xf3] |
| vshl.u64 d16, d17, d16 |
| @ CHECK: vshl.i8 d16, d16, #7 @ encoding: [0x30,0x05,0xcf,0xf2] |
| vshl.i8 d16, d16, #7 |
| @ CHECK: vshl.i16 d16, d16, #15 @ encoding: [0x30,0x05,0xdf,0xf2] |
| vshl.i16 d16, d16, #15 |
| @ CHECK: vshl.i32 d16, d16, #31 @ encoding: [0x30,0x05,0xff,0xf2] |
| vshl.i32 d16, d16, #31 |
| @ CHECK: vshl.i64 d16, d16, #63 @ encoding: [0xb0,0x05,0xff,0xf2] |
| vshl.i64 d16, d16, #63 |
| @ CHECK: vshl.u8 q8, q9, q8 @ encoding: [0xe2,0x04,0x40,0xf3] |
| vshl.u8 q8, q9, q8 |
| @ CHECK: vshl.u16 q8, q9, q8 @ encoding: [0xe2,0x04,0x50,0xf3] |
| vshl.u16 q8, q9, q8 |
| @ CHECK: vshl.u32 q8, q9, q8 @ encoding: [0xe2,0x04,0x60,0xf3] |
| vshl.u32 q8, q9, q8 |
| @ CHECK: vshl.u64 q8, q9, q8 @ encoding: [0xe2,0x04,0x70,0xf3] |
| vshl.u64 q8, q9, q8 |
| @ CHECK: vshl.i8 q8, q8, #7 @ encoding: [0x70,0x05,0xcf,0xf2] |
| vshl.i8 q8, q8, #7 |
| @ CHECK: vshl.i16 q8, q8, #15 @ encoding: [0x70,0x05,0xdf,0xf2] |
| vshl.i16 q8, q8, #15 |
| @ CHECK: vshl.i32 q8, q8, #31 @ encoding: [0x70,0x05,0xff,0xf2] |
| vshl.i32 q8, q8, #31 |
| @ CHECK: vshl.i64 q8, q8, #63 @ encoding: [0xf0,0x05,0xff,0xf2] |
| vshl.i64 q8, q8, #63 |
| @ CHECK: vshr.u8 d16, d16, #7 @ encoding: [0x30,0x00,0xc9,0xf3] |
| vshr.u8 d16, d16, #7 |
| @ CHECK: vshr.u16 d16, d16, #15 @ encoding: [0x30,0x00,0xd1,0xf3] |
| vshr.u16 d16, d16, #15 |
| @ CHECK: vshr.u32 d16, d16, #31 @ encoding: [0x30,0x00,0xe1,0xf3] |
| vshr.u32 d16, d16, #31 |
| @ CHECK: vshr.u64 d16, d16, #63 @ encoding: [0xb0,0x00,0xc1,0xf3] |
| vshr.u64 d16, d16, #63 |
| @ CHECK: vshr.u8 q8, q8, #7 @ encoding: [0x70,0x00,0xc9,0xf3] |
| vshr.u8 q8, q8, #7 |
| @ CHECK: vshr.u16 q8, q8, #15 @ encoding: [0x70,0x00,0xd1,0xf3] |
| vshr.u16 q8, q8, #15 |
| @ CHECK: vshr.u32 q8, q8, #31 @ encoding: [0x70,0x00,0xe1,0xf3] |
| vshr.u32 q8, q8, #31 |
| @ CHECK: vshr.u64 q8, q8, #63 @ encoding: [0xf0,0x00,0xc1,0xf3] |
| vshr.u64 q8, q8, #63 |
| @ CHECK: vshr.s8 d16, d16, #7 @ encoding: [0x30,0x00,0xc9,0xf2] |
| vshr.s8 d16, d16, #7 |
| @ CHECK: vshr.s16 d16, d16, #15 @ encoding: [0x30,0x00,0xd1,0xf2] |
| vshr.s16 d16, d16, #15 |
| @ CHECK: vshr.s32 d16, d16, #31 @ encoding: [0x30,0x00,0xe1,0xf2] |
| vshr.s32 d16, d16, #31 |
| @ CHECK: vshr.s64 d16, d16, #63 @ encoding: [0xb0,0x00,0xc1,0xf2] |
| vshr.s64 d16, d16, #63 |
| @ CHECK: vshr.s8 q8, q8, #7 @ encoding: [0x70,0x00,0xc9,0xf2] |
| vshr.s8 q8, q8, #7 |
| @ CHECK: vshr.s16 q8, q8, #15 @ encoding: [0x70,0x00,0xd1,0xf2] |
| vshr.s16 q8, q8, #15 |
| @ CHECK: vshr.s32 q8, q8, #31 @ encoding: [0x70,0x00,0xe1,0xf2] |
| vshr.s32 q8, q8, #31 |
| @ CHECK: vshr.s64 q8, q8, #63 @ encoding: [0xf0,0x00,0xc1,0xf2] |
| vshr.s64 q8, q8, #63 |
| @ CHECK: vsra.u8 d16, d16, #7 @ encoding: [0x30,0x01,0xc9,0xf3] |
| vsra.u8 d16, d16, #7 |
| @ CHECK: vsra.u16 d16, d16, #15 @ encoding: [0x30,0x01,0xd1,0xf3] |
| vsra.u16 d16, d16, #15 |
| @ CHECK: vsra.u32 d16, d16, #31 @ encoding: [0x30,0x01,0xe1,0xf3] |
| vsra.u32 d16, d16, #31 |
| @ CHECK: vsra.u64 d16, d16, #63 @ encoding: [0xb0,0x01,0xc1,0xf3] |
| vsra.u64 d16, d16, #63 |
| @ CHECK: vsra.u8 q8, q8, #7 @ encoding: [0x70,0x01,0xc9,0xf3] |
| vsra.u8 q8, q8, #7 |
| @ CHECK: vsra.u16 q8, q8, #15 @ encoding: [0x70,0x01,0xd1,0xf3] |
| vsra.u16 q8, q8, #15 |
| @ CHECK: vsra.u32 q8, q8, #31 @ encoding: [0x70,0x01,0xe1,0xf3] |
| vsra.u32 q8, q8, #31 |
| @ CHECK: vsra.u64 q8, q8, #63 @ encoding: [0xf0,0x01,0xc1,0xf3] |
| vsra.u64 q8, q8, #63 |
| @ CHECK: vsra.s8 d16, d16, #7 @ encoding: [0x30,0x01,0xc9,0xf2] |
| vsra.s8 d16, d16, #7 |
| @ CHECK: vsra.s16 d16, d16, #15 @ encoding: [0x30,0x01,0xd1,0xf2] |
| vsra.s16 d16, d16, #15 |
| @ CHECK: vsra.s32 d16, d16, #31 @ encoding: [0x30,0x01,0xe1,0xf2] |
| vsra.s32 d16, d16, #31 |
| @ CHECK: vsra.s64 d16, d16, #63 @ encoding: [0xb0,0x01,0xc1,0xf2] |
| vsra.s64 d16, d16, #63 |
| @ CHECK: vsra.s8 q8, q8, #7 @ encoding: [0x70,0x01,0xc9,0xf2] |
| vsra.s8 q8, q8, #7 |
| @ CHECK: vsra.s16 q8, q8, #15 @ encoding: [0x70,0x01,0xd1,0xf2] |
| vsra.s16 q8, q8, #15 |
| @ CHECK: vsra.s32 q8, q8, #31 @ encoding: [0x70,0x01,0xe1,0xf2] |
| vsra.s32 q8, q8, #31 |
| @ CHECK: vsra.s64 q8, q8, #63 @ encoding: [0xf0,0x01,0xc1,0xf2] |
| vsra.s64 q8, q8, #63 |
| @ CHECK: vsri.8 d16, d16, #7 @ encoding: [0x30,0x04,0xc9,0xf3] |
| vsri.8 d16, d16, #7 |
| @ CHECK: vsri.16 d16, d16, #15 @ encoding: [0x30,0x04,0xd1,0xf3] |
| vsri.16 d16, d16, #15 |
| @ CHECK: vsri.32 d16, d16, #31 @ encoding: [0x30,0x04,0xe1,0xf3] |
| vsri.32 d16, d16, #31 |
| @ CHECK: vsri.64 d16, d16, #63 @ encoding: [0xb0,0x04,0xc1,0xf3] |
| vsri.64 d16, d16, #63 |
| @ CHECK: vsri.8 q8, q8, #7 @ encoding: [0x70,0x04,0xc9,0xf3] |
| vsri.8 q8, q8, #7 |
| @ CHECK: vsri.16 q8, q8, #15 @ encoding: [0x70,0x04,0xd1,0xf3] |
| vsri.16 q8, q8, #15 |
| @ CHECK: vsri.32 q8, q8, #31 @ encoding: [0x70,0x04,0xe1,0xf3] |
| vsri.32 q8, q8, #31 |
| @ CHECK: vsri.64 q8, q8, #63 @ encoding: [0xf0,0x04,0xc1,0xf3] |
| vsri.64 q8, q8, #63 |
| @ CHECK: vsli.8 d16, d16, #7 @ encoding: [0x30,0x05,0xcf,0xf3] |
| vsli.8 d16, d16, #7 |
| @ CHECK: vsli.16 d16, d16, #15 @ encoding: [0x30,0x05,0xdf,0xf3] |
| vsli.16 d16, d16, #15 |
| @ CHECK: vsli.32 d16, d16, #31 @ encoding: [0x30,0x05,0xff,0xf3] |
| vsli.32 d16, d16, #31 |
| @ CHECK: vsli.64 d16, d16, #63 @ encoding: [0xb0,0x05,0xff,0xf3] |
| vsli.64 d16, d16, #63 |
| @ CHECK: vsli.8 q8, q8, #7 @ encoding: [0x70,0x05,0xcf,0xf3] |
| vsli.8 q8, q8, #7 |
| @ CHECK: vsli.16 q8, q8, #15 @ encoding: [0x70,0x05,0xdf,0xf3] |
| vsli.16 q8, q8, #15 |
| @ CHECK: vsli.32 q8, q8, #31 @ encoding: [0x70,0x05,0xff,0xf3] |
| vsli.32 q8, q8, #31 |
| @ CHECK: vsli.64 q8, q8, #63 @ encoding: [0xf0,0x05,0xff,0xf3] |
| vsli.64 q8, q8, #63 |
| @ CHECK: vshll.s8 q8, d16, #7 @ encoding: [0x30,0x0a,0xcf,0xf2] |
| vshll.s8 q8, d16, #7 |
| @ CHECK: vshll.s16 q8, d16, #15 @ encoding: [0x30,0x0a,0xdf,0xf2] |
| vshll.s16 q8, d16, #15 |
| @ CHECK: vshll.s32 q8, d16, #31 @ encoding: [0x30,0x0a,0xff,0xf2] |
| vshll.s32 q8, d16, #31 |
| @ CHECK: vshll.u8 q8, d16, #7 @ encoding: [0x30,0x0a,0xcf,0xf3] |
| vshll.u8 q8, d16, #7 |
| @ CHECK: vshll.u16 q8, d16, #15 @ encoding: [0x30,0x0a,0xdf,0xf3] |
| vshll.u16 q8, d16, #15 |
| @ CHECK: vshll.u32 q8, d16, #31 @ encoding: [0x30,0x0a,0xff,0xf3] |
| vshll.u32 q8, d16, #31 |
| @ CHECK: vshll.i8 q8, d16, #8 @ encoding: [0x20,0x03,0xf2,0xf3] |
| vshll.i8 q8, d16, #8 |
| @ CHECK: vshll.i16 q8, d16, #16 @ encoding: [0x20,0x03,0xf6,0xf3] |
| vshll.i16 q8, d16, #16 |
| @ CHECK: vshll.i32 q8, d16, #32 @ encoding: [0x20,0x03,0xfa,0xf3] |
| vshll.i32 q8, d16, #32 |
| @ CHECK: vshrn.i16 d16, q8, #8 @ encoding: [0x30,0x08,0xc8,0xf2] |
| vshrn.i16 d16, q8, #8 |
| @ CHECK: vshrn.i32 d16, q8, #16 @ encoding: [0x30,0x08,0xd0,0xf2] |
| vshrn.i32 d16, q8, #16 |
| @ CHECK: vshrn.i64 d16, q8, #32 @ encoding: [0x30,0x08,0xe0,0xf2] |
| vshrn.i64 d16, q8, #32 |
| @ CHECK: vrshl.s8 d16, d17, d16 @ encoding: [0xa1,0x05,0x40,0xf2] |
| vrshl.s8 d16, d17, d16 |
| @ CHECK: vrshl.s16 d16, d17, d16 @ encoding: [0xa1,0x05,0x50,0xf2] |
| vrshl.s16 d16, d17, d16 |
| @ CHECK: vrshl.s32 d16, d17, d16 @ encoding: [0xa1,0x05,0x60,0xf2] |
| vrshl.s32 d16, d17, d16 |
| @ CHECK: vrshl.s64 d16, d17, d16 @ encoding: [0xa1,0x05,0x70,0xf2] |
| vrshl.s64 d16, d17, d16 |
| @ CHECK: vrshl.u8 d16, d17, d16 @ encoding: [0xa1,0x05,0x40,0xf3] |
| vrshl.u8 d16, d17, d16 |
| @ CHECK: vrshl.u16 d16, d17, d16 @ encoding: [0xa1,0x05,0x50,0xf3] |
| vrshl.u16 d16, d17, d16 |
| @ CHECK: vrshl.u32 d16, d17, d16 @ encoding: [0xa1,0x05,0x60,0xf3] |
| vrshl.u32 d16, d17, d16 |
| @ CHECK: vrshl.u64 d16, d17, d16 @ encoding: [0xa1,0x05,0x70,0xf3] |
| vrshl.u64 d16, d17, d16 |
| @ CHECK: vrshl.s8 q8, q9, q8 @ encoding: [0xe2,0x05,0x40,0xf2] |
| vrshl.s8 q8, q9, q8 |
| @ CHECK: vrshl.s16 q8, q9, q8 @ encoding: [0xe2,0x05,0x50,0xf2] |
| vrshl.s16 q8, q9, q8 |
| @ CHECK: vrshl.s32 q8, q9, q8 @ encoding: [0xe2,0x05,0x60,0xf2] |
| vrshl.s32 q8, q9, q8 |
| @ CHECK: vrshl.s64 q8, q9, q8 @ encoding: [0xe2,0x05,0x70,0xf2] |
| vrshl.s64 q8, q9, q8 |
| @ CHECK: vrshl.u8 q8, q9, q8 @ encoding: [0xe2,0x05,0x40,0xf3] |
| vrshl.u8 q8, q9, q8 |
| @ CHECK: vrshl.u16 q8, q9, q8 @ encoding: [0xe2,0x05,0x50,0xf3] |
| vrshl.u16 q8, q9, q8 |
| @ CHECK: vrshl.u32 q8, q9, q8 @ encoding: [0xe2,0x05,0x60,0xf3] |
| vrshl.u32 q8, q9, q8 |
| @ CHECK: vrshl.u64 q8, q9, q8 @ encoding: [0xe2,0x05,0x70,0xf3] |
| vrshl.u64 q8, q9, q8 |
| @ CHECK: vrshr.s8 d16, d16, #8 @ encoding: [0x30,0x02,0xc8,0xf2] |
| vrshr.s8 d16, d16, #8 |
| @ CHECK: vrshr.s16 d16, d16, #16 @ encoding: [0x30,0x02,0xd0,0xf2] |
| vrshr.s16 d16, d16, #16 |
| @ CHECK: vrshr.s32 d16, d16, #32 @ encoding: [0x30,0x02,0xe0,0xf2] |
| vrshr.s32 d16, d16, #32 |
| @ CHECK: vrshr.s64 d16, d16, #64 @ encoding: [0xb0,0x02,0xc0,0xf2] |
| vrshr.s64 d16, d16, #64 |
| @ CHECK: vrshr.u8 d16, d16, #8 @ encoding: [0x30,0x02,0xc8,0xf3] |
| vrshr.u8 d16, d16, #8 |
| @ CHECK: vrshr.u16 d16, d16, #16 @ encoding: [0x30,0x02,0xd0,0xf3] |
| vrshr.u16 d16, d16, #16 |
| @ CHECK: vrshr.u32 d16, d16, #32 @ encoding: [0x30,0x02,0xe0,0xf3] |
| vrshr.u32 d16, d16, #32 |
| @ CHECK: vrshr.u64 d16, d16, #64 @ encoding: [0xb0,0x02,0xc0,0xf3] |
| vrshr.u64 d16, d16, #64 |
| @ CHECK: vrshr.s8 q8, q8, #8 @ encoding: [0x70,0x02,0xc8,0xf2] |
| vrshr.s8 q8, q8, #8 |
| @ CHECK: vrshr.s16 q8, q8, #16 @ encoding: [0x70,0x02,0xd0,0xf2] |
| vrshr.s16 q8, q8, #16 |
| @ CHECK: vrshr.s32 q8, q8, #32 @ encoding: [0x70,0x02,0xe0,0xf2] |
| vrshr.s32 q8, q8, #32 |
| @ CHECK: vrshr.s64 q8, q8, #64 @ encoding: [0xf0,0x02,0xc0,0xf2] |
| vrshr.s64 q8, q8, #64 |
| @ CHECK: vrshr.u8 q8, q8, #8 @ encoding: [0x70,0x02,0xc8,0xf3] |
| vrshr.u8 q8, q8, #8 |
| @ CHECK: vrshr.u16 q8, q8, #16 @ encoding: [0x70,0x02,0xd0,0xf3] |
| vrshr.u16 q8, q8, #16 |
| @ CHECK: vrshr.u32 q8, q8, #32 @ encoding: [0x70,0x02,0xe0,0xf3] |
| vrshr.u32 q8, q8, #32 |
| @ CHECK: vrshr.u64 q8, q8, #64 @ encoding: [0xf0,0x02,0xc0,0xf3] |
| vrshr.u64 q8, q8, #64 |
| @ CHECK: vrshrn.i16 d16, q8, #8 @ encoding: [0x70,0x08,0xc8,0xf2] |
| vrshrn.i16 d16, q8, #8 |
| @ CHECK: vrshrn.i32 d16, q8, #16 @ encoding: [0x70,0x08,0xd0,0xf2] |
| vrshrn.i32 d16, q8, #16 |
| @ CHECK: vrshrn.i64 d16, q8, #32 @ encoding: [0x70,0x08,0xe0,0xf2] |
| vrshrn.i64 d16, q8, #32 |
| @ CHECK: vqrshrn.s16 d16, q8, #4 @ encoding: [0x70,0x09,0xcc,0xf2] |
| vqrshrn.s16 d16, q8, #4 |
| @ CHECK: vqrshrn.s32 d16, q8, #13 @ encoding: [0x70,0x09,0xd3,0xf2] |
| vqrshrn.s32 d16, q8, #13 |
| @ CHECK: vqrshrn.s64 d16, q8, #13 @ encoding: [0x70,0x09,0xf3,0xf2] |
| vqrshrn.s64 d16, q8, #13 |
| @ CHECK: vqrshrn.u16 d16, q8, #4 @ encoding: [0x70,0x09,0xcc,0xf3] |
| vqrshrn.u16 d16, q8, #4 |
| @ CHECK: vqrshrn.u32 d16, q8, #13 @ encoding: [0x70,0x09,0xd3,0xf3] |
| vqrshrn.u32 d16, q8, #13 |
| @ CHECK: vqrshrn.u64 d16, q8, #13 @ encoding: [0x70,0x09,0xf3,0xf3] |
| vqrshrn.u64 d16, q8, #13 |