| ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=tahiti -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,MOVREL,PREGFX9 %s |
| ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=tonga -mattr=-flat-for-global -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,MOVREL,PREGFX9 %s |
| ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=tonga -mattr=-flat-for-global -amdgpu-vgpr-index-mode -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,IDXMODE,PREGFX9 %s |
| ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=gfx900 -mattr=-flat-for-global -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,IDXMODE,GFX9 %s |
| |
| ; Tests for indirect addressing on SI, which is implemented using dynamic |
| ; indexing of vectors. |
| |
| ; GCN-LABEL: {{^}}extract_w_offset: |
| ; GCN-DAG: s_load_dword [[IN:s[0-9]+]] |
| ; GCN-DAG: v_mov_b32_e32 v{{[0-9]+}}, 4.0 |
| ; GCN-DAG: v_mov_b32_e32 v{{[0-9]+}}, 0x40400000 |
| ; GCN-DAG: v_mov_b32_e32 [[BASEREG:v[0-9]+]], 2.0 |
| ; GCN-DAG: v_mov_b32_e32 v{{[0-9]+}}, 1.0 |
| |
| ; MOVREL-DAG: s_mov_b32 m0, [[IN]] |
| ; MOVREL: v_movrels_b32_e32 v{{[0-9]+}}, [[BASEREG]] |
| |
| ; IDXMODE: s_set_gpr_idx_on [[IN]], src0{{$}} |
| ; IDXMODE-NEXT: v_mov_b32_e32 v{{[0-9]+}}, [[BASEREG]] |
| ; IDXMODE-NEXT: s_set_gpr_idx_off |
| define amdgpu_kernel void @extract_w_offset(float addrspace(1)* %out, i32 %in) { |
| entry: |
| %idx = add i32 %in, 1 |
| %elt = extractelement <4 x float> <float 1.0, float 2.0, float 3.0, float 4.0>, i32 %idx |
| store float %elt, float addrspace(1)* %out |
| ret void |
| } |
| |
| ; XXX: Could do v_or_b32 directly |
| ; GCN-LABEL: {{^}}extract_w_offset_salu_use_vector: |
| ; MOVREL: s_mov_b32 m0 |
| ; GCN-DAG: s_or_b32 |
| ; GCN-DAG: s_or_b32 |
| ; GCN-DAG: s_or_b32 |
| ; GCN-DAG: s_or_b32 |
| ; GCN-DAG: v_mov_b32_e32 v{{[0-9]+}}, s{{[0-9]+}} |
| ; GCN-DAG: v_mov_b32_e32 v{{[0-9]+}}, s{{[0-9]+}} |
| ; GCN-DAG: v_mov_b32_e32 v{{[0-9]+}}, s{{[0-9]+}} |
| ; GCN-DAG: v_mov_b32_e32 v{{[0-9]+}}, s{{[0-9]+}} |
| |
| ; MOVREL: v_movrels_b32_e32 |
| |
| ; IDXMODE: s_set_gpr_idx_on s{{[0-9]+}}, src0{{$}} |
| ; IDXMODE-NEXT: v_mov_b32_e32 v{{[0-9]+}}, v{{[0-9]+}} |
| ; IDXMODE-NEXT: s_set_gpr_idx_off |
| define amdgpu_kernel void @extract_w_offset_salu_use_vector(i32 addrspace(1)* %out, i32 %in, <4 x i32> %or.val) { |
| entry: |
| %idx = add i32 %in, 1 |
| %vec = or <4 x i32> %or.val, <i32 1, i32 2, i32 3, i32 4> |
| %elt = extractelement <4 x i32> %vec, i32 %idx |
| store i32 %elt, i32 addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}extract_wo_offset: |
| ; GCN-DAG: s_load_dword [[IN:s[0-9]+]] |
| ; GCN-DAG: v_mov_b32_e32 v{{[0-9]+}}, 4.0 |
| ; GCN-DAG: v_mov_b32_e32 v{{[0-9]+}}, 0x40400000 |
| ; GCN-DAG: v_mov_b32_e32 v{{[0-9]+}}, 2.0 |
| ; GCN-DAG: v_mov_b32_e32 [[BASEREG:v[0-9]+]], 1.0 |
| |
| ; MOVREL-DAG: s_mov_b32 m0, [[IN]] |
| ; MOVREL: v_movrels_b32_e32 v{{[0-9]+}}, [[BASEREG]] |
| |
| ; IDXMODE: s_set_gpr_idx_on [[IN]], src0{{$}} |
| ; IDXMODE-NEXT: v_mov_b32_e32 v{{[0-9]+}}, [[BASEREG]] |
| ; IDXMODE-NEXT: s_set_gpr_idx_off |
| define amdgpu_kernel void @extract_wo_offset(float addrspace(1)* %out, i32 %in) { |
| entry: |
| %elt = extractelement <4 x float> <float 1.0, float 2.0, float 3.0, float 4.0>, i32 %in |
| store float %elt, float addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}extract_neg_offset_sgpr: |
| ; The offset depends on the register that holds the first element of the vector. |
| ; MOVREL: s_add_i32 m0, s{{[0-9]+}}, 0xfffffe{{[0-9a-z]+}} |
| ; MOVREL: v_movrels_b32_e32 v{{[0-9]}}, v0 |
| |
| ; IDXMODE: s_addk_i32 [[ADD_IDX:s[0-9]+]], 0xfe00{{$}} |
| ; IDXMODE: v_mov_b32_e32 v2, 2 |
| ; IDXMODE: v_mov_b32_e32 v3, 3 |
| ; IDXMODE-NEXT: s_set_gpr_idx_on [[ADD_IDX]], src0{{$}} |
| ; IDXMODE-NEXT: v_mov_b32_e32 v{{[0-9]+}}, v{{[0-9]+}} |
| ; IDXMODE-NEXT: s_set_gpr_idx_off |
| define amdgpu_kernel void @extract_neg_offset_sgpr(i32 addrspace(1)* %out, i32 %offset) { |
| entry: |
| %index = add i32 %offset, -512 |
| %value = extractelement <4 x i32> <i32 0, i32 1, i32 2, i32 3>, i32 %index |
| store i32 %value, i32 addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}extract_neg_offset_sgpr_loaded: |
| ; The offset depends on the register that holds the first element of the vector. |
| ; MOVREL: s_add_i32 m0, s{{[0-9]+}}, 0xfffffe{{[0-9a-z]+}} |
| ; MOVREL: v_movrels_b32_e32 v{{[0-9]}}, v0 |
| |
| ; IDXMODE: s_addk_i32 [[ADD_IDX:s[0-9]+]], 0xfe00{{$}} |
| ; IDXMODE: v_mov_b32_e32 v0, |
| ; IDXMODE: v_mov_b32_e32 v1, |
| ; IDXMODE: v_mov_b32_e32 v2, |
| ; IDXMODE: v_mov_b32_e32 v3, |
| ; IDXMODE-NEXT: s_set_gpr_idx_on [[ADD_IDX]], src0{{$}} |
| ; IDXMODE-NEXT: v_mov_b32_e32 v{{[0-9]+}}, v{{[0-9]+}} |
| ; IDXMODE-NEXT: s_set_gpr_idx_off |
| define amdgpu_kernel void @extract_neg_offset_sgpr_loaded(i32 addrspace(1)* %out, <4 x i32> %vec0, <4 x i32> %vec1, i32 %offset) { |
| entry: |
| %index = add i32 %offset, -512 |
| %or = or <4 x i32> %vec0, %vec1 |
| %value = extractelement <4 x i32> %or, i32 %index |
| store i32 %value, i32 addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}extract_neg_offset_vgpr: |
| ; The offset depends on the register that holds the first element of the vector. |
| |
| ; FIXME: The waitcnt for the argument load can go after the loop |
| ; GCN: s_mov_b64 s{{\[[0-9]+:[0-9]+\]}}, exec |
| ; GCN: [[LOOPBB:BB[0-9]+_[0-9]+]]: |
| ; GCN: v_readfirstlane_b32 [[READLANE:s[0-9]+]], v{{[0-9]+}} |
| ; GCN: s_and_saveexec_b64 vcc, vcc |
| |
| ; MOVREL: s_add_i32 m0, [[READLANE]], 0xfffffe0 |
| ; MOVREL: v_movrels_b32_e32 [[RESULT:v[0-9]+]], v1 |
| |
| ; IDXMODE: s_addk_i32 [[ADD_IDX:s[0-9]+]], 0xfe00 |
| ; IDXMODE: s_set_gpr_idx_on [[ADD_IDX]], src0 |
| ; IDXMODE: v_mov_b32_e32 [[RESULT:v[0-9]+]], v1 |
| ; IDXMODE: s_set_gpr_idx_off |
| |
| ; GCN: s_cbranch_execnz |
| |
| ; GCN: buffer_store_dword [[RESULT]] |
| define amdgpu_kernel void @extract_neg_offset_vgpr(i32 addrspace(1)* %out) { |
| entry: |
| %id = call i32 @llvm.amdgcn.workitem.id.x() #1 |
| %index = add i32 %id, -512 |
| %value = extractelement <4 x i32> <i32 0, i32 1, i32 2, i32 3>, i32 %index |
| store i32 %value, i32 addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}extract_undef_offset_sgpr: |
| ; undefined behavior, but shouldn't crash compiler |
| define amdgpu_kernel void @extract_undef_offset_sgpr(i32 addrspace(1)* %out, <4 x i32> addrspace(1)* %in) { |
| entry: |
| %ld = load volatile <4 x i32>, <4 x i32> addrspace(1)* %in |
| %value = extractelement <4 x i32> %ld, i32 undef |
| store i32 %value, i32 addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}insert_undef_offset_sgpr_vector_src: |
| ; undefined behavior, but shouldn't crash compiler |
| define amdgpu_kernel void @insert_undef_offset_sgpr_vector_src(<4 x i32> addrspace(1)* %out, <4 x i32> addrspace(1)* %in) { |
| entry: |
| %ld = load <4 x i32>, <4 x i32> addrspace(1)* %in |
| %value = insertelement <4 x i32> %ld, i32 5, i32 undef |
| store <4 x i32> %value, <4 x i32> addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}insert_w_offset: |
| ; GCN-DAG: s_load_dword [[IN:s[0-9]+]] |
| ; MOVREL-DAG: s_mov_b32 m0, [[IN]] |
| ; GCN-DAG: v_mov_b32_e32 v[[ELT0:[0-9]+]], 1.0 |
| ; GCN-DAG: v_mov_b32_e32 v[[ELT1:[0-9]+]], 2.0 |
| ; GCN-DAG: v_mov_b32_e32 v[[ELT2:[0-9]+]], 0x40400000 |
| ; GCN-DAG: v_mov_b32_e32 v[[ELT3:[0-9]+]], 4.0 |
| ; GCN-DAG: v_mov_b32_e32 v[[INS:[0-9]+]], 0x40a00000 |
| |
| ; MOVREL: v_movreld_b32_e32 v[[ELT1]], v[[INS]] |
| ; MOVREL: buffer_store_dwordx4 v{{\[}}[[ELT0]]:[[ELT3]]{{\]}} |
| define amdgpu_kernel void @insert_w_offset(<4 x float> addrspace(1)* %out, i32 %in) { |
| entry: |
| %0 = add i32 %in, 1 |
| %1 = insertelement <4 x float> <float 1.0, float 2.0, float 3.0, float 4.0>, float 5.0, i32 %0 |
| store <4 x float> %1, <4 x float> addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}insert_wo_offset: |
| ; GCN: s_load_dword [[IN:s[0-9]+]] |
| |
| ; MOVREL: s_mov_b32 m0, [[IN]] |
| ; MOVREL: v_movreld_b32_e32 v[[ELT0:[0-9]+]] |
| |
| ; IDXMODE: s_set_gpr_idx_on [[IN]], dst |
| ; IDXMODE-NEXT: v_mov_b32_e32 v[[ELT0:[0-9]+]], v{{[0-9]+}} |
| ; IDXMODE-NEXT: s_set_gpr_idx_off |
| |
| ; GCN: buffer_store_dwordx4 v{{\[}}[[ELT0]]: |
| define amdgpu_kernel void @insert_wo_offset(<4 x float> addrspace(1)* %out, i32 %in) { |
| entry: |
| %0 = insertelement <4 x float> <float 1.0, float 2.0, float 3.0, float 4.0>, float 5.0, i32 %in |
| store <4 x float> %0, <4 x float> addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}insert_neg_offset_sgpr: |
| ; The offset depends on the register that holds the first element of the vector. |
| ; MOVREL: s_add_i32 m0, s{{[0-9]+}}, 0xfffffe{{[0-9a-z]+}} |
| ; MOVREL: v_movreld_b32_e32 v0, 5 |
| |
| ; IDXMODE: s_addk_i32 [[ADD_IDX:s[0-9]+]], 0xfe00{{$}} |
| ; IDXMODE: s_set_gpr_idx_on [[ADD_IDX]], dst |
| ; IDXMODE-NEXT: v_mov_b32_e32 v0, 5 |
| ; IDXMODE-NEXT: s_set_gpr_idx_off |
| define amdgpu_kernel void @insert_neg_offset_sgpr(i32 addrspace(1)* %in, <4 x i32> addrspace(1)* %out, i32 %offset) { |
| entry: |
| %index = add i32 %offset, -512 |
| %value = insertelement <4 x i32> <i32 0, i32 1, i32 2, i32 3>, i32 5, i32 %index |
| store <4 x i32> %value, <4 x i32> addrspace(1)* %out |
| ret void |
| } |
| |
| ; The vector indexed into is originally loaded into an SGPR rather |
| ; than built with a reg_sequence |
| |
| ; GCN-LABEL: {{^}}insert_neg_offset_sgpr_loadreg: |
| ; The offset depends on the register that holds the first element of the vector. |
| ; MOVREL: s_add_i32 m0, s{{[0-9]+}}, 0xfffffe{{[0-9a-z]+}} |
| ; MOVREL: v_movreld_b32_e32 v0, 5 |
| |
| ; IDXMODE: s_addk_i32 [[ADD_IDX:s[0-9]+]], 0xfe00{{$}} |
| ; IDXMODE: s_set_gpr_idx_on [[ADD_IDX]], dst |
| ; IDXMODE-NEXT: v_mov_b32_e32 v0, 5 |
| ; IDXMODE-NEXT: s_set_gpr_idx_off |
| define amdgpu_kernel void @insert_neg_offset_sgpr_loadreg(i32 addrspace(1)* %in, <4 x i32> addrspace(1)* %out, <4 x i32> %vec, i32 %offset) { |
| entry: |
| %index = add i32 %offset, -512 |
| %value = insertelement <4 x i32> %vec, i32 5, i32 %index |
| store <4 x i32> %value, <4 x i32> addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}insert_neg_offset_vgpr: |
| ; The offset depends on the register that holds the first element of the vector. |
| |
| ; GCN-DAG: v_mov_b32_e32 [[VEC_ELT0:v[0-9]+]], 1{{$}} |
| ; GCN-DAG: v_mov_b32_e32 [[VEC_ELT1:v[0-9]+]], 2{{$}} |
| ; GCN-DAG: v_mov_b32_e32 [[VEC_ELT2:v[0-9]+]], 3{{$}} |
| ; GCN-DAG: v_mov_b32_e32 [[VEC_ELT3:v[0-9]+]], 4{{$}} |
| |
| ; GCN: s_mov_b64 [[SAVEEXEC:s\[[0-9]+:[0-9]+\]]], exec |
| ; GCN: [[LOOPBB:BB[0-9]+_[0-9]+]]: |
| ; GCN: v_readfirstlane_b32 [[READLANE:s[0-9]+]] |
| ; GCN: s_and_saveexec_b64 vcc, vcc |
| |
| ; MOVREL: s_add_i32 m0, [[READLANE]], 0xfffffe00 |
| ; MOVREL: v_movreld_b32_e32 [[VEC_ELT0]], 5 |
| |
| ; IDXMODE: s_addk_i32 [[ADD_IDX:s[0-9]+]], 0xfe00{{$}} |
| ; IDXMODE: s_set_gpr_idx_on [[ADD_IDX]], dst |
| ; IDXMODE: v_mov_b32_e32 v{{[0-9]+}}, 5 |
| ; IDXMODE: s_set_gpr_idx_off |
| |
| ; GCN: s_cbranch_execnz [[LOOPBB]] |
| ; GCN: s_mov_b64 exec, [[SAVEEXEC]] |
| |
| ; GCN: buffer_store_dword |
| define amdgpu_kernel void @insert_neg_offset_vgpr(i32 addrspace(1)* %in, <4 x i32> addrspace(1)* %out) { |
| entry: |
| %id = call i32 @llvm.amdgcn.workitem.id.x() #1 |
| %index = add i32 %id, -512 |
| %value = insertelement <4 x i32> <i32 1, i32 2, i32 3, i32 4>, i32 5, i32 %index |
| store <4 x i32> %value, <4 x i32> addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}insert_neg_inline_offset_vgpr: |
| |
| ; GCN-DAG: v_mov_b32_e32 [[VEC_ELT0:v[0-9]+]], 1{{$}} |
| ; GCN-DAG: v_mov_b32_e32 [[VEC_ELT1:v[0-9]+]], 2{{$}} |
| ; GCN-DAG: v_mov_b32_e32 [[VEC_ELT2:v[0-9]+]], 3{{$}} |
| ; GCN-DAG: v_mov_b32_e32 [[VEC_ELT3:v[0-9]+]], 4{{$}} |
| ; GCN-DAG: v_mov_b32_e32 [[VAL:v[0-9]+]], 0x1f4{{$}} |
| |
| ; GCN: s_mov_b64 [[SAVEEXEC:s\[[0-9]+:[0-9]+\]]], exec |
| |
| ; The offset depends on the register that holds the first element of the vector. |
| ; GCN: v_readfirstlane_b32 [[READLANE:s[0-9]+]] |
| |
| ; MOVREL: s_add_i32 m0, [[READLANE]], -16 |
| ; MOVREL: v_movreld_b32_e32 [[VEC_ELT0]], [[VAL]] |
| |
| ; IDXMODE: s_add_i32 [[ADD_IDX:s[0-9]+]], [[READLANE]], -16 |
| ; IDXMODE: s_set_gpr_idx_on [[ADD_IDX]], dst |
| ; IDXMODE: v_mov_b32_e32 [[VEC_ELT0]], [[VAL]] |
| ; IDXMODE: s_set_gpr_idx_off |
| |
| ; GCN: s_cbranch_execnz |
| define amdgpu_kernel void @insert_neg_inline_offset_vgpr(i32 addrspace(1)* %in, <4 x i32> addrspace(1)* %out) { |
| entry: |
| %id = call i32 @llvm.amdgcn.workitem.id.x() #1 |
| %index = add i32 %id, -16 |
| %value = insertelement <4 x i32> <i32 1, i32 2, i32 3, i32 4>, i32 500, i32 %index |
| store <4 x i32> %value, <4 x i32> addrspace(1)* %out |
| ret void |
| } |
| |
| ; When the block is split to insert the loop, make sure any other |
| ; places that need to be expanded in the same block are also handled. |
| |
| ; GCN-LABEL: {{^}}extract_vgpr_offset_multiple_in_block: |
| |
| ; FIXME: Why is vector copied in between? |
| |
| ; GCN-DAG: {{buffer|flat|global}}_load_dword [[IDX0:v[0-9]+]] |
| ; GCN-DAG: s_mov_b32 [[S_ELT1:s[0-9]+]], 9 |
| ; GCN-DAG: s_mov_b32 [[S_ELT0:s[0-9]+]], 7 |
| ; GCN-DAG: v_mov_b32_e32 [[VEC_ELT0:v[0-9]+]], [[S_ELT0]] |
| ; GCN-DAG: v_mov_b32_e32 [[VEC_ELT1:v[0-9]+]], [[S_ELT1]] |
| |
| ; GCN: s_mov_b64 [[MASK:s\[[0-9]+:[0-9]+\]]], exec |
| |
| ; GCN: [[LOOP0:BB[0-9]+_[0-9]+]]: |
| ; GCN-NEXT: s_waitcnt vmcnt(0) |
| ; GCN-NEXT: v_readfirstlane_b32 [[READLANE:s[0-9]+]], [[IDX0]] |
| ; GCN: v_cmp_eq_u32_e32 vcc, [[READLANE]], [[IDX0]] |
| ; GCN: s_and_saveexec_b64 vcc, vcc |
| |
| ; MOVREL: s_mov_b32 m0, [[READLANE]] |
| ; MOVREL: v_movrels_b32_e32 [[MOVREL0:v[0-9]+]], [[VEC_ELT0]] |
| |
| ; IDXMODE: s_set_gpr_idx_on [[READLANE]], src0 |
| ; IDXMODE: v_mov_b32_e32 [[MOVREL0:v[0-9]+]], [[VEC_ELT0]] |
| ; IDXMODE: s_set_gpr_idx_off |
| |
| ; GCN-NEXT: s_xor_b64 exec, exec, vcc |
| ; GCN-NEXT: s_cbranch_execnz [[LOOP0]] |
| |
| ; FIXME: Redundant copy |
| ; GCN: s_mov_b64 exec, [[MASK]] |
| |
| ; GCN: v_mov_b32_e32 [[VEC_ELT1_2:v[0-9]+]], [[S_ELT1]] |
| |
| ; GCN: s_mov_b64 [[MASK2:s\[[0-9]+:[0-9]+\]]], exec |
| |
| ; GCN: [[LOOP1:BB[0-9]+_[0-9]+]]: |
| ; GCN-NEXT: v_readfirstlane_b32 [[READLANE:s[0-9]+]], [[IDX0]] |
| ; GCN: v_cmp_eq_u32_e32 vcc, [[READLANE]], [[IDX0]] |
| ; GCN: s_and_saveexec_b64 vcc, vcc |
| |
| ; MOVREL: s_mov_b32 m0, [[READLANE]] |
| ; MOVREL-NEXT: v_movrels_b32_e32 [[MOVREL1:v[0-9]+]], [[VEC_ELT1_2]] |
| |
| ; IDXMODE: s_set_gpr_idx_on [[READLANE]], src0 |
| ; IDXMODE-NEXT: v_mov_b32_e32 [[MOVREL1:v[0-9]+]], [[VEC_ELT1_2]] |
| ; IDXMODE: s_set_gpr_idx_off |
| |
| ; GCN-NEXT: s_xor_b64 exec, exec, vcc |
| ; GCN: s_cbranch_execnz [[LOOP1]] |
| |
| ; GCN: buffer_store_dword [[MOVREL0]] |
| ; GCN: buffer_store_dword [[MOVREL1]] |
| define amdgpu_kernel void @extract_vgpr_offset_multiple_in_block(i32 addrspace(1)* %out0, i32 addrspace(1)* %out1, i32 addrspace(1)* %in) #0 { |
| entry: |
| %id = call i32 @llvm.amdgcn.workitem.id.x() #1 |
| %id.ext = zext i32 %id to i64 |
| %gep = getelementptr inbounds i32, i32 addrspace(1)* %in, i64 %id.ext |
| %idx0 = load volatile i32, i32 addrspace(1)* %gep |
| %idx1 = add i32 %idx0, 1 |
| %val0 = extractelement <4 x i32> <i32 7, i32 9, i32 11, i32 13>, i32 %idx0 |
| %live.out.reg = call i32 asm sideeffect "s_mov_b32 $0, 17", "={s4}" () |
| %val1 = extractelement <4 x i32> <i32 7, i32 9, i32 11, i32 13>, i32 %idx1 |
| store volatile i32 %val0, i32 addrspace(1)* %out0 |
| store volatile i32 %val1, i32 addrspace(1)* %out0 |
| %cmp = icmp eq i32 %id, 0 |
| br i1 %cmp, label %bb1, label %bb2 |
| |
| bb1: |
| store volatile i32 %live.out.reg, i32 addrspace(1)* undef |
| br label %bb2 |
| |
| bb2: |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}insert_vgpr_offset_multiple_in_block: |
| ; GCN-DAG: s_load_dwordx4 s{{\[}}[[S_ELT0:[0-9]+]]:[[S_ELT3:[0-9]+]]{{\]}} |
| ; GCN-DAG: {{buffer|flat|global}}_load_dword [[IDX0:v[0-9]+]] |
| ; GCN-DAG: v_mov_b32 [[INS0:v[0-9]+]], 62 |
| |
| ; GCN-DAG: v_mov_b32_e32 v[[VEC_ELT3:[0-9]+]], s[[S_ELT3]] |
| ; GCN: v_mov_b32_e32 v[[VEC_ELT2:[0-9]+]], s{{[0-9]+}} |
| ; GCN: v_mov_b32_e32 v[[VEC_ELT1:[0-9]+]], s{{[0-9]+}} |
| ; GCN: v_mov_b32_e32 v[[VEC_ELT0:[0-9]+]], s[[S_ELT0]] |
| |
| ; GCN: [[LOOP0:BB[0-9]+_[0-9]+]]: |
| ; GCN-NEXT: s_waitcnt vmcnt(0) |
| ; GCN-NEXT: v_readfirstlane_b32 [[READLANE:s[0-9]+]], [[IDX0]] |
| ; GCN: v_cmp_eq_u32_e32 vcc, [[READLANE]], [[IDX0]] |
| ; GCN: s_and_saveexec_b64 vcc, vcc |
| |
| ; MOVREL: s_mov_b32 m0, [[READLANE]] |
| ; MOVREL-NEXT: v_movreld_b32_e32 v[[VEC_ELT0]], [[INS0]] |
| |
| ; IDXMODE: s_set_gpr_idx_on [[READLANE]], dst |
| ; IDXMODE-NEXT: v_mov_b32_e32 v[[VEC_ELT0]], [[INS0]] |
| ; IDXMODE: s_set_gpr_idx_off |
| |
| ; GCN-NEXT: s_xor_b64 exec, exec, vcc |
| ; GCN: s_cbranch_execnz [[LOOP0]] |
| |
| ; FIXME: Redundant copy |
| ; GCN: s_mov_b64 exec, [[MASK:s\[[0-9]+:[0-9]+\]]] |
| |
| ; GCN: s_mov_b64 [[MASK]], exec |
| |
| ; GCN: [[LOOP1:BB[0-9]+_[0-9]+]]: |
| ; GCN-NEXT: v_readfirstlane_b32 [[READLANE:s[0-9]+]], [[IDX0]] |
| ; GCN: v_cmp_eq_u32_e32 vcc, [[READLANE]], [[IDX0]] |
| ; GCN: s_and_saveexec_b64 vcc, vcc |
| |
| ; MOVREL: s_mov_b32 m0, [[READLANE]] |
| ; MOVREL-NEXT: v_movreld_b32_e32 v[[VEC_ELT1]], 63 |
| |
| ; IDXMODE: s_set_gpr_idx_on [[READLANE]], dst |
| ; IDXMODE-NEXT: v_mov_b32_e32 v[[VEC_ELT1]], 63 |
| ; IDXMODE: s_set_gpr_idx_off |
| |
| ; GCN-NEXT: s_xor_b64 exec, exec, vcc |
| ; GCN: s_cbranch_execnz [[LOOP1]] |
| |
| ; GCN: buffer_store_dwordx4 v{{\[}}[[VEC_ELT0]]: |
| |
| ; GCN: buffer_store_dword [[INS0]] |
| define amdgpu_kernel void @insert_vgpr_offset_multiple_in_block(<4 x i32> addrspace(1)* %out0, <4 x i32> addrspace(1)* %out1, i32 addrspace(1)* %in, <4 x i32> %vec0) #0 { |
| entry: |
| %id = call i32 @llvm.amdgcn.workitem.id.x() #1 |
| %id.ext = zext i32 %id to i64 |
| %gep = getelementptr inbounds i32, i32 addrspace(1)* %in, i64 %id.ext |
| %idx0 = load volatile i32, i32 addrspace(1)* %gep |
| %idx1 = add i32 %idx0, 1 |
| %live.out.val = call i32 asm sideeffect "v_mov_b32 $0, 62", "=v"() |
| %vec1 = insertelement <4 x i32> %vec0, i32 %live.out.val, i32 %idx0 |
| %vec2 = insertelement <4 x i32> %vec1, i32 63, i32 %idx1 |
| store volatile <4 x i32> %vec2, <4 x i32> addrspace(1)* %out0 |
| %cmp = icmp eq i32 %id, 0 |
| br i1 %cmp, label %bb1, label %bb2 |
| |
| bb1: |
| store volatile i32 %live.out.val, i32 addrspace(1)* undef |
| br label %bb2 |
| |
| bb2: |
| ret void |
| } |
| |
| |
| ; GCN-LABEL: {{^}}insert_adjacent_blocks: |
| define amdgpu_kernel void @insert_adjacent_blocks(i32 %arg, float %val0) #0 { |
| bb: |
| %tmp = icmp eq i32 %arg, 0 |
| br i1 %tmp, label %bb1, label %bb4 |
| |
| bb1: ; preds = %bb |
| %tmp2 = load volatile <4 x float>, <4 x float> addrspace(1)* undef |
| %tmp3 = insertelement <4 x float> %tmp2, float %val0, i32 undef |
| call void asm sideeffect "; reg use $0", "v"(<4 x float> %tmp3) #0 ; Prevent block optimize out |
| br label %bb7 |
| |
| bb4: ; preds = %bb |
| %tmp5 = load volatile <4 x float>, <4 x float> addrspace(1)* undef |
| %tmp6 = insertelement <4 x float> %tmp5, float %val0, i32 undef |
| call void asm sideeffect "; reg use $0", "v"(<4 x float> %tmp6) #0 ; Prevent block optimize out |
| br label %bb7 |
| |
| bb7: ; preds = %bb4, %bb1 |
| %tmp8 = phi <4 x float> [ %tmp3, %bb1 ], [ %tmp6, %bb4 ] |
| store volatile <4 x float> %tmp8, <4 x float> addrspace(1)* undef |
| ret void |
| } |
| |
| ; FIXME: Should be able to fold zero input to movreld to inline imm? |
| |
| ; GCN-LABEL: {{^}}multi_same_block: |
| |
| ; GCN: s_load_dword [[ARG:s[0-9]+]] |
| |
| ; MOVREL: v_mov_b32_e32 v{{[0-9]+}}, 0x41900000 |
| ; MOVREL: v_mov_b32_e32 v{{[0-9]+}}, 0x41b0cccd |
| ; MOVREL: s_waitcnt |
| ; MOVREL: s_add_i32 m0, [[ARG]], -16 |
| ; MOVREL: v_movreld_b32_e32 v{{[0-9]+}}, 4.0 |
| ; MOVREL: v_movreld_b32_e32 v{{[0-9]+}}, -4.0 |
| ; MOVREL: s_mov_b32 m0, -1 |
| |
| |
| ; IDXMODE: v_mov_b32_e32 v{{[0-9]+}}, 0x41900000 |
| ; IDXMODE: s_waitcnt |
| ; IDXMODE: s_add_i32 [[ARG]], [[ARG]], -16 |
| ; IDXMODE: s_set_gpr_idx_on [[ARG]], dst |
| ; IDXMODE: v_mov_b32_e32 v{{[0-9]+}}, 4.0 |
| ; IDXMODE: s_set_gpr_idx_off |
| ; IDXMODE: v_mov_b32_e32 v{{[0-9]+}}, 0x41b0cccd |
| ; IDXMODE: s_set_gpr_idx_on [[ARG]], dst |
| ; IDXMODE: v_mov_b32_e32 v{{[0-9]+}}, -4.0 |
| ; IDXMODE: s_set_gpr_idx_off |
| |
| ; GCN: ds_write_b32 |
| ; GCN: ds_write_b32 |
| ; GCN: s_endpgm |
| define amdgpu_kernel void @multi_same_block(i32 %arg) #0 { |
| bb: |
| %tmp1 = add i32 %arg, -16 |
| %tmp2 = insertelement <6 x float> <float 1.700000e+01, float 1.800000e+01, float 1.900000e+01, float 2.000000e+01, float 2.100000e+01, float 2.200000e+01>, float 4.000000e+00, i32 %tmp1 |
| %tmp3 = add i32 %arg, -16 |
| %tmp4 = insertelement <6 x float> <float 0x40311999A0000000, float 0x40321999A0000000, float 0x40331999A0000000, float 0x40341999A0000000, float 0x40351999A0000000, float 0x40361999A0000000>, float -4.0, i32 %tmp3 |
| %tmp5 = bitcast <6 x float> %tmp2 to <6 x i32> |
| %tmp6 = extractelement <6 x i32> %tmp5, i32 1 |
| %tmp7 = bitcast <6 x float> %tmp4 to <6 x i32> |
| %tmp8 = extractelement <6 x i32> %tmp7, i32 5 |
| store volatile i32 %tmp6, i32 addrspace(3)* undef, align 4 |
| store volatile i32 %tmp8, i32 addrspace(3)* undef, align 4 |
| ret void |
| } |
| |
| ; offset puts outside of superegister bounaries, so clamp to 1st element. |
| ; GCN-LABEL: {{^}}extract_largest_inbounds_offset: |
| ; GCN-DAG: buffer_load_dwordx4 v{{\[}}[[LO_ELT:[0-9]+]]:[[HI_ELT:[0-9]+]]{{\]}} |
| ; GCN-DAG: s_load_dword [[IDX:s[0-9]+]] |
| ; MOVREL: s_mov_b32 m0, [[IDX]] |
| ; MOVREL: v_movrels_b32_e32 [[EXTRACT:v[0-9]+]], v[[HI_ELT]] |
| |
| ; IDXMODE: s_set_gpr_idx_on [[IDX]], src0 |
| ; IDXMODE: v_mov_b32_e32 [[EXTRACT:v[0-9]+]], v[[HI_ELT]] |
| ; IDXMODE: s_set_gpr_idx_off |
| |
| ; GCN: buffer_store_dword [[EXTRACT]] |
| define amdgpu_kernel void @extract_largest_inbounds_offset(i32 addrspace(1)* %out, <4 x i32> addrspace(1)* %in, i32 %idx) { |
| entry: |
| %ld = load volatile <4 x i32>, <4 x i32> addrspace(1)* %in |
| %offset = add i32 %idx, 3 |
| %value = extractelement <4 x i32> %ld, i32 %offset |
| store i32 %value, i32 addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}extract_out_of_bounds_offset: |
| ; GCN-DAG: buffer_load_dwordx4 v{{\[}}[[LO_ELT:[0-9]+]]:[[HI_ELT:[0-9]+]]{{\]}} |
| ; GCN-DAG: s_load_dword [[IDX:s[0-9]+]] |
| ; MOVREL: s_add_i32 m0, [[IDX]], 4 |
| ; MOVREL: v_movrels_b32_e32 [[EXTRACT:v[0-9]+]], v[[LO_ELT]] |
| |
| ; IDXMODE: s_add_i32 [[ADD_IDX:s[0-9]+]], [[IDX]], 4 |
| ; IDXMODE: s_set_gpr_idx_on [[ADD_IDX]], src0 |
| ; IDXMODE: v_mov_b32_e32 [[EXTRACT:v[0-9]+]], v[[LO_ELT]] |
| ; IDXMODE: s_set_gpr_idx_off |
| |
| ; GCN: buffer_store_dword [[EXTRACT]] |
| define amdgpu_kernel void @extract_out_of_bounds_offset(i32 addrspace(1)* %out, <4 x i32> addrspace(1)* %in, i32 %idx) { |
| entry: |
| %ld = load volatile <4 x i32>, <4 x i32> addrspace(1)* %in |
| %offset = add i32 %idx, 4 |
| %value = extractelement <4 x i32> %ld, i32 %offset |
| store i32 %value, i32 addrspace(1)* %out |
| ret void |
| } |
| |
| ; Test that the or is folded into the base address register instead of |
| ; added to m0 |
| |
| ; GCN-LABEL: {{^}}extractelement_v4i32_or_index: |
| ; GCN: s_load_dword [[IDX_IN:s[0-9]+]] |
| ; GCN: s_lshl_b32 [[IDX_SHL:s[0-9]+]], [[IDX_IN]] |
| ; GCN-NOT: [[IDX_SHL]] |
| |
| ; MOVREL: s_mov_b32 m0, [[IDX_SHL]] |
| ; MOVREL: v_movrels_b32_e32 v{{[0-9]+}}, v{{[0-9]+}} |
| |
| ; IDXMODE: s_set_gpr_idx_on [[IDX_SHL]], src0 |
| ; IDXMODE: v_mov_b32_e32 v{{[0-9]+}}, v{{[0-9]+}} |
| ; IDXMODE: s_set_gpr_idx_off |
| define amdgpu_kernel void @extractelement_v4i32_or_index(i32 addrspace(1)* %out, <4 x i32> addrspace(1)* %in, i32 %idx.in) { |
| entry: |
| %ld = load volatile <4 x i32>, <4 x i32> addrspace(1)* %in |
| %idx.shl = shl i32 %idx.in, 2 |
| %idx = or i32 %idx.shl, 1 |
| %value = extractelement <4 x i32> %ld, i32 %idx |
| store i32 %value, i32 addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}insertelement_v4f32_or_index: |
| ; GCN: s_load_dword [[IDX_IN:s[0-9]+]] |
| ; GCN: s_lshl_b32 [[IDX_SHL:s[0-9]+]], [[IDX_IN]] |
| ; GCN-NOT: [[IDX_SHL]] |
| |
| ; MOVREL: s_mov_b32 m0, [[IDX_SHL]] |
| ; MOVREL: v_movreld_b32_e32 v{{[0-9]+}}, v{{[0-9]+}} |
| |
| ; IDXMODE: s_set_gpr_idx_on [[IDX_SHL]], dst |
| ; IDXMODE: v_mov_b32_e32 v{{[0-9]+}}, v{{[0-9]+}} |
| ; IDXMODE: s_set_gpr_idx_off |
| define amdgpu_kernel void @insertelement_v4f32_or_index(<4 x float> addrspace(1)* %out, <4 x float> %a, i32 %idx.in) nounwind { |
| %idx.shl = shl i32 %idx.in, 2 |
| %idx = or i32 %idx.shl, 1 |
| %vecins = insertelement <4 x float> %a, float 5.000000e+00, i32 %idx |
| store <4 x float> %vecins, <4 x float> addrspace(1)* %out, align 16 |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}broken_phi_bb: |
| ; GCN: v_mov_b32_e32 [[PHIREG:v[0-9]+]], 8 |
| |
| ; GCN: s_branch [[BB2:BB[0-9]+_[0-9]+]] |
| |
| ; GCN: {{^BB[0-9]+_[0-9]+}}: |
| ; GCN: s_mov_b64 exec, |
| |
| ; GCN: [[BB2]]: |
| ; GCN: v_cmp_le_i32_e32 vcc, s{{[0-9]+}}, [[PHIREG]] |
| ; GCN: buffer_load_dword |
| |
| ; GCN: [[REGLOOP:BB[0-9]+_[0-9]+]]: |
| ; MOVREL: v_movreld_b32_e32 |
| |
| ; IDXMODE: s_set_gpr_idx_on |
| ; IDXMODE: v_mov_b32_e32 |
| ; IDXMODE: s_set_gpr_idx_off |
| |
| ; GCN: s_cbranch_execnz [[REGLOOP]] |
| define amdgpu_kernel void @broken_phi_bb(i32 %arg, i32 %arg1) #0 { |
| bb: |
| br label %bb2 |
| |
| bb2: ; preds = %bb4, %bb |
| %tmp = phi i32 [ 8, %bb ], [ %tmp7, %bb4 ] |
| %tmp3 = icmp slt i32 %tmp, %arg |
| br i1 %tmp3, label %bb4, label %bb8 |
| |
| bb4: ; preds = %bb2 |
| %vgpr = load volatile i32, i32 addrspace(1)* undef |
| %tmp5 = insertelement <8 x i32> undef, i32 undef, i32 %vgpr |
| %tmp6 = insertelement <8 x i32> %tmp5, i32 %arg1, i32 %vgpr |
| %tmp7 = extractelement <8 x i32> %tmp6, i32 0 |
| br label %bb2 |
| |
| bb8: ; preds = %bb2 |
| ret void |
| } |
| |
| declare i32 @llvm.amdgcn.workitem.id.x() #1 |
| declare void @llvm.amdgcn.s.barrier() #2 |
| |
| attributes #0 = { nounwind } |
| attributes #1 = { nounwind readnone } |
| attributes #2 = { nounwind convergent } |