| ; XUN: llc -mtriple=amdgcn-amd-amdhsa -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefixes=GCN,SI %s |
| ; RUN: llc -mtriple=amdgcn-amd-amdhsa -mcpu=tonga -mattr=+fp64-fp16-denormals,-flat-for-global -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefixes=GCN,VI,VI-DENORM %s |
| ; RUN: llc -mtriple=amdgcn-amd-amdhsa -mcpu=tonga -mattr=-fp64-fp16-denormals,-flat-for-global -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefixes=GCN,VI,VI-FLUSH %s |
| |
| |
| ; Make sure (fmul (fadd x, x), c) -> (fmul x, (fmul 2.0, c)) doesn't |
| ; make add an instruction if the fadd has more than one use. |
| |
| declare half @llvm.fabs.f16(half) #1 |
| declare float @llvm.fabs.f32(float) #1 |
| |
| ; GCN-LABEL: {{^}}multiple_fadd_use_test_f32: |
| ; SI: v_max_legacy_f32_e64 [[A16:v[0-9]+]], |
| ; SI: v_add_f32_e32 [[A17:v[0-9]+]], [[A16]], [[A16]] |
| ; SI: v_mul_f32_e32 [[A18:v[0-9]+]], [[A17]], [[A17]] |
| ; SI: v_mad_f32 [[A20:v[0-9]+]], -[[A18]], [[A17]], 1.0 |
| ; SI: buffer_store_dword [[A20]] |
| |
| ; VI: v_add_f32_e64 v{{[0-9]+}}, s{{[0-9]+}}, -1.0 |
| ; VI: v_add_f32_e64 v{{[0-9]+}}, s{{[0-9]+}}, -1.0 |
| ; VI: v_cmp_gt_f32_e64 vcc, |v{{[0-9]+}}|, |v{{[0-9]+}}| |
| ; VI: v_cndmask_b32_e32 |
| ; VI: v_add_f32_e64 v{{[0-9]+}}, |v{{[0-9]+}}|, |v{{[0-9]+}}| |
| ; VI: v_mul_f32_e32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}} |
| ; VI: v_mad_f32 v{{[0-9]+}}, -v{{[0-9]+}}, v{{[0-9]+}}, 1.0 |
| define amdgpu_kernel void @multiple_fadd_use_test_f32(float addrspace(1)* %out, float %x, float %y, float %z) #0 { |
| %a11 = fadd fast float %y, -1.0 |
| %a12 = call float @llvm.fabs.f32(float %a11) |
| %a13 = fadd fast float %x, -1.0 |
| %a14 = call float @llvm.fabs.f32(float %a13) |
| %a15 = fcmp ogt float %a12, %a14 |
| %a16 = select i1 %a15, float %a12, float %a14 |
| %a17 = fmul fast float %a16, 2.0 |
| %a18 = fmul fast float %a17, %a17 |
| %a19 = fmul fast float %a18, %a17 |
| %a20 = fsub fast float 1.0, %a19 |
| store float %a20, float addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}multiple_use_fadd_fmac_f32: |
| ; GCN-DAG: v_add_f32_e64 [[MUL2:v[0-9]+]], [[X:s[0-9]+]], s{{[0-9]+}} |
| ; GCN-DAG: v_mac_f32_e64 [[MAD:v[0-9]+]], [[X]], 2.0 |
| ; GCN-DAG: buffer_store_dword [[MUL2]] |
| ; GCN-DAG: buffer_store_dword [[MAD]] |
| ; GCN: s_endpgm |
| define amdgpu_kernel void @multiple_use_fadd_fmac_f32(float addrspace(1)* %out, float %x, [8 x i32], float %y) #0 { |
| %out.gep.1 = getelementptr float, float addrspace(1)* %out, i32 1 |
| %mul2 = fmul fast float %x, 2.0 |
| %mad = fadd fast float %mul2, %y |
| store volatile float %mul2, float addrspace(1)* %out |
| store volatile float %mad, float addrspace(1)* %out.gep.1 |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}multiple_use_fadd_fmad_f32: |
| ; GCN-DAG: v_add_f32_e64 [[MUL2:v[0-9]+]], |[[X:s[0-9]+]]|, |s{{[0-9]+}}| |
| ; GCN-DAG: v_mad_f32 [[MAD:v[0-9]+]], |[[X]]|, 2.0, v{{[0-9]+}} |
| ; GCN-DAG: buffer_store_dword [[MUL2]] |
| ; GCN-DAG: buffer_store_dword [[MAD]] |
| ; GCN: s_endpgm |
| define amdgpu_kernel void @multiple_use_fadd_fmad_f32(float addrspace(1)* %out, float %x, float %y) #0 { |
| %out.gep.1 = getelementptr float, float addrspace(1)* %out, i32 1 |
| %x.abs = call float @llvm.fabs.f32(float %x) |
| %mul2 = fmul fast float %x.abs, 2.0 |
| %mad = fadd fast float %mul2, %y |
| store volatile float %mul2, float addrspace(1)* %out |
| store volatile float %mad, float addrspace(1)* %out.gep.1 |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}multiple_use_fadd_multi_fmad_f32: |
| ; GCN: v_mad_f32 {{v[0-9]+}}, |[[X:s[0-9]+]]|, 2.0, v{{[0-9]+}} |
| ; GCN: v_mad_f32 {{v[0-9]+}}, |[[X]]|, 2.0, v{{[0-9]+}} |
| define amdgpu_kernel void @multiple_use_fadd_multi_fmad_f32(float addrspace(1)* %out, float %x, float %y, float %z) #0 { |
| %out.gep.1 = getelementptr float, float addrspace(1)* %out, i32 1 |
| %x.abs = call float @llvm.fabs.f32(float %x) |
| %mul2 = fmul fast float %x.abs, 2.0 |
| %mad0 = fadd fast float %mul2, %y |
| %mad1 = fadd fast float %mul2, %z |
| store volatile float %mad0, float addrspace(1)* %out |
| store volatile float %mad1, float addrspace(1)* %out.gep.1 |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}fmul_x2_xn2_f32: |
| ; GCN: v_mul_f32_e64 [[TMP0:v[0-9]+]], [[X:s[0-9]+]], -4.0 |
| ; GCN: v_mul_f32_e32 [[RESULT:v[0-9]+]], [[X]], [[TMP0]] |
| ; GCN: buffer_store_dword [[RESULT]] |
| define amdgpu_kernel void @fmul_x2_xn2_f32(float addrspace(1)* %out, float %x, float %y) #0 { |
| %out.gep.1 = getelementptr float, float addrspace(1)* %out, i32 1 |
| %mul2 = fmul fast float %x, 2.0 |
| %muln2 = fmul fast float %x, -2.0 |
| %mul = fmul fast float %mul2, %muln2 |
| store volatile float %mul, float addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}fmul_x2_xn3_f32: |
| ; GCN: v_mov_b32_e32 [[K:v[0-9]+]], 0xc0c00000 |
| ; GCN: v_mul_f32_e32 [[TMP0:v[0-9]+]], [[X:s[0-9]+]], [[K]] |
| ; GCN: v_mul_f32_e32 [[RESULT:v[0-9]+]], [[X]], [[TMP0]] |
| ; GCN: buffer_store_dword [[RESULT]] |
| define amdgpu_kernel void @fmul_x2_xn3_f32(float addrspace(1)* %out, float %x, float %y) #0 { |
| %out.gep.1 = getelementptr float, float addrspace(1)* %out, i32 1 |
| %mul2 = fmul fast float %x, 2.0 |
| %muln2 = fmul fast float %x, -3.0 |
| %mul = fmul fast float %mul2, %muln2 |
| store volatile float %mul, float addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}multiple_fadd_use_test_f16: |
| ; VI: v_add_f16_e64 v{{[0-9]+}}, s{{[0-9]+}}, -1.0 |
| ; VI: v_add_f16_e64 v{{[0-9]+}}, s{{[0-9]+}}, -1.0 |
| ; VI: v_cmp_gt_f16_e64 vcc, |v{{[0-9]+}}|, |v{{[0-9]+}}| |
| ; VI: v_cndmask_b32_e32 |
| ; VI: v_add_f16_e64 v{{[0-9]+}}, |v{{[0-9]+}}|, |v{{[0-9]+}}| |
| ; VI: v_mul_f16_e32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}} |
| ; VI-FLUSH: v_mad_f16 v{{[0-9]+}}, -v{{[0-9]+}}, v{{[0-9]+}}, 1.0 |
| ; VI-DENORM: v_fma_f16 v{{[0-9]+}}, -v{{[0-9]+}}, v{{[0-9]+}}, 1.0 |
| define amdgpu_kernel void @multiple_fadd_use_test_f16(half addrspace(1)* %out, i16 zeroext %x.arg, i16 zeroext %y.arg, i16 zeroext %z.arg) #0 { |
| %x = bitcast i16 %x.arg to half |
| %y = bitcast i16 %y.arg to half |
| %z = bitcast i16 %z.arg to half |
| %a11 = fadd fast half %y, -1.0 |
| %a12 = call half @llvm.fabs.f16(half %a11) |
| %a13 = fadd fast half %x, -1.0 |
| %a14 = call half @llvm.fabs.f16(half %a13) |
| %a15 = fcmp ogt half %a12, %a14 |
| %a16 = select i1 %a15, half %a12, half %a14 |
| %a17 = fmul fast half %a16, 2.0 |
| %a18 = fmul fast half %a17, %a17 |
| %a19 = fmul fast half %a18, %a17 |
| %a20 = fsub fast half 1.0, %a19 |
| store half %a20, half addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}multiple_use_fadd_fmac_f16: |
| ; GCN-DAG: v_add_f16_e64 [[MUL2:v[0-9]+]], [[X:s[0-9]+]], s{{[0-9]+}} |
| |
| ; VI-FLUSH-DAG: v_mac_f16_e64 [[MAD:v[0-9]+]], [[X]], 2.0 |
| ; VI-DENORM-DAG: v_fma_f16 [[MAD:v[0-9]+]], [[X]], 2.0, v{{[0-9]+}} |
| |
| ; GCN-DAG: buffer_store_short [[MUL2]] |
| ; GCN-DAG: buffer_store_short [[MAD]] |
| ; GCN: s_endpgm |
| define amdgpu_kernel void @multiple_use_fadd_fmac_f16(half addrspace(1)* %out, i16 zeroext %x.arg, i16 zeroext %y.arg) #0 { |
| %x = bitcast i16 %x.arg to half |
| %y = bitcast i16 %y.arg to half |
| %out.gep.1 = getelementptr half, half addrspace(1)* %out, i32 1 |
| %mul2 = fmul fast half %x, 2.0 |
| %mad = fadd fast half %mul2, %y |
| store volatile half %mul2, half addrspace(1)* %out |
| store volatile half %mad, half addrspace(1)* %out.gep.1 |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}multiple_use_fadd_fmad_f16: |
| ; GCN-DAG: v_add_f16_e64 [[MUL2:v[0-9]+]], |[[X:s[0-9]+]]|, |s{{[0-9]+}}| |
| |
| ; VI-FLUSH-DAG: v_mad_f16 [[MAD:v[0-9]+]], |[[X]]|, 2.0, v{{[0-9]+}} |
| ; VI-DENORM-DAG: v_fma_f16 [[MAD:v[0-9]+]], |[[X]]|, 2.0, v{{[0-9]+}} |
| |
| ; GCN-DAG: buffer_store_short [[MUL2]] |
| ; GCN-DAG: buffer_store_short [[MAD]] |
| ; GCN: s_endpgm |
| define amdgpu_kernel void @multiple_use_fadd_fmad_f16(half addrspace(1)* %out, i16 zeroext %x.arg, i16 zeroext %y.arg) #0 { |
| %x = bitcast i16 %x.arg to half |
| %y = bitcast i16 %y.arg to half |
| %out.gep.1 = getelementptr half, half addrspace(1)* %out, i32 1 |
| %x.abs = call half @llvm.fabs.f16(half %x) |
| %mul2 = fmul fast half %x.abs, 2.0 |
| %mad = fadd fast half %mul2, %y |
| store volatile half %mul2, half addrspace(1)* %out |
| store volatile half %mad, half addrspace(1)* %out.gep.1 |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}multiple_use_fadd_multi_fmad_f16: |
| ; VI-FLUSH: v_mad_f16 {{v[0-9]+}}, |[[X:s[0-9]+]]|, 2.0, v{{[0-9]+}} |
| ; VI-FLUSH: v_mad_f16 {{v[0-9]+}}, |[[X]]|, 2.0, v{{[0-9]+}} |
| |
| ; VI-DENORM: v_fma_f16 {{v[0-9]+}}, |[[X:s[0-9]+]]|, 2.0, v{{[0-9]+}} |
| ; VI-DENORM: v_fma_f16 {{v[0-9]+}}, |[[X]]|, 2.0, v{{[0-9]+}} |
| |
| define amdgpu_kernel void @multiple_use_fadd_multi_fmad_f16(half addrspace(1)* %out, i16 zeroext %x.arg, i16 zeroext %y.arg, i16 zeroext %z.arg) #0 { |
| %x = bitcast i16 %x.arg to half |
| %y = bitcast i16 %y.arg to half |
| %z = bitcast i16 %z.arg to half |
| %out.gep.1 = getelementptr half, half addrspace(1)* %out, i32 1 |
| %x.abs = call half @llvm.fabs.f16(half %x) |
| %mul2 = fmul fast half %x.abs, 2.0 |
| %mad0 = fadd fast half %mul2, %y |
| %mad1 = fadd fast half %mul2, %z |
| store volatile half %mad0, half addrspace(1)* %out |
| store volatile half %mad1, half addrspace(1)* %out.gep.1 |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}fmul_x2_xn2_f16: |
| ; GCN: v_mul_f16_e64 [[TMP0:v[0-9]+]], [[X:s[0-9]+]], -4.0 |
| ; GCN: v_mul_f16_e32 [[RESULT:v[0-9]+]], [[X]], [[TMP0]] |
| ; GCN: buffer_store_short [[RESULT]] |
| define amdgpu_kernel void @fmul_x2_xn2_f16(half addrspace(1)* %out, i16 zeroext %x.arg, i16 zeroext %y.arg) #0 { |
| %x = bitcast i16 %x.arg to half |
| %y = bitcast i16 %y.arg to half |
| %out.gep.1 = getelementptr half, half addrspace(1)* %out, i32 1 |
| %mul2 = fmul fast half %x, 2.0 |
| %muln2 = fmul fast half %x, -2.0 |
| %mul = fmul fast half %mul2, %muln2 |
| store volatile half %mul, half addrspace(1)* %out |
| ret void |
| } |
| |
| ; GCN-LABEL: {{^}}fmul_x2_xn3_f16: |
| ; GCN: v_mov_b32_e32 [[K:v[0-9]+]], 0xc600 |
| ; GCN: v_mul_f16_e32 [[TMP0:v[0-9]+]], [[X:s[0-9]+]], [[K]] |
| ; GCN: v_mul_f16_e32 [[RESULT:v[0-9]+]], [[X]], [[TMP0]] |
| ; GCN: buffer_store_short [[RESULT]] |
| define amdgpu_kernel void @fmul_x2_xn3_f16(half addrspace(1)* %out, i16 zeroext %x.arg, i16 zeroext %y.arg) #0 { |
| %x = bitcast i16 %x.arg to half |
| %y = bitcast i16 %y.arg to half |
| %out.gep.1 = getelementptr half, half addrspace(1)* %out, i32 1 |
| %mul2 = fmul fast half %x, 2.0 |
| %muln2 = fmul fast half %x, -3.0 |
| %mul = fmul fast half %mul2, %muln2 |
| store volatile half %mul, half addrspace(1)* %out |
| ret void |
| } |
| |
| attributes #0 = { nounwind "unsafe-fp-math"="true" } |
| attributes #1 = { nounwind readnone } |