| ; RUN: llc -march=hexagon -O2 < %s |
| ; REQUIRES: asserts |
| |
| define inreg <16 x i32> @f0(i32 %a0, <16 x i32>* nocapture %a1) #0 { |
| b0: |
| %v0 = tail call <512 x i1> @llvm.hexagon.V6.pred.scalar2(i32 %a0) |
| %v1 = tail call <512 x i1> @llvm.hexagon.V6.pred.not(<512 x i1> %v0) |
| %v2 = icmp ult i32 %a0, 48 |
| br i1 %v2, label %b1, label %b2 |
| |
| b1: ; preds = %b0 |
| %v3 = add nuw nsw i32 %a0, 16 |
| %v4 = tail call <512 x i1> @llvm.hexagon.V6.pred.scalar2(i32 %v3) |
| %v5 = tail call <512 x i1> @llvm.hexagon.V6.pred.and(<512 x i1> %v4, <512 x i1> %v1) |
| br label %b2 |
| |
| b2: ; preds = %b1, %b0 |
| %v6 = phi <512 x i1> [ %v5, %b1 ], [ %v1, %b0 ] |
| %v7 = bitcast <512 x i1> %v6 to <16 x i32> |
| %v8 = getelementptr inbounds <16 x i32>, <16 x i32>* %a1, i32 1 |
| %v9 = load <16 x i32>, <16 x i32>* %v8, align 64 |
| %v10 = getelementptr inbounds <16 x i32>, <16 x i32>* %a1, i32 2 |
| %v11 = load <16 x i32>, <16 x i32>* %v10, align 64 |
| %v12 = tail call <16 x i32> @llvm.hexagon.V6.vmux(<512 x i1> %v6, <16 x i32> %v9, <16 x i32> %v11) |
| store <16 x i32> %v12, <16 x i32>* %a1, align 64 |
| ret <16 x i32> %v7 |
| } |
| |
| ; Function Attrs: nounwind readnone |
| declare <512 x i1> @llvm.hexagon.V6.pred.not(<512 x i1>) #1 |
| |
| ; Function Attrs: nounwind readnone |
| declare <512 x i1> @llvm.hexagon.V6.pred.scalar2(i32) #1 |
| |
| ; Function Attrs: nounwind readnone |
| declare <512 x i1> @llvm.hexagon.V6.pred.and(<512 x i1>, <512 x i1>) #1 |
| |
| ; Function Attrs: nounwind readnone |
| declare <16 x i32> @llvm.hexagon.V6.vmux(<512 x i1>, <16 x i32>, <16 x i32>) #1 |
| |
| attributes #0 = { nounwind "target-cpu"="hexagonv60" "target-features"="+hvxv60,+hvx-length64b" } |
| attributes #1 = { nounwind readnone } |