| ; RUN: opt -basicaa -scoped-noalias -loop-vectorize -licm -force-vector-width=2 \ |
| ; RUN: -force-vector-interleave=1 -S < %s | FileCheck %s |
| |
| target datalayout = "e-m:o-i64:64-f80:128-n8:16:32:64-S128" |
| |
| ; In order to vectorize the inner loop, it needs to be versioned with |
| ; memchecks between {A} x {B, C} first: |
| ; |
| ; for (i = 0; i < n; i++) |
| ; for (j = 0; j < m; j++) |
| ; A[j] += B[i] + C[j]; |
| ; |
| ; Since in the versioned vector loop A and B can no longer alias, B[i] can be |
| ; LICM'ed from the inner loop. |
| |
| |
| define void @f(i32* %a, i32* %b, i32* %c) { |
| entry: |
| br label %outer |
| |
| outer: |
| %i.2 = phi i64 [ 0, %entry ], [ %i, %inner.end ] |
| %arrayidxB = getelementptr inbounds i32, i32* %b, i64 %i.2 |
| br label %inner.ph |
| |
| inner.ph: |
| ; CHECK: vector.ph: |
| ; CHECK: load i32, i32* %arrayidxB, |
| ; CHECK: br label %vector.body |
| br label %inner |
| |
| inner: |
| %j.2 = phi i64 [ 0, %inner.ph ], [ %j, %inner ] |
| |
| %arrayidxA = getelementptr inbounds i32, i32* %a, i64 %j.2 |
| %loadA = load i32, i32* %arrayidxA, align 4 |
| |
| %loadB = load i32, i32* %arrayidxB, align 4 |
| |
| %arrayidxC = getelementptr inbounds i32, i32* %c, i64 %j.2 |
| %loadC = load i32, i32* %arrayidxC, align 4 |
| |
| %add = add nuw i32 %loadA, %loadB |
| %add2 = add nuw i32 %add, %loadC |
| |
| store i32 %add2, i32* %arrayidxA, align 4 |
| |
| %j = add nuw nsw i64 %j.2, 1 |
| %cond1 = icmp eq i64 %j, 20 |
| br i1 %cond1, label %inner.end, label %inner |
| |
| inner.end: |
| %i = add nuw nsw i64 %i.2, 1 |
| %cond2 = icmp eq i64 %i, 30 |
| br i1 %cond2, label %outer.end, label %outer |
| |
| outer.end: |
| ret void |
| } |