| ; REQUIRES: asserts |
| ; RUN: opt -inline -mtriple=aarch64--linux-gnu -S -debug-only=inline-cost < %s 2>&1 | FileCheck %s |
| |
| target datalayout = "e-m:e-i8:8:32-i16:16:32-i64:64-i128:128-n32:64-S128" |
| target triple = "aarch64--linux-gnu" |
| |
| define i32 @outer1(i32* %ptr, i32 %i) { |
| %C = call i32 @inner1(i32* %ptr, i32 %i) |
| ret i32 %C |
| } |
| |
| ; sext can be folded into gep. |
| ; CHECK: Analyzing call of inner1 |
| ; CHECK: NumInstructionsSimplified: 3 |
| ; CHECK: NumInstructions: 4 |
| define i32 @inner1(i32* %ptr, i32 %i) { |
| %E = sext i32 %i to i64 |
| %G = getelementptr inbounds i32, i32* %ptr, i64 %E |
| %L = load i32, i32* %G |
| ret i32 %L |
| } |
| |
| define i32 @outer2(i32* %ptr, i32 %i) { |
| %C = call i32 @inner2(i32* %ptr, i32 %i) |
| ret i32 %C |
| } |
| |
| ; zext from i32 to i64 is free. |
| ; CHECK: Analyzing call of inner2 |
| ; CHECK: NumInstructionsSimplified: 3 |
| ; CHECK: NumInstructions: 4 |
| define i32 @inner2(i32* %ptr, i32 %i) { |
| %E = zext i32 %i to i64 |
| %G = getelementptr inbounds i32, i32* %ptr, i64 %E |
| %L = load i32, i32* %G |
| ret i32 %L |
| } |
| |
| define i32 @outer3(i32* %ptr, i16 %i) { |
| %C = call i32 @inner3(i32* %ptr, i16 %i) |
| ret i32 %C |
| } |
| |
| ; zext can be folded into gep. |
| ; CHECK: Analyzing call of inner3 |
| ; CHECK: NumInstructionsSimplified: 3 |
| ; CHECK: NumInstructions: 4 |
| define i32 @inner3(i32* %ptr, i16 %i) { |
| %E = zext i16 %i to i64 |
| %G = getelementptr inbounds i32, i32* %ptr, i64 %E |
| %L = load i32, i32* %G |
| ret i32 %L |
| } |
| |
| define i16 @outer4(i8* %ptr) { |
| %C = call i16 @inner4(i8* %ptr) |
| ret i16 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner4 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i16 @inner4(i8* %ptr) { |
| %L = load i8, i8* %ptr |
| %E = zext i8 %L to i16 |
| ret i16 %E |
| } |
| |
| define i16 @outer5(i8* %ptr) { |
| %C = call i16 @inner5(i8* %ptr) |
| ret i16 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner5 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i16 @inner5(i8* %ptr) { |
| %L = load i8, i8* %ptr |
| %E = sext i8 %L to i16 |
| ret i16 %E |
| } |
| |
| define i32 @outer6(i8* %ptr) { |
| %C = call i32 @inner6(i8* %ptr) |
| ret i32 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner6 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i32 @inner6(i8* %ptr) { |
| %L = load i8, i8* %ptr |
| %E = zext i8 %L to i32 |
| ret i32 %E |
| } |
| |
| define i32 @outer7(i8* %ptr) { |
| %C = call i32 @inner7(i8* %ptr) |
| ret i32 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner7 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i32 @inner7(i8* %ptr) { |
| %L = load i8, i8* %ptr |
| %E = sext i8 %L to i32 |
| ret i32 %E |
| } |
| |
| define i32 @outer8(i16* %ptr) { |
| %C = call i32 @inner8(i16* %ptr) |
| ret i32 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner8 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i32 @inner8(i16* %ptr) { |
| %L = load i16, i16* %ptr |
| %E = zext i16 %L to i32 |
| ret i32 %E |
| } |
| |
| define i32 @outer9(i16* %ptr) { |
| %C = call i32 @inner9(i16* %ptr) |
| ret i32 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner9 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i32 @inner9(i16* %ptr) { |
| %L = load i16, i16* %ptr |
| %E = sext i16 %L to i32 |
| ret i32 %E |
| } |
| |
| define i64 @outer10(i8* %ptr) { |
| %C = call i64 @inner10(i8* %ptr) |
| ret i64 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner10 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i64 @inner10(i8* %ptr) { |
| %L = load i8, i8* %ptr |
| %E = zext i8 %L to i64 |
| ret i64 %E |
| } |
| |
| define i64 @outer11(i8* %ptr) { |
| %C = call i64 @inner11(i8* %ptr) |
| ret i64 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner11 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i64 @inner11(i8* %ptr) { |
| %L = load i8, i8* %ptr |
| %E = sext i8 %L to i64 |
| ret i64 %E |
| } |
| |
| define i64 @outer12(i16* %ptr) { |
| %C = call i64 @inner12(i16* %ptr) |
| ret i64 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner12 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i64 @inner12(i16* %ptr) { |
| %L = load i16, i16* %ptr |
| %E = zext i16 %L to i64 |
| ret i64 %E |
| } |
| |
| define i64 @outer13(i16* %ptr) { |
| %C = call i64 @inner13(i16* %ptr) |
| ret i64 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner13 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i64 @inner13(i16* %ptr) { |
| %L = load i16, i16* %ptr |
| %E = sext i16 %L to i64 |
| ret i64 %E |
| } |
| |
| define i64 @outer14(i32* %ptr) { |
| %C = call i64 @inner14(i32* %ptr) |
| ret i64 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner14 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i64 @inner14(i32* %ptr) { |
| %L = load i32, i32* %ptr |
| %E = zext i32 %L to i64 |
| ret i64 %E |
| } |
| |
| define i64 @outer15(i32* %ptr) { |
| %C = call i64 @inner15(i32* %ptr) |
| ret i64 %C |
| } |
| |
| ; It is an ExtLoad. |
| ; CHECK: Analyzing call of inner15 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 3 |
| define i64 @inner15(i32* %ptr) { |
| %L = load i32, i32* %ptr |
| %E = sext i32 %L to i64 |
| ret i64 %E |
| } |
| |
| define i64 @outer16(i32 %V1, i64 %V2) { |
| %C = call i64 @inner16(i32 %V1, i64 %V2) |
| ret i64 %C |
| } |
| |
| ; sext can be folded into shl. |
| ; CHECK: Analyzing call of inner16 |
| ; CHECK: NumInstructionsSimplified: 2 |
| ; CHECK: NumInstructions: 4 |
| define i64 @inner16(i32 %V1, i64 %V2) { |
| %E = sext i32 %V1 to i64 |
| %S = shl i64 %E, 3 |
| %A = add i64 %V2, %S |
| ret i64 %A |
| } |