| ; Test that direct loads and stores of local variables are not checked. | 
 | ; Also test that redundant checks of the same variable are elided. | 
 |  | 
 | ; REQUIRES: allow_dump | 
 |  | 
 | ; RUN: %p2i -i %s --args -verbose=inst -threads=0 -fsanitize-address \ | 
 | ; RUN:     | FileCheck --check-prefix=DUMP %s | 
 |  | 
 | define internal void @foo() { | 
 |   %ptr8 = alloca i8, i32 1, align 4 | 
 |   %ptr16 = alloca i8, i32 2, align 4 | 
 |   %ptr32 = alloca i8, i32 4, align 4 | 
 |   %ptr64 = alloca i8, i32 8, align 4 | 
 |   %ptr128 = alloca i8, i32 16, align 4 | 
 |  | 
 |   %target8 = bitcast i8* %ptr8 to i8* | 
 |   %target16 = bitcast i8* %ptr16 to i16* | 
 |   %target32 = bitcast i8* %ptr32 to i32* | 
 |   %target64 = bitcast i8* %ptr64 to i64* | 
 |   %target128 = bitcast i8* %ptr128 to <4 x i32>* | 
 |  | 
 |   ; unchecked loads | 
 |   %loaded8 = load i8, i8* %target8, align 1 | 
 |   %loaded16 = load i16, i16* %target16, align 1 | 
 |   %loaded32 = load i32, i32* %target32, align 1 | 
 |   %loaded64 = load i64, i64* %target64, align 1 | 
 |   %loaded128 = load <4 x i32>, <4 x i32>* %target128, align 4 | 
 |  | 
 |   ; unchecked stores | 
 |   store i8 %loaded8, i8* %target8, align 1 | 
 |   store i16 %loaded16, i16* %target16, align 1 | 
 |   store i32 %loaded32, i32* %target32, align 1 | 
 |   store i64 %loaded64, i64* %target64, align 1 | 
 |   store <4 x i32> %loaded128, <4 x i32>* %target128, align 4 | 
 |  | 
 |   %addr8 = ptrtoint i8* %ptr8 to i32 | 
 |   %addr16 = ptrtoint i8* %ptr16 to i32 | 
 |   %addr32 = ptrtoint i8* %ptr32 to i32 | 
 |   %addr64 = ptrtoint i8* %ptr64 to i32 | 
 |   %addr128 = ptrtoint i8* %ptr128 to i32 | 
 |  | 
 |   %off8 = add i32 %addr8, -1 | 
 |   %off16 = add i32 %addr16, -1 | 
 |   %off32 = add i32 %addr32, -1 | 
 |   %off64 = add i32 %addr64, -1 | 
 |   %off128 = add i32 %addr128, -1 | 
 |  | 
 |   %offtarget8 = inttoptr i32 %off8 to i8* | 
 |   %offtarget16 = inttoptr i32 %off16 to i16* | 
 |   %offtarget32 = inttoptr i32 %off32 to i32* | 
 |   %offtarget64 = inttoptr i32 %off64 to i64* | 
 |   %offtarget128 = inttoptr i32 %off128 to <4 x i32>* | 
 |  | 
 |   ; checked stores | 
 |   store i8 42, i8* %offtarget8, align 1 | 
 |   store i16 42, i16* %offtarget16, align 1 | 
 |   store i32 42, i32* %offtarget32, align 1 | 
 |  | 
 |   ; checked loads | 
 |   %offloaded64 = load i64, i64* %offtarget64, align 1 | 
 |   %offloaded128 = load <4 x i32>, <4 x i32>* %offtarget128, align 4 | 
 |  | 
 |   ; loads and stores with elided redundant checks | 
 |   %offloaded8 = load i8, i8* %offtarget8, align 1 | 
 |   %offloaded16 = load i16, i16* %offtarget16, align 1 | 
 |   %offloaded32 = load i32, i32* %offtarget32, align 1 | 
 |   store i64 %offloaded64, i64* %offtarget64, align 1 | 
 |   store <4 x i32> %offloaded128, <4 x i32>* %offtarget128, align 4 | 
 |  | 
 |   ret void | 
 | } | 
 |  | 
 | ; DUMP-LABEL: ================ Instrumented CFG ================ | 
 | ; DUMP-NEXT: define internal void @foo() { | 
 |  | 
 | ; Direct unchecked loads and stores | 
 | ; DUMP: %loaded8 = load i8, i8* %ptr8, align 1 | 
 | ; DUMP-NEXT: %loaded16 = load i16, i16* %ptr16, align 1 | 
 | ; DUMP-NEXT: %loaded32 = load i32, i32* %ptr32, align 1 | 
 | ; DUMP-NEXT: %loaded64 = load i64, i64* %ptr64, align 1 | 
 | ; DUMP-NEXT: %loaded128 = load <4 x i32>, <4 x i32>* %ptr128, align 4 | 
 | ; DUMP-NEXT: store i8 %loaded8, i8* %ptr8, align 1 | 
 | ; DUMP-NEXT: store i16 %loaded16, i16* %ptr16, align 1 | 
 | ; DUMP-NEXT: store i32 %loaded32, i32* %ptr32, align 1 | 
 | ; DUMP-NEXT: store i64 %loaded64, i64* %ptr64, align 1 | 
 | ; DUMP-NEXT: store <4 x i32> %loaded128, <4 x i32>* %ptr128, align 4 | 
 |  | 
 | ; Checked stores | 
 | ; DUMP: call void @__asan_check_store(i32 %off8, i32 1) | 
 | ; DUMP-NEXT: store i8 42, i8* %off8, align 1 | 
 | ; DUMP-NEXT: call void @__asan_check_store(i32 %off16, i32 2) | 
 | ; DUMP-NEXT: store i16 42, i16* %off16, align 1 | 
 | ; DUMP-NEXT: call void @__asan_check_store(i32 %off32, i32 4) | 
 | ; DUMP-NEXT: store i32 42, i32* %off32, align 1 | 
 |  | 
 | ; Checked loads | 
 | ; DUMP-NEXT: call void @__asan_check_load(i32 %off64, i32 8) | 
 | ; DUMP-NEXT: %offloaded64 = load i64, i64* %off64, align 1 | 
 | ; DUMP-NEXT: call void @__asan_check_load(i32 %off128, i32 16) | 
 | ; DUMP-NEXT: %offloaded128 = load <4 x i32>, <4 x i32>* %off128, align 4 | 
 |  | 
 | ; Loads and stores with elided redundant checks | 
 | ; DUMP-NEXT: %offloaded8 = load i8, i8* %off8, align 1 | 
 | ; DUMP-NEXT: %offloaded16 = load i16, i16* %off16, align 1 | 
 | ; DUMP-NEXT: %offloaded32 = load i32, i32* %off32, align 1 | 
 | ; DUMP-NEXT: store i64 %offloaded64, i64* %off64, align 1, beacon %offloaded64 | 
 | ; DUMP-NEXT: store <4 x i32> %offloaded128, <4 x i32>* %off128, align 4, beacon %offloaded128 |