mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2025-02-07 14:33:15 +00:00
88990248d3
and stores capture) to permit the caller to see each capture point and decide whether to continue looking. Use this inside memdep to do an analysis that basicaa won't do. This lets us solve another devirtualization case, fixing PR8908! git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@144580 91177308-0d34-0410-b5e6-96231b3b80d8
670 lines
16 KiB
LLVM
670 lines
16 KiB
LLVM
; RUN: opt < %s -basicaa -gvn -S -die | FileCheck %s
|
|
|
|
; 32-bit little endian target.
|
|
target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:64:64-v128:128:128-a0:0:64-f80:128:128-n8:16:32"
|
|
|
|
;; Trivial RLE test.
|
|
define i32 @test0(i32 %V, i32* %P) {
|
|
store i32 %V, i32* %P
|
|
|
|
%A = load i32* %P
|
|
ret i32 %A
|
|
; CHECK: @test0
|
|
; CHECK: ret i32 %V
|
|
}
|
|
|
|
|
|
;;===----------------------------------------------------------------------===;;
|
|
;; Tests for crashers
|
|
;;===----------------------------------------------------------------------===;;
|
|
|
|
;; PR5016
|
|
define i8 @crash0({i32, i32} %A, {i32, i32}* %P) {
|
|
store {i32, i32} %A, {i32, i32}* %P
|
|
%X = bitcast {i32, i32}* %P to i8*
|
|
%Y = load i8* %X
|
|
ret i8 %Y
|
|
}
|
|
|
|
|
|
;;===----------------------------------------------------------------------===;;
|
|
;; Store -> Load and Load -> Load forwarding where src and dst are different
|
|
;; types, but where the base pointer is a must alias.
|
|
;;===----------------------------------------------------------------------===;;
|
|
|
|
;; i32 -> f32 forwarding.
|
|
define float @coerce_mustalias1(i32 %V, i32* %P) {
|
|
store i32 %V, i32* %P
|
|
|
|
%P2 = bitcast i32* %P to float*
|
|
|
|
%A = load float* %P2
|
|
ret float %A
|
|
; CHECK: @coerce_mustalias1
|
|
; CHECK-NOT: load
|
|
; CHECK: ret float
|
|
}
|
|
|
|
;; i32* -> float forwarding.
|
|
define float @coerce_mustalias2(i32* %V, i32** %P) {
|
|
store i32* %V, i32** %P
|
|
|
|
%P2 = bitcast i32** %P to float*
|
|
|
|
%A = load float* %P2
|
|
ret float %A
|
|
; CHECK: @coerce_mustalias2
|
|
; CHECK-NOT: load
|
|
; CHECK: ret float
|
|
}
|
|
|
|
;; float -> i32* forwarding.
|
|
define i32* @coerce_mustalias3(float %V, float* %P) {
|
|
store float %V, float* %P
|
|
|
|
%P2 = bitcast float* %P to i32**
|
|
|
|
%A = load i32** %P2
|
|
ret i32* %A
|
|
; CHECK: @coerce_mustalias3
|
|
; CHECK-NOT: load
|
|
; CHECK: ret i32*
|
|
}
|
|
|
|
;; i32 -> f32 load forwarding.
|
|
define float @coerce_mustalias4(i32* %P, i1 %cond) {
|
|
%A = load i32* %P
|
|
|
|
%P2 = bitcast i32* %P to float*
|
|
%B = load float* %P2
|
|
br i1 %cond, label %T, label %F
|
|
T:
|
|
ret float %B
|
|
|
|
F:
|
|
%X = bitcast i32 %A to float
|
|
ret float %X
|
|
|
|
; CHECK: @coerce_mustalias4
|
|
; CHECK: %A = load i32* %P
|
|
; CHECK-NOT: load
|
|
; CHECK: ret float
|
|
; CHECK: F:
|
|
}
|
|
|
|
;; i32 -> i8 forwarding
|
|
define i8 @coerce_mustalias5(i32 %V, i32* %P) {
|
|
store i32 %V, i32* %P
|
|
|
|
%P2 = bitcast i32* %P to i8*
|
|
|
|
%A = load i8* %P2
|
|
ret i8 %A
|
|
; CHECK: @coerce_mustalias5
|
|
; CHECK-NOT: load
|
|
; CHECK: ret i8
|
|
}
|
|
|
|
;; i64 -> float forwarding
|
|
define float @coerce_mustalias6(i64 %V, i64* %P) {
|
|
store i64 %V, i64* %P
|
|
|
|
%P2 = bitcast i64* %P to float*
|
|
|
|
%A = load float* %P2
|
|
ret float %A
|
|
; CHECK: @coerce_mustalias6
|
|
; CHECK-NOT: load
|
|
; CHECK: ret float
|
|
}
|
|
|
|
;; i64 -> i8* (32-bit) forwarding
|
|
define i8* @coerce_mustalias7(i64 %V, i64* %P) {
|
|
store i64 %V, i64* %P
|
|
|
|
%P2 = bitcast i64* %P to i8**
|
|
|
|
%A = load i8** %P2
|
|
ret i8* %A
|
|
; CHECK: @coerce_mustalias7
|
|
; CHECK-NOT: load
|
|
; CHECK: ret i8*
|
|
}
|
|
|
|
; memset -> i16 forwarding.
|
|
define signext i16 @memset_to_i16_local(i16* %A) nounwind ssp {
|
|
entry:
|
|
%conv = bitcast i16* %A to i8*
|
|
tail call void @llvm.memset.p0i8.i64(i8* %conv, i8 1, i64 200, i32 1, i1 false)
|
|
%arrayidx = getelementptr inbounds i16* %A, i64 42
|
|
%tmp2 = load i16* %arrayidx
|
|
ret i16 %tmp2
|
|
; CHECK: @memset_to_i16_local
|
|
; CHECK-NOT: load
|
|
; CHECK: ret i16 257
|
|
}
|
|
|
|
; memset -> float forwarding.
|
|
define float @memset_to_float_local(float* %A, i8 %Val) nounwind ssp {
|
|
entry:
|
|
%conv = bitcast float* %A to i8* ; <i8*> [#uses=1]
|
|
tail call void @llvm.memset.p0i8.i64(i8* %conv, i8 %Val, i64 400, i32 1, i1 false)
|
|
%arrayidx = getelementptr inbounds float* %A, i64 42 ; <float*> [#uses=1]
|
|
%tmp2 = load float* %arrayidx ; <float> [#uses=1]
|
|
ret float %tmp2
|
|
; CHECK: @memset_to_float_local
|
|
; CHECK-NOT: load
|
|
; CHECK: zext
|
|
; CHECK-NEXT: shl
|
|
; CHECK-NEXT: or
|
|
; CHECK-NEXT: shl
|
|
; CHECK-NEXT: or
|
|
; CHECK-NEXT: bitcast
|
|
; CHECK-NEXT: ret float
|
|
}
|
|
|
|
;; non-local memset -> i16 load forwarding.
|
|
define i16 @memset_to_i16_nonlocal0(i16* %P, i1 %cond) {
|
|
%P3 = bitcast i16* %P to i8*
|
|
br i1 %cond, label %T, label %F
|
|
T:
|
|
tail call void @llvm.memset.p0i8.i64(i8* %P3, i8 1, i64 400, i32 1, i1 false)
|
|
br label %Cont
|
|
|
|
F:
|
|
tail call void @llvm.memset.p0i8.i64(i8* %P3, i8 2, i64 400, i32 1, i1 false)
|
|
br label %Cont
|
|
|
|
Cont:
|
|
%P2 = getelementptr i16* %P, i32 4
|
|
%A = load i16* %P2
|
|
ret i16 %A
|
|
|
|
; CHECK: @memset_to_i16_nonlocal0
|
|
; CHECK: Cont:
|
|
; CHECK-NEXT: %A = phi i16 [ 514, %F ], [ 257, %T ]
|
|
; CHECK-NOT: load
|
|
; CHECK: ret i16 %A
|
|
}
|
|
|
|
@GCst = constant {i32, float, i32 } { i32 42, float 14., i32 97 }
|
|
|
|
; memset -> float forwarding.
|
|
define float @memcpy_to_float_local(float* %A) nounwind ssp {
|
|
entry:
|
|
%conv = bitcast float* %A to i8* ; <i8*> [#uses=1]
|
|
tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %conv, i8* bitcast ({i32, float, i32 }* @GCst to i8*), i64 12, i32 1, i1 false)
|
|
%arrayidx = getelementptr inbounds float* %A, i64 1 ; <float*> [#uses=1]
|
|
%tmp2 = load float* %arrayidx ; <float> [#uses=1]
|
|
ret float %tmp2
|
|
; CHECK: @memcpy_to_float_local
|
|
; CHECK-NOT: load
|
|
; CHECK: ret float 1.400000e+01
|
|
}
|
|
|
|
|
|
|
|
;; non-local i32/float -> i8 load forwarding.
|
|
define i8 @coerce_mustalias_nonlocal0(i32* %P, i1 %cond) {
|
|
%P2 = bitcast i32* %P to float*
|
|
%P3 = bitcast i32* %P to i8*
|
|
br i1 %cond, label %T, label %F
|
|
T:
|
|
store i32 42, i32* %P
|
|
br label %Cont
|
|
|
|
F:
|
|
store float 1.0, float* %P2
|
|
br label %Cont
|
|
|
|
Cont:
|
|
%A = load i8* %P3
|
|
ret i8 %A
|
|
|
|
; CHECK: @coerce_mustalias_nonlocal0
|
|
; CHECK: Cont:
|
|
; CHECK: %A = phi i8 [
|
|
; CHECK-NOT: load
|
|
; CHECK: ret i8 %A
|
|
}
|
|
|
|
|
|
;; non-local i32/float -> i8 load forwarding. This also tests that the "P3"
|
|
;; bitcast equivalence can be properly phi translated.
|
|
define i8 @coerce_mustalias_nonlocal1(i32* %P, i1 %cond) {
|
|
%P2 = bitcast i32* %P to float*
|
|
br i1 %cond, label %T, label %F
|
|
T:
|
|
store i32 42, i32* %P
|
|
br label %Cont
|
|
|
|
F:
|
|
store float 1.0, float* %P2
|
|
br label %Cont
|
|
|
|
Cont:
|
|
%P3 = bitcast i32* %P to i8*
|
|
%A = load i8* %P3
|
|
ret i8 %A
|
|
|
|
;; FIXME: This is disabled because this caused a miscompile in the llvm-gcc
|
|
;; bootstrap, see r82411
|
|
;
|
|
; HECK: @coerce_mustalias_nonlocal1
|
|
; HECK: Cont:
|
|
; HECK: %A = phi i8 [
|
|
; HECK-NOT: load
|
|
; HECK: ret i8 %A
|
|
}
|
|
|
|
|
|
;; non-local i32 -> i8 partial redundancy load forwarding.
|
|
define i8 @coerce_mustalias_pre0(i32* %P, i1 %cond) {
|
|
%P3 = bitcast i32* %P to i8*
|
|
br i1 %cond, label %T, label %F
|
|
T:
|
|
store i32 42, i32* %P
|
|
br label %Cont
|
|
|
|
F:
|
|
br label %Cont
|
|
|
|
Cont:
|
|
%A = load i8* %P3
|
|
ret i8 %A
|
|
|
|
; CHECK: @coerce_mustalias_pre0
|
|
; CHECK: F:
|
|
; CHECK: load i8* %P3
|
|
; CHECK: Cont:
|
|
; CHECK: %A = phi i8 [
|
|
; CHECK-NOT: load
|
|
; CHECK: ret i8 %A
|
|
}
|
|
|
|
;;===----------------------------------------------------------------------===;;
|
|
;; Store -> Load and Load -> Load forwarding where src and dst are different
|
|
;; types, and the reload is an offset from the store pointer.
|
|
;;===----------------------------------------------------------------------===;;
|
|
|
|
;; i32 -> i8 forwarding.
|
|
;; PR4216
|
|
define i8 @coerce_offset0(i32 %V, i32* %P) {
|
|
store i32 %V, i32* %P
|
|
|
|
%P2 = bitcast i32* %P to i8*
|
|
%P3 = getelementptr i8* %P2, i32 2
|
|
|
|
%A = load i8* %P3
|
|
ret i8 %A
|
|
; CHECK: @coerce_offset0
|
|
; CHECK-NOT: load
|
|
; CHECK: ret i8
|
|
}
|
|
|
|
;; non-local i32/float -> i8 load forwarding.
|
|
define i8 @coerce_offset_nonlocal0(i32* %P, i1 %cond) {
|
|
%P2 = bitcast i32* %P to float*
|
|
%P3 = bitcast i32* %P to i8*
|
|
%P4 = getelementptr i8* %P3, i32 2
|
|
br i1 %cond, label %T, label %F
|
|
T:
|
|
store i32 42, i32* %P
|
|
br label %Cont
|
|
|
|
F:
|
|
store float 1.0, float* %P2
|
|
br label %Cont
|
|
|
|
Cont:
|
|
%A = load i8* %P4
|
|
ret i8 %A
|
|
|
|
; CHECK: @coerce_offset_nonlocal0
|
|
; CHECK: Cont:
|
|
; CHECK: %A = phi i8 [
|
|
; CHECK-NOT: load
|
|
; CHECK: ret i8 %A
|
|
}
|
|
|
|
|
|
;; non-local i32 -> i8 partial redundancy load forwarding.
|
|
define i8 @coerce_offset_pre0(i32* %P, i1 %cond) {
|
|
%P3 = bitcast i32* %P to i8*
|
|
%P4 = getelementptr i8* %P3, i32 2
|
|
br i1 %cond, label %T, label %F
|
|
T:
|
|
store i32 42, i32* %P
|
|
br label %Cont
|
|
|
|
F:
|
|
br label %Cont
|
|
|
|
Cont:
|
|
%A = load i8* %P4
|
|
ret i8 %A
|
|
|
|
; CHECK: @coerce_offset_pre0
|
|
; CHECK: F:
|
|
; CHECK: load i8* %P4
|
|
; CHECK: Cont:
|
|
; CHECK: %A = phi i8 [
|
|
; CHECK-NOT: load
|
|
; CHECK: ret i8 %A
|
|
}
|
|
|
|
define i32 @chained_load(i32** %p) {
|
|
block1:
|
|
%A = alloca i32*
|
|
|
|
%z = load i32** %p
|
|
store i32* %z, i32** %A
|
|
br i1 true, label %block2, label %block3
|
|
|
|
block2:
|
|
%a = load i32** %p
|
|
br label %block4
|
|
|
|
block3:
|
|
%b = load i32** %p
|
|
br label %block4
|
|
|
|
block4:
|
|
%c = load i32** %p
|
|
%d = load i32* %c
|
|
ret i32 %d
|
|
|
|
; CHECK: @chained_load
|
|
; CHECK: %z = load i32** %p
|
|
; CHECK-NOT: load
|
|
; CHECK: %d = load i32* %z
|
|
; CHECK-NEXT: ret i32 %d
|
|
}
|
|
|
|
|
|
declare i1 @cond() readonly
|
|
declare i1 @cond2() readonly
|
|
|
|
define i32 @phi_trans2() {
|
|
; CHECK: @phi_trans2
|
|
entry:
|
|
%P = alloca i32, i32 400
|
|
br label %F1
|
|
|
|
F1:
|
|
%A = phi i32 [1, %entry], [2, %F]
|
|
%cond2 = call i1 @cond()
|
|
br i1 %cond2, label %T1, label %TY
|
|
|
|
T1:
|
|
%P2 = getelementptr i32* %P, i32 %A
|
|
%x = load i32* %P2
|
|
%cond = call i1 @cond2()
|
|
br i1 %cond, label %TX, label %F
|
|
|
|
F:
|
|
%P3 = getelementptr i32* %P, i32 2
|
|
store i32 17, i32* %P3
|
|
|
|
store i32 42, i32* %P2 ; Provides "P[A]".
|
|
br label %F1
|
|
|
|
TX:
|
|
; This load should not be compiled to 'ret i32 42'. An overly clever
|
|
; implementation of GVN would see that we're returning 17 if the loop
|
|
; executes once or 42 if it executes more than that, but we'd have to do
|
|
; loop restructuring to expose this, and GVN shouldn't do this sort of CFG
|
|
; transformation.
|
|
|
|
; CHECK: TX:
|
|
; CHECK: ret i32 %x
|
|
ret i32 %x
|
|
TY:
|
|
ret i32 0
|
|
}
|
|
|
|
define i32 @phi_trans3(i32* %p) {
|
|
; CHECK: @phi_trans3
|
|
block1:
|
|
br i1 true, label %block2, label %block3
|
|
|
|
block2:
|
|
store i32 87, i32* %p
|
|
br label %block4
|
|
|
|
block3:
|
|
%p2 = getelementptr i32* %p, i32 43
|
|
store i32 97, i32* %p2
|
|
br label %block4
|
|
|
|
block4:
|
|
%A = phi i32 [-1, %block2], [42, %block3]
|
|
br i1 true, label %block5, label %exit
|
|
|
|
; CHECK: block4:
|
|
; CHECK-NEXT: %D = phi i32 [ 87, %block2 ], [ 97, %block3 ]
|
|
; CHECK-NOT: load
|
|
|
|
block5:
|
|
%B = add i32 %A, 1
|
|
br i1 true, label %block6, label %exit
|
|
|
|
block6:
|
|
%C = getelementptr i32* %p, i32 %B
|
|
br i1 true, label %block7, label %exit
|
|
|
|
block7:
|
|
%D = load i32* %C
|
|
ret i32 %D
|
|
|
|
; CHECK: block7:
|
|
; CHECK-NEXT: ret i32 %D
|
|
|
|
exit:
|
|
ret i32 -1
|
|
}
|
|
|
|
define i8 @phi_trans4(i8* %p) {
|
|
; CHECK: @phi_trans4
|
|
entry:
|
|
%X3 = getelementptr i8* %p, i32 192
|
|
store i8 192, i8* %X3
|
|
|
|
%X = getelementptr i8* %p, i32 4
|
|
%Y = load i8* %X
|
|
br label %loop
|
|
|
|
loop:
|
|
%i = phi i32 [4, %entry], [192, %loop]
|
|
%X2 = getelementptr i8* %p, i32 %i
|
|
%Y2 = load i8* %X2
|
|
|
|
; CHECK: loop:
|
|
; CHECK-NEXT: %Y2 = phi i8 [ %Y, %entry ], [ 0, %loop ]
|
|
; CHECK-NOT: load i8
|
|
|
|
%cond = call i1 @cond2()
|
|
|
|
%Z = bitcast i8 *%X3 to i32*
|
|
store i32 0, i32* %Z
|
|
br i1 %cond, label %loop, label %out
|
|
|
|
out:
|
|
%R = add i8 %Y, %Y2
|
|
ret i8 %R
|
|
}
|
|
|
|
define i8 @phi_trans5(i8* %p) {
|
|
; CHECK: @phi_trans5
|
|
entry:
|
|
|
|
%X4 = getelementptr i8* %p, i32 2
|
|
store i8 19, i8* %X4
|
|
|
|
%X = getelementptr i8* %p, i32 4
|
|
%Y = load i8* %X
|
|
br label %loop
|
|
|
|
loop:
|
|
%i = phi i32 [4, %entry], [3, %cont]
|
|
%X2 = getelementptr i8* %p, i32 %i
|
|
%Y2 = load i8* %X2 ; Ensure this load is not being incorrectly replaced.
|
|
%cond = call i1 @cond2()
|
|
br i1 %cond, label %cont, label %out
|
|
|
|
cont:
|
|
%Z = getelementptr i8* %X2, i32 -1
|
|
%Z2 = bitcast i8 *%Z to i32*
|
|
store i32 50462976, i32* %Z2 ;; (1 << 8) | (2 << 16) | (3 << 24)
|
|
|
|
|
|
; CHECK: store i32
|
|
; CHECK-NEXT: getelementptr i8* %p, i32 3
|
|
; CHECK-NEXT: load i8*
|
|
br label %loop
|
|
|
|
out:
|
|
%R = add i8 %Y, %Y2
|
|
ret i8 %R
|
|
}
|
|
|
|
|
|
; PR6642
|
|
define i32 @memset_to_load() nounwind readnone {
|
|
entry:
|
|
%x = alloca [256 x i32], align 4 ; <[256 x i32]*> [#uses=2]
|
|
%tmp = bitcast [256 x i32]* %x to i8* ; <i8*> [#uses=1]
|
|
call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 1024, i32 4, i1 false)
|
|
%arraydecay = getelementptr inbounds [256 x i32]* %x, i32 0, i32 0 ; <i32*>
|
|
%tmp1 = load i32* %arraydecay ; <i32> [#uses=1]
|
|
ret i32 %tmp1
|
|
; CHECK: @memset_to_load
|
|
; CHECK: ret i32 0
|
|
}
|
|
|
|
|
|
;;===----------------------------------------------------------------------===;;
|
|
;; Load -> Load forwarding in partial alias case.
|
|
;;===----------------------------------------------------------------------===;;
|
|
|
|
define i32 @load_load_partial_alias(i8* %P) nounwind ssp {
|
|
entry:
|
|
%0 = bitcast i8* %P to i32*
|
|
%tmp2 = load i32* %0
|
|
%add.ptr = getelementptr inbounds i8* %P, i64 1
|
|
%tmp5 = load i8* %add.ptr
|
|
%conv = zext i8 %tmp5 to i32
|
|
%add = add nsw i32 %tmp2, %conv
|
|
ret i32 %add
|
|
|
|
; TEMPORARILYDISABLED: @load_load_partial_alias
|
|
; TEMPORARILYDISABLED: load i32*
|
|
; TEMPORARILYDISABLED-NOT: load
|
|
; TEMPORARILYDISABLED: lshr i32 {{.*}}, 8
|
|
; TEMPORARILYDISABLED-NOT: load
|
|
; TEMPORARILYDISABLED: trunc i32 {{.*}} to i8
|
|
; TEMPORARILYDISABLED-NOT: load
|
|
; TEMPORARILYDISABLED: ret i32
|
|
}
|
|
|
|
|
|
; Cross block partial alias case.
|
|
define i32 @load_load_partial_alias_cross_block(i8* %P) nounwind ssp {
|
|
entry:
|
|
%xx = bitcast i8* %P to i32*
|
|
%x1 = load i32* %xx, align 4
|
|
%cmp = icmp eq i32 %x1, 127
|
|
br i1 %cmp, label %land.lhs.true, label %if.end
|
|
|
|
land.lhs.true: ; preds = %entry
|
|
%arrayidx4 = getelementptr inbounds i8* %P, i64 1
|
|
%tmp5 = load i8* %arrayidx4, align 1
|
|
%conv6 = zext i8 %tmp5 to i32
|
|
ret i32 %conv6
|
|
|
|
if.end:
|
|
ret i32 52
|
|
; TEMPORARILY_DISABLED: @load_load_partial_alias_cross_block
|
|
; TEMPORARILY_DISABLED: land.lhs.true:
|
|
; TEMPORARILY_DISABLED-NOT: load i8
|
|
; TEMPORARILY_DISABLED: ret i32 %conv6
|
|
}
|
|
|
|
|
|
;;===----------------------------------------------------------------------===;;
|
|
;; Load Widening
|
|
;;===----------------------------------------------------------------------===;;
|
|
|
|
%widening1 = type { i32, i8, i8, i8, i8 }
|
|
|
|
@f = global %widening1 zeroinitializer, align 4
|
|
|
|
define i32 @test_widening1(i8* %P) nounwind ssp noredzone {
|
|
entry:
|
|
%tmp = load i8* getelementptr inbounds (%widening1* @f, i64 0, i32 1), align 4
|
|
%conv = zext i8 %tmp to i32
|
|
%tmp1 = load i8* getelementptr inbounds (%widening1* @f, i64 0, i32 2), align 1
|
|
%conv2 = zext i8 %tmp1 to i32
|
|
%add = add nsw i32 %conv, %conv2
|
|
ret i32 %add
|
|
; CHECK: @test_widening1
|
|
; CHECK-NOT: load
|
|
; CHECK: load i16*
|
|
; CHECK-NOT: load
|
|
; CHECK-ret i32
|
|
}
|
|
|
|
define i32 @test_widening2() nounwind ssp noredzone {
|
|
entry:
|
|
%tmp = load i8* getelementptr inbounds (%widening1* @f, i64 0, i32 1), align 4
|
|
%conv = zext i8 %tmp to i32
|
|
%tmp1 = load i8* getelementptr inbounds (%widening1* @f, i64 0, i32 2), align 1
|
|
%conv2 = zext i8 %tmp1 to i32
|
|
%add = add nsw i32 %conv, %conv2
|
|
|
|
%tmp2 = load i8* getelementptr inbounds (%widening1* @f, i64 0, i32 3), align 2
|
|
%conv3 = zext i8 %tmp2 to i32
|
|
%add2 = add nsw i32 %add, %conv3
|
|
|
|
%tmp3 = load i8* getelementptr inbounds (%widening1* @f, i64 0, i32 4), align 1
|
|
%conv4 = zext i8 %tmp3 to i32
|
|
%add3 = add nsw i32 %add2, %conv3
|
|
|
|
ret i32 %add3
|
|
; CHECK: @test_widening2
|
|
; CHECK-NOT: load
|
|
; CHECK: load i32*
|
|
; CHECK-NOT: load
|
|
; CHECK-ret i32
|
|
}
|
|
|
|
declare void @llvm.memset.p0i8.i64(i8* nocapture, i8, i64, i32, i1) nounwind
|
|
|
|
declare void @llvm.memcpy.p0i8.p0i8.i64(i8* nocapture, i8* nocapture, i64, i32, i1) nounwind
|
|
|
|
;;===----------------------------------------------------------------------===;;
|
|
;; Load -> Store dependency which isn't interfered with by a call that happens
|
|
;; before the pointer was captured.
|
|
;;===----------------------------------------------------------------------===;;
|
|
|
|
%class.X = type { [8 x i8] }
|
|
|
|
@_ZTV1X = weak_odr constant [5 x i8*] zeroinitializer
|
|
@_ZTV1Y = weak_odr constant [5 x i8*] zeroinitializer
|
|
|
|
declare void @use()
|
|
declare void @use3(i8***, i8**)
|
|
|
|
; PR8908
|
|
define void @test_escape1() nounwind {
|
|
%x = alloca i8**, align 8
|
|
store i8** getelementptr inbounds ([5 x i8*]* @_ZTV1X, i64 0, i64 2), i8*** %x, align 8
|
|
call void @use() nounwind
|
|
%DEAD = load i8*** %x, align 8
|
|
call void @use3(i8*** %x, i8** %DEAD) nounwind
|
|
ret void
|
|
; CHECK: test_escape1
|
|
; CHECK-NOT: DEAD
|
|
; CHECK: ret
|
|
}
|