llvm-6502/test/Transforms/ObjCARC/no-objc-arc-exceptions.ll
Michael Gottesman 0d92a3c600 [ObjCARC] Strength reduce objc_retainBlock -> objc_retain if the objc_retainBlock is optimizable.
If an objc_retainBlock has the copy_on_escape metadata attached to it
AND if the block pointer argument only escapes down the stack, we are
allowed to strength reduce the objc_retainBlock to to an objc_retain and
thus optimize it.

Current there is logic in the ARC data flow analysis to handle
this case which is complicated and involved making distinctions in
between objc_retainBlock and objc_retain in certain places and
considering them the same in others.

This patch simplifies said code by:

1. Performing the strength reduction in the initial ARC peephole
analysis (ObjCARCOpts::OptimizeIndividualCalls).

2. Changes the ARC dataflow analysis (which runs after the peephole
analysis) to consider all objc_retainBlock calls to not be optimizable
(since if the call was optimizable, we would have strength reduced it
already).

This patch leaves in the infrastructure in the ARC dataflow analysis to
handle this case, which due to 2 will just be dead code. I am doing this
on purpose to separate the removal of the old code from the testing of
the new code.

<rdar://problem/13249661>.

git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@178284 91177308-0d34-0410-b5e6-96231b3b80d8
2013-03-28 20:11:19 +00:00

124 lines
6.9 KiB
LLVM

; RUN: opt -S -objc-arc < %s | FileCheck %s
target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64-S128"
%struct.__block_byref_x = type { i8*, %struct.__block_byref_x*, i32, i32, i32 }
%struct.__block_descriptor = type { i64, i64 }
@_NSConcreteStackBlock = external global i8*
@__block_descriptor_tmp = external hidden constant { i64, i64, i8*, i8*, i8*, i8* }
; The optimizer should make use of the !clang.arc.no_objc_arc_exceptions
; metadata and eliminate the retainBlock+release pair here.
; rdar://10803830.
; CHECK: define void @test0(
; CHECK-NOT: @objc
; CHECK: }
define void @test0() {
entry:
%x = alloca %struct.__block_byref_x, align 8
%block = alloca <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>, align 8
%byref.isa = getelementptr inbounds %struct.__block_byref_x* %x, i64 0, i32 0
store i8* null, i8** %byref.isa, align 8
%byref.forwarding = getelementptr inbounds %struct.__block_byref_x* %x, i64 0, i32 1
store %struct.__block_byref_x* %x, %struct.__block_byref_x** %byref.forwarding, align 8
%byref.flags = getelementptr inbounds %struct.__block_byref_x* %x, i64 0, i32 2
store i32 0, i32* %byref.flags, align 8
%byref.size = getelementptr inbounds %struct.__block_byref_x* %x, i64 0, i32 3
store i32 32, i32* %byref.size, align 4
%block.isa = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 0
store i8* bitcast (i8** @_NSConcreteStackBlock to i8*), i8** %block.isa, align 8
%block.flags = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 1
store i32 1107296256, i32* %block.flags, align 8
%block.reserved = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 2
store i32 0, i32* %block.reserved, align 4
%block.invoke = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 3
store i8* bitcast (void (i8*)* @__foo_block_invoke_0 to i8*), i8** %block.invoke, align 8
%block.descriptor = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 4
store %struct.__block_descriptor* bitcast ({ i64, i64, i8*, i8*, i8*, i8* }* @__block_descriptor_tmp to %struct.__block_descriptor*), %struct.__block_descriptor** %block.descriptor, align 8
%block.captured = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 5
%t1 = bitcast %struct.__block_byref_x* %x to i8*
store i8* %t1, i8** %block.captured, align 8
%t2 = bitcast <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block to i8*
%t3 = call i8* @objc_retainBlock(i8* %t2) nounwind, !clang.arc.copy_on_escape !4
%t4 = getelementptr inbounds i8* %t3, i64 16
%t5 = bitcast i8* %t4 to i8**
%t6 = load i8** %t5, align 8
%t7 = bitcast i8* %t6 to void (i8*)*
invoke void %t7(i8* %t3)
to label %invoke.cont unwind label %lpad, !clang.arc.no_objc_arc_exceptions !4
invoke.cont: ; preds = %entry
call void @objc_release(i8* %t3) nounwind, !clang.imprecise_release !4
call void @_Block_object_dispose(i8* %t1, i32 8)
ret void
lpad: ; preds = %entry
%t8 = landingpad { i8*, i32 } personality i8* bitcast (i32 (...)* @__objc_personality_v0 to i8*)
cleanup
call void @_Block_object_dispose(i8* %t1, i32 8)
resume { i8*, i32 } %t8
}
; There is no !clang.arc.no_objc_arc_exceptions metadata here, so the optimizer
; shouldn't eliminate anything, but *CAN* strength reduce the objc_retainBlock
; to an objc_retain.
; CHECK: define void @test0_no_metadata(
; CHECK: call i8* @objc_retain(
; CHECK: invoke
; CHECK: call void @objc_release(
; CHECK: }
define void @test0_no_metadata() {
entry:
%x = alloca %struct.__block_byref_x, align 8
%block = alloca <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>, align 8
%byref.isa = getelementptr inbounds %struct.__block_byref_x* %x, i64 0, i32 0
store i8* null, i8** %byref.isa, align 8
%byref.forwarding = getelementptr inbounds %struct.__block_byref_x* %x, i64 0, i32 1
store %struct.__block_byref_x* %x, %struct.__block_byref_x** %byref.forwarding, align 8
%byref.flags = getelementptr inbounds %struct.__block_byref_x* %x, i64 0, i32 2
store i32 0, i32* %byref.flags, align 8
%byref.size = getelementptr inbounds %struct.__block_byref_x* %x, i64 0, i32 3
store i32 32, i32* %byref.size, align 4
%block.isa = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 0
store i8* bitcast (i8** @_NSConcreteStackBlock to i8*), i8** %block.isa, align 8
%block.flags = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 1
store i32 1107296256, i32* %block.flags, align 8
%block.reserved = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 2
store i32 0, i32* %block.reserved, align 4
%block.invoke = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 3
store i8* bitcast (void (i8*)* @__foo_block_invoke_0 to i8*), i8** %block.invoke, align 8
%block.descriptor = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 4
store %struct.__block_descriptor* bitcast ({ i64, i64, i8*, i8*, i8*, i8* }* @__block_descriptor_tmp to %struct.__block_descriptor*), %struct.__block_descriptor** %block.descriptor, align 8
%block.captured = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block, i64 0, i32 5
%t1 = bitcast %struct.__block_byref_x* %x to i8*
store i8* %t1, i8** %block.captured, align 8
%t2 = bitcast <{ i8*, i32, i32, i8*, %struct.__block_descriptor*, i8* }>* %block to i8*
%t3 = call i8* @objc_retainBlock(i8* %t2) nounwind, !clang.arc.copy_on_escape !4
%t4 = getelementptr inbounds i8* %t3, i64 16
%t5 = bitcast i8* %t4 to i8**
%t6 = load i8** %t5, align 8
%t7 = bitcast i8* %t6 to void (i8*)*
invoke void %t7(i8* %t3)
to label %invoke.cont unwind label %lpad
invoke.cont: ; preds = %entry
call void @objc_release(i8* %t3) nounwind, !clang.imprecise_release !4
call void @_Block_object_dispose(i8* %t1, i32 8)
ret void
lpad: ; preds = %entry
%t8 = landingpad { i8*, i32 } personality i8* bitcast (i32 (...)* @__objc_personality_v0 to i8*)
cleanup
call void @_Block_object_dispose(i8* %t1, i32 8)
resume { i8*, i32 } %t8
}
declare i8* @objc_retainBlock(i8*)
declare void @objc_release(i8*)
declare void @_Block_object_dispose(i8*, i32)
declare i32 @__objc_personality_v0(...)
declare void @__foo_block_invoke_0(i8* nocapture) uwtable ssp
!4 = metadata !{}