blob: 7d72e37f80341fa5a69d53062243078001c0c790 [file] [log] [blame]
Dan Gohman12130272011-08-12 00:26:31 +00001; RUN: opt -objc-arc -S < %s | FileCheck %s
2
3%struct.__objcFastEnumerationState = type { i64, i8**, i64*, [5 x i64] }
4
5@"\01L_OBJC_METH_VAR_NAME_" = internal global [43 x i8] c"countByEnumeratingWithState:objects:count:\00", section "__TEXT,__objc_methname,cstring_literals", align 1
6@"\01L_OBJC_SELECTOR_REFERENCES_" = internal global i8* getelementptr inbounds ([43 x i8]* @"\01L_OBJC_METH_VAR_NAME_", i64 0, i64 0), section "__DATA, __objc_selrefs, literal_pointers, no_dead_strip"
7@g = common global i8* null, align 8
8@"\01L_OBJC_IMAGE_INFO" = internal constant [2 x i32] [i32 0, i32 16], section "__DATA, __objc_imageinfo, regular, no_dead_strip"
9
10declare void @callee()
11declare i8* @returner()
12declare i8* @objc_retainAutoreleasedReturnValue(i8*)
13declare i8* @objc_retain(i8*)
14declare void @objc_enumerationMutation(i8*)
15declare void @llvm.memset.p0i8.i64(i8* nocapture, i8, i64, i32, i1) nounwind
16declare i8* @objc_msgSend(i8*, i8*, ...) nonlazybind
17declare void @use(i8*)
18declare void @objc_release(i8*)
Dan Gohmandf476e52012-09-04 23:16:20 +000019declare i8* @def()
20declare void @__crasher_block_invoke(i8* nocapture)
21declare i8* @objc_retainBlock(i8*)
22declare void @__crasher_block_invoke1(i8* nocapture)
Dan Gohman12130272011-08-12 00:26:31 +000023
Duncan P. N. Exon Smithbe7ea192014-12-15 19:07:53 +000024!0 = !{}
Dan Gohman12130272011-08-12 00:26:31 +000025
26; Delete a nested retain+release pair.
27
Stephen Lina76289a2013-07-14 01:50:49 +000028; CHECK-LABEL: define void @test0(
Dan Gohman12130272011-08-12 00:26:31 +000029; CHECK: call i8* @objc_retain
30; CHECK-NOT: @objc_retain
31; CHECK: }
32define void @test0(i8* %a) nounwind {
33entry:
34 %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
35 %items.ptr = alloca [16 x i8*], align 8
36 %0 = call i8* @objc_retain(i8* %a) nounwind
37 %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
38 call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 64, i32 8, i1 false)
39 %1 = call i8* @objc_retain(i8* %0) nounwind
40 %tmp2 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
41 %call = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
42 %iszero = icmp eq i64 %call, 0
43 br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
44
45forcoll.loopinit:
46 %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
47 %mutationsptr = load i64** %mutationsptr.ptr, align 8
48 %forcoll.initial-mutations = load i64* %mutationsptr, align 8
49 %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
50 br label %forcoll.loopbody.outer
51
52forcoll.loopbody.outer:
53 %forcoll.count.ph = phi i64 [ %call, %forcoll.loopinit ], [ %call6, %forcoll.refetch ]
54 %tmp7 = icmp ugt i64 %forcoll.count.ph, 1
55 %umax = select i1 %tmp7, i64 %forcoll.count.ph, i64 1
56 br label %forcoll.loopbody
57
58forcoll.loopbody:
59 %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
60 %mutationsptr3 = load i64** %mutationsptr.ptr, align 8
61 %statemutations = load i64* %mutationsptr3, align 8
62 %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
63 br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
64
65forcoll.mutated:
66 call void @objc_enumerationMutation(i8* %1)
67 br label %forcoll.notmutated
68
69forcoll.notmutated:
70 %stateitems = load i8*** %stateitems.ptr, align 8
71 %currentitem.ptr = getelementptr i8** %stateitems, i64 %forcoll.index
72 %3 = load i8** %currentitem.ptr, align 8
73 call void @use(i8* %3)
74 %4 = add i64 %forcoll.index, 1
75 %exitcond = icmp eq i64 %4, %umax
76 br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
77
78forcoll.refetch:
79 %tmp5 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
80 %call6 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp5, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
81 %5 = icmp eq i64 %call6, 0
82 br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
83
84forcoll.empty:
85 call void @objc_release(i8* %1) nounwind
86 call void @objc_release(i8* %0) nounwind, !clang.imprecise_release !0
87 ret void
88}
89
90; Delete a nested retain+release pair.
91
Stephen Lina76289a2013-07-14 01:50:49 +000092; CHECK-LABEL: define void @test2(
Dan Gohman12130272011-08-12 00:26:31 +000093; CHECK: call i8* @objc_retain
94; CHECK-NOT: @objc_retain
95; CHECK: }
96define void @test2() nounwind {
97entry:
98 %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
99 %items.ptr = alloca [16 x i8*], align 8
100 %call = call i8* @returner()
101 %0 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call) nounwind
102 %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
103 call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 64, i32 8, i1 false)
104 %1 = call i8* @objc_retain(i8* %0) nounwind
105 %tmp2 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
106 %call3 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
107 %iszero = icmp eq i64 %call3, 0
108 br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
109
110forcoll.loopinit:
111 %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
112 %mutationsptr = load i64** %mutationsptr.ptr, align 8
113 %forcoll.initial-mutations = load i64* %mutationsptr, align 8
114 %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
115 br label %forcoll.loopbody.outer
116
117forcoll.loopbody.outer:
118 %forcoll.count.ph = phi i64 [ %call3, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
119 %tmp8 = icmp ugt i64 %forcoll.count.ph, 1
120 %umax = select i1 %tmp8, i64 %forcoll.count.ph, i64 1
121 br label %forcoll.loopbody
122
123forcoll.loopbody:
124 %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
125 %mutationsptr4 = load i64** %mutationsptr.ptr, align 8
126 %statemutations = load i64* %mutationsptr4, align 8
127 %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
128 br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
129
130forcoll.mutated:
131 call void @objc_enumerationMutation(i8* %1)
132 br label %forcoll.notmutated
133
134forcoll.notmutated:
135 %stateitems = load i8*** %stateitems.ptr, align 8
136 %currentitem.ptr = getelementptr i8** %stateitems, i64 %forcoll.index
137 %3 = load i8** %currentitem.ptr, align 8
138 call void @use(i8* %3)
139 %4 = add i64 %forcoll.index, 1
140 %exitcond = icmp eq i64 %4, %umax
141 br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
142
143forcoll.refetch:
144 %tmp6 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
145 %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
146 %5 = icmp eq i64 %call7, 0
147 br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
148
149forcoll.empty:
150 call void @objc_release(i8* %1) nounwind
151 call void @objc_release(i8* %0) nounwind, !clang.imprecise_release !0
152 ret void
153}
154
155; Delete a nested retain+release pair.
156
Stephen Lina76289a2013-07-14 01:50:49 +0000157; CHECK-LABEL: define void @test4(
Dan Gohman12130272011-08-12 00:26:31 +0000158; CHECK: call i8* @objc_retain
159; CHECK-NOT: @objc_retain
160; CHECK: }
161define void @test4() nounwind {
162entry:
163 %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
164 %items.ptr = alloca [16 x i8*], align 8
165 %tmp = load i8** @g, align 8
166 %0 = call i8* @objc_retain(i8* %tmp) nounwind
167 %tmp2 = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
168 call void @llvm.memset.p0i8.i64(i8* %tmp2, i8 0, i64 64, i32 8, i1 false)
169 %1 = call i8* @objc_retain(i8* %0) nounwind
170 %tmp4 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
171 %call = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp4, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
172 %iszero = icmp eq i64 %call, 0
173 br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
174
175forcoll.loopinit:
176 %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
177 %mutationsptr = load i64** %mutationsptr.ptr, align 8
178 %forcoll.initial-mutations = load i64* %mutationsptr, align 8
179 %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
180 br label %forcoll.loopbody.outer
181
182forcoll.loopbody.outer:
183 %forcoll.count.ph = phi i64 [ %call, %forcoll.loopinit ], [ %call8, %forcoll.refetch ]
184 %tmp9 = icmp ugt i64 %forcoll.count.ph, 1
185 %umax = select i1 %tmp9, i64 %forcoll.count.ph, i64 1
186 br label %forcoll.loopbody
187
188forcoll.loopbody:
189 %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
190 %mutationsptr5 = load i64** %mutationsptr.ptr, align 8
191 %statemutations = load i64* %mutationsptr5, align 8
192 %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
193 br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
194
195forcoll.mutated:
196 call void @objc_enumerationMutation(i8* %1)
197 br label %forcoll.notmutated
198
199forcoll.notmutated:
200 %stateitems = load i8*** %stateitems.ptr, align 8
201 %currentitem.ptr = getelementptr i8** %stateitems, i64 %forcoll.index
202 %3 = load i8** %currentitem.ptr, align 8
203 call void @use(i8* %3)
204 %4 = add i64 %forcoll.index, 1
205 %exitcond = icmp eq i64 %4, %umax
206 br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
207
208forcoll.refetch:
209 %tmp7 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
210 %call8 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp7, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
211 %5 = icmp eq i64 %call8, 0
212 br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
213
214forcoll.empty:
215 call void @objc_release(i8* %1) nounwind
216 call void @objc_release(i8* %0) nounwind, !clang.imprecise_release !0
217 ret void
218}
219
220; Delete a nested retain+release pair.
221
Stephen Lina76289a2013-07-14 01:50:49 +0000222; CHECK-LABEL: define void @test5(
Dan Gohman12130272011-08-12 00:26:31 +0000223; CHECK: call i8* @objc_retain
224; CHECK-NOT: @objc_retain
225; CHECK: }
226define void @test5() nounwind {
227entry:
228 %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
229 %items.ptr = alloca [16 x i8*], align 8
230 %call = call i8* @returner()
231 %0 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call) nounwind
232 call void @callee()
233 %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
234 call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 64, i32 8, i1 false)
235 %1 = call i8* @objc_retain(i8* %0) nounwind
236 %tmp2 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
237 %call3 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
238 %iszero = icmp eq i64 %call3, 0
239 br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
240
241forcoll.loopinit:
242 %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
243 %mutationsptr = load i64** %mutationsptr.ptr, align 8
244 %forcoll.initial-mutations = load i64* %mutationsptr, align 8
245 %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
246 br label %forcoll.loopbody.outer
247
248forcoll.loopbody.outer:
249 %forcoll.count.ph = phi i64 [ %call3, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
250 %tmp8 = icmp ugt i64 %forcoll.count.ph, 1
251 %umax = select i1 %tmp8, i64 %forcoll.count.ph, i64 1
252 br label %forcoll.loopbody
253
254forcoll.loopbody:
255 %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
256 %mutationsptr4 = load i64** %mutationsptr.ptr, align 8
257 %statemutations = load i64* %mutationsptr4, align 8
258 %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
259 br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
260
261forcoll.mutated:
262 call void @objc_enumerationMutation(i8* %1)
263 br label %forcoll.notmutated
264
265forcoll.notmutated:
266 %stateitems = load i8*** %stateitems.ptr, align 8
267 %currentitem.ptr = getelementptr i8** %stateitems, i64 %forcoll.index
268 %3 = load i8** %currentitem.ptr, align 8
269 call void @use(i8* %3)
270 %4 = add i64 %forcoll.index, 1
271 %exitcond = icmp eq i64 %4, %umax
272 br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
273
274forcoll.refetch:
275 %tmp6 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
276 %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
277 %5 = icmp eq i64 %call7, 0
278 br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
279
280forcoll.empty:
281 call void @objc_release(i8* %1) nounwind
282 call void @objc_release(i8* %0) nounwind, !clang.imprecise_release !0
283 ret void
284}
285
Dan Gohmandf476e52012-09-04 23:16:20 +0000286; TODO: Delete a nested retain+release pair.
287; The optimizer currently can't do this, because isn't isn't sophisticated enough in
288; reasnoning about nesting.
Dan Gohman12130272011-08-12 00:26:31 +0000289
Stephen Lina76289a2013-07-14 01:50:49 +0000290; CHECK-LABEL: define void @test6(
Dan Gohman12130272011-08-12 00:26:31 +0000291; CHECK: call i8* @objc_retain
Dan Gohmandf476e52012-09-04 23:16:20 +0000292; CHECK: @objc_retain
Dan Gohman12130272011-08-12 00:26:31 +0000293; CHECK: }
294define void @test6() nounwind {
295entry:
296 %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
297 %items.ptr = alloca [16 x i8*], align 8
298 %call = call i8* @returner()
299 %0 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call) nounwind
300 %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
301 call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 64, i32 8, i1 false)
302 %1 = call i8* @objc_retain(i8* %0) nounwind
303 %tmp2 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
304 %call3 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
305 %iszero = icmp eq i64 %call3, 0
306 br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
307
308forcoll.loopinit:
309 %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
310 %mutationsptr = load i64** %mutationsptr.ptr, align 8
311 %forcoll.initial-mutations = load i64* %mutationsptr, align 8
312 %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
313 br label %forcoll.loopbody.outer
314
315forcoll.loopbody.outer:
316 %forcoll.count.ph = phi i64 [ %call3, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
317 %tmp8 = icmp ugt i64 %forcoll.count.ph, 1
318 %umax = select i1 %tmp8, i64 %forcoll.count.ph, i64 1
319 br label %forcoll.loopbody
320
321forcoll.loopbody:
322 %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
323 %mutationsptr4 = load i64** %mutationsptr.ptr, align 8
324 %statemutations = load i64* %mutationsptr4, align 8
325 %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
326 br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
327
328forcoll.mutated:
329 call void @objc_enumerationMutation(i8* %1)
330 br label %forcoll.notmutated
331
332forcoll.notmutated:
333 %stateitems = load i8*** %stateitems.ptr, align 8
334 %currentitem.ptr = getelementptr i8** %stateitems, i64 %forcoll.index
335 %3 = load i8** %currentitem.ptr, align 8
336 call void @use(i8* %3)
337 %4 = add i64 %forcoll.index, 1
338 %exitcond = icmp eq i64 %4, %umax
339 br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
340
341forcoll.refetch:
342 %tmp6 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
343 %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
344 %5 = icmp eq i64 %call7, 0
345 br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
346
347forcoll.empty:
348 call void @objc_release(i8* %1) nounwind
349 call void @callee()
350 call void @objc_release(i8* %0) nounwind, !clang.imprecise_release !0
351 ret void
352}
353
Dan Gohmandf476e52012-09-04 23:16:20 +0000354; TODO: Delete a nested retain+release pair.
355; The optimizer currently can't do this, because isn't isn't sophisticated enough in
356; reasnoning about nesting.
Dan Gohman12130272011-08-12 00:26:31 +0000357
Stephen Lina76289a2013-07-14 01:50:49 +0000358; CHECK-LABEL: define void @test7(
Dan Gohman12130272011-08-12 00:26:31 +0000359; CHECK: call i8* @objc_retain
Dan Gohmandf476e52012-09-04 23:16:20 +0000360; CHECK: @objc_retain
Dan Gohman12130272011-08-12 00:26:31 +0000361; CHECK: }
362define void @test7() nounwind {
363entry:
364 %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
365 %items.ptr = alloca [16 x i8*], align 8
366 %call = call i8* @returner()
367 %0 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call) nounwind
368 call void @callee()
369 %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
370 call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 64, i32 8, i1 false)
371 %1 = call i8* @objc_retain(i8* %0) nounwind
372 %tmp2 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
373 %call3 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
374 %iszero = icmp eq i64 %call3, 0
375 br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
376
377forcoll.loopinit:
378 %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
379 %mutationsptr = load i64** %mutationsptr.ptr, align 8
380 %forcoll.initial-mutations = load i64* %mutationsptr, align 8
381 %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
382 br label %forcoll.loopbody.outer
383
384forcoll.loopbody.outer:
385 %forcoll.count.ph = phi i64 [ %call3, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
386 %tmp8 = icmp ugt i64 %forcoll.count.ph, 1
387 %umax = select i1 %tmp8, i64 %forcoll.count.ph, i64 1
388 br label %forcoll.loopbody
389
390forcoll.loopbody:
391 %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
392 %mutationsptr4 = load i64** %mutationsptr.ptr, align 8
393 %statemutations = load i64* %mutationsptr4, align 8
394 %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
395 br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
396
397forcoll.mutated:
398 call void @objc_enumerationMutation(i8* %1)
399 br label %forcoll.notmutated
400
401forcoll.notmutated:
402 %stateitems = load i8*** %stateitems.ptr, align 8
403 %currentitem.ptr = getelementptr i8** %stateitems, i64 %forcoll.index
404 %3 = load i8** %currentitem.ptr, align 8
405 call void @use(i8* %3)
406 %4 = add i64 %forcoll.index, 1
407 %exitcond = icmp eq i64 %4, %umax
408 br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
409
410forcoll.refetch:
411 %tmp6 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
412 %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
413 %5 = icmp eq i64 %call7, 0
414 br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
415
416forcoll.empty:
417 call void @objc_release(i8* %1) nounwind
418 call void @callee()
419 call void @objc_release(i8* %0) nounwind, !clang.imprecise_release !0
420 ret void
421}
422
423; Delete a nested retain+release pair.
424
Stephen Lina76289a2013-07-14 01:50:49 +0000425; CHECK-LABEL: define void @test8(
Dan Gohman12130272011-08-12 00:26:31 +0000426; CHECK: call i8* @objc_retain
427; CHECK-NOT: @objc_retain
428; CHECK: }
429define void @test8() nounwind {
430entry:
431 %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
432 %items.ptr = alloca [16 x i8*], align 8
433 %call = call i8* @returner()
434 %0 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call) nounwind
435 %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
436 call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 64, i32 8, i1 false)
437 %1 = call i8* @objc_retain(i8* %0) nounwind
438 %tmp2 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
439 %call3 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
440 %iszero = icmp eq i64 %call3, 0
441 br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
442
443forcoll.loopinit:
444 %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
445 %mutationsptr = load i64** %mutationsptr.ptr, align 8
446 %forcoll.initial-mutations = load i64* %mutationsptr, align 8
447 %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
448 br label %forcoll.loopbody.outer
449
450forcoll.loopbody.outer:
451 %forcoll.count.ph = phi i64 [ %call3, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
452 %tmp8 = icmp ugt i64 %forcoll.count.ph, 1
453 %umax = select i1 %tmp8, i64 %forcoll.count.ph, i64 1
454 br label %forcoll.loopbody
455
456forcoll.loopbody:
457 %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.next ]
458 %mutationsptr4 = load i64** %mutationsptr.ptr, align 8
459 %statemutations = load i64* %mutationsptr4, align 8
460 %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
461 br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
462
463forcoll.mutated:
464 call void @objc_enumerationMutation(i8* %1)
465 br label %forcoll.notmutated
466
467forcoll.notmutated:
468 %stateitems = load i8*** %stateitems.ptr, align 8
469 %currentitem.ptr = getelementptr i8** %stateitems, i64 %forcoll.index
470 %3 = load i8** %currentitem.ptr, align 8
471 %tobool = icmp eq i8* %3, null
472 br i1 %tobool, label %forcoll.next, label %if.then
473
474if.then:
475 call void @callee()
476 br label %forcoll.next
477
478forcoll.next:
479 %4 = add i64 %forcoll.index, 1
480 %exitcond = icmp eq i64 %4, %umax
481 br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
482
483forcoll.refetch:
484 %tmp6 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
485 %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
486 %5 = icmp eq i64 %call7, 0
487 br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
488
489forcoll.empty:
490 call void @objc_release(i8* %1) nounwind
491 call void @objc_release(i8* %0) nounwind, !clang.imprecise_release !0
492 ret void
493}
494
Dan Gohman500b5982012-03-09 18:50:52 +0000495; TODO: Delete a nested retain+release pair.
496; The optimizer currently can't do this, because of a split loop backedge.
497; See test9b for the same testcase without a split backedge.
Dan Gohman12130272011-08-12 00:26:31 +0000498
Stephen Lina76289a2013-07-14 01:50:49 +0000499; CHECK-LABEL: define void @test9(
Dan Gohman12130272011-08-12 00:26:31 +0000500; CHECK: call i8* @objc_retain
501; CHECK: call i8* @objc_retain
Dan Gohman500b5982012-03-09 18:50:52 +0000502; CHECK: call i8* @objc_retain
Dan Gohman12130272011-08-12 00:26:31 +0000503; CHECK: }
504define void @test9() nounwind {
505entry:
506 %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
507 %items.ptr = alloca [16 x i8*], align 8
508 %call = call i8* @returner()
509 %0 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call) nounwind
510 %call1 = call i8* @returner()
511 %1 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call1) nounwind
512 %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
513 call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 64, i32 8, i1 false)
514 %2 = call i8* @objc_retain(i8* %0) nounwind
515 %tmp3 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
516 %call4 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp3, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
517 %iszero = icmp eq i64 %call4, 0
518 br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
519
520forcoll.loopinit:
521 %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
522 %mutationsptr = load i64** %mutationsptr.ptr, align 8
523 %forcoll.initial-mutations = load i64* %mutationsptr, align 8
524 br label %forcoll.loopbody.outer
525
526forcoll.loopbody.outer:
527 %forcoll.count.ph = phi i64 [ %call4, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
528 %tmp9 = icmp ugt i64 %forcoll.count.ph, 1
529 %umax = select i1 %tmp9, i64 %forcoll.count.ph, i64 1
530 br label %forcoll.loopbody
531
532forcoll.loopbody:
533 %forcoll.index = phi i64 [ %phitmp, %forcoll.notmutated.forcoll.loopbody_crit_edge ], [ 1, %forcoll.loopbody.outer ]
534 %mutationsptr5 = load i64** %mutationsptr.ptr, align 8
535 %statemutations = load i64* %mutationsptr5, align 8
536 %3 = icmp eq i64 %statemutations, %forcoll.initial-mutations
537 br i1 %3, label %forcoll.notmutated, label %forcoll.mutated
538
539forcoll.mutated:
540 call void @objc_enumerationMutation(i8* %2)
541 br label %forcoll.notmutated
542
543forcoll.notmutated:
544 %exitcond = icmp eq i64 %forcoll.index, %umax
545 br i1 %exitcond, label %forcoll.refetch, label %forcoll.notmutated.forcoll.loopbody_crit_edge
546
547forcoll.notmutated.forcoll.loopbody_crit_edge:
548 %phitmp = add i64 %forcoll.index, 1
549 br label %forcoll.loopbody
550
551forcoll.refetch:
552 %tmp6 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
553 %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
554 %4 = icmp eq i64 %call7, 0
555 br i1 %4, label %forcoll.empty, label %forcoll.loopbody.outer
556
557forcoll.empty:
558 call void @objc_release(i8* %2) nounwind
559 call void @objc_release(i8* %1) nounwind, !clang.imprecise_release !0
560 call void @objc_release(i8* %0) nounwind, !clang.imprecise_release !0
561 ret void
562}
563
Dan Gohmandf476e52012-09-04 23:16:20 +0000564; Like test9, but without a split backedge. TODO: optimize this.
Dan Gohman500b5982012-03-09 18:50:52 +0000565
Stephen Lina76289a2013-07-14 01:50:49 +0000566; CHECK-LABEL: define void @test9b(
Dan Gohman500b5982012-03-09 18:50:52 +0000567; CHECK: call i8* @objc_retain
568; CHECK: call i8* @objc_retain
Dan Gohmandf476e52012-09-04 23:16:20 +0000569; CHECK: @objc_retain
Dan Gohman500b5982012-03-09 18:50:52 +0000570; CHECK: }
571define void @test9b() nounwind {
572entry:
573 %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
574 %items.ptr = alloca [16 x i8*], align 8
575 %call = call i8* @returner()
576 %0 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call) nounwind
577 %call1 = call i8* @returner()
578 %1 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call1) nounwind
579 %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
580 call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 64, i32 8, i1 false)
581 %2 = call i8* @objc_retain(i8* %0) nounwind
582 %tmp3 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
583 %call4 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp3, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
584 %iszero = icmp eq i64 %call4, 0
585 br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
586
587forcoll.loopinit:
588 %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
589 %mutationsptr = load i64** %mutationsptr.ptr, align 8
590 %forcoll.initial-mutations = load i64* %mutationsptr, align 8
591 br label %forcoll.loopbody.outer
592
593forcoll.loopbody.outer:
594 %forcoll.count.ph = phi i64 [ %call4, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
595 %tmp9 = icmp ugt i64 %forcoll.count.ph, 1
596 %umax = select i1 %tmp9, i64 %forcoll.count.ph, i64 1
597 br label %forcoll.loopbody
598
599forcoll.loopbody:
600 %forcoll.index = phi i64 [ %phitmp, %forcoll.notmutated ], [ 0, %forcoll.loopbody.outer ]
601 %mutationsptr5 = load i64** %mutationsptr.ptr, align 8
602 %statemutations = load i64* %mutationsptr5, align 8
603 %3 = icmp eq i64 %statemutations, %forcoll.initial-mutations
604 br i1 %3, label %forcoll.notmutated, label %forcoll.mutated
605
606forcoll.mutated:
607 call void @objc_enumerationMutation(i8* %2)
608 br label %forcoll.notmutated
609
610forcoll.notmutated:
611 %phitmp = add i64 %forcoll.index, 1
612 %exitcond = icmp eq i64 %phitmp, %umax
613 br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
614
615forcoll.refetch:
616 %tmp6 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
617 %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
618 %4 = icmp eq i64 %call7, 0
619 br i1 %4, label %forcoll.empty, label %forcoll.loopbody.outer
620
621forcoll.empty:
622 call void @objc_release(i8* %2) nounwind
623 call void @objc_release(i8* %1) nounwind, !clang.imprecise_release !0
624 call void @objc_release(i8* %0) nounwind, !clang.imprecise_release !0
625 ret void
626}
627
628; TODO: Delete a nested retain+release pair.
629; The optimizer currently can't do this, because of a split loop backedge.
630; See test10b for the same testcase without a split backedge.
Dan Gohman12130272011-08-12 00:26:31 +0000631
Stephen Lina76289a2013-07-14 01:50:49 +0000632; CHECK-LABEL: define void @test10(
Dan Gohman12130272011-08-12 00:26:31 +0000633; CHECK: call i8* @objc_retain
634; CHECK: call i8* @objc_retain
Dan Gohman500b5982012-03-09 18:50:52 +0000635; CHECK: call i8* @objc_retain
Dan Gohman12130272011-08-12 00:26:31 +0000636; CHECK: }
637define void @test10() nounwind {
638entry:
639 %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
640 %items.ptr = alloca [16 x i8*], align 8
641 %call = call i8* @returner()
642 %0 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call) nounwind
643 %call1 = call i8* @returner()
644 %1 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call1) nounwind
645 call void @callee()
646 %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
647 call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 64, i32 8, i1 false)
648 %2 = call i8* @objc_retain(i8* %0) nounwind
649 %tmp3 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
650 %call4 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp3, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
651 %iszero = icmp eq i64 %call4, 0
652 br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
653
654forcoll.loopinit:
655 %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
656 %mutationsptr = load i64** %mutationsptr.ptr, align 8
657 %forcoll.initial-mutations = load i64* %mutationsptr, align 8
658 br label %forcoll.loopbody.outer
659
660forcoll.loopbody.outer:
661 %forcoll.count.ph = phi i64 [ %call4, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
662 %tmp9 = icmp ugt i64 %forcoll.count.ph, 1
663 %umax = select i1 %tmp9, i64 %forcoll.count.ph, i64 1
664 br label %forcoll.loopbody
665
666forcoll.loopbody:
667 %forcoll.index = phi i64 [ %phitmp, %forcoll.notmutated.forcoll.loopbody_crit_edge ], [ 1, %forcoll.loopbody.outer ]
668 %mutationsptr5 = load i64** %mutationsptr.ptr, align 8
669 %statemutations = load i64* %mutationsptr5, align 8
670 %3 = icmp eq i64 %statemutations, %forcoll.initial-mutations
671 br i1 %3, label %forcoll.notmutated, label %forcoll.mutated
672
673forcoll.mutated:
674 call void @objc_enumerationMutation(i8* %2)
675 br label %forcoll.notmutated
676
677forcoll.notmutated:
678 %exitcond = icmp eq i64 %forcoll.index, %umax
679 br i1 %exitcond, label %forcoll.refetch, label %forcoll.notmutated.forcoll.loopbody_crit_edge
680
681forcoll.notmutated.forcoll.loopbody_crit_edge:
682 %phitmp = add i64 %forcoll.index, 1
683 br label %forcoll.loopbody
684
685forcoll.refetch:
686 %tmp6 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
687 %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
688 %4 = icmp eq i64 %call7, 0
689 br i1 %4, label %forcoll.empty, label %forcoll.loopbody.outer
690
691forcoll.empty:
692 call void @objc_release(i8* %2) nounwind
693 call void @objc_release(i8* %1) nounwind, !clang.imprecise_release !0
694 call void @objc_release(i8* %0) nounwind, !clang.imprecise_release !0
695 ret void
696}
Dan Gohman500b5982012-03-09 18:50:52 +0000697
Dan Gohmandf476e52012-09-04 23:16:20 +0000698; Like test10, but without a split backedge. TODO: optimize this.
Dan Gohman500b5982012-03-09 18:50:52 +0000699
Stephen Lina76289a2013-07-14 01:50:49 +0000700; CHECK-LABEL: define void @test10b(
Dan Gohman500b5982012-03-09 18:50:52 +0000701; CHECK: call i8* @objc_retain
702; CHECK: call i8* @objc_retain
Dan Gohmandf476e52012-09-04 23:16:20 +0000703; CHECK: @objc_retain
Dan Gohman500b5982012-03-09 18:50:52 +0000704; CHECK: }
705define void @test10b() nounwind {
706entry:
707 %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
708 %items.ptr = alloca [16 x i8*], align 8
709 %call = call i8* @returner()
710 %0 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call) nounwind
711 %call1 = call i8* @returner()
712 %1 = call i8* @objc_retainAutoreleasedReturnValue(i8* %call1) nounwind
713 call void @callee()
714 %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
715 call void @llvm.memset.p0i8.i64(i8* %tmp, i8 0, i64 64, i32 8, i1 false)
716 %2 = call i8* @objc_retain(i8* %0) nounwind
717 %tmp3 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
718 %call4 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp3, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
719 %iszero = icmp eq i64 %call4, 0
720 br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
721
722forcoll.loopinit:
723 %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
724 %mutationsptr = load i64** %mutationsptr.ptr, align 8
725 %forcoll.initial-mutations = load i64* %mutationsptr, align 8
726 br label %forcoll.loopbody.outer
727
728forcoll.loopbody.outer:
729 %forcoll.count.ph = phi i64 [ %call4, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
730 %tmp9 = icmp ugt i64 %forcoll.count.ph, 1
731 %umax = select i1 %tmp9, i64 %forcoll.count.ph, i64 1
732 br label %forcoll.loopbody
733
734forcoll.loopbody:
735 %forcoll.index = phi i64 [ %phitmp, %forcoll.notmutated ], [ 0, %forcoll.loopbody.outer ]
736 %mutationsptr5 = load i64** %mutationsptr.ptr, align 8
737 %statemutations = load i64* %mutationsptr5, align 8
738 %3 = icmp eq i64 %statemutations, %forcoll.initial-mutations
739 br i1 %3, label %forcoll.notmutated, label %forcoll.mutated
740
741forcoll.mutated:
742 call void @objc_enumerationMutation(i8* %2)
743 br label %forcoll.notmutated
744
745forcoll.notmutated:
746 %phitmp = add i64 %forcoll.index, 1
747 %exitcond = icmp eq i64 %phitmp, %umax
748 br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
749
750forcoll.refetch:
751 %tmp6 = load i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
752 %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @objc_msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
753 %4 = icmp eq i64 %call7, 0
754 br i1 %4, label %forcoll.empty, label %forcoll.loopbody.outer
755
756forcoll.empty:
757 call void @objc_release(i8* %2) nounwind
758 call void @objc_release(i8* %1) nounwind, !clang.imprecise_release !0
759 call void @objc_release(i8* %0) nounwind, !clang.imprecise_release !0
760 ret void
761}
Dan Gohmandf476e52012-09-04 23:16:20 +0000762
763; Pointers to strong pointers can obscure provenance relationships. Be conservative
764; in the face of escaping pointers. rdar://12150909.
765
766%struct.__block_d = type { i64, i64 }
767
768@_NSConcreteStackBlock = external global i8*
769@__block_d_tmp = external hidden constant { i64, i64, i8*, i8*, i8*, i8* }
770@__block_d_tmp5 = external hidden constant { i64, i64, i8*, i8*, i8*, i8* }
771
Stephen Lina76289a2013-07-14 01:50:49 +0000772; CHECK-LABEL: define void @test11(
Bill Wendlinga0323742013-02-22 09:09:42 +0000773; CHECK: tail call i8* @objc_retain(i8* %call) [[NUW:#[0-9]+]]
774; CHECK: tail call i8* @objc_retain(i8* %call) [[NUW]]
775; CHECK: call void @objc_release(i8* %call) [[NUW]], !clang.imprecise_release !0
Dan Gohmandf476e52012-09-04 23:16:20 +0000776; CHECK: }
777define void @test11() {
778entry:
779 %block = alloca <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, align 8
780 %block9 = alloca <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, align 8
781 %call = call i8* @def(), !clang.arc.no_objc_arc_exceptions !0
782 %foo = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 5
783 %block.isa = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 0
784 store i8* bitcast (i8** @_NSConcreteStackBlock to i8*), i8** %block.isa, align 8
785 %block.flags = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 1
786 store i32 1107296256, i32* %block.flags, align 8
787 %block.reserved = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 2
788 store i32 0, i32* %block.reserved, align 4
789 %block.invoke = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 3
790 store i8* bitcast (void (i8*)* @__crasher_block_invoke to i8*), i8** %block.invoke, align 8
791 %block.d = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 4
792 store %struct.__block_d* bitcast ({ i64, i64, i8*, i8*, i8*, i8* }* @__block_d_tmp to %struct.__block_d*), %struct.__block_d** %block.d, align 8
793 %foo2 = tail call i8* @objc_retain(i8* %call) nounwind
794 store i8* %foo2, i8** %foo, align 8
795 %foo4 = bitcast <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block to i8*
796 %foo5 = call i8* @objc_retainBlock(i8* %foo4) nounwind
797 call void @use(i8* %foo5), !clang.arc.no_objc_arc_exceptions !0
798 call void @objc_release(i8* %foo5) nounwind
799 %strongdestroy = load i8** %foo, align 8
800 call void @objc_release(i8* %strongdestroy) nounwind, !clang.imprecise_release !0
801 %foo10 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 5
802 %block.isa11 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 0
803 store i8* bitcast (i8** @_NSConcreteStackBlock to i8*), i8** %block.isa11, align 8
804 %block.flags12 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 1
805 store i32 1107296256, i32* %block.flags12, align 8
806 %block.reserved13 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 2
807 store i32 0, i32* %block.reserved13, align 4
808 %block.invoke14 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 3
809 store i8* bitcast (void (i8*)* @__crasher_block_invoke1 to i8*), i8** %block.invoke14, align 8
810 %block.d15 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 4
811 store %struct.__block_d* bitcast ({ i64, i64, i8*, i8*, i8*, i8* }* @__block_d_tmp5 to %struct.__block_d*), %struct.__block_d** %block.d15, align 8
812 %foo18 = call i8* @objc_retain(i8* %call) nounwind
813 store i8* %call, i8** %foo10, align 8
814 %foo20 = bitcast <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9 to i8*
815 %foo21 = call i8* @objc_retainBlock(i8* %foo20) nounwind
816 call void @use(i8* %foo21), !clang.arc.no_objc_arc_exceptions !0
817 call void @objc_release(i8* %foo21) nounwind
818 %strongdestroy25 = load i8** %foo10, align 8
819 call void @objc_release(i8* %strongdestroy25) nounwind, !clang.imprecise_release !0
820 call void @objc_release(i8* %call) nounwind, !clang.imprecise_release !0
821 ret void
822}
Bill Wendlinga0323742013-02-22 09:09:42 +0000823
824; CHECK: attributes [[NUW]] = { nounwind }
825; CHECK: attributes #1 = { nonlazybind }