blob: 07672584362a5ed4d99043311f9cf9e6a03109f7 [file] [log] [blame]
Momchil Velikovd501a602018-04-18 13:37:12 +00001; Test shrink wrapping placement is correct with respect to calls to llvm.{stacksave,stackrestore}
2
3; void f(int n, int x[]) {
4; if (n < 0)
5; return;
6;
7; int a[n];
8;
9; for (int i = 0; i < n; i++)
10; a[i] = x[n - i - 1];
11;
12; for (int i = 0; i < n; i++)
13; x[i] = a[i] + 1;
14; }
15;
16; RUN: llc -mtriple aarch64-linux %s -o - | FileCheck %s
17
18define dso_local void @f(i32 %n, i32* nocapture %x) {
19entry:
20 %cmp = icmp slt i32 %n, 0
21 br i1 %cmp, label %return, label %if.end
22
23if.end: ; preds = %entry
24 %0 = zext i32 %n to i64
25 %1 = tail call i8* @llvm.stacksave()
26 %vla = alloca i32, i64 %0, align 16
27 %cmp132 = icmp eq i32 %n, 0
28 br i1 %cmp132, label %for.cond.cleanup8, label %for.body.lr.ph
29
30for.body.lr.ph: ; preds = %if.end
31 %sub = add i32 %n, -1
32 br label %for.body
33
34for.cond6.preheader: ; preds = %for.body
35 %cmp730 = icmp sgt i32 %n, 0
36 br i1 %cmp730, label %for.body9, label %for.cond.cleanup8
37
38for.body: ; preds = %for.body, %for.body.lr.ph
39 %indvars.iv34 = phi i64 [ 0, %for.body.lr.ph ], [ %indvars.iv.next35, %for.body ]
40 %2 = trunc i64 %indvars.iv34 to i32
41 %sub2 = sub i32 %sub, %2
42 %idxprom = sext i32 %sub2 to i64
43 %arrayidx = getelementptr inbounds i32, i32* %x, i64 %idxprom
44 %3 = load i32, i32* %arrayidx, align 4
45 %arrayidx4 = getelementptr inbounds i32, i32* %vla, i64 %indvars.iv34
46 store i32 %3, i32* %arrayidx4, align 4
47 %indvars.iv.next35 = add nuw nsw i64 %indvars.iv34, 1
48 %exitcond37 = icmp eq i64 %indvars.iv.next35, %0
49 br i1 %exitcond37, label %for.cond6.preheader, label %for.body
50
51for.cond.cleanup8: ; preds = %for.body9, %if.end, %for.cond6.preheader
52 tail call void @llvm.stackrestore(i8* %1)
53 br label %return
54
55for.body9: ; preds = %for.cond6.preheader, %for.body9
56 %indvars.iv = phi i64 [ %indvars.iv.next, %for.body9 ], [ 0, %for.cond6.preheader ]
57 %arrayidx11 = getelementptr inbounds i32, i32* %vla, i64 %indvars.iv
58 %4 = load i32, i32* %arrayidx11, align 4
59 %add = add nsw i32 %4, 1
60 %arrayidx13 = getelementptr inbounds i32, i32* %x, i64 %indvars.iv
61 store i32 %add, i32* %arrayidx13, align 4
62 %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1
63 %exitcond = icmp eq i64 %indvars.iv.next, %0
64 br i1 %exitcond, label %for.cond.cleanup8, label %for.body9
65
66return: ; preds = %entry, %for.cond.cleanup8
67 ret void
68}
69
70; Function Attrs: nounwind
71declare i8* @llvm.stacksave()
72
73; Function Attrs: nounwind
74declare void @llvm.stackrestore(i8*)
75
76; Check that llvm.stackrestore() happens before CSRs are popped off the stack
77
78; CHECK-LABEL: f
79
80; CHECK: stp x29, x30, [sp, #-16]!
81; CHECK-NEXT: mov x29, sp
82
83; VLA allocation
84; CHECK: add [[X1:x[0-9]+]], [[X1]], #15
85; CHECK: mov [[X2:x[0-9]+]], sp
86; CHECK: and [[X1]], [[X1]], #0x7fffffff0
87; Saving the SP via llvm.stacksave()
88; CHECK: mov [[SAVE:x[0-9]+]], sp
89; CHECK: sub [[X2]], [[X2]], [[X1]]
90
91; The next instruction comes from llvm.stackrestore()
92; CHECK: mov sp, [[SAVE]]
93; Epilogue
94; CHECK-NEXT: mov sp, x29
95; CHECK-NEXT: ldp x29, x30, [sp], #16