blob: f9c36e3c77cfd3be4dc66d0f60f0d4a12adede56 [file] [log] [blame]
Chad Rosier3d3c75c2012-07-10 17:49:39 +00001; RUN: llc < %s -march=x86-64 -mattr=+avx -mtriple=i686-apple-darwin10 | FileCheck %s
2; RUN: llc < %s -force-align-stack -stack-alignment=32 -march=x86-64 -mattr=+avx -mtriple=i686-apple-darwin10 | FileCheck %s -check-prefix=FORCE-ALIGN
3; rdar://11496434
4
5; no VLAs or dynamic alignment
6define i32 @t1() nounwind uwtable ssp {
7entry:
8 %a = alloca i32, align 4
9 call void @t1_helper(i32* %a) nounwind
10 %0 = load i32* %a, align 4
11 %add = add nsw i32 %0, 13
12 ret i32 %add
13
14; CHECK: _t1
15; CHECK-NOT: andq $-{{[0-9]+}}, %rsp
16; CHECK: leaq [[OFFSET:[0-9]*]](%rsp), %rdi
17; CHECK: callq _t1_helper
18; CHECK: movl [[OFFSET]](%rsp), %eax
19; CHECK: addl $13, %eax
20}
21
22declare void @t1_helper(i32*)
23
24; dynamic realignment
25define i32 @t2() nounwind uwtable ssp {
26entry:
27 %a = alloca i32, align 4
28 %v = alloca <8 x float>, align 32
29 call void @t2_helper(i32* %a, <8 x float>* %v) nounwind
30 %0 = load i32* %a, align 4
31 %add = add nsw i32 %0, 13
32 ret i32 %add
33
34; CHECK: _t2
35; CHECK: pushq %rbp
36; CHECK: movq %rsp, %rbp
37; CHECK: andq $-32, %rsp
38; CHECK: subq ${{[0-9]+}}, %rsp
39;
40; CHECK: leaq {{[0-9]*}}(%rsp), %rdi
41; CHECK: leaq {{[0-9]*}}(%rsp), %rsi
42; CHECK: callq _t2_helper
43;
44; CHECK: movq %rbp, %rsp
45; CHECK: popq %rbp
46}
47
48declare void @t2_helper(i32*, <8 x float>*)
49
50; VLAs
51define i32 @t3(i64 %sz) nounwind uwtable ssp {
52entry:
53 %a = alloca i32, align 4
54 %vla = alloca i32, i64 %sz, align 16
55 call void @t3_helper(i32* %a, i32* %vla) nounwind
56 %0 = load i32* %a, align 4
57 %add = add nsw i32 %0, 13
58 ret i32 %add
59
60; CHECK: _t3
61; CHECK: pushq %rbp
62; CHECK: movq %rsp, %rbp
63; CHECK: pushq %rbx
64; CHECK-NOT: andq $-{{[0-9]+}}, %rsp
65; CHECK: subq ${{[0-9]+}}, %rsp
66;
67; CHECK: leaq -{{[0-9]+}}(%rbp), %rsp
68; CHECK: popq %rbx
69; CHECK: popq %rbp
70}
71
72declare void @t3_helper(i32*, i32*)
73
74; VLAs + Dynamic realignment
75define i32 @t4(i64 %sz) nounwind uwtable ssp {
76entry:
77 %a = alloca i32, align 4
78 %v = alloca <8 x float>, align 32
79 %vla = alloca i32, i64 %sz, align 16
80 call void @t4_helper(i32* %a, i32* %vla, <8 x float>* %v) nounwind
81 %0 = load i32* %a, align 4
82 %add = add nsw i32 %0, 13
83 ret i32 %add
84
85; CHECK: _t4
86; CHECK: pushq %rbp
87; CHECK: movq %rsp, %rbp
88; CHECK: andq $-32, %rsp
89; CHECK: pushq %r14
90; CHECK: pushq %rbx
91; CHECK: subq $[[STACKADJ:[0-9]+]], %rsp
92; CHECK: movq %rsp, %rbx
93;
94; CHECK: leaq {{[0-9]*}}(%rbx), %rdi
95; CHECK: leaq {{[0-9]*}}(%rbx), %rdx
96; CHECK: callq _t4_helper
97;
98; CHECK: addq $[[STACKADJ]], %rsp
99; CHECK: popq %rbx
100; CHECK: popq %r14
101; CHECK: movq %rbp, %rsp
102; CHECK: popq %rbp
103}
104
105declare void @t4_helper(i32*, i32*, <8 x float>*)
106
107; Dynamic realignment + Spill
108define i32 @t5(float* nocapture %f) nounwind uwtable ssp {
109entry:
110 %a = alloca i32, align 4
111 %0 = bitcast float* %f to <8 x float>*
112 %1 = load <8 x float>* %0, align 32
113 call void @t5_helper1(i32* %a) nounwind
114 call void @t5_helper2(<8 x float> %1) nounwind
115 %2 = load i32* %a, align 4
116 %add = add nsw i32 %2, 13
117 ret i32 %add
118
119; CHECK: _t5
120; CHECK: pushq %rbp
121; CHECK: movq %rsp, %rbp
122; CHECK: andq $-32, %rsp
123; CHECK: subq ${{[0-9]+}}, %rsp
124;
125; CHECK: vmovaps (%rdi), [[AVXREG:%ymm[0-9]+]]
126; CHECK: vmovaps [[AVXREG]], (%rsp)
127; CHECK: leaq {{[0-9]+}}(%rsp), %rdi
128; CHECK: callq _t5_helper1
129; CHECK: vmovaps (%rsp), %ymm0
130; CHECK: callq _t5_helper2
131; CHECK: movl {{[0-9]+}}(%rsp), %eax
132;
133; CHECK: movq %rbp, %rsp
134; CHECK: popq %rbp
135}
136
137declare void @t5_helper1(i32*)
138
139declare void @t5_helper2(<8 x float>)
140
141; VLAs + Dynamic realignment + Spill
142; FIXME: RA has already reserved RBX, so we can't do dynamic realignment.
143define i32 @t6(i64 %sz, float* nocapture %f) nounwind uwtable ssp {
144entry:
145; CHECK: _t6
146 %a = alloca i32, align 4
147 %0 = bitcast float* %f to <8 x float>*
148 %1 = load <8 x float>* %0, align 32
149 %vla = alloca i32, i64 %sz, align 16
150 call void @t6_helper1(i32* %a, i32* %vla) nounwind
151 call void @t6_helper2(<8 x float> %1) nounwind
152 %2 = load i32* %a, align 4
153 %add = add nsw i32 %2, 13
154 ret i32 %add
155}
156
157declare void @t6_helper1(i32*, i32*)
158
159declare void @t6_helper2(<8 x float>)
160
161; VLAs + Dynamic realignment + byval
162; The byval adjust the sp after the prolog, but if we're restoring the sp from
163; the base pointer we use the original adjustment.
164%struct.struct_t = type { [5 x i32] }
165
166define void @t7(i32 %size, %struct.struct_t* byval align 8 %arg1) nounwind uwtable {
167entry:
168 %x = alloca i32, align 32
169 store i32 0, i32* %x, align 32
170 %0 = zext i32 %size to i64
171 %vla = alloca i32, i64 %0, align 16
172 %1 = load i32* %x, align 32
173 call void @bar(i32 %1, i32* %vla, %struct.struct_t* byval align 8 %arg1)
174 ret void
175
176; CHECK: _t7
177; CHECK: pushq %rbp
178; CHECK: movq %rsp, %rbp
179; CHECK: andq $-32, %rsp
180; CHECK: pushq %rbx
181; CHECK: subq $[[ADJ:[0-9]+]], %rsp
182; CHECK: movq %rsp, %rbx
183
184; Stack adjustment for byval
185; CHECK: subq {{.*}}, %rsp
186; CHECK: callq _bar
187; CHECK-NOT: addq {{.*}}, %rsp
188; CHECK: movq %rbx, %rsp
189; CHECK: addq $[[ADJ]], %rsp
190; CHECK: popq %rbx
191; CHECK: movq %rbp, %rsp
192; CHECK: popq %rbp
193}
194
195declare i8* @llvm.stacksave() nounwind
196
197declare void @bar(i32, i32*, %struct.struct_t* byval align 8)
198
199declare void @llvm.stackrestore(i8*) nounwind
200
201
202; Test when forcing stack alignment
203define i32 @t8() nounwind uwtable {
204entry:
205 %a = alloca i32, align 4
206 call void @t1_helper(i32* %a) nounwind
207 %0 = load i32* %a, align 4
208 %add = add nsw i32 %0, 13
209 ret i32 %add
210
211; FORCE-ALIGN: _t8
212; FORCE-ALIGN: movq %rsp, %rbp
213; FORCE-ALIGN: andq $-32, %rsp
214; FORCE-ALIGN-NEXT: subq $32, %rsp
215; FORCE-ALIGN: movq %rbp, %rsp
216; FORCE-ALIGN: popq %rbp
217}
218
219; VLAs
220define i32 @t9(i64 %sz) nounwind uwtable {
221entry:
222 %a = alloca i32, align 4
223 %vla = alloca i32, i64 %sz, align 16
224 call void @t3_helper(i32* %a, i32* %vla) nounwind
225 %0 = load i32* %a, align 4
226 %add = add nsw i32 %0, 13
227 ret i32 %add
228
229; FORCE-ALIGN: _t9
230; FORCE-ALIGN: pushq %rbp
231; FORCE-ALIGN: movq %rsp, %rbp
232; FORCE-ALIGN: andq $-32, %rsp
233; FORCE-ALIGN: pushq %rbx
234; FORCE-ALIGN: subq $24, %rsp
235; FORCE-ALIGN: movq %rsp, %rbx
236
237; FORCE-ALIGN: movq %rbx, %rsp
238; FORCE-ALIGN: addq $24, %rsp
239; FORCE-ALIGN: popq %rbx
240; FORCE-ALIGN: movq %rbp, %rsp
241; FORCE-ALIGN: popq %rbp
242}