blob: 8f9a4b793a9ef3d4fc770be9fb0e6b9499e78836 [file] [log] [blame]
Alex Bradburyc85be0d2018-01-10 19:41:03 +00001; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
2; RUN: llc -mtriple=riscv32 -verify-machineinstrs < %s \
3; RUN: | FileCheck -check-prefix=RV32I %s
4
5declare void @llvm.va_start(i8*)
6declare void @llvm.va_end(i8*)
7
8declare void @notdead(i8*)
9
10; Although frontends are recommended to not generate va_arg due to the lack of
11; support for aggregate types, we test simple cases here to ensure they are
12; lowered correctly
13
14define i32 @va1(i8* %fmt, ...) nounwind {
15; RV32I-LABEL: va1:
16; RV32I: # %bb.0:
17; RV32I-NEXT: addi sp, sp, -48
18; RV32I-NEXT: sw ra, 12(sp)
19; RV32I-NEXT: sw s0, 8(sp)
20; RV32I-NEXT: addi s0, sp, 16
21; RV32I-NEXT: sw a1, 4(s0)
22; RV32I-NEXT: sw a7, 28(s0)
23; RV32I-NEXT: sw a6, 24(s0)
24; RV32I-NEXT: sw a5, 20(s0)
25; RV32I-NEXT: sw a4, 16(s0)
26; RV32I-NEXT: sw a3, 12(s0)
27; RV32I-NEXT: sw a2, 8(s0)
28; RV32I-NEXT: addi a0, s0, 8
29; RV32I-NEXT: sw a0, -12(s0)
30; RV32I-NEXT: lw a0, 4(s0)
31; RV32I-NEXT: lw s0, 8(sp)
32; RV32I-NEXT: lw ra, 12(sp)
33; RV32I-NEXT: addi sp, sp, 48
34; RV32I-NEXT: ret
35 %va = alloca i8*, align 4
36 %1 = bitcast i8** %va to i8*
37 call void @llvm.va_start(i8* %1)
38 %argp.cur = load i8*, i8** %va, align 4
39 %argp.next = getelementptr inbounds i8, i8* %argp.cur, i32 4
40 store i8* %argp.next, i8** %va, align 4
41 %2 = bitcast i8* %argp.cur to i32*
42 %3 = load i32, i32* %2, align 4
43 call void @llvm.va_end(i8* %1)
44 ret i32 %3
45}
46
47define i32 @va1_va_arg(i8* %fmt, ...) nounwind {
48; RV32I-LABEL: va1_va_arg:
49; RV32I: # %bb.0:
50; RV32I-NEXT: addi sp, sp, -48
51; RV32I-NEXT: sw ra, 12(sp)
52; RV32I-NEXT: sw s0, 8(sp)
53; RV32I-NEXT: addi s0, sp, 16
54; RV32I-NEXT: sw a1, 4(s0)
55; RV32I-NEXT: sw a7, 28(s0)
56; RV32I-NEXT: sw a6, 24(s0)
57; RV32I-NEXT: sw a5, 20(s0)
58; RV32I-NEXT: sw a4, 16(s0)
59; RV32I-NEXT: sw a3, 12(s0)
60; RV32I-NEXT: sw a2, 8(s0)
61; RV32I-NEXT: addi a0, s0, 8
62; RV32I-NEXT: sw a0, -12(s0)
63; RV32I-NEXT: lw a0, 4(s0)
64; RV32I-NEXT: lw s0, 8(sp)
65; RV32I-NEXT: lw ra, 12(sp)
66; RV32I-NEXT: addi sp, sp, 48
67; RV32I-NEXT: ret
68 %va = alloca i8*, align 4
69 %1 = bitcast i8** %va to i8*
70 call void @llvm.va_start(i8* %1)
71 %2 = va_arg i8** %va, i32
72 call void @llvm.va_end(i8* %1)
73 ret i32 %2
74}
75
76; Ensure the adjustment when restoring the stack pointer using the frame
77; pointer is correct
78define i32 @va1_va_arg_alloca(i8* %fmt, ...) nounwind {
79; RV32I-LABEL: va1_va_arg_alloca:
80; RV32I: # %bb.0:
81; RV32I-NEXT: addi sp, sp, -48
82; RV32I-NEXT: sw ra, 12(sp)
83; RV32I-NEXT: sw s0, 8(sp)
84; RV32I-NEXT: sw s1, 4(sp)
85; RV32I-NEXT: addi s0, sp, 16
86; RV32I-NEXT: sw a1, 4(s0)
87; RV32I-NEXT: sw a7, 28(s0)
88; RV32I-NEXT: sw a6, 24(s0)
89; RV32I-NEXT: sw a5, 20(s0)
90; RV32I-NEXT: sw a4, 16(s0)
91; RV32I-NEXT: sw a3, 12(s0)
92; RV32I-NEXT: sw a2, 8(s0)
93; RV32I-NEXT: addi a0, s0, 8
94; RV32I-NEXT: sw a0, -16(s0)
95; RV32I-NEXT: lw s1, 4(s0)
96; RV32I-NEXT: addi a0, s1, 15
97; RV32I-NEXT: andi a0, a0, -16
98; RV32I-NEXT: sub a0, sp, a0
99; RV32I-NEXT: mv sp, a0
100; RV32I-NEXT: lui a1, %hi(notdead)
101; RV32I-NEXT: addi a1, a1, %lo(notdead)
102; RV32I-NEXT: jalr a1
103; RV32I-NEXT: mv a0, s1
104; RV32I-NEXT: addi sp, s0, -16
105; RV32I-NEXT: lw s1, 4(sp)
106; RV32I-NEXT: lw s0, 8(sp)
107; RV32I-NEXT: lw ra, 12(sp)
108; RV32I-NEXT: addi sp, sp, 48
109; RV32I-NEXT: ret
110 %va = alloca i8*, align 4
111 %1 = bitcast i8** %va to i8*
112 call void @llvm.va_start(i8* %1)
113 %2 = va_arg i8** %va, i32
114 %3 = alloca i8, i32 %2
115 call void @notdead(i8* %3)
116 call void @llvm.va_end(i8* %1)
117 ret i32 %2
118}
119
120define void @va1_caller() nounwind {
121; RV32I-LABEL: va1_caller:
122; RV32I: # %bb.0:
123; RV32I-NEXT: addi sp, sp, -16
124; RV32I-NEXT: sw ra, 12(sp)
125; RV32I-NEXT: sw s0, 8(sp)
126; RV32I-NEXT: addi s0, sp, 16
127; RV32I-NEXT: lui a0, 261888
128; RV32I-NEXT: mv a3, a0
129; RV32I-NEXT: lui a0, %hi(va1)
130; RV32I-NEXT: addi a0, a0, %lo(va1)
131; RV32I-NEXT: addi a4, zero, 2
132; RV32I-NEXT: mv a2, zero
133; RV32I-NEXT: jalr a0
134; RV32I-NEXT: lw s0, 8(sp)
135; RV32I-NEXT: lw ra, 12(sp)
136; RV32I-NEXT: addi sp, sp, 16
137; RV32I-NEXT: ret
138; Pass a double, as a float would be promoted by a C/C++ frontend
139 %1 = call i32 (i8*, ...) @va1(i8* undef, double 1.0, i32 2)
140 ret void
141}
142
143; Ensure that 2x xlen size+alignment varargs are accessed via an "aligned"
144; register pair (where the first register is even-numbered).
145
146define double @va2(i8 *%fmt, ...) nounwind {
147; RV32I-LABEL: va2:
148; RV32I: # %bb.0:
149; RV32I-NEXT: addi sp, sp, -48
150; RV32I-NEXT: sw ra, 12(sp)
151; RV32I-NEXT: sw s0, 8(sp)
152; RV32I-NEXT: addi s0, sp, 16
153; RV32I-NEXT: sw a7, 28(s0)
154; RV32I-NEXT: sw a6, 24(s0)
155; RV32I-NEXT: sw a5, 20(s0)
156; RV32I-NEXT: sw a4, 16(s0)
157; RV32I-NEXT: sw a3, 12(s0)
158; RV32I-NEXT: sw a2, 8(s0)
159; RV32I-NEXT: sw a1, 4(s0)
160; RV32I-NEXT: addi a0, s0, 19
161; RV32I-NEXT: sw a0, -12(s0)
162; RV32I-NEXT: addi a0, s0, 11
163; RV32I-NEXT: andi a1, a0, -8
164; RV32I-NEXT: lw a0, 0(a1)
165; RV32I-NEXT: ori a1, a1, 4
166; RV32I-NEXT: lw a1, 0(a1)
167; RV32I-NEXT: lw s0, 8(sp)
168; RV32I-NEXT: lw ra, 12(sp)
169; RV32I-NEXT: addi sp, sp, 48
170; RV32I-NEXT: ret
171 %va = alloca i8*, align 4
172 %1 = bitcast i8** %va to i8*
173 call void @llvm.va_start(i8* %1)
174 %2 = bitcast i8** %va to i32*
175 %argp.cur = load i32, i32* %2, align 4
176 %3 = add i32 %argp.cur, 7
177 %4 = and i32 %3, -8
178 %argp.cur.aligned = inttoptr i32 %3 to i8*
179 %argp.next = getelementptr inbounds i8, i8* %argp.cur.aligned, i32 8
180 store i8* %argp.next, i8** %va, align 4
181 %5 = inttoptr i32 %4 to double*
182 %6 = load double, double* %5, align 8
183 call void @llvm.va_end(i8* %1)
184 ret double %6
185}
186
187define double @va2_va_arg(i8 *%fmt, ...) nounwind {
188; RV32I-LABEL: va2_va_arg:
189; RV32I: # %bb.0:
190; RV32I-NEXT: addi sp, sp, -48
191; RV32I-NEXT: sw ra, 12(sp)
192; RV32I-NEXT: sw s0, 8(sp)
193; RV32I-NEXT: addi s0, sp, 16
194; RV32I-NEXT: sw a7, 28(s0)
195; RV32I-NEXT: sw a6, 24(s0)
196; RV32I-NEXT: sw a5, 20(s0)
197; RV32I-NEXT: sw a4, 16(s0)
198; RV32I-NEXT: sw a3, 12(s0)
199; RV32I-NEXT: sw a2, 8(s0)
200; RV32I-NEXT: sw a1, 4(s0)
201; RV32I-NEXT: addi a0, s0, 11
202; RV32I-NEXT: andi a0, a0, -8
203; RV32I-NEXT: ori a1, a0, 4
204; RV32I-NEXT: sw a1, -12(s0)
205; RV32I-NEXT: lw a0, 0(a0)
206; RV32I-NEXT: addi a2, a1, 4
207; RV32I-NEXT: sw a2, -12(s0)
208; RV32I-NEXT: lw a1, 0(a1)
209; RV32I-NEXT: lw s0, 8(sp)
210; RV32I-NEXT: lw ra, 12(sp)
211; RV32I-NEXT: addi sp, sp, 48
212; RV32I-NEXT: ret
213 %va = alloca i8*, align 4
214 %1 = bitcast i8** %va to i8*
215 call void @llvm.va_start(i8* %1)
216 %2 = va_arg i8** %va, double
217 call void @llvm.va_end(i8* %1)
218 ret double %2
219}
220
221define void @va2_caller() nounwind {
222; RV32I-LABEL: va2_caller:
223; RV32I: # %bb.0:
224; RV32I-NEXT: addi sp, sp, -16
225; RV32I-NEXT: sw ra, 12(sp)
226; RV32I-NEXT: sw s0, 8(sp)
227; RV32I-NEXT: addi s0, sp, 16
228; RV32I-NEXT: lui a0, 261888
229; RV32I-NEXT: mv a3, a0
230; RV32I-NEXT: lui a0, %hi(va2)
231; RV32I-NEXT: addi a0, a0, %lo(va2)
232; RV32I-NEXT: mv a2, zero
233; RV32I-NEXT: jalr a0
234; RV32I-NEXT: lw s0, 8(sp)
235; RV32I-NEXT: lw ra, 12(sp)
236; RV32I-NEXT: addi sp, sp, 16
237; RV32I-NEXT: ret
238 %1 = call double (i8*, ...) @va2(i8* undef, double 1.000000e+00)
239 ret void
240}
241
242; Ensure a named double argument is passed in a1 and a2, while the vararg
243; double is passed in a4 and a5 (rather than a3 and a4)
244
245define double @va3(i32 %a, double %b, ...) nounwind {
246; RV32I-LABEL: va3:
247; RV32I: # %bb.0:
248; RV32I-NEXT: addi sp, sp, -48
249; RV32I-NEXT: sw ra, 20(sp)
250; RV32I-NEXT: sw s0, 16(sp)
251; RV32I-NEXT: addi s0, sp, 24
252; RV32I-NEXT: sw a7, 20(s0)
253; RV32I-NEXT: sw a6, 16(s0)
254; RV32I-NEXT: sw a5, 12(s0)
255; RV32I-NEXT: sw a4, 8(s0)
256; RV32I-NEXT: sw a3, 4(s0)
257; RV32I-NEXT: addi a0, s0, 19
258; RV32I-NEXT: sw a0, -12(s0)
259; RV32I-NEXT: lui a0, %hi(__adddf3)
260; RV32I-NEXT: addi a5, a0, %lo(__adddf3)
261; RV32I-NEXT: addi a0, s0, 11
262; RV32I-NEXT: andi a0, a0, -8
263; RV32I-NEXT: lw a4, 0(a0)
264; RV32I-NEXT: ori a0, a0, 4
265; RV32I-NEXT: lw a3, 0(a0)
266; RV32I-NEXT: mv a0, a1
267; RV32I-NEXT: mv a1, a2
268; RV32I-NEXT: mv a2, a4
269; RV32I-NEXT: jalr a5
270; RV32I-NEXT: lw s0, 16(sp)
271; RV32I-NEXT: lw ra, 20(sp)
272; RV32I-NEXT: addi sp, sp, 48
273; RV32I-NEXT: ret
274 %va = alloca i8*, align 4
275 %1 = bitcast i8** %va to i8*
276 call void @llvm.va_start(i8* %1)
277 %2 = bitcast i8** %va to i32*
278 %argp.cur = load i32, i32* %2, align 4
279 %3 = add i32 %argp.cur, 7
280 %4 = and i32 %3, -8
281 %argp.cur.aligned = inttoptr i32 %3 to i8*
282 %argp.next = getelementptr inbounds i8, i8* %argp.cur.aligned, i32 8
283 store i8* %argp.next, i8** %va, align 4
284 %5 = inttoptr i32 %4 to double*
285 %6 = load double, double* %5, align 8
286 call void @llvm.va_end(i8* %1)
287 %7 = fadd double %b, %6
288 ret double %7
289}
290
291define double @va3_va_arg(i32 %a, double %b, ...) nounwind {
292; RV32I-LABEL: va3_va_arg:
293; RV32I: # %bb.0:
294; RV32I-NEXT: addi sp, sp, -48
295; RV32I-NEXT: sw ra, 20(sp)
296; RV32I-NEXT: sw s0, 16(sp)
297; RV32I-NEXT: addi s0, sp, 24
298; RV32I-NEXT: sw a7, 20(s0)
299; RV32I-NEXT: sw a6, 16(s0)
300; RV32I-NEXT: sw a5, 12(s0)
301; RV32I-NEXT: sw a4, 8(s0)
302; RV32I-NEXT: sw a3, 4(s0)
303; RV32I-NEXT: addi a0, s0, 11
304; RV32I-NEXT: andi a0, a0, -8
305; RV32I-NEXT: ori a3, a0, 4
306; RV32I-NEXT: sw a3, -12(s0)
307; RV32I-NEXT: lw a4, 0(a0)
308; RV32I-NEXT: addi a0, a3, 4
309; RV32I-NEXT: sw a0, -12(s0)
310; RV32I-NEXT: lui a0, %hi(__adddf3)
311; RV32I-NEXT: addi a5, a0, %lo(__adddf3)
312; RV32I-NEXT: lw a3, 0(a3)
313; RV32I-NEXT: mv a0, a1
314; RV32I-NEXT: mv a1, a2
315; RV32I-NEXT: mv a2, a4
316; RV32I-NEXT: jalr a5
317; RV32I-NEXT: lw s0, 16(sp)
318; RV32I-NEXT: lw ra, 20(sp)
319; RV32I-NEXT: addi sp, sp, 48
320; RV32I-NEXT: ret
321 %va = alloca i8*, align 4
322 %1 = bitcast i8** %va to i8*
323 call void @llvm.va_start(i8* %1)
324 %2 = va_arg i8** %va, double
325 call void @llvm.va_end(i8* %1)
326 %3 = fadd double %b, %2
327 ret double %3
328}
329
330define void @va3_caller() nounwind {
331; RV32I-LABEL: va3_caller:
332; RV32I: # %bb.0:
333; RV32I-NEXT: addi sp, sp, -16
334; RV32I-NEXT: sw ra, 12(sp)
335; RV32I-NEXT: sw s0, 8(sp)
336; RV32I-NEXT: addi s0, sp, 16
337; RV32I-NEXT: lui a0, 261888
338; RV32I-NEXT: mv a2, a0
339; RV32I-NEXT: lui a0, 262144
340; RV32I-NEXT: mv a5, a0
341; RV32I-NEXT: lui a0, %hi(va3)
342; RV32I-NEXT: addi a3, a0, %lo(va3)
343; RV32I-NEXT: addi a0, zero, 2
344; RV32I-NEXT: mv a1, zero
345; RV32I-NEXT: mv a4, zero
346; RV32I-NEXT: jalr a3
347; RV32I-NEXT: lw s0, 8(sp)
348; RV32I-NEXT: lw ra, 12(sp)
349; RV32I-NEXT: addi sp, sp, 16
350; RV32I-NEXT: ret
351 %1 = call double (i32, double, ...) @va3(i32 2, double 1.000000e+00, double 2.000000e+00)
352 ret void
353}
354
355declare void @llvm.va_copy(i8*, i8*)
356
357define i32 @va4_va_copy(i32 %argno, ...) nounwind {
358; RV32I-LABEL: va4_va_copy:
359; RV32I: # %bb.0:
360; RV32I-NEXT: addi sp, sp, -64
361; RV32I-NEXT: sw ra, 28(sp)
362; RV32I-NEXT: sw s0, 24(sp)
363; RV32I-NEXT: sw s1, 20(sp)
364; RV32I-NEXT: addi s0, sp, 32
365; RV32I-NEXT: sw a1, 4(s0)
366; RV32I-NEXT: sw a7, 28(s0)
367; RV32I-NEXT: sw a6, 24(s0)
368; RV32I-NEXT: sw a5, 20(s0)
369; RV32I-NEXT: sw a4, 16(s0)
370; RV32I-NEXT: sw a3, 12(s0)
371; RV32I-NEXT: sw a2, 8(s0)
372; RV32I-NEXT: addi a0, s0, 8
373; RV32I-NEXT: sw a0, -16(s0)
374; RV32I-NEXT: sw a0, -20(s0)
375; RV32I-NEXT: lw s1, 4(s0)
376; RV32I-NEXT: lui a1, %hi(notdead)
377; RV32I-NEXT: addi a1, a1, %lo(notdead)
378; RV32I-NEXT: jalr a1
379; RV32I-NEXT: lw a0, -16(s0)
380; RV32I-NEXT: addi a0, a0, 3
381; RV32I-NEXT: andi a0, a0, -4
382; RV32I-NEXT: addi a1, a0, 4
383; RV32I-NEXT: sw a1, -16(s0)
384; RV32I-NEXT: lw a1, 0(a0)
385; RV32I-NEXT: addi a0, a0, 7
386; RV32I-NEXT: andi a0, a0, -4
387; RV32I-NEXT: addi a2, a0, 4
388; RV32I-NEXT: sw a2, -16(s0)
389; RV32I-NEXT: lw a2, 0(a0)
390; RV32I-NEXT: addi a0, a0, 7
391; RV32I-NEXT: andi a0, a0, -4
392; RV32I-NEXT: addi a3, a0, 4
393; RV32I-NEXT: sw a3, -16(s0)
394; RV32I-NEXT: add a1, a1, s1
395; RV32I-NEXT: add a1, a1, a2
396; RV32I-NEXT: lw a0, 0(a0)
397; RV32I-NEXT: add a0, a1, a0
398; RV32I-NEXT: lw s1, 20(sp)
399; RV32I-NEXT: lw s0, 24(sp)
400; RV32I-NEXT: lw ra, 28(sp)
401; RV32I-NEXT: addi sp, sp, 64
402; RV32I-NEXT: ret
403 %vargs = alloca i8*, align 4
404 %wargs = alloca i8*, align 4
405 %1 = bitcast i8** %vargs to i8*
406 %2 = bitcast i8** %wargs to i8*
407 call void @llvm.va_start(i8* %1)
408 %3 = va_arg i8** %vargs, i32
409 call void @llvm.va_copy(i8* %2, i8* %1)
410 %4 = load i8*, i8** %wargs, align 4
411 call void @notdead(i8* %4)
412 %5 = va_arg i8** %vargs, i32
413 %6 = va_arg i8** %vargs, i32
414 %7 = va_arg i8** %vargs, i32
415 call void @llvm.va_end(i8* %1)
416 call void @llvm.va_end(i8* %2)
417 %add1 = add i32 %5, %3
418 %add2 = add i32 %add1, %6
419 %add3 = add i32 %add2, %7
420 ret i32 %add3
421}
422
423; Check 2x*xlen values are aligned appropriately when passed on the stack in a vararg call
424
425define i32 @va5_aligned_stack_callee(i32 %a, ...) nounwind {
426; RV32I-LABEL: va5_aligned_stack_callee:
427; RV32I: # %bb.0:
428; RV32I-NEXT: addi sp, sp, -48
429; RV32I-NEXT: sw ra, 12(sp)
430; RV32I-NEXT: sw s0, 8(sp)
431; RV32I-NEXT: addi s0, sp, 16
432; RV32I-NEXT: sw a7, 28(s0)
433; RV32I-NEXT: sw a6, 24(s0)
434; RV32I-NEXT: sw a5, 20(s0)
435; RV32I-NEXT: sw a4, 16(s0)
436; RV32I-NEXT: sw a3, 12(s0)
437; RV32I-NEXT: sw a2, 8(s0)
438; RV32I-NEXT: sw a1, 4(s0)
439; RV32I-NEXT: addi a0, zero, 1
440; RV32I-NEXT: lw s0, 8(sp)
441; RV32I-NEXT: lw ra, 12(sp)
442; RV32I-NEXT: addi sp, sp, 48
443; RV32I-NEXT: ret
444 ret i32 1
445}
446
447define void @va5_aligned_stack_caller() nounwind {
448; The double should be 8-byte aligned on the stack, but the two-element array
449; should only be 4-byte aligned
450; RV32I-LABEL: va5_aligned_stack_caller:
451; RV32I: # %bb.0:
452; RV32I-NEXT: addi sp, sp, -64
453; RV32I-NEXT: sw ra, 60(sp)
454; RV32I-NEXT: sw s0, 56(sp)
455; RV32I-NEXT: addi s0, sp, 64
456; RV32I-NEXT: addi a0, zero, 17
457; RV32I-NEXT: sw a0, 24(sp)
458; RV32I-NEXT: addi a0, zero, 16
459; RV32I-NEXT: sw a0, 20(sp)
460; RV32I-NEXT: addi a0, zero, 15
461; RV32I-NEXT: sw a0, 16(sp)
462; RV32I-NEXT: lui a0, 262236
463; RV32I-NEXT: addi a0, a0, 655
464; RV32I-NEXT: sw a0, 12(sp)
465; RV32I-NEXT: lui a0, 377487
466; RV32I-NEXT: addi a0, a0, 1475
467; RV32I-NEXT: sw a0, 8(sp)
468; RV32I-NEXT: addi a0, zero, 14
469; RV32I-NEXT: sw a0, 0(sp)
470; RV32I-NEXT: lui a0, 262153
471; RV32I-NEXT: addi a0, a0, 491
472; RV32I-NEXT: sw a0, -20(s0)
473; RV32I-NEXT: lui a0, 545260
474; RV32I-NEXT: addi a0, a0, -1967
475; RV32I-NEXT: sw a0, -24(s0)
476; RV32I-NEXT: lui a0, 964690
477; RV32I-NEXT: addi a0, a0, -328
478; RV32I-NEXT: sw a0, -28(s0)
479; RV32I-NEXT: lui a0, 335544
480; RV32I-NEXT: addi a0, a0, 1311
481; RV32I-NEXT: sw a0, -32(s0)
482; RV32I-NEXT: lui a0, 688509
483; RV32I-NEXT: addi a6, a0, -2048
484; RV32I-NEXT: lui a0, %hi(va5_aligned_stack_callee)
485; RV32I-NEXT: addi a5, a0, %lo(va5_aligned_stack_callee)
486; RV32I-NEXT: addi a0, zero, 1
487; RV32I-NEXT: addi a1, zero, 11
488; RV32I-NEXT: addi a2, s0, -32
489; RV32I-NEXT: addi a3, zero, 12
490; RV32I-NEXT: addi a4, zero, 13
491; RV32I-NEXT: addi a7, zero, 4
492; RV32I-NEXT: jalr a5
493; RV32I-NEXT: lw s0, 56(sp)
494; RV32I-NEXT: lw ra, 60(sp)
495; RV32I-NEXT: addi sp, sp, 64
496; RV32I-NEXT: ret
497 %1 = call i32 (i32, ...) @va5_aligned_stack_callee(i32 1, i32 11,
498 fp128 0xLEB851EB851EB851F400091EB851EB851, i32 12, i32 13, i64 20000000000,
499 i32 14, double 2.720000e+00, i32 15, [2 x i32] [i32 16, i32 17])
500 ret void
501}
502
503; A function with no fixed arguments is not valid C, but can be
504; specified in LLVM IR. We must ensure the vararg save area is
505; still set up correctly.
506
507define i32 @va6_no_fixed_args(...) nounwind {
508; RV32I-LABEL: va6_no_fixed_args:
509; RV32I: # %bb.0:
510; RV32I-NEXT: addi sp, sp, -48
511; RV32I-NEXT: sw ra, 12(sp)
512; RV32I-NEXT: sw s0, 8(sp)
513; RV32I-NEXT: addi s0, sp, 16
514; RV32I-NEXT: sw a0, 0(s0)
515; RV32I-NEXT: sw a7, 28(s0)
516; RV32I-NEXT: sw a6, 24(s0)
517; RV32I-NEXT: sw a5, 20(s0)
518; RV32I-NEXT: sw a4, 16(s0)
519; RV32I-NEXT: sw a3, 12(s0)
520; RV32I-NEXT: sw a2, 8(s0)
521; RV32I-NEXT: sw a1, 4(s0)
522; RV32I-NEXT: addi a0, s0, 4
523; RV32I-NEXT: sw a0, -12(s0)
524; RV32I-NEXT: lw a0, 0(s0)
525; RV32I-NEXT: lw s0, 8(sp)
526; RV32I-NEXT: lw ra, 12(sp)
527; RV32I-NEXT: addi sp, sp, 48
528; RV32I-NEXT: ret
529 %va = alloca i8*, align 4
530 %1 = bitcast i8** %va to i8*
531 call void @llvm.va_start(i8* %1)
532 %2 = va_arg i8** %va, i32
533 call void @llvm.va_end(i8* %1)
534 ret i32 %2
535}