Alex Bradbury | c85be0d | 2018-01-10 19:41:03 +0000 | [diff] [blame^] | 1 | ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py |
| 2 | ; RUN: llc -mtriple=riscv32 -verify-machineinstrs < %s \ |
| 3 | ; RUN: | FileCheck -check-prefix=RV32I %s |
| 4 | |
| 5 | declare void @llvm.va_start(i8*) |
| 6 | declare void @llvm.va_end(i8*) |
| 7 | |
| 8 | declare void @notdead(i8*) |
| 9 | |
| 10 | ; Although frontends are recommended to not generate va_arg due to the lack of |
| 11 | ; support for aggregate types, we test simple cases here to ensure they are |
| 12 | ; lowered correctly |
| 13 | |
| 14 | define i32 @va1(i8* %fmt, ...) nounwind { |
| 15 | ; RV32I-LABEL: va1: |
| 16 | ; RV32I: # %bb.0: |
| 17 | ; RV32I-NEXT: addi sp, sp, -48 |
| 18 | ; RV32I-NEXT: sw ra, 12(sp) |
| 19 | ; RV32I-NEXT: sw s0, 8(sp) |
| 20 | ; RV32I-NEXT: addi s0, sp, 16 |
| 21 | ; RV32I-NEXT: sw a1, 4(s0) |
| 22 | ; RV32I-NEXT: sw a7, 28(s0) |
| 23 | ; RV32I-NEXT: sw a6, 24(s0) |
| 24 | ; RV32I-NEXT: sw a5, 20(s0) |
| 25 | ; RV32I-NEXT: sw a4, 16(s0) |
| 26 | ; RV32I-NEXT: sw a3, 12(s0) |
| 27 | ; RV32I-NEXT: sw a2, 8(s0) |
| 28 | ; RV32I-NEXT: addi a0, s0, 8 |
| 29 | ; RV32I-NEXT: sw a0, -12(s0) |
| 30 | ; RV32I-NEXT: lw a0, 4(s0) |
| 31 | ; RV32I-NEXT: lw s0, 8(sp) |
| 32 | ; RV32I-NEXT: lw ra, 12(sp) |
| 33 | ; RV32I-NEXT: addi sp, sp, 48 |
| 34 | ; RV32I-NEXT: ret |
| 35 | %va = alloca i8*, align 4 |
| 36 | %1 = bitcast i8** %va to i8* |
| 37 | call void @llvm.va_start(i8* %1) |
| 38 | %argp.cur = load i8*, i8** %va, align 4 |
| 39 | %argp.next = getelementptr inbounds i8, i8* %argp.cur, i32 4 |
| 40 | store i8* %argp.next, i8** %va, align 4 |
| 41 | %2 = bitcast i8* %argp.cur to i32* |
| 42 | %3 = load i32, i32* %2, align 4 |
| 43 | call void @llvm.va_end(i8* %1) |
| 44 | ret i32 %3 |
| 45 | } |
| 46 | |
| 47 | define i32 @va1_va_arg(i8* %fmt, ...) nounwind { |
| 48 | ; RV32I-LABEL: va1_va_arg: |
| 49 | ; RV32I: # %bb.0: |
| 50 | ; RV32I-NEXT: addi sp, sp, -48 |
| 51 | ; RV32I-NEXT: sw ra, 12(sp) |
| 52 | ; RV32I-NEXT: sw s0, 8(sp) |
| 53 | ; RV32I-NEXT: addi s0, sp, 16 |
| 54 | ; RV32I-NEXT: sw a1, 4(s0) |
| 55 | ; RV32I-NEXT: sw a7, 28(s0) |
| 56 | ; RV32I-NEXT: sw a6, 24(s0) |
| 57 | ; RV32I-NEXT: sw a5, 20(s0) |
| 58 | ; RV32I-NEXT: sw a4, 16(s0) |
| 59 | ; RV32I-NEXT: sw a3, 12(s0) |
| 60 | ; RV32I-NEXT: sw a2, 8(s0) |
| 61 | ; RV32I-NEXT: addi a0, s0, 8 |
| 62 | ; RV32I-NEXT: sw a0, -12(s0) |
| 63 | ; RV32I-NEXT: lw a0, 4(s0) |
| 64 | ; RV32I-NEXT: lw s0, 8(sp) |
| 65 | ; RV32I-NEXT: lw ra, 12(sp) |
| 66 | ; RV32I-NEXT: addi sp, sp, 48 |
| 67 | ; RV32I-NEXT: ret |
| 68 | %va = alloca i8*, align 4 |
| 69 | %1 = bitcast i8** %va to i8* |
| 70 | call void @llvm.va_start(i8* %1) |
| 71 | %2 = va_arg i8** %va, i32 |
| 72 | call void @llvm.va_end(i8* %1) |
| 73 | ret i32 %2 |
| 74 | } |
| 75 | |
| 76 | ; Ensure the adjustment when restoring the stack pointer using the frame |
| 77 | ; pointer is correct |
| 78 | define i32 @va1_va_arg_alloca(i8* %fmt, ...) nounwind { |
| 79 | ; RV32I-LABEL: va1_va_arg_alloca: |
| 80 | ; RV32I: # %bb.0: |
| 81 | ; RV32I-NEXT: addi sp, sp, -48 |
| 82 | ; RV32I-NEXT: sw ra, 12(sp) |
| 83 | ; RV32I-NEXT: sw s0, 8(sp) |
| 84 | ; RV32I-NEXT: sw s1, 4(sp) |
| 85 | ; RV32I-NEXT: addi s0, sp, 16 |
| 86 | ; RV32I-NEXT: sw a1, 4(s0) |
| 87 | ; RV32I-NEXT: sw a7, 28(s0) |
| 88 | ; RV32I-NEXT: sw a6, 24(s0) |
| 89 | ; RV32I-NEXT: sw a5, 20(s0) |
| 90 | ; RV32I-NEXT: sw a4, 16(s0) |
| 91 | ; RV32I-NEXT: sw a3, 12(s0) |
| 92 | ; RV32I-NEXT: sw a2, 8(s0) |
| 93 | ; RV32I-NEXT: addi a0, s0, 8 |
| 94 | ; RV32I-NEXT: sw a0, -16(s0) |
| 95 | ; RV32I-NEXT: lw s1, 4(s0) |
| 96 | ; RV32I-NEXT: addi a0, s1, 15 |
| 97 | ; RV32I-NEXT: andi a0, a0, -16 |
| 98 | ; RV32I-NEXT: sub a0, sp, a0 |
| 99 | ; RV32I-NEXT: mv sp, a0 |
| 100 | ; RV32I-NEXT: lui a1, %hi(notdead) |
| 101 | ; RV32I-NEXT: addi a1, a1, %lo(notdead) |
| 102 | ; RV32I-NEXT: jalr a1 |
| 103 | ; RV32I-NEXT: mv a0, s1 |
| 104 | ; RV32I-NEXT: addi sp, s0, -16 |
| 105 | ; RV32I-NEXT: lw s1, 4(sp) |
| 106 | ; RV32I-NEXT: lw s0, 8(sp) |
| 107 | ; RV32I-NEXT: lw ra, 12(sp) |
| 108 | ; RV32I-NEXT: addi sp, sp, 48 |
| 109 | ; RV32I-NEXT: ret |
| 110 | %va = alloca i8*, align 4 |
| 111 | %1 = bitcast i8** %va to i8* |
| 112 | call void @llvm.va_start(i8* %1) |
| 113 | %2 = va_arg i8** %va, i32 |
| 114 | %3 = alloca i8, i32 %2 |
| 115 | call void @notdead(i8* %3) |
| 116 | call void @llvm.va_end(i8* %1) |
| 117 | ret i32 %2 |
| 118 | } |
| 119 | |
| 120 | define void @va1_caller() nounwind { |
| 121 | ; RV32I-LABEL: va1_caller: |
| 122 | ; RV32I: # %bb.0: |
| 123 | ; RV32I-NEXT: addi sp, sp, -16 |
| 124 | ; RV32I-NEXT: sw ra, 12(sp) |
| 125 | ; RV32I-NEXT: sw s0, 8(sp) |
| 126 | ; RV32I-NEXT: addi s0, sp, 16 |
| 127 | ; RV32I-NEXT: lui a0, 261888 |
| 128 | ; RV32I-NEXT: mv a3, a0 |
| 129 | ; RV32I-NEXT: lui a0, %hi(va1) |
| 130 | ; RV32I-NEXT: addi a0, a0, %lo(va1) |
| 131 | ; RV32I-NEXT: addi a4, zero, 2 |
| 132 | ; RV32I-NEXT: mv a2, zero |
| 133 | ; RV32I-NEXT: jalr a0 |
| 134 | ; RV32I-NEXT: lw s0, 8(sp) |
| 135 | ; RV32I-NEXT: lw ra, 12(sp) |
| 136 | ; RV32I-NEXT: addi sp, sp, 16 |
| 137 | ; RV32I-NEXT: ret |
| 138 | ; Pass a double, as a float would be promoted by a C/C++ frontend |
| 139 | %1 = call i32 (i8*, ...) @va1(i8* undef, double 1.0, i32 2) |
| 140 | ret void |
| 141 | } |
| 142 | |
| 143 | ; Ensure that 2x xlen size+alignment varargs are accessed via an "aligned" |
| 144 | ; register pair (where the first register is even-numbered). |
| 145 | |
| 146 | define double @va2(i8 *%fmt, ...) nounwind { |
| 147 | ; RV32I-LABEL: va2: |
| 148 | ; RV32I: # %bb.0: |
| 149 | ; RV32I-NEXT: addi sp, sp, -48 |
| 150 | ; RV32I-NEXT: sw ra, 12(sp) |
| 151 | ; RV32I-NEXT: sw s0, 8(sp) |
| 152 | ; RV32I-NEXT: addi s0, sp, 16 |
| 153 | ; RV32I-NEXT: sw a7, 28(s0) |
| 154 | ; RV32I-NEXT: sw a6, 24(s0) |
| 155 | ; RV32I-NEXT: sw a5, 20(s0) |
| 156 | ; RV32I-NEXT: sw a4, 16(s0) |
| 157 | ; RV32I-NEXT: sw a3, 12(s0) |
| 158 | ; RV32I-NEXT: sw a2, 8(s0) |
| 159 | ; RV32I-NEXT: sw a1, 4(s0) |
| 160 | ; RV32I-NEXT: addi a0, s0, 19 |
| 161 | ; RV32I-NEXT: sw a0, -12(s0) |
| 162 | ; RV32I-NEXT: addi a0, s0, 11 |
| 163 | ; RV32I-NEXT: andi a1, a0, -8 |
| 164 | ; RV32I-NEXT: lw a0, 0(a1) |
| 165 | ; RV32I-NEXT: ori a1, a1, 4 |
| 166 | ; RV32I-NEXT: lw a1, 0(a1) |
| 167 | ; RV32I-NEXT: lw s0, 8(sp) |
| 168 | ; RV32I-NEXT: lw ra, 12(sp) |
| 169 | ; RV32I-NEXT: addi sp, sp, 48 |
| 170 | ; RV32I-NEXT: ret |
| 171 | %va = alloca i8*, align 4 |
| 172 | %1 = bitcast i8** %va to i8* |
| 173 | call void @llvm.va_start(i8* %1) |
| 174 | %2 = bitcast i8** %va to i32* |
| 175 | %argp.cur = load i32, i32* %2, align 4 |
| 176 | %3 = add i32 %argp.cur, 7 |
| 177 | %4 = and i32 %3, -8 |
| 178 | %argp.cur.aligned = inttoptr i32 %3 to i8* |
| 179 | %argp.next = getelementptr inbounds i8, i8* %argp.cur.aligned, i32 8 |
| 180 | store i8* %argp.next, i8** %va, align 4 |
| 181 | %5 = inttoptr i32 %4 to double* |
| 182 | %6 = load double, double* %5, align 8 |
| 183 | call void @llvm.va_end(i8* %1) |
| 184 | ret double %6 |
| 185 | } |
| 186 | |
| 187 | define double @va2_va_arg(i8 *%fmt, ...) nounwind { |
| 188 | ; RV32I-LABEL: va2_va_arg: |
| 189 | ; RV32I: # %bb.0: |
| 190 | ; RV32I-NEXT: addi sp, sp, -48 |
| 191 | ; RV32I-NEXT: sw ra, 12(sp) |
| 192 | ; RV32I-NEXT: sw s0, 8(sp) |
| 193 | ; RV32I-NEXT: addi s0, sp, 16 |
| 194 | ; RV32I-NEXT: sw a7, 28(s0) |
| 195 | ; RV32I-NEXT: sw a6, 24(s0) |
| 196 | ; RV32I-NEXT: sw a5, 20(s0) |
| 197 | ; RV32I-NEXT: sw a4, 16(s0) |
| 198 | ; RV32I-NEXT: sw a3, 12(s0) |
| 199 | ; RV32I-NEXT: sw a2, 8(s0) |
| 200 | ; RV32I-NEXT: sw a1, 4(s0) |
| 201 | ; RV32I-NEXT: addi a0, s0, 11 |
| 202 | ; RV32I-NEXT: andi a0, a0, -8 |
| 203 | ; RV32I-NEXT: ori a1, a0, 4 |
| 204 | ; RV32I-NEXT: sw a1, -12(s0) |
| 205 | ; RV32I-NEXT: lw a0, 0(a0) |
| 206 | ; RV32I-NEXT: addi a2, a1, 4 |
| 207 | ; RV32I-NEXT: sw a2, -12(s0) |
| 208 | ; RV32I-NEXT: lw a1, 0(a1) |
| 209 | ; RV32I-NEXT: lw s0, 8(sp) |
| 210 | ; RV32I-NEXT: lw ra, 12(sp) |
| 211 | ; RV32I-NEXT: addi sp, sp, 48 |
| 212 | ; RV32I-NEXT: ret |
| 213 | %va = alloca i8*, align 4 |
| 214 | %1 = bitcast i8** %va to i8* |
| 215 | call void @llvm.va_start(i8* %1) |
| 216 | %2 = va_arg i8** %va, double |
| 217 | call void @llvm.va_end(i8* %1) |
| 218 | ret double %2 |
| 219 | } |
| 220 | |
| 221 | define void @va2_caller() nounwind { |
| 222 | ; RV32I-LABEL: va2_caller: |
| 223 | ; RV32I: # %bb.0: |
| 224 | ; RV32I-NEXT: addi sp, sp, -16 |
| 225 | ; RV32I-NEXT: sw ra, 12(sp) |
| 226 | ; RV32I-NEXT: sw s0, 8(sp) |
| 227 | ; RV32I-NEXT: addi s0, sp, 16 |
| 228 | ; RV32I-NEXT: lui a0, 261888 |
| 229 | ; RV32I-NEXT: mv a3, a0 |
| 230 | ; RV32I-NEXT: lui a0, %hi(va2) |
| 231 | ; RV32I-NEXT: addi a0, a0, %lo(va2) |
| 232 | ; RV32I-NEXT: mv a2, zero |
| 233 | ; RV32I-NEXT: jalr a0 |
| 234 | ; RV32I-NEXT: lw s0, 8(sp) |
| 235 | ; RV32I-NEXT: lw ra, 12(sp) |
| 236 | ; RV32I-NEXT: addi sp, sp, 16 |
| 237 | ; RV32I-NEXT: ret |
| 238 | %1 = call double (i8*, ...) @va2(i8* undef, double 1.000000e+00) |
| 239 | ret void |
| 240 | } |
| 241 | |
| 242 | ; Ensure a named double argument is passed in a1 and a2, while the vararg |
| 243 | ; double is passed in a4 and a5 (rather than a3 and a4) |
| 244 | |
| 245 | define double @va3(i32 %a, double %b, ...) nounwind { |
| 246 | ; RV32I-LABEL: va3: |
| 247 | ; RV32I: # %bb.0: |
| 248 | ; RV32I-NEXT: addi sp, sp, -48 |
| 249 | ; RV32I-NEXT: sw ra, 20(sp) |
| 250 | ; RV32I-NEXT: sw s0, 16(sp) |
| 251 | ; RV32I-NEXT: addi s0, sp, 24 |
| 252 | ; RV32I-NEXT: sw a7, 20(s0) |
| 253 | ; RV32I-NEXT: sw a6, 16(s0) |
| 254 | ; RV32I-NEXT: sw a5, 12(s0) |
| 255 | ; RV32I-NEXT: sw a4, 8(s0) |
| 256 | ; RV32I-NEXT: sw a3, 4(s0) |
| 257 | ; RV32I-NEXT: addi a0, s0, 19 |
| 258 | ; RV32I-NEXT: sw a0, -12(s0) |
| 259 | ; RV32I-NEXT: lui a0, %hi(__adddf3) |
| 260 | ; RV32I-NEXT: addi a5, a0, %lo(__adddf3) |
| 261 | ; RV32I-NEXT: addi a0, s0, 11 |
| 262 | ; RV32I-NEXT: andi a0, a0, -8 |
| 263 | ; RV32I-NEXT: lw a4, 0(a0) |
| 264 | ; RV32I-NEXT: ori a0, a0, 4 |
| 265 | ; RV32I-NEXT: lw a3, 0(a0) |
| 266 | ; RV32I-NEXT: mv a0, a1 |
| 267 | ; RV32I-NEXT: mv a1, a2 |
| 268 | ; RV32I-NEXT: mv a2, a4 |
| 269 | ; RV32I-NEXT: jalr a5 |
| 270 | ; RV32I-NEXT: lw s0, 16(sp) |
| 271 | ; RV32I-NEXT: lw ra, 20(sp) |
| 272 | ; RV32I-NEXT: addi sp, sp, 48 |
| 273 | ; RV32I-NEXT: ret |
| 274 | %va = alloca i8*, align 4 |
| 275 | %1 = bitcast i8** %va to i8* |
| 276 | call void @llvm.va_start(i8* %1) |
| 277 | %2 = bitcast i8** %va to i32* |
| 278 | %argp.cur = load i32, i32* %2, align 4 |
| 279 | %3 = add i32 %argp.cur, 7 |
| 280 | %4 = and i32 %3, -8 |
| 281 | %argp.cur.aligned = inttoptr i32 %3 to i8* |
| 282 | %argp.next = getelementptr inbounds i8, i8* %argp.cur.aligned, i32 8 |
| 283 | store i8* %argp.next, i8** %va, align 4 |
| 284 | %5 = inttoptr i32 %4 to double* |
| 285 | %6 = load double, double* %5, align 8 |
| 286 | call void @llvm.va_end(i8* %1) |
| 287 | %7 = fadd double %b, %6 |
| 288 | ret double %7 |
| 289 | } |
| 290 | |
| 291 | define double @va3_va_arg(i32 %a, double %b, ...) nounwind { |
| 292 | ; RV32I-LABEL: va3_va_arg: |
| 293 | ; RV32I: # %bb.0: |
| 294 | ; RV32I-NEXT: addi sp, sp, -48 |
| 295 | ; RV32I-NEXT: sw ra, 20(sp) |
| 296 | ; RV32I-NEXT: sw s0, 16(sp) |
| 297 | ; RV32I-NEXT: addi s0, sp, 24 |
| 298 | ; RV32I-NEXT: sw a7, 20(s0) |
| 299 | ; RV32I-NEXT: sw a6, 16(s0) |
| 300 | ; RV32I-NEXT: sw a5, 12(s0) |
| 301 | ; RV32I-NEXT: sw a4, 8(s0) |
| 302 | ; RV32I-NEXT: sw a3, 4(s0) |
| 303 | ; RV32I-NEXT: addi a0, s0, 11 |
| 304 | ; RV32I-NEXT: andi a0, a0, -8 |
| 305 | ; RV32I-NEXT: ori a3, a0, 4 |
| 306 | ; RV32I-NEXT: sw a3, -12(s0) |
| 307 | ; RV32I-NEXT: lw a4, 0(a0) |
| 308 | ; RV32I-NEXT: addi a0, a3, 4 |
| 309 | ; RV32I-NEXT: sw a0, -12(s0) |
| 310 | ; RV32I-NEXT: lui a0, %hi(__adddf3) |
| 311 | ; RV32I-NEXT: addi a5, a0, %lo(__adddf3) |
| 312 | ; RV32I-NEXT: lw a3, 0(a3) |
| 313 | ; RV32I-NEXT: mv a0, a1 |
| 314 | ; RV32I-NEXT: mv a1, a2 |
| 315 | ; RV32I-NEXT: mv a2, a4 |
| 316 | ; RV32I-NEXT: jalr a5 |
| 317 | ; RV32I-NEXT: lw s0, 16(sp) |
| 318 | ; RV32I-NEXT: lw ra, 20(sp) |
| 319 | ; RV32I-NEXT: addi sp, sp, 48 |
| 320 | ; RV32I-NEXT: ret |
| 321 | %va = alloca i8*, align 4 |
| 322 | %1 = bitcast i8** %va to i8* |
| 323 | call void @llvm.va_start(i8* %1) |
| 324 | %2 = va_arg i8** %va, double |
| 325 | call void @llvm.va_end(i8* %1) |
| 326 | %3 = fadd double %b, %2 |
| 327 | ret double %3 |
| 328 | } |
| 329 | |
| 330 | define void @va3_caller() nounwind { |
| 331 | ; RV32I-LABEL: va3_caller: |
| 332 | ; RV32I: # %bb.0: |
| 333 | ; RV32I-NEXT: addi sp, sp, -16 |
| 334 | ; RV32I-NEXT: sw ra, 12(sp) |
| 335 | ; RV32I-NEXT: sw s0, 8(sp) |
| 336 | ; RV32I-NEXT: addi s0, sp, 16 |
| 337 | ; RV32I-NEXT: lui a0, 261888 |
| 338 | ; RV32I-NEXT: mv a2, a0 |
| 339 | ; RV32I-NEXT: lui a0, 262144 |
| 340 | ; RV32I-NEXT: mv a5, a0 |
| 341 | ; RV32I-NEXT: lui a0, %hi(va3) |
| 342 | ; RV32I-NEXT: addi a3, a0, %lo(va3) |
| 343 | ; RV32I-NEXT: addi a0, zero, 2 |
| 344 | ; RV32I-NEXT: mv a1, zero |
| 345 | ; RV32I-NEXT: mv a4, zero |
| 346 | ; RV32I-NEXT: jalr a3 |
| 347 | ; RV32I-NEXT: lw s0, 8(sp) |
| 348 | ; RV32I-NEXT: lw ra, 12(sp) |
| 349 | ; RV32I-NEXT: addi sp, sp, 16 |
| 350 | ; RV32I-NEXT: ret |
| 351 | %1 = call double (i32, double, ...) @va3(i32 2, double 1.000000e+00, double 2.000000e+00) |
| 352 | ret void |
| 353 | } |
| 354 | |
| 355 | declare void @llvm.va_copy(i8*, i8*) |
| 356 | |
| 357 | define i32 @va4_va_copy(i32 %argno, ...) nounwind { |
| 358 | ; RV32I-LABEL: va4_va_copy: |
| 359 | ; RV32I: # %bb.0: |
| 360 | ; RV32I-NEXT: addi sp, sp, -64 |
| 361 | ; RV32I-NEXT: sw ra, 28(sp) |
| 362 | ; RV32I-NEXT: sw s0, 24(sp) |
| 363 | ; RV32I-NEXT: sw s1, 20(sp) |
| 364 | ; RV32I-NEXT: addi s0, sp, 32 |
| 365 | ; RV32I-NEXT: sw a1, 4(s0) |
| 366 | ; RV32I-NEXT: sw a7, 28(s0) |
| 367 | ; RV32I-NEXT: sw a6, 24(s0) |
| 368 | ; RV32I-NEXT: sw a5, 20(s0) |
| 369 | ; RV32I-NEXT: sw a4, 16(s0) |
| 370 | ; RV32I-NEXT: sw a3, 12(s0) |
| 371 | ; RV32I-NEXT: sw a2, 8(s0) |
| 372 | ; RV32I-NEXT: addi a0, s0, 8 |
| 373 | ; RV32I-NEXT: sw a0, -16(s0) |
| 374 | ; RV32I-NEXT: sw a0, -20(s0) |
| 375 | ; RV32I-NEXT: lw s1, 4(s0) |
| 376 | ; RV32I-NEXT: lui a1, %hi(notdead) |
| 377 | ; RV32I-NEXT: addi a1, a1, %lo(notdead) |
| 378 | ; RV32I-NEXT: jalr a1 |
| 379 | ; RV32I-NEXT: lw a0, -16(s0) |
| 380 | ; RV32I-NEXT: addi a0, a0, 3 |
| 381 | ; RV32I-NEXT: andi a0, a0, -4 |
| 382 | ; RV32I-NEXT: addi a1, a0, 4 |
| 383 | ; RV32I-NEXT: sw a1, -16(s0) |
| 384 | ; RV32I-NEXT: lw a1, 0(a0) |
| 385 | ; RV32I-NEXT: addi a0, a0, 7 |
| 386 | ; RV32I-NEXT: andi a0, a0, -4 |
| 387 | ; RV32I-NEXT: addi a2, a0, 4 |
| 388 | ; RV32I-NEXT: sw a2, -16(s0) |
| 389 | ; RV32I-NEXT: lw a2, 0(a0) |
| 390 | ; RV32I-NEXT: addi a0, a0, 7 |
| 391 | ; RV32I-NEXT: andi a0, a0, -4 |
| 392 | ; RV32I-NEXT: addi a3, a0, 4 |
| 393 | ; RV32I-NEXT: sw a3, -16(s0) |
| 394 | ; RV32I-NEXT: add a1, a1, s1 |
| 395 | ; RV32I-NEXT: add a1, a1, a2 |
| 396 | ; RV32I-NEXT: lw a0, 0(a0) |
| 397 | ; RV32I-NEXT: add a0, a1, a0 |
| 398 | ; RV32I-NEXT: lw s1, 20(sp) |
| 399 | ; RV32I-NEXT: lw s0, 24(sp) |
| 400 | ; RV32I-NEXT: lw ra, 28(sp) |
| 401 | ; RV32I-NEXT: addi sp, sp, 64 |
| 402 | ; RV32I-NEXT: ret |
| 403 | %vargs = alloca i8*, align 4 |
| 404 | %wargs = alloca i8*, align 4 |
| 405 | %1 = bitcast i8** %vargs to i8* |
| 406 | %2 = bitcast i8** %wargs to i8* |
| 407 | call void @llvm.va_start(i8* %1) |
| 408 | %3 = va_arg i8** %vargs, i32 |
| 409 | call void @llvm.va_copy(i8* %2, i8* %1) |
| 410 | %4 = load i8*, i8** %wargs, align 4 |
| 411 | call void @notdead(i8* %4) |
| 412 | %5 = va_arg i8** %vargs, i32 |
| 413 | %6 = va_arg i8** %vargs, i32 |
| 414 | %7 = va_arg i8** %vargs, i32 |
| 415 | call void @llvm.va_end(i8* %1) |
| 416 | call void @llvm.va_end(i8* %2) |
| 417 | %add1 = add i32 %5, %3 |
| 418 | %add2 = add i32 %add1, %6 |
| 419 | %add3 = add i32 %add2, %7 |
| 420 | ret i32 %add3 |
| 421 | } |
| 422 | |
| 423 | ; Check 2x*xlen values are aligned appropriately when passed on the stack in a vararg call |
| 424 | |
| 425 | define i32 @va5_aligned_stack_callee(i32 %a, ...) nounwind { |
| 426 | ; RV32I-LABEL: va5_aligned_stack_callee: |
| 427 | ; RV32I: # %bb.0: |
| 428 | ; RV32I-NEXT: addi sp, sp, -48 |
| 429 | ; RV32I-NEXT: sw ra, 12(sp) |
| 430 | ; RV32I-NEXT: sw s0, 8(sp) |
| 431 | ; RV32I-NEXT: addi s0, sp, 16 |
| 432 | ; RV32I-NEXT: sw a7, 28(s0) |
| 433 | ; RV32I-NEXT: sw a6, 24(s0) |
| 434 | ; RV32I-NEXT: sw a5, 20(s0) |
| 435 | ; RV32I-NEXT: sw a4, 16(s0) |
| 436 | ; RV32I-NEXT: sw a3, 12(s0) |
| 437 | ; RV32I-NEXT: sw a2, 8(s0) |
| 438 | ; RV32I-NEXT: sw a1, 4(s0) |
| 439 | ; RV32I-NEXT: addi a0, zero, 1 |
| 440 | ; RV32I-NEXT: lw s0, 8(sp) |
| 441 | ; RV32I-NEXT: lw ra, 12(sp) |
| 442 | ; RV32I-NEXT: addi sp, sp, 48 |
| 443 | ; RV32I-NEXT: ret |
| 444 | ret i32 1 |
| 445 | } |
| 446 | |
| 447 | define void @va5_aligned_stack_caller() nounwind { |
| 448 | ; The double should be 8-byte aligned on the stack, but the two-element array |
| 449 | ; should only be 4-byte aligned |
| 450 | ; RV32I-LABEL: va5_aligned_stack_caller: |
| 451 | ; RV32I: # %bb.0: |
| 452 | ; RV32I-NEXT: addi sp, sp, -64 |
| 453 | ; RV32I-NEXT: sw ra, 60(sp) |
| 454 | ; RV32I-NEXT: sw s0, 56(sp) |
| 455 | ; RV32I-NEXT: addi s0, sp, 64 |
| 456 | ; RV32I-NEXT: addi a0, zero, 17 |
| 457 | ; RV32I-NEXT: sw a0, 24(sp) |
| 458 | ; RV32I-NEXT: addi a0, zero, 16 |
| 459 | ; RV32I-NEXT: sw a0, 20(sp) |
| 460 | ; RV32I-NEXT: addi a0, zero, 15 |
| 461 | ; RV32I-NEXT: sw a0, 16(sp) |
| 462 | ; RV32I-NEXT: lui a0, 262236 |
| 463 | ; RV32I-NEXT: addi a0, a0, 655 |
| 464 | ; RV32I-NEXT: sw a0, 12(sp) |
| 465 | ; RV32I-NEXT: lui a0, 377487 |
| 466 | ; RV32I-NEXT: addi a0, a0, 1475 |
| 467 | ; RV32I-NEXT: sw a0, 8(sp) |
| 468 | ; RV32I-NEXT: addi a0, zero, 14 |
| 469 | ; RV32I-NEXT: sw a0, 0(sp) |
| 470 | ; RV32I-NEXT: lui a0, 262153 |
| 471 | ; RV32I-NEXT: addi a0, a0, 491 |
| 472 | ; RV32I-NEXT: sw a0, -20(s0) |
| 473 | ; RV32I-NEXT: lui a0, 545260 |
| 474 | ; RV32I-NEXT: addi a0, a0, -1967 |
| 475 | ; RV32I-NEXT: sw a0, -24(s0) |
| 476 | ; RV32I-NEXT: lui a0, 964690 |
| 477 | ; RV32I-NEXT: addi a0, a0, -328 |
| 478 | ; RV32I-NEXT: sw a0, -28(s0) |
| 479 | ; RV32I-NEXT: lui a0, 335544 |
| 480 | ; RV32I-NEXT: addi a0, a0, 1311 |
| 481 | ; RV32I-NEXT: sw a0, -32(s0) |
| 482 | ; RV32I-NEXT: lui a0, 688509 |
| 483 | ; RV32I-NEXT: addi a6, a0, -2048 |
| 484 | ; RV32I-NEXT: lui a0, %hi(va5_aligned_stack_callee) |
| 485 | ; RV32I-NEXT: addi a5, a0, %lo(va5_aligned_stack_callee) |
| 486 | ; RV32I-NEXT: addi a0, zero, 1 |
| 487 | ; RV32I-NEXT: addi a1, zero, 11 |
| 488 | ; RV32I-NEXT: addi a2, s0, -32 |
| 489 | ; RV32I-NEXT: addi a3, zero, 12 |
| 490 | ; RV32I-NEXT: addi a4, zero, 13 |
| 491 | ; RV32I-NEXT: addi a7, zero, 4 |
| 492 | ; RV32I-NEXT: jalr a5 |
| 493 | ; RV32I-NEXT: lw s0, 56(sp) |
| 494 | ; RV32I-NEXT: lw ra, 60(sp) |
| 495 | ; RV32I-NEXT: addi sp, sp, 64 |
| 496 | ; RV32I-NEXT: ret |
| 497 | %1 = call i32 (i32, ...) @va5_aligned_stack_callee(i32 1, i32 11, |
| 498 | fp128 0xLEB851EB851EB851F400091EB851EB851, i32 12, i32 13, i64 20000000000, |
| 499 | i32 14, double 2.720000e+00, i32 15, [2 x i32] [i32 16, i32 17]) |
| 500 | ret void |
| 501 | } |
| 502 | |
| 503 | ; A function with no fixed arguments is not valid C, but can be |
| 504 | ; specified in LLVM IR. We must ensure the vararg save area is |
| 505 | ; still set up correctly. |
| 506 | |
| 507 | define i32 @va6_no_fixed_args(...) nounwind { |
| 508 | ; RV32I-LABEL: va6_no_fixed_args: |
| 509 | ; RV32I: # %bb.0: |
| 510 | ; RV32I-NEXT: addi sp, sp, -48 |
| 511 | ; RV32I-NEXT: sw ra, 12(sp) |
| 512 | ; RV32I-NEXT: sw s0, 8(sp) |
| 513 | ; RV32I-NEXT: addi s0, sp, 16 |
| 514 | ; RV32I-NEXT: sw a0, 0(s0) |
| 515 | ; RV32I-NEXT: sw a7, 28(s0) |
| 516 | ; RV32I-NEXT: sw a6, 24(s0) |
| 517 | ; RV32I-NEXT: sw a5, 20(s0) |
| 518 | ; RV32I-NEXT: sw a4, 16(s0) |
| 519 | ; RV32I-NEXT: sw a3, 12(s0) |
| 520 | ; RV32I-NEXT: sw a2, 8(s0) |
| 521 | ; RV32I-NEXT: sw a1, 4(s0) |
| 522 | ; RV32I-NEXT: addi a0, s0, 4 |
| 523 | ; RV32I-NEXT: sw a0, -12(s0) |
| 524 | ; RV32I-NEXT: lw a0, 0(s0) |
| 525 | ; RV32I-NEXT: lw s0, 8(sp) |
| 526 | ; RV32I-NEXT: lw ra, 12(sp) |
| 527 | ; RV32I-NEXT: addi sp, sp, 48 |
| 528 | ; RV32I-NEXT: ret |
| 529 | %va = alloca i8*, align 4 |
| 530 | %1 = bitcast i8** %va to i8* |
| 531 | call void @llvm.va_start(i8* %1) |
| 532 | %2 = va_arg i8** %va, i32 |
| 533 | call void @llvm.va_end(i8* %1) |
| 534 | ret i32 %2 |
| 535 | } |