Tim Northover | 3b0846e | 2014-05-24 12:50:23 +0000 | [diff] [blame] | 1 | ; RUN: llc < %s -march=arm64 -verify-machineinstrs | FileCheck %s |
| 2 | ; RUN: llc < %s -march=arm64 -aarch64-unscaled-mem-op=true\ |
| 3 | ; RUN: -verify-machineinstrs | FileCheck -check-prefix=LDUR_CHK %s |
| 4 | |
| 5 | ; CHECK: ldp_int |
| 6 | ; CHECK: ldp |
| 7 | define i32 @ldp_int(i32* %p) nounwind { |
| 8 | %tmp = load i32* %p, align 4 |
| 9 | %add.ptr = getelementptr inbounds i32* %p, i64 1 |
| 10 | %tmp1 = load i32* %add.ptr, align 4 |
| 11 | %add = add nsw i32 %tmp1, %tmp |
| 12 | ret i32 %add |
| 13 | } |
| 14 | |
| 15 | ; CHECK: ldp_long |
| 16 | ; CHECK: ldp |
| 17 | define i64 @ldp_long(i64* %p) nounwind { |
| 18 | %tmp = load i64* %p, align 8 |
| 19 | %add.ptr = getelementptr inbounds i64* %p, i64 1 |
| 20 | %tmp1 = load i64* %add.ptr, align 8 |
| 21 | %add = add nsw i64 %tmp1, %tmp |
| 22 | ret i64 %add |
| 23 | } |
| 24 | |
| 25 | ; CHECK: ldp_float |
| 26 | ; CHECK: ldp |
| 27 | define float @ldp_float(float* %p) nounwind { |
| 28 | %tmp = load float* %p, align 4 |
| 29 | %add.ptr = getelementptr inbounds float* %p, i64 1 |
| 30 | %tmp1 = load float* %add.ptr, align 4 |
| 31 | %add = fadd float %tmp, %tmp1 |
| 32 | ret float %add |
| 33 | } |
| 34 | |
| 35 | ; CHECK: ldp_double |
| 36 | ; CHECK: ldp |
| 37 | define double @ldp_double(double* %p) nounwind { |
| 38 | %tmp = load double* %p, align 8 |
| 39 | %add.ptr = getelementptr inbounds double* %p, i64 1 |
| 40 | %tmp1 = load double* %add.ptr, align 8 |
| 41 | %add = fadd double %tmp, %tmp1 |
| 42 | ret double %add |
| 43 | } |
| 44 | |
| 45 | ; Test the load/store optimizer---combine ldurs into a ldp, if appropriate |
| 46 | define i32 @ldur_int(i32* %a) nounwind { |
| 47 | ; LDUR_CHK: ldur_int |
| 48 | ; LDUR_CHK: ldp [[DST1:w[0-9]+]], [[DST2:w[0-9]+]], [x0, #-8] |
| 49 | ; LDUR_CHK-NEXT: add w{{[0-9]+}}, [[DST2]], [[DST1]] |
| 50 | ; LDUR_CHK-NEXT: ret |
| 51 | %p1 = getelementptr inbounds i32* %a, i32 -1 |
| 52 | %tmp1 = load i32* %p1, align 2 |
| 53 | %p2 = getelementptr inbounds i32* %a, i32 -2 |
| 54 | %tmp2 = load i32* %p2, align 2 |
| 55 | %tmp3 = add i32 %tmp1, %tmp2 |
| 56 | ret i32 %tmp3 |
| 57 | } |
| 58 | |
| 59 | define i64 @ldur_long(i64* %a) nounwind ssp { |
| 60 | ; LDUR_CHK: ldur_long |
| 61 | ; LDUR_CHK: ldp [[DST1:x[0-9]+]], [[DST2:x[0-9]+]], [x0, #-16] |
| 62 | ; LDUR_CHK-NEXT: add x{{[0-9]+}}, [[DST2]], [[DST1]] |
| 63 | ; LDUR_CHK-NEXT: ret |
| 64 | %p1 = getelementptr inbounds i64* %a, i64 -1 |
| 65 | %tmp1 = load i64* %p1, align 2 |
| 66 | %p2 = getelementptr inbounds i64* %a, i64 -2 |
| 67 | %tmp2 = load i64* %p2, align 2 |
| 68 | %tmp3 = add i64 %tmp1, %tmp2 |
| 69 | ret i64 %tmp3 |
| 70 | } |
| 71 | |
| 72 | define float @ldur_float(float* %a) { |
| 73 | ; LDUR_CHK: ldur_float |
| 74 | ; LDUR_CHK: ldp [[DST1:s[0-9]+]], [[DST2:s[0-9]+]], [x0, #-8] |
| 75 | ; LDUR_CHK-NEXT: add s{{[0-9]+}}, [[DST2]], [[DST1]] |
| 76 | ; LDUR_CHK-NEXT: ret |
| 77 | %p1 = getelementptr inbounds float* %a, i64 -1 |
| 78 | %tmp1 = load float* %p1, align 2 |
| 79 | %p2 = getelementptr inbounds float* %a, i64 -2 |
| 80 | %tmp2 = load float* %p2, align 2 |
| 81 | %tmp3 = fadd float %tmp1, %tmp2 |
| 82 | ret float %tmp3 |
| 83 | } |
| 84 | |
| 85 | define double @ldur_double(double* %a) { |
| 86 | ; LDUR_CHK: ldur_double |
| 87 | ; LDUR_CHK: ldp [[DST1:d[0-9]+]], [[DST2:d[0-9]+]], [x0, #-16] |
| 88 | ; LDUR_CHK-NEXT: add d{{[0-9]+}}, [[DST2]], [[DST1]] |
| 89 | ; LDUR_CHK-NEXT: ret |
| 90 | %p1 = getelementptr inbounds double* %a, i64 -1 |
| 91 | %tmp1 = load double* %p1, align 2 |
| 92 | %p2 = getelementptr inbounds double* %a, i64 -2 |
| 93 | %tmp2 = load double* %p2, align 2 |
| 94 | %tmp3 = fadd double %tmp1, %tmp2 |
| 95 | ret double %tmp3 |
| 96 | } |
| 97 | |
| 98 | ; Now check some boundary conditions |
| 99 | define i64 @pairUpBarelyIn(i64* %a) nounwind ssp { |
| 100 | ; LDUR_CHK: pairUpBarelyIn |
| 101 | ; LDUR_CHK-NOT: ldur |
| 102 | ; LDUR_CHK: ldp [[DST1:x[0-9]+]], [[DST2:x[0-9]+]], [x0, #-256] |
| 103 | ; LDUR_CHK-NEXT: add x{{[0-9]+}}, [[DST2]], [[DST1]] |
| 104 | ; LDUR_CHK-NEXT: ret |
| 105 | %p1 = getelementptr inbounds i64* %a, i64 -31 |
| 106 | %tmp1 = load i64* %p1, align 2 |
| 107 | %p2 = getelementptr inbounds i64* %a, i64 -32 |
| 108 | %tmp2 = load i64* %p2, align 2 |
| 109 | %tmp3 = add i64 %tmp1, %tmp2 |
| 110 | ret i64 %tmp3 |
| 111 | } |
| 112 | |
| 113 | define i64 @pairUpBarelyOut(i64* %a) nounwind ssp { |
| 114 | ; LDUR_CHK: pairUpBarelyOut |
| 115 | ; LDUR_CHK-NOT: ldp |
| 116 | ; Don't be fragile about which loads or manipulations of the base register |
| 117 | ; are used---just check that there isn't an ldp before the add |
| 118 | ; LDUR_CHK: add |
| 119 | ; LDUR_CHK-NEXT: ret |
| 120 | %p1 = getelementptr inbounds i64* %a, i64 -32 |
| 121 | %tmp1 = load i64* %p1, align 2 |
| 122 | %p2 = getelementptr inbounds i64* %a, i64 -33 |
| 123 | %tmp2 = load i64* %p2, align 2 |
| 124 | %tmp3 = add i64 %tmp1, %tmp2 |
| 125 | ret i64 %tmp3 |
| 126 | } |
| 127 | |
| 128 | define i64 @pairUpNotAligned(i64* %a) nounwind ssp { |
| 129 | ; LDUR_CHK: pairUpNotAligned |
| 130 | ; LDUR_CHK-NOT: ldp |
| 131 | ; LDUR_CHK: ldur |
| 132 | ; LDUR_CHK-NEXT: ldur |
| 133 | ; LDUR_CHK-NEXT: add |
| 134 | ; LDUR_CHK-NEXT: ret |
| 135 | %p1 = getelementptr inbounds i64* %a, i64 -18 |
| 136 | %bp1 = bitcast i64* %p1 to i8* |
| 137 | %bp1p1 = getelementptr inbounds i8* %bp1, i64 1 |
| 138 | %dp1 = bitcast i8* %bp1p1 to i64* |
| 139 | %tmp1 = load i64* %dp1, align 1 |
| 140 | |
| 141 | %p2 = getelementptr inbounds i64* %a, i64 -17 |
| 142 | %bp2 = bitcast i64* %p2 to i8* |
| 143 | %bp2p1 = getelementptr inbounds i8* %bp2, i64 1 |
| 144 | %dp2 = bitcast i8* %bp2p1 to i64* |
| 145 | %tmp2 = load i64* %dp2, align 1 |
| 146 | |
| 147 | %tmp3 = add i64 %tmp1, %tmp2 |
| 148 | ret i64 %tmp3 |
| 149 | } |