Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 1 | ; Test 64-bit conditional stores that are presented as selects. |
| 2 | ; |
Richard Sandiford | a68e6f5 | 2013-07-25 08:57:02 +0000 | [diff] [blame] | 3 | ; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z10 | FileCheck %s |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 4 | |
| 5 | declare void @foo(i64 *) |
| 6 | |
| 7 | ; Test with the loaded value first. |
| 8 | define void @f1(i64 *%ptr, i64 %alt, i32 %limit) { |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 9 | ; CHECK-LABEL: f1: |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 10 | ; CHECK-NOT: %r2 |
Ulrich Weigand | 2eb027d | 2016-04-07 16:11:44 +0000 | [diff] [blame] | 11 | ; CHECK: blr %r14 |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 12 | ; CHECK-NOT: %r2 |
| 13 | ; CHECK: stg %r3, 0(%r2) |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 14 | ; CHECK: br %r14 |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 15 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 16 | %orig = load i64, i64 *%ptr |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 17 | %res = select i1 %cond, i64 %orig, i64 %alt |
| 18 | store i64 %res, i64 *%ptr |
| 19 | ret void |
| 20 | } |
| 21 | |
| 22 | ; ...and with the loaded value second |
| 23 | define void @f2(i64 *%ptr, i64 %alt, i32 %limit) { |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 24 | ; CHECK-LABEL: f2: |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 25 | ; CHECK-NOT: %r2 |
Ulrich Weigand | 2eb027d | 2016-04-07 16:11:44 +0000 | [diff] [blame] | 26 | ; CHECK: bher %r14 |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 27 | ; CHECK-NOT: %r2 |
| 28 | ; CHECK: stg %r3, 0(%r2) |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 29 | ; CHECK: br %r14 |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 30 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 31 | %orig = load i64, i64 *%ptr |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 32 | %res = select i1 %cond, i64 %alt, i64 %orig |
| 33 | store i64 %res, i64 *%ptr |
| 34 | ret void |
| 35 | } |
| 36 | |
| 37 | ; Check the high end of the aligned STG range. |
| 38 | define void @f3(i64 *%base, i64 %alt, i32 %limit) { |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 39 | ; CHECK-LABEL: f3: |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 40 | ; CHECK-NOT: %r2 |
Ulrich Weigand | 2eb027d | 2016-04-07 16:11:44 +0000 | [diff] [blame] | 41 | ; CHECK: blr %r14 |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 42 | ; CHECK-NOT: %r2 |
| 43 | ; CHECK: stg %r3, 524280(%r2) |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 44 | ; CHECK: br %r14 |
David Blaikie | 79e6c74 | 2015-02-27 19:29:02 +0000 | [diff] [blame] | 45 | %ptr = getelementptr i64, i64 *%base, i64 65535 |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 46 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 47 | %orig = load i64, i64 *%ptr |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 48 | %res = select i1 %cond, i64 %orig, i64 %alt |
| 49 | store i64 %res, i64 *%ptr |
| 50 | ret void |
| 51 | } |
| 52 | |
| 53 | ; Check the next doubleword up, which needs separate address logic. |
| 54 | ; Other sequences besides this one would be OK. |
| 55 | define void @f4(i64 *%base, i64 %alt, i32 %limit) { |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 56 | ; CHECK-LABEL: f4: |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 57 | ; CHECK-NOT: %r2 |
Ulrich Weigand | 2eb027d | 2016-04-07 16:11:44 +0000 | [diff] [blame] | 58 | ; CHECK: blr %r14 |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 59 | ; CHECK-NOT: %r2 |
| 60 | ; CHECK: agfi %r2, 524288 |
| 61 | ; CHECK: stg %r3, 0(%r2) |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 62 | ; CHECK: br %r14 |
David Blaikie | 79e6c74 | 2015-02-27 19:29:02 +0000 | [diff] [blame] | 63 | %ptr = getelementptr i64, i64 *%base, i64 65536 |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 64 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 65 | %orig = load i64, i64 *%ptr |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 66 | %res = select i1 %cond, i64 %orig, i64 %alt |
| 67 | store i64 %res, i64 *%ptr |
| 68 | ret void |
| 69 | } |
| 70 | |
| 71 | ; Check the low end of the STG range. |
| 72 | define void @f5(i64 *%base, i64 %alt, i32 %limit) { |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 73 | ; CHECK-LABEL: f5: |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 74 | ; CHECK-NOT: %r2 |
Ulrich Weigand | 2eb027d | 2016-04-07 16:11:44 +0000 | [diff] [blame] | 75 | ; CHECK: blr %r14 |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 76 | ; CHECK-NOT: %r2 |
| 77 | ; CHECK: stg %r3, -524288(%r2) |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 78 | ; CHECK: br %r14 |
David Blaikie | 79e6c74 | 2015-02-27 19:29:02 +0000 | [diff] [blame] | 79 | %ptr = getelementptr i64, i64 *%base, i64 -65536 |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 80 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 81 | %orig = load i64, i64 *%ptr |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 82 | %res = select i1 %cond, i64 %orig, i64 %alt |
| 83 | store i64 %res, i64 *%ptr |
| 84 | ret void |
| 85 | } |
| 86 | |
| 87 | ; Check the next doubleword down, which needs separate address logic. |
| 88 | ; Other sequences besides this one would be OK. |
| 89 | define void @f6(i64 *%base, i64 %alt, i32 %limit) { |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 90 | ; CHECK-LABEL: f6: |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 91 | ; CHECK-NOT: %r2 |
Ulrich Weigand | 2eb027d | 2016-04-07 16:11:44 +0000 | [diff] [blame] | 92 | ; CHECK: blr %r14 |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 93 | ; CHECK-NOT: %r2 |
| 94 | ; CHECK: agfi %r2, -524296 |
| 95 | ; CHECK: stg %r3, 0(%r2) |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 96 | ; CHECK: br %r14 |
David Blaikie | 79e6c74 | 2015-02-27 19:29:02 +0000 | [diff] [blame] | 97 | %ptr = getelementptr i64, i64 *%base, i64 -65537 |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 98 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 99 | %orig = load i64, i64 *%ptr |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 100 | %res = select i1 %cond, i64 %orig, i64 %alt |
| 101 | store i64 %res, i64 *%ptr |
| 102 | ret void |
| 103 | } |
| 104 | |
| 105 | ; Check that STG allows an index. |
| 106 | define void @f7(i64 %base, i64 %index, i64 %alt, i32 %limit) { |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 107 | ; CHECK-LABEL: f7: |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 108 | ; CHECK-NOT: %r2 |
Ulrich Weigand | 2eb027d | 2016-04-07 16:11:44 +0000 | [diff] [blame] | 109 | ; CHECK: blr %r14 |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 110 | ; CHECK-NOT: %r2 |
| 111 | ; CHECK: stg %r4, 524287(%r3,%r2) |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 112 | ; CHECK: br %r14 |
| 113 | %add1 = add i64 %base, %index |
| 114 | %add2 = add i64 %add1, 524287 |
| 115 | %ptr = inttoptr i64 %add2 to i64 * |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 116 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 117 | %orig = load i64, i64 *%ptr |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 118 | %res = select i1 %cond, i64 %orig, i64 %alt |
| 119 | store i64 %res, i64 *%ptr |
| 120 | ret void |
| 121 | } |
| 122 | |
| 123 | ; Check that volatile loads are not matched. |
| 124 | define void @f8(i64 *%ptr, i64 %alt, i32 %limit) { |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 125 | ; CHECK-LABEL: f8: |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 126 | ; CHECK: lg {{%r[0-5]}}, 0(%r2) |
Kyle Butt | efe56fe | 2017-01-11 19:55:19 +0000 | [diff] [blame] | 127 | ; CHECK: {{jl|jnl}} [[LABEL:[^ ]*]] |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 128 | ; CHECK: [[LABEL]]: |
| 129 | ; CHECK: stg {{%r[0-5]}}, 0(%r2) |
| 130 | ; CHECK: br %r14 |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 131 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 132 | %orig = load volatile i64, i64 *%ptr |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 133 | %res = select i1 %cond, i64 %orig, i64 %alt |
| 134 | store i64 %res, i64 *%ptr |
| 135 | ret void |
| 136 | } |
| 137 | |
| 138 | ; ...likewise stores. In this case we should have a conditional load into %r3. |
| 139 | define void @f9(i64 *%ptr, i64 %alt, i32 %limit) { |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 140 | ; CHECK-LABEL: f9: |
Richard Sandiford | 3d768e3 | 2013-07-31 12:30:20 +0000 | [diff] [blame] | 141 | ; CHECK: jhe [[LABEL:[^ ]*]] |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 142 | ; CHECK: lg %r3, 0(%r2) |
| 143 | ; CHECK: [[LABEL]]: |
| 144 | ; CHECK: stg %r3, 0(%r2) |
| 145 | ; CHECK: br %r14 |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 146 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 147 | %orig = load i64, i64 *%ptr |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 148 | %res = select i1 %cond, i64 %orig, i64 %alt |
| 149 | store volatile i64 %res, i64 *%ptr |
| 150 | ret void |
| 151 | } |
| 152 | |
| 153 | ; Check that atomic loads are not matched. The transformation is OK for |
| 154 | ; the "unordered" case tested here, but since we don't try to handle atomic |
| 155 | ; operations at all in this context, it seems better to assert that than |
| 156 | ; to restrict the test to a stronger ordering. |
| 157 | define void @f10(i64 *%ptr, i64 %alt, i32 %limit) { |
| 158 | ; FIXME: should use a normal load instead of CSG. |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 159 | ; CHECK-LABEL: f10: |
Richard Sandiford | bef3d7a | 2013-12-10 10:49:34 +0000 | [diff] [blame] | 160 | ; CHECK: lg {{%r[0-5]}}, 0(%r2) |
Kyle Butt | efe56fe | 2017-01-11 19:55:19 +0000 | [diff] [blame] | 161 | ; CHECK: {{jl|jnl}} [[LABEL:[^ ]*]] |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 162 | ; CHECK: [[LABEL]]: |
| 163 | ; CHECK: stg {{%r[0-5]}}, 0(%r2) |
| 164 | ; CHECK: br %r14 |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 165 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 166 | %orig = load atomic i64, i64 *%ptr unordered, align 8 |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 167 | %res = select i1 %cond, i64 %orig, i64 %alt |
| 168 | store i64 %res, i64 *%ptr |
| 169 | ret void |
| 170 | } |
| 171 | |
| 172 | ; ...likewise stores. |
| 173 | define void @f11(i64 *%ptr, i64 %alt, i32 %limit) { |
| 174 | ; FIXME: should use a normal store instead of CSG. |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 175 | ; CHECK-LABEL: f11: |
Richard Sandiford | 3d768e3 | 2013-07-31 12:30:20 +0000 | [diff] [blame] | 176 | ; CHECK: jhe [[LABEL:[^ ]*]] |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 177 | ; CHECK: lg %r3, 0(%r2) |
| 178 | ; CHECK: [[LABEL]]: |
Richard Sandiford | bef3d7a | 2013-12-10 10:49:34 +0000 | [diff] [blame] | 179 | ; CHECK: stg %r3, 0(%r2) |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 180 | ; CHECK: br %r14 |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 181 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 182 | %orig = load i64, i64 *%ptr |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 183 | %res = select i1 %cond, i64 %orig, i64 %alt |
| 184 | store atomic i64 %res, i64 *%ptr unordered, align 8 |
| 185 | ret void |
| 186 | } |
| 187 | |
| 188 | ; Try a frame index base. |
| 189 | define void @f12(i64 %alt, i32 %limit) { |
Stephen Lin | d24ab20 | 2013-07-14 06:24:09 +0000 | [diff] [blame] | 190 | ; CHECK-LABEL: f12: |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 191 | ; CHECK: brasl %r14, foo@PLT |
| 192 | ; CHECK-NOT: %r15 |
| 193 | ; CHECK: jl [[LABEL:[^ ]*]] |
| 194 | ; CHECK-NOT: %r15 |
| 195 | ; CHECK: stg {{%r[0-9]+}}, {{[0-9]+}}(%r15) |
| 196 | ; CHECK: [[LABEL]]: |
| 197 | ; CHECK: brasl %r14, foo@PLT |
| 198 | ; CHECK: br %r14 |
| 199 | %ptr = alloca i64 |
| 200 | call void @foo(i64 *%ptr) |
Richard Sandiford | 93183ee | 2013-09-18 09:56:40 +0000 | [diff] [blame] | 201 | %cond = icmp ult i32 %limit, 420 |
Ulrich Weigand | 9dd23b8 | 2018-07-20 12:12:10 +0000 | [diff] [blame] | 202 | %orig = load i64, i64 *%ptr |
Richard Sandiford | b86a834 | 2013-06-27 09:27:40 +0000 | [diff] [blame] | 203 | %res = select i1 %cond, i64 %orig, i64 %alt |
| 204 | store i64 %res, i64 *%ptr |
| 205 | call void @foo(i64 *%ptr) |
| 206 | ret void |
| 207 | } |