Cong Hou | 4ce0280 | 2016-02-24 23:40:36 +0000 | [diff] [blame] | 1 | ; RUN: llc < %s -debug-only=isel -o /dev/null 2>&1 | FileCheck %s |
| 2 | ; REQUIRES: asserts |
| 3 | |
Cong Hou | 4ce0280 | 2016-02-24 23:40:36 +0000 | [diff] [blame] | 4 | @a = global [1024 x i32] zeroinitializer, align 16 |
| 5 | |
| 6 | define i32 @reduce_add() { |
| 7 | ; CHECK-LABEL: reduce_add |
| 8 | ; CHECK: Detected a reduction operation: {{.*}} add |
| 9 | ; CHECK: Detected a reduction operation: {{.*}} add |
| 10 | ; CHECK: Detected a reduction operation: {{.*}} add |
| 11 | ; CHECK: Detected a reduction operation: {{.*}} add |
| 12 | ; CHECK: Detected a reduction operation: {{.*}} add |
| 13 | ; CHECK: Detected a reduction operation: {{.*}} add |
| 14 | ; CHECK: Detected a reduction operation: {{.*}} add |
| 15 | ; CHECK: Detected a reduction operation: {{.*}} add |
| 16 | ; CHECK: Detected a reduction operation: {{.*}} add |
| 17 | ; CHECK: Detected a reduction operation: {{.*}} add |
| 18 | ; CHECK: Detected a reduction operation: {{.*}} add |
| 19 | |
| 20 | min.iters.checked: |
| 21 | br label %vector.body |
| 22 | |
| 23 | vector.body: |
| 24 | %index = phi i64 [ 0, %min.iters.checked ], [ %index.next.4, %vector.body ] |
| 25 | %vec.phi = phi <4 x i32> [ zeroinitializer, %min.iters.checked ], [ %28, %vector.body ] |
| 26 | %vec.phi4 = phi <4 x i32> [ zeroinitializer, %min.iters.checked ], [ %29, %vector.body ] |
| 27 | %0 = getelementptr inbounds [1024 x i32], [1024 x i32]* @a, i64 0, i64 %index |
| 28 | %1 = bitcast i32* %0 to <4 x i32>* |
| 29 | %wide.load = load <4 x i32>, <4 x i32>* %1, align 16 |
| 30 | %2 = getelementptr i32, i32* %0, i64 4 |
| 31 | %3 = bitcast i32* %2 to <4 x i32>* |
| 32 | %wide.load5 = load <4 x i32>, <4 x i32>* %3, align 16 |
| 33 | %4 = add nsw <4 x i32> %wide.load, %vec.phi |
| 34 | %5 = add nsw <4 x i32> %wide.load5, %vec.phi4 |
| 35 | %index.next = add nuw nsw i64 %index, 8 |
| 36 | %6 = getelementptr inbounds [1024 x i32], [1024 x i32]* @a, i64 0, i64 %index.next |
| 37 | %7 = bitcast i32* %6 to <4 x i32>* |
| 38 | %wide.load.1 = load <4 x i32>, <4 x i32>* %7, align 16 |
| 39 | %8 = getelementptr i32, i32* %6, i64 4 |
| 40 | %9 = bitcast i32* %8 to <4 x i32>* |
| 41 | %wide.load5.1 = load <4 x i32>, <4 x i32>* %9, align 16 |
| 42 | %10 = add nsw <4 x i32> %wide.load.1, %4 |
| 43 | %11 = add nsw <4 x i32> %wide.load5.1, %5 |
| 44 | %index.next.1 = add nsw i64 %index, 16 |
| 45 | %12 = getelementptr inbounds [1024 x i32], [1024 x i32]* @a, i64 0, i64 %index.next.1 |
| 46 | %13 = bitcast i32* %12 to <4 x i32>* |
| 47 | %wide.load.2 = load <4 x i32>, <4 x i32>* %13, align 16 |
| 48 | %14 = getelementptr i32, i32* %12, i64 4 |
| 49 | %15 = bitcast i32* %14 to <4 x i32>* |
| 50 | %wide.load5.2 = load <4 x i32>, <4 x i32>* %15, align 16 |
| 51 | %16 = add nsw <4 x i32> %wide.load.2, %10 |
| 52 | %17 = add nsw <4 x i32> %wide.load5.2, %11 |
| 53 | %index.next.2 = add nsw i64 %index, 24 |
| 54 | %18 = getelementptr inbounds [1024 x i32], [1024 x i32]* @a, i64 0, i64 %index.next.2 |
| 55 | %19 = bitcast i32* %18 to <4 x i32>* |
| 56 | %wide.load.3 = load <4 x i32>, <4 x i32>* %19, align 16 |
| 57 | %20 = getelementptr i32, i32* %18, i64 4 |
| 58 | %21 = bitcast i32* %20 to <4 x i32>* |
| 59 | %wide.load5.3 = load <4 x i32>, <4 x i32>* %21, align 16 |
| 60 | %22 = add nsw <4 x i32> %wide.load.3, %16 |
| 61 | %23 = add nsw <4 x i32> %wide.load5.3, %17 |
| 62 | %index.next.3 = add nsw i64 %index, 32 |
| 63 | %24 = getelementptr inbounds [1024 x i32], [1024 x i32]* @a, i64 0, i64 %index.next.3 |
| 64 | %25 = bitcast i32* %24 to <4 x i32>* |
| 65 | %wide.load.4 = load <4 x i32>, <4 x i32>* %25, align 16 |
| 66 | %26 = getelementptr i32, i32* %24, i64 4 |
| 67 | %27 = bitcast i32* %26 to <4 x i32>* |
| 68 | %wide.load5.4 = load <4 x i32>, <4 x i32>* %27, align 16 |
| 69 | %28 = add nsw <4 x i32> %wide.load.4, %22 |
| 70 | %29 = add nsw <4 x i32> %wide.load5.4, %23 |
| 71 | %index.next.4 = add nsw i64 %index, 40 |
| 72 | %30 = icmp eq i64 %index.next.4, 1000 |
| 73 | br i1 %30, label %middle.block, label %vector.body |
| 74 | |
| 75 | middle.block: |
| 76 | %.lcssa10 = phi <4 x i32> [ %29, %vector.body ] |
| 77 | %.lcssa = phi <4 x i32> [ %28, %vector.body ] |
| 78 | %bin.rdx = add <4 x i32> %.lcssa10, %.lcssa |
| 79 | %rdx.shuf = shufflevector <4 x i32> %bin.rdx, <4 x i32> undef, <4 x i32> <i32 2, i32 3, i32 undef, i32 undef> |
| 80 | %bin.rdx6 = add <4 x i32> %bin.rdx, %rdx.shuf |
| 81 | %rdx.shuf7 = shufflevector <4 x i32> %bin.rdx6, <4 x i32> undef, <4 x i32> <i32 1, i32 undef, i32 undef, i32 undef> |
| 82 | %bin.rdx8 = add <4 x i32> %bin.rdx6, %rdx.shuf7 |
| 83 | %31 = extractelement <4 x i32> %bin.rdx8, i32 0 |
| 84 | ret i32 %31 |
| 85 | } |
| 86 | |
| 87 | define i32 @reduce_and() { |
| 88 | ; CHECK-LABEL: reduce_and |
| 89 | ; CHECK: Detected a reduction operation: {{.*}} and |
| 90 | ; CHECK: Detected a reduction operation: {{.*}} and |
| 91 | ; CHECK: Detected a reduction operation: {{.*}} and |
| 92 | ; CHECK: Detected a reduction operation: {{.*}} and |
| 93 | ; CHECK: Detected a reduction operation: {{.*}} and |
| 94 | ; CHECK: Detected a reduction operation: {{.*}} and |
| 95 | ; CHECK: Detected a reduction operation: {{.*}} and |
| 96 | ; CHECK: Detected a reduction operation: {{.*}} and |
| 97 | ; CHECK: Detected a reduction operation: {{.*}} and |
| 98 | |
| 99 | entry: |
| 100 | br label %vector.body |
| 101 | |
| 102 | vector.body: |
| 103 | %lsr.iv = phi i64 [ %lsr.iv.next, %vector.body ], [ -4096, %entry ] |
| 104 | %vec.phi = phi <4 x i32> [ <i32 -1, i32 -1, i32 -1, i32 -1>, %entry ], [ %6, %vector.body ] |
| 105 | %vec.phi9 = phi <4 x i32> [ <i32 -1, i32 -1, i32 -1, i32 -1>, %entry ], [ %7, %vector.body ] |
| 106 | %uglygep33 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv |
| 107 | %uglygep3334 = bitcast i8* %uglygep33 to <4 x i32>* |
| 108 | %scevgep35 = getelementptr <4 x i32>, <4 x i32>* %uglygep3334, i64 256 |
| 109 | %wide.load = load <4 x i32>, <4 x i32>* %scevgep35, align 16 |
| 110 | %scevgep36 = getelementptr <4 x i32>, <4 x i32>* %uglygep3334, i64 257 |
| 111 | %wide.load10 = load <4 x i32>, <4 x i32>* %scevgep36, align 16 |
| 112 | %0 = and <4 x i32> %wide.load, %vec.phi |
| 113 | %1 = and <4 x i32> %wide.load10, %vec.phi9 |
| 114 | %uglygep30 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv |
| 115 | %uglygep3031 = bitcast i8* %uglygep30 to <4 x i32>* |
| 116 | %scevgep32 = getelementptr <4 x i32>, <4 x i32>* %uglygep3031, i64 258 |
| 117 | %wide.load.1 = load <4 x i32>, <4 x i32>* %scevgep32, align 16 |
| 118 | %uglygep27 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv |
| 119 | %uglygep2728 = bitcast i8* %uglygep27 to <4 x i32>* |
| 120 | %scevgep29 = getelementptr <4 x i32>, <4 x i32>* %uglygep2728, i64 259 |
| 121 | %wide.load10.1 = load <4 x i32>, <4 x i32>* %scevgep29, align 16 |
| 122 | %2 = and <4 x i32> %wide.load.1, %0 |
| 123 | %3 = and <4 x i32> %wide.load10.1, %1 |
| 124 | %uglygep24 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv |
| 125 | %uglygep2425 = bitcast i8* %uglygep24 to <4 x i32>* |
| 126 | %scevgep26 = getelementptr <4 x i32>, <4 x i32>* %uglygep2425, i64 260 |
| 127 | %wide.load.2 = load <4 x i32>, <4 x i32>* %scevgep26, align 16 |
| 128 | %uglygep21 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv |
| 129 | %uglygep2122 = bitcast i8* %uglygep21 to <4 x i32>* |
| 130 | %scevgep23 = getelementptr <4 x i32>, <4 x i32>* %uglygep2122, i64 261 |
| 131 | %wide.load10.2 = load <4 x i32>, <4 x i32>* %scevgep23, align 16 |
| 132 | %4 = and <4 x i32> %wide.load.2, %2 |
| 133 | %5 = and <4 x i32> %wide.load10.2, %3 |
| 134 | %uglygep18 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv |
| 135 | %uglygep1819 = bitcast i8* %uglygep18 to <4 x i32>* |
| 136 | %scevgep20 = getelementptr <4 x i32>, <4 x i32>* %uglygep1819, i64 262 |
| 137 | %wide.load.3 = load <4 x i32>, <4 x i32>* %scevgep20, align 16 |
| 138 | %uglygep = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv |
| 139 | %uglygep17 = bitcast i8* %uglygep to <4 x i32>* |
| 140 | %scevgep = getelementptr <4 x i32>, <4 x i32>* %uglygep17, i64 263 |
| 141 | %wide.load10.3 = load <4 x i32>, <4 x i32>* %scevgep, align 16 |
| 142 | %6 = and <4 x i32> %wide.load.3, %4 |
| 143 | %7 = and <4 x i32> %wide.load10.3, %5 |
| 144 | %lsr.iv.next = add nsw i64 %lsr.iv, 128 |
| 145 | %8 = icmp eq i64 %lsr.iv.next, 0 |
| 146 | br i1 %8, label %middle.block, label %vector.body |
| 147 | |
| 148 | middle.block: |
| 149 | %bin.rdx = and <4 x i32> %7, %6 |
| 150 | %rdx.shuf = shufflevector <4 x i32> %bin.rdx, <4 x i32> undef, <4 x i32> <i32 2, i32 3, i32 undef, i32 undef> |
| 151 | %bin.rdx11 = and <4 x i32> %bin.rdx, %rdx.shuf |
| 152 | %rdx.shuf12 = shufflevector <4 x i32> %bin.rdx11, <4 x i32> undef, <4 x i32> <i32 1, i32 undef, i32 undef, i32 undef> |
| 153 | %bin.rdx13 = and <4 x i32> %bin.rdx11, %rdx.shuf12 |
| 154 | %9 = extractelement <4 x i32> %bin.rdx13, i32 0 |
| 155 | ret i32 %9 |
| 156 | } |
| 157 | |
| 158 | define float @reduce_add_float(float* nocapture readonly %a) { |
| 159 | ; CHECK-LABEL: reduce_add_float |
| 160 | ; CHECK: Detected a reduction operation: {{.*}} fadd fast |
| 161 | ; CHECK: Detected a reduction operation: {{.*}} fadd fast |
| 162 | ; CHECK: Detected a reduction operation: {{.*}} fadd fast |
| 163 | ; CHECK: Detected a reduction operation: {{.*}} fadd fast |
| 164 | ; CHECK: Detected a reduction operation: {{.*}} fadd fast |
| 165 | ; CHECK: Detected a reduction operation: {{.*}} fadd fast |
| 166 | ; CHECK: Detected a reduction operation: {{.*}} fadd fast |
| 167 | ; CHECK: Detected a reduction operation: {{.*}} fadd fast |
| 168 | ; CHECK: Detected a reduction operation: {{.*}} fadd fast |
| 169 | ; CHECK: Detected a reduction operation: {{.*}} fadd fast |
| 170 | ; CHECK: Detected a reduction operation: {{.*}} fadd fast |
| 171 | ; |
| 172 | entry: |
| 173 | br label %vector.body |
| 174 | |
| 175 | vector.body: |
| 176 | %index = phi i64 [ 0, %entry ], [ %index.next.4, %vector.body ] |
| 177 | %vec.phi = phi <4 x float> [ zeroinitializer, %entry ], [ %28, %vector.body ] |
| 178 | %vec.phi9 = phi <4 x float> [ zeroinitializer, %entry ], [ %29, %vector.body ] |
| 179 | %0 = getelementptr inbounds float, float* %a, i64 %index |
| 180 | %1 = bitcast float* %0 to <4 x float>* |
| 181 | %wide.load = load <4 x float>, <4 x float>* %1, align 4 |
| 182 | %2 = getelementptr float, float* %0, i64 4 |
| 183 | %3 = bitcast float* %2 to <4 x float>* |
| 184 | %wide.load10 = load <4 x float>, <4 x float>* %3, align 4 |
| 185 | %4 = fadd fast <4 x float> %wide.load, %vec.phi |
| 186 | %5 = fadd fast <4 x float> %wide.load10, %vec.phi9 |
| 187 | %index.next = add nuw nsw i64 %index, 8 |
| 188 | %6 = getelementptr inbounds float, float* %a, i64 %index.next |
| 189 | %7 = bitcast float* %6 to <4 x float>* |
| 190 | %wide.load.1 = load <4 x float>, <4 x float>* %7, align 4 |
| 191 | %8 = getelementptr float, float* %6, i64 4 |
| 192 | %9 = bitcast float* %8 to <4 x float>* |
| 193 | %wide.load10.1 = load <4 x float>, <4 x float>* %9, align 4 |
| 194 | %10 = fadd fast <4 x float> %wide.load.1, %4 |
| 195 | %11 = fadd fast <4 x float> %wide.load10.1, %5 |
| 196 | %index.next.1 = add nsw i64 %index, 16 |
| 197 | %12 = getelementptr inbounds float, float* %a, i64 %index.next.1 |
| 198 | %13 = bitcast float* %12 to <4 x float>* |
| 199 | %wide.load.2 = load <4 x float>, <4 x float>* %13, align 4 |
| 200 | %14 = getelementptr float, float* %12, i64 4 |
| 201 | %15 = bitcast float* %14 to <4 x float>* |
| 202 | %wide.load10.2 = load <4 x float>, <4 x float>* %15, align 4 |
| 203 | %16 = fadd fast <4 x float> %wide.load.2, %10 |
| 204 | %17 = fadd fast <4 x float> %wide.load10.2, %11 |
| 205 | %index.next.2 = add nsw i64 %index, 24 |
| 206 | %18 = getelementptr inbounds float, float* %a, i64 %index.next.2 |
| 207 | %19 = bitcast float* %18 to <4 x float>* |
| 208 | %wide.load.3 = load <4 x float>, <4 x float>* %19, align 4 |
| 209 | %20 = getelementptr float, float* %18, i64 4 |
| 210 | %21 = bitcast float* %20 to <4 x float>* |
| 211 | %wide.load10.3 = load <4 x float>, <4 x float>* %21, align 4 |
| 212 | %22 = fadd fast <4 x float> %wide.load.3, %16 |
| 213 | %23 = fadd fast <4 x float> %wide.load10.3, %17 |
| 214 | %index.next.3 = add nsw i64 %index, 32 |
| 215 | %24 = getelementptr inbounds float, float* %a, i64 %index.next.3 |
| 216 | %25 = bitcast float* %24 to <4 x float>* |
| 217 | %wide.load.4 = load <4 x float>, <4 x float>* %25, align 4 |
| 218 | %26 = getelementptr float, float* %24, i64 4 |
| 219 | %27 = bitcast float* %26 to <4 x float>* |
| 220 | %wide.load10.4 = load <4 x float>, <4 x float>* %27, align 4 |
| 221 | %28 = fadd fast <4 x float> %wide.load.4, %22 |
| 222 | %29 = fadd fast <4 x float> %wide.load10.4, %23 |
| 223 | %index.next.4 = add nsw i64 %index, 40 |
| 224 | %30 = icmp eq i64 %index.next.4, 1000 |
| 225 | br i1 %30, label %middle.block, label %vector.body |
| 226 | |
| 227 | middle.block: |
| 228 | %.lcssa15 = phi <4 x float> [ %29, %vector.body ] |
| 229 | %.lcssa = phi <4 x float> [ %28, %vector.body ] |
| 230 | %bin.rdx = fadd fast <4 x float> %.lcssa15, %.lcssa |
| 231 | %rdx.shuf = shufflevector <4 x float> %bin.rdx, <4 x float> undef, <4 x i32> <i32 2, i32 3, i32 undef, i32 undef> |
| 232 | %bin.rdx11 = fadd fast <4 x float> %bin.rdx, %rdx.shuf |
| 233 | %rdx.shuf12 = shufflevector <4 x float> %bin.rdx11, <4 x float> undef, <4 x i32> <i32 1, i32 undef, i32 undef, i32 undef> |
| 234 | %bin.rdx13 = fadd fast <4 x float> %bin.rdx11, %rdx.shuf12 |
| 235 | %31 = extractelement <4 x float> %bin.rdx13, i32 0 |
| 236 | ret float %31 |
| 237 | } |