| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 1 | ; RUN: llc -march=r600 -mcpu=bonaire -verify-machineinstrs -mattr=+load-store-opt -enable-misched < %s | FileCheck -check-prefix=SI %s | 
|  | 2 |  | 
| Matt Arsenault | cc8d3b8 | 2014-11-13 19:56:13 +0000 | [diff] [blame] | 3 | @lds = addrspace(3) global [512 x float] undef, align 4 | 
|  | 4 | @lds.f64 = addrspace(3) global [512 x double] undef, align 8 | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 5 |  | 
|  | 6 |  | 
|  | 7 | ; SI-LABEL: @simple_read2st64_f32_0_1 | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 8 | ; SI: ds_read2st64_b32 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:0 offset1:1 | 
|  | 9 | ; SI: s_waitcnt lgkmcnt(0) | 
|  | 10 | ; SI: v_add_f32_e32 [[RESULT:v[0-9]+]], v[[HI_VREG]], v[[LO_VREG]] | 
|  | 11 | ; SI: buffer_store_dword [[RESULT]] | 
|  | 12 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 13 | define void @simple_read2st64_f32_0_1(float addrspace(1)* %out) #0 { | 
|  | 14 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 15 | %arrayidx0 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %x.i | 
|  | 16 | %val0 = load float addrspace(3)* %arrayidx0, align 4 | 
|  | 17 | %add.x = add nsw i32 %x.i, 64 | 
|  | 18 | %arrayidx1 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %add.x | 
|  | 19 | %val1 = load float addrspace(3)* %arrayidx1, align 4 | 
|  | 20 | %sum = fadd float %val0, %val1 | 
|  | 21 | %out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i | 
|  | 22 | store float %sum, float addrspace(1)* %out.gep, align 4 | 
|  | 23 | ret void | 
|  | 24 | } | 
|  | 25 |  | 
|  | 26 | ; SI-LABEL: @simple_read2st64_f32_1_2 | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 27 | ; SI: ds_read2st64_b32 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:1 offset1:2 | 
|  | 28 | ; SI: s_waitcnt lgkmcnt(0) | 
|  | 29 | ; SI: v_add_f32_e32 [[RESULT:v[0-9]+]], v[[HI_VREG]], v[[LO_VREG]] | 
|  | 30 | ; SI: buffer_store_dword [[RESULT]] | 
|  | 31 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 32 | define void @simple_read2st64_f32_1_2(float addrspace(1)* %out, float addrspace(3)* %lds) #0 { | 
|  | 33 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 34 | %add.x.0 = add nsw i32 %x.i, 64 | 
|  | 35 | %arrayidx0 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.0 | 
|  | 36 | %val0 = load float addrspace(3)* %arrayidx0, align 4 | 
|  | 37 | %add.x.1 = add nsw i32 %x.i, 128 | 
|  | 38 | %arrayidx1 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.1 | 
|  | 39 | %val1 = load float addrspace(3)* %arrayidx1, align 4 | 
|  | 40 | %sum = fadd float %val0, %val1 | 
|  | 41 | %out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i | 
|  | 42 | store float %sum, float addrspace(1)* %out.gep, align 4 | 
|  | 43 | ret void | 
|  | 44 | } | 
|  | 45 |  | 
|  | 46 | ; SI-LABEL: @simple_read2st64_f32_max_offset | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 47 | ; SI: ds_read2st64_b32 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:1 offset1:255 | 
|  | 48 | ; SI: s_waitcnt lgkmcnt(0) | 
|  | 49 | ; SI: v_add_f32_e32 [[RESULT:v[0-9]+]], v[[HI_VREG]], v[[LO_VREG]] | 
|  | 50 | ; SI: buffer_store_dword [[RESULT]] | 
|  | 51 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 52 | define void @simple_read2st64_f32_max_offset(float addrspace(1)* %out, float addrspace(3)* %lds) #0 { | 
|  | 53 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 54 | %add.x.0 = add nsw i32 %x.i, 64 | 
|  | 55 | %arrayidx0 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.0 | 
|  | 56 | %val0 = load float addrspace(3)* %arrayidx0, align 4 | 
|  | 57 | %add.x.1 = add nsw i32 %x.i, 16320 | 
|  | 58 | %arrayidx1 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.1 | 
|  | 59 | %val1 = load float addrspace(3)* %arrayidx1, align 4 | 
|  | 60 | %sum = fadd float %val0, %val1 | 
|  | 61 | %out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i | 
|  | 62 | store float %sum, float addrspace(1)* %out.gep, align 4 | 
|  | 63 | ret void | 
|  | 64 | } | 
|  | 65 |  | 
|  | 66 | ; SI-LABEL: @simple_read2st64_f32_over_max_offset | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 67 | ; SI-NOT: ds_read2st64_b32 | 
|  | 68 | ; SI: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:256 | 
|  | 69 | ; SI: v_add_i32_e32 [[BIGADD:v[0-9]+]], 0x10000, {{v[0-9]+}} | 
|  | 70 | ; SI: ds_read_b32 {{v[0-9]+}}, [[BIGADD]] | 
|  | 71 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 72 | define void @simple_read2st64_f32_over_max_offset(float addrspace(1)* %out, float addrspace(3)* %lds) #0 { | 
|  | 73 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 74 | %add.x.0 = add nsw i32 %x.i, 64 | 
|  | 75 | %arrayidx0 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.0 | 
|  | 76 | %val0 = load float addrspace(3)* %arrayidx0, align 4 | 
|  | 77 | %add.x.1 = add nsw i32 %x.i, 16384 | 
|  | 78 | %arrayidx1 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.1 | 
|  | 79 | %val1 = load float addrspace(3)* %arrayidx1, align 4 | 
|  | 80 | %sum = fadd float %val0, %val1 | 
|  | 81 | %out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i | 
|  | 82 | store float %sum, float addrspace(1)* %out.gep, align 4 | 
|  | 83 | ret void | 
|  | 84 | } | 
|  | 85 |  | 
|  | 86 | ; SI-LABEL: @odd_invalid_read2st64_f32_0 | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 87 | ; SI-NOT: ds_read2st64_b32 | 
|  | 88 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 89 | define void @odd_invalid_read2st64_f32_0(float addrspace(1)* %out) #0 { | 
|  | 90 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 91 | %arrayidx0 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %x.i | 
|  | 92 | %val0 = load float addrspace(3)* %arrayidx0, align 4 | 
|  | 93 | %add.x = add nsw i32 %x.i, 63 | 
|  | 94 | %arrayidx1 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %add.x | 
|  | 95 | %val1 = load float addrspace(3)* %arrayidx1, align 4 | 
|  | 96 | %sum = fadd float %val0, %val1 | 
|  | 97 | %out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i | 
|  | 98 | store float %sum, float addrspace(1)* %out.gep, align 4 | 
|  | 99 | ret void | 
|  | 100 | } | 
|  | 101 |  | 
|  | 102 | ; SI-LABEL: @odd_invalid_read2st64_f32_1 | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 103 | ; SI-NOT: ds_read2st64_b32 | 
|  | 104 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 105 | define void @odd_invalid_read2st64_f32_1(float addrspace(1)* %out) #0 { | 
|  | 106 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 107 | %add.x.0 = add nsw i32 %x.i, 64 | 
|  | 108 | %arrayidx0 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %add.x.0 | 
|  | 109 | %val0 = load float addrspace(3)* %arrayidx0, align 4 | 
|  | 110 | %add.x.1 = add nsw i32 %x.i, 127 | 
|  | 111 | %arrayidx1 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %add.x.1 | 
|  | 112 | %val1 = load float addrspace(3)* %arrayidx1, align 4 | 
|  | 113 | %sum = fadd float %val0, %val1 | 
|  | 114 | %out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i | 
|  | 115 | store float %sum, float addrspace(1)* %out.gep, align 4 | 
|  | 116 | ret void | 
|  | 117 | } | 
|  | 118 |  | 
|  | 119 | ; SI-LABEL: @simple_read2st64_f64_0_1 | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 120 | ; SI: ds_read2st64_b64 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:0 offset1:1 | 
|  | 121 | ; SI: s_waitcnt lgkmcnt(0) | 
|  | 122 | ; SI: v_add_f64 [[RESULT:v\[[0-9]+:[0-9]+\]]], v{{\[}}[[LO_VREG]]:{{[0-9]+\]}}, v{{\[[0-9]+}}:[[HI_VREG]]{{\]}} | 
|  | 123 | ; SI: buffer_store_dwordx2 [[RESULT]] | 
|  | 124 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 125 | define void @simple_read2st64_f64_0_1(double addrspace(1)* %out) #0 { | 
|  | 126 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 127 | %arrayidx0 = getelementptr inbounds [512 x double] addrspace(3)* @lds.f64, i32 0, i32 %x.i | 
|  | 128 | %val0 = load double addrspace(3)* %arrayidx0, align 8 | 
|  | 129 | %add.x = add nsw i32 %x.i, 64 | 
|  | 130 | %arrayidx1 = getelementptr inbounds [512 x double] addrspace(3)* @lds.f64, i32 0, i32 %add.x | 
|  | 131 | %val1 = load double addrspace(3)* %arrayidx1, align 8 | 
|  | 132 | %sum = fadd double %val0, %val1 | 
|  | 133 | %out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i | 
|  | 134 | store double %sum, double addrspace(1)* %out.gep, align 8 | 
|  | 135 | ret void | 
|  | 136 | } | 
|  | 137 |  | 
|  | 138 | ; SI-LABEL: @simple_read2st64_f64_1_2 | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 139 | ; SI: ds_read2st64_b64 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:1 offset1:2 | 
|  | 140 | ; SI: s_waitcnt lgkmcnt(0) | 
|  | 141 | ; SI: v_add_f64 [[RESULT:v\[[0-9]+:[0-9]+\]]], v{{\[}}[[LO_VREG]]:{{[0-9]+\]}}, v{{\[[0-9]+}}:[[HI_VREG]]{{\]}} | 
|  | 142 | ; SI: buffer_store_dwordx2 [[RESULT]] | 
|  | 143 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 144 | define void @simple_read2st64_f64_1_2(double addrspace(1)* %out, double addrspace(3)* %lds) #0 { | 
|  | 145 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 146 | %add.x.0 = add nsw i32 %x.i, 64 | 
|  | 147 | %arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.0 | 
|  | 148 | %val0 = load double addrspace(3)* %arrayidx0, align 8 | 
|  | 149 | %add.x.1 = add nsw i32 %x.i, 128 | 
|  | 150 | %arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.1 | 
|  | 151 | %val1 = load double addrspace(3)* %arrayidx1, align 8 | 
|  | 152 | %sum = fadd double %val0, %val1 | 
|  | 153 | %out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i | 
|  | 154 | store double %sum, double addrspace(1)* %out.gep, align 8 | 
|  | 155 | ret void | 
|  | 156 | } | 
|  | 157 |  | 
|  | 158 | ; Alignment only | 
|  | 159 |  | 
|  | 160 | ; SI-LABEL: @misaligned_read2st64_f64 | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 161 | ; SI: ds_read2_b32 v{{\[[0-9]+:[0-9]+\]}}, {{v[0-9]+}} offset0:0 offset1:1 | 
|  | 162 | ; SI: ds_read2_b32 v{{\[[0-9]+:[0-9]+\]}}, {{v[0-9]+}} offset0:128 offset1:129 | 
|  | 163 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 164 | define void @misaligned_read2st64_f64(double addrspace(1)* %out, double addrspace(3)* %lds) #0 { | 
|  | 165 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 166 | %arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %x.i | 
|  | 167 | %val0 = load double addrspace(3)* %arrayidx0, align 4 | 
|  | 168 | %add.x = add nsw i32 %x.i, 64 | 
|  | 169 | %arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x | 
|  | 170 | %val1 = load double addrspace(3)* %arrayidx1, align 4 | 
|  | 171 | %sum = fadd double %val0, %val1 | 
|  | 172 | %out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i | 
|  | 173 | store double %sum, double addrspace(1)* %out.gep, align 4 | 
|  | 174 | ret void | 
|  | 175 | } | 
|  | 176 |  | 
|  | 177 | ; The maximum is not the usual 0xff because 0xff * 8 * 64 > 0xffff | 
|  | 178 | ; SI-LABEL: @simple_read2st64_f64_max_offset | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 179 | ; SI: ds_read2st64_b64 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:4 offset1:127 | 
|  | 180 | ; SI: s_waitcnt lgkmcnt(0) | 
|  | 181 | ; SI: v_add_f64 [[RESULT:v\[[0-9]+:[0-9]+\]]], v{{\[}}[[LO_VREG]]:{{[0-9]+\]}}, v{{\[[0-9]+}}:[[HI_VREG]]{{\]}} | 
|  | 182 | ; SI: buffer_store_dwordx2 [[RESULT]] | 
|  | 183 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 184 | define void @simple_read2st64_f64_max_offset(double addrspace(1)* %out, double addrspace(3)* %lds) #0 { | 
|  | 185 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 186 | %add.x.0 = add nsw i32 %x.i, 256 | 
|  | 187 | %arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.0 | 
|  | 188 | %val0 = load double addrspace(3)* %arrayidx0, align 8 | 
|  | 189 | %add.x.1 = add nsw i32 %x.i, 8128 | 
|  | 190 | %arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.1 | 
|  | 191 | %val1 = load double addrspace(3)* %arrayidx1, align 8 | 
|  | 192 | %sum = fadd double %val0, %val1 | 
|  | 193 | %out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i | 
|  | 194 | store double %sum, double addrspace(1)* %out.gep, align 8 | 
|  | 195 | ret void | 
|  | 196 | } | 
|  | 197 |  | 
|  | 198 | ; SI-LABEL: @simple_read2st64_f64_over_max_offset | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 199 | ; SI-NOT: ds_read2st64_b64 | 
|  | 200 | ; SI: ds_read_b64 {{v\[[0-9]+:[0-9]+\]}}, {{v[0-9]+}} offset:512 | 
|  | 201 | ; SI: v_add_i32_e32 [[BIGADD:v[0-9]+]], 0x10000, {{v[0-9]+}} | 
|  | 202 | ; SI: ds_read_b64 {{v\[[0-9]+:[0-9]+\]}}, [[BIGADD]] | 
|  | 203 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 204 | define void @simple_read2st64_f64_over_max_offset(double addrspace(1)* %out, double addrspace(3)* %lds) #0 { | 
|  | 205 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 206 | %add.x.0 = add nsw i32 %x.i, 64 | 
|  | 207 | %arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.0 | 
|  | 208 | %val0 = load double addrspace(3)* %arrayidx0, align 8 | 
|  | 209 | %add.x.1 = add nsw i32 %x.i, 8192 | 
|  | 210 | %arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.1 | 
|  | 211 | %val1 = load double addrspace(3)* %arrayidx1, align 8 | 
|  | 212 | %sum = fadd double %val0, %val1 | 
|  | 213 | %out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i | 
|  | 214 | store double %sum, double addrspace(1)* %out.gep, align 8 | 
|  | 215 | ret void | 
|  | 216 | } | 
|  | 217 |  | 
|  | 218 | ; SI-LABEL: @invalid_read2st64_f64_odd_offset | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 219 | ; SI-NOT: ds_read2st64_b64 | 
|  | 220 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 221 | define void @invalid_read2st64_f64_odd_offset(double addrspace(1)* %out, double addrspace(3)* %lds) #0 { | 
|  | 222 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 223 | %add.x.0 = add nsw i32 %x.i, 64 | 
|  | 224 | %arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.0 | 
|  | 225 | %val0 = load double addrspace(3)* %arrayidx0, align 8 | 
|  | 226 | %add.x.1 = add nsw i32 %x.i, 8129 | 
|  | 227 | %arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.1 | 
|  | 228 | %val1 = load double addrspace(3)* %arrayidx1, align 8 | 
|  | 229 | %sum = fadd double %val0, %val1 | 
|  | 230 | %out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i | 
|  | 231 | store double %sum, double addrspace(1)* %out.gep, align 8 | 
|  | 232 | ret void | 
|  | 233 | } | 
|  | 234 |  | 
|  | 235 | ; The stride of 8 elements is 8 * 8 bytes. We need to make sure the | 
|  | 236 | ; stride in elements, not bytes, is a multiple of 64. | 
|  | 237 |  | 
|  | 238 | ; SI-LABEL: @byte_size_only_divisible_64_read2_f64 | 
| Tom Stellard | 326d6ec | 2014-11-05 14:50:53 +0000 | [diff] [blame] | 239 | ; SI-NOT: ds_read2st_b64 | 
|  | 240 | ; SI: ds_read2_b64 v{{\[[0-9]+:[0-9]+\]}}, v{{[0-9]+}} offset0:0 offset1:8 | 
|  | 241 | ; SI: s_endpgm | 
| Matt Arsenault | fe0a2e6 | 2014-10-10 22:12:32 +0000 | [diff] [blame] | 242 | define void @byte_size_only_divisible_64_read2_f64(double addrspace(1)* %out, double addrspace(3)* %lds) #0 { | 
|  | 243 | %x.i = tail call i32 @llvm.r600.read.tidig.x() #1 | 
|  | 244 | %arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %x.i | 
|  | 245 | %val0 = load double addrspace(3)* %arrayidx0, align 8 | 
|  | 246 | %add.x = add nsw i32 %x.i, 8 | 
|  | 247 | %arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x | 
|  | 248 | %val1 = load double addrspace(3)* %arrayidx1, align 8 | 
|  | 249 | %sum = fadd double %val0, %val1 | 
|  | 250 | %out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i | 
|  | 251 | store double %sum, double addrspace(1)* %out.gep, align 4 | 
|  | 252 | ret void | 
|  | 253 | } | 
|  | 254 |  | 
|  | 255 | ; Function Attrs: nounwind readnone | 
|  | 256 | declare i32 @llvm.r600.read.tgid.x() #1 | 
|  | 257 |  | 
|  | 258 | ; Function Attrs: nounwind readnone | 
|  | 259 | declare i32 @llvm.r600.read.tgid.y() #1 | 
|  | 260 |  | 
|  | 261 | ; Function Attrs: nounwind readnone | 
|  | 262 | declare i32 @llvm.r600.read.tidig.x() #1 | 
|  | 263 |  | 
|  | 264 | ; Function Attrs: nounwind readnone | 
|  | 265 | declare i32 @llvm.r600.read.tidig.y() #1 | 
|  | 266 |  | 
|  | 267 | ; Function Attrs: noduplicate nounwind | 
|  | 268 | declare void @llvm.AMDGPU.barrier.local() #2 | 
|  | 269 |  | 
|  | 270 | attributes #0 = { nounwind "less-precise-fpmad"="false" "no-frame-pointer-elim"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "no-realign-stack" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" } | 
|  | 271 | attributes #1 = { nounwind readnone } | 
|  | 272 | attributes #2 = { noduplicate nounwind } |