blob: 2c2ce4c5d35189a2bf1d52f2d6cfa9542db2cf27 [file] [log] [blame]
Jan Vesely70293a02017-02-23 16:12:21 +00001; RUN: llc -march=amdgcn -verify-machineinstrs< %s | FileCheck -check-prefix=GCN -check-prefix=SI %s
2; RUN: llc -march=amdgcn -mcpu=fiji -verify-machineinstrs< %s | FileCheck -check-prefix=GCN -check-prefix=VI %s
Matt Arsenault6f243792013-09-05 19:41:10 +00003; RUN: llc -march=r600 -mcpu=cypress < %s | FileCheck -check-prefix=EG %s
4
Matt Arsenaultabd271b2015-02-05 06:05:13 +00005declare i32 @llvm.r600.read.tidig.x() nounwind readnone
6
Matt Arsenault6f243792013-09-05 19:41:10 +00007define void @trunc_i64_to_i32_store(i32 addrspace(1)* %out, i64 %in) {
Jan Vesely70293a02017-02-23 16:12:21 +00008; GCN-LABEL: {{^}}trunc_i64_to_i32_store:
9; GCN: s_load_dword [[SLOAD:s[0-9]+]], s[0:1],
10; GCN: v_mov_b32_e32 [[VLOAD:v[0-9]+]], [[SLOAD]]
Tom Stellard326d6ec2014-11-05 14:50:53 +000011; SI: buffer_store_dword [[VLOAD]]
Jan Vesely70293a02017-02-23 16:12:21 +000012; VI: flat_store_dword v[{{[0-9:]+}}], [[VLOAD]]
Matt Arsenault6f243792013-09-05 19:41:10 +000013
Tom Stellard79243d92014-10-01 17:15:17 +000014; EG-LABEL: {{^}}trunc_i64_to_i32_store:
Matt Arsenault6f243792013-09-05 19:41:10 +000015; EG: MEM_RAT_CACHELESS STORE_RAW T0.X, T1.X, 1
16; EG: LSHR
17; EG-NEXT: 2(
18
19 %result = trunc i64 %in to i32 store i32 %result, i32 addrspace(1)* %out, align 4
20 ret void
21}
22
Jan Vesely70293a02017-02-23 16:12:21 +000023; GCN-LABEL: {{^}}trunc_load_shl_i64:
24; GCN-DAG: s_load_dwordx2
25; GCN-DAG: s_load_dword [[SREG:s[0-9]+]],
26; GCN: s_lshl_b32 [[SHL:s[0-9]+]], [[SREG]], 2
27; GCN: v_mov_b32_e32 [[VSHL:v[0-9]+]], [[SHL]]
28; SI: buffer_store_dword [[VSHL]]
29; VI: flat_store_dword v[{{[0-9:]+}}], [[VSHL]]
30
Matt Arsenaulta7f1e0c2014-03-24 19:43:31 +000031define void @trunc_load_shl_i64(i32 addrspace(1)* %out, i64 %a) {
32 %b = shl i64 %a, 2
33 %result = trunc i64 %b to i32
34 store i32 %result, i32 addrspace(1)* %out, align 4
35 ret void
36}
37
Jan Vesely70293a02017-02-23 16:12:21 +000038; GCN-LABEL: {{^}}trunc_shl_i64:
Tom Stellard326d6ec2014-11-05 14:50:53 +000039; SI: s_load_dwordx2 s{{\[}}[[LO_SREG:[0-9]+]]:{{[0-9]+\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0xd
Jan Vesely70293a02017-02-23 16:12:21 +000040; VI: s_load_dwordx2 s{{\[}}[[LO_SREG:[0-9]+]]:{{[0-9]+\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0x34
41; GCN: s_lshl_b64 s{{\[}}[[LO_SHL:[0-9]+]]:{{[0-9]+\]}}, s{{\[}}[[LO_SREG]]:{{[0-9]+\]}}, 2
42; GCN: s_add_u32 s[[LO_SREG2:[0-9]+]], s[[LO_SHL]],
43; GCN: v_mov_b32_e32 v[[LO_VREG:[0-9]+]], s[[LO_SREG2]]
44; GCN: s_addc_u32
Konstantin Zhuravlyov0a1a7b62016-11-17 16:41:49 +000045; SI: buffer_store_dword v[[LO_VREG]],
Jan Vesely70293a02017-02-23 16:12:21 +000046; VI: flat_store_dword v[{{[0-9:]+}}], v[[LO_VREG]]
47; GCN: v_mov_b32_e32
48; GCN: v_mov_b32_e32
Matt Arsenaultb517c812014-03-27 17:23:31 +000049define void @trunc_shl_i64(i64 addrspace(1)* %out2, i32 addrspace(1)* %out, i64 %a) {
Matt Arsenaulta7f1e0c2014-03-24 19:43:31 +000050 %aa = add i64 %a, 234 ; Prevent shrinking store.
51 %b = shl i64 %aa, 2
Matt Arsenault204cfa62013-10-10 18:04:16 +000052 %result = trunc i64 %b to i32
53 store i32 %result, i32 addrspace(1)* %out, align 4
Matt Arsenaultb517c812014-03-27 17:23:31 +000054 store i64 %b, i64 addrspace(1)* %out2, align 8 ; Prevent reducing ops to 32-bits
Matt Arsenault204cfa62013-10-10 18:04:16 +000055 ret void
56}
Michel Danzerbf1a6412014-01-28 03:01:16 +000057
Jan Vesely70293a02017-02-23 16:12:21 +000058; GCN-LABEL: {{^}}trunc_i32_to_i1:
59; GCN: v_and_b32_e32 v{{[0-9]+}}, 1, v{{[0-9]+}}
60; GCN: v_cmp_eq_u32
Matt Arsenault49dd4282014-09-15 17:15:02 +000061define void @trunc_i32_to_i1(i32 addrspace(1)* %out, i32 addrspace(1)* %ptr) {
David Blaikiea79ac142015-02-27 21:17:42 +000062 %a = load i32, i32 addrspace(1)* %ptr, align 4
Matt Arsenault49dd4282014-09-15 17:15:02 +000063 %trunc = trunc i32 %a to i1
64 %result = select i1 %trunc, i32 1, i32 0
65 store i32 %result, i32 addrspace(1)* %out, align 4
66 ret void
67}
68
Jan Vesely70293a02017-02-23 16:12:21 +000069; GCN-LABEL: {{^}}trunc_i8_to_i1:
70; GCN: v_and_b32_e32 v{{[0-9]+}}, 1, v{{[0-9]+}}
71; GCN: v_cmp_eq_u32
72define void @trunc_i8_to_i1(i8 addrspace(1)* %out, i8 addrspace(1)* %ptr) {
73 %a = load i8, i8 addrspace(1)* %ptr, align 4
74 %trunc = trunc i8 %a to i1
75 %result = select i1 %trunc, i8 1, i8 0
76 store i8 %result, i8 addrspace(1)* %out, align 4
77 ret void
78}
79
80; GCN-LABEL: {{^}}sgpr_trunc_i16_to_i1:
81; GCN: s_and_b32 s{{[0-9]+}}, 1, s{{[0-9]+}}
82; GCN: v_cmp_eq_u32
83define void @sgpr_trunc_i16_to_i1(i16 addrspace(1)* %out, i16 %a) {
84 %trunc = trunc i16 %a to i1
85 %result = select i1 %trunc, i16 1, i16 0
86 store i16 %result, i16 addrspace(1)* %out, align 4
87 ret void
88}
89
90; GCN-LABEL: {{^}}sgpr_trunc_i32_to_i1:
91; GCN: s_and_b32 s{{[0-9]+}}, 1, s{{[0-9]+}}
92; GCN: v_cmp_eq_u32
Matt Arsenault49dd4282014-09-15 17:15:02 +000093define void @sgpr_trunc_i32_to_i1(i32 addrspace(1)* %out, i32 %a) {
Michel Danzerbf1a6412014-01-28 03:01:16 +000094 %trunc = trunc i32 %a to i1
95 %result = select i1 %trunc, i32 1, i32 0
96 store i32 %result, i32 addrspace(1)* %out, align 4
97 ret void
98}
Matt Arsenaultabd271b2015-02-05 06:05:13 +000099
Jan Vesely70293a02017-02-23 16:12:21 +0000100; GCN-LABEL: {{^}}s_trunc_i64_to_i1:
Matt Arsenaultabd271b2015-02-05 06:05:13 +0000101; SI: s_load_dwordx2 s{{\[}}[[SLO:[0-9]+]]:{{[0-9]+\]}}, {{s\[[0-9]+:[0-9]+\]}}, 0xb
Jan Vesely70293a02017-02-23 16:12:21 +0000102; VI: s_load_dwordx2 s{{\[}}[[SLO:[0-9]+]]:{{[0-9]+\]}}, {{s\[[0-9]+:[0-9]+\]}}, 0x2c
103; GCN: s_and_b32 [[MASKED:s[0-9]+]], 1, s[[SLO]]
104; GCN: v_cmp_eq_u32_e64 s{{\[}}[[VLO:[0-9]+]]:[[VHI:[0-9]+]]], [[MASKED]], 1{{$}}
105; GCN: v_cndmask_b32_e64 {{v[0-9]+}}, -12, 63, s{{\[}}[[VLO]]:[[VHI]]]
Matt Arsenaultabd271b2015-02-05 06:05:13 +0000106define void @s_trunc_i64_to_i1(i32 addrspace(1)* %out, i64 %x) {
107 %trunc = trunc i64 %x to i1
108 %sel = select i1 %trunc, i32 63, i32 -12
109 store i32 %sel, i32 addrspace(1)* %out
110 ret void
111}
112
Jan Vesely70293a02017-02-23 16:12:21 +0000113; GCN-LABEL: {{^}}v_trunc_i64_to_i1:
Matt Arsenaultabd271b2015-02-05 06:05:13 +0000114; SI: buffer_load_dwordx2 v{{\[}}[[VLO:[0-9]+]]:{{[0-9]+\]}}
Jan Vesely70293a02017-02-23 16:12:21 +0000115; VI: flat_load_dwordx2 v{{\[}}[[VLO:[0-9]+]]:{{[0-9]+\]}}
116; GCN: v_and_b32_e32 [[MASKED:v[0-9]+]], 1, v[[VLO]]
117; GCN: v_cmp_eq_u32_e32 vcc, 1, [[MASKED]]
118; GCN: v_cndmask_b32_e64 {{v[0-9]+}}, -12, 63, vcc
Matt Arsenaultabd271b2015-02-05 06:05:13 +0000119define void @v_trunc_i64_to_i1(i32 addrspace(1)* %out, i64 addrspace(1)* %in) {
120 %tid = call i32 @llvm.r600.read.tidig.x() nounwind readnone
David Blaikie79e6c742015-02-27 19:29:02 +0000121 %gep = getelementptr i64, i64 addrspace(1)* %in, i32 %tid
122 %out.gep = getelementptr i32, i32 addrspace(1)* %out, i32 %tid
David Blaikiea79ac142015-02-27 21:17:42 +0000123 %x = load i64, i64 addrspace(1)* %gep
Matt Arsenaultabd271b2015-02-05 06:05:13 +0000124
125 %trunc = trunc i64 %x to i1
126 %sel = select i1 %trunc, i32 63, i32 -12
127 store i32 %sel, i32 addrspace(1)* %out.gep
128 ret void
129}