blob: 97f89be8d2a8b1a04a6c35abdb84179eef2d4e62 [file] [log] [blame]
Tim Northovere0e3aef2013-01-31 12:12:40 +00001; RUN: llc -verify-machineinstrs < %s -march=aarch64 | FileCheck %s
2
3@var32_0 = global i32 0
4@var32_1 = global i32 0
5@var64_0 = global i64 0
6@var64_1 = global i64 0
7
8define void @rorv_i64() {
9; CHECK: rorv_i64:
10 %val0_tmp = load i64* @var64_0
11 %val1_tmp = load i64* @var64_1
12 %val2_tmp = sub i64 64, %val1_tmp
13 %val3_tmp = shl i64 %val0_tmp, %val2_tmp
14 %val4_tmp = lshr i64 %val0_tmp, %val1_tmp
15 %val5_tmp = or i64 %val3_tmp, %val4_tmp
16; CHECK: ror {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}
17 store volatile i64 %val5_tmp, i64* @var64_0
18 ret void
19}
20
21define void @asrv_i64() {
22; CHECK: asrv_i64:
23 %val0_tmp = load i64* @var64_0
24 %val1_tmp = load i64* @var64_1
25 %val4_tmp = ashr i64 %val0_tmp, %val1_tmp
26; CHECK: asr {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}
27 store volatile i64 %val4_tmp, i64* @var64_1
28 ret void
29}
30
31define void @lsrv_i64() {
32; CHECK: lsrv_i64:
33 %val0_tmp = load i64* @var64_0
34 %val1_tmp = load i64* @var64_1
35 %val4_tmp = lshr i64 %val0_tmp, %val1_tmp
36; CHECK: lsr {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}
37 store volatile i64 %val4_tmp, i64* @var64_0
38 ret void
39}
40
41define void @lslv_i64() {
42; CHECK: lslv_i64:
43 %val0_tmp = load i64* @var64_0
44 %val1_tmp = load i64* @var64_1
45 %val4_tmp = shl i64 %val0_tmp, %val1_tmp
46; CHECK: lsl {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}
47 store volatile i64 %val4_tmp, i64* @var64_1
48 ret void
49}
50
51define void @udiv_i64() {
52; CHECK: udiv_i64:
53 %val0_tmp = load i64* @var64_0
54 %val1_tmp = load i64* @var64_1
55 %val4_tmp = udiv i64 %val0_tmp, %val1_tmp
56; CHECK: udiv {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}
57 store volatile i64 %val4_tmp, i64* @var64_0
58 ret void
59}
60
61define void @sdiv_i64() {
62; CHECK: sdiv_i64:
63 %val0_tmp = load i64* @var64_0
64 %val1_tmp = load i64* @var64_1
65 %val4_tmp = sdiv i64 %val0_tmp, %val1_tmp
66; CHECK: sdiv {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}
67 store volatile i64 %val4_tmp, i64* @var64_1
68 ret void
69}
70
71
72define void @lsrv_i32() {
73; CHECK: lsrv_i32:
74 %val0_tmp = load i32* @var32_0
75 %val1_tmp = load i32* @var32_1
76 %val2_tmp = add i32 1, %val1_tmp
77 %val4_tmp = lshr i32 %val0_tmp, %val2_tmp
78; CHECK: lsr {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}
79 store volatile i32 %val4_tmp, i32* @var32_0
80 ret void
81}
82
83define void @lslv_i32() {
84; CHECK: lslv_i32:
85 %val0_tmp = load i32* @var32_0
86 %val1_tmp = load i32* @var32_1
87 %val2_tmp = add i32 1, %val1_tmp
88 %val4_tmp = shl i32 %val0_tmp, %val2_tmp
89; CHECK: lsl {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}
90 store volatile i32 %val4_tmp, i32* @var32_1
91 ret void
92}
93
94define void @rorv_i32() {
95; CHECK: rorv_i32:
96 %val0_tmp = load i32* @var32_0
97 %val6_tmp = load i32* @var32_1
98 %val1_tmp = add i32 1, %val6_tmp
99 %val2_tmp = sub i32 32, %val1_tmp
100 %val3_tmp = shl i32 %val0_tmp, %val2_tmp
101 %val4_tmp = lshr i32 %val0_tmp, %val1_tmp
102 %val5_tmp = or i32 %val3_tmp, %val4_tmp
103; CHECK: ror {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}
104 store volatile i32 %val5_tmp, i32* @var32_0
105 ret void
106}
107
108define void @asrv_i32() {
109; CHECK: asrv_i32:
110 %val0_tmp = load i32* @var32_0
111 %val1_tmp = load i32* @var32_1
112 %val2_tmp = add i32 1, %val1_tmp
113 %val4_tmp = ashr i32 %val0_tmp, %val2_tmp
114; CHECK: asr {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}
115 store volatile i32 %val4_tmp, i32* @var32_1
116 ret void
117}
118
119define void @sdiv_i32() {
120; CHECK: sdiv_i32:
121 %val0_tmp = load i32* @var32_0
122 %val1_tmp = load i32* @var32_1
123 %val4_tmp = sdiv i32 %val0_tmp, %val1_tmp
124; CHECK: sdiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}
125 store volatile i32 %val4_tmp, i32* @var32_1
126 ret void
127}
128
129define void @udiv_i32() {
130; CHECK: udiv_i32:
131 %val0_tmp = load i32* @var32_0
132 %val1_tmp = load i32* @var32_1
133 %val4_tmp = udiv i32 %val0_tmp, %val1_tmp
134; CHECK: udiv {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}
135 store volatile i32 %val4_tmp, i32* @var32_0
136 ret void
137}
138
139; The point of this test is that we may not actually see (shl GPR32:$Val, (zext GPR32:$Val2))
140; in the DAG (the RHS may be natively 64-bit), but we should still use the lsl instructions.
141define i32 @test_lsl32() {
142; CHECK: test_lsl32:
143
144 %val = load i32* @var32_0
145 %ret = shl i32 1, %val
146; CHECK: lsl {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}
147
148 ret i32 %ret
149}
150
151define i32 @test_lsr32() {
152; CHECK: test_lsr32:
153
154 %val = load i32* @var32_0
155 %ret = lshr i32 1, %val
156; CHECK: lsr {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}
157
158 ret i32 %ret
159}
160
161define i32 @test_asr32(i32 %in) {
162; CHECK: test_asr32:
163
164 %val = load i32* @var32_0
165 %ret = ashr i32 %in, %val
166; CHECK: asr {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}
167
168 ret i32 %ret
169}