blob: 2b70a7878ced5d9dbadf70ff95c28948435f11f2 [file] [log] [blame]
Bob Wilsonb2cae812009-10-07 22:30:19 +00001; RUN: llc < %s -march=arm -mattr=+neon | FileCheck %s
Bob Wilson5bafff32009-06-22 23:27:02 +00002
3define <8 x i8> @vmlsi8(<8 x i8>* %A, <8 x i8>* %B, <8 x i8> * %C) nounwind {
Bob Wilsonb2cae812009-10-07 22:30:19 +00004;CHECK: vmlsi8:
5;CHECK: vmls.i8
Bob Wilson5bafff32009-06-22 23:27:02 +00006 %tmp1 = load <8 x i8>* %A
7 %tmp2 = load <8 x i8>* %B
8 %tmp3 = load <8 x i8>* %C
9 %tmp4 = mul <8 x i8> %tmp2, %tmp3
10 %tmp5 = sub <8 x i8> %tmp1, %tmp4
11 ret <8 x i8> %tmp5
12}
13
14define <4 x i16> @vmlsi16(<4 x i16>* %A, <4 x i16>* %B, <4 x i16>* %C) nounwind {
Bob Wilsonb2cae812009-10-07 22:30:19 +000015;CHECK: vmlsi16:
16;CHECK: vmls.i16
Bob Wilson5bafff32009-06-22 23:27:02 +000017 %tmp1 = load <4 x i16>* %A
18 %tmp2 = load <4 x i16>* %B
19 %tmp3 = load <4 x i16>* %C
20 %tmp4 = mul <4 x i16> %tmp2, %tmp3
21 %tmp5 = sub <4 x i16> %tmp1, %tmp4
22 ret <4 x i16> %tmp5
23}
24
25define <2 x i32> @vmlsi32(<2 x i32>* %A, <2 x i32>* %B, <2 x i32>* %C) nounwind {
Bob Wilsonb2cae812009-10-07 22:30:19 +000026;CHECK: vmlsi32:
27;CHECK: vmls.i32
Bob Wilson5bafff32009-06-22 23:27:02 +000028 %tmp1 = load <2 x i32>* %A
29 %tmp2 = load <2 x i32>* %B
30 %tmp3 = load <2 x i32>* %C
31 %tmp4 = mul <2 x i32> %tmp2, %tmp3
32 %tmp5 = sub <2 x i32> %tmp1, %tmp4
33 ret <2 x i32> %tmp5
34}
35
36define <2 x float> @vmlsf32(<2 x float>* %A, <2 x float>* %B, <2 x float>* %C) nounwind {
Bob Wilsonb2cae812009-10-07 22:30:19 +000037;CHECK: vmlsf32:
38;CHECK: vmls.f32
Bob Wilson5bafff32009-06-22 23:27:02 +000039 %tmp1 = load <2 x float>* %A
40 %tmp2 = load <2 x float>* %B
41 %tmp3 = load <2 x float>* %C
Dan Gohmand4d01152010-05-03 22:36:46 +000042 %tmp4 = fmul <2 x float> %tmp2, %tmp3
43 %tmp5 = fsub <2 x float> %tmp1, %tmp4
Bob Wilson5bafff32009-06-22 23:27:02 +000044 ret <2 x float> %tmp5
45}
46
47define <16 x i8> @vmlsQi8(<16 x i8>* %A, <16 x i8>* %B, <16 x i8> * %C) nounwind {
Bob Wilsonb2cae812009-10-07 22:30:19 +000048;CHECK: vmlsQi8:
49;CHECK: vmls.i8
Bob Wilson5bafff32009-06-22 23:27:02 +000050 %tmp1 = load <16 x i8>* %A
51 %tmp2 = load <16 x i8>* %B
52 %tmp3 = load <16 x i8>* %C
53 %tmp4 = mul <16 x i8> %tmp2, %tmp3
54 %tmp5 = sub <16 x i8> %tmp1, %tmp4
55 ret <16 x i8> %tmp5
56}
57
58define <8 x i16> @vmlsQi16(<8 x i16>* %A, <8 x i16>* %B, <8 x i16>* %C) nounwind {
Bob Wilsonb2cae812009-10-07 22:30:19 +000059;CHECK: vmlsQi16:
60;CHECK: vmls.i16
Bob Wilson5bafff32009-06-22 23:27:02 +000061 %tmp1 = load <8 x i16>* %A
62 %tmp2 = load <8 x i16>* %B
63 %tmp3 = load <8 x i16>* %C
64 %tmp4 = mul <8 x i16> %tmp2, %tmp3
65 %tmp5 = sub <8 x i16> %tmp1, %tmp4
66 ret <8 x i16> %tmp5
67}
68
69define <4 x i32> @vmlsQi32(<4 x i32>* %A, <4 x i32>* %B, <4 x i32>* %C) nounwind {
Bob Wilsonb2cae812009-10-07 22:30:19 +000070;CHECK: vmlsQi32:
71;CHECK: vmls.i32
Bob Wilson5bafff32009-06-22 23:27:02 +000072 %tmp1 = load <4 x i32>* %A
73 %tmp2 = load <4 x i32>* %B
74 %tmp3 = load <4 x i32>* %C
75 %tmp4 = mul <4 x i32> %tmp2, %tmp3
76 %tmp5 = sub <4 x i32> %tmp1, %tmp4
77 ret <4 x i32> %tmp5
78}
79
80define <4 x float> @vmlsQf32(<4 x float>* %A, <4 x float>* %B, <4 x float>* %C) nounwind {
Bob Wilsonb2cae812009-10-07 22:30:19 +000081;CHECK: vmlsQf32:
82;CHECK: vmls.f32
Bob Wilson5bafff32009-06-22 23:27:02 +000083 %tmp1 = load <4 x float>* %A
84 %tmp2 = load <4 x float>* %B
85 %tmp3 = load <4 x float>* %C
Dan Gohmand4d01152010-05-03 22:36:46 +000086 %tmp4 = fmul <4 x float> %tmp2, %tmp3
87 %tmp5 = fsub <4 x float> %tmp1, %tmp4
Bob Wilson5bafff32009-06-22 23:27:02 +000088 ret <4 x float> %tmp5
89}
Bob Wilson83815ae2009-10-09 20:20:54 +000090
91define <8 x i16> @vmlsls8(<8 x i16>* %A, <8 x i8>* %B, <8 x i8>* %C) nounwind {
92;CHECK: vmlsls8:
93;CHECK: vmlsl.s8
94 %tmp1 = load <8 x i16>* %A
95 %tmp2 = load <8 x i8>* %B
96 %tmp3 = load <8 x i8>* %C
97 %tmp4 = call <8 x i16> @llvm.arm.neon.vmlsls.v8i16(<8 x i16> %tmp1, <8 x i8> %tmp2, <8 x i8> %tmp3)
98 ret <8 x i16> %tmp4
99}
100
101define <4 x i32> @vmlsls16(<4 x i32>* %A, <4 x i16>* %B, <4 x i16>* %C) nounwind {
102;CHECK: vmlsls16:
103;CHECK: vmlsl.s16
104 %tmp1 = load <4 x i32>* %A
105 %tmp2 = load <4 x i16>* %B
106 %tmp3 = load <4 x i16>* %C
107 %tmp4 = call <4 x i32> @llvm.arm.neon.vmlsls.v4i32(<4 x i32> %tmp1, <4 x i16> %tmp2, <4 x i16> %tmp3)
108 ret <4 x i32> %tmp4
109}
110
111define <2 x i64> @vmlsls32(<2 x i64>* %A, <2 x i32>* %B, <2 x i32>* %C) nounwind {
112;CHECK: vmlsls32:
113;CHECK: vmlsl.s32
114 %tmp1 = load <2 x i64>* %A
115 %tmp2 = load <2 x i32>* %B
116 %tmp3 = load <2 x i32>* %C
117 %tmp4 = call <2 x i64> @llvm.arm.neon.vmlsls.v2i64(<2 x i64> %tmp1, <2 x i32> %tmp2, <2 x i32> %tmp3)
118 ret <2 x i64> %tmp4
119}
120
121define <8 x i16> @vmlslu8(<8 x i16>* %A, <8 x i8>* %B, <8 x i8>* %C) nounwind {
122;CHECK: vmlslu8:
123;CHECK: vmlsl.u8
124 %tmp1 = load <8 x i16>* %A
125 %tmp2 = load <8 x i8>* %B
126 %tmp3 = load <8 x i8>* %C
127 %tmp4 = call <8 x i16> @llvm.arm.neon.vmlslu.v8i16(<8 x i16> %tmp1, <8 x i8> %tmp2, <8 x i8> %tmp3)
128 ret <8 x i16> %tmp4
129}
130
131define <4 x i32> @vmlslu16(<4 x i32>* %A, <4 x i16>* %B, <4 x i16>* %C) nounwind {
132;CHECK: vmlslu16:
133;CHECK: vmlsl.u16
134 %tmp1 = load <4 x i32>* %A
135 %tmp2 = load <4 x i16>* %B
136 %tmp3 = load <4 x i16>* %C
137 %tmp4 = call <4 x i32> @llvm.arm.neon.vmlslu.v4i32(<4 x i32> %tmp1, <4 x i16> %tmp2, <4 x i16> %tmp3)
138 ret <4 x i32> %tmp4
139}
140
141define <2 x i64> @vmlslu32(<2 x i64>* %A, <2 x i32>* %B, <2 x i32>* %C) nounwind {
142;CHECK: vmlslu32:
143;CHECK: vmlsl.u32
144 %tmp1 = load <2 x i64>* %A
145 %tmp2 = load <2 x i32>* %B
146 %tmp3 = load <2 x i32>* %C
147 %tmp4 = call <2 x i64> @llvm.arm.neon.vmlslu.v2i64(<2 x i64> %tmp1, <2 x i32> %tmp2, <2 x i32> %tmp3)
148 ret <2 x i64> %tmp4
149}
150
151define arm_aapcs_vfpcc <4 x i32> @test_vmlsl_lanes16(<4 x i32> %arg0_int32x4_t, <4 x i16> %arg1_int16x4_t, <4 x i16> %arg2_int16x4_t) nounwind readnone {
152entry:
153; CHECK: test_vmlsl_lanes16
154; CHECK: vmlsl.s16 q0, d2, d3[1]
155 %0 = shufflevector <4 x i16> %arg2_int16x4_t, <4 x i16> undef, <4 x i32> <i32 1, i32 1, i32 1, i32 1> ; <<4 x i16>> [#uses=1]
156 %1 = tail call <4 x i32> @llvm.arm.neon.vmlsls.v4i32(<4 x i32> %arg0_int32x4_t, <4 x i16> %arg1_int16x4_t, <4 x i16> %0) ; <<4 x i32>> [#uses=1]
157 ret <4 x i32> %1
158}
159
160define arm_aapcs_vfpcc <2 x i64> @test_vmlsl_lanes32(<2 x i64> %arg0_int64x2_t, <2 x i32> %arg1_int32x2_t, <2 x i32> %arg2_int32x2_t) nounwind readnone {
161entry:
162; CHECK: test_vmlsl_lanes32
163; CHECK: vmlsl.s32 q0, d2, d3[1]
164 %0 = shufflevector <2 x i32> %arg2_int32x2_t, <2 x i32> undef, <2 x i32> <i32 1, i32 1> ; <<2 x i32>> [#uses=1]
165 %1 = tail call <2 x i64> @llvm.arm.neon.vmlsls.v2i64(<2 x i64> %arg0_int64x2_t, <2 x i32> %arg1_int32x2_t, <2 x i32> %0) ; <<2 x i64>> [#uses=1]
166 ret <2 x i64> %1
167}
168
169define arm_aapcs_vfpcc <4 x i32> @test_vmlsl_laneu16(<4 x i32> %arg0_uint32x4_t, <4 x i16> %arg1_uint16x4_t, <4 x i16> %arg2_uint16x4_t) nounwind readnone {
170entry:
171; CHECK: test_vmlsl_laneu16
172; CHECK: vmlsl.u16 q0, d2, d3[1]
173 %0 = shufflevector <4 x i16> %arg2_uint16x4_t, <4 x i16> undef, <4 x i32> <i32 1, i32 1, i32 1, i32 1> ; <<4 x i16>> [#uses=1]
174 %1 = tail call <4 x i32> @llvm.arm.neon.vmlslu.v4i32(<4 x i32> %arg0_uint32x4_t, <4 x i16> %arg1_uint16x4_t, <4 x i16> %0) ; <<4 x i32>> [#uses=1]
175 ret <4 x i32> %1
176}
177
178define arm_aapcs_vfpcc <2 x i64> @test_vmlsl_laneu32(<2 x i64> %arg0_uint64x2_t, <2 x i32> %arg1_uint32x2_t, <2 x i32> %arg2_uint32x2_t) nounwind readnone {
179entry:
180; CHECK: test_vmlsl_laneu32
181; CHECK: vmlsl.u32 q0, d2, d3[1]
182 %0 = shufflevector <2 x i32> %arg2_uint32x2_t, <2 x i32> undef, <2 x i32> <i32 1, i32 1> ; <<2 x i32>> [#uses=1]
183 %1 = tail call <2 x i64> @llvm.arm.neon.vmlslu.v2i64(<2 x i64> %arg0_uint64x2_t, <2 x i32> %arg1_uint32x2_t, <2 x i32> %0) ; <<2 x i64>> [#uses=1]
184 ret <2 x i64> %1
185}
186
187declare <8 x i16> @llvm.arm.neon.vmlsls.v8i16(<8 x i16>, <8 x i8>, <8 x i8>) nounwind readnone
188declare <4 x i32> @llvm.arm.neon.vmlsls.v4i32(<4 x i32>, <4 x i16>, <4 x i16>) nounwind readnone
189declare <2 x i64> @llvm.arm.neon.vmlsls.v2i64(<2 x i64>, <2 x i32>, <2 x i32>) nounwind readnone
190
191declare <8 x i16> @llvm.arm.neon.vmlslu.v8i16(<8 x i16>, <8 x i8>, <8 x i8>) nounwind readnone
192declare <4 x i32> @llvm.arm.neon.vmlslu.v4i32(<4 x i32>, <4 x i16>, <4 x i16>) nounwind readnone
193declare <2 x i64> @llvm.arm.neon.vmlslu.v2i64(<2 x i64>, <2 x i32>, <2 x i32>) nounwind readnone