blob: 3a851d68f0a4fe9450f7e18ac698afb551d1ac9c [file] [log] [blame]
Saleem Abdulrasool72587352014-04-03 16:01:44 +00001; RUN: llc -mtriple=arm-eabi -mattr=+neon %s -o - | FileCheck %s
Sebastian Popa204f722012-11-30 19:08:04 +00002
3; PR12281
4; Test generataion of code for vmull instruction when multiplying 128-bit
5; vectors that were created by sign-extending smaller vector sizes.
6;
7; The vmull operation requires 64-bit vectors, so we must extend the original
8; vector size to 64 bits for vmull operation.
9; Previously failed with an assertion because the <4 x i8> vector was too small
10; for vmull.
11
12; Vector x Constant
13; v4i8
14;
15define void @sextload_v4i8_c(<4 x i8>* %v) nounwind {
Stephen Lind24ab202013-07-14 06:24:09 +000016;CHECK-LABEL: sextload_v4i8_c:
Sebastian Popa204f722012-11-30 19:08:04 +000017entry:
David Blaikiea79ac142015-02-27 21:17:42 +000018 %0 = load <4 x i8>, <4 x i8>* %v, align 8
Sebastian Popa204f722012-11-30 19:08:04 +000019 %v0 = sext <4 x i8> %0 to <4 x i32>
20;CHECK: vmull
21 %v1 = mul <4 x i32> %v0, <i32 3, i32 3, i32 3, i32 3>
22 store <4 x i32> %v1, <4 x i32>* undef, align 8
23 ret void;
24}
25
26; v2i8
27;
28define void @sextload_v2i8_c(<2 x i8>* %v) nounwind {
Stephen Lind24ab202013-07-14 06:24:09 +000029;CHECK-LABEL: sextload_v2i8_c:
Sebastian Popa204f722012-11-30 19:08:04 +000030entry:
David Blaikiea79ac142015-02-27 21:17:42 +000031 %0 = load <2 x i8>, <2 x i8>* %v, align 8
Sebastian Popa204f722012-11-30 19:08:04 +000032 %v0 = sext <2 x i8> %0 to <2 x i64>
33;CHECK: vmull
34 %v1 = mul <2 x i64> %v0, <i64 3, i64 3>
35 store <2 x i64> %v1, <2 x i64>* undef, align 8
36 ret void;
37}
38
39; v2i16
40;
41define void @sextload_v2i16_c(<2 x i16>* %v) nounwind {
Stephen Lind24ab202013-07-14 06:24:09 +000042;CHECK-LABEL: sextload_v2i16_c:
Sebastian Popa204f722012-11-30 19:08:04 +000043entry:
David Blaikiea79ac142015-02-27 21:17:42 +000044 %0 = load <2 x i16>, <2 x i16>* %v, align 8
Sebastian Popa204f722012-11-30 19:08:04 +000045 %v0 = sext <2 x i16> %0 to <2 x i64>
46;CHECK: vmull
47 %v1 = mul <2 x i64> %v0, <i64 3, i64 3>
48 store <2 x i64> %v1, <2 x i64>* undef, align 8
49 ret void;
50}
51
52
53; Vector x Vector
54; v4i8
55;
56define void @sextload_v4i8_v(<4 x i8>* %v, <4 x i8>* %p) nounwind {
Stephen Lind24ab202013-07-14 06:24:09 +000057;CHECK-LABEL: sextload_v4i8_v:
Sebastian Popa204f722012-11-30 19:08:04 +000058entry:
David Blaikiea79ac142015-02-27 21:17:42 +000059 %0 = load <4 x i8>, <4 x i8>* %v, align 8
Sebastian Popa204f722012-11-30 19:08:04 +000060 %v0 = sext <4 x i8> %0 to <4 x i32>
61
David Blaikiea79ac142015-02-27 21:17:42 +000062 %1 = load <4 x i8>, <4 x i8>* %p, align 8
Sebastian Popa204f722012-11-30 19:08:04 +000063 %v2 = sext <4 x i8> %1 to <4 x i32>
64;CHECK: vmull
65 %v1 = mul <4 x i32> %v0, %v2
66 store <4 x i32> %v1, <4 x i32>* undef, align 8
67 ret void;
68}
69
70; v2i8
71;
72define void @sextload_v2i8_v(<2 x i8>* %v, <2 x i8>* %p) nounwind {
Stephen Lind24ab202013-07-14 06:24:09 +000073;CHECK-LABEL: sextload_v2i8_v:
Sebastian Popa204f722012-11-30 19:08:04 +000074entry:
David Blaikiea79ac142015-02-27 21:17:42 +000075 %0 = load <2 x i8>, <2 x i8>* %v, align 8
Sebastian Popa204f722012-11-30 19:08:04 +000076 %v0 = sext <2 x i8> %0 to <2 x i64>
77
David Blaikiea79ac142015-02-27 21:17:42 +000078 %1 = load <2 x i8>, <2 x i8>* %p, align 8
Sebastian Popa204f722012-11-30 19:08:04 +000079 %v2 = sext <2 x i8> %1 to <2 x i64>
80;CHECK: vmull
81 %v1 = mul <2 x i64> %v0, %v2
82 store <2 x i64> %v1, <2 x i64>* undef, align 8
83 ret void;
84}
85
86; v2i16
87;
88define void @sextload_v2i16_v(<2 x i16>* %v, <2 x i16>* %p) nounwind {
Stephen Lind24ab202013-07-14 06:24:09 +000089;CHECK-LABEL: sextload_v2i16_v:
Sebastian Popa204f722012-11-30 19:08:04 +000090entry:
David Blaikiea79ac142015-02-27 21:17:42 +000091 %0 = load <2 x i16>, <2 x i16>* %v, align 8
Sebastian Popa204f722012-11-30 19:08:04 +000092 %v0 = sext <2 x i16> %0 to <2 x i64>
93
David Blaikiea79ac142015-02-27 21:17:42 +000094 %1 = load <2 x i16>, <2 x i16>* %p, align 8
Sebastian Popa204f722012-11-30 19:08:04 +000095 %v2 = sext <2 x i16> %1 to <2 x i64>
96;CHECK: vmull
97 %v1 = mul <2 x i64> %v0, %v2
98 store <2 x i64> %v1, <2 x i64>* undef, align 8
99 ret void;
100}
101
102
103; Vector(small) x Vector(big)
104; v4i8 x v4i16
105;
106define void @sextload_v4i8_vs(<4 x i8>* %v, <4 x i16>* %p) nounwind {
Stephen Lind24ab202013-07-14 06:24:09 +0000107;CHECK-LABEL: sextload_v4i8_vs:
Sebastian Popa204f722012-11-30 19:08:04 +0000108entry:
David Blaikiea79ac142015-02-27 21:17:42 +0000109 %0 = load <4 x i8>, <4 x i8>* %v, align 8
Sebastian Popa204f722012-11-30 19:08:04 +0000110 %v0 = sext <4 x i8> %0 to <4 x i32>
111
David Blaikiea79ac142015-02-27 21:17:42 +0000112 %1 = load <4 x i16>, <4 x i16>* %p, align 8
Sebastian Popa204f722012-11-30 19:08:04 +0000113 %v2 = sext <4 x i16> %1 to <4 x i32>
114;CHECK: vmull
115 %v1 = mul <4 x i32> %v0, %v2
116 store <4 x i32> %v1, <4 x i32>* undef, align 8
117 ret void;
118}
119
120; v2i8
121; v2i8 x v2i16
122define void @sextload_v2i8_vs(<2 x i8>* %v, <2 x i16>* %p) nounwind {
Stephen Lind24ab202013-07-14 06:24:09 +0000123;CHECK-LABEL: sextload_v2i8_vs:
Sebastian Popa204f722012-11-30 19:08:04 +0000124entry:
David Blaikiea79ac142015-02-27 21:17:42 +0000125 %0 = load <2 x i8>, <2 x i8>* %v, align 8
Sebastian Popa204f722012-11-30 19:08:04 +0000126 %v0 = sext <2 x i8> %0 to <2 x i64>
127
David Blaikiea79ac142015-02-27 21:17:42 +0000128 %1 = load <2 x i16>, <2 x i16>* %p, align 8
Sebastian Popa204f722012-11-30 19:08:04 +0000129 %v2 = sext <2 x i16> %1 to <2 x i64>
130;CHECK: vmull
131 %v1 = mul <2 x i64> %v0, %v2
132 store <2 x i64> %v1, <2 x i64>* undef, align 8
133 ret void;
134}
135
136; v2i16
137; v2i16 x v2i32
138define void @sextload_v2i16_vs(<2 x i16>* %v, <2 x i32>* %p) nounwind {
Stephen Lind24ab202013-07-14 06:24:09 +0000139;CHECK-LABEL: sextload_v2i16_vs:
Sebastian Popa204f722012-11-30 19:08:04 +0000140entry:
David Blaikiea79ac142015-02-27 21:17:42 +0000141 %0 = load <2 x i16>, <2 x i16>* %v, align 8
Sebastian Popa204f722012-11-30 19:08:04 +0000142 %v0 = sext <2 x i16> %0 to <2 x i64>
143
David Blaikiea79ac142015-02-27 21:17:42 +0000144 %1 = load <2 x i32>, <2 x i32>* %p, align 8
Sebastian Popa204f722012-11-30 19:08:04 +0000145 %v2 = sext <2 x i32> %1 to <2 x i64>
146;CHECK: vmull
147 %v1 = mul <2 x i64> %v0, %v2
148 store <2 x i64> %v1, <2 x i64>* undef, align 8
149 ret void;
150}