blob: 37111ef0d89beb820a647f8c414dee83a08372e4 [file] [log] [blame]
Nemanja Ivanovicea1db8a2015-06-11 06:21:25 +00001; RUN: llc -mtriple=powerpc64-unknown-linux-gnu -mcpu=pwr8 -mattr=-vsx < %s | FileCheck %s
2; RUN: llc -mtriple=powerpc64le-unknown-linux-gnu -mcpu=pwr8 -mattr=-vsx < %s | FileCheck %s
3; RUN: llc -mtriple=powerpc64-unknown-linux-gnu -mcpu=pwr7 -mattr=+power8-vector -mattr=-vsx < %s | FileCheck %s
4; RUN: llc -mtriple=powerpc64le-unknown-linux-gnu -mcpu=pwr8 < %s | FileCheck %s -check-prefix=CHECK-VSX
5
6@vsc = global <16 x i8> <i8 0, i8 1, i8 2, i8 3, i8 4, i8 5, i8 6, i8 7, i8 8, i8 9, i8 0, i8 1, i8 2, i8 3, i8 4, i8 5>, align 16
7@vuc = global <16 x i8> <i8 0, i8 1, i8 2, i8 3, i8 4, i8 5, i8 6, i8 7, i8 8, i8 9, i8 0, i8 1, i8 2, i8 3, i8 4, i8 5>, align 16
8@res_vll = common global <2 x i64> zeroinitializer, align 16
9@res_vull = common global <2 x i64> zeroinitializer, align 16
10@res_vsc = common global <16 x i8> zeroinitializer, align 16
11@res_vuc = common global <16 x i8> zeroinitializer, align 16
12
13; Function Attrs: nounwind
14define void @test1() {
15entry:
16 %__a.addr.i = alloca <16 x i8>, align 16
17 %__b.addr.i = alloca <16 x i8>, align 16
18 %0 = load <16 x i8>, <16 x i8>* @vsc, align 16
19 %1 = load <16 x i8>, <16 x i8>* @vsc, align 16
20 store <16 x i8> %0, <16 x i8>* %__a.addr.i, align 16
21 store <16 x i8> %1, <16 x i8>* %__b.addr.i, align 16
22 %2 = load <16 x i8>, <16 x i8>* %__a.addr.i, align 16
23 %3 = load <16 x i8>, <16 x i8>* %__b.addr.i, align 16
24 %4 = call <2 x i64> @llvm.ppc.altivec.vbpermq(<16 x i8> %2, <16 x i8> %3)
25 store <2 x i64> %4, <2 x i64>* @res_vll, align 16
26 ret void
27; CHECK-LABEL: @test1
28; CHECK: lvx [[REG1:[0-9]+]],
29; CHECK: lvx [[REG2:[0-9]+]],
30; CHECK: vbpermq {{[0-9]+}}, [[REG2]], [[REG1]]
31; CHECK-VSX: vbpermq {{[0-9]+}}, {{[0-9]+}}, {{[0-9]+}}
32}
33
34; Function Attrs: nounwind
35define void @test2() {
36entry:
37 %__a.addr.i = alloca <16 x i8>, align 16
38 %__b.addr.i = alloca <16 x i8>, align 16
39 %0 = load <16 x i8>, <16 x i8>* @vuc, align 16
40 %1 = load <16 x i8>, <16 x i8>* @vuc, align 16
41 store <16 x i8> %0, <16 x i8>* %__a.addr.i, align 16
42 store <16 x i8> %1, <16 x i8>* %__b.addr.i, align 16
43 %2 = load <16 x i8>, <16 x i8>* %__a.addr.i, align 16
44 %3 = load <16 x i8>, <16 x i8>* %__b.addr.i, align 16
45 %4 = call <2 x i64> @llvm.ppc.altivec.vbpermq(<16 x i8> %2, <16 x i8> %3)
46 store <2 x i64> %4, <2 x i64>* @res_vull, align 16
47 ret void
48; CHECK-LABEL: @test2
49; CHECK: lvx [[REG1:[0-9]+]],
50; CHECK: lvx [[REG2:[0-9]+]],
51; CHECK: vbpermq {{[0-9]+}}, [[REG2]], [[REG1]]
52; CHECK-VSX: vbpermq {{[0-9]+}}, {{[0-9]+}}, {{[0-9]+}}
53}
54
55; Function Attrs: nounwind
56define void @test3() {
57entry:
58 %__a.addr.i = alloca <16 x i8>, align 16
59 %0 = load <16 x i8>, <16 x i8>* @vsc, align 16
60 store <16 x i8> %0, <16 x i8>* %__a.addr.i, align 16
61 %1 = load <16 x i8>, <16 x i8>* %__a.addr.i, align 16
62 %2 = call <16 x i8> @llvm.ppc.altivec.vgbbd(<16 x i8> %1)
63 store <16 x i8> %2, <16 x i8>* @res_vsc, align 16
64 ret void
65; CHECK-LABEL: @test3
66; CHECK: lvx [[REG1:[0-9]+]],
67; CHECK: vgbbd {{[0-9]+}}, [[REG1]]
68; CHECK-VSX: vgbbd {{[0-9]+}}, {{[0-9]+}}
69}
70
71; Function Attrs: nounwind
72define void @test4() {
73entry:
74 %__a.addr.i = alloca <16 x i8>, align 16
75 %0 = load <16 x i8>, <16 x i8>* @vuc, align 16
76 store <16 x i8> %0, <16 x i8>* %__a.addr.i, align 16
77 %1 = load <16 x i8>, <16 x i8>* %__a.addr.i, align 16
78 %2 = call <16 x i8> @llvm.ppc.altivec.vgbbd(<16 x i8> %1)
79 store <16 x i8> %2, <16 x i8>* @res_vuc, align 16
80 ret void
81; CHECK-LABEL: @test4
82; CHECK: lvx [[REG1:[0-9]+]],
83; CHECK: vgbbd {{[0-9]+}}, [[REG1]]
84; CHECK-VSX: vgbbd {{[0-9]+}}, {{[0-9]+}}
85}
86
87; Function Attrs: nounwind readnone
88declare <2 x i64> @llvm.ppc.altivec.vbpermq(<16 x i8>, <16 x i8>)
89
90; Function Attrs: nounwind readnone
91declare <16 x i8> @llvm.ppc.altivec.vgbbd(<16 x i8>)