blob: e458d7ab621567c8413b41b613e857bbae9bbe82 [file] [log] [blame]
Simon Pilgrim153b4082016-09-09 13:31:52 +00001; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
2; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx | FileCheck %s --check-prefix=X32
3; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx | FileCheck %s --check-prefix=X64
4
5define void @knownbits_zext_in_reg(i8*) nounwind {
6; X32-LABEL: knownbits_zext_in_reg:
7; X32: # BB#0: # %BB
8; X32-NEXT: pushl %ebp
9; X32-NEXT: pushl %ebx
10; X32-NEXT: pushl %edi
11; X32-NEXT: pushl %esi
12; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
13; X32-NEXT: movzbl (%eax), %eax
14; X32-NEXT: imull $101, %eax, %eax
15; X32-NEXT: andl $16384, %eax # imm = 0x4000
16; X32-NEXT: shrl $14, %eax
17; X32-NEXT: movzbl %al, %eax
18; X32-NEXT: vmovd %eax, %xmm0
Simon Pilgrim1e4d8702016-12-01 15:41:40 +000019; X32-NEXT: vpshufb {{.*#+}} xmm0 = zero,zero,zero,zero,xmm0[0],zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero
Simon Pilgrim153b4082016-09-09 13:31:52 +000020; X32-NEXT: vpextrd $1, %xmm0, %ebp
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000021; X32-NEXT: xorl %ecx, %ecx
Simon Pilgrim153b4082016-09-09 13:31:52 +000022; X32-NEXT: vmovd %xmm0, %esi
23; X32-NEXT: vpextrd $2, %xmm0, %edi
24; X32-NEXT: vpextrd $3, %xmm0, %ebx
Simon Pilgrim153b4082016-09-09 13:31:52 +000025; X32-NEXT: .p2align 4, 0x90
26; X32-NEXT: .LBB0_1: # %CF
27; X32-NEXT: # =>This Loop Header: Depth=1
28; X32-NEXT: # Child Loop BB0_2 Depth 2
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000029; X32-NEXT: xorl %edx, %edx
Simon Pilgrim153b4082016-09-09 13:31:52 +000030; X32-NEXT: movl %ebp, %eax
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000031; X32-NEXT: divl %ebp
32; X32-NEXT: xorl %edx, %edx
Simon Pilgrim153b4082016-09-09 13:31:52 +000033; X32-NEXT: movl %esi, %eax
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000034; X32-NEXT: divl %esi
35; X32-NEXT: xorl %edx, %edx
Simon Pilgrim153b4082016-09-09 13:31:52 +000036; X32-NEXT: movl %edi, %eax
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000037; X32-NEXT: divl %edi
38; X32-NEXT: xorl %edx, %edx
Simon Pilgrim153b4082016-09-09 13:31:52 +000039; X32-NEXT: movl %ebx, %eax
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000040; X32-NEXT: divl %ebx
Simon Pilgrim153b4082016-09-09 13:31:52 +000041; X32-NEXT: .p2align 4, 0x90
42; X32-NEXT: .LBB0_2: # %CF237
43; X32-NEXT: # Parent Loop BB0_1 Depth=1
44; X32-NEXT: # => This Inner Loop Header: Depth=2
45; X32-NEXT: testb %cl, %cl
46; X32-NEXT: jne .LBB0_2
47; X32-NEXT: jmp .LBB0_1
48;
49; X64-LABEL: knownbits_zext_in_reg:
50; X64: # BB#0: # %BB
51; X64-NEXT: movzbl (%rdi), %eax
52; X64-NEXT: imull $101, %eax, %eax
53; X64-NEXT: andl $16384, %eax # imm = 0x4000
54; X64-NEXT: shrl $14, %eax
55; X64-NEXT: movzbl %al, %eax
56; X64-NEXT: vmovd %eax, %xmm0
Simon Pilgrim1e4d8702016-12-01 15:41:40 +000057; X64-NEXT: vpshufb {{.*#+}} xmm0 = zero,zero,zero,zero,xmm0[0],zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero
Simon Pilgrim153b4082016-09-09 13:31:52 +000058; X64-NEXT: vpextrd $1, %xmm0, %r8d
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000059; X64-NEXT: xorl %esi, %esi
Simon Pilgrim153b4082016-09-09 13:31:52 +000060; X64-NEXT: vmovd %xmm0, %r9d
61; X64-NEXT: vpextrd $2, %xmm0, %edi
62; X64-NEXT: vpextrd $3, %xmm0, %ecx
Simon Pilgrim153b4082016-09-09 13:31:52 +000063; X64-NEXT: .p2align 4, 0x90
64; X64-NEXT: .LBB0_1: # %CF
65; X64-NEXT: # =>This Loop Header: Depth=1
66; X64-NEXT: # Child Loop BB0_2 Depth 2
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000067; X64-NEXT: xorl %edx, %edx
Simon Pilgrim153b4082016-09-09 13:31:52 +000068; X64-NEXT: movl %r8d, %eax
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000069; X64-NEXT: divl %r8d
70; X64-NEXT: xorl %edx, %edx
Simon Pilgrim153b4082016-09-09 13:31:52 +000071; X64-NEXT: movl %r9d, %eax
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000072; X64-NEXT: divl %r9d
73; X64-NEXT: xorl %edx, %edx
Simon Pilgrim153b4082016-09-09 13:31:52 +000074; X64-NEXT: movl %edi, %eax
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000075; X64-NEXT: divl %edi
76; X64-NEXT: xorl %edx, %edx
Simon Pilgrim153b4082016-09-09 13:31:52 +000077; X64-NEXT: movl %ecx, %eax
Simon Pilgrimdaf82f52016-10-25 21:24:33 +000078; X64-NEXT: divl %ecx
Simon Pilgrim153b4082016-09-09 13:31:52 +000079; X64-NEXT: .p2align 4, 0x90
80; X64-NEXT: .LBB0_2: # %CF237
81; X64-NEXT: # Parent Loop BB0_1 Depth=1
82; X64-NEXT: # => This Inner Loop Header: Depth=2
83; X64-NEXT: testb %sil, %sil
84; X64-NEXT: jne .LBB0_2
85; X64-NEXT: jmp .LBB0_1
86BB:
87 %L5 = load i8, i8* %0
88 %Sl9 = select i1 true, i8 %L5, i8 undef
89 %B21 = udiv i8 %Sl9, -93
90 br label %CF
91
92CF: ; preds = %CF246, %BB
93 %I40 = insertelement <4 x i8> zeroinitializer, i8 %B21, i32 1
94 %B41 = srem <4 x i8> %I40, %I40
95 br label %CF237
96
97CF237: ; preds = %CF237, %CF
98 %Cmp73 = icmp ne i1 undef, undef
99 br i1 %Cmp73, label %CF237, label %CF246
100
101CF246: ; preds = %CF237
102 %Cmp117 = icmp ult <4 x i8> %B41, undef
103 %E156 = extractelement <4 x i1> %Cmp117, i32 2
104 br label %CF
105}
Simon Pilgrimb070ce82017-02-06 14:06:57 +0000106
107define i32 @knownbits_mask_add_lshr(i32 %a0, i32 %a1) nounwind {
108; X32-LABEL: knownbits_mask_add_lshr:
109; X32: # BB#0:
110; X32-NEXT: xorl %eax, %eax
111; X32-NEXT: retl
112;
113; X64-LABEL: knownbits_mask_add_lshr:
114; X64: # BB#0:
115; X64-NEXT: xorl %eax, %eax
116; X64-NEXT: retq
117 %1 = and i32 %a0, 32767
118 %2 = and i32 %a1, 32766
119 %3 = add i32 %1, %2
120 %4 = lshr i32 %3, 17
121 ret i32 %4
122}
123
124define i128 @knownbits_mask_addc_shl(i64 %a0, i64 %a1, i64 %a2) nounwind {
125; X32-LABEL: knownbits_mask_addc_shl:
126; X32: # BB#0:
127; X32-NEXT: pushl %edi
128; X32-NEXT: pushl %esi
129; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
130; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
131; X32-NEXT: movl {{[0-9]+}}(%esp), %edx
132; X32-NEXT: movl $-1024, %esi # imm = 0xFC00
133; X32-NEXT: movl {{[0-9]+}}(%esp), %edi
134; X32-NEXT: andl %esi, %edi
135; X32-NEXT: andl {{[0-9]+}}(%esp), %esi
136; X32-NEXT: addl %edi, %esi
137; X32-NEXT: adcl {{[0-9]+}}(%esp), %edx
138; X32-NEXT: adcl $0, %ecx
139; X32-NEXT: shldl $22, %edx, %ecx
140; X32-NEXT: shldl $22, %esi, %edx
Simon Pilgrimb070ce82017-02-06 14:06:57 +0000141; X32-NEXT: movl %edx, 8(%eax)
142; X32-NEXT: movl %ecx, 12(%eax)
Amaury Sechete674f5c2017-02-06 14:59:06 +0000143; X32-NEXT: movl $0, 4(%eax)
Simon Pilgrimb070ce82017-02-06 14:06:57 +0000144; X32-NEXT: movl $0, (%eax)
145; X32-NEXT: popl %esi
146; X32-NEXT: popl %edi
147; X32-NEXT: retl $4
148;
149; X64-LABEL: knownbits_mask_addc_shl:
150; X64: # BB#0:
151; X64-NEXT: andq $-1024, %rdi # imm = 0xFC00
152; X64-NEXT: andq $-1024, %rsi # imm = 0xFC00
153; X64-NEXT: addq %rdi, %rsi
Amaury Sechet4b946912017-02-08 00:32:36 +0000154; X64-NEXT: adcl $0, %edx
Simon Pilgrimb070ce82017-02-06 14:06:57 +0000155; X64-NEXT: shldq $54, %rsi, %rdx
Amaury Sechete674f5c2017-02-06 14:59:06 +0000156; X64-NEXT: xorl %eax, %eax
Simon Pilgrimb070ce82017-02-06 14:06:57 +0000157; X64-NEXT: retq
158 %1 = and i64 %a0, -1024
159 %2 = zext i64 %1 to i128
160 %3 = and i64 %a1, -1024
161 %4 = zext i64 %3 to i128
162 %5 = add i128 %2, %4
163 %6 = zext i64 %a2 to i128
164 %7 = shl i128 %6, 64
165 %8 = add i128 %5, %7
166 %9 = shl i128 %8, 54
167 ret i128 %9
168}