blob: 47705190b0f4a95b2d29b9f7fd55e51161fc4082 [file] [log] [blame]
NAKAMURA Takumi20cee652011-02-22 07:20:44 +00001; RUN: llc < %s -mtriple=x86_64-linux | FileCheck %s
2; RUN: llc < %s -mtriple=x86_64-win32 | FileCheck %s
Dan Gohmanc2385a02010-02-16 01:42:53 +00003
4; Full strength reduction wouldn't reduce register pressure, so LSR should
5; stick with indexing here.
6
Jakob Stoklund Olesenb0e47cd2011-04-05 00:32:44 +00007; CHECK: movaps (%{{rsi|rdx}},%rax,4), [[X3:%xmm[0-9]+]]
8; CHECK: movaps
9; CHECK: [[X3]], (%{{rdi|rcx}},%rax,4)
Dan Gohmanc2385a02010-02-16 01:42:53 +000010; CHECK: addq $4, %rax
NAKAMURA Takumi20cee652011-02-22 07:20:44 +000011; CHECK: cmpl %eax, (%{{rdx|r8}})
Dan Gohmanc2385a02010-02-16 01:42:53 +000012; CHECK-NEXT: jg
13
14define void @vvfloorf(float* nocapture %y, float* nocapture %x, i32* nocapture %n) nounwind {
15entry:
16 %0 = load i32* %n, align 4
17 %1 = icmp sgt i32 %0, 0
18 br i1 %1, label %bb, label %return
19
20bb:
21 %indvar = phi i64 [ %indvar.next, %bb ], [ 0, %entry ]
22 %tmp = shl i64 %indvar, 2
23 %scevgep = getelementptr float* %y, i64 %tmp
24 %scevgep9 = bitcast float* %scevgep to <4 x float>*
25 %scevgep10 = getelementptr float* %x, i64 %tmp
26 %scevgep1011 = bitcast float* %scevgep10 to <4 x float>*
27 %2 = load <4 x float>* %scevgep1011, align 16
28 %3 = bitcast <4 x float> %2 to <4 x i32>
29 %4 = and <4 x i32> %3, <i32 2147483647, i32 2147483647, i32 2147483647, i32 2147483647>
30 %5 = bitcast <4 x i32> %4 to <4 x float>
31 %6 = and <4 x i32> %3, <i32 -2147483648, i32 -2147483648, i32 -2147483648, i32 -2147483648>
32 %7 = tail call <4 x float> @llvm.x86.sse.cmp.ps(<4 x float> %5, <4 x float> <float 8.388608e+06, float 8.388608e+06, float 8.388608e+06, float 8.388608e+06>, i8 5) nounwind
33 %tmp.i4 = bitcast <4 x float> %7 to <4 x i32>
34 %8 = xor <4 x i32> %tmp.i4, <i32 -1, i32 -1, i32 -1, i32 -1>
35 %9 = and <4 x i32> %8, <i32 1258291200, i32 1258291200, i32 1258291200, i32 1258291200>
36 %10 = or <4 x i32> %9, %6
37 %11 = bitcast <4 x i32> %10 to <4 x float>
38 %12 = fadd <4 x float> %2, %11
39 %13 = fsub <4 x float> %12, %11
40 %14 = tail call <4 x float> @llvm.x86.sse.cmp.ps(<4 x float> %2, <4 x float> %13, i8 1) nounwind
41 %15 = bitcast <4 x float> %14 to <4 x i32>
42 %16 = tail call <4 x float> @llvm.x86.sse2.cvtdq2ps(<4 x i32> %15) nounwind readnone
43 %17 = fadd <4 x float> %13, %16
44 %tmp.i = bitcast <4 x float> %17 to <4 x i32>
45 %18 = or <4 x i32> %tmp.i, %6
46 %19 = bitcast <4 x i32> %18 to <4 x float>
47 store <4 x float> %19, <4 x float>* %scevgep9, align 16
48 %tmp12 = add i64 %tmp, 4
49 %tmp13 = trunc i64 %tmp12 to i32
50 %20 = load i32* %n, align 4
51 %21 = icmp sgt i32 %20, %tmp13
52 %indvar.next = add i64 %indvar, 1
53 br i1 %21, label %bb, label %return
54
55return:
56 ret void
57}
58
59declare <4 x float> @llvm.x86.sse.cmp.ps(<4 x float>, <4 x float>, i8) nounwind readnone
60
61declare <4 x float> @llvm.x86.sse2.cvtdq2ps(<4 x i32>) nounwind readnone