For PR761:
Remove "target endian/pointersize" or add "target datalayout" to make
the test parse properly or set the datalayout because defaults changes.
For PR645:
Make global names use the @ prefix.
For llvm-upgrade changes:
Fix test cases or completely remove use of llvm-upgrade for test cases
that cannot survive the new renaming or upgrade capabilities.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@33533 91177308-0d34-0410-b5e6-96231b3b80d8
diff --git a/test/CodeGen/ARM/2007-01-19-InfiniteLoop.ll b/test/CodeGen/ARM/2007-01-19-InfiniteLoop.ll
index c88a867..0dbf795 100644
--- a/test/CodeGen/ARM/2007-01-19-InfiniteLoop.ll
+++ b/test/CodeGen/ARM/2007-01-19-InfiniteLoop.ll
@@ -1,10 +1,10 @@
; RUN: llvm-as < %s | llc -march=arm -mattr=+v6,+vfp2
-%quant_coef = external global [6 x [4 x [4 x i32]]] ; <[6 x [4 x [4 x i32]]]*> [#uses=1]
-%dequant_coef = external global [6 x [4 x [4 x i32]]] ; <[6 x [4 x [4 x i32]]]*> [#uses=1]
-%A = external global [4 x [4 x i32]] ; <[4 x [4 x i32]]*> [#uses=1]
+@quant_coef = external global [6 x [4 x [4 x i32]]] ; <[6 x [4 x [4 x i32]]]*> [#uses=1]
+@dequant_coef = external global [6 x [4 x [4 x i32]]] ; <[6 x [4 x [4 x i32]]]*> [#uses=1]
+@A = external global [4 x [4 x i32]] ; <[4 x [4 x i32]]*> [#uses=1]
-define fastcc i32 %dct_luma_sp(i32 %block_x, i32 %block_y, i32* %coeff_cost) {
+define fastcc i32 @dct_luma_sp(i32 %block_x, i32 %block_y, i32* %coeff_cost) {
entry:
%predicted_block = alloca [4 x [4 x i32]], align 4 ; <[4 x [4 x i32]]*> [#uses=1]
br label %cond_next489
@@ -17,7 +17,7 @@
%tmp495 = getelementptr [4 x [4 x i32]]* %predicted_block, i32 0, i32 %i.8, i32 %j.7 ; <i32*> [#uses=2]
%tmp496 = load i32* %tmp495 ; <i32> [#uses=2]
%tmp502 = load i32* null ; <i32> [#uses=1]
- %tmp542 = getelementptr [6 x [4 x [4 x i32]]]* %quant_coef, i32 0, i32 0, i32 %i.8, i32 %j.7 ; <i32*> [#uses=1]
+ %tmp542 = getelementptr [6 x [4 x [4 x i32]]]* @quant_coef, i32 0, i32 0, i32 %i.8, i32 %j.7 ; <i32*> [#uses=1]
%tmp543 = load i32* %tmp542 ; <i32> [#uses=1]
%tmp548 = ashr i32 0, i8 0 ; <i32> [#uses=3]
%tmp561 = sub i32 0, %tmp496 ; <i32> [#uses=3]
@@ -30,14 +30,14 @@
br i1 %tmp579, label %bb712, label %cond_next589
cond_next589: ; preds = %cond_next489
- %tmp605 = getelementptr [6 x [4 x [4 x i32]]]* %dequant_coef, i32 0, i32 0, i32 %i.8, i32 %j.7 ; <i32*> [#uses=1]
+ %tmp605 = getelementptr [6 x [4 x [4 x i32]]]* @dequant_coef, i32 0, i32 0, i32 %i.8, i32 %j.7 ; <i32*> [#uses=1]
%tmp606 = load i32* %tmp605 ; <i32> [#uses=1]
%tmp612 = load i32* null ; <i32> [#uses=1]
%tmp629 = load i32* null ; <i32> [#uses=1]
%tmp629 = sitofp i32 %tmp629 to double ; <double> [#uses=1]
%tmp631 = mul double %tmp629, 0.000000e+00 ; <double> [#uses=1]
%tmp632 = add double 0.000000e+00, %tmp631 ; <double> [#uses=1]
- %tmp642 = call fastcc i32 %sign( i32 %tmp576, i32 %tmp561 ) ; <i32> [#uses=1]
+ %tmp642 = call fastcc i32 @sign( i32 %tmp576, i32 %tmp561 ) ; <i32> [#uses=1]
%tmp650 = mul i32 %tmp606, %tmp642 ; <i32> [#uses=1]
%tmp656 = mul i32 %tmp650, %tmp612 ; <i32> [#uses=1]
%tmp658 = shl i32 %tmp656, i8 0 ; <i32> [#uses=1]
@@ -45,7 +45,7 @@
%tmp660 = sub i32 0, %tmp659 ; <i32> [#uses=1]
%tmp666 = sub i32 %tmp660, %tmp496 ; <i32> [#uses=1]
%tmp666 = sitofp i32 %tmp666 to double ; <double> [#uses=2]
- call void %levrun_linfo_inter( i32 %tmp576, i32 0, i32* null, i32* null )
+ call void @levrun_linfo_inter( i32 %tmp576, i32 0, i32* null, i32* null )
%tmp671 = mul double %tmp666, %tmp666 ; <double> [#uses=1]
%tmp675 = add double %tmp671, 0.000000e+00 ; <double> [#uses=1]
%tmp678 = fcmp oeq double %tmp632, %tmp675 ; <i1> [#uses=1]
@@ -79,9 +79,9 @@
br i1 %tmp739, label %cond_next791, label %cond_true740
cond_true740: ; preds = %bb737
- %tmp761 = call fastcc i32 %sign( i32 %tmp576, i32 0 ) ; <i32> [#uses=1]
+ %tmp761 = call fastcc i32 @sign( i32 %tmp576, i32 0 ) ; <i32> [#uses=1]
%tmp780 = load i32* null ; <i32> [#uses=1]
- %tmp785 = getelementptr [4 x [4 x i32]]* %A, i32 0, i32 %i.8, i32 %j.7 ; <i32*> [#uses=1]
+ %tmp785 = getelementptr [4 x [4 x i32]]* @A, i32 0, i32 %i.8, i32 %j.7 ; <i32*> [#uses=1]
%tmp786 = load i32* %tmp785 ; <i32> [#uses=1]
%tmp781 = mul i32 %tmp780, %tmp761 ; <i32> [#uses=1]
%tmp787 = mul i32 %tmp781, %tmp786 ; <i32> [#uses=1]
@@ -94,10 +94,10 @@
%tmp796 = load i32* %tmp495 ; <i32> [#uses=1]
%tmp798 = add i32 %tmp796, %ilev.1 ; <i32> [#uses=1]
%tmp812 = mul i32 0, %tmp502 ; <i32> [#uses=0]
- %tmp818 = call fastcc i32 %sign( i32 0, i32 %tmp798 ) ; <i32> [#uses=0]
+ %tmp818 = call fastcc i32 @sign( i32 0, i32 %tmp798 ) ; <i32> [#uses=0]
unreachable
}
-declare i32 %sign(i32, i32)
+declare i32 @sign(i32, i32)
-declare void %levrun_linfo_inter(i32, i32, i32*, i32*)
+declare void @levrun_linfo_inter(i32, i32, i32*, i32*)
diff --git a/test/CodeGen/ARM/fcopysign.ll b/test/CodeGen/ARM/fcopysign.ll
index ddc0fae..73c6670 100644
--- a/test/CodeGen/ARM/fcopysign.ll
+++ b/test/CodeGen/ARM/fcopysign.ll
@@ -3,17 +3,17 @@
; RUN: llvm-as < %s | llc -march=arm -mattr=+v6,+vfp2 &&
; RUN: llvm-as < %s | llc -march=arm -mattr=+v6,+vfp2 | grep fneg | wc -l | grep 2
-define float %test1(float %x, double %y) {
+define float @test1(float %x, double %y) {
%tmp = fpext float %x to double
- %tmp2 = tail call double %copysign( double %tmp, double %y )
+ %tmp2 = tail call double @copysign( double %tmp, double %y )
%tmp2 = fptrunc double %tmp2 to float
ret float %tmp2
}
-define double %test2(double %x, float %y) {
+define double @test2(double %x, float %y) {
%tmp = fpext float %y to double
- %tmp2 = tail call double %copysign( double %x, double %tmp )
+ %tmp2 = tail call double @copysign( double %x, double %tmp )
ret double %tmp2
}
-declare double %copysign(double, double)
+declare double @copysign(double, double)
diff --git a/test/CodeGen/ARM/fpcmp_ueq.ll b/test/CodeGen/ARM/fpcmp_ueq.ll
index 186c9f3..3ff824a 100644
--- a/test/CodeGen/ARM/fpcmp_ueq.ll
+++ b/test/CodeGen/ARM/fpcmp_ueq.ll
@@ -2,7 +2,7 @@
; RUN: llvm-as < %s | llc -march=arm | grep moveq &&
; RUN: llvm-as < %s | llc -march=arm -mattr=+vfp2 | grep movvs
-define i32 %f7(float %a, float %b) {
+define i32 @f7(float %a, float %b) {
entry:
%tmp = fcmp ueq float %a,%b
%retval = select i1 %tmp, i32 666, i32 42
diff --git a/test/CodeGen/ARM/large-stack.ll b/test/CodeGen/ARM/large-stack.ll
index 2516522..4402659 100644
--- a/test/CodeGen/ARM/large-stack.ll
+++ b/test/CodeGen/ARM/large-stack.ll
@@ -1,12 +1,12 @@
; RUN: llvm-as < %s | llc -march=arm &&
; RUN: llvm-as < %s | llc -march=arm -enable-thumb
-define void %test1() {
+define void @test1() {
%tmp = alloca [ 64 x i32 ] , align 4
ret void
}
-define void %test2() {
+define void @test2() {
%tmp = alloca [ 4168 x i8 ] , align 4
ret void
}
diff --git a/test/CodeGen/ARM/ldr_ext.ll b/test/CodeGen/ARM/ldr_ext.ll
index 739447c..3eabda3 100644
--- a/test/CodeGen/ARM/ldr_ext.ll
+++ b/test/CodeGen/ARM/ldr_ext.ll
@@ -8,25 +8,25 @@
; RUN: llvm-as < %s | llc -march=arm -enable-thumb | grep "ldrsb" | wc -l | grep 1 &&
; RUN: llvm-as < %s | llc -march=arm -enable-thumb | grep "ldrsh" | wc -l | grep 1
-define i32 %test1(i8* %v.pntr.s0.u1) {
+define i32 @test1(i8* %v.pntr.s0.u1) {
%tmp.u = load i8* %v.pntr.s0.u1
%tmp1.s = zext i8 %tmp.u to i32
ret i32 %tmp1.s
}
-define i32 %test2(i16* %v.pntr.s0.u1) {
+define i32 @test2(i16* %v.pntr.s0.u1) {
%tmp.u = load i16* %v.pntr.s0.u1
%tmp1.s = zext i16 %tmp.u to i32
ret i32 %tmp1.s
}
-define i32 %test3(i8* %v.pntr.s1.u0) {
+define i32 @test3(i8* %v.pntr.s1.u0) {
%tmp.s = load i8* %v.pntr.s1.u0
%tmp1.s = sext i8 %tmp.s to i32
ret i32 %tmp1.s
}
-define i32 %test4() {
+define i32 @test4() {
%tmp.s = load i16* null
%tmp1.s = sext i16 %tmp.s to i32
ret i32 %tmp1.s
diff --git a/test/CodeGen/ARM/ldr_frame.ll b/test/CodeGen/ARM/ldr_frame.ll
index 23407c0..3db9d79 100644
--- a/test/CodeGen/ARM/ldr_frame.ll
+++ b/test/CodeGen/ARM/ldr_frame.ll
@@ -3,14 +3,14 @@
; RUN: llvm-as < %s | llc -march=arm -enable-thumb &&
; RUN: llvm-as < %s | llc -march=arm -enable-thumb | grep cpy | wc -l | grep 2
-define i32 %f1() {
+define i32 @f1() {
%buf = alloca [32 x i32], align 4
%tmp = getelementptr [32 x i32]* %buf, i32 0, i32 0
%tmp1 = load i32* %tmp
ret i32 %tmp1
}
-define i32 %f2() {
+define i32 @f2() {
%buf = alloca [32 x i8], align 4
%tmp = getelementptr [32 x i8]* %buf, i32 0, i32 0
%tmp1 = load i8* %tmp
@@ -18,14 +18,14 @@
ret i32 %tmp2
}
-define i32 %f3() {
+define i32 @f3() {
%buf = alloca [32 x i32], align 4
%tmp = getelementptr [32 x i32]* %buf, i32 0, i32 32
%tmp1 = load i32* %tmp
ret i32 %tmp1
}
-define i32 %f4() {
+define i32 @f4() {
%buf = alloca [32 x i8], align 4
%tmp = getelementptr [32 x i8]* %buf, i32 0, i32 2
%tmp1 = load i8* %tmp
diff --git a/test/CodeGen/ARM/mul.ll b/test/CodeGen/ARM/mul.ll
index 435e59f..f1a0a5a 100644
--- a/test/CodeGen/ARM/mul.ll
+++ b/test/CodeGen/ARM/mul.ll
@@ -4,22 +4,22 @@
; RUN: llvm-as < %s | llc -march=arm -enable-thumb | grep mul | wc -l | grep 3 &&
; RUN: llvm-as < %s | llc -march=arm -enable-thumb | grep lsl | wc -l | grep 1
-define i32 %f1(i32 %u) {
+define i32 @f1(i32 %u) {
%tmp = mul i32 %u, %u
ret i32 %tmp
}
-define i32 %f2(i32 %u, i32 %v) {
+define i32 @f2(i32 %u, i32 %v) {
%tmp = mul i32 %u, %v
ret i32 %tmp
}
-define i32 %f3(i32 %u) {
+define i32 @f3(i32 %u) {
%tmp = mul i32 %u, 5
ret i32 %tmp
}
-define i32 %f4(i32 %u) {
+define i32 @f4(i32 %u) {
%tmp = mul i32 %u, 4
ret i32 %tmp
}
diff --git a/test/CodeGen/ARM/select.ll b/test/CodeGen/ARM/select.ll
index aca7b7d..82f1ac8 100644
--- a/test/CodeGen/ARM/select.ll
+++ b/test/CodeGen/ARM/select.ll
@@ -14,49 +14,49 @@
; RUN: llvm-as < %s | llc -march=arm -enable-thumb | grep bhi | wc -l | grep 1 &&
; RUN: llvm-as < %s | llc -march=arm -enable-thumb | grep __ltdf2
-define i32 %f1(i32 %a.s) {
+define i32 @f1(i32 %a.s) {
entry:
%tmp = icmp eq i32 %a.s, 4
%tmp1.s = select i1 %tmp, i32 2, i32 3
ret i32 %tmp1.s
}
-define i32 %f2(i32 %a.s) {
+define i32 @f2(i32 %a.s) {
entry:
%tmp = icmp sgt i32 %a.s, 4
%tmp1.s = select i1 %tmp, i32 2, i32 3
ret i32 %tmp1.s
}
-define i32 %f3(i32 %a.s, i32 %b.s) {
+define i32 @f3(i32 %a.s, i32 %b.s) {
entry:
%tmp = icmp slt i32 %a.s, %b.s
%tmp1.s = select i1 %tmp, i32 2, i32 3
ret i32 %tmp1.s
}
-define i32 %f4(i32 %a.s, i32 %b.s) {
+define i32 @f4(i32 %a.s, i32 %b.s) {
entry:
%tmp = icmp sle i32 %a.s, %b.s
%tmp1.s = select i1 %tmp, i32 2, i32 3
ret i32 %tmp1.s
}
-define i32 %f5(i32 %a.u, i32 %b.u) {
+define i32 @f5(i32 %a.u, i32 %b.u) {
entry:
%tmp = icmp ule i32 %a.u, %b.u
%tmp1.s = select i1 %tmp, i32 2, i32 3
ret i32 %tmp1.s
}
-define i32 %f6(i32 %a.u, i32 %b.u) {
+define i32 @f6(i32 %a.u, i32 %b.u) {
entry:
%tmp = icmp ugt i32 %a.u, %b.u
%tmp1.s = select i1 %tmp, i32 2, i32 3
ret i32 %tmp1.s
}
-define double %f7(double %a, double %b) {
+define double @f7(double %a, double %b) {
%tmp = fcmp olt double %a, 1.234e+00
%tmp1 = select i1 %tmp, double -1.000e+00, double %b
ret double %tmp1
diff --git a/test/CodeGen/ARM/sxt_rot.ll b/test/CodeGen/ARM/sxt_rot.ll
index 1019b94..3f1483b 100644
--- a/test/CodeGen/ARM/sxt_rot.ll
+++ b/test/CodeGen/ARM/sxt_rot.ll
@@ -3,7 +3,7 @@
; RUN: llvm-as < %s | llc -march=arm -mattr=+v6 | grep "sxtb" | wc -l | grep 1 &&
; RUN: llvm-as < %s | llc -march=arm -mattr=+v6 | grep "sxtab" | wc -l | grep 1
-define i8 %test1(i32 %A) sext {
+define i8 @test1(i32 %A) sext {
%B = lshr i32 %A, i8 8
%C = shl i32 %A, i8 24
%D = or i32 %B, %C
@@ -11,7 +11,7 @@
ret i8 %E
}
-define i32 %test2(i32 %A, i32 %X) sext {
+define i32 @test2(i32 %A, i32 %X) sext {
%B = lshr i32 %A, i8 8
%C = shl i32 %A, i8 24
%D = or i32 %B, %C
diff --git a/test/CodeGen/ARM/thumb-imm.ll b/test/CodeGen/ARM/thumb-imm.ll
index 1e50a89..4bd0b16 100644
--- a/test/CodeGen/ARM/thumb-imm.ll
+++ b/test/CodeGen/ARM/thumb-imm.ll
@@ -2,10 +2,10 @@
; RUN: llvm-as < %s | llc -march=arm -enable-thumb | not grep CPI
-define i32 %test1() {
+define i32 @test1() {
ret i32 1000
}
-define i32 %test2() {
+define i32 @test2() {
ret i32 -256
}
diff --git a/test/CodeGen/ARM/uxt_rot.ll b/test/CodeGen/ARM/uxt_rot.ll
index 47f907e..0c7516f 100644
--- a/test/CodeGen/ARM/uxt_rot.ll
+++ b/test/CodeGen/ARM/uxt_rot.ll
@@ -4,19 +4,19 @@
; RUN: llvm-as < %s | llc -march=arm -mattr=+v6 | grep "uxtab" | wc -l | grep 1 &&
; RUN: llvm-as < %s | llc -march=arm -mattr=+v6 | grep "uxth" | wc -l | grep 1
-define i8 %test1(i32 %A.u) zext {
+define i8 @test1(i32 %A.u) zext {
%B.u = trunc i32 %A.u to i8
ret i8 %B.u
}
-define i32 %test2(i32 %A.u, i32 %B.u) zext {
+define i32 @test2(i32 %A.u, i32 %B.u) zext {
%C.u = trunc i32 %B.u to i8
%D.u = zext i8 %C.u to i32
%E.u = add i32 %A.u, %D.u
ret i32 %E.u
}
-define i32 %test3(i32 %A.u) zext {
+define i32 @test3(i32 %A.u) zext {
%B.u = lshr i32 %A.u, i8 8
%C.u = shl i32 %A.u, i8 24
%D.u = or i32 %B.u, %C.u
diff --git a/test/CodeGen/Alpha/add.ll b/test/CodeGen/Alpha/add.ll
index e385dee..7d1bd0c 100644
--- a/test/CodeGen/Alpha/add.ll
+++ b/test/CodeGen/Alpha/add.ll
@@ -19,160 +19,160 @@
implementation ; Functions:
-define i32 %al(i32 sext %x.s, i32 sext %y.s) sext {
+define i32 @al(i32 sext %x.s, i32 sext %y.s) sext {
entry:
%tmp.3.s = add i32 %y.s, %x.s ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i32 %ali(i32 sext %x.s) sext {
+define i32 @ali(i32 sext %x.s) sext {
entry:
%tmp.3.s = add i32 100, %x.s ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i64 %aq(i64 sext %x.s, i64 sext %y.s) sext {
+define i64 @aq(i64 sext %x.s, i64 sext %y.s) sext {
entry:
%tmp.3.s = add i64 %y.s, %x.s ; <i64> [#uses=1]
ret i64 %tmp.3.s
}
-define i64 %aqi(i64 %x.s) {
+define i64 @aqi(i64 %x.s) {
entry:
%tmp.3.s = add i64 100, %x.s ; <i64> [#uses=1]
ret i64 %tmp.3.s
}
-define i32 %sl(i32 sext %x.s, i32 sext %y.s) sext {
+define i32 @sl(i32 sext %x.s, i32 sext %y.s) sext {
entry:
%tmp.3.s = sub i32 %y.s, %x.s ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i32 %sli(i32 sext %x.s) sext {
+define i32 @sli(i32 sext %x.s) sext {
entry:
%tmp.3.s = sub i32 %x.s, 100 ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i64 %sq(i64 %x.s, i64 %y.s) {
+define i64 @sq(i64 %x.s, i64 %y.s) {
entry:
%tmp.3.s = sub i64 %y.s, %x.s ; <i64> [#uses=1]
ret i64 %tmp.3.s
}
-define i64 %sqi(i64 %x.s) {
+define i64 @sqi(i64 %x.s) {
entry:
%tmp.3.s = sub i64 %x.s, 100 ; <i64> [#uses=1]
ret i64 %tmp.3.s
}
-define i32 %a4l(i32 sext %x.s, i32 sext %y.s) sext {
+define i32 @a4l(i32 sext %x.s, i32 sext %y.s) sext {
entry:
%tmp.1.s = shl i32 %y.s, i8 2 ; <i32> [#uses=1]
%tmp.3.s = add i32 %tmp.1.s, %x.s ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i32 %a8l(i32 sext %x.s, i32 sext %y.s) sext {
+define i32 @a8l(i32 sext %x.s, i32 sext %y.s) sext {
entry:
%tmp.1.s = shl i32 %y.s, i8 3 ; <i32> [#uses=1]
%tmp.3.s = add i32 %tmp.1.s, %x.s ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i64 %a4q(i64 %x.s, i64 %y.s) {
+define i64 @a4q(i64 %x.s, i64 %y.s) {
entry:
%tmp.1.s = shl i64 %y.s, i8 2 ; <i64> [#uses=1]
%tmp.3.s = add i64 %tmp.1.s, %x.s ; <i64> [#uses=1]
ret i64 %tmp.3.s
}
-define i64 %a8q(i64 %x.s, i64 %y.s) {
+define i64 @a8q(i64 %x.s, i64 %y.s) {
entry:
%tmp.1.s = shl i64 %y.s, i8 3 ; <i64> [#uses=1]
%tmp.3.s = add i64 %tmp.1.s, %x.s ; <i64> [#uses=1]
ret i64 %tmp.3.s
}
-define i32 %a4li(i32 sext %y.s) sext {
+define i32 @a4li(i32 sext %y.s) sext {
entry:
%tmp.1.s = shl i32 %y.s, i8 2 ; <i32> [#uses=1]
%tmp.3.s = add i32 100, %tmp.1.s ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i32 %a8li(i32 sext %y.s) sext {
+define i32 @a8li(i32 sext %y.s) sext {
entry:
%tmp.1.s = shl i32 %y.s, i8 3 ; <i32> [#uses=1]
%tmp.3.s = add i32 100, %tmp.1.s ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i64 %a4qi(i64 %y.s) {
+define i64 @a4qi(i64 %y.s) {
entry:
%tmp.1.s = shl i64 %y.s, i8 2 ; <i64> [#uses=1]
%tmp.3.s = add i64 100, %tmp.1.s ; <i64> [#uses=1]
ret i64 %tmp.3.s
}
-define i64 %a8qi(i64 %y.s) {
+define i64 @a8qi(i64 %y.s) {
entry:
%tmp.1.s = shl i64 %y.s, i8 3 ; <i64> [#uses=1]
%tmp.3.s = add i64 100, %tmp.1.s ; <i64> [#uses=1]
ret i64 %tmp.3.s
}
-define i32 %s4l(i32 sext %x.s, i32 sext %y.s) sext {
+define i32 @s4l(i32 sext %x.s, i32 sext %y.s) sext {
entry:
%tmp.1.s = shl i32 %y.s, i8 2 ; <i32> [#uses=1]
%tmp.3.s = sub i32 %tmp.1.s, %x.s ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i32 %s8l(i32 sext %x.s, i32 sext %y.s) sext {
+define i32 @s8l(i32 sext %x.s, i32 sext %y.s) sext {
entry:
%tmp.1.s = shl i32 %y.s, i8 3 ; <i32> [#uses=1]
%tmp.3.s = sub i32 %tmp.1.s, %x.s ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i64 %s4q(i64 %x.s, i64 %y.s) {
+define i64 @s4q(i64 %x.s, i64 %y.s) {
entry:
%tmp.1.s = shl i64 %y.s, i8 2 ; <i64> [#uses=1]
%tmp.3.s = sub i64 %tmp.1.s, %x.s ; <i64> [#uses=1]
ret i64 %tmp.3.s
}
-define i64 %s8q(i64 %x.s, i64 %y.s) {
+define i64 @s8q(i64 %x.s, i64 %y.s) {
entry:
%tmp.1.s = shl i64 %y.s, i8 3 ; <i64> [#uses=1]
%tmp.3.s = sub i64 %tmp.1.s, %x.s ; <i64> [#uses=1]
ret i64 %tmp.3.s
}
-define i32 %s4li(i32 sext %y.s) sext {
+define i32 @s4li(i32 sext %y.s) sext {
entry:
%tmp.1.s = shl i32 %y.s, i8 2 ; <i32> [#uses=1]
%tmp.3.s = sub i32 %tmp.1.s, 100 ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i32 %s8li(i32 sext %y.s) sext {
+define i32 @s8li(i32 sext %y.s) sext {
entry:
%tmp.1.s = shl i32 %y.s, i8 3 ; <i32> [#uses=1]
%tmp.3.s = sub i32 %tmp.1.s, 100 ; <i32> [#uses=1]
ret i32 %tmp.3.s
}
-define i64 %s4qi(i64 %y.s) {
+define i64 @s4qi(i64 %y.s) {
entry:
%tmp.1.s = shl i64 %y.s, i8 2 ; <i64> [#uses=1]
%tmp.3.s = sub i64 %tmp.1.s, 100 ; <i64> [#uses=1]
ret i64 %tmp.3.s
}
-define i64 %s8qi(i64 %y.s) {
+define i64 @s8qi(i64 %y.s) {
entry:
%tmp.1.s = shl i64 %y.s, i8 3 ; <i64> [#uses=1]
%tmp.3.s = sub i64 %tmp.1.s, 100 ; <i64> [#uses=1]
diff --git a/test/CodeGen/Alpha/i32_sub_1.ll b/test/CodeGen/Alpha/i32_sub_1.ll
index 9d8cb09..91ce648 100644
--- a/test/CodeGen/Alpha/i32_sub_1.ll
+++ b/test/CodeGen/Alpha/i32_sub_1.ll
@@ -3,7 +3,7 @@
implementation ; Functions:
-define i32 %foo(i32 sext %x) sext {
+define i32 @foo(i32 sext %x) sext {
entry:
%tmp.1 = add i32 %x, -1 ; <int> [#uses=1]
ret i32 %tmp.1
diff --git a/test/CodeGen/Alpha/zapnot.ll b/test/CodeGen/Alpha/zapnot.ll
index 8d1fa11..09c31a4 100644
--- a/test/CodeGen/Alpha/zapnot.ll
+++ b/test/CodeGen/Alpha/zapnot.ll
@@ -3,7 +3,7 @@
implementation ; Functions:
-define i16 %foo(i64 %y) zext {
+define i16 @foo(i64 %y) zext {
entry:
%tmp.1 = trunc i64 %y to i16 ; <ushort> [#uses=1]
ret i16 %tmp.1
diff --git a/test/CodeGen/CBackend/2007-01-08-ParamAttr-ICmp.ll b/test/CodeGen/CBackend/2007-01-08-ParamAttr-ICmp.ll
index e6760e3..ca04f9b 100644
--- a/test/CodeGen/CBackend/2007-01-08-ParamAttr-ICmp.ll
+++ b/test/CodeGen/CBackend/2007-01-08-ParamAttr-ICmp.ll
@@ -4,14 +4,12 @@
; XFAIL: *
target datalayout = "e-p:32:32"
-target endian = little
-target pointersize = 32
target triple = "i686-apple-darwin8"
%struct.Connector = type { i16, i16, i8, i8, %struct.Connector*, i8* }
implementation ; Functions:
-define bool %prune_match_entry_2E_ce(%struct.Connector* %a, i16 %b.0.0.val) {
+define bool @prune_match_entry_2E_ce(%struct.Connector* %a, i16 %b.0.0.val) {
newFuncRoot:
br label %entry.ce
diff --git a/test/CodeGen/CBackend/2007-01-17-StackSaveNRestore.ll b/test/CodeGen/CBackend/2007-01-17-StackSaveNRestore.ll
index 5694d2a..8fe06b7 100644
--- a/test/CodeGen/CBackend/2007-01-17-StackSaveNRestore.ll
+++ b/test/CodeGen/CBackend/2007-01-17-StackSaveNRestore.ll
@@ -2,11 +2,11 @@
; RUN: llvm-as < %s | llc -march=c | grep __builtin_stack_restore
; PR1028
-declare i8* %llvm.stacksave()
-declare void %llvm.stackrestore(i8*)
+declare i8* @llvm.stacksave()
+declare void @llvm.stackrestore(i8*)
-define i8* %test() {
- %s = call i8* %llvm.stacksave()
- call void %llvm.stackrestore(i8* %s)
+define i8* @test() {
+ %s = call i8* @llvm.stacksave()
+ call void @llvm.stackrestore(i8* %s)
ret i8* %s
}
diff --git a/test/CodeGen/Generic/2007-01-15-LoadSelectCycle.ll b/test/CodeGen/Generic/2007-01-15-LoadSelectCycle.ll
index 6eb22cc..49203d9 100644
--- a/test/CodeGen/Generic/2007-01-15-LoadSelectCycle.ll
+++ b/test/CodeGen/Generic/2007-01-15-LoadSelectCycle.ll
@@ -1,12 +1,12 @@
; RUN: llvm-as < %s | llc
; PR1114
-declare i1 %foo()
+declare i1 @foo()
-define i32 %test(i32* %A, i32* %B) {
+define i32 @test(i32* %A, i32* %B) {
%a = load i32* %A
%b = load i32* %B
- %cond = call i1 %foo()
+ %cond = call i1 @foo()
%c = select i1 %cond, i32 %a, i32 %b
ret i32 %c
}
diff --git a/test/CodeGen/PowerPC/2007-01-04-ArgExtension.ll b/test/CodeGen/PowerPC/2007-01-04-ArgExtension.ll
index 06875f3..a1dfc12 100644
--- a/test/CodeGen/PowerPC/2007-01-04-ArgExtension.ll
+++ b/test/CodeGen/PowerPC/2007-01-04-ArgExtension.ll
@@ -1,7 +1,7 @@
; RUN: llvm-as < %s | llc -march=ppc32 | grep extsb &&
; RUN: llvm-as < %s | llc -march=ppc32 | grep extsh
-define i32 %p1(i8 %c, i16 %s) {
+define i32 @p1(i8 %c, i16 %s) {
entry:
%tmp = sext i8 %c to i32 ; <i32> [#uses=1]
%tmp1 = sext i16 %s to i32 ; <i32> [#uses=1]
diff --git a/test/CodeGen/PowerPC/2007-01-15-AsmDialect.ll b/test/CodeGen/PowerPC/2007-01-15-AsmDialect.ll
index eecbb81..cfbac7d 100644
--- a/test/CodeGen/PowerPC/2007-01-15-AsmDialect.ll
+++ b/test/CodeGen/PowerPC/2007-01-15-AsmDialect.ll
@@ -1,6 +1,6 @@
; RUN: llvm-as < %s | llc -march=ppc32 -mtriple=powerpc-apple-darwin8 | grep cntlzw
-define i32 %foo() {
+define i32 @foo() {
entry:
%retval = alloca i32, align 4 ; <i32*> [#uses=2]
%tmp = alloca i32, align 4 ; <i32*> [#uses=2]
diff --git a/test/CodeGen/PowerPC/and-elim.ll b/test/CodeGen/PowerPC/and-elim.ll
index 8f423d8..b7fe9d2 100644
--- a/test/CodeGen/PowerPC/and-elim.ll
+++ b/test/CodeGen/PowerPC/and-elim.ll
@@ -1,7 +1,7 @@
; RUN: llvm-as < %s | llc -march=ppc32 &&
; RUN: llvm-as < %s | llc -march=ppc32 | not grep rlwin
-define void %test(i8* %P) {
+define void @test(i8* %P) {
%W = load i8* %P
%X = shl i8 %W, i8 1
%Y = add i8 %X, 2
@@ -10,7 +10,7 @@
ret void
}
-define i16 %test2(i16 zext %crc) zext {
+define i16 @test2(i16 zext %crc) zext {
; No and's should be needed for the i16s here.
%tmp.1 = lshr i16 %crc, i8 1
%tmp.7 = xor i16 %tmp.1, 40961
diff --git a/test/CodeGen/PowerPC/and_sext.ll b/test/CodeGen/PowerPC/and_sext.ll
index f025e21..7f4c5f0 100644
--- a/test/CodeGen/PowerPC/and_sext.ll
+++ b/test/CodeGen/PowerPC/and_sext.ll
@@ -3,14 +3,14 @@
; RUN: llvm-as < %s | llc -march=ppc32 | not grep extsh &&
; RUN: llvm-as < %s | llc -march=ppc32 | not grep extsb
-define i32 %test1(i32 %mode.0.i.0) {
+define i32 @test1(i32 %mode.0.i.0) {
%tmp.79 = trunc i32 %mode.0.i.0 to i16
%tmp.80 = sext i16 %tmp.79 to i32
%tmp.81 = and i32 %tmp.80, 24
ret i32 %tmp.81
}
-define i16 %test2(i16 sext %X, i16 sext %x) sext {
+define i16 @test2(i16 sext %X, i16 sext %x) sext {
%tmp = sext i16 %X to i32
%tmp1 = sext i16 %x to i32
%tmp2 = add i32 %tmp, %tmp1
@@ -21,7 +21,7 @@
ret i16 %retval
}
-define i16 %test3(i32 zext %X) sext {
+define i16 @test3(i32 zext %X) sext {
%tmp1 = lshr i32 %X, i8 16
%tmp1 = trunc i32 %tmp1 to i16
ret i16 %tmp1
diff --git a/test/CodeGen/PowerPC/small-arguments.ll b/test/CodeGen/PowerPC/small-arguments.ll
index 0068427..3ab92d1 100644
--- a/test/CodeGen/PowerPC/small-arguments.ll
+++ b/test/CodeGen/PowerPC/small-arguments.ll
@@ -1,39 +1,39 @@
; RUN: llvm-as < %s | llc -march=ppc32 &&
; RUN: llvm-as < %s | llc -march=ppc32 | not grep 'extsh\|rlwinm'
-declare i16 %foo() sext
+declare i16 @foo() sext
-define i32 %test1(i16 sext %X) {
+define i32 @test1(i16 sext %X) {
%Y = sext i16 %X to i32 ;; dead
ret i32 %Y
}
-define i32 %test2(i16 zext %X) {
+define i32 @test2(i16 zext %X) {
%Y = sext i16 %X to i32
%Z = and i32 %Y, 65535 ;; dead
ret i32 %Z
}
-define void %test3() {
- %tmp.0 = call i16 %foo() sext ;; no extsh!
+define void @test3() {
+ %tmp.0 = call i16 @foo() sext ;; no extsh!
%tmp.1 = icmp slt i16 %tmp.0, 1234
br i1 %tmp.1, label %then, label %UnifiedReturnBlock
then:
- call i32 %test1(i16 0 sext)
+ call i32 @test1(i16 0 sext)
ret void
UnifiedReturnBlock:
ret void
}
-define i32 %test4(i16* %P) {
+define i32 @test4(i16* %P) {
%tmp.1 = load i16* %P
%tmp.2 = zext i16 %tmp.1 to i32
%tmp.3 = and i32 %tmp.2, 255
ret i32 %tmp.3
}
-define i32 %test5(i16* %P) {
+define i32 @test5(i16* %P) {
%tmp.1 = load i16* %P
%tmp.2 = bitcast i16 %tmp.1 to i16
%tmp.3 = zext i16 %tmp.2 to i32
@@ -41,13 +41,13 @@
ret i32 %tmp.4
}
-define i32 %test6(i32* %P) {
+define i32 @test6(i32* %P) {
%tmp.1 = load i32* %P
%tmp.2 = and i32 %tmp.1, 255
ret i32 %tmp.2
}
-define i16 %test7(float %a) zext {
+define i16 @test7(float %a) zext {
%tmp.1 = fptoui float %a to i16
ret i16 %tmp.1
}
diff --git a/test/CodeGen/PowerPC/unsafe-math.ll b/test/CodeGen/PowerPC/unsafe-math.ll
index 29a4598..583c542 100644
--- a/test/CodeGen/PowerPC/unsafe-math.ll
+++ b/test/CodeGen/PowerPC/unsafe-math.ll
@@ -1,7 +1,7 @@
; RUN: llvm-as < %s | llc -march=ppc32 | grep fmul | wc -l | grep 2 &&
; RUN: llvm-as < %s | llc -march=ppc32 -enable-unsafe-fp-math | grep fmul | wc -l | grep 1
-define double %foo(double %X) {
+define double @foo(double %X) {
%tmp1 = mul double %X, 1.23
%tmp2 = mul double %tmp1, 4.124
ret double %tmp2
diff --git a/test/CodeGen/X86/2006-05-11-InstrSched.ll b/test/CodeGen/X86/2006-05-11-InstrSched.ll
index 40dd61a..e924142 100644
--- a/test/CodeGen/X86/2006-05-11-InstrSched.ll
+++ b/test/CodeGen/X86/2006-05-11-InstrSched.ll
@@ -1,5 +1,7 @@
-; RUN: llvm-upgrade < %s | llvm-as | llc -march=x86 -mattr=+sse2 -stats 2>&1 | grep 'asm-printer' | grep 39
+; RUN: llvm-upgrade < %s | llvm-as | llc -march=x86 -mattr=+sse2 -stats 2>&1 |\
+; RUN: grep 'asm-printer' | grep 39
+target datalayout = "e-p:32:32"
void %foo(int* %mc, int* %bp, int* %ms, int* %xmb, int* %mpp, int* %tpmm, int* %ip, int* %tpim, int* %dpp, int* %tpdm, int* %bpi, int %M) {
entry:
%tmp9 = setlt int %M, 5 ; <bool> [#uses=1]
diff --git a/test/CodeGen/X86/2007-01-08-InstrSched.ll b/test/CodeGen/X86/2007-01-08-InstrSched.ll
index 800eec2..daf283e 100644
--- a/test/CodeGen/X86/2007-01-08-InstrSched.ll
+++ b/test/CodeGen/X86/2007-01-08-InstrSched.ll
@@ -2,7 +2,7 @@
; RUN: %prcontext 'mulss LCPI1_3' 1 | grep mulss | wc -l | grep 1
; PR1075
-define float %foo(float %x) {
+define float @foo(float %x) {
%tmp1 = mul float %x, 3.000000e+00
%tmp3 = mul float %x, 5.000000e+00
%tmp5 = mul float %x, 7.000000e+00
diff --git a/test/CodeGen/X86/2007-01-13-StackPtrIndex.ll b/test/CodeGen/X86/2007-01-13-StackPtrIndex.ll
index 9a73130..9e218b5 100644
--- a/test/CodeGen/X86/2007-01-13-StackPtrIndex.ll
+++ b/test/CodeGen/X86/2007-01-13-StackPtrIndex.ll
@@ -3,13 +3,11 @@
; PR1103
target datalayout = "e-p:64:64"
-target endian = little
-target pointersize = 64
-%i6000 = global [128 x i64] zeroinitializer, align 16
+@i6000 = global [128 x i64] zeroinitializer, align 16
implementation
-define void %foo(i32* %a0, i32* %a1, i32* %a2, i32* %a3, i32* %a4, i32* %a5) {
+define void @foo(i32* %a0, i32* %a1, i32* %a2, i32* %a3, i32* %a4, i32* %a5) {
b:
%r = load i32* %a0
%r2 = load i32* %a1
@@ -135,14 +133,14 @@
%d753 = bitcast i64 %w1874 to i64
%r343 = add i64 %s661, 0
%r346 = add i64 %r343, 0
- %r347 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r346
+ %r347 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r346
%r348 = load float* %r347
%r352 = add i64 %r343, 0
- %r353 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r352
+ %r353 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r352
%r354 = load float* %r353
- %r362 = load float* bitcast ([128 x i64]* %i6000 to float*)
+ %r362 = load float* bitcast ([128 x i64]* @i6000 to float*)
%r363 = add float 0.000000e+00, %r362
- %r370 = load float* bitcast ([128 x i64]* %i6000 to float*)
+ %r370 = load float* bitcast ([128 x i64]* @i6000 to float*)
%r376 = icmp slt i64 %r16, 0
br i1 %r376, label %b377, label %a35b
b377:
@@ -186,9 +184,9 @@
%s923 = phi i64 [ 0, %b514 ], [ %r799, %b712 ]
%s933 = phi i64 [ %r533, %b514 ], [ %r795, %b712 ]
%r538 = add i64 %w1855, 0
- %r539 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r538
+ %r539 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r538
%r540 = load float* %r539
- %r551 = load float* bitcast ([128 x i64]* %i6000 to float*)
+ %r551 = load float* bitcast ([128 x i64]* @i6000 to float*)
%r562 = sub i64 %s933, 0
%r564 = icmp slt i64 %r512, 0
br i1 %r564, label %b565, label %a45b
@@ -215,22 +213,22 @@
a45b714:
%r717 = add i64 %e944, 0
%r720 = add i64 %r717, 0
- %r721 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r720
+ %r721 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r720
%r722 = load float* %r721
%r726 = add i64 %r717, 0
- %r727 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r726
+ %r727 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r726
%r728 = load float* %r727
%r732 = add i64 %r717, 0
- %r733 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r732
+ %r733 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r732
%r734 = load float* %r733
%r738 = add i64 %r717, 0
- %r739 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r738
+ %r739 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r738
%r740 = load float* %r739
%r744 = add i64 %r717, 0
- %r745 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r744
+ %r745 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r744
%r746 = load float* %r745
%r750 = add i64 %r717, 0
- %r751 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r750
+ %r751 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r750
%r752 = load float* %r751
%r753 = add float %r752, %r746
%r754 = add float %r728, %r722
@@ -239,10 +237,10 @@
%r757 = add float %r753, %r756
%r759 = add float %r757, %r540
%r770 = add i64 %r717, 0
- %r771 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r770
+ %r771 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r770
%r772 = load float* %r771
%r776 = add i64 %r717, 0
- %r777 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r776
+ %r777 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r776
%r778 = load float* %r777
%r781 = add float %r363, %r772
%r782 = add float %r781, %r778
@@ -256,7 +254,7 @@
%r844 = add i64 %r16, 0
%r846 = sext i32 %r60 to i64
%r847 = add i64 %r846, 0
- %r851 = load float* bitcast ([128 x i64]* %i6000 to float*)
+ %r851 = load float* bitcast ([128 x i64]* @i6000 to float*)
%r856 = sub i64 %r, 0
br label %b858
b858:
@@ -268,10 +266,10 @@
%s1173 = add i64 %b1902, 0
%r859 = add i64 %r856, 0
%r862 = add i64 %w1891, 0
- %r863 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r862
+ %r863 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r862
%r864 = load float* %r863
%r868 = add i64 %w1891, 0
- %r869 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r868
+ %r869 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r868
%r870 = load float* %r869
%r873 = sub i64 %r859, 0
%r876 = sub i64 %s1173, 0
@@ -338,7 +336,7 @@
%d1353 = bitcast i64 %w1915 to i64
%r1120 = add i64 %s661, 0
%r1121 = add i64 %r1120, 0
- %r1122 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r1121
+ %r1122 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r1121
%r1123 = load float* %r1122
%r1132 = bitcast i8* %c22012 to float*
%r1134 = getelementptr float* %r1132, i64 %w1915
@@ -412,7 +410,7 @@
%r1352 = add i64 %s1523, 0
%r1355 = sub i64 %r1352, 0
%r1370 = add i64 %d1533, 0
- %r1371 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r1370
+ %r1371 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r1370
%r1372 = load float* %r1371
br label %a74b
a74b:
@@ -446,10 +444,10 @@
%r1756 = getelementptr float* %r1754, i64 %w1970
%r1757 = load float* %r1756
%r1761 = add i64 %r1622, 0
- %r1762 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r1761
+ %r1762 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r1761
%r1763 = load float* %r1762
%r1767 = add i64 %r1622, 0
- %r1768 = getelementptr float* bitcast ([128 x i64]* %i6000 to float*), i64 %r1767
+ %r1768 = getelementptr float* bitcast ([128 x i64]* @i6000 to float*), i64 %r1767
%r1772 = add float %r1763, 0.000000e+00
%r1773 = add float %r1772, 0.000000e+00
%r1809 = add float %r1757, 0.000000e+00
diff --git a/test/CodeGen/X86/sse-fcopysign.ll b/test/CodeGen/X86/sse-fcopysign.ll
index 50f6b9a..37d8643 100644
--- a/test/CodeGen/X86/sse-fcopysign.ll
+++ b/test/CodeGen/X86/sse-fcopysign.ll
@@ -1,17 +1,17 @@
; RUN: llvm-as < %s | llc -march=x86 -mattr=+sse2 &&
; RUN: llvm-as < %s | llc -march=x86 -mattr=+sse2 | not getp test
-define float %test1(float %a, float %b) {
- %tmp = tail call float %copysignf( float %b, float %a )
+define float @test1(float %a, float %b) {
+ %tmp = tail call float @copysignf( float %b, float %a )
ret float %tmp
}
-define double %test2(double %a, float %b, float %c) {
+define double @test2(double %a, float %b, float %c) {
%tmp1 = add float %b, %c
%tmp2 = fpext float %tmp1 to double
- %tmp = tail call double %copysign( double %a, double %tmp2 )
+ %tmp = tail call double @copysign( double %a, double %tmp2 )
ret double %tmp
}
-declare float %copysignf(float, float)
-declare double %copysign(double, double)
+declare float @copysignf(float, float)
+declare double @copysign(double, double)
diff --git a/test/CodeGen/X86/test-hidden.ll b/test/CodeGen/X86/test-hidden.ll
index 915c3e8..27f890f 100644
--- a/test/CodeGen/X86/test-hidden.ll
+++ b/test/CodeGen/X86/test-hidden.ll
@@ -3,18 +3,18 @@
; RUN: llvm-as < %s | llc -mtriple=i686-apple-darwin8.8.0 | grep ".private_extern" | wc -l | grep 2
%struct.Person = type { i32 }
-%a = hidden global i32 0
-%b = external global i32
+@a = hidden global i32 0
+@b = external global i32
implementation ; Functions:
-define weak hidden void %_ZN6Person13privateMethodEv(%struct.Person* %this) {
+define weak hidden void @_ZN6Person13privateMethodEv(%struct.Person* %this) {
ret void
}
-declare void %function(i32)
+declare void @function(i32)
-define weak void %_ZN6PersonC1Ei(%struct.Person* %this, i32 %_c) {
+define weak void @_ZN6PersonC1Ei(%struct.Person* %this, i32 %_c) {
ret void
}
diff --git a/test/CodeGen/X86/test-pic-1.ll b/test/CodeGen/X86/test-pic-1.ll
index a12e742..81e283f 100644
--- a/test/CodeGen/X86/test-pic-1.ll
+++ b/test/CodeGen/X86/test-pic-1.ll
@@ -4,15 +4,15 @@
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep GOT | wc -l | grep 3 &&
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep GOTOFF | wc -l | grep 0
-%ptr = external global i32*
-%dst = external global i32
-%src = external global i32
+@ptr = external global i32*
+@dst = external global i32
+@src = external global i32
-define void %foo() {
+define void @foo() {
entry:
- store i32* %dst, i32** %ptr
- %tmp.s = load i32* %src
- store i32 %tmp.s, i32* %dst
+ store i32* @dst, i32** @ptr
+ %tmp.s = load i32* @src
+ store i32 %tmp.s, i32* @dst
ret void
}
diff --git a/test/CodeGen/X86/test-pic-2.ll b/test/CodeGen/X86/test-pic-2.ll
index 04fdbda..792c6d2 100644
--- a/test/CodeGen/X86/test-pic-2.ll
+++ b/test/CodeGen/X86/test-pic-2.ll
@@ -3,15 +3,15 @@
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep piclabel | wc -l | grep 3 &&
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep GOTOFF | wc -l | grep 4
-%ptr = internal global i32* null
-%dst = internal global i32 0
-%src = internal global i32 0
+@ptr = internal global i32* null
+@dst = internal global i32 0
+@src = internal global i32 0
-define void %foo() {
+define void @foo() {
entry:
- store i32* %dst, i32** %ptr
- %tmp.s = load i32* %src
- store i32 %tmp.s, i32* %dst
+ store i32* @dst, i32** @ptr
+ %tmp.s = load i32* @src
+ store i32 %tmp.s, i32* @dst
ret void
}
diff --git a/test/CodeGen/X86/test-pic-3.ll b/test/CodeGen/X86/test-pic-3.ll
index 726a53b..79dce93 100644
--- a/test/CodeGen/X86/test-pic-3.ll
+++ b/test/CodeGen/X86/test-pic-3.ll
@@ -3,12 +3,12 @@
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep piclabel | wc -l | grep 3 &&
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep PLT | wc -l | grep 1
-define void %bar() {
+define void @bar() {
entry:
- call void(...)* %foo()
+ call void(...)* @foo()
br label %return
return:
ret void
}
-declare void %foo(...)
+declare void @foo(...)
diff --git a/test/CodeGen/X86/test-pic-4.ll b/test/CodeGen/X86/test-pic-4.ll
index 13abb2e..8a12bcf 100644
--- a/test/CodeGen/X86/test-pic-4.ll
+++ b/test/CodeGen/X86/test-pic-4.ll
@@ -5,17 +5,17 @@
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep "GOT" | wc -l | grep 1 &&
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep "GOTOFF" | wc -l | grep 0
-%pfoo = external global void(...)*
+@pfoo = external global void(...)*
-define void %bar() {
+define void @bar() {
entry:
- %tmp = call void(...)*(...)* %afoo()
- store void(...)* %tmp, void(...)** %pfoo
- %tmp1 = load void(...)** %pfoo
+ %tmp = call void(...)*(...)* @afoo()
+ store void(...)* %tmp, void(...)** @pfoo
+ %tmp1 = load void(...)** @pfoo
call void(...)* %tmp1()
br label %return
return:
ret void
}
-declare void(...)* %afoo(...)
+declare void(...)* @afoo(...)
diff --git a/test/CodeGen/X86/test-pic-5.ll b/test/CodeGen/X86/test-pic-5.ll
index 4517e09..202ac9a 100644
--- a/test/CodeGen/X86/test-pic-5.ll
+++ b/test/CodeGen/X86/test-pic-5.ll
@@ -3,9 +3,9 @@
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep piclabel | wc -l | grep 3 &&
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep PLT | wc -l | grep 1
-%ptr = external global i32*
+@ptr = external global i32*
-define void %foo() {
+define void @foo() {
entry:
%ptr = malloc i32, i32 10
ret void
diff --git a/test/CodeGen/X86/test-pic-6.ll b/test/CodeGen/X86/test-pic-6.ll
index 9f29531..fa3190f 100644
--- a/test/CodeGen/X86/test-pic-6.ll
+++ b/test/CodeGen/X86/test-pic-6.ll
@@ -3,15 +3,15 @@
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep piclabel | wc -l | grep 3 &&
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep GOT | wc -l | grep 3
-%ptr = global i32* null
-%dst = global i32 0
-%src = global i32 0
+@ptr = global i32* null
+@dst = global i32 0
+@src = global i32 0
-define void %foo() {
+define void @foo() {
entry:
- store i32* %dst, i32** %ptr
- %tmp.s = load i32* %src
- store i32 %tmp.s, i32* %dst
+ store i32* @dst, i32** @ptr
+ %tmp.s = load i32* @src
+ store i32 %tmp.s, i32* @dst
ret void
}
diff --git a/test/CodeGen/X86/test-pic-cpool.ll b/test/CodeGen/X86/test-pic-cpool.ll
index 3911c69..f52c5ca 100644
--- a/test/CodeGen/X86/test-pic-cpool.ll
+++ b/test/CodeGen/X86/test-pic-cpool.ll
@@ -4,7 +4,7 @@
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep GOTOFF | wc -l | grep 2 &&
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep CPI | wc -l | grep 4
-define double %foo(i32 %a.u) {
+define double @foo(i32 %a.u) {
entry:
%tmp = icmp eq i32 %a.u,0
%retval = select i1 %tmp, double 4.561230e+02, double 1.234560e+02
diff --git a/test/CodeGen/X86/test-pic-jtbl.ll b/test/CodeGen/X86/test-pic-jtbl.ll
index ff3c4d6..b58e6de 100644
--- a/test/CodeGen/X86/test-pic-jtbl.ll
+++ b/test/CodeGen/X86/test-pic-jtbl.ll
@@ -5,53 +5,53 @@
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep GOTOFF | wc -l | grep 1 &&
; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic | grep JTI | wc -l | grep 15
-define void %bar(i32 %n.u) {
+define void @bar(i32 %n.u) {
entry:
switch i32 %n.u, label %bb12 [i32 1, label %bb i32 2, label %bb6 i32 4, label %bb7 i32 5, label %bb8 i32 6, label %bb10 i32 7, label %bb1 i32 8, label %bb3 i32 9, label %bb4 i32 10, label %bb9 i32 11, label %bb2 i32 12, label %bb5 i32 13, label %bb11 ]
bb:
- tail call void(...)* %foo1()
+ tail call void(...)* @foo1()
ret void
bb1:
- tail call void(...)* %foo2()
+ tail call void(...)* @foo2()
ret void
bb2:
- tail call void(...)* %foo6()
+ tail call void(...)* @foo6()
ret void
bb3:
- tail call void(...)* %foo3()
+ tail call void(...)* @foo3()
ret void
bb4:
- tail call void(...)* %foo4()
+ tail call void(...)* @foo4()
ret void
bb5:
- tail call void(...)* %foo5()
+ tail call void(...)* @foo5()
ret void
bb6:
- tail call void(...)* %foo1()
+ tail call void(...)* @foo1()
ret void
bb7:
- tail call void(...)* %foo2()
+ tail call void(...)* @foo2()
ret void
bb8:
- tail call void(...)* %foo6()
+ tail call void(...)* @foo6()
ret void
bb9:
- tail call void(...)* %foo3()
+ tail call void(...)* @foo3()
ret void
bb10:
- tail call void(...)* %foo4()
+ tail call void(...)* @foo4()
ret void
bb11:
- tail call void(...)* %foo5()
+ tail call void(...)* @foo5()
ret void
bb12:
- tail call void(...)* %foo6()
+ tail call void(...)* @foo6()
ret void
}
-declare void %foo1(...)
-declare void %foo2(...)
-declare void %foo6(...)
-declare void %foo3(...)
-declare void %foo4(...)
-declare void %foo5(...)
+declare void @foo1(...)
+declare void @foo2(...)
+declare void @foo6(...)
+declare void @foo3(...)
+declare void @foo4(...)
+declare void @foo5(...)
diff --git a/test/CodeGen/X86/trunc-to-bool.ll b/test/CodeGen/X86/trunc-to-bool.ll
index a9ac6c1..f2871f7 100644
--- a/test/CodeGen/X86/trunc-to-bool.ll
+++ b/test/CodeGen/X86/trunc-to-bool.ll
@@ -5,12 +5,12 @@
; RUN: llvm-as < %s | llc -march=x86 | grep '\(and\)\|\(test.*\$1\)' | \
; RUN: wc -l | grep 6
-define i1 %test1(i32 %X) zext {
+define i1 @test1(i32 %X) zext {
%Y = trunc i32 %X to i1
ret i1 %Y
}
-define i1 %test2(i32 %val, i32 %mask) {
+define i1 @test2(i32 %val, i32 %mask) {
entry:
%mask = trunc i32 %mask to i8
%shifted = ashr i32 %val, i8 %mask
@@ -23,7 +23,7 @@
ret i1 false
}
-define i32 %test3(i8* %ptr) {
+define i32 @test3(i8* %ptr) {
%val = load i8* %ptr
%tmp = trunc i8 %val to i1
br i1 %tmp, label %cond_true, label %cond_false
@@ -33,7 +33,7 @@
ret i32 42
}
-define i32 %test4(i8* %ptr) {
+define i32 @test4(i8* %ptr) {
%tmp = ptrtoint i8* %ptr to i1
br i1 %tmp, label %cond_true, label %cond_false
cond_true:
@@ -42,7 +42,7 @@
ret i32 42
}
-define i32 %test5(float %f) {
+define i32 @test5(float %f) {
%tmp = fptoui float %f to i1
br i1 %tmp, label %cond_true, label %cond_false
cond_true:
@@ -51,7 +51,7 @@
ret i32 42
}
-define i32 %test6(double %d) {
+define i32 @test6(double %d) {
%tmp = fptosi double %d to i1
br i1 %tmp, label %cond_true, label %cond_false
cond_true: