[X86] Remove the Fv packed logical operation alias instructions. Replace them with patterns to the regular instructions.
This enables execution domain fixing which is why the tests changed.
llvm-svn: 278090
diff --git a/llvm/test/CodeGen/X86/copysign-constant-magnitude.ll b/llvm/test/CodeGen/X86/copysign-constant-magnitude.ll
index 6c577a2..539061b 100644
--- a/llvm/test/CodeGen/X86/copysign-constant-magnitude.ll
+++ b/llvm/test/CodeGen/X86/copysign-constant-magnitude.ll
@@ -17,28 +17,28 @@
; CHECK: id
%iX = call double @id_d(double %X)
-; CHECK-NEXT: andpd [[SIGNMASK]](%rip), %xmm0
+; CHECK-NEXT: andps [[SIGNMASK]](%rip), %xmm0
%d0 = call double @copysign(double 0.000000e+00, double %iX)
; CHECK-NEXT: id
%id0 = call double @id_d(double %d0)
-; CHECK-NEXT: andpd [[SIGNMASK]](%rip), %xmm0
-; CHECK-NEXT: orpd [[ZERO]](%rip), %xmm0
+; CHECK-NEXT: andps [[SIGNMASK]](%rip), %xmm0
+; CHECK-NEXT: orps [[ZERO]](%rip), %xmm0
%dn0 = call double @copysign(double -0.000000e+00, double %id0)
; CHECK-NEXT: id
%idn0 = call double @id_d(double %dn0)
-; CHECK-NEXT: andpd [[SIGNMASK]](%rip), %xmm0
-; CHECK-NEXT: orpd [[ONE]](%rip), %xmm0
+; CHECK-NEXT: andps [[SIGNMASK]](%rip), %xmm0
+; CHECK-NEXT: orps [[ONE]](%rip), %xmm0
%d1 = call double @copysign(double 1.000000e+00, double %idn0)
; CHECK-NEXT: id
%id1 = call double @id_d(double %d1)
-; CHECK-NEXT: andpd [[SIGNMASK]](%rip), %xmm0
-; CHECK-NEXT: orpd [[ONE]](%rip), %xmm0
+; CHECK-NEXT: andps [[SIGNMASK]](%rip), %xmm0
+; CHECK-NEXT: orps [[ONE]](%rip), %xmm0
%dn1 = call double @copysign(double -1.000000e+00, double %id1)
; CHECK-NEXT: id
diff --git a/llvm/test/CodeGen/X86/fp-logic.ll b/llvm/test/CodeGen/X86/fp-logic.ll
index 9ab6751..ce60d80 100644
--- a/llvm/test/CodeGen/X86/fp-logic.ll
+++ b/llvm/test/CodeGen/X86/fp-logic.ll
@@ -265,7 +265,7 @@
define double @bitcast_fabs(double %x) {
; CHECK-LABEL: bitcast_fabs:
; CHECK: # BB#0:
-; CHECK-NEXT: andpd {{.*}}(%rip), %xmm0
+; CHECK-NEXT: andps {{.*}}(%rip), %xmm0
; CHECK-NEXT: retq
;
%bc1 = bitcast double %x to i64
diff --git a/llvm/test/CodeGen/X86/fp128-cast.ll b/llvm/test/CodeGen/X86/fp128-cast.ll
index 2d87249..5bae3cb 100644
--- a/llvm/test/CodeGen/X86/fp128-cast.ll
+++ b/llvm/test/CodeGen/X86/fp128-cast.ll
@@ -351,8 +351,8 @@
;
; X64-LABEL: TestTruncCopysign:
; X64: callq __trunctfdf2
-; X64-NEXT: andpd {{.*}}, %xmm0
-; X64-NEXT: orpd {{.*}}, %xmm0
+; X64-NEXT: andps {{.*}}, %xmm0
+; X64-NEXT: orps {{.*}}, %xmm0
; X64-NEXT: callq __extenddftf2
; X64: retq
}
diff --git a/llvm/test/CodeGen/X86/sse-fcopysign.ll b/llvm/test/CodeGen/X86/sse-fcopysign.ll
index b01bbac..b84a589 100644
--- a/llvm/test/CodeGen/X86/sse-fcopysign.ll
+++ b/llvm/test/CodeGen/X86/sse-fcopysign.ll
@@ -108,9 +108,9 @@
; X64: # BB#0:
; X64-NEXT: addss %xmm2, %xmm1
; X64-NEXT: cvtss2sd %xmm1, %xmm1
-; X64-NEXT: andpd {{.*}}(%rip), %xmm1
-; X64-NEXT: andpd {{.*}}(%rip), %xmm0
-; X64-NEXT: orpd %xmm1, %xmm0
+; X64-NEXT: andps {{.*}}(%rip), %xmm1
+; X64-NEXT: andps {{.*}}(%rip), %xmm0
+; X64-NEXT: orps %xmm1, %xmm0
; X64-NEXT: retq
%tmp1 = fadd float %b, %c
%tmp2 = fpext float %tmp1 to double
diff --git a/llvm/test/CodeGen/X86/vec_fabs.ll b/llvm/test/CodeGen/X86/vec_fabs.ll
index 0f5e099..768c5ec 100644
--- a/llvm/test/CodeGen/X86/vec_fabs.ll
+++ b/llvm/test/CodeGen/X86/vec_fabs.ll
@@ -7,12 +7,12 @@
define <2 x double> @fabs_v2f64(<2 x double> %p) {
; X32-LABEL: fabs_v2f64:
; X32: # BB#0:
-; X32-NEXT: vandpd .LCPI0_0, %xmm0, %xmm0
+; X32-NEXT: vandps {{\.LCPI.*}}, %xmm0, %xmm0
; X32-NEXT: retl
;
; X64-LABEL: fabs_v2f64:
; X64: # BB#0:
-; X64-NEXT: vandpd {{.*}}(%rip), %xmm0, %xmm0
+; X64-NEXT: vandps {{.*}}(%rip), %xmm0, %xmm0
; X64-NEXT: retq
%t = call <2 x double> @llvm.fabs.v2f64(<2 x double> %p)
ret <2 x double> %t
@@ -22,7 +22,7 @@
define <4 x float> @fabs_v4f32(<4 x float> %p) {
; X32-LABEL: fabs_v4f32:
; X32: # BB#0:
-; X32-NEXT: vandps .LCPI1_0, %xmm0, %xmm0
+; X32-NEXT: vandps {{\.LCPI.*}}, %xmm0, %xmm0
; X32-NEXT: retl
;
; X64-LABEL: fabs_v4f32:
@@ -37,12 +37,12 @@
define <4 x double> @fabs_v4f64(<4 x double> %p) {
; X32-LABEL: fabs_v4f64:
; X32: # BB#0:
-; X32-NEXT: vandpd .LCPI2_0, %ymm0, %ymm0
+; X32-NEXT: vandps {{\.LCPI.*}}, %ymm0, %ymm0
; X32-NEXT: retl
;
; X64-LABEL: fabs_v4f64:
; X64: # BB#0:
-; X64-NEXT: vandpd {{.*}}(%rip), %ymm0, %ymm0
+; X64-NEXT: vandps {{.*}}(%rip), %ymm0, %ymm0
; X64-NEXT: retq
%t = call <4 x double> @llvm.fabs.v4f64(<4 x double> %p)
ret <4 x double> %t
@@ -52,7 +52,7 @@
define <8 x float> @fabs_v8f32(<8 x float> %p) {
; X32-LABEL: fabs_v8f32:
; X32: # BB#0:
-; X32-NEXT: vandps .LCPI3_0, %ymm0, %ymm0
+; X32-NEXT: vandps {{\.LCPI.*}}, %ymm0, %ymm0
; X32-NEXT: retl
;
; X64-LABEL: fabs_v8f32: