My compiler complains that "x always evaluates to true"
in this call:
Result.IntVal = APInt(80, 2, x);
What is x?
uint16_t x[8];
I deduce that the APInt constructor being used is this one:
APInt(uint32_t numBits, uint64_t val, bool isSigned = false);
rather than this one:
APInt(uint32_t numBits, uint32_t numWords, const uint64_t bigVal[]);
That doesn't seem right! This fix compiles but is otherwise completely
untested.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@44400 91177308-0d34-0410-b5e6-96231b3b80d8
diff --git a/lib/ExecutionEngine/ExecutionEngine.cpp b/lib/ExecutionEngine/ExecutionEngine.cpp
index 72db4e4..cc3cc38 100644
--- a/lib/ExecutionEngine/ExecutionEngine.cpp
+++ b/lib/ExecutionEngine/ExecutionEngine.cpp
@@ -712,13 +712,17 @@
break;
case Type::X86_FP80TyID: {
// This is endian dependent, but it will only work on x86 anyway.
- uint16_t x[8], *p = (uint16_t*)Ptr;
+ uint16_t *p = (uint16_t*)Ptr;
+ union {
+ uint16_t x[8];
+ uint64_t y[2];
+ };
x[0] = p[1];
x[1] = p[2];
x[2] = p[3];
x[3] = p[4];
x[4] = p[0];
- Result.IntVal = APInt(80, 2, x);
+ Result.IntVal = APInt(80, 2, y);
break;
}
default: