blob: 86721a20dd87012c94e5091c2c3d98a866dc345f [file] [log] [blame]
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -07001//===- subzero/src/IceTargetLowering.h - Lowering interface -----*- C++ -*-===//
2//
3// The Subzero Code Generator
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
Andrew Scull9612d322015-07-06 14:53:25 -07009///
10/// \file
Jim Stichnoth92a6e5b2015-12-02 16:52:44 -080011/// \brief Declares the TargetLowering, LoweringContext, and TargetDataLowering
12/// classes.
13///
14/// TargetLowering is an abstract class used to drive the translation/lowering
15/// process. LoweringContext maintains a context for lowering each instruction,
16/// offering conveniences such as iterating over non-deleted instructions.
17/// TargetDataLowering is an abstract class used to drive the lowering/emission
18/// of global initializers, external global declarations, and internal constant
19/// pools.
Andrew Scull9612d322015-07-06 14:53:25 -070020///
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -070021//===----------------------------------------------------------------------===//
22
23#ifndef SUBZERO_SRC_ICETARGETLOWERING_H
24#define SUBZERO_SRC_ICETARGETLOWERING_H
25
26#include "IceDefs.h"
John Portoe82b5602016-02-24 15:58:55 -080027#include "IceBitVector.h"
28#include "IceCfgNode.h"
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -070029#include "IceInst.h" // for the names of the Inst subtypes
Jan Voung76bb0be2015-05-14 09:26:19 -070030#include "IceOperand.h"
Jim Stichnotha18cc9c2014-09-30 19:10:22 -070031#include "IceTypes.h"
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -070032
John Porto1d937a82015-12-17 06:19:34 -080033#include <utility>
34
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -070035namespace Ice {
36
Karl Schimpfc5abdc12015-10-09 13:29:13 -070037// UnimplementedError is defined as a macro so that we can get actual line
38// numbers.
39#define UnimplementedError(Flags) \
40 do { \
41 if (!static_cast<const ClFlags &>(Flags).getSkipUnimplemented()) { \
42 /* Use llvm_unreachable instead of report_fatal_error, which gives \
43 better stack traces. */ \
44 llvm_unreachable("Not yet implemented"); \
45 abort(); \
46 } \
47 } while (0)
48
Jim Stichnoth91c773e2016-01-19 09:52:22 -080049// UnimplementedLoweringError is similar in style to UnimplementedError. Given
50// a TargetLowering object pointer and an Inst pointer, it adds appropriate
51// FakeDef and FakeUse instructions to try maintain liveness consistency.
52#define UnimplementedLoweringError(Target, Instr) \
53 do { \
54 if ((Target)->Ctx->getFlags().getSkipUnimplemented()) { \
55 (Target)->addFakeDefUses(Instr); \
56 } else { \
57 /* Use llvm_unreachable instead of report_fatal_error, which gives \
58 better stack traces. */ \
Eric Holke37076a2016-01-27 14:06:35 -080059 llvm_unreachable( \
60 ("Not yet implemented: " + Instr->getInstName()).c_str()); \
Jim Stichnoth91c773e2016-01-19 09:52:22 -080061 abort(); \
62 } \
63 } while (0)
64
Andrew Scull57e12682015-09-16 11:30:19 -070065/// LoweringContext makes it easy to iterate through non-deleted instructions in
66/// a node, and insert new (lowered) instructions at the current point. Along
67/// with the instruction list container and associated iterators, it holds the
68/// current node, which is needed when inserting new instructions in order to
69/// track whether variables are used as single-block or multi-block.
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -070070class LoweringContext {
Jim Stichnoth7b451a92014-10-15 14:39:23 -070071 LoweringContext(const LoweringContext &) = delete;
72 LoweringContext &operator=(const LoweringContext &) = delete;
73
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -070074public:
Jim Stichnotheafb56c2015-06-22 10:35:22 -070075 LoweringContext() = default;
76 ~LoweringContext() = default;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -070077 void init(CfgNode *Node);
78 Inst *getNextInst() const {
79 if (Next == End)
Jim Stichnothae953202014-12-20 06:17:49 -080080 return nullptr;
Jim Stichnoth607e9f02014-11-06 13:32:05 -080081 return Next;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -070082 }
Jan Voungc820ddf2014-07-29 14:38:51 -070083 Inst *getNextInst(InstList::iterator &Iter) const {
Jan Vounge6e497d2014-07-30 10:06:03 -070084 advanceForward(Iter);
Jan Voungc820ddf2014-07-29 14:38:51 -070085 if (Iter == End)
Jim Stichnothae953202014-12-20 06:17:49 -080086 return nullptr;
Jim Stichnoth607e9f02014-11-06 13:32:05 -080087 return Iter;
Jan Voungc820ddf2014-07-29 14:38:51 -070088 }
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -070089 CfgNode *getNode() const { return Node; }
90 bool atEnd() const { return Cur == End; }
91 InstList::iterator getCur() const { return Cur; }
Jim Stichnoth5d2fa0c2014-12-01 09:30:55 -080092 InstList::iterator getNext() const { return Next; }
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -070093 InstList::iterator getEnd() const { return End; }
Jim Stichnoth8cfeb692016-02-05 09:50:02 -080094 void insert(Inst *Instr);
John Porto1d937a82015-12-17 06:19:34 -080095 template <typename Inst, typename... Args> Inst *insert(Args &&... A) {
96 auto *New = Inst::create(Node->getCfg(), std::forward<Args>(A)...);
97 insert(New);
98 return New;
99 }
Jan Vounge6e497d2014-07-30 10:06:03 -0700100 Inst *getLastInserted() const;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700101 void advanceCur() { Cur = Next; }
Jan Vounge6e497d2014-07-30 10:06:03 -0700102 void advanceNext() { advanceForward(Next); }
Jim Stichnotha3f57b92015-07-30 12:46:04 -0700103 void setCur(InstList::iterator C) { Cur = C; }
104 void setNext(InstList::iterator N) { Next = N; }
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700105 void rewind();
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700106 void setInsertPoint(const InstList::iterator &Position) { Next = Position; }
Jim Stichnoth318f4cd2015-10-01 21:02:37 -0700107 void availabilityReset();
108 void availabilityUpdate();
109 Variable *availabilityGet(Operand *Src) const;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700110
111private:
Andrew Scull9612d322015-07-06 14:53:25 -0700112 /// Node is the argument to Inst::updateVars().
Jim Stichnotheafb56c2015-06-22 10:35:22 -0700113 CfgNode *Node = nullptr;
114 Inst *LastInserted = nullptr;
Andrew Scull57e12682015-09-16 11:30:19 -0700115 /// Cur points to the current instruction being considered. It is guaranteed
116 /// to point to a non-deleted instruction, or to be End.
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700117 InstList::iterator Cur;
Andrew Scull57e12682015-09-16 11:30:19 -0700118 /// Next doubles as a pointer to the next valid instruction (if any), and the
119 /// new-instruction insertion point. It is also updated for the caller in case
120 /// the lowering consumes more than one high-level instruction. It is
121 /// guaranteed to point to a non-deleted instruction after Cur, or to be End.
122 // TODO: Consider separating the notion of "next valid instruction" and "new
123 // instruction insertion point", to avoid confusion when previously-deleted
124 // instructions come between the two points.
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700125 InstList::iterator Next;
Andrew Scull9612d322015-07-06 14:53:25 -0700126 /// Begin is a copy of Insts.begin(), used if iterators are moved backward.
Jan Vounge6e497d2014-07-30 10:06:03 -0700127 InstList::iterator Begin;
Andrew Scull9612d322015-07-06 14:53:25 -0700128 /// End is a copy of Insts.end(), used if Next needs to be advanced.
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700129 InstList::iterator End;
Jim Stichnoth318f4cd2015-10-01 21:02:37 -0700130 /// LastDest and LastSrc capture the parameters of the last "Dest=Src" simple
131 /// assignment inserted (provided Src is a variable). This is used for simple
132 /// availability analysis.
133 Variable *LastDest = nullptr;
134 Variable *LastSrc = nullptr;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700135
Jan Voungc820ddf2014-07-29 14:38:51 -0700136 void skipDeleted(InstList::iterator &I) const;
Jan Vounge6e497d2014-07-30 10:06:03 -0700137 void advanceForward(InstList::iterator &I) const;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700138};
139
Jan Voung28068ad2015-07-31 12:58:46 -0700140/// A helper class to advance the LoweringContext at each loop iteration.
141class PostIncrLoweringContext {
142 PostIncrLoweringContext() = delete;
143 PostIncrLoweringContext(const PostIncrLoweringContext &) = delete;
144 PostIncrLoweringContext &operator=(const PostIncrLoweringContext &) = delete;
145
146public:
147 explicit PostIncrLoweringContext(LoweringContext &Context)
148 : Context(Context) {}
149 ~PostIncrLoweringContext() {
150 Context.advanceCur();
151 Context.advanceNext();
152 }
153
154private:
155 LoweringContext &Context;
156};
157
John Porto53611e22015-12-30 07:30:10 -0800158/// TargetLowering is the base class for all backends in Subzero. In addition to
159/// implementing the abstract methods in this class, each concrete target must
160/// also implement a named constructor in its own namespace. For instance, for
161/// X8632 we have:
162///
163/// namespace X8632 {
164/// void createTargetLowering(Cfg *Func);
165/// }
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700166class TargetLowering {
Jim Stichnothc6ead202015-02-24 09:30:30 -0800167 TargetLowering() = delete;
Jim Stichnoth7b451a92014-10-15 14:39:23 -0700168 TargetLowering(const TargetLowering &) = delete;
169 TargetLowering &operator=(const TargetLowering &) = delete;
170
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700171public:
Karl Schimpf5403f5d2016-01-15 11:07:46 -0800172 static void staticInit(GlobalContext *Ctx);
Jim Stichnoth8ff4b282016-01-04 15:39:06 -0800173 // Each target must define a public static method:
Karl Schimpf5403f5d2016-01-15 11:07:46 -0800174 // static void staticInit(GlobalContext *Ctx);
John Porto53611e22015-12-30 07:30:10 -0800175
176 static std::unique_ptr<TargetLowering> createLowering(TargetArch Target,
177 Cfg *Func);
178
179 virtual std::unique_ptr<Assembler> createAssembler() const = 0;
180
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700181 void translate() {
Jan Voung1f47ad02015-03-20 15:01:26 -0700182 switch (Ctx->getFlags().getOptLevel()) {
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700183 case Opt_m1:
184 translateOm1();
185 break;
186 case Opt_0:
187 translateO0();
188 break;
189 case Opt_1:
190 translateO1();
191 break;
192 case Opt_2:
193 translateO2();
194 break;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700195 }
196 }
197 virtual void translateOm1() {
198 Func->setError("Target doesn't specify Om1 lowering steps.");
199 }
200 virtual void translateO0() {
201 Func->setError("Target doesn't specify O0 lowering steps.");
202 }
203 virtual void translateO1() {
204 Func->setError("Target doesn't specify O1 lowering steps.");
205 }
206 virtual void translateO2() {
207 Func->setError("Target doesn't specify O2 lowering steps.");
208 }
209
John Porto5e0a8a72015-11-20 13:50:36 -0800210 /// Generates calls to intrinsics for operations the Target can't handle.
211 void genTargetHelperCalls();
Andrew Scull9612d322015-07-06 14:53:25 -0700212 /// Tries to do address mode optimization on a single instruction.
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700213 void doAddressOpt();
Andrew Scull9612d322015-07-06 14:53:25 -0700214 /// Randomly insert NOPs.
Qining Luaee5fa82015-08-20 14:59:03 -0700215 void doNopInsertion(RandomNumberGenerator &RNG);
Andrew Scull9612d322015-07-06 14:53:25 -0700216 /// Lowers a single non-Phi instruction.
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700217 void lower();
Jim Stichnotha3f57b92015-07-30 12:46:04 -0700218 /// Inserts and lowers a single high-level instruction at a specific insertion
219 /// point.
220 void lowerInst(CfgNode *Node, InstList::iterator Next, InstHighLevel *Instr);
Andrew Scull57e12682015-09-16 11:30:19 -0700221 /// Does preliminary lowering of the set of Phi instructions in the current
222 /// node. The main intention is to do what's needed to keep the unlowered Phi
223 /// instructions consistent with the lowered non-Phi instructions, e.g. to
224 /// lower 64-bit operands on a 32-bit target.
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700225 virtual void prelowerPhis() {}
Andrew Scull57e12682015-09-16 11:30:19 -0700226 /// Tries to do branch optimization on a single instruction. Returns true if
227 /// some optimization was done.
Jim Stichnothff9c7062014-09-18 04:50:49 -0700228 virtual bool doBranchOpt(Inst * /*I*/, const CfgNode * /*NextNode*/) {
229 return false;
230 }
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700231
Jim Stichnoth3d44fe82014-11-01 10:10:18 -0700232 virtual SizeT getNumRegisters() const = 0;
Andrew Scull57e12682015-09-16 11:30:19 -0700233 /// Returns a variable pre-colored to the specified physical register. This is
234 /// generally used to get very direct access to the register such as in the
235 /// prolog or epilog or for marking scratch registers as killed by a call. If
236 /// a Type is not provided, a target-specific default type is used.
Jim Stichnoth8aa39662016-02-10 11:20:30 -0800237 virtual Variable *getPhysicalRegister(RegNumT RegNum,
Jim Stichnoth98712a32014-10-24 10:59:02 -0700238 Type Ty = IceType_void) = 0;
Andrew Scull9612d322015-07-06 14:53:25 -0700239 /// Returns a printable name for the register.
Jim Stichnoth8aa39662016-02-10 11:20:30 -0800240 virtual IceString getRegName(RegNumT RegNum, Type Ty) const = 0;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700241
242 virtual bool hasFramePointer() const { return false; }
David Sehre39d0ca2015-11-06 11:25:41 -0800243 virtual void setHasFramePointer() = 0;
Jim Stichnoth8aa39662016-02-10 11:20:30 -0800244 virtual RegNumT getStackReg() const = 0;
245 virtual RegNumT getFrameReg() const = 0;
246 virtual RegNumT getFrameOrStackReg() const = 0;
Matt Walad4799f42014-08-14 14:24:12 -0700247 virtual size_t typeWidthInBytesOnStack(Type Ty) const = 0;
David Sehre39d0ca2015-11-06 11:25:41 -0800248 virtual uint32_t getStackAlignment() const = 0;
David Sehr2f3b8ec2015-11-16 16:51:39 -0800249 virtual void reserveFixedAllocaArea(size_t Size, size_t Align) = 0;
250 virtual int32_t getFrameFixedAllocaOffset() const = 0;
John Porto614140e2015-11-23 11:43:13 -0800251 virtual uint32_t maxOutArgsSizeBytes() const { return 0; }
Jan Voung0fa6c5a2015-06-01 11:04:04 -0700252
Andrew Scull6d47bcd2015-09-17 17:10:05 -0700253 /// Return whether a 64-bit Variable should be split into a Variable64On32.
254 virtual bool shouldSplitToVariable64On32(Type Ty) const = 0;
255
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700256 bool hasComputedFrame() const { return HasComputedFrame; }
Andrew Scull57e12682015-09-16 11:30:19 -0700257 /// Returns true if this function calls a function that has the "returns
258 /// twice" attribute.
Jan Voung44d53e12014-09-11 19:18:03 -0700259 bool callsReturnsTwice() const { return CallsReturnsTwice; }
Jim Stichnothdd842db2015-01-27 12:53:53 -0800260 void setCallsReturnsTwice(bool RetTwice) { CallsReturnsTwice = RetTwice; }
Jan Voungb36ad9b2015-04-21 17:01:49 -0700261 SizeT makeNextLabelNumber() { return NextLabelNumber++; }
Andrew Scull86df4e92015-07-30 13:54:44 -0700262 SizeT makeNextJumpTableNumber() { return NextJumpTableNumber++; }
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700263 LoweringContext &getContext() { return Context; }
Jim Stichnoth8ff4b282016-01-04 15:39:06 -0800264 Cfg *getFunc() const { return Func; }
265 GlobalContext *getGlobalContext() const { return Ctx; }
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700266
267 enum RegSet {
268 RegSet_None = 0,
269 RegSet_CallerSave = 1 << 0,
270 RegSet_CalleeSave = 1 << 1,
271 RegSet_StackPointer = 1 << 2,
272 RegSet_FramePointer = 1 << 3,
273 RegSet_All = ~RegSet_None
274 };
Andrew Scull8072bae2015-09-14 16:01:26 -0700275 using RegSetMask = uint32_t;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700276
John Portoe82b5602016-02-24 15:58:55 -0800277 virtual SmallBitVector getRegisterSet(RegSetMask Include,
278 RegSetMask Exclude) const = 0;
Jim Stichnothb40595a2016-01-29 06:14:31 -0800279 /// Get the set of physical registers available for the specified Variable's
280 /// register class, applying register restrictions from the command line.
John Portoe82b5602016-02-24 15:58:55 -0800281 virtual const SmallBitVector &
Jim Stichnothc59288b2015-11-09 11:38:40 -0800282 getRegistersForVariable(const Variable *Var) const = 0;
Jim Stichnothb40595a2016-01-29 06:14:31 -0800283 /// Get the set of *all* physical registers available for the specified
284 /// Variable's register class, *not* applying register restrictions from the
285 /// command line.
John Portoe82b5602016-02-24 15:58:55 -0800286 virtual const SmallBitVector &
Jim Stichnothb40595a2016-01-29 06:14:31 -0800287 getAllRegistersForVariable(const Variable *Var) const = 0;
John Portoe82b5602016-02-24 15:58:55 -0800288 virtual const SmallBitVector &getAliasesForRegister(RegNumT) const = 0;
John Portobb0a5fe2015-09-04 11:23:41 -0700289
Jim Stichnoth70d0a052014-11-14 15:53:46 -0800290 void regAlloc(RegAllocKind Kind);
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700291
Qining Luaee5fa82015-08-20 14:59:03 -0700292 virtual void
Jim Stichnoth8aa39662016-02-10 11:20:30 -0800293 makeRandomRegisterPermutation(llvm::SmallVectorImpl<RegNumT> &Permutation,
John Portoe82b5602016-02-24 15:58:55 -0800294 const SmallBitVector &ExcludeRegisters,
Qining Luaee5fa82015-08-20 14:59:03 -0700295 uint64_t Salt) const = 0;
Jim Stichnothe6d24782014-12-19 05:42:24 -0800296
Andrew Scull87f80c12015-07-20 10:19:16 -0700297 /// Get the minimum number of clusters required for a jump table to be
298 /// considered.
299 virtual SizeT getMinJumpTableSize() const = 0;
Andrew Scull86df4e92015-07-30 13:54:44 -0700300 virtual void emitJumpTable(const Cfg *Func,
301 const InstJumpTable *JumpTable) const = 0;
Andrew Scull87f80c12015-07-20 10:19:16 -0700302
Jim Stichnoth144cdce2014-09-22 16:02:59 -0700303 virtual void emitVariable(const Variable *Var) const = 0;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700304
Jim Stichnoth8ff4b282016-01-04 15:39:06 -0800305 void emitWithoutPrefix(const ConstantRelocatable *CR,
306 const char *Suffix = "") const;
Jan Voung76bb0be2015-05-14 09:26:19 -0700307
Jan Voung76bb0be2015-05-14 09:26:19 -0700308 virtual void emit(const ConstantInteger32 *C) const = 0;
309 virtual void emit(const ConstantInteger64 *C) const = 0;
310 virtual void emit(const ConstantFloat *C) const = 0;
311 virtual void emit(const ConstantDouble *C) const = 0;
Jim Stichnoth8ff4b282016-01-04 15:39:06 -0800312 virtual void emit(const ConstantUndef *C) const = 0;
313 virtual void emit(const ConstantRelocatable *CR) const = 0;
Jan Voung76bb0be2015-05-14 09:26:19 -0700314
Andrew Scull9612d322015-07-06 14:53:25 -0700315 /// Performs target-specific argument lowering.
Matt Wala45a06232014-07-09 16:33:22 -0700316 virtual void lowerArguments() = 0;
317
Jim Stichnotha59ae6f2015-05-17 10:11:41 -0700318 virtual void initNodeForLowering(CfgNode *) {}
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700319 virtual void addProlog(CfgNode *Node) = 0;
320 virtual void addEpilog(CfgNode *Node) = 0;
321
Jim Stichnotheafb56c2015-06-22 10:35:22 -0700322 virtual ~TargetLowering() = default;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700323
John Porto3bf335f2016-01-15 11:17:55 -0800324private:
325 // This control variable is used by AutoBundle (RAII-style bundle
326 // locking/unlocking) to prevent nested bundles.
327 bool AutoBundling = false;
328
Eric Holkd6cf6b32016-02-17 11:09:48 -0800329 /// This indicates whether we are in the genTargetHelperCalls phase, and
330 /// therefore can do things like scalarization.
331 bool GeneratingTargetHelpers = false;
332
John Porto3bf335f2016-01-15 11:17:55 -0800333 // _bundle_lock(), and _bundle_unlock(), were made private to force subtargets
334 // to use the AutoBundle helper.
335 void
336 _bundle_lock(InstBundleLock::Option BundleOption = InstBundleLock::Opt_None) {
337 Context.insert<InstBundleLock>(BundleOption);
338 }
339 void _bundle_unlock() { Context.insert<InstBundleUnlock>(); }
340
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700341protected:
John Porto3bf335f2016-01-15 11:17:55 -0800342 /// AutoBundle provides RIAA-style bundling. Sub-targets are expected to use
343 /// it when emitting NaCl Bundles to ensure proper bundle_unlocking, and
344 /// prevent nested bundles.
345 ///
346 /// AutoBundle objects will emit a _bundle_lock during construction (but only
347 /// if sandboxed code generation was requested), and a bundle_unlock() during
348 /// destruction. By carefully scoping objects of this type, Subtargets can
349 /// ensure proper bundle emission.
350 class AutoBundle {
351 AutoBundle() = delete;
352 AutoBundle(const AutoBundle &) = delete;
353 AutoBundle &operator=(const AutoBundle &) = delete;
354
355 public:
356 explicit AutoBundle(TargetLowering *Target, InstBundleLock::Option Option =
357 InstBundleLock::Opt_None);
358 ~AutoBundle();
359
360 private:
361 TargetLowering *const Target;
362 const bool NeedSandboxing;
363 };
364
Jim Stichnothc6ead202015-02-24 09:30:30 -0800365 explicit TargetLowering(Cfg *Func);
Karl Schimpf5403f5d2016-01-15 11:07:46 -0800366 // Applies command line filters to TypeToRegisterSet array.
367 static void
368 filterTypeToRegisterSet(GlobalContext *Ctx, int32_t NumRegs,
John Portoe82b5602016-02-24 15:58:55 -0800369 SmallBitVector TypeToRegisterSet[],
Karl Schimpf5403f5d2016-01-15 11:07:46 -0800370 size_t TypeToRegisterSetSize,
Jim Stichnoth8aa39662016-02-10 11:20:30 -0800371 std::function<IceString(RegNumT)> getRegName,
Jim Stichnoth2544d4d2016-01-22 13:07:46 -0800372 std::function<IceString(RegClass)> getRegClassName);
Jim Stichnoth8cfeb692016-02-05 09:50:02 -0800373 virtual void lowerAlloca(const InstAlloca *Instr) = 0;
374 virtual void lowerArithmetic(const InstArithmetic *Instr) = 0;
375 virtual void lowerAssign(const InstAssign *Instr) = 0;
376 virtual void lowerBr(const InstBr *Instr) = 0;
377 virtual void lowerCall(const InstCall *Instr) = 0;
378 virtual void lowerCast(const InstCast *Instr) = 0;
379 virtual void lowerFcmp(const InstFcmp *Instr) = 0;
380 virtual void lowerExtractElement(const InstExtractElement *Instr) = 0;
381 virtual void lowerIcmp(const InstIcmp *Instr) = 0;
382 virtual void lowerInsertElement(const InstInsertElement *Instr) = 0;
383 virtual void lowerIntrinsicCall(const InstIntrinsicCall *Instr) = 0;
384 virtual void lowerLoad(const InstLoad *Instr) = 0;
385 virtual void lowerPhi(const InstPhi *Instr) = 0;
386 virtual void lowerRet(const InstRet *Instr) = 0;
387 virtual void lowerSelect(const InstSelect *Instr) = 0;
388 virtual void lowerStore(const InstStore *Instr) = 0;
389 virtual void lowerSwitch(const InstSwitch *Instr) = 0;
390 virtual void lowerUnreachable(const InstUnreachable *Instr) = 0;
Jim Stichnothe4f65d82015-06-17 22:16:02 -0700391 virtual void lowerOther(const Inst *Instr);
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700392
John Porto5e0a8a72015-11-20 13:50:36 -0800393 virtual void genTargetHelperCallFor(Inst *Instr) = 0;
John Portof4198542015-11-20 14:17:23 -0800394 virtual uint32_t getCallStackArgumentsSizeBytes(const InstCall *Instr) = 0;
John Porto5e0a8a72015-11-20 13:50:36 -0800395
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700396 virtual void doAddressOptLoad() {}
397 virtual void doAddressOptStore() {}
Jim Stichnothad2989b2015-09-15 10:21:42 -0700398 virtual void doMockBoundsCheck(Operand *) {}
Qining Luaee5fa82015-08-20 14:59:03 -0700399 virtual void randomlyInsertNop(float Probability,
400 RandomNumberGenerator &RNG) = 0;
Andrew Scull57e12682015-09-16 11:30:19 -0700401 /// This gives the target an opportunity to post-process the lowered expansion
402 /// before returning.
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700403 virtual void postLower() {}
404
Jim Stichnoth91c773e2016-01-19 09:52:22 -0800405 /// When the SkipUnimplemented flag is set, addFakeDefUses() gets invoked by
406 /// the UnimplementedLoweringError macro to insert fake uses of all the
407 /// instruction variables and a fake def of the instruction dest, in order to
408 /// preserve integrity of liveness analysis.
409 void addFakeDefUses(const Inst *Instr);
410
Jim Stichnoth230d4102015-09-25 17:40:32 -0700411 /// Find (non-SSA) instructions where the Dest variable appears in some source
412 /// operand, and set the IsDestRedefined flag. This keeps liveness analysis
413 /// consistent.
414 void markRedefinitions();
Jan Voungb3401d22015-05-18 09:38:21 -0700415
Andrew Scull57e12682015-09-16 11:30:19 -0700416 /// Make a pass over the Cfg to determine which variables need stack slots and
417 /// place them in a sorted list (SortedSpilledVariables). Among those, vars,
418 /// classify the spill variables as local to the basic block vs global
419 /// (multi-block) in order to compute the parameters GlobalsSize and
420 /// SpillAreaSizeBytes (represents locals or general vars if the coalescing of
421 /// locals is disallowed) along with alignments required for variables in each
422 /// area. We rely on accurate VMetadata in order to classify a variable as
423 /// global vs local (otherwise the variable is conservatively global). The
424 /// in-args should be initialized to 0.
Andrew Scull9612d322015-07-06 14:53:25 -0700425 ///
Andrew Scull57e12682015-09-16 11:30:19 -0700426 /// This is only a pre-pass and the actual stack slot assignment is handled
427 /// separately.
Andrew Scull9612d322015-07-06 14:53:25 -0700428 ///
Andrew Scull57e12682015-09-16 11:30:19 -0700429 /// There may be target-specific Variable types, which will be handled by
430 /// TargetVarHook. If the TargetVarHook returns true, then the variable is
431 /// skipped and not considered with the rest of the spilled variables.
Jan Voung0fa6c5a2015-06-01 11:04:04 -0700432 void getVarStackSlotParams(VarList &SortedSpilledVariables,
John Portoe82b5602016-02-24 15:58:55 -0800433 SmallBitVector &RegsUsed, size_t *GlobalsSize,
434 size_t *SpillAreaSizeBytes,
Jan Voung0fa6c5a2015-06-01 11:04:04 -0700435 uint32_t *SpillAreaAlignmentBytes,
436 uint32_t *LocalsSlotsAlignmentBytes,
437 std::function<bool(Variable *)> TargetVarHook);
438
Andrew Scull57e12682015-09-16 11:30:19 -0700439 /// Calculate the amount of padding needed to align the local and global areas
440 /// to the required alignment. This assumes the globals/locals layout used by
441 /// getVarStackSlotParams and assignVarStackSlots.
Jan Voung0fa6c5a2015-06-01 11:04:04 -0700442 void alignStackSpillAreas(uint32_t SpillAreaStartOffset,
443 uint32_t SpillAreaAlignmentBytes,
444 size_t GlobalsSize,
445 uint32_t LocalsSlotsAlignmentBytes,
446 uint32_t *SpillAreaPaddingBytes,
447 uint32_t *LocalsSlotsPaddingBytes);
448
Andrew Scull57e12682015-09-16 11:30:19 -0700449 /// Make a pass through the SortedSpilledVariables and actually assign stack
450 /// slots. SpillAreaPaddingBytes takes into account stack alignment padding.
451 /// The SpillArea starts after that amount of padding. This matches the scheme
452 /// in getVarStackSlotParams, where there may be a separate multi-block global
453 /// var spill area and a local var spill area.
Jan Voung0fa6c5a2015-06-01 11:04:04 -0700454 void assignVarStackSlots(VarList &SortedSpilledVariables,
455 size_t SpillAreaPaddingBytes,
456 size_t SpillAreaSizeBytes,
457 size_t GlobalsAndSubsequentPaddingSize,
458 bool UsesFramePointer);
459
Andrew Scull57e12682015-09-16 11:30:19 -0700460 /// Sort the variables in Source based on required alignment. The variables
461 /// with the largest alignment need are placed in the front of the Dest list.
Jan Voung0fa6c5a2015-06-01 11:04:04 -0700462 void sortVarsByAlignment(VarList &Dest, const VarList &Source) const;
463
Andrew Scull9612d322015-07-06 14:53:25 -0700464 /// Make a call to an external helper function.
Jan Voungb36ad9b2015-04-21 17:01:49 -0700465 InstCall *makeHelperCall(const IceString &Name, Variable *Dest,
466 SizeT MaxSrcs);
467
Jim Stichnoth230d4102015-09-25 17:40:32 -0700468 void _set_dest_redefined() { Context.getLastInserted()->setDestRedefined(); }
Jan Voung0fa6c5a2015-06-01 11:04:04 -0700469
Andrew Scullcfa628b2015-08-20 14:23:05 -0700470 bool shouldOptimizeMemIntrins();
471
Eric Holkcfc25532016-02-09 17:47:58 -0800472 void scalarizeArithmetic(InstArithmetic::OpKind K, Variable *Dest,
473 Operand *Src0, Operand *Src1);
474
Eric Holkcc69fa22016-02-10 13:07:06 -0800475 /// Generalizes scalarizeArithmetic to support other instruction types.
476 ///
Eric Holkd6cf6b32016-02-17 11:09:48 -0800477 /// insertScalarInstruction is a function-like object with signature
Eric Holkcc69fa22016-02-10 13:07:06 -0800478 /// (Variable *Dest, Variable *Src0, Variable *Src1) -> Instr *.
Eric Holkd6cf6b32016-02-17 11:09:48 -0800479 template <typename... Operands,
480 typename F = std::function<Inst *(Variable *, Operands *...)>>
481 void scalarizeInstruction(Variable *Dest, F insertScalarInstruction,
482 Operands *... Srcs) {
483 assert(GeneratingTargetHelpers &&
484 "scalarizeInstruction called during incorrect phase");
Eric Holkcc69fa22016-02-10 13:07:06 -0800485 const Type DestTy = Dest->getType();
486 assert(isVectorType(DestTy));
487 const Type DestElementTy = typeElementType(DestTy);
488 const SizeT NumElements = typeNumElements(DestTy);
Eric Holkcc69fa22016-02-10 13:07:06 -0800489
490 Variable *T = Func->makeVariable(DestTy);
491 Context.insert<InstFakeDef>(T);
Eric Holkcc69fa22016-02-10 13:07:06 -0800492
Eric Holkd6cf6b32016-02-17 11:09:48 -0800493 for (SizeT I = 0; I < NumElements; ++I) {
494 auto *Index = Ctx->getConstantInt32(I);
495
496 auto makeExtractThunk = [this, Index, NumElements](Operand *Src) {
497 return [this, Index, NumElements, Src]() {
498 assert(typeNumElements(Src->getType()) == NumElements);
499
500 const auto ElementTy = typeElementType(Src->getType());
501 auto *Op = Func->makeVariable(ElementTy);
502 Context.insert<InstExtractElement>(Op, Src, Index);
503 return Op;
504 };
505 };
Eric Holkcc69fa22016-02-10 13:07:06 -0800506
507 // Perform the operation as a scalar operation.
Eric Holkd6cf6b32016-02-17 11:09:48 -0800508 auto *Res = Func->makeVariable(DestElementTy);
509 auto *Arith = applyToThunkedArgs(insertScalarInstruction, Res,
510 makeExtractThunk(Srcs)...);
Eric Holkcc69fa22016-02-10 13:07:06 -0800511 genTargetHelperCallFor(Arith);
512
Eric Holkcc69fa22016-02-10 13:07:06 -0800513 Variable *DestT = Func->makeVariable(DestTy);
514 Context.insert<InstInsertElement>(DestT, T, Res, Index);
515 T = DestT;
516 }
517 Context.insert<InstAssign>(Dest, T);
518 }
519
Eric Holkd6cf6b32016-02-17 11:09:48 -0800520 // applyToThunkedArgs is used by scalarizeInstruction. Ideally, we would just
521 // call insertScalarInstruction(Res, Srcs...), but C++ does not specify
522 // evaluation order which means this leads to an unpredictable final
523 // output. Instead, we wrap each of the Srcs in a thunk and these
524 // applyToThunkedArgs functions apply the thunks in a well defined order so we
525 // still get well-defined output.
526 Inst *applyToThunkedArgs(
527 std::function<Inst *(Variable *, Variable *)> insertScalarInstruction,
528 Variable *Res, std::function<Variable *()> thunk0) {
529 auto *Src0 = thunk0();
530 return insertScalarInstruction(Res, Src0);
531 }
Eric Holkcc69fa22016-02-10 13:07:06 -0800532
Eric Holkd6cf6b32016-02-17 11:09:48 -0800533 Inst *
534 applyToThunkedArgs(std::function<Inst *(Variable *, Variable *, Variable *)>
535 insertScalarInstruction,
536 Variable *Res, std::function<Variable *()> thunk0,
537 std::function<Variable *()> thunk1) {
538 auto *Src0 = thunk0();
539 auto *Src1 = thunk1();
540 return insertScalarInstruction(Res, Src0, Src1);
541 }
Eric Holkcc69fa22016-02-10 13:07:06 -0800542
Eric Holkd6cf6b32016-02-17 11:09:48 -0800543 Inst *applyToThunkedArgs(
544 std::function<Inst *(Variable *, Variable *, Variable *, Variable *)>
545 insertScalarInstruction,
546 Variable *Res, std::function<Variable *()> thunk0,
547 std::function<Variable *()> thunk1, std::function<Variable *()> thunk2) {
548 auto *Src0 = thunk0();
549 auto *Src1 = thunk1();
550 auto *Src2 = thunk2();
551 return insertScalarInstruction(Res, Src0, Src1, Src2);
Eric Holkcc69fa22016-02-10 13:07:06 -0800552 }
553
John Portoac2388c2016-01-22 07:10:56 -0800554 /// SandboxType enumerates all possible sandboxing strategies that
555 enum SandboxType {
556 ST_None,
557 ST_NaCl,
558 ST_Nonsfi,
559 };
560
561 static SandboxType determineSandboxTypeFromFlags(const ClFlags &Flags);
562
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700563 Cfg *Func;
564 GlobalContext *Ctx;
Jim Stichnotheafb56c2015-06-22 10:35:22 -0700565 bool HasComputedFrame = false;
566 bool CallsReturnsTwice = false;
Jim Stichnotheafb56c2015-06-22 10:35:22 -0700567 SizeT NextLabelNumber = 0;
Andrew Scull86df4e92015-07-30 13:54:44 -0700568 SizeT NextJumpTableNumber = 0;
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700569 LoweringContext Context;
John Portoac2388c2016-01-22 07:10:56 -0800570 const SandboxType SandboxingType = ST_None;
Jim Stichnoth9738a9e2015-02-23 16:39:06 -0800571
Jim Stichnothc4508792015-03-01 23:12:55 -0800572 // Runtime helper function names
573 const static constexpr char *H_bitcast_16xi1_i16 = "__Sz_bitcast_16xi1_i16";
574 const static constexpr char *H_bitcast_8xi1_i8 = "__Sz_bitcast_8xi1_i8";
575 const static constexpr char *H_bitcast_i16_16xi1 = "__Sz_bitcast_i16_16xi1";
576 const static constexpr char *H_bitcast_i8_8xi1 = "__Sz_bitcast_i8_8xi1";
577 const static constexpr char *H_call_ctpop_i32 = "__popcountsi2";
578 const static constexpr char *H_call_ctpop_i64 = "__popcountdi2";
579 const static constexpr char *H_call_longjmp = "longjmp";
580 const static constexpr char *H_call_memcpy = "memcpy";
581 const static constexpr char *H_call_memmove = "memmove";
582 const static constexpr char *H_call_memset = "memset";
583 const static constexpr char *H_call_read_tp = "__nacl_read_tp";
584 const static constexpr char *H_call_setjmp = "setjmp";
585 const static constexpr char *H_fptosi_f32_i64 = "__Sz_fptosi_f32_i64";
586 const static constexpr char *H_fptosi_f64_i64 = "__Sz_fptosi_f64_i64";
587 const static constexpr char *H_fptoui_4xi32_f32 = "__Sz_fptoui_4xi32_f32";
588 const static constexpr char *H_fptoui_f32_i32 = "__Sz_fptoui_f32_i32";
589 const static constexpr char *H_fptoui_f32_i64 = "__Sz_fptoui_f32_i64";
590 const static constexpr char *H_fptoui_f64_i32 = "__Sz_fptoui_f64_i32";
591 const static constexpr char *H_fptoui_f64_i64 = "__Sz_fptoui_f64_i64";
592 const static constexpr char *H_frem_f32 = "fmodf";
593 const static constexpr char *H_frem_f64 = "fmod";
Jim Stichnoth8ff4b282016-01-04 15:39:06 -0800594 const static constexpr char *H_getIP_prefix = "__Sz_getIP_";
Jan Voung6ec369e2015-06-30 11:03:15 -0700595 const static constexpr char *H_sdiv_i32 = "__divsi3";
Jim Stichnothc4508792015-03-01 23:12:55 -0800596 const static constexpr char *H_sdiv_i64 = "__divdi3";
597 const static constexpr char *H_sitofp_i64_f32 = "__Sz_sitofp_i64_f32";
598 const static constexpr char *H_sitofp_i64_f64 = "__Sz_sitofp_i64_f64";
Jan Voung6ec369e2015-06-30 11:03:15 -0700599 const static constexpr char *H_srem_i32 = "__modsi3";
Jim Stichnothc4508792015-03-01 23:12:55 -0800600 const static constexpr char *H_srem_i64 = "__moddi3";
Jan Voung6ec369e2015-06-30 11:03:15 -0700601 const static constexpr char *H_udiv_i32 = "__udivsi3";
Jim Stichnothc4508792015-03-01 23:12:55 -0800602 const static constexpr char *H_udiv_i64 = "__udivdi3";
603 const static constexpr char *H_uitofp_4xi32_4xf32 = "__Sz_uitofp_4xi32_4xf32";
604 const static constexpr char *H_uitofp_i32_f32 = "__Sz_uitofp_i32_f32";
605 const static constexpr char *H_uitofp_i32_f64 = "__Sz_uitofp_i32_f64";
606 const static constexpr char *H_uitofp_i64_f32 = "__Sz_uitofp_i64_f32";
607 const static constexpr char *H_uitofp_i64_f64 = "__Sz_uitofp_i64_f64";
Jan Voung6ec369e2015-06-30 11:03:15 -0700608 const static constexpr char *H_urem_i32 = "__umodsi3";
Jim Stichnothc4508792015-03-01 23:12:55 -0800609 const static constexpr char *H_urem_i64 = "__umoddi3";
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700610};
611
Andrew Scull57e12682015-09-16 11:30:19 -0700612/// TargetDataLowering is used for "lowering" data including initializers for
613/// global variables, and the internal constant pools. It is separated out from
614/// TargetLowering because it does not require a Cfg.
Jan Voung72984d82015-01-29 14:42:38 -0800615class TargetDataLowering {
616 TargetDataLowering() = delete;
617 TargetDataLowering(const TargetDataLowering &) = delete;
618 TargetDataLowering &operator=(const TargetDataLowering &) = delete;
Jim Stichnoth7b451a92014-10-15 14:39:23 -0700619
Jim Stichnothde4ca712014-06-29 08:13:48 -0700620public:
Jim Stichnothbbca7542015-02-11 16:08:31 -0800621 static std::unique_ptr<TargetDataLowering> createLowering(GlobalContext *Ctx);
Jan Voung72984d82015-01-29 14:42:38 -0800622 virtual ~TargetDataLowering();
Jan Voung839c4ce2014-07-28 15:19:43 -0700623
John Porto8b1a7052015-06-17 13:20:08 -0700624 virtual void lowerGlobals(const VariableDeclarationList &Vars,
625 const IceString &SectionSuffix) = 0;
John Porto0f86d032015-06-15 07:44:27 -0700626 virtual void lowerConstants() = 0;
Andrew Scull86df4e92015-07-30 13:54:44 -0700627 virtual void lowerJumpTables() = 0;
Jim Stichnothde4ca712014-06-29 08:13:48 -0700628
629protected:
John Porto8b1a7052015-06-17 13:20:08 -0700630 void emitGlobal(const VariableDeclaration &Var,
631 const IceString &SectionSuffix);
Jan Voung58eea4d2015-06-15 15:11:56 -0700632
Andrew Scull57e12682015-09-16 11:30:19 -0700633 /// For now, we assume .long is the right directive for emitting 4 byte emit
634 /// global relocations. However, LLVM MIPS usually uses .4byte instead.
Andrew Scull9612d322015-07-06 14:53:25 -0700635 /// Perhaps there is some difference when the location is unaligned.
John Porto8b1a7052015-06-17 13:20:08 -0700636 static const char *getEmit32Directive() { return ".long"; }
Jan Voung58eea4d2015-06-15 15:11:56 -0700637
Jim Stichnothc6ead202015-02-24 09:30:30 -0800638 explicit TargetDataLowering(GlobalContext *Ctx) : Ctx(Ctx) {}
Jim Stichnothde4ca712014-06-29 08:13:48 -0700639 GlobalContext *Ctx;
Jim Stichnothde4ca712014-06-29 08:13:48 -0700640};
641
Andrew Scull57e12682015-09-16 11:30:19 -0700642/// TargetHeaderLowering is used to "lower" the header of an output file. It
643/// writes out the target-specific header attributes. E.g., for ARM this writes
644/// out the build attributes (float ABI, etc.).
Jan Voungfb792842015-06-11 15:27:50 -0700645class TargetHeaderLowering {
646 TargetHeaderLowering() = delete;
647 TargetHeaderLowering(const TargetHeaderLowering &) = delete;
648 TargetHeaderLowering &operator=(const TargetHeaderLowering &) = delete;
649
650public:
651 static std::unique_ptr<TargetHeaderLowering>
652 createLowering(GlobalContext *Ctx);
653 virtual ~TargetHeaderLowering();
654
655 virtual void lower() {}
656
657protected:
658 explicit TargetHeaderLowering(GlobalContext *Ctx) : Ctx(Ctx) {}
659 GlobalContext *Ctx;
660};
661
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700662} // end of namespace Ice
663
664#endif // SUBZERO_SRC_ICETARGETLOWERING_H