Chris Lattner | b0cfa6d | 2002-08-09 18:55:18 +0000 | [diff] [blame] | 1 | //===- SchedGraph.cpp - Scheduling Graph Implementation -------------------===// |
John Criswell | b576c94 | 2003-10-20 19:43:21 +0000 | [diff] [blame] | 2 | // |
| 3 | // The LLVM Compiler Infrastructure |
| 4 | // |
| 5 | // This file was developed by the LLVM research group and is distributed under |
| 6 | // the University of Illinois Open Source License. See LICENSE.TXT for details. |
| 7 | // |
| 8 | //===----------------------------------------------------------------------===// |
Chris Lattner | b0cfa6d | 2002-08-09 18:55:18 +0000 | [diff] [blame] | 9 | // |
| 10 | // Scheduling graph based on SSA graph plus extra dependence edges capturing |
| 11 | // dependences due to machine resources (machine registers, CC registers, and |
| 12 | // any others). |
| 13 | // |
| 14 | //===----------------------------------------------------------------------===// |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 15 | |
Chris Lattner | 46cbff6 | 2001-09-14 16:56:32 +0000 | [diff] [blame] | 16 | #include "SchedGraph.h" |
Chris Lattner | 2fbfdcf | 2002-04-07 20:49:59 +0000 | [diff] [blame] | 17 | #include "llvm/Function.h" |
Chris Lattner | b00c582 | 2001-10-02 03:41:24 +0000 | [diff] [blame] | 18 | #include "llvm/iOther.h" |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 19 | #include "llvm/CodeGen/MachineCodeForInstruction.h" |
| 20 | #include "llvm/CodeGen/MachineFunction.h" |
| 21 | #include "llvm/Target/TargetInstrInfo.h" |
| 22 | #include "llvm/Target/TargetMachine.h" |
| 23 | #include "llvm/Target/TargetRegInfo.h" |
| 24 | #include "Support/STLExtras.h" |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 25 | |
Brian Gaeke | d0fde30 | 2003-11-11 22:41:34 +0000 | [diff] [blame] | 26 | namespace llvm { |
| 27 | |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 28 | //*********************** Internal Data Structures *************************/ |
| 29 | |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 30 | // The following two types need to be classes, not typedefs, so we can use |
| 31 | // opaque declarations in SchedGraph.h |
| 32 | // |
Misha Brukman | c2312df | 2003-05-22 21:24:35 +0000 | [diff] [blame] | 33 | struct RefVec: public std::vector<std::pair<SchedGraphNode*, int> > { |
| 34 | typedef std::vector<std::pair<SchedGraphNode*,int> >::iterator iterator; |
| 35 | typedef |
| 36 | std::vector<std::pair<SchedGraphNode*,int> >::const_iterator const_iterator; |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 37 | }; |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 38 | |
Chris Lattner | 80c685f | 2001-10-13 06:51:01 +0000 | [diff] [blame] | 39 | struct RegToRefVecMap: public hash_map<int, RefVec> { |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 40 | typedef hash_map<int, RefVec>:: iterator iterator; |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 41 | typedef hash_map<int, RefVec>::const_iterator const_iterator; |
| 42 | }; |
| 43 | |
Vikram S. Adve | 74d15d3 | 2003-07-02 01:16:01 +0000 | [diff] [blame] | 44 | struct ValueToDefVecMap: public hash_map<const Value*, RefVec> { |
| 45 | typedef hash_map<const Value*, RefVec>:: iterator iterator; |
| 46 | typedef hash_map<const Value*, RefVec>::const_iterator const_iterator; |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 47 | }; |
| 48 | |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 49 | |
| 50 | // |
| 51 | // class SchedGraphNode |
| 52 | // |
| 53 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 54 | SchedGraphNode::SchedGraphNode(unsigned NID, MachineBasicBlock *mbb, |
| 55 | int indexInBB, const TargetMachine& Target) |
Tanya Lattner | 8dc9982 | 2003-08-28 15:30:40 +0000 | [diff] [blame] | 56 | : SchedGraphNodeCommon(NID,indexInBB), MBB(mbb), MI(mbb ? (*mbb)[indexInBB] : 0) { |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 57 | if (MI) { |
| 58 | MachineOpCode mopCode = MI->getOpCode(); |
| 59 | latency = Target.getInstrInfo().hasResultInterlock(mopCode) |
| 60 | ? Target.getInstrInfo().minLatency(mopCode) |
| 61 | : Target.getInstrInfo().maxLatency(mopCode); |
| 62 | } |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 63 | } |
| 64 | |
John Criswell | c9afb49 | 2003-08-28 21:43:17 +0000 | [diff] [blame] | 65 | // |
| 66 | // Method: SchedGraphNode Destructor |
| 67 | // |
| 68 | // Description: |
| 69 | // Free memory allocated by the SchedGraphNode object. |
| 70 | // |
| 71 | // Notes: |
| 72 | // Do not delete the edges here. The base class will take care of that. |
| 73 | // Only handle subclass specific stuff here (where currently there is |
| 74 | // none). |
| 75 | // |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 76 | SchedGraphNode::~SchedGraphNode() { |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 77 | } |
| 78 | |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 79 | // |
| 80 | // class SchedGraph |
| 81 | // |
Chris Lattner | fb3a0aed | 2002-10-28 18:50:08 +0000 | [diff] [blame] | 82 | SchedGraph::SchedGraph(MachineBasicBlock &mbb, const TargetMachine& target) |
| 83 | : MBB(mbb) { |
Chris Lattner | 697954c | 2002-01-20 22:54:45 +0000 | [diff] [blame] | 84 | buildGraph(target); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 85 | } |
| 86 | |
John Criswell | c9afb49 | 2003-08-28 21:43:17 +0000 | [diff] [blame] | 87 | // |
| 88 | // Method: SchedGraph Destructor |
| 89 | // |
| 90 | // Description: |
| 91 | // This method deletes memory allocated by the SchedGraph object. |
| 92 | // |
| 93 | // Notes: |
| 94 | // Do not delete the graphRoot or graphLeaf here. The base class handles |
| 95 | // that bit of work. |
| 96 | // |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 97 | SchedGraph::~SchedGraph() { |
Chris Lattner | 697954c | 2002-01-20 22:54:45 +0000 | [diff] [blame] | 98 | for (const_iterator I = begin(); I != end(); ++I) |
Chris Lattner | f3dd05c | 2002-04-09 05:15:33 +0000 | [diff] [blame] | 99 | delete I->second; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 100 | } |
| 101 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 102 | void SchedGraph::dump() const { |
Misha Brukman | c2312df | 2003-05-22 21:24:35 +0000 | [diff] [blame] | 103 | std::cerr << " Sched Graph for Basic Block: "; |
| 104 | std::cerr << MBB.getBasicBlock()->getName() |
| 105 | << " (" << MBB.getBasicBlock() << ")"; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 106 | |
Misha Brukman | c2312df | 2003-05-22 21:24:35 +0000 | [diff] [blame] | 107 | std::cerr << "\n\n Actual Root nodes : "; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 108 | for (unsigned i=0, N=graphRoot->outEdges.size(); i < N; i++) |
Misha Brukman | c2312df | 2003-05-22 21:24:35 +0000 | [diff] [blame] | 109 | std::cerr << graphRoot->outEdges[i]->getSink()->getNodeId() |
| 110 | << ((i == N-1)? "" : ", "); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 111 | |
Misha Brukman | c2312df | 2003-05-22 21:24:35 +0000 | [diff] [blame] | 112 | std::cerr << "\n Graph Nodes:\n"; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 113 | for (const_iterator I=begin(); I != end(); ++I) |
Misha Brukman | c2312df | 2003-05-22 21:24:35 +0000 | [diff] [blame] | 114 | std::cerr << "\n" << *I->second; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 115 | |
Misha Brukman | c2312df | 2003-05-22 21:24:35 +0000 | [diff] [blame] | 116 | std::cerr << "\n"; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 117 | } |
| 118 | |
| 119 | |
Vikram S. Adve | 8b6d245 | 2001-09-18 12:50:40 +0000 | [diff] [blame] | 120 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 121 | void SchedGraph::addDummyEdges() { |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 122 | assert(graphRoot->outEdges.size() == 0); |
| 123 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 124 | for (const_iterator I=begin(); I != end(); ++I) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 125 | SchedGraphNode* node = (*I).second; |
| 126 | assert(node != graphRoot && node != graphLeaf); |
| 127 | if (node->beginInEdges() == node->endInEdges()) |
| 128 | (void) new SchedGraphEdge(graphRoot, node, SchedGraphEdge::CtrlDep, |
| 129 | SchedGraphEdge::NonDataDep, 0); |
| 130 | if (node->beginOutEdges() == node->endOutEdges()) |
| 131 | (void) new SchedGraphEdge(node, graphLeaf, SchedGraphEdge::CtrlDep, |
| 132 | SchedGraphEdge::NonDataDep, 0); |
| 133 | } |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 134 | } |
| 135 | |
| 136 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 137 | void SchedGraph::addCDEdges(const TerminatorInst* term, |
| 138 | const TargetMachine& target) { |
Chris Lattner | 3501fea | 2003-01-14 22:00:31 +0000 | [diff] [blame] | 139 | const TargetInstrInfo& mii = target.getInstrInfo(); |
Chris Lattner | 0861b0c | 2002-02-03 07:29:45 +0000 | [diff] [blame] | 140 | MachineCodeForInstruction &termMvec = MachineCodeForInstruction::get(term); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 141 | |
| 142 | // Find the first branch instr in the sequence of machine instrs for term |
| 143 | // |
| 144 | unsigned first = 0; |
Vikram S. Adve | acf0f70 | 2002-10-13 00:39:22 +0000 | [diff] [blame] | 145 | while (! mii.isBranch(termMvec[first]->getOpCode()) && |
| 146 | ! mii.isReturn(termMvec[first]->getOpCode())) |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 147 | ++first; |
| 148 | assert(first < termMvec.size() && |
Vikram S. Adve | acf0f70 | 2002-10-13 00:39:22 +0000 | [diff] [blame] | 149 | "No branch instructions for terminator? Ok, but weird!"); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 150 | if (first == termMvec.size()) |
| 151 | return; |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 152 | |
Chris Lattner | b0cfa6d | 2002-08-09 18:55:18 +0000 | [diff] [blame] | 153 | SchedGraphNode* firstBrNode = getGraphNodeForInstr(termMvec[first]); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 154 | |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 155 | // Add CD edges from each instruction in the sequence to the |
| 156 | // *last preceding* branch instr. in the sequence |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 157 | // Use a latency of 0 because we only need to prevent out-of-order issue. |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 158 | // |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 159 | for (unsigned i = termMvec.size(); i > first+1; --i) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 160 | SchedGraphNode* toNode = getGraphNodeForInstr(termMvec[i-1]); |
| 161 | assert(toNode && "No node for instr generated for branch/ret?"); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 162 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 163 | for (unsigned j = i-1; j != 0; --j) |
| 164 | if (mii.isBranch(termMvec[j-1]->getOpCode()) || |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 165 | mii.isReturn(termMvec[j-1]->getOpCode())) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 166 | SchedGraphNode* brNode = getGraphNodeForInstr(termMvec[j-1]); |
| 167 | assert(brNode && "No node for instr generated for branch/ret?"); |
| 168 | (void) new SchedGraphEdge(brNode, toNode, SchedGraphEdge::CtrlDep, |
| 169 | SchedGraphEdge::NonDataDep, 0); |
| 170 | break; // only one incoming edge is enough |
| 171 | } |
| 172 | } |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 173 | |
| 174 | // Add CD edges from each instruction preceding the first branch |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 175 | // to the first branch. Use a latency of 0 as above. |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 176 | // |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 177 | for (unsigned i = first; i != 0; --i) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 178 | SchedGraphNode* fromNode = getGraphNodeForInstr(termMvec[i-1]); |
| 179 | assert(fromNode && "No node for instr generated for branch?"); |
| 180 | (void) new SchedGraphEdge(fromNode, firstBrNode, SchedGraphEdge::CtrlDep, |
| 181 | SchedGraphEdge::NonDataDep, 0); |
| 182 | } |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 183 | |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 184 | // Now add CD edges to the first branch instruction in the sequence from |
| 185 | // all preceding instructions in the basic block. Use 0 latency again. |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 186 | // |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 187 | for (unsigned i=0, N=MBB.size(); i < N; i++) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 188 | if (MBB[i] == termMvec[first]) // reached the first branch |
| 189 | break; |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 190 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 191 | SchedGraphNode* fromNode = this->getGraphNodeForInstr(MBB[i]); |
| 192 | if (fromNode == NULL) |
| 193 | continue; // dummy instruction, e.g., PHI |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 194 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 195 | (void) new SchedGraphEdge(fromNode, firstBrNode, |
| 196 | SchedGraphEdge::CtrlDep, |
| 197 | SchedGraphEdge::NonDataDep, 0); |
| 198 | |
| 199 | // If we find any other machine instructions (other than due to |
| 200 | // the terminator) that also have delay slots, add an outgoing edge |
| 201 | // from the instruction to the instructions in the delay slots. |
| 202 | // |
| 203 | unsigned d = mii.getNumDelaySlots(MBB[i]->getOpCode()); |
| 204 | assert(i+d < N && "Insufficient delay slots for instruction?"); |
| 205 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 206 | for (unsigned j=1; j <= d; j++) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 207 | SchedGraphNode* toNode = this->getGraphNodeForInstr(MBB[i+j]); |
| 208 | assert(toNode && "No node for machine instr in delay slot?"); |
| 209 | (void) new SchedGraphEdge(fromNode, toNode, |
Vikram S. Adve | 200a435 | 2001-11-12 18:53:43 +0000 | [diff] [blame] | 210 | SchedGraphEdge::CtrlDep, |
| 211 | SchedGraphEdge::NonDataDep, 0); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 212 | } |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 213 | } |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 214 | } |
| 215 | |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 216 | static const int SG_LOAD_REF = 0; |
| 217 | static const int SG_STORE_REF = 1; |
| 218 | static const int SG_CALL_REF = 2; |
| 219 | |
| 220 | static const unsigned int SG_DepOrderArray[][3] = { |
| 221 | { SchedGraphEdge::NonDataDep, |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 222 | SchedGraphEdge::AntiDep, |
| 223 | SchedGraphEdge::AntiDep }, |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 224 | { SchedGraphEdge::TrueDep, |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 225 | SchedGraphEdge::OutputDep, |
| 226 | SchedGraphEdge::TrueDep | SchedGraphEdge::OutputDep }, |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 227 | { SchedGraphEdge::TrueDep, |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 228 | SchedGraphEdge::AntiDep | SchedGraphEdge::OutputDep, |
| 229 | SchedGraphEdge::TrueDep | SchedGraphEdge::AntiDep |
| 230 | | SchedGraphEdge::OutputDep } |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 231 | }; |
| 232 | |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 233 | |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 234 | // Add a dependence edge between every pair of machine load/store/call |
| 235 | // instructions, where at least one is a store or a call. |
| 236 | // Use latency 1 just to ensure that memory operations are ordered; |
| 237 | // latency does not otherwise matter (true dependences enforce that). |
| 238 | // |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 239 | void SchedGraph::addMemEdges(const std::vector<SchedGraphNode*>& memNodeVec, |
| 240 | const TargetMachine& target) { |
Chris Lattner | 3501fea | 2003-01-14 22:00:31 +0000 | [diff] [blame] | 241 | const TargetInstrInfo& mii = target.getInstrInfo(); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 242 | |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 243 | // Instructions in memNodeVec are in execution order within the basic block, |
| 244 | // so simply look at all pairs <memNodeVec[i], memNodeVec[j: j > i]>. |
| 245 | // |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 246 | for (unsigned im=0, NM=memNodeVec.size(); im < NM; im++) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 247 | MachineOpCode fromOpCode = memNodeVec[im]->getOpCode(); |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 248 | int fromType = (mii.isCall(fromOpCode)? SG_CALL_REF |
| 249 | : (mii.isLoad(fromOpCode)? SG_LOAD_REF |
| 250 | : SG_STORE_REF)); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 251 | for (unsigned jm=im+1; jm < NM; jm++) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 252 | MachineOpCode toOpCode = memNodeVec[jm]->getOpCode(); |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 253 | int toType = (mii.isCall(toOpCode)? SG_CALL_REF |
| 254 | : (mii.isLoad(toOpCode)? SG_LOAD_REF |
| 255 | : SG_STORE_REF)); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 256 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 257 | if (fromType != SG_LOAD_REF || toType != SG_LOAD_REF) |
| 258 | (void) new SchedGraphEdge(memNodeVec[im], memNodeVec[jm], |
| 259 | SchedGraphEdge::MemoryDep, |
| 260 | SG_DepOrderArray[fromType][toType], 1); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 261 | } |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 262 | } |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 263 | } |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 264 | |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 265 | // Add edges from/to CC reg instrs to/from call instrs. |
| 266 | // Essentially this prevents anything that sets or uses a CC reg from being |
| 267 | // reordered w.r.t. a call. |
| 268 | // Use a latency of 0 because we only need to prevent out-of-order issue, |
| 269 | // like with control dependences. |
| 270 | // |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 271 | void SchedGraph::addCallDepEdges(const std::vector<SchedGraphNode*>& callDepNodeVec, |
| 272 | const TargetMachine& target) { |
Chris Lattner | 3501fea | 2003-01-14 22:00:31 +0000 | [diff] [blame] | 273 | const TargetInstrInfo& mii = target.getInstrInfo(); |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 274 | |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 275 | // Instructions in memNodeVec are in execution order within the basic block, |
| 276 | // so simply look at all pairs <memNodeVec[i], memNodeVec[j: j > i]>. |
| 277 | // |
| 278 | for (unsigned ic=0, NC=callDepNodeVec.size(); ic < NC; ic++) |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 279 | if (mii.isCall(callDepNodeVec[ic]->getOpCode())) { |
| 280 | // Add SG_CALL_REF edges from all preds to this instruction. |
| 281 | for (unsigned jc=0; jc < ic; jc++) |
| 282 | (void) new SchedGraphEdge(callDepNodeVec[jc], callDepNodeVec[ic], |
| 283 | SchedGraphEdge::MachineRegister, |
| 284 | MachineIntRegsRID, 0); |
| 285 | |
| 286 | // And do the same from this instruction to all successors. |
| 287 | for (unsigned jc=ic+1; jc < NC; jc++) |
| 288 | (void) new SchedGraphEdge(callDepNodeVec[ic], callDepNodeVec[jc], |
| 289 | SchedGraphEdge::MachineRegister, |
| 290 | MachineIntRegsRID, 0); |
| 291 | } |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 292 | |
| 293 | #ifdef CALL_DEP_NODE_VEC_CANNOT_WORK |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 294 | // Find the call instruction nodes and put them in a vector. |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 295 | std::vector<SchedGraphNode*> callNodeVec; |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 296 | for (unsigned im=0, NM=memNodeVec.size(); im < NM; im++) |
| 297 | if (mii.isCall(memNodeVec[im]->getOpCode())) |
| 298 | callNodeVec.push_back(memNodeVec[im]); |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 299 | |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 300 | // Now walk the entire basic block, looking for CC instructions *and* |
| 301 | // call instructions, and keep track of the order of the instructions. |
| 302 | // Use the call node vec to quickly find earlier and later call nodes |
| 303 | // relative to the current CC instruction. |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 304 | // |
| 305 | int lastCallNodeIdx = -1; |
| 306 | for (unsigned i=0, N=bbMvec.size(); i < N; i++) |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 307 | if (mii.isCall(bbMvec[i]->getOpCode())) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 308 | ++lastCallNodeIdx; |
| 309 | for ( ; lastCallNodeIdx < (int)callNodeVec.size(); ++lastCallNodeIdx) |
| 310 | if (callNodeVec[lastCallNodeIdx]->getMachineInstr() == bbMvec[i]) |
| 311 | break; |
| 312 | assert(lastCallNodeIdx < (int)callNodeVec.size() && "Missed Call?"); |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 313 | } |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 314 | else if (mii.isCCInstr(bbMvec[i]->getOpCode())) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 315 | // Add incoming/outgoing edges from/to preceding/later calls |
| 316 | SchedGraphNode* ccNode = this->getGraphNodeForInstr(bbMvec[i]); |
| 317 | int j=0; |
| 318 | for ( ; j <= lastCallNodeIdx; j++) |
| 319 | (void) new SchedGraphEdge(callNodeVec[j], ccNode, |
| 320 | MachineCCRegsRID, 0); |
| 321 | for ( ; j < (int) callNodeVec.size(); j++) |
| 322 | (void) new SchedGraphEdge(ccNode, callNodeVec[j], |
| 323 | MachineCCRegsRID, 0); |
| 324 | } |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 325 | #endif |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 326 | } |
| 327 | |
| 328 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 329 | void SchedGraph::addMachineRegEdges(RegToRefVecMap& regToRefVecMap, |
| 330 | const TargetMachine& target) { |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 331 | // This code assumes that two registers with different numbers are |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 332 | // not aliased! |
| 333 | // |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 334 | for (RegToRefVecMap::iterator I = regToRefVecMap.begin(); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 335 | I != regToRefVecMap.end(); ++I) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 336 | int regNum = (*I).first; |
| 337 | RefVec& regRefVec = (*I).second; |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 338 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 339 | // regRefVec is ordered by control flow order in the basic block |
| 340 | for (unsigned i=0; i < regRefVec.size(); ++i) { |
| 341 | SchedGraphNode* node = regRefVec[i].first; |
| 342 | unsigned int opNum = regRefVec[i].second; |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 343 | const MachineOperand& mop = |
| 344 | node->getMachineInstr()->getExplOrImplOperand(opNum); |
Alkis Evlogimenos | 4d7af65 | 2003-12-14 13:24:17 +0000 | [diff] [blame^] | 345 | bool isDef = mop.isDef() && !mop.isUse(); |
| 346 | bool isDefAndUse = mop.isDef() && mop.isUse(); |
Vikram S. Adve | 0baf1c0 | 2002-07-08 22:59:23 +0000 | [diff] [blame] | 347 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 348 | for (unsigned p=0; p < i; ++p) { |
| 349 | SchedGraphNode* prevNode = regRefVec[p].first; |
| 350 | if (prevNode != node) { |
| 351 | unsigned int prevOpNum = regRefVec[p].second; |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 352 | const MachineOperand& prevMop = |
| 353 | prevNode->getMachineInstr()->getExplOrImplOperand(prevOpNum); |
Alkis Evlogimenos | 4d7af65 | 2003-12-14 13:24:17 +0000 | [diff] [blame^] | 354 | bool prevIsDef = prevMop.isDef() && !prevMop.isUse(); |
| 355 | bool prevIsDefAndUse = prevMop.isDef() && prevMop.isUse(); |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 356 | if (isDef) { |
| 357 | if (prevIsDef) |
| 358 | new SchedGraphEdge(prevNode, node, regNum, |
| 359 | SchedGraphEdge::OutputDep); |
| 360 | if (!prevIsDef || prevIsDefAndUse) |
| 361 | new SchedGraphEdge(prevNode, node, regNum, |
| 362 | SchedGraphEdge::AntiDep); |
| 363 | } |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 364 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 365 | if (prevIsDef) |
| 366 | if (!isDef || isDefAndUse) |
| 367 | new SchedGraphEdge(prevNode, node, regNum, |
| 368 | SchedGraphEdge::TrueDep); |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 369 | } |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 370 | } |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 371 | } |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 372 | } |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 373 | } |
| 374 | |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 375 | |
Vikram S. Adve | 0baf1c0 | 2002-07-08 22:59:23 +0000 | [diff] [blame] | 376 | // Adds dependences to/from refNode from/to all other defs |
| 377 | // in the basic block. refNode may be a use, a def, or both. |
| 378 | // We do not consider other uses because we are not building use-use deps. |
| 379 | // |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 380 | void SchedGraph::addEdgesForValue(SchedGraphNode* refNode, |
| 381 | const RefVec& defVec, |
| 382 | const Value* defValue, |
| 383 | bool refNodeIsDef, |
Alkis Evlogimenos | 4d7af65 | 2003-12-14 13:24:17 +0000 | [diff] [blame^] | 384 | bool refNodeIsUse, |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 385 | const TargetMachine& target) { |
Vikram S. Adve | 200a435 | 2001-11-12 18:53:43 +0000 | [diff] [blame] | 386 | // Add true or output dep edges from all def nodes before refNode in BB. |
| 387 | // Add anti or output dep edges to all def nodes after refNode. |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 388 | for (RefVec::const_iterator I=defVec.begin(), E=defVec.end(); I != E; ++I) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 389 | if ((*I).first == refNode) |
| 390 | continue; // Dont add any self-loops |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 391 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 392 | if ((*I).first->getOrigIndexInBB() < refNode->getOrigIndexInBB()) { |
| 393 | // (*).first is before refNode |
Alkis Evlogimenos | 4d7af65 | 2003-12-14 13:24:17 +0000 | [diff] [blame^] | 394 | if (refNodeIsDef && !refNodeIsUse) |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 395 | (void) new SchedGraphEdge((*I).first, refNode, defValue, |
| 396 | SchedGraphEdge::OutputDep); |
| 397 | if (refNodeIsUse) |
| 398 | (void) new SchedGraphEdge((*I).first, refNode, defValue, |
| 399 | SchedGraphEdge::TrueDep); |
| 400 | } else { |
| 401 | // (*).first is after refNode |
Alkis Evlogimenos | 4d7af65 | 2003-12-14 13:24:17 +0000 | [diff] [blame^] | 402 | if (refNodeIsDef && !refNodeIsUse) |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 403 | (void) new SchedGraphEdge(refNode, (*I).first, defValue, |
| 404 | SchedGraphEdge::OutputDep); |
| 405 | if (refNodeIsUse) |
| 406 | (void) new SchedGraphEdge(refNode, (*I).first, defValue, |
| 407 | SchedGraphEdge::AntiDep); |
Vikram S. Adve | 200a435 | 2001-11-12 18:53:43 +0000 | [diff] [blame] | 408 | } |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 409 | } |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 410 | } |
| 411 | |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 412 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 413 | void SchedGraph::addEdgesForInstruction(const MachineInstr& MI, |
| 414 | const ValueToDefVecMap& valueToDefVecMap, |
| 415 | const TargetMachine& target) { |
Chris Lattner | 133f079 | 2002-10-28 04:45:29 +0000 | [diff] [blame] | 416 | SchedGraphNode* node = getGraphNodeForInstr(&MI); |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 417 | if (node == NULL) |
| 418 | return; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 419 | |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 420 | // Add edges for all operands of the machine instruction. |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 421 | // |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 422 | for (unsigned i = 0, numOps = MI.getNumOperands(); i != numOps; ++i) { |
| 423 | switch (MI.getOperand(i).getType()) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 424 | case MachineOperand::MO_VirtualRegister: |
| 425 | case MachineOperand::MO_CCRegister: |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 426 | if (const Value* srcI = MI.getOperand(i).getVRegValue()) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 427 | ValueToDefVecMap::const_iterator I = valueToDefVecMap.find(srcI); |
| 428 | if (I != valueToDefVecMap.end()) |
| 429 | addEdgesForValue(node, I->second, srcI, |
Alkis Evlogimenos | 4d7af65 | 2003-12-14 13:24:17 +0000 | [diff] [blame^] | 430 | MI.getOperand(i).isDef(), MI.getOperand(i).isUse(), |
| 431 | target); |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 432 | } |
| 433 | break; |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 434 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 435 | case MachineOperand::MO_MachineRegister: |
| 436 | break; |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 437 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 438 | case MachineOperand::MO_SignExtendedImmed: |
| 439 | case MachineOperand::MO_UnextendedImmed: |
| 440 | case MachineOperand::MO_PCRelativeDisp: |
Misha Brukman | e2bf0a2 | 2003-11-06 00:04:11 +0000 | [diff] [blame] | 441 | case MachineOperand::MO_ConstantPoolIndex: |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 442 | break; // nothing to do for immediate fields |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 443 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 444 | default: |
| 445 | assert(0 && "Unknown machine operand type in SchedGraph builder"); |
| 446 | break; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 447 | } |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 448 | } |
Vikram S. Adve | 8d0ffa5 | 2001-10-11 04:22:45 +0000 | [diff] [blame] | 449 | |
| 450 | // Add edges for values implicitly used by the machine instruction. |
| 451 | // Examples include function arguments to a Call instructions or the return |
| 452 | // value of a Ret instruction. |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 453 | // |
Chris Lattner | 133f079 | 2002-10-28 04:45:29 +0000 | [diff] [blame] | 454 | for (unsigned i=0, N=MI.getNumImplicitRefs(); i < N; ++i) |
Alkis Evlogimenos | 4d7af65 | 2003-12-14 13:24:17 +0000 | [diff] [blame^] | 455 | if (MI.getImplicitOp(i).isUse()) |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 456 | if (const Value* srcI = MI.getImplicitRef(i)) { |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 457 | ValueToDefVecMap::const_iterator I = valueToDefVecMap.find(srcI); |
| 458 | if (I != valueToDefVecMap.end()) |
| 459 | addEdgesForValue(node, I->second, srcI, |
Alkis Evlogimenos | 4d7af65 | 2003-12-14 13:24:17 +0000 | [diff] [blame^] | 460 | MI.getImplicitOp(i).isDef(), |
| 461 | MI.getImplicitOp(i).isUse(), target); |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 462 | } |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 463 | } |
| 464 | |
| 465 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 466 | void SchedGraph::findDefUseInfoAtInstr(const TargetMachine& target, |
| 467 | SchedGraphNode* node, |
| 468 | std::vector<SchedGraphNode*>& memNodeVec, |
| 469 | std::vector<SchedGraphNode*>& callDepNodeVec, |
| 470 | RegToRefVecMap& regToRefVecMap, |
| 471 | ValueToDefVecMap& valueToDefVecMap) { |
Chris Lattner | 3501fea | 2003-01-14 22:00:31 +0000 | [diff] [blame] | 472 | const TargetInstrInfo& mii = target.getInstrInfo(); |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 473 | |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 474 | MachineOpCode opCode = node->getOpCode(); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 475 | |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 476 | if (mii.isCall(opCode) || mii.isCCInstr(opCode)) |
| 477 | callDepNodeVec.push_back(node); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 478 | |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 479 | if (mii.isLoad(opCode) || mii.isStore(opCode) || mii.isCall(opCode)) |
| 480 | memNodeVec.push_back(node); |
| 481 | |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 482 | // Collect the register references and value defs. for explicit operands |
| 483 | // |
Tanya Lattner | b6489f3 | 2003-08-25 22:42:20 +0000 | [diff] [blame] | 484 | const MachineInstr& MI = *node->getMachineInstr(); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 485 | for (int i=0, numOps = (int) MI.getNumOperands(); i < numOps; i++) { |
Tanya Lattner | b6489f3 | 2003-08-25 22:42:20 +0000 | [diff] [blame] | 486 | const MachineOperand& mop = MI.getOperand(i); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 487 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 488 | // if this references a register other than the hardwired |
| 489 | // "zero" register, record the reference. |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 490 | if (mop.hasAllocatedReg()) { |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 491 | int regNum = mop.getAllocatedRegNum(); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 492 | |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 493 | // If this is not a dummy zero register, record the reference in order |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 494 | if (regNum != target.getRegInfo().getZeroRegNum()) |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 495 | regToRefVecMap[mop.getAllocatedRegNum()] |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 496 | .push_back(std::make_pair(node, i)); |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 497 | |
| 498 | // If this is a volatile register, add the instruction to callDepVec |
| 499 | // (only if the node is not already on the callDepVec!) |
| 500 | if (callDepNodeVec.size() == 0 || callDepNodeVec.back() != node) |
| 501 | { |
| 502 | unsigned rcid; |
| 503 | int regInClass = target.getRegInfo().getClassRegNum(regNum, rcid); |
| 504 | if (target.getRegInfo().getMachineRegClass(rcid) |
| 505 | ->isRegVolatile(regInClass)) |
| 506 | callDepNodeVec.push_back(node); |
| 507 | } |
| 508 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 509 | continue; // nothing more to do |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 510 | } |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 511 | |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 512 | // ignore all other non-def operands |
Alkis Evlogimenos | 4d7af65 | 2003-12-14 13:24:17 +0000 | [diff] [blame^] | 513 | if (!MI.getOperand(i).isDef()) |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 514 | continue; |
| 515 | |
| 516 | // We must be defining a value. |
| 517 | assert((mop.getType() == MachineOperand::MO_VirtualRegister || |
| 518 | mop.getType() == MachineOperand::MO_CCRegister) |
| 519 | && "Do not expect any other kind of operand to be defined!"); |
Vikram S. Adve | 74d15d3 | 2003-07-02 01:16:01 +0000 | [diff] [blame] | 520 | assert(mop.getVRegValue() != NULL && "Null value being defined?"); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 521 | |
Vikram S. Adve | 74d15d3 | 2003-07-02 01:16:01 +0000 | [diff] [blame] | 522 | valueToDefVecMap[mop.getVRegValue()].push_back(std::make_pair(node, i)); |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 523 | } |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 524 | |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 525 | // |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 526 | // Collect value defs. for implicit operands. They may have allocated |
| 527 | // physical registers also. |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 528 | // |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 529 | for (unsigned i=0, N = MI.getNumImplicitRefs(); i != N; ++i) { |
Tanya Lattner | b6489f3 | 2003-08-25 22:42:20 +0000 | [diff] [blame] | 530 | const MachineOperand& mop = MI.getImplicitOp(i); |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 531 | if (mop.hasAllocatedReg()) { |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 532 | int regNum = mop.getAllocatedRegNum(); |
| 533 | if (regNum != target.getRegInfo().getZeroRegNum()) |
| 534 | regToRefVecMap[mop.getAllocatedRegNum()] |
Tanya Lattner | b6489f3 | 2003-08-25 22:42:20 +0000 | [diff] [blame] | 535 | .push_back(std::make_pair(node, i + MI.getNumOperands())); |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 536 | continue; // nothing more to do |
| 537 | } |
| 538 | |
Alkis Evlogimenos | 4d7af65 | 2003-12-14 13:24:17 +0000 | [diff] [blame^] | 539 | if (mop.isDef()) { |
Tanya Lattner | b6489f3 | 2003-08-25 22:42:20 +0000 | [diff] [blame] | 540 | assert(MI.getImplicitRef(i) != NULL && "Null value being defined?"); |
Alkis Evlogimenos | 4d7af65 | 2003-12-14 13:24:17 +0000 | [diff] [blame^] | 541 | valueToDefVecMap[MI.getImplicitRef(i)].push_back( |
| 542 | std::make_pair(node, -i)); |
Vikram S. Adve | 74d15d3 | 2003-07-02 01:16:01 +0000 | [diff] [blame] | 543 | } |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 544 | } |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 545 | } |
| 546 | |
| 547 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 548 | void SchedGraph::buildNodesForBB(const TargetMachine& target, |
| 549 | MachineBasicBlock& MBB, |
| 550 | std::vector<SchedGraphNode*>& memNodeVec, |
| 551 | std::vector<SchedGraphNode*>& callDepNodeVec, |
| 552 | RegToRefVecMap& regToRefVecMap, |
| 553 | ValueToDefVecMap& valueToDefVecMap) { |
Chris Lattner | 3501fea | 2003-01-14 22:00:31 +0000 | [diff] [blame] | 554 | const TargetInstrInfo& mii = target.getInstrInfo(); |
Vikram S. Adve | 5b43af9 | 2001-11-11 01:23:27 +0000 | [diff] [blame] | 555 | |
| 556 | // Build graph nodes for each VM instruction and gather def/use info. |
| 557 | // Do both those together in a single pass over all machine instructions. |
Chris Lattner | fb3a0aed | 2002-10-28 18:50:08 +0000 | [diff] [blame] | 558 | for (unsigned i=0; i < MBB.size(); i++) |
| 559 | if (!mii.isDummyPhiInstr(MBB[i]->getOpCode())) { |
| 560 | SchedGraphNode* node = new SchedGraphNode(getNumNodes(), &MBB, i, target); |
| 561 | noteGraphNodeForInstr(MBB[i], node); |
| 562 | |
| 563 | // Remember all register references and value defs |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 564 | findDefUseInfoAtInstr(target, node, memNodeVec, callDepNodeVec, |
| 565 | regToRefVecMap, valueToDefVecMap); |
Chris Lattner | fb3a0aed | 2002-10-28 18:50:08 +0000 | [diff] [blame] | 566 | } |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 567 | } |
| 568 | |
| 569 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 570 | void SchedGraph::buildGraph(const TargetMachine& target) { |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 571 | // Use this data structure to note all machine operands that compute |
| 572 | // ordinary LLVM values. These must be computed defs (i.e., instructions). |
| 573 | // Note that there may be multiple machine instructions that define |
| 574 | // each Value. |
| 575 | ValueToDefVecMap valueToDefVecMap; |
| 576 | |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 577 | // Use this data structure to note all memory instructions. |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 578 | // We use this to add memory dependence edges without a second full walk. |
Misha Brukman | c2312df | 2003-05-22 21:24:35 +0000 | [diff] [blame] | 579 | std::vector<SchedGraphNode*> memNodeVec; |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 580 | |
| 581 | // Use this data structure to note all instructions that access physical |
| 582 | // registers that can be modified by a call (including call instructions) |
| 583 | std::vector<SchedGraphNode*> callDepNodeVec; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 584 | |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 585 | // Use this data structure to note any uses or definitions of |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 586 | // machine registers so we can add edges for those later without |
| 587 | // extra passes over the nodes. |
| 588 | // The vector holds an ordered list of references to the machine reg, |
| 589 | // ordered according to control-flow order. This only works for a |
| 590 | // single basic block, hence the assertion. Each reference is identified |
| 591 | // by the pair: <node, operand-number>. |
| 592 | // |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 593 | RegToRefVecMap regToRefVecMap; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 594 | |
| 595 | // Make a dummy root node. We'll add edges to the real roots later. |
Chris Lattner | fb3a0aed | 2002-10-28 18:50:08 +0000 | [diff] [blame] | 596 | graphRoot = new SchedGraphNode(0, NULL, -1, target); |
| 597 | graphLeaf = new SchedGraphNode(1, NULL, -1, target); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 598 | |
| 599 | //---------------------------------------------------------------- |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 600 | // First add nodes for all the machine instructions in the basic block |
| 601 | // because this greatly simplifies identifying which edges to add. |
| 602 | // Do this one VM instruction at a time since the SchedGraphNode needs that. |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 603 | // Also, remember the load/store instructions to add memory deps later. |
| 604 | //---------------------------------------------------------------- |
Chris Lattner | fb3a0aed | 2002-10-28 18:50:08 +0000 | [diff] [blame] | 605 | |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 606 | buildNodesForBB(target, MBB, memNodeVec, callDepNodeVec, |
| 607 | regToRefVecMap, valueToDefVecMap); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 608 | |
| 609 | //---------------------------------------------------------------- |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 610 | // Now add edges for the following (all are incoming edges except (4)): |
| 611 | // (1) operands of the machine instruction, including hidden operands |
| 612 | // (2) machine register dependences |
| 613 | // (3) memory load/store dependences |
| 614 | // (3) other resource dependences for the machine instruction, if any |
| 615 | // (4) output dependences when multiple machine instructions define the |
| 616 | // same value; all must have been generated from a single VM instrn |
| 617 | // (5) control dependences to branch instructions generated for the |
| 618 | // terminator instruction of the BB. Because of delay slots and |
| 619 | // 2-way conditional branches, multiple CD edges are needed |
| 620 | // (see addCDEdges for details). |
| 621 | // Also, note any uses or defs of machine registers. |
| 622 | // |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 623 | //---------------------------------------------------------------- |
| 624 | |
| 625 | // First, add edges to the terminator instruction of the basic block. |
Chris Lattner | fb3a0aed | 2002-10-28 18:50:08 +0000 | [diff] [blame] | 626 | this->addCDEdges(MBB.getBasicBlock()->getTerminator(), target); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 627 | |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 628 | // Then add memory dep edges: store->load, load->store, and store->store. |
| 629 | // Call instructions are treated as both load and store. |
Vikram S. Adve | e64574c | 2001-11-08 05:20:23 +0000 | [diff] [blame] | 630 | this->addMemEdges(memNodeVec, target); |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 631 | |
| 632 | // Then add edges between call instructions and CC set/use instructions |
Vikram S. Adve | 7952d60 | 2003-05-31 07:37:05 +0000 | [diff] [blame] | 633 | this->addCallDepEdges(callDepNodeVec, target); |
Vikram S. Adve | a93bbac | 2001-10-28 21:43:33 +0000 | [diff] [blame] | 634 | |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 635 | // Then add incoming def-use (SSA) edges for each machine instruction. |
Chris Lattner | fb3a0aed | 2002-10-28 18:50:08 +0000 | [diff] [blame] | 636 | for (unsigned i=0, N=MBB.size(); i < N; i++) |
| 637 | addEdgesForInstruction(*MBB[i], valueToDefVecMap, target); |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 638 | |
Vikram S. Adve | 200a435 | 2001-11-12 18:53:43 +0000 | [diff] [blame] | 639 | #ifdef NEED_SEPARATE_NONSSA_EDGES_CODE |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 640 | // Then add non-SSA edges for all VM instructions in the block. |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 641 | // We assume that all machine instructions that define a value are |
| 642 | // generated from the VM instruction corresponding to that value. |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 643 | // TODO: This could probably be done much more efficiently. |
Vikram S. Adve | 5316f8f | 2001-09-30 23:36:58 +0000 | [diff] [blame] | 644 | for (BasicBlock::const_iterator II = bb->begin(); II != bb->end(); ++II) |
Vikram S. Adve | c352d2c | 2001-11-05 04:04:23 +0000 | [diff] [blame] | 645 | this->addNonSSAEdgesForValue(*II, target); |
Chris Lattner | 4ed17ba | 2001-11-26 18:56:52 +0000 | [diff] [blame] | 646 | #endif //NEED_SEPARATE_NONSSA_EDGES_CODE |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 647 | |
| 648 | // Then add edges for dependences on machine registers |
| 649 | this->addMachineRegEdges(regToRefVecMap, target); |
| 650 | |
| 651 | // Finally, add edges from the dummy root and to dummy leaf |
| 652 | this->addDummyEdges(); |
| 653 | } |
| 654 | |
| 655 | |
| 656 | // |
| 657 | // class SchedGraphSet |
| 658 | // |
Chris Lattner | 2fbfdcf | 2002-04-07 20:49:59 +0000 | [diff] [blame] | 659 | SchedGraphSet::SchedGraphSet(const Function* _function, |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 660 | const TargetMachine& target) : |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 661 | function(_function) { |
Tanya Lattner | b6489f3 | 2003-08-25 22:42:20 +0000 | [diff] [blame] | 662 | buildGraphsForMethod(function, target); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 663 | } |
| 664 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 665 | SchedGraphSet::~SchedGraphSet() { |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 666 | // delete all the graphs |
Chris Lattner | f3dd05c | 2002-04-09 05:15:33 +0000 | [diff] [blame] | 667 | for(iterator I = begin(), E = end(); I != E; ++I) |
| 668 | delete *I; // destructor is a friend |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 669 | } |
| 670 | |
| 671 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 672 | void SchedGraphSet::dump() const { |
Tanya Lattner | b6489f3 | 2003-08-25 22:42:20 +0000 | [diff] [blame] | 673 | std::cerr << "======== Sched graphs for function `" << function->getName() |
Misha Brukman | c2312df | 2003-05-22 21:24:35 +0000 | [diff] [blame] | 674 | << "' ========\n\n"; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 675 | |
| 676 | for (const_iterator I=begin(); I != end(); ++I) |
Vikram S. Adve | cf8a98f | 2002-03-24 03:40:59 +0000 | [diff] [blame] | 677 | (*I)->dump(); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 678 | |
Tanya Lattner | b6489f3 | 2003-08-25 22:42:20 +0000 | [diff] [blame] | 679 | std::cerr << "\n====== End graphs for function `" << function->getName() |
Misha Brukman | c2312df | 2003-05-22 21:24:35 +0000 | [diff] [blame] | 680 | << "' ========\n\n"; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 681 | } |
| 682 | |
| 683 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 684 | void SchedGraphSet::buildGraphsForMethod(const Function *F, |
| 685 | const TargetMachine& target) { |
Chris Lattner | fb3a0aed | 2002-10-28 18:50:08 +0000 | [diff] [blame] | 686 | MachineFunction &MF = MachineFunction::get(F); |
| 687 | for (MachineFunction::iterator I = MF.begin(), E = MF.end(); I != E; ++I) |
| 688 | addGraph(new SchedGraph(*I, target)); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 689 | } |
| 690 | |
| 691 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 692 | void SchedGraphEdge::print(std::ostream &os) const { |
| 693 | os << "edge [" << src->getNodeId() << "] -> [" |
| 694 | << sink->getNodeId() << "] : "; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 695 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 696 | switch(depType) { |
| 697 | case SchedGraphEdge::CtrlDep: |
| 698 | os<< "Control Dep"; |
| 699 | break; |
| 700 | case SchedGraphEdge::ValueDep: |
| 701 | os<< "Reg Value " << val; |
| 702 | break; |
| 703 | case SchedGraphEdge::MemoryDep: |
| 704 | os<< "Memory Dep"; |
| 705 | break; |
| 706 | case SchedGraphEdge::MachineRegister: |
| 707 | os<< "Reg " << machineRegNum; |
| 708 | break; |
| 709 | case SchedGraphEdge::MachineResource: |
| 710 | os<<"Resource "<< resourceId; |
| 711 | break; |
| 712 | default: |
| 713 | assert(0); |
| 714 | break; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 715 | } |
| 716 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 717 | os << " : delay = " << minDelay << "\n"; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 718 | } |
| 719 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 720 | void SchedGraphNode::print(std::ostream &os) const { |
Chris Lattner | 697954c | 2002-01-20 22:54:45 +0000 | [diff] [blame] | 721 | os << std::string(8, ' ') |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 722 | << "Node " << ID << " : " |
| 723 | << "latency = " << latency << "\n" << std::string(12, ' '); |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 724 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 725 | if (getMachineInstr() == NULL) |
Chris Lattner | 697954c | 2002-01-20 22:54:45 +0000 | [diff] [blame] | 726 | os << "(Dummy node)\n"; |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 727 | else { |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 728 | os << *getMachineInstr() << "\n" << std::string(12, ' '); |
| 729 | os << inEdges.size() << " Incoming Edges:\n"; |
| 730 | for (unsigned i=0, N = inEdges.size(); i < N; i++) |
| 731 | os << std::string(16, ' ') << *inEdges[i]; |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 732 | |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 733 | os << std::string(12, ' ') << outEdges.size() |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 734 | << " Outgoing Edges:\n"; |
Tanya Lattner | c50ee55 | 2003-08-27 02:42:58 +0000 | [diff] [blame] | 735 | for (unsigned i=0, N= outEdges.size(); i < N; i++) |
| 736 | os << std::string(16, ' ') << *outEdges[i]; |
Misha Brukman | 6b77ec4 | 2003-05-22 21:49:18 +0000 | [diff] [blame] | 737 | } |
Vikram S. Adve | 78ef139 | 2001-08-28 23:06:02 +0000 | [diff] [blame] | 738 | } |
Brian Gaeke | d0fde30 | 2003-11-11 22:41:34 +0000 | [diff] [blame] | 739 | |
| 740 | } // End llvm namespace |