blob: 480893a4ea7fe62c475eefc79312a0ed611b83f3 [file] [log] [blame]
Jim Stichnothf7c9a142014-04-29 10:52:43 -07001//===- subzero/src/IceCfg.cpp - Control flow graph implementation ---------===//
2//
3// The Subzero Code Generator
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
Andrew Scull9612d322015-07-06 14:53:25 -07009///
10/// \file
Jim Stichnoth92a6e5b2015-12-02 16:52:44 -080011/// \brief Implements the Cfg class.
Andrew Scull9612d322015-07-06 14:53:25 -070012///
Jim Stichnothf7c9a142014-04-29 10:52:43 -070013//===----------------------------------------------------------------------===//
14
15#include "IceCfg.h"
John Porto67f8de92015-06-25 10:14:17 -070016
17#include "IceAssembler.h"
John Porto36d6aa62016-02-26 07:19:59 -080018#include "IceBitVector.h"
Jim Stichnothf7c9a142014-04-29 10:52:43 -070019#include "IceCfgNode.h"
Jim Stichnoth989a7032014-08-08 10:13:44 -070020#include "IceClFlags.h"
Jim Stichnothf7c9a142014-04-29 10:52:43 -070021#include "IceDefs.h"
Jan Voungec270732015-01-12 17:00:22 -080022#include "IceELFObjectWriter.h"
John Portof8b4cc82015-06-09 18:06:19 -070023#include "IceGlobalInits.h"
Jim Stichnothf7c9a142014-04-29 10:52:43 -070024#include "IceInst.h"
John Portoec3f5652015-08-31 15:07:09 -070025#include "IceInstVarIter.h"
Jim Stichnothd97c7df2014-06-04 11:57:08 -070026#include "IceLiveness.h"
Andrew Scullaa6c1092015-09-03 17:50:30 -070027#include "IceLoopAnalyzer.h"
Jim Stichnothf7c9a142014-04-29 10:52:43 -070028#include "IceOperand.h"
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -070029#include "IceTargetLowering.h"
Jim Stichnothf7c9a142014-04-29 10:52:43 -070030
31namespace Ice {
32
Jim Stichnothbbca7542015-02-11 16:08:31 -080033Cfg::Cfg(GlobalContext *Ctx, uint32_t SequenceNumber)
Jan Voung1f47ad02015-03-20 15:01:26 -070034 : Ctx(Ctx), SequenceNumber(SequenceNumber),
Jim Stichnotheafb56c2015-06-22 10:35:22 -070035 VMask(Ctx->getFlags().getVerbose()), NextInstNumber(Inst::NumberInitial),
John Porto44b3ce82016-02-26 13:10:55 -080036 Live(nullptr) {
37 Allocator.reset(new ArenaAllocator());
38 CfgLocalAllocatorScope _(this);
39 Target =
40 TargetLowering::createLowering(Ctx->getFlags().getTargetArch(), this);
41 VMetadata.reset(new VariablesMetadata(this));
42 TargetAssembler = Target->createAssembler();
43
Qining Luaee5fa82015-08-20 14:59:03 -070044 if (Ctx->getFlags().getRandomizeAndPoolImmediatesOption() == RPI_Randomize) {
Andrew Scull57e12682015-09-16 11:30:19 -070045 // If -randomize-pool-immediates=randomize, create a random number
46 // generator to generate a cookie for constant blinding.
Qining Luaee5fa82015-08-20 14:59:03 -070047 RandomNumberGenerator RNG(Ctx->getFlags().getRandomSeed(),
Qining Lu360e3192015-08-20 17:13:07 -070048 RPE_ConstantBlinding, this->SequenceNumber);
Qining Luaee5fa82015-08-20 14:59:03 -070049 ConstantBlindingCookie =
Qining Lu360e3192015-08-20 17:13:07 -070050 (uint32_t)RNG.next((uint64_t)std::numeric_limits<uint32_t>::max() + 1);
Qining Luaee5fa82015-08-20 14:59:03 -070051 }
Karl Schimpf6fcbddd2014-11-06 09:49:24 -080052}
Jim Stichnothf7c9a142014-04-29 10:52:43 -070053
John Portoe82b5602016-02-24 15:58:55 -080054Cfg::~Cfg() { assert(CfgAllocatorTraits::current() == nullptr); }
Jim Stichnothf7c9a142014-04-29 10:52:43 -070055
Jim Stichnothb40595a2016-01-29 06:14:31 -080056/// Create a string like "foo(i=123:b=9)" indicating the function name, number
57/// of high-level instructions, and number of basic blocks. This string is only
58/// used for dumping and other diagnostics, and the idea is that given a set of
59/// functions to debug a problem on, it's easy to find the smallest or simplest
60/// function to attack. Note that the counts may change somewhat depending on
61/// what point it is called during the translation passes.
62IceString Cfg::getFunctionNameAndSize() const {
63 if (!BuildDefs::dump())
64 return getFunctionName();
65 SizeT NodeCount = 0;
66 SizeT InstCount = 0;
67 for (CfgNode *Node : getNodes()) {
68 ++NodeCount;
69 // Note: deleted instructions are *not* ignored.
70 InstCount += Node->getPhis().size();
71 for (Inst &I : Node->getInsts()) {
72 if (!llvm::isa<InstTarget>(&I))
73 ++InstCount;
74 }
75 }
76 return getFunctionName() + "(i=" + std::to_string(InstCount) + ":b=" +
77 std::to_string(NodeCount) + ")";
78}
79
Jim Stichnothf7c9a142014-04-29 10:52:43 -070080void Cfg::setError(const IceString &Message) {
81 HasError = true;
82 ErrorMessage = Message;
Jim Stichnothf7c9a142014-04-29 10:52:43 -070083}
84
Jim Stichnoth668a7a32014-12-10 15:32:25 -080085CfgNode *Cfg::makeNode() {
Jim Stichnothf7c9a142014-04-29 10:52:43 -070086 SizeT LabelIndex = Nodes.size();
Jim Stichnoth54f3d512015-12-11 09:53:00 -080087 auto *Node = CfgNode::create(this, LabelIndex);
Jim Stichnothf7c9a142014-04-29 10:52:43 -070088 Nodes.push_back(Node);
89 return Node;
90}
91
Karl Schimpfac7d7342015-08-06 12:55:23 -070092void Cfg::swapNodes(NodeList &NewNodes) {
Jim Stichnothe7dbc0b2015-09-15 10:09:24 -070093 assert(Nodes.size() == NewNodes.size());
Karl Schimpfac7d7342015-08-06 12:55:23 -070094 Nodes.swap(NewNodes);
95 for (SizeT I = 0, NumNodes = getNumNodes(); I < NumNodes; ++I)
96 Nodes[I]->resetIndex(I);
97}
98
John Portoa83bfde2015-09-18 08:43:02 -070099template <> Variable *Cfg::makeVariable<Variable>(Type Ty) {
Andrew Scull6d47bcd2015-09-17 17:10:05 -0700100 SizeT Index = Variables.size();
101 Variable *Var = Target->shouldSplitToVariable64On32(Ty)
102 ? Variable64On32::create(this, Ty, Index)
103 : Variable::create(this, Ty, Index);
104 Variables.push_back(Var);
105 return Var;
106}
107
Jim Stichnothf7c9a142014-04-29 10:52:43 -0700108void Cfg::addArg(Variable *Arg) {
Jim Stichnoth144cdce2014-09-22 16:02:59 -0700109 Arg->setIsArg();
Jim Stichnothf7c9a142014-04-29 10:52:43 -0700110 Args.push_back(Arg);
111}
112
Jim Stichnoth144cdce2014-09-22 16:02:59 -0700113void Cfg::addImplicitArg(Variable *Arg) {
114 Arg->setIsImplicitArg();
115 ImplicitArgs.push_back(Arg);
116}
117
Andrew Scull57e12682015-09-16 11:30:19 -0700118// Returns whether the stack frame layout has been computed yet. This is used
119// for dumping the stack frame location of Variables.
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700120bool Cfg::hasComputedFrame() const { return getTarget()->hasComputedFrame(); }
121
John Portof8b4cc82015-06-09 18:06:19 -0700122namespace {
123constexpr char BlockNameGlobalPrefix[] = ".L$profiler$block_name$";
124constexpr char BlockStatsGlobalPrefix[] = ".L$profiler$block_info$";
125
John Porto1bec8bc2015-06-22 10:51:13 -0700126VariableDeclaration *nodeNameDeclaration(GlobalContext *Ctx,
127 const IceString &NodeAsmName) {
Jim Stichnoth54f3d512015-12-11 09:53:00 -0800128 auto *Var = VariableDeclaration::create(Ctx);
John Portof8b4cc82015-06-09 18:06:19 -0700129 Var->setName(BlockNameGlobalPrefix + NodeAsmName);
130 Var->setIsConstant(true);
John Porto1bec8bc2015-06-22 10:51:13 -0700131 Var->addInitializer(VariableDeclaration::DataInitializer::create(
John Portof8b4cc82015-06-09 18:06:19 -0700132 NodeAsmName.data(), NodeAsmName.size() + 1));
133 const SizeT Int64ByteSize = typeWidthInBytes(IceType_i64);
134 Var->setAlignment(Int64ByteSize); // Wasteful, 32-bit could use 4 bytes.
135 return Var;
136}
137
138VariableDeclaration *
John Porto1bec8bc2015-06-22 10:51:13 -0700139blockProfilingInfoDeclaration(GlobalContext *Ctx, const IceString &NodeAsmName,
John Portof8b4cc82015-06-09 18:06:19 -0700140 VariableDeclaration *NodeNameDeclaration) {
Jim Stichnoth54f3d512015-12-11 09:53:00 -0800141 auto *Var = VariableDeclaration::create(Ctx);
John Portof8b4cc82015-06-09 18:06:19 -0700142 Var->setName(BlockStatsGlobalPrefix + NodeAsmName);
143 const SizeT Int64ByteSize = typeWidthInBytes(IceType_i64);
John Porto1bec8bc2015-06-22 10:51:13 -0700144 Var->addInitializer(
145 VariableDeclaration::ZeroInitializer::create(Int64ByteSize));
John Portof8b4cc82015-06-09 18:06:19 -0700146
147 const RelocOffsetT NodeNameDeclarationOffset = 0;
John Porto1bec8bc2015-06-22 10:51:13 -0700148 Var->addInitializer(VariableDeclaration::RelocInitializer::create(
John Porto844211e2016-02-04 08:42:48 -0800149 NodeNameDeclaration,
150 {RelocOffset::create(Ctx, NodeNameDeclarationOffset)}));
John Portof8b4cc82015-06-09 18:06:19 -0700151 Var->setAlignment(Int64ByteSize);
152 return Var;
153}
John Portof8b4cc82015-06-09 18:06:19 -0700154} // end of anonymous namespace
155
156void Cfg::profileBlocks() {
157 if (GlobalInits == nullptr)
158 GlobalInits.reset(new VariableDeclarationList());
159
160 for (CfgNode *Node : Nodes) {
161 IceString NodeAsmName = Node->getAsmName();
John Porto1bec8bc2015-06-22 10:51:13 -0700162 GlobalInits->push_back(nodeNameDeclaration(Ctx, NodeAsmName));
John Portof8b4cc82015-06-09 18:06:19 -0700163 GlobalInits->push_back(
John Porto1bec8bc2015-06-22 10:51:13 -0700164 blockProfilingInfoDeclaration(Ctx, NodeAsmName, GlobalInits->back()));
John Portof8b4cc82015-06-09 18:06:19 -0700165 Node->profileExecutionCount(GlobalInits->back());
166 }
167}
168
169bool Cfg::isProfileGlobal(const VariableDeclaration &Var) {
170 return Var.getName().find(BlockStatsGlobalPrefix) == 0;
171}
172
173void Cfg::addCallToProfileSummary() {
174 // The call(s) to __Sz_profile_summary are added by the profiler in functions
175 // that cause the program to exit. This function is defined in
176 // runtime/szrt_profiler.c.
177 Constant *ProfileSummarySym =
178 Ctx->getConstantExternSym("__Sz_profile_summary");
179 constexpr SizeT NumArgs = 0;
180 constexpr Variable *Void = nullptr;
181 constexpr bool HasTailCall = false;
182 auto *Call =
183 InstCall::create(this, NumArgs, Void, ProfileSummarySym, HasTailCall);
184 getEntryNode()->getInsts().push_front(Call);
185}
186
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700187void Cfg::translate() {
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700188 if (hasError())
189 return;
Andrew Scull57e12682015-09-16 11:30:19 -0700190 // FunctionTimer conditionally pushes/pops a TimerMarker if TimeEachFunction
191 // is enabled.
Jim Stichnoth380d7b92015-01-30 13:10:39 -0800192 std::unique_ptr<TimerMarker> FunctionTimer;
Jim Stichnoth20b71f52015-06-24 15:52:24 -0700193 if (BuildDefs::dump()) {
Karl Schimpfdf80eb82015-02-09 14:20:22 -0800194 const IceString &TimingFocusOn =
195 getContext()->getFlags().getTimingFocusOn();
Jim Stichnoth380d7b92015-01-30 13:10:39 -0800196 const IceString &Name = getFunctionName();
197 if (TimingFocusOn == "*" || TimingFocusOn == Name) {
Jim Stichnoth1c44d812014-12-08 14:57:52 -0800198 setFocusedTiming();
199 getContext()->resetTimer(GlobalContext::TSK_Default);
Jim Stichnoth380d7b92015-01-30 13:10:39 -0800200 getContext()->setTimerName(GlobalContext::TSK_Default, Name);
Jim Stichnoth1c44d812014-12-08 14:57:52 -0800201 }
Karl Schimpfdf80eb82015-02-09 14:20:22 -0800202 if (getContext()->getFlags().getTimeEachFunction())
Jim Stichnoth380d7b92015-01-30 13:10:39 -0800203 FunctionTimer.reset(new TimerMarker(
204 getContext()->getTimerID(GlobalContext::TSK_Funcs, Name),
205 getContext(), GlobalContext::TSK_Funcs));
Jim Stichnotha5229652015-11-12 10:25:21 -0800206 if (isVerbose(IceV_Status))
207 getContext()->getStrDump() << ">>>Translating " << Name << "\n";
Jim Stichnothd14b1a02014-10-08 08:28:36 -0700208 }
Jim Stichnoth8363a062014-10-07 10:02:38 -0700209 TimerMarker T(TimerStack::TT_translate, this);
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700210
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700211 dump("Initial CFG");
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700212
John Portof8b4cc82015-06-09 18:06:19 -0700213 if (getContext()->getFlags().getEnableBlockProfile()) {
214 profileBlocks();
Andrew Scull57e12682015-09-16 11:30:19 -0700215 // TODO(jpp): this is fragile, at best. Figure out a better way of
216 // detecting exit functions.
John Portof8b4cc82015-06-09 18:06:19 -0700217 if (GlobalContext::matchSymbolName(getFunctionName(), "exit")) {
218 addCallToProfileSummary();
219 }
220 dump("Profiled CFG");
221 }
222
Jim Stichnothbe87b2e2015-09-18 15:43:59 -0700223 // Create the Hi and Lo variables where a split was needed
224 for (Variable *Var : Variables)
Jim Stichnoth5bff61c2015-10-28 09:26:00 -0700225 if (auto *Var64On32 = llvm::dyn_cast<Variable64On32>(Var))
Jim Stichnothbe87b2e2015-09-18 15:43:59 -0700226 Var64On32->initHiLo(this);
227
Andrew Scull57e12682015-09-16 11:30:19 -0700228 // The set of translation passes and their order are determined by the
229 // target.
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700230 getTarget()->translate();
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700231
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700232 dump("Final output");
Jim Stichnoth8363a062014-10-07 10:02:38 -0700233 if (getFocusedTiming())
234 getContext()->dumpTimers();
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700235}
236
Jim Stichnoth69d3f9c2015-03-23 10:33:38 -0700237void Cfg::computeInOutEdges() {
238 // Compute the out-edges.
Jim Stichnothf44f3712014-10-01 14:05:51 -0700239 for (CfgNode *Node : Nodes)
Jim Stichnoth69d3f9c2015-03-23 10:33:38 -0700240 Node->computeSuccessors();
241
242 // Prune any unreachable nodes before computing in-edges.
243 SizeT NumNodes = getNumNodes();
John Porto36d6aa62016-02-26 07:19:59 -0800244 BitVector Reachable(NumNodes);
245 BitVector Pending(NumNodes);
Jim Stichnoth69d3f9c2015-03-23 10:33:38 -0700246 Pending.set(getEntryNode()->getIndex());
247 while (true) {
248 int Index = Pending.find_first();
249 if (Index == -1)
250 break;
251 Pending.reset(Index);
252 Reachable.set(Index);
253 CfgNode *Node = Nodes[Index];
254 assert(Node->getIndex() == (SizeT)Index);
255 for (CfgNode *Succ : Node->getOutEdges()) {
256 SizeT SuccIndex = Succ->getIndex();
257 if (!Reachable.test(SuccIndex))
258 Pending.set(SuccIndex);
259 }
260 }
261 SizeT Dest = 0;
262 for (SizeT Source = 0; Source < NumNodes; ++Source) {
263 if (Reachable.test(Source)) {
264 Nodes[Dest] = Nodes[Source];
265 Nodes[Dest]->resetIndex(Dest);
266 // Compute the in-edges.
267 Nodes[Dest]->computePredecessors();
268 ++Dest;
269 }
270 }
271 Nodes.resize(Dest);
Jim Stichnoth1aca2302015-09-16 11:25:22 -0700272
273 TimerMarker T(TimerStack::TT_phiValidation, this);
274 for (CfgNode *Node : Nodes)
275 Node->validatePhis();
Jim Stichnothf7c9a142014-04-29 10:52:43 -0700276}
277
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700278void Cfg::renumberInstructions() {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700279 TimerMarker T(TimerStack::TT_renumberInstructions, this);
Jim Stichnothe5b73e62014-12-15 09:58:51 -0800280 NextInstNumber = Inst::NumberInitial;
Jim Stichnothf44f3712014-10-01 14:05:51 -0700281 for (CfgNode *Node : Nodes)
282 Node->renumberInstructions();
Jim Stichnoth6f7ad6c2016-01-15 09:52:54 -0800283 // Make sure the entry node is the first node and therefore got the lowest
284 // instruction numbers, to facilitate live range computation of function
285 // arguments. We want to model function arguments as being live on entry to
286 // the function, otherwise an argument whose only use is in the first
287 // instruction will be assigned a trivial live range and the register
288 // allocator will not recognize its live range as overlapping another
289 // variable's live range.
290 assert(Nodes.empty() || (*Nodes.begin() == getEntryNode()));
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700291}
292
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700293// placePhiLoads() must be called before placePhiStores().
294void Cfg::placePhiLoads() {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700295 TimerMarker T(TimerStack::TT_placePhiLoads, this);
Jim Stichnothf44f3712014-10-01 14:05:51 -0700296 for (CfgNode *Node : Nodes)
297 Node->placePhiLoads();
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700298}
299
300// placePhiStores() must be called after placePhiLoads().
301void Cfg::placePhiStores() {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700302 TimerMarker T(TimerStack::TT_placePhiStores, this);
Jim Stichnothf44f3712014-10-01 14:05:51 -0700303 for (CfgNode *Node : Nodes)
304 Node->placePhiStores();
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700305}
306
307void Cfg::deletePhis() {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700308 TimerMarker T(TimerStack::TT_deletePhis, this);
Jim Stichnothf44f3712014-10-01 14:05:51 -0700309 for (CfgNode *Node : Nodes)
310 Node->deletePhis();
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700311}
312
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700313void Cfg::advancedPhiLowering() {
314 TimerMarker T(TimerStack::TT_advancedPhiLowering, this);
Jim Stichnotha3f57b92015-07-30 12:46:04 -0700315 // Clear all previously computed live ranges (but not live-in/live-out bit
316 // vectors or last-use markers), because the follow-on register allocation is
317 // only concerned with live ranges across the newly created blocks.
318 for (Variable *Var : Variables) {
319 Var->getLiveRange().reset();
320 }
Andrew Scull57e12682015-09-16 11:30:19 -0700321 // This splits edges and appends new nodes to the end of the node list. This
322 // can invalidate iterators, so don't use an iterator.
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700323 SizeT NumNodes = getNumNodes();
Jim Stichnotha3f57b92015-07-30 12:46:04 -0700324 SizeT NumVars = getNumVariables();
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700325 for (SizeT I = 0; I < NumNodes; ++I)
326 Nodes[I]->advancedPhiLowering();
Jim Stichnotha3f57b92015-07-30 12:46:04 -0700327
328 TimerMarker TT(TimerStack::TT_lowerPhiAssignments, this);
329 if (true) {
Andrew Scull57e12682015-09-16 11:30:19 -0700330 // The following code does an in-place update of liveness and live ranges
331 // as a result of adding the new phi edge split nodes.
Jim Stichnotha3f57b92015-07-30 12:46:04 -0700332 getLiveness()->initPhiEdgeSplits(Nodes.begin() + NumNodes,
333 Variables.begin() + NumVars);
334 TimerMarker TTT(TimerStack::TT_liveness, this);
335 // Iterate over the newly added nodes to add their liveness info.
336 for (auto I = Nodes.begin() + NumNodes, E = Nodes.end(); I != E; ++I) {
337 InstNumberT FirstInstNum = getNextInstNumber();
338 (*I)->renumberInstructions();
339 InstNumberT LastInstNum = getNextInstNumber() - 1;
Jim Stichnotha3f57b92015-07-30 12:46:04 -0700340 (*I)->liveness(getLiveness());
341 (*I)->livenessAddIntervals(getLiveness(), FirstInstNum, LastInstNum);
342 }
343 } else {
344 // The following code does a brute-force recalculation of live ranges as a
Andrew Scull57e12682015-09-16 11:30:19 -0700345 // result of adding the new phi edge split nodes. The liveness calculation
Jim Stichnotha3f57b92015-07-30 12:46:04 -0700346 // is particularly expensive because the new nodes are not yet in a proper
347 // topological order and so convergence is slower.
348 //
349 // This code is kept here for reference and can be temporarily enabled in
350 // case the incremental code is under suspicion.
351 renumberInstructions();
352 liveness(Liveness_Intervals);
353 getVMetadata()->init(VMK_All);
354 }
355 Target->regAlloc(RAK_Phi);
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700356}
357
Andrew Scull57e12682015-09-16 11:30:19 -0700358// Find a reasonable placement for nodes that have not yet been placed, while
359// maintaining the same relative ordering among already placed nodes.
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700360void Cfg::reorderNodes() {
Andrew Scull57e12682015-09-16 11:30:19 -0700361 // TODO(ascull): it would be nice if the switch tests were always followed by
362 // the default case to allow for fall through.
Andrew Scull00741a02015-09-16 19:04:09 -0700363 using PlacedList = CfgList<CfgNode *>;
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700364 PlacedList Placed; // Nodes with relative placement locked down
365 PlacedList Unreachable; // Unreachable nodes
366 PlacedList::iterator NoPlace = Placed.end();
Andrew Scull57e12682015-09-16 11:30:19 -0700367 // Keep track of where each node has been tentatively placed so that we can
368 // manage insertions into the middle.
Andrew Scull00741a02015-09-16 19:04:09 -0700369 CfgVector<PlacedList::iterator> PlaceIndex(Nodes.size(), NoPlace);
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700370 for (CfgNode *Node : Nodes) {
Andrew Scull57e12682015-09-16 11:30:19 -0700371 // The "do ... while(0);" construct is to factor out the --PlaceIndex and
372 // assert() statements before moving to the next node.
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700373 do {
Andrew Scull713278a2015-07-22 17:17:02 -0700374 if (Node != getEntryNode() && Node->getInEdges().empty()) {
Andrew Scull57e12682015-09-16 11:30:19 -0700375 // The node has essentially been deleted since it is not a successor of
376 // any other node.
Andrew Scull713278a2015-07-22 17:17:02 -0700377 Unreachable.push_back(Node);
378 PlaceIndex[Node->getIndex()] = Unreachable.end();
379 Node->setNeedsPlacement(false);
380 continue;
381 }
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700382 if (!Node->needsPlacement()) {
383 // Add to the end of the Placed list.
384 Placed.push_back(Node);
385 PlaceIndex[Node->getIndex()] = Placed.end();
386 continue;
387 }
388 Node->setNeedsPlacement(false);
Andrew Scull57e12682015-09-16 11:30:19 -0700389 // Assume for now that the unplaced node is from edge-splitting and
390 // therefore has 1 in-edge and 1 out-edge (actually, possibly more than 1
391 // in-edge if the predecessor node was contracted). If this changes in
392 // the future, rethink the strategy.
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700393 assert(Node->getInEdges().size() >= 1);
394 assert(Node->getOutEdges().size() == 1);
395
396 // If it's a (non-critical) edge where the successor has a single
397 // in-edge, then place it before the successor.
Jim Stichnothbfb410d2014-11-05 16:04:05 -0800398 CfgNode *Succ = Node->getOutEdges().front();
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700399 if (Succ->getInEdges().size() == 1 &&
400 PlaceIndex[Succ->getIndex()] != NoPlace) {
401 Placed.insert(PlaceIndex[Succ->getIndex()], Node);
402 PlaceIndex[Node->getIndex()] = PlaceIndex[Succ->getIndex()];
403 continue;
404 }
405
406 // Otherwise, place it after the (first) predecessor.
Jim Stichnothbfb410d2014-11-05 16:04:05 -0800407 CfgNode *Pred = Node->getInEdges().front();
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700408 auto PredPosition = PlaceIndex[Pred->getIndex()];
Andrew Scull57e12682015-09-16 11:30:19 -0700409 // It shouldn't be the case that PredPosition==NoPlace, but if that
410 // somehow turns out to be true, we just insert Node before
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700411 // PredPosition=NoPlace=Placed.end() .
412 if (PredPosition != NoPlace)
413 ++PredPosition;
414 Placed.insert(PredPosition, Node);
415 PlaceIndex[Node->getIndex()] = PredPosition;
416 } while (0);
417
418 --PlaceIndex[Node->getIndex()];
419 assert(*PlaceIndex[Node->getIndex()] == Node);
420 }
421
422 // Reorder Nodes according to the built-up lists.
Karl Schimpfac7d7342015-08-06 12:55:23 -0700423 NodeList Reordered;
424 Reordered.reserve(Placed.size() + Unreachable.size());
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700425 for (CfgNode *Node : Placed)
Karl Schimpfac7d7342015-08-06 12:55:23 -0700426 Reordered.push_back(Node);
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700427 for (CfgNode *Node : Unreachable)
Karl Schimpfac7d7342015-08-06 12:55:23 -0700428 Reordered.push_back(Node);
429 assert(getNumNodes() == Reordered.size());
430 swapNodes(Reordered);
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700431}
432
Qining Lu969f6a32015-07-31 09:58:34 -0700433namespace {
John Porto36d6aa62016-02-26 07:19:59 -0800434void getRandomPostOrder(CfgNode *Node, BitVector &ToVisit,
Qining Lu969f6a32015-07-31 09:58:34 -0700435 Ice::NodeList &PostOrder,
436 Ice::RandomNumberGenerator *RNG) {
437 assert(ToVisit[Node->getIndex()]);
438 ToVisit[Node->getIndex()] = false;
439 NodeList Outs = Node->getOutEdges();
440 Ice::RandomShuffle(Outs.begin(), Outs.end(),
441 [RNG](int N) { return RNG->next(N); });
442 for (CfgNode *Next : Outs) {
443 if (ToVisit[Next->getIndex()])
444 getRandomPostOrder(Next, ToVisit, PostOrder, RNG);
445 }
446 PostOrder.push_back(Node);
447}
448} // end of anonymous namespace
449
450void Cfg::shuffleNodes() {
451 if (!Ctx->getFlags().shouldReorderBasicBlocks())
452 return;
453
454 NodeList ReversedReachable;
455 NodeList Unreachable;
John Porto36d6aa62016-02-26 07:19:59 -0800456 BitVector ToVisit(Nodes.size(), true);
Qining Luaee5fa82015-08-20 14:59:03 -0700457 // Create Random number generator for function reordering
458 RandomNumberGenerator RNG(Ctx->getFlags().getRandomSeed(),
459 RPE_BasicBlockReordering, SequenceNumber);
Qining Lu969f6a32015-07-31 09:58:34 -0700460 // Traverse from entry node.
Qining Luaee5fa82015-08-20 14:59:03 -0700461 getRandomPostOrder(getEntryNode(), ToVisit, ReversedReachable, &RNG);
Qining Lu969f6a32015-07-31 09:58:34 -0700462 // Collect the unreachable nodes.
463 for (CfgNode *Node : Nodes)
464 if (ToVisit[Node->getIndex()])
465 Unreachable.push_back(Node);
466 // Copy the layout list to the Nodes.
Karl Schimpfac7d7342015-08-06 12:55:23 -0700467 NodeList Shuffled;
468 Shuffled.reserve(ReversedReachable.size() + Unreachable.size());
Qining Lu969f6a32015-07-31 09:58:34 -0700469 for (CfgNode *Node : reverse_range(ReversedReachable))
Karl Schimpfac7d7342015-08-06 12:55:23 -0700470 Shuffled.push_back(Node);
471 for (CfgNode *Node : Unreachable)
472 Shuffled.push_back(Node);
473 assert(Nodes.size() == Shuffled.size());
474 swapNodes(Shuffled);
Qining Lu969f6a32015-07-31 09:58:34 -0700475
476 dump("After basic block shuffling");
477}
478
Matt Wala45a06232014-07-09 16:33:22 -0700479void Cfg::doArgLowering() {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700480 TimerMarker T(TimerStack::TT_doArgLowering, this);
Matt Wala45a06232014-07-09 16:33:22 -0700481 getTarget()->lowerArguments();
482}
483
David Sehr4318a412015-11-11 15:01:55 -0800484void Cfg::sortAndCombineAllocas(CfgVector<Inst *> &Allocas,
485 uint32_t CombinedAlignment, InstList &Insts,
486 AllocaBaseVariableType BaseVariableType) {
David Sehre39d0ca2015-11-06 11:25:41 -0800487 if (Allocas.empty())
488 return;
David Sehr4318a412015-11-11 15:01:55 -0800489 // Sort by decreasing alignment.
Jim Stichnothc59288b2015-11-09 11:38:40 -0800490 std::sort(Allocas.begin(), Allocas.end(), [](Inst *I1, Inst *I2) {
491 auto *A1 = llvm::dyn_cast<InstAlloca>(I1);
492 auto *A2 = llvm::dyn_cast<InstAlloca>(I2);
493 return A1->getAlignInBytes() > A2->getAlignInBytes();
494 });
David Sehr4318a412015-11-11 15:01:55 -0800495 // Process the allocas in order of decreasing stack alignment. This allows
496 // us to pack less-aligned pieces after more-aligned ones, resulting in less
497 // stack growth. It also allows there to be at most one stack alignment "and"
498 // instruction for a whole list of allocas.
499 uint32_t CurrentOffset = 0;
500 CfgVector<int32_t> Offsets;
Jim Stichnothc59288b2015-11-09 11:38:40 -0800501 for (Inst *Instr : Allocas) {
David Sehre39d0ca2015-11-06 11:25:41 -0800502 auto *Alloca = llvm::cast<InstAlloca>(Instr);
David Sehr4318a412015-11-11 15:01:55 -0800503 // Adjust the size of the allocation up to the next multiple of the
504 // object's alignment.
505 uint32_t Alignment = std::max(Alloca->getAlignInBytes(), 1u);
506 auto *ConstSize =
507 llvm::dyn_cast<ConstantInteger32>(Alloca->getSizeInBytes());
508 uint32_t Size = Utils::applyAlignment(ConstSize->getValue(), Alignment);
509 if (BaseVariableType == BVT_FramePointer) {
510 // Addressing is relative to the frame pointer. Subtract the offset after
511 // adding the size of the alloca, because it grows downwards from the
512 // frame pointer.
513 Offsets.push_back(-(CurrentOffset + Size));
514 } else {
515 // Addressing is relative to the stack pointer or to a user pointer. Add
516 // the offset before adding the size of the object, because it grows
John Porto614140e2015-11-23 11:43:13 -0800517 // upwards from the stack pointer. In addition, if the addressing is
518 // relative to the stack pointer, we need to add the pre-computed max out
519 // args size bytes.
520 const uint32_t OutArgsOffsetOrZero =
521 (BaseVariableType == BVT_StackPointer)
522 ? getTarget()->maxOutArgsSizeBytes()
523 : 0;
524 Offsets.push_back(CurrentOffset + OutArgsOffsetOrZero);
David Sehr4318a412015-11-11 15:01:55 -0800525 }
526 // Update the running offset of the fused alloca region.
527 CurrentOffset += Size;
528 }
529 // Round the offset up to the alignment granularity to use as the size.
530 uint32_t TotalSize = Utils::applyAlignment(CurrentOffset, CombinedAlignment);
531 // Ensure every alloca was assigned an offset.
532 assert(Allocas.size() == Offsets.size());
David Sehr2f3b8ec2015-11-16 16:51:39 -0800533
534 switch (BaseVariableType) {
535 case BVT_UserPointer: {
536 Variable *BaseVariable = makeVariable(IceType_i32);
537 for (SizeT i = 0; i < Allocas.size(); ++i) {
538 auto *Alloca = llvm::cast<InstAlloca>(Allocas[i]);
David Sehr4318a412015-11-11 15:01:55 -0800539 // Emit a new addition operation to replace the alloca.
540 Operand *AllocaOffset = Ctx->getConstantInt32(Offsets[i]);
541 InstArithmetic *Add =
542 InstArithmetic::create(this, InstArithmetic::Add, Alloca->getDest(),
543 BaseVariable, AllocaOffset);
544 Insts.push_front(Add);
David Sehr2f3b8ec2015-11-16 16:51:39 -0800545 Alloca->setDeleted();
546 }
547 Operand *AllocaSize = Ctx->getConstantInt32(TotalSize);
548 InstAlloca *CombinedAlloca =
549 InstAlloca::create(this, BaseVariable, AllocaSize, CombinedAlignment);
550 CombinedAlloca->setKnownFrameOffset();
551 Insts.push_front(CombinedAlloca);
552 } break;
553 case BVT_StackPointer:
554 case BVT_FramePointer: {
555 for (SizeT i = 0; i < Allocas.size(); ++i) {
556 auto *Alloca = llvm::cast<InstAlloca>(Allocas[i]);
David Sehr4318a412015-11-11 15:01:55 -0800557 // Emit a fake definition of the rematerializable variable.
558 Variable *Dest = Alloca->getDest();
Jim Stichnoth54f3d512015-12-11 09:53:00 -0800559 auto *Def = InstFakeDef::create(this, Dest);
David Sehr2f3b8ec2015-11-16 16:51:39 -0800560 if (BaseVariableType == BVT_StackPointer)
561 Dest->setRematerializable(getTarget()->getStackReg(), Offsets[i]);
562 else
563 Dest->setRematerializable(getTarget()->getFrameReg(), Offsets[i]);
David Sehr4318a412015-11-11 15:01:55 -0800564 Insts.push_front(Def);
David Sehr2f3b8ec2015-11-16 16:51:39 -0800565 Alloca->setDeleted();
David Sehr4318a412015-11-11 15:01:55 -0800566 }
David Sehr2f3b8ec2015-11-16 16:51:39 -0800567 // Allocate the fixed area in the function prolog.
568 getTarget()->reserveFixedAllocaArea(TotalSize, CombinedAlignment);
David Sehr4318a412015-11-11 15:01:55 -0800569 } break;
David Sehr4318a412015-11-11 15:01:55 -0800570 }
David Sehre39d0ca2015-11-06 11:25:41 -0800571}
572
David Sehr4318a412015-11-11 15:01:55 -0800573void Cfg::processAllocas(bool SortAndCombine) {
David Sehre39d0ca2015-11-06 11:25:41 -0800574 const uint32_t StackAlignment = getTarget()->getStackAlignment();
575 CfgNode *EntryNode = getEntryNode();
David Sehre39d0ca2015-11-06 11:25:41 -0800576 // LLVM enforces power of 2 alignment.
577 assert(llvm::isPowerOf2_32(StackAlignment));
David Sehr4318a412015-11-11 15:01:55 -0800578 // Determine if there are large alignment allocations in the entry block or
579 // dynamic allocations (variable size in the entry block).
580 bool HasLargeAlignment = false;
581 bool HasDynamicAllocation = false;
David Sehre39d0ca2015-11-06 11:25:41 -0800582 for (Inst &Instr : EntryNode->getInsts()) {
583 if (auto *Alloca = llvm::dyn_cast<InstAlloca>(&Instr)) {
David Sehre39d0ca2015-11-06 11:25:41 -0800584 uint32_t AlignmentParam = Alloca->getAlignInBytes();
David Sehr4318a412015-11-11 15:01:55 -0800585 if (AlignmentParam > StackAlignment)
586 HasLargeAlignment = true;
587 if (llvm::isa<Constant>(Alloca->getSizeInBytes()))
588 Alloca->setKnownFrameOffset();
589 else {
590 HasDynamicAllocation = true;
591 // If Allocas are not sorted, the first dynamic allocation causes
592 // later Allocas to be at unknown offsets relative to the stack/frame.
593 if (!SortAndCombine)
594 break;
595 }
David Sehre39d0ca2015-11-06 11:25:41 -0800596 }
597 }
David Sehr4318a412015-11-11 15:01:55 -0800598 // Don't do the heavyweight sorting and layout for low optimization levels.
599 if (!SortAndCombine)
600 return;
601 // Any alloca outside the entry block is a dynamic allocation.
David Sehre39d0ca2015-11-06 11:25:41 -0800602 for (CfgNode *Node : Nodes) {
603 if (Node == EntryNode)
604 continue;
605 for (Inst &Instr : Node->getInsts()) {
606 if (llvm::isa<InstAlloca>(&Instr)) {
607 // Allocations outside the entry block require a frame pointer.
David Sehr4318a412015-11-11 15:01:55 -0800608 HasDynamicAllocation = true;
David Sehre39d0ca2015-11-06 11:25:41 -0800609 break;
610 }
611 }
David Sehr4318a412015-11-11 15:01:55 -0800612 if (HasDynamicAllocation)
David Sehre39d0ca2015-11-06 11:25:41 -0800613 break;
614 }
615 // Mark the target as requiring a frame pointer.
David Sehr4318a412015-11-11 15:01:55 -0800616 if (HasLargeAlignment || HasDynamicAllocation)
David Sehre39d0ca2015-11-06 11:25:41 -0800617 getTarget()->setHasFramePointer();
David Sehr4318a412015-11-11 15:01:55 -0800618 // Collect the Allocas into the two vectors.
619 // Allocas in the entry block that have constant size and alignment less
620 // than or equal to the function's stack alignment.
621 CfgVector<Inst *> FixedAllocas;
622 // Allocas in the entry block that have constant size and alignment greater
623 // than the function's stack alignment.
624 CfgVector<Inst *> AlignedAllocas;
David Sehr2f3b8ec2015-11-16 16:51:39 -0800625 // Maximum alignment used by any alloca.
David Sehr4318a412015-11-11 15:01:55 -0800626 uint32_t MaxAlignment = StackAlignment;
627 for (Inst &Instr : EntryNode->getInsts()) {
628 if (auto *Alloca = llvm::dyn_cast<InstAlloca>(&Instr)) {
629 if (!llvm::isa<Constant>(Alloca->getSizeInBytes()))
630 continue;
631 uint32_t AlignmentParam = Alloca->getAlignInBytes();
632 // For default align=0, set it to the real value 1, to avoid any
633 // bit-manipulation problems below.
634 AlignmentParam = std::max(AlignmentParam, 1u);
635 assert(llvm::isPowerOf2_32(AlignmentParam));
636 if (HasDynamicAllocation && AlignmentParam > StackAlignment) {
637 // If we have both dynamic allocations and large stack alignments,
638 // high-alignment allocations are pulled out with their own base.
639 AlignedAllocas.push_back(Alloca);
640 } else {
641 FixedAllocas.push_back(Alloca);
642 }
643 MaxAlignment = std::max(AlignmentParam, MaxAlignment);
644 }
645 }
David Sehre39d0ca2015-11-06 11:25:41 -0800646 // Add instructions to the head of the entry block in reverse order.
647 InstList &Insts = getEntryNode()->getInsts();
David Sehr4318a412015-11-11 15:01:55 -0800648 if (HasDynamicAllocation && HasLargeAlignment) {
649 // We are using a frame pointer, but fixed large-alignment alloca addresses,
650 // do not have a known offset from either the stack or frame pointer.
651 // They grow up from a user pointer from an alloca.
652 sortAndCombineAllocas(AlignedAllocas, MaxAlignment, Insts, BVT_UserPointer);
David Sehr2f3b8ec2015-11-16 16:51:39 -0800653 // Fixed size allocas are addressed relative to the frame pointer.
654 sortAndCombineAllocas(FixedAllocas, StackAlignment, Insts,
655 BVT_FramePointer);
656 } else {
657 // Otherwise, fixed size allocas are addressed relative to the stack unless
658 // there are dynamic allocas.
659 const AllocaBaseVariableType BasePointerType =
660 (HasDynamicAllocation ? BVT_FramePointer : BVT_StackPointer);
661 sortAndCombineAllocas(FixedAllocas, MaxAlignment, Insts, BasePointerType);
David Sehr4318a412015-11-11 15:01:55 -0800662 }
Jim Stichnoth3607b6c2015-11-13 14:28:23 -0800663 if (!FixedAllocas.empty() || !AlignedAllocas.empty())
664 // No use calling findRematerializable() unless there is some
665 // rematerializable alloca instruction to seed it.
666 findRematerializable();
667}
668
669namespace {
670
671// Helpers for findRematerializable(). For each of them, if a suitable
672// rematerialization is found, the instruction's Dest variable is set to be
673// rematerializable and it returns true, otherwise it returns false.
674
675bool rematerializeArithmetic(const Inst *Instr) {
676 // Check that it's an Arithmetic instruction with an Add operation.
677 auto *Arith = llvm::dyn_cast<InstArithmetic>(Instr);
678 if (Arith == nullptr || Arith->getOp() != InstArithmetic::Add)
679 return false;
680 // Check that Src(0) is rematerializable.
681 auto *Src0Var = llvm::dyn_cast<Variable>(Arith->getSrc(0));
682 if (Src0Var == nullptr || !Src0Var->isRematerializable())
683 return false;
684 // Check that Src(1) is an immediate.
685 auto *Src1Imm = llvm::dyn_cast<ConstantInteger32>(Arith->getSrc(1));
686 if (Src1Imm == nullptr)
687 return false;
688 Arith->getDest()->setRematerializable(
689 Src0Var->getRegNum(), Src0Var->getStackOffset() + Src1Imm->getValue());
690 return true;
691}
692
693bool rematerializeAssign(const Inst *Instr) {
694 // An InstAssign only originates from an inttoptr or ptrtoint instruction,
695 // which never occurs in a MINIMAL build.
696 if (BuildDefs::minimal())
697 return false;
698 // Check that it's an Assign instruction.
699 if (!llvm::isa<InstAssign>(Instr))
700 return false;
701 // Check that Src(0) is rematerializable.
702 auto *Src0Var = llvm::dyn_cast<Variable>(Instr->getSrc(0));
703 if (Src0Var == nullptr || !Src0Var->isRematerializable())
704 return false;
705 Instr->getDest()->setRematerializable(Src0Var->getRegNum(),
706 Src0Var->getStackOffset());
707 return true;
708}
709
710bool rematerializeCast(const Inst *Instr) {
711 // An pointer-type bitcast never occurs in a MINIMAL build.
712 if (BuildDefs::minimal())
713 return false;
714 // Check that it's a Cast instruction with a Bitcast operation.
715 auto *Cast = llvm::dyn_cast<InstCast>(Instr);
716 if (Cast == nullptr || Cast->getCastKind() != InstCast::Bitcast)
717 return false;
718 // Check that Src(0) is rematerializable.
719 auto *Src0Var = llvm::dyn_cast<Variable>(Cast->getSrc(0));
720 if (Src0Var == nullptr || !Src0Var->isRematerializable())
721 return false;
722 // Check that Dest and Src(0) have the same type.
723 Variable *Dest = Cast->getDest();
724 if (Dest->getType() != Src0Var->getType())
725 return false;
726 Dest->setRematerializable(Src0Var->getRegNum(), Src0Var->getStackOffset());
727 return true;
728}
729
730} // end of anonymous namespace
731
732/// Scan the function to find additional rematerializable variables. This is
733/// possible when the source operand of an InstAssignment is a rematerializable
734/// variable, or the same for a pointer-type InstCast::Bitcast, or when an
735/// InstArithmetic is an add of a rematerializable variable and an immediate.
736/// Note that InstAssignment instructions and pointer-type InstCast::Bitcast
737/// instructions generally only come about from the IceConverter's treatment of
738/// inttoptr, ptrtoint, and bitcast instructions. TODO(stichnot): Consider
739/// other possibilities, however unlikely, such as InstArithmetic::Sub, or
740/// commutativity.
741void Cfg::findRematerializable() {
742 // Scan the instructions in order, and repeat until no new opportunities are
743 // found. It may take more than one iteration because a variable's defining
744 // block may happen to come after a block where it is used, depending on the
745 // CfgNode linearization order.
746 bool FoundNewAssignment;
747 do {
748 FoundNewAssignment = false;
749 for (CfgNode *Node : getNodes()) {
750 // No need to process Phi instructions.
751 for (Inst &Instr : Node->getInsts()) {
752 if (Instr.isDeleted())
753 continue;
754 Variable *Dest = Instr.getDest();
755 if (Dest == nullptr || Dest->isRematerializable())
756 continue;
757 if (rematerializeArithmetic(&Instr) || rematerializeAssign(&Instr) ||
758 rematerializeCast(&Instr)) {
759 FoundNewAssignment = true;
760 }
761 }
762 }
763 } while (FoundNewAssignment);
David Sehre39d0ca2015-11-06 11:25:41 -0800764}
765
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700766void Cfg::doAddressOpt() {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700767 TimerMarker T(TimerStack::TT_doAddressOpt, this);
Jim Stichnothf44f3712014-10-01 14:05:51 -0700768 for (CfgNode *Node : Nodes)
769 Node->doAddressOpt();
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700770}
771
Matt Walac3302742014-08-15 16:21:56 -0700772void Cfg::doNopInsertion() {
Qining Lu969f6a32015-07-31 09:58:34 -0700773 if (!Ctx->getFlags().shouldDoNopInsertion())
774 return;
Jim Stichnoth8363a062014-10-07 10:02:38 -0700775 TimerMarker T(TimerStack::TT_doNopInsertion, this);
Qining Luaee5fa82015-08-20 14:59:03 -0700776 RandomNumberGenerator RNG(Ctx->getFlags().getRandomSeed(), RPE_NopInsertion,
777 SequenceNumber);
Jim Stichnothf44f3712014-10-01 14:05:51 -0700778 for (CfgNode *Node : Nodes)
Qining Luaee5fa82015-08-20 14:59:03 -0700779 Node->doNopInsertion(RNG);
Matt Walac3302742014-08-15 16:21:56 -0700780}
781
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700782void Cfg::genCode() {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700783 TimerMarker T(TimerStack::TT_genCode, this);
Jim Stichnothf44f3712014-10-01 14:05:51 -0700784 for (CfgNode *Node : Nodes)
785 Node->genCode();
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700786}
787
788// Compute the stack frame layout.
789void Cfg::genFrame() {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700790 TimerMarker T(TimerStack::TT_genFrame, this);
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700791 getTarget()->addProlog(Entry);
Jim Stichnothf44f3712014-10-01 14:05:51 -0700792 for (CfgNode *Node : Nodes)
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700793 if (Node->getHasReturn())
794 getTarget()->addEpilog(Node);
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -0700795}
796
Andrew Scullaa6c1092015-09-03 17:50:30 -0700797void Cfg::computeLoopNestDepth() {
798 TimerMarker T(TimerStack::TT_computeLoopNestDepth, this);
799 LoopAnalyzer LA(this);
800 LA.computeLoopNestDepth();
801}
802
Andrew Scull57e12682015-09-16 11:30:19 -0700803// This is a lightweight version of live-range-end calculation. Marks the last
Andrew Scullaa6c1092015-09-03 17:50:30 -0700804// use of only those variables whose definition and uses are completely with a
Andrew Scull57e12682015-09-16 11:30:19 -0700805// single block. It is a quick single pass and doesn't need to iterate until
Andrew Scullaa6c1092015-09-03 17:50:30 -0700806// convergence.
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700807void Cfg::livenessLightweight() {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700808 TimerMarker T(TimerStack::TT_livenessLightweight, this);
Jim Stichnoth877b04e2014-10-15 15:13:06 -0700809 getVMetadata()->init(VMK_Uses);
Jim Stichnothf44f3712014-10-01 14:05:51 -0700810 for (CfgNode *Node : Nodes)
811 Node->livenessLightweight();
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700812}
813
814void Cfg::liveness(LivenessMode Mode) {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700815 TimerMarker T(TimerStack::TT_liveness, this);
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700816 Live.reset(new Liveness(this, Mode));
Jim Stichnoth877b04e2014-10-15 15:13:06 -0700817 getVMetadata()->init(VMK_Uses);
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700818 Live->init();
819 // Initialize with all nodes needing to be processed.
John Porto36d6aa62016-02-26 07:19:59 -0800820 BitVector NeedToProcess(Nodes.size(), true);
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700821 while (NeedToProcess.any()) {
822 // Iterate in reverse topological order to speed up convergence.
Jim Stichnoth7e571362015-01-09 11:43:26 -0800823 for (CfgNode *Node : reverse_range(Nodes)) {
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700824 if (NeedToProcess[Node->getIndex()]) {
825 NeedToProcess[Node->getIndex()] = false;
826 bool Changed = Node->liveness(getLiveness());
827 if (Changed) {
828 // If the beginning-of-block liveness changed since the last
829 // iteration, mark all in-edges as needing to be processed.
Jim Stichnothf44f3712014-10-01 14:05:51 -0700830 for (CfgNode *Pred : Node->getInEdges())
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700831 NeedToProcess[Pred->getIndex()] = true;
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700832 }
833 }
834 }
835 }
836 if (Mode == Liveness_Intervals) {
837 // Reset each variable's live range.
Jim Stichnothf44f3712014-10-01 14:05:51 -0700838 for (Variable *Var : Variables)
839 Var->resetLiveRange();
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700840 }
Andrew Scull57e12682015-09-16 11:30:19 -0700841 // Make a final pass over each node to delete dead instructions, collect the
842 // first and last instruction numbers, and add live range segments for that
843 // node.
Jim Stichnothe5b73e62014-12-15 09:58:51 -0800844 for (CfgNode *Node : Nodes) {
845 InstNumberT FirstInstNum = Inst::NumberSentinel;
846 InstNumberT LastInstNum = Inst::NumberSentinel;
Jim Stichnoth29841e82014-12-23 12:26:24 -0800847 for (Inst &I : Node->getPhis()) {
848 I.deleteIfDead();
849 if (Mode == Liveness_Intervals && !I.isDeleted()) {
Jim Stichnothe5b73e62014-12-15 09:58:51 -0800850 if (FirstInstNum == Inst::NumberSentinel)
Jim Stichnoth29841e82014-12-23 12:26:24 -0800851 FirstInstNum = I.getNumber();
852 assert(I.getNumber() > LastInstNum);
853 LastInstNum = I.getNumber();
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700854 }
Jim Stichnothe5b73e62014-12-15 09:58:51 -0800855 }
Jim Stichnoth29841e82014-12-23 12:26:24 -0800856 for (Inst &I : Node->getInsts()) {
857 I.deleteIfDead();
858 if (Mode == Liveness_Intervals && !I.isDeleted()) {
Jim Stichnothe5b73e62014-12-15 09:58:51 -0800859 if (FirstInstNum == Inst::NumberSentinel)
Jim Stichnoth29841e82014-12-23 12:26:24 -0800860 FirstInstNum = I.getNumber();
861 assert(I.getNumber() > LastInstNum);
862 LastInstNum = I.getNumber();
Jim Stichnothe5b73e62014-12-15 09:58:51 -0800863 }
864 }
865 if (Mode == Liveness_Intervals) {
Andrew Scull57e12682015-09-16 11:30:19 -0700866 // Special treatment for live in-args. Their liveness needs to extend
867 // beyond the beginning of the function, otherwise an arg whose only use
868 // is in the first instruction will end up having the trivial live range
869 // [2,2) and will *not* interfere with other arguments. So if the first
870 // instruction of the method is "r=arg1+arg2", both args may be assigned
871 // the same register. This is accomplished by extending the entry block's
872 // instruction range from [2,n) to [1,n) which will transform the
Jim Stichnoth6f7ad6c2016-01-15 09:52:54 -0800873 // problematic [2,2) live ranges into [1,2). This extension works because
874 // the entry node is guaranteed to have the lowest instruction numbers.
Jim Stichnothe4f65d82015-06-17 22:16:02 -0700875 if (Node == getEntryNode()) {
Jim Stichnothe5b73e62014-12-15 09:58:51 -0800876 FirstInstNum = Inst::NumberExtended;
Jim Stichnoth6f7ad6c2016-01-15 09:52:54 -0800877 // Just in case the entry node somehow contains no instructions...
878 if (LastInstNum == Inst::NumberSentinel)
879 LastInstNum = FirstInstNum;
Jim Stichnothe4f65d82015-06-17 22:16:02 -0700880 }
Jim Stichnoth6f7ad6c2016-01-15 09:52:54 -0800881 // If this node somehow contains no instructions, don't bother trying to
882 // add liveness intervals for it, because variables that are live-in and
883 // live-out will have a bogus interval added.
884 if (FirstInstNum != Inst::NumberSentinel)
885 Node->livenessAddIntervals(getLiveness(), FirstInstNum, LastInstNum);
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700886 }
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700887 }
888}
889
Andrew Scull57e12682015-09-16 11:30:19 -0700890// Traverse every Variable of every Inst and verify that it appears within the
891// Variable's computed live range.
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700892bool Cfg::validateLiveness() const {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700893 TimerMarker T(TimerStack::TT_validateLiveness, this);
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700894 bool Valid = true;
Jim Stichnothe4a8f402015-01-20 12:52:51 -0800895 OstreamLocker L(Ctx);
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700896 Ostream &Str = Ctx->getStrDump();
Jim Stichnothf44f3712014-10-01 14:05:51 -0700897 for (CfgNode *Node : Nodes) {
Jim Stichnothae953202014-12-20 06:17:49 -0800898 Inst *FirstInst = nullptr;
Jim Stichnoth8cfeb692016-02-05 09:50:02 -0800899 for (Inst &Instr : Node->getInsts()) {
900 if (Instr.isDeleted())
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700901 continue;
Jim Stichnothae953202014-12-20 06:17:49 -0800902 if (FirstInst == nullptr)
Jim Stichnoth8cfeb692016-02-05 09:50:02 -0800903 FirstInst = &Instr;
904 InstNumberT InstNumber = Instr.getNumber();
905 if (Variable *Dest = Instr.getDest()) {
Jim Stichnoth47752552014-10-13 17:15:08 -0700906 if (!Dest->getIgnoreLiveness()) {
907 bool Invalid = false;
Jim Stichnoth5bff61c2015-10-28 09:26:00 -0700908 constexpr bool IsDest = true;
Jim Stichnoth47752552014-10-13 17:15:08 -0700909 if (!Dest->getLiveRange().containsValue(InstNumber, IsDest))
910 Invalid = true;
Andrew Scull57e12682015-09-16 11:30:19 -0700911 // Check that this instruction actually *begins* Dest's live range,
912 // by checking that Dest is not live in the previous instruction. As
913 // a special exception, we don't check this for the first instruction
914 // of the block, because a Phi temporary may be live at the end of
915 // the previous block, and if it is also assigned in the first
916 // instruction of this block, the adjacent live ranges get merged.
Jim Stichnoth2d6c8262016-02-07 09:50:27 -0800917 if (&Instr != FirstInst && !Instr.isDestRedefined() &&
Jim Stichnoth47752552014-10-13 17:15:08 -0700918 Dest->getLiveRange().containsValue(InstNumber - 1, IsDest))
919 Invalid = true;
920 if (Invalid) {
921 Valid = false;
Jim Stichnoth8cfeb692016-02-05 09:50:02 -0800922 Str << "Liveness error: inst " << Instr.getNumber() << " dest ";
Jim Stichnoth47752552014-10-13 17:15:08 -0700923 Dest->dump(this);
924 Str << " live range " << Dest->getLiveRange() << "\n";
925 }
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700926 }
927 }
Jim Stichnoth8cfeb692016-02-05 09:50:02 -0800928 FOREACH_VAR_IN_INST(Var, Instr) {
John Portoec3f5652015-08-31 15:07:09 -0700929 static constexpr bool IsDest = false;
930 if (!Var->getIgnoreLiveness() &&
931 !Var->getLiveRange().containsValue(InstNumber, IsDest)) {
932 Valid = false;
Jim Stichnoth8cfeb692016-02-05 09:50:02 -0800933 Str << "Liveness error: inst " << Instr.getNumber() << " var ";
John Portoec3f5652015-08-31 15:07:09 -0700934 Var->dump(this);
935 Str << " live range " << Var->getLiveRange() << "\n";
Jim Stichnothd97c7df2014-06-04 11:57:08 -0700936 }
937 }
938 }
939 }
940 return Valid;
941}
942
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700943void Cfg::contractEmptyNodes() {
Andrew Scullaa6c1092015-09-03 17:50:30 -0700944 // If we're decorating the asm output with register liveness info, this
945 // information may become corrupted or incorrect after contracting nodes that
946 // contain only redundant assignments. As such, we disable this pass when
947 // DecorateAsm is specified. This may make the resulting code look more
948 // branchy, but it should have no effect on the register assignments.
Karl Schimpfdf80eb82015-02-09 14:20:22 -0800949 if (Ctx->getFlags().getDecorateAsm())
Jim Stichnoth3d44fe82014-11-01 10:10:18 -0700950 return;
Jim Stichnoth336f6c42014-10-30 15:01:31 -0700951 for (CfgNode *Node : Nodes) {
952 Node->contractIfEmpty();
953 }
954}
955
Jim Stichnothff9c7062014-09-18 04:50:49 -0700956void Cfg::doBranchOpt() {
Jim Stichnoth8363a062014-10-07 10:02:38 -0700957 TimerMarker T(TimerStack::TT_doBranchOpt, this);
Jim Stichnothf44f3712014-10-01 14:05:51 -0700958 for (auto I = Nodes.begin(), E = Nodes.end(); I != E; ++I) {
Andrew Scull713278a2015-07-22 17:17:02 -0700959 auto NextNode = I + 1;
Jim Stichnothae953202014-12-20 06:17:49 -0800960 (*I)->doBranchOpt(NextNode == E ? nullptr : *NextNode);
Jim Stichnothff9c7062014-09-18 04:50:49 -0700961 }
962}
963
Andrew Scull86df4e92015-07-30 13:54:44 -0700964void Cfg::markNodesForSandboxing() {
965 for (const InstJumpTable *JT : JumpTables)
966 for (SizeT I = 0; I < JT->getNumTargets(); ++I)
967 JT->getTarget(I)->setNeedsAlignment();
968}
969
Jim Stichnothf7c9a142014-04-29 10:52:43 -0700970// ======================== Dump routines ======================== //
971
Andrew Scull57e12682015-09-16 11:30:19 -0700972// emitTextHeader() is not target-specific (apart from what is abstracted by
973// the Assembler), so it is defined here rather than in the target lowering
974// class.
Jim Stichnothbbca7542015-02-11 16:08:31 -0800975void Cfg::emitTextHeader(const IceString &MangledName, GlobalContext *Ctx,
976 const Assembler *Asm) {
Jim Stichnoth20b71f52015-06-24 15:52:24 -0700977 if (!BuildDefs::dump())
Jim Stichnoth729dbd02015-02-25 14:48:43 -0800978 return;
Jan Voung0faec4c2014-11-05 17:29:56 -0800979 Ostream &Str = Ctx->getStrEmit();
980 Str << "\t.text\n";
Karl Schimpfdf80eb82015-02-09 14:20:22 -0800981 if (Ctx->getFlags().getFunctionSections())
John Portoafc92af2015-10-16 10:34:04 -0700982 Str << "\t.section\t.text." << MangledName << ",\"ax\",%progbits\n";
Jim Stichnothbbca7542015-02-11 16:08:31 -0800983 if (!Asm->getInternal() || Ctx->getFlags().getDisableInternal()) {
Jan Voung0faec4c2014-11-05 17:29:56 -0800984 Str << "\t.globl\t" << MangledName << "\n";
Jan Voungb2d50842015-05-12 09:53:50 -0700985 Str << "\t.type\t" << MangledName << ",%function\n";
Jan Voung0faec4c2014-11-05 17:29:56 -0800986 }
Andrew Scull86df4e92015-07-30 13:54:44 -0700987 Str << "\t" << Asm->getAlignDirective() << " "
Jan Voungb2d50842015-05-12 09:53:50 -0700988 << Asm->getBundleAlignLog2Bytes() << ",0x";
Jan Voung08c3bcd2014-12-01 17:55:16 -0800989 for (uint8_t I : Asm->getNonExecBundlePadding())
Jan Voung0faec4c2014-11-05 17:29:56 -0800990 Str.write_hex(I);
991 Str << "\n";
992 Str << MangledName << ":\n";
993}
994
Andrew Scull86df4e92015-07-30 13:54:44 -0700995void Cfg::deleteJumpTableInsts() {
996 for (InstJumpTable *JumpTable : JumpTables)
997 JumpTable->setDeleted();
998}
999
1000void Cfg::emitJumpTables() {
1001 switch (Ctx->getFlags().getOutFileType()) {
1002 case FT_Elf:
1003 case FT_Iasm: {
Andrew Scull57e12682015-09-16 11:30:19 -07001004 // The emission needs to be delayed until the after the text section so
1005 // save the offsets in the global context.
Andrew Scull86df4e92015-07-30 13:54:44 -07001006 IceString MangledName = Ctx->mangleName(getFunctionName());
1007 for (const InstJumpTable *JumpTable : JumpTables) {
1008 SizeT NumTargets = JumpTable->getNumTargets();
David Sehr0fe6b542015-11-19 21:47:15 -08001009 JumpTableData::TargetList TargetList;
Andrew Scull86df4e92015-07-30 13:54:44 -07001010 for (SizeT I = 0; I < NumTargets; ++I) {
1011 SizeT Index = JumpTable->getTarget(I)->getIndex();
David Sehr0fe6b542015-11-19 21:47:15 -08001012 TargetList.emplace_back(
1013 getAssembler()->getCfgNodeLabel(Index)->getPosition());
Andrew Scull86df4e92015-07-30 13:54:44 -07001014 }
David Sehr0fe6b542015-11-19 21:47:15 -08001015 Ctx->addJumpTable(MangledName, JumpTable->getId(), TargetList);
Andrew Scull86df4e92015-07-30 13:54:44 -07001016 }
1017 } break;
1018 case FT_Asm: {
1019 // Emit the assembly directly so we don't need to hang on to all the names
1020 for (const InstJumpTable *JumpTable : JumpTables)
1021 getTarget()->emitJumpTable(this, JumpTable);
1022 } break;
Andrew Scull86df4e92015-07-30 13:54:44 -07001023 }
1024}
1025
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -07001026void Cfg::emit() {
Jim Stichnoth20b71f52015-06-24 15:52:24 -07001027 if (!BuildDefs::dump())
Karl Schimpfb6c96af2014-11-17 10:58:39 -08001028 return;
Jim Stichnoth8363a062014-10-07 10:02:38 -07001029 TimerMarker T(TimerStack::TT_emit, this);
Karl Schimpfdf80eb82015-02-09 14:20:22 -08001030 if (Ctx->getFlags().getDecorateAsm()) {
Jim Stichnoth3d44fe82014-11-01 10:10:18 -07001031 renumberInstructions();
1032 getVMetadata()->init(VMK_Uses);
1033 liveness(Liveness_Basic);
1034 dump("After recomputing liveness for -decorate-asm");
1035 }
Jim Stichnothe4a8f402015-01-20 12:52:51 -08001036 OstreamLocker L(Ctx);
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -07001037 Ostream &Str = Ctx->getStrEmit();
Andrew Scull86df4e92015-07-30 13:54:44 -07001038 IceString MangledName = Ctx->mangleName(getFunctionName());
1039 const Assembler *Asm = getAssembler<>();
1040 const bool NeedSandboxing = Ctx->getFlags().getUseSandboxing();
1041
1042 emitTextHeader(MangledName, Ctx, Asm);
1043 deleteJumpTableInsts();
Jim Stichnoth238b4c12015-10-01 07:46:38 -07001044 if (Ctx->getFlags().getDecorateAsm()) {
1045 for (Variable *Var : getVariables()) {
Jim Stichnoth3607b6c2015-11-13 14:28:23 -08001046 if (Var->getStackOffset() && !Var->isRematerializable()) {
Jim Stichnoth238b4c12015-10-01 07:46:38 -07001047 Str << "\t" << Var->getSymbolicStackOffset(this) << " = "
1048 << Var->getStackOffset() << "\n";
1049 }
1050 }
1051 }
Andrew Scull86df4e92015-07-30 13:54:44 -07001052 for (CfgNode *Node : Nodes) {
1053 if (NeedSandboxing && Node->needsAlignment()) {
1054 Str << "\t" << Asm->getAlignDirective() << " "
1055 << Asm->getBundleAlignLog2Bytes() << "\n";
1056 }
Jim Stichnothf44f3712014-10-01 14:05:51 -07001057 Node->emit(this);
Andrew Scull86df4e92015-07-30 13:54:44 -07001058 }
1059 emitJumpTables();
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -07001060 Str << "\n";
Jim Stichnoth5bc2b1d2014-05-22 13:38:48 -07001061}
1062
Jan Voung0faec4c2014-11-05 17:29:56 -08001063void Cfg::emitIAS() {
1064 TimerMarker T(TimerStack::TT_emit, this);
Andrew Scull57e12682015-09-16 11:30:19 -07001065 // The emitIAS() routines emit into the internal assembler buffer, so there's
1066 // no need to lock the streams.
Andrew Scull86df4e92015-07-30 13:54:44 -07001067 deleteJumpTableInsts();
1068 const bool NeedSandboxing = Ctx->getFlags().getUseSandboxing();
1069 for (CfgNode *Node : Nodes) {
1070 if (NeedSandboxing && Node->needsAlignment())
1071 getAssembler()->alignCfgNode();
Jan Voung0faec4c2014-11-05 17:29:56 -08001072 Node->emitIAS(this);
Andrew Scull86df4e92015-07-30 13:54:44 -07001073 }
1074 emitJumpTables();
Jan Voung0faec4c2014-11-05 17:29:56 -08001075}
1076
Jim Stichnothd97c7df2014-06-04 11:57:08 -07001077// Dumps the IR with an optional introductory message.
1078void Cfg::dump(const IceString &Message) {
Jim Stichnoth20b71f52015-06-24 15:52:24 -07001079 if (!BuildDefs::dump())
Karl Schimpfb6c96af2014-11-17 10:58:39 -08001080 return;
Jim Stichnothfa4efea2015-01-27 05:06:03 -08001081 if (!isVerbose())
Jim Stichnothd97c7df2014-06-04 11:57:08 -07001082 return;
Jim Stichnothe4a8f402015-01-20 12:52:51 -08001083 OstreamLocker L(Ctx);
Jim Stichnothf7c9a142014-04-29 10:52:43 -07001084 Ostream &Str = Ctx->getStrDump();
Jim Stichnothd97c7df2014-06-04 11:57:08 -07001085 if (!Message.empty())
1086 Str << "================ " << Message << " ================\n";
Jim Stichnothf7c9a142014-04-29 10:52:43 -07001087 setCurrentNode(getEntryNode());
1088 // Print function name+args
Jim Stichnothfa4efea2015-01-27 05:06:03 -08001089 if (isVerbose(IceV_Instructions)) {
Jim Stichnothf7c9a142014-04-29 10:52:43 -07001090 Str << "define ";
Karl Schimpfdf80eb82015-02-09 14:20:22 -08001091 if (getInternal() && !Ctx->getFlags().getDisableInternal())
Jim Stichnothf7c9a142014-04-29 10:52:43 -07001092 Str << "internal ";
Karl Schimpfdf6f9d12014-10-20 14:09:00 -07001093 Str << ReturnType << " @" << Ctx->mangleName(getFunctionName()) << "(";
Jim Stichnothf7c9a142014-04-29 10:52:43 -07001094 for (SizeT i = 0; i < Args.size(); ++i) {
1095 if (i > 0)
1096 Str << ", ";
1097 Str << Args[i]->getType() << " ";
1098 Args[i]->dump(this);
1099 }
Jim Stichnothb40595a2016-01-29 06:14:31 -08001100 // Append an extra copy of the function name here, in order to print its
1101 // size stats but not mess up lit tests.
1102 Str << ") { # " << getFunctionNameAndSize() << "\n";
Jim Stichnothf7c9a142014-04-29 10:52:43 -07001103 }
Jim Stichnoth800dab22014-09-20 12:25:02 -07001104 resetCurrentNode();
Jim Stichnothfa4efea2015-01-27 05:06:03 -08001105 if (isVerbose(IceV_Liveness)) {
Jim Stichnothd97c7df2014-06-04 11:57:08 -07001106 // Print summary info about variables
Jim Stichnothf44f3712014-10-01 14:05:51 -07001107 for (Variable *Var : Variables) {
Jim Stichnoth144cdce2014-09-22 16:02:59 -07001108 Str << "// multiblock=";
1109 if (getVMetadata()->isTracked(Var))
1110 Str << getVMetadata()->isMultiBlock(Var);
1111 else
1112 Str << "?";
Jim Stichnoth48e3ae52015-10-01 13:33:35 -07001113 Str << " defs=";
1114 bool FirstPrint = true;
1115 if (VMetadata->getKind() != VMK_Uses) {
1116 if (const Inst *FirstDef = VMetadata->getFirstDefinition(Var)) {
1117 Str << FirstDef->getNumber();
1118 FirstPrint = false;
1119 }
1120 }
1121 if (VMetadata->getKind() == VMK_All) {
1122 for (const Inst *Instr : VMetadata->getLatterDefinitions(Var)) {
1123 if (!FirstPrint)
1124 Str << ",";
1125 Str << Instr->getNumber();
1126 FirstPrint = false;
1127 }
1128 }
Andrew Scull11c9a322015-08-28 14:24:14 -07001129 Str << " weight=" << Var->getWeight(this) << " ";
Jim Stichnothd97c7df2014-06-04 11:57:08 -07001130 Var->dump(this);
1131 Str << " LIVE=" << Var->getLiveRange() << "\n";
1132 }
1133 }
Jim Stichnothf7c9a142014-04-29 10:52:43 -07001134 // Print each basic block
Jim Stichnothf44f3712014-10-01 14:05:51 -07001135 for (CfgNode *Node : Nodes)
1136 Node->dump(this);
Jim Stichnothfa4efea2015-01-27 05:06:03 -08001137 if (isVerbose(IceV_Instructions))
Jim Stichnothf7c9a142014-04-29 10:52:43 -07001138 Str << "}\n";
Jim Stichnothf7c9a142014-04-29 10:52:43 -07001139}
1140
1141} // end of namespace Ice