1 /*
2 * Copyright 2011 Christoph Bumiller
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
21 */
22
23 #include "nv50_ir.h"
24 #include "nv50_ir_target.h"
25
26 #include <algorithm>
27 #include <stack>
28 #include <limits>
29 #include <unordered_map>
30
31 namespace nv50_ir {
32
33 #define MAX_REGISTER_FILE_SIZE 256
34
35 class RegisterSet
36 {
37 public:
38 RegisterSet(const Target *);
39
40 void init(const Target *);
41 void reset(DataFile, bool resetMax = false);
42
43 void periodicMask(DataFile f, uint32_t lock, uint32_t unlock);
44 void intersect(DataFile f, const RegisterSet *);
45
46 bool assign(int32_t& reg, DataFile f, unsigned int size, unsigned int maxReg);
47 void release(DataFile f, int32_t reg, unsigned int size);
48 void occupy(DataFile f, int32_t reg, unsigned int size);
49 void occupy(const Value *);
50 void occupyMask(DataFile f, int32_t reg, uint8_t mask);
51 bool isOccupied(DataFile f, int32_t reg, unsigned int size) const;
52 bool testOccupy(const Value *);
53 bool testOccupy(DataFile f, int32_t reg, unsigned int size);
54
getMaxAssigned(DataFile f) const55 inline int getMaxAssigned(DataFile f) const { return fill[f]; }
56
getFileSize(DataFile f) const57 inline unsigned int getFileSize(DataFile f) const
58 {
59 return last[f] + 1;
60 }
61
units(DataFile f, unsigned int size) const62 inline unsigned int units(DataFile f, unsigned int size) const
63 {
64 return size >> unit[f];
65 }
66 // for regs of size >= 4, id is counted in 4-byte words (like nv50/c0 binary)
idToBytes(const Value *v) const67 inline unsigned int idToBytes(const Value *v) const
68 {
69 return v->reg.data.id * MIN2(v->reg.size, 4);
70 }
idToUnits(const Value *v) const71 inline unsigned int idToUnits(const Value *v) const
72 {
73 return units(v->reg.file, idToBytes(v));
74 }
bytesToId(Value *v, unsigned int bytes) const75 inline int bytesToId(Value *v, unsigned int bytes) const
76 {
77 if (v->reg.size < 4)
78 return units(v->reg.file, bytes);
79 return bytes / 4;
80 }
unitsToId(DataFile f, int u, uint8_t size) const81 inline int unitsToId(DataFile f, int u, uint8_t size) const
82 {
83 if (u < 0)
84 return -1;
85 return (size < 4) ? u : ((u << unit[f]) / 4);
86 }
87
88 void print(DataFile f) const;
89
90 const bool restrictedGPR16Range;
91
92 private:
93 BitSet bits[LAST_REGISTER_FILE + 1];
94
95 int unit[LAST_REGISTER_FILE + 1]; // log2 of allocation granularity
96
97 int last[LAST_REGISTER_FILE + 1];
98 int fill[LAST_REGISTER_FILE + 1];
99 };
100
101 void
reset(DataFile f, bool resetMax)102 RegisterSet::reset(DataFile f, bool resetMax)
103 {
104 bits[f].fill(0);
105 if (resetMax)
106 fill[f] = -1;
107 }
108
109 void
init(const Target *targ)110 RegisterSet::init(const Target *targ)
111 {
112 for (unsigned int rf = 0; rf <= LAST_REGISTER_FILE; ++rf) {
113 DataFile f = static_cast<DataFile>(rf);
114 last[rf] = targ->getFileSize(f) - 1;
115 unit[rf] = targ->getFileUnit(f);
116 fill[rf] = -1;
117 assert(last[rf] < MAX_REGISTER_FILE_SIZE);
118 bits[rf].allocate(last[rf] + 1, true);
119 }
120 }
121
122 RegisterSet::RegisterSet(const Target *targ)
123 : restrictedGPR16Range(targ->getChipset() < 0xc0)
124 {
125 init(targ);
126 for (unsigned int i = 0; i <= LAST_REGISTER_FILE; ++i)
127 reset(static_cast<DataFile>(i));
128 }
129
130 void
periodicMask(DataFile f, uint32_t lock, uint32_t unlock)131 RegisterSet::periodicMask(DataFile f, uint32_t lock, uint32_t unlock)
132 {
133 bits[f].periodicMask32(lock, unlock);
134 }
135
136 void
intersect(DataFile f, const RegisterSet *set)137 RegisterSet::intersect(DataFile f, const RegisterSet *set)
138 {
139 bits[f] |= set->bits[f];
140 }
141
142 void
print(DataFile f) const143 RegisterSet::print(DataFile f) const
144 {
145 INFO("GPR:");
146 bits[f].print();
147 INFO("\n");
148 }
149
150 bool
assign(int32_t& reg, DataFile f, unsigned int size, unsigned int maxReg)151 RegisterSet::assign(int32_t& reg, DataFile f, unsigned int size, unsigned int maxReg)
152 {
153 reg = bits[f].findFreeRange(size, maxReg);
154 if (reg < 0)
155 return false;
156 fill[f] = MAX2(fill[f], (int32_t)(reg + size - 1));
157 return true;
158 }
159
160 bool
isOccupied(DataFile f, int32_t reg, unsigned int size) const161 RegisterSet::isOccupied(DataFile f, int32_t reg, unsigned int size) const
162 {
163 return bits[f].testRange(reg, size);
164 }
165
166 void
occupy(const Value *v)167 RegisterSet::occupy(const Value *v)
168 {
169 occupy(v->reg.file, idToUnits(v), v->reg.size >> unit[v->reg.file]);
170 }
171
172 void
occupyMask(DataFile f, int32_t reg, uint8_t mask)173 RegisterSet::occupyMask(DataFile f, int32_t reg, uint8_t mask)
174 {
175 bits[f].setMask(reg & ~31, static_cast<uint32_t>(mask) << (reg % 32));
176 }
177
178 void
occupy(DataFile f, int32_t reg, unsigned int size)179 RegisterSet::occupy(DataFile f, int32_t reg, unsigned int size)
180 {
181 bits[f].setRange(reg, size);
182
183 INFO_DBG(0, REG_ALLOC, "reg occupy: %u[%i] %u\n", f, reg, size);
184
185 fill[f] = MAX2(fill[f], (int32_t)(reg + size - 1));
186 }
187
188 bool
testOccupy(const Value *v)189 RegisterSet::testOccupy(const Value *v)
190 {
191 return testOccupy(v->reg.file,
192 idToUnits(v), v->reg.size >> unit[v->reg.file]);
193 }
194
195 bool
testOccupy(DataFile f, int32_t reg, unsigned int size)196 RegisterSet::testOccupy(DataFile f, int32_t reg, unsigned int size)
197 {
198 if (isOccupied(f, reg, size))
199 return false;
200 occupy(f, reg, size);
201 return true;
202 }
203
204 void
release(DataFile f, int32_t reg, unsigned int size)205 RegisterSet::release(DataFile f, int32_t reg, unsigned int size)
206 {
207 bits[f].clrRange(reg, size);
208
209 INFO_DBG(0, REG_ALLOC, "reg release: %u[%i] %u\n", f, reg, size);
210 }
211
212 class RegAlloc
213 {
214 public:
RegAlloc(Program *program)215 RegAlloc(Program *program) : prog(program), func(NULL), sequence(0) { }
216
217 bool exec();
218 bool execFunc();
219
220 private:
221 class PhiMovesPass : public Pass {
222 private:
223 virtual bool visit(BasicBlock *);
224 inline bool needNewElseBlock(BasicBlock *b, BasicBlock *p);
225 inline void splitEdges(BasicBlock *b);
226 };
227
228 class ArgumentMovesPass : public Pass {
229 private:
230 virtual bool visit(BasicBlock *);
231 };
232
233 class BuildIntervalsPass : public Pass {
234 private:
235 virtual bool visit(BasicBlock *);
236 void collectLiveValues(BasicBlock *);
237 void addLiveRange(Value *, const BasicBlock *, int end);
238 };
239
240 class InsertConstraintsPass : public Pass {
241 public:
InsertConstraintsPass()242 InsertConstraintsPass() : targ(NULL) { }
243 bool exec(Function *func);
244 private:
245 virtual bool visit(BasicBlock *);
246
247 void insertConstraintMove(Instruction *, int s);
248 bool insertConstraintMoves();
249
250 void condenseDefs(Instruction *);
251 void condenseDefs(Instruction *, const int first, const int last);
252 void condenseSrcs(Instruction *, const int first, const int last);
253
254 void addHazard(Instruction *i, const ValueRef *src);
255 void textureMask(TexInstruction *);
256 void addConstraint(Instruction *, int s, int n);
257 bool detectConflict(Instruction *, int s);
258
259 // target specific functions, TODO: put in subclass or Target
260 void texConstraintNV50(TexInstruction *);
261 void texConstraintNVC0(TexInstruction *);
262 void texConstraintNVE0(TexInstruction *);
263 void texConstraintGM107(TexInstruction *);
264
265 bool isScalarTexGM107(TexInstruction *);
266 void handleScalarTexGM107(TexInstruction *);
267
268 std::list<Instruction *> constrList;
269
270 const Target *targ;
271 };
272
273 bool buildLiveSets(BasicBlock *);
274
275 private:
276 Program *prog;
277 Function *func;
278
279 // instructions in control flow / chronological order
280 ArrayList insns;
281
282 int sequence; // for manual passes through CFG
283 };
284
285 typedef std::pair<Value *, Value *> ValuePair;
286
287 class MergedDefs
288 {
289 private:
entry(Value *val)290 std::list<ValueDef *>& entry(Value *val) {
291 auto it = defs.find(val);
292
293 if (it == defs.end()) {
294 std::list<ValueDef *> &res = defs[val];
295 res = val->defs;
296 return res;
297 } else {
298 return (*it).second;
299 }
300 }
301
302 std::unordered_map<Value *, std::list<ValueDef *> > defs;
303
304 public:
operator ()(Value *val)305 std::list<ValueDef *>& operator()(Value *val) {
306 return entry(val);
307 }
308
add(Value *val, const std::list<ValueDef *> &vals)309 void add(Value *val, const std::list<ValueDef *> &vals) {
310 assert(val);
311 std::list<ValueDef *> &valdefs = entry(val);
312 valdefs.insert(valdefs.end(), vals.begin(), vals.end());
313 }
314
removeDefsOfInstruction(Instruction *insn)315 void removeDefsOfInstruction(Instruction *insn) {
316 for (int d = 0; insn->defExists(d); ++d) {
317 ValueDef *def = &insn->def(d);
318 defs.erase(def->get());
319 for (auto &p : defs)
320 p.second.remove(def);
321 }
322 }
323
merge()324 void merge() {
325 for (auto &p : defs)
326 p.first->defs = p.second;
327 }
328 };
329
330 class SpillCodeInserter
331 {
332 public:
SpillCodeInserter(Function *fn, MergedDefs &mergedDefs)333 SpillCodeInserter(Function *fn, MergedDefs &mergedDefs) : func(fn), mergedDefs(mergedDefs), stackSize(0), stackBase(0) { }
334
335 bool run(const std::list<ValuePair>&);
336
337 Symbol *assignSlot(const Interval&, const unsigned int size);
338 Value *offsetSlot(Value *, const LValue *);
getStackSize() const339 inline int32_t getStackSize() const { return stackSize; }
340
341 private:
342 Function *func;
343 MergedDefs &mergedDefs;
344
345 struct SpillSlot
346 {
347 Interval occup;
348 std::list<Value *> residents; // needed to recalculate occup
349 Symbol *sym;
350 int32_t offset;
sizenv50_ir::SpillCodeInserter::SpillSlot351 inline uint8_t size() const { return sym->reg.size; }
352 };
353 std::list<SpillSlot> slots;
354 int32_t stackSize;
355 int32_t stackBase;
356
357 LValue *unspill(Instruction *usei, LValue *, Value *slot);
358 void spill(Instruction *defi, Value *slot, LValue *);
359 };
360
361 void
addLiveRange(Value *val, const BasicBlock *bb, int end)362 RegAlloc::BuildIntervalsPass::addLiveRange(Value *val,
363 const BasicBlock *bb,
364 int end)
365 {
366 Instruction *insn = val->getUniqueInsn();
367
368 if (!insn)
369 insn = bb->getFirst();
370
371 assert(bb->getFirst()->serial <= bb->getExit()->serial);
372 assert(bb->getExit()->serial + 1 >= end);
373
374 int begin = insn->serial;
375 if (begin < bb->getEntry()->serial || begin > bb->getExit()->serial)
376 begin = bb->getEntry()->serial;
377
378 INFO_DBG(prog->dbgFlags, REG_ALLOC, "%%%i <- live range [%i(%i), %i)\n",
379 val->id, begin, insn->serial, end);
380
381 if (begin != end) // empty ranges are only added as hazards for fixed regs
382 val->livei.extend(begin, end);
383 }
384
385 bool
needNewElseBlock(BasicBlock *b, BasicBlock *p)386 RegAlloc::PhiMovesPass::needNewElseBlock(BasicBlock *b, BasicBlock *p)
387 {
388 if (b->cfg.incidentCount() <= 1)
389 return false;
390
391 int n = 0;
392 for (Graph::EdgeIterator ei = p->cfg.outgoing(); !ei.end(); ei.next())
393 if (ei.getType() == Graph::Edge::TREE ||
394 ei.getType() == Graph::Edge::FORWARD)
395 ++n;
396 return (n == 2);
397 }
398
399 struct PhiMapHash {
operator ()nv50_ir::PhiMapHash400 size_t operator()(const std::pair<Instruction *, BasicBlock *>& val) const {
401 return std::hash<Instruction*>()(val.first) * 31 +
402 std::hash<BasicBlock*>()(val.second);
403 }
404 };
405
406 typedef std::unordered_map<
407 std::pair<Instruction *, BasicBlock *>, Value *, PhiMapHash> PhiMap;
408
409 // Critical edges need to be split up so that work can be inserted along
410 // specific edge transitions. Unfortunately manipulating incident edges into a
411 // BB invalidates all the PHI nodes since their sources are implicitly ordered
412 // by incident edge order.
413 //
414 // TODO: Make it so that that is not the case, and PHI nodes store pointers to
415 // the original BBs.
416 void
splitEdges(BasicBlock *bb)417 RegAlloc::PhiMovesPass::splitEdges(BasicBlock *bb)
418 {
419 BasicBlock *pb, *pn;
420 Instruction *phi;
421 Graph::EdgeIterator ei;
422 std::stack<BasicBlock *> stack;
423 int j = 0;
424
425 for (ei = bb->cfg.incident(); !ei.end(); ei.next()) {
426 pb = BasicBlock::get(ei.getNode());
427 assert(pb);
428 if (needNewElseBlock(bb, pb))
429 stack.push(pb);
430 }
431
432 // No critical edges were found, no need to perform any work.
433 if (stack.empty())
434 return;
435
436 // We're about to, potentially, reorder the inbound edges. This means that
437 // we need to hold on to the (phi, bb) -> src mapping, and fix up the phi
438 // nodes after the graph has been modified.
439 PhiMap phis;
440
441 j = 0;
442 for (ei = bb->cfg.incident(); !ei.end(); ei.next(), j++) {
443 pb = BasicBlock::get(ei.getNode());
444 for (phi = bb->getPhi(); phi && phi->op == OP_PHI; phi = phi->next)
445 phis.insert(std::make_pair(std::make_pair(phi, pb), phi->getSrc(j)));
446 }
447
448 while (!stack.empty()) {
449 pb = stack.top();
450 pn = new BasicBlock(func);
451 stack.pop();
452
453 pb->cfg.detach(&bb->cfg);
454 pb->cfg.attach(&pn->cfg, Graph::Edge::TREE);
455 pn->cfg.attach(&bb->cfg, Graph::Edge::FORWARD);
456
457 assert(pb->getExit()->op != OP_CALL);
458 if (pb->getExit()->asFlow()->target.bb == bb)
459 pb->getExit()->asFlow()->target.bb = pn;
460
461 for (phi = bb->getPhi(); phi && phi->op == OP_PHI; phi = phi->next) {
462 PhiMap::iterator it = phis.find(std::make_pair(phi, pb));
463 assert(it != phis.end());
464 phis.insert(std::make_pair(std::make_pair(phi, pn), it->second));
465 phis.erase(it);
466 }
467 }
468
469 // Now go through and fix up all of the phi node sources.
470 j = 0;
471 for (ei = bb->cfg.incident(); !ei.end(); ei.next(), j++) {
472 pb = BasicBlock::get(ei.getNode());
473 for (phi = bb->getPhi(); phi && phi->op == OP_PHI; phi = phi->next) {
474 PhiMap::const_iterator it = phis.find(std::make_pair(phi, pb));
475 assert(it != phis.end());
476
477 phi->setSrc(j, it->second);
478 }
479 }
480 }
481
482 // For each operand of each PHI in b, generate a new value by inserting a MOV
483 // at the end of the block it is coming from and replace the operand with its
484 // result. This eliminates liveness conflicts and enables us to let values be
485 // copied to the right register if such a conflict exists nonetheless.
486 //
487 // These MOVs are also crucial in making sure the live intervals of phi srces
488 // are extended until the end of the loop, since they are not included in the
489 // live-in sets.
490 bool
visit(BasicBlock *bb)491 RegAlloc::PhiMovesPass::visit(BasicBlock *bb)
492 {
493 Instruction *phi, *mov;
494
495 splitEdges(bb);
496
497 // insert MOVs (phi->src(j) should stem from j-th in-BB)
498 int j = 0;
499 for (Graph::EdgeIterator ei = bb->cfg.incident(); !ei.end(); ei.next()) {
500 BasicBlock *pb = BasicBlock::get(ei.getNode());
501 if (!pb->isTerminated())
502 pb->insertTail(new_FlowInstruction(func, OP_BRA, bb));
503
504 for (phi = bb->getPhi(); phi && phi->op == OP_PHI; phi = phi->next) {
505 LValue *tmp = new_LValue(func, phi->getDef(0)->asLValue());
506 mov = new_Instruction(func, OP_MOV, typeOfSize(tmp->reg.size));
507
508 mov->setSrc(0, phi->getSrc(j));
509 mov->setDef(0, tmp);
510 phi->setSrc(j, tmp);
511
512 pb->insertBefore(pb->getExit(), mov);
513 }
514 ++j;
515 }
516
517 return true;
518 }
519
520 bool
visit(BasicBlock *bb)521 RegAlloc::ArgumentMovesPass::visit(BasicBlock *bb)
522 {
523 // Bind function call inputs/outputs to the same physical register
524 // the callee uses, inserting moves as appropriate for the case a
525 // conflict arises.
526 for (Instruction *i = bb->getEntry(); i; i = i->next) {
527 FlowInstruction *cal = i->asFlow();
528 // TODO: Handle indirect calls.
529 // Right now they should only be generated for builtins.
530 if (!cal || cal->op != OP_CALL || cal->builtin || cal->indirect)
531 continue;
532 RegisterSet clobberSet(prog->getTarget());
533
534 // Bind input values.
535 for (int s = cal->indirect ? 1 : 0; cal->srcExists(s); ++s) {
536 const int t = cal->indirect ? (s - 1) : s;
537 LValue *tmp = new_LValue(func, cal->getSrc(s)->asLValue());
538 tmp->reg.data.id = cal->target.fn->ins[t].rep()->reg.data.id;
539
540 Instruction *mov =
541 new_Instruction(func, OP_MOV, typeOfSize(tmp->reg.size));
542 mov->setDef(0, tmp);
543 mov->setSrc(0, cal->getSrc(s));
544 cal->setSrc(s, tmp);
545
546 bb->insertBefore(cal, mov);
547 }
548
549 // Bind output values.
550 for (int d = 0; cal->defExists(d); ++d) {
551 LValue *tmp = new_LValue(func, cal->getDef(d)->asLValue());
552 tmp->reg.data.id = cal->target.fn->outs[d].rep()->reg.data.id;
553
554 Instruction *mov =
555 new_Instruction(func, OP_MOV, typeOfSize(tmp->reg.size));
556 mov->setSrc(0, tmp);
557 mov->setDef(0, cal->getDef(d));
558 cal->setDef(d, tmp);
559
560 bb->insertAfter(cal, mov);
561 clobberSet.occupy(tmp);
562 }
563
564 // Bind clobbered values.
565 for (std::deque<Value *>::iterator it = cal->target.fn->clobbers.begin();
566 it != cal->target.fn->clobbers.end();
567 ++it) {
568 if (clobberSet.testOccupy(*it)) {
569 Value *tmp = new_LValue(func, (*it)->asLValue());
570 tmp->reg.data.id = (*it)->reg.data.id;
571 cal->setDef(cal->defCount(), tmp);
572 }
573 }
574 }
575
576 // Update the clobber set of the function.
577 if (BasicBlock::get(func->cfgExit) == bb) {
578 func->buildDefSets();
579 for (unsigned int i = 0; i < bb->defSet.getSize(); ++i)
580 if (bb->defSet.test(i))
581 func->clobbers.push_back(func->getLValue(i));
582 }
583
584 return true;
585 }
586
587 // Build the set of live-in variables of bb.
588 bool
buildLiveSets(BasicBlock *bb)589 RegAlloc::buildLiveSets(BasicBlock *bb)
590 {
591 Function *f = bb->getFunction();
592 BasicBlock *bn;
593 Instruction *i;
594 unsigned int s, d;
595
596 INFO_DBG(prog->dbgFlags, REG_ALLOC, "buildLiveSets(BB:%i)\n", bb->getId());
597
598 bb->liveSet.allocate(func->allLValues.getSize(), false);
599
600 int n = 0;
601 for (Graph::EdgeIterator ei = bb->cfg.outgoing(); !ei.end(); ei.next()) {
602 bn = BasicBlock::get(ei.getNode());
603 if (bn == bb)
604 continue;
605 if (bn->cfg.visit(sequence))
606 if (!buildLiveSets(bn))
607 return false;
608 if (n++ || bb->liveSet.marker)
609 bb->liveSet |= bn->liveSet;
610 else
611 bb->liveSet = bn->liveSet;
612 }
613 if (!n && !bb->liveSet.marker)
614 bb->liveSet.fill(0);
615 bb->liveSet.marker = true;
616
617 if (prog->dbgFlags & NV50_IR_DEBUG_REG_ALLOC) {
618 INFO("BB:%i live set of out blocks:\n", bb->getId());
619 bb->liveSet.print();
620 }
621
622 // if (!bb->getEntry())
623 // return true;
624
625 if (bb == BasicBlock::get(f->cfgExit)) {
626 for (std::deque<ValueRef>::iterator it = f->outs.begin();
627 it != f->outs.end(); ++it) {
628 assert(it->get()->asLValue());
629 bb->liveSet.set(it->get()->id);
630 }
631 }
632
633 for (i = bb->getExit(); i && i != bb->getEntry()->prev; i = i->prev) {
634 for (d = 0; i->defExists(d); ++d)
635 bb->liveSet.clr(i->getDef(d)->id);
636 for (s = 0; i->srcExists(s); ++s)
637 if (i->getSrc(s)->asLValue())
638 bb->liveSet.set(i->getSrc(s)->id);
639 }
640 for (i = bb->getPhi(); i && i->op == OP_PHI; i = i->next)
641 bb->liveSet.clr(i->getDef(0)->id);
642
643 if (prog->dbgFlags & NV50_IR_DEBUG_REG_ALLOC) {
644 INFO("BB:%i live set after propagation:\n", bb->getId());
645 bb->liveSet.print();
646 }
647
648 return true;
649 }
650
651 void
collectLiveValues(BasicBlock *bb)652 RegAlloc::BuildIntervalsPass::collectLiveValues(BasicBlock *bb)
653 {
654 BasicBlock *bbA = NULL, *bbB = NULL;
655
656 if (bb->cfg.outgoingCount()) {
657 // trickery to save a loop of OR'ing liveSets
658 // aliasing works fine with BitSet::setOr
659 for (Graph::EdgeIterator ei = bb->cfg.outgoing(); !ei.end(); ei.next()) {
660 if (bbA) {
661 bb->liveSet.setOr(&bbA->liveSet, &bbB->liveSet);
662 bbA = bb;
663 } else {
664 bbA = bbB;
665 }
666 bbB = BasicBlock::get(ei.getNode());
667 }
668 bb->liveSet.setOr(&bbB->liveSet, bbA ? &bbA->liveSet : NULL);
669 } else
670 if (bb->cfg.incidentCount()) {
671 bb->liveSet.fill(0);
672 }
673 }
674
675 bool
visit(BasicBlock *bb)676 RegAlloc::BuildIntervalsPass::visit(BasicBlock *bb)
677 {
678 collectLiveValues(bb);
679
680 INFO_DBG(prog->dbgFlags, REG_ALLOC, "BuildIntervals(BB:%i)\n", bb->getId());
681
682 // go through out blocks and delete phi sources that do not originate from
683 // the current block from the live set
684 for (Graph::EdgeIterator ei = bb->cfg.outgoing(); !ei.end(); ei.next()) {
685 BasicBlock *out = BasicBlock::get(ei.getNode());
686
687 for (Instruction *i = out->getPhi(); i && i->op == OP_PHI; i = i->next) {
688 bb->liveSet.clr(i->getDef(0)->id);
689
690 for (int s = 0; i->srcExists(s); ++s) {
691 assert(i->src(s).getInsn());
692 if (i->getSrc(s)->getUniqueInsn()->bb == bb) // XXX: reachableBy ?
693 bb->liveSet.set(i->getSrc(s)->id);
694 else
695 bb->liveSet.clr(i->getSrc(s)->id);
696 }
697 }
698 }
699
700 // remaining live-outs are live until end
701 if (bb->getExit()) {
702 for (unsigned int j = 0; j < bb->liveSet.getSize(); ++j)
703 if (bb->liveSet.test(j))
704 addLiveRange(func->getLValue(j), bb, bb->getExit()->serial + 1);
705 }
706
707 for (Instruction *i = bb->getExit(); i && i->op != OP_PHI; i = i->prev) {
708 for (int d = 0; i->defExists(d); ++d) {
709 bb->liveSet.clr(i->getDef(d)->id);
710 if (i->getDef(d)->reg.data.id >= 0) // add hazard for fixed regs
711 i->getDef(d)->livei.extend(i->serial, i->serial);
712 }
713
714 for (int s = 0; i->srcExists(s); ++s) {
715 if (!i->getSrc(s)->asLValue())
716 continue;
717 if (!bb->liveSet.test(i->getSrc(s)->id)) {
718 bb->liveSet.set(i->getSrc(s)->id);
719 addLiveRange(i->getSrc(s), bb, i->serial);
720 }
721 }
722 }
723
724 if (bb == BasicBlock::get(func->cfg.getRoot())) {
725 for (std::deque<ValueDef>::iterator it = func->ins.begin();
726 it != func->ins.end(); ++it) {
727 if (it->get()->reg.data.id >= 0) // add hazard for fixed regs
728 it->get()->livei.extend(0, 1);
729 }
730 }
731
732 return true;
733 }
734
735
736 #define JOIN_MASK_PHI (1 << 0)
737 #define JOIN_MASK_UNION (1 << 1)
738 #define JOIN_MASK_MOV (1 << 2)
739 #define JOIN_MASK_TEX (1 << 3)
740
741 class GCRA
742 {
743 public:
744 GCRA(Function *, SpillCodeInserter&, MergedDefs&);
745 ~GCRA();
746
747 bool allocateRegisters(ArrayList& insns);
748
749 void printNodeInfo() const;
750
751 private:
752 class RIG_Node : public Graph::Node
753 {
754 public:
755 RIG_Node();
756
757 void init(const RegisterSet&, LValue *);
758
759 void addInterference(RIG_Node *);
760 void addRegPreference(RIG_Node *);
761
getValue() const762 inline LValue *getValue() const
763 {
764 return reinterpret_cast<LValue *>(data);
765 }
setValue(LValue *lval)766 inline void setValue(LValue *lval) { data = lval; }
767
getCompMask() const768 inline uint8_t getCompMask() const
769 {
770 return ((1 << colors) - 1) << (reg & 7);
771 }
772
get(const Graph::EdgeIterator& ei)773 static inline RIG_Node *get(const Graph::EdgeIterator& ei)
774 {
775 return static_cast<RIG_Node *>(ei.getNode());
776 }
777
778 public:
779 uint32_t degree;
780 uint16_t degreeLimit; // if deg < degLimit, node is trivially colourable
781 uint16_t maxReg;
782 uint16_t colors;
783
784 DataFile f;
785 int32_t reg;
786
787 float weight;
788
789 // list pointers for simplify() phase
790 RIG_Node *next;
791 RIG_Node *prev;
792
793 // union of the live intervals of all coalesced values (we want to retain
794 // the separate intervals for testing interference of compound values)
795 Interval livei;
796
797 std::list<RIG_Node *> prefRegs;
798 };
799
800 private:
getNode(const LValue *v) const801 inline RIG_Node *getNode(const LValue *v) const { return &nodes[v->id]; }
802
803 void buildRIG(ArrayList&);
804 bool coalesce(ArrayList&);
805 bool doCoalesce(ArrayList&, unsigned int mask);
806 void calculateSpillWeights();
807 bool simplify();
808 bool selectRegisters();
809 void cleanup(const bool success);
810
811 void simplifyEdge(RIG_Node *, RIG_Node *);
812 void simplifyNode(RIG_Node *);
813
814 void copyCompound(Value *dst, Value *src);
815 bool coalesceValues(Value *, Value *, bool force);
816 void resolveSplitsAndMerges();
817 void makeCompound(Instruction *, bool isSplit);
818
819 inline void checkInterference(const RIG_Node *, Graph::EdgeIterator&);
820
821 inline void insertOrderedTail(std::list<RIG_Node *>&, RIG_Node *);
822 void checkList(std::list<RIG_Node *>&);
823
824 private:
825 std::stack<uint32_t> stack;
826
827 // list headers for simplify() phase
828 RIG_Node lo[2];
829 RIG_Node hi;
830
831 Graph RIG;
832 RIG_Node *nodes;
833 unsigned int nodeCount;
834
835 Function *func;
836 Program *prog;
837
838 struct RelDegree {
839 uint8_t data[17][17];
840
RelDegreenv50_ir::GCRA::RelDegree841 RelDegree() {
842 for (int i = 1; i <= 16; ++i)
843 for (int j = 1; j <= 16; ++j)
844 data[i][j] = j * ((i + j - 1) / j);
845 }
846
operator []nv50_ir::GCRA::RelDegree847 const uint8_t* operator[](std::size_t i) const {
848 return data[i];
849 }
850 };
851
852 static const RelDegree relDegree;
853
854 RegisterSet regs;
855
856 // need to fixup register id for participants of OP_MERGE/SPLIT
857 std::list<Instruction *> merges;
858 std::list<Instruction *> splits;
859
860 SpillCodeInserter& spill;
861 std::list<ValuePair> mustSpill;
862
863 MergedDefs &mergedDefs;
864 };
865
866 const GCRA::RelDegree GCRA::relDegree;
867
RIG_Node()868 GCRA::RIG_Node::RIG_Node() : Node(NULL), degree(0), degreeLimit(0), maxReg(0),
869 colors(0), f(FILE_NULL), reg(0), weight(0), next(this), prev(this)
870 {
871 }
872
873 void
printNodeInfo() const874 GCRA::printNodeInfo() const
875 {
876 for (unsigned int i = 0; i < nodeCount; ++i) {
877 if (!nodes[i].colors)
878 continue;
879 INFO("RIG_Node[%%%i]($[%u]%i): %u colors, weight %f, deg %u/%u\n X",
880 i,
881 nodes[i].f,nodes[i].reg,nodes[i].colors,
882 nodes[i].weight,
883 nodes[i].degree, nodes[i].degreeLimit);
884
885 for (Graph::EdgeIterator ei = nodes[i].outgoing(); !ei.end(); ei.next())
886 INFO(" %%%i", RIG_Node::get(ei)->getValue()->id);
887 for (Graph::EdgeIterator ei = nodes[i].incident(); !ei.end(); ei.next())
888 INFO(" %%%i", RIG_Node::get(ei)->getValue()->id);
889 INFO("\n");
890 }
891 }
892
893 static bool
isShortRegOp(Instruction *insn)894 isShortRegOp(Instruction *insn)
895 {
896 // Immediates are always in src1 (except zeroes, which end up getting
897 // replaced with a zero reg). Every other situation can be resolved by
898 // using a long encoding.
899 return insn->srcExists(1) && insn->src(1).getFile() == FILE_IMMEDIATE &&
900 insn->getSrc(1)->reg.data.u64;
901 }
902
903 // Check if this LValue is ever used in an instruction that can't be encoded
904 // with long registers (i.e. > r63)
905 static bool
isShortRegVal(LValue *lval)906 isShortRegVal(LValue *lval)
907 {
908 if (lval->getInsn() == NULL)
909 return false;
910 for (Value::DefCIterator def = lval->defs.begin();
911 def != lval->defs.end(); ++def)
912 if (isShortRegOp((*def)->getInsn()))
913 return true;
914 for (Value::UseCIterator use = lval->uses.begin();
915 use != lval->uses.end(); ++use)
916 if (isShortRegOp((*use)->getInsn()))
917 return true;
918 return false;
919 }
920
921 void
init(const RegisterSet& regs, LValue *lval)922 GCRA::RIG_Node::init(const RegisterSet& regs, LValue *lval)
923 {
924 setValue(lval);
925 if (lval->reg.data.id >= 0)
926 lval->noSpill = lval->fixedReg = 1;
927
928 colors = regs.units(lval->reg.file, lval->reg.size);
929 f = lval->reg.file;
930 reg = -1;
931 if (lval->reg.data.id >= 0)
932 reg = regs.idToUnits(lval);
933
934 weight = std::numeric_limits<float>::infinity();
935 degree = 0;
936 maxReg = regs.getFileSize(f);
937 // On nv50, we lose a bit of gpr encoding when there's an embedded
938 // immediate.
939 if (regs.restrictedGPR16Range && f == FILE_GPR && (lval->reg.size == 2 || isShortRegVal(lval)))
940 maxReg /= 2;
941 degreeLimit = maxReg;
942 degreeLimit -= relDegree[1][colors] - 1;
943
944 livei.insert(lval->livei);
945 }
946
947 // Used when coalescing moves. The non-compound value will become one, e.g.:
948 // mov b32 $r0 $r2 / merge b64 $r0d { $r0 $r1 }
949 // split b64 { $r0 $r1 } $r0d / mov b64 $r0d f64 $r2d
950 void
copyCompound(Value *dst, Value *src)951 GCRA::copyCompound(Value *dst, Value *src)
952 {
953 LValue *ldst = dst->asLValue();
954 LValue *lsrc = src->asLValue();
955
956 if (ldst->compound && !lsrc->compound) {
957 LValue *swap = lsrc;
958 lsrc = ldst;
959 ldst = swap;
960 }
961
962 assert(!ldst->compound);
963
964 if (lsrc->compound) {
965 for (ValueDef *d : mergedDefs(ldst->join)) {
966 LValue *ldst = d->get()->asLValue();
967 if (!ldst->compound)
968 ldst->compMask = 0xff;
969 ldst->compound = 1;
970 ldst->compMask &= lsrc->compMask;
971 }
972 }
973 }
974
975 bool
coalesceValues(Value *dst, Value *src, bool force)976 GCRA::coalesceValues(Value *dst, Value *src, bool force)
977 {
978 LValue *rep = dst->join->asLValue();
979 LValue *val = src->join->asLValue();
980
981 if (!force && val->reg.data.id >= 0) {
982 rep = src->join->asLValue();
983 val = dst->join->asLValue();
984 }
985 RIG_Node *nRep = &nodes[rep->id];
986 RIG_Node *nVal = &nodes[val->id];
987
988 if (src->reg.file != dst->reg.file) {
989 if (!force)
990 return false;
991 WARN("forced coalescing of values in different files !\n");
992 }
993 if (!force && dst->reg.size != src->reg.size)
994 return false;
995
996 if ((rep->reg.data.id >= 0) && (rep->reg.data.id != val->reg.data.id)) {
997 if (force) {
998 if (val->reg.data.id >= 0)
999 WARN("forced coalescing of values in different fixed regs !\n");
1000 } else {
1001 if (val->reg.data.id >= 0)
1002 return false;
1003 // make sure that there is no overlap with the fixed register of rep
1004 for (ArrayList::Iterator it = func->allLValues.iterator();
1005 !it.end(); it.next()) {
1006 Value *reg = reinterpret_cast<Value *>(it.get())->asLValue();
1007 assert(reg);
1008 if (reg->interfers(rep) && reg->livei.overlaps(nVal->livei))
1009 return false;
1010 }
1011 }
1012 }
1013
1014 if (!force && nRep->livei.overlaps(nVal->livei))
1015 return false;
1016
1017 // TODO: Handle this case properly.
1018 if (!force && rep->compound && val->compound)
1019 return false;
1020
1021 INFO_DBG(prog->dbgFlags, REG_ALLOC, "joining %%%i($%i) <- %%%i\n",
1022 rep->id, rep->reg.data.id, val->id);
1023
1024 if (!force)
1025 copyCompound(dst, src);
1026
1027 // set join pointer of all values joined with val
1028 const std::list<ValueDef *> &defs = mergedDefs(val);
1029 for (ValueDef *def : defs)
1030 def->get()->join = rep;
1031 assert(rep->join == rep && val->join == rep);
1032
1033 // add val's definitions to rep and extend the live interval of its RIG node
1034 mergedDefs.add(rep, defs);
1035 nRep->livei.unify(nVal->livei);
1036 nRep->degreeLimit = MIN2(nRep->degreeLimit, nVal->degreeLimit);
1037 nRep->maxReg = MIN2(nRep->maxReg, nVal->maxReg);
1038 return true;
1039 }
1040
1041 bool
coalesce(ArrayList& insns)1042 GCRA::coalesce(ArrayList& insns)
1043 {
1044 bool ret = doCoalesce(insns, JOIN_MASK_PHI);
1045 if (!ret)
1046 return false;
1047 switch (func->getProgram()->getTarget()->getChipset() & ~0xf) {
1048 case 0x50:
1049 case 0x80:
1050 case 0x90:
1051 case 0xa0:
1052 ret = doCoalesce(insns, JOIN_MASK_UNION | JOIN_MASK_TEX);
1053 break;
1054 case 0xc0:
1055 case 0xd0:
1056 case 0xe0:
1057 case 0xf0:
1058 case 0x100:
1059 case 0x110:
1060 case 0x120:
1061 case 0x130:
1062 case 0x140:
1063 case 0x160:
1064 case 0x170:
1065 ret = doCoalesce(insns, JOIN_MASK_UNION);
1066 break;
1067 default:
1068 break;
1069 }
1070 if (!ret)
1071 return false;
1072 return doCoalesce(insns, JOIN_MASK_MOV);
1073 }
1074
makeCompMask(int compSize, int base, int size)1075 static inline uint8_t makeCompMask(int compSize, int base, int size)
1076 {
1077 uint8_t m = ((1 << size) - 1) << base;
1078
1079 switch (compSize) {
1080 case 1:
1081 return 0xff;
1082 case 2:
1083 m |= (m << 2);
1084 return (m << 4) | m;
1085 case 3:
1086 case 4:
1087 return (m << 4) | m;
1088 default:
1089 assert(compSize <= 8);
1090 return m;
1091 }
1092 }
1093
1094 void
makeCompound(Instruction *insn, bool split)1095 GCRA::makeCompound(Instruction *insn, bool split)
1096 {
1097 LValue *rep = (split ? insn->getSrc(0) : insn->getDef(0))->asLValue();
1098
1099 if (prog->dbgFlags & NV50_IR_DEBUG_REG_ALLOC) {
1100 INFO("makeCompound(split = %i): ", split);
1101 insn->print();
1102 }
1103
1104 const unsigned int size = getNode(rep)->colors;
1105 unsigned int base = 0;
1106
1107 if (!rep->compound)
1108 rep->compMask = 0xff;
1109 rep->compound = 1;
1110
1111 for (int c = 0; split ? insn->defExists(c) : insn->srcExists(c); ++c) {
1112 LValue *val = (split ? insn->getDef(c) : insn->getSrc(c))->asLValue();
1113
1114 val->compound = 1;
1115 if (!val->compMask)
1116 val->compMask = 0xff;
1117 val->compMask &= makeCompMask(size, base, getNode(val)->colors);
1118 assert(val->compMask);
1119
1120 INFO_DBG(prog->dbgFlags, REG_ALLOC, "compound: %%%i:%02x <- %%%i:%02x\n",
1121 rep->id, rep->compMask, val->id, val->compMask);
1122
1123 base += getNode(val)->colors;
1124 }
1125 assert(base == size);
1126 }
1127
1128 bool
doCoalesce(ArrayList& insns, unsigned int mask)1129 GCRA::doCoalesce(ArrayList& insns, unsigned int mask)
1130 {
1131 int c, n;
1132
1133 for (n = 0; n < insns.getSize(); ++n) {
1134 Instruction *i;
1135 Instruction *insn = reinterpret_cast<Instruction *>(insns.get(n));
1136
1137 switch (insn->op) {
1138 case OP_PHI:
1139 if (!(mask & JOIN_MASK_PHI))
1140 break;
1141 for (c = 0; insn->srcExists(c); ++c)
1142 if (!coalesceValues(insn->getDef(0), insn->getSrc(c), false)) {
1143 // this is bad
1144 ERROR("failed to coalesce phi operands\n");
1145 return false;
1146 }
1147 break;
1148 case OP_UNION:
1149 case OP_MERGE:
1150 if (!(mask & JOIN_MASK_UNION))
1151 break;
1152 for (c = 0; insn->srcExists(c); ++c)
1153 coalesceValues(insn->getDef(0), insn->getSrc(c), true);
1154 if (insn->op == OP_MERGE) {
1155 merges.push_back(insn);
1156 if (insn->srcExists(1))
1157 makeCompound(insn, false);
1158 }
1159 break;
1160 case OP_SPLIT:
1161 if (!(mask & JOIN_MASK_UNION))
1162 break;
1163 splits.push_back(insn);
1164 for (c = 0; insn->defExists(c); ++c)
1165 coalesceValues(insn->getSrc(0), insn->getDef(c), true);
1166 makeCompound(insn, true);
1167 break;
1168 case OP_MOV:
1169 if (!(mask & JOIN_MASK_MOV))
1170 break;
1171 i = NULL;
1172 if (!insn->getDef(0)->uses.empty())
1173 i = (*insn->getDef(0)->uses.begin())->getInsn();
1174 // if this is a contraint-move there will only be a single use
1175 if (i && i->op == OP_MERGE) // do we really still need this ?
1176 break;
1177 i = insn->getSrc(0)->getUniqueInsn();
1178 if (i && !i->constrainedDefs()) {
1179 coalesceValues(insn->getDef(0), insn->getSrc(0), false);
1180 }
1181 break;
1182 case OP_TEX:
1183 case OP_TXB:
1184 case OP_TXL:
1185 case OP_TXF:
1186 case OP_TXQ:
1187 case OP_TXD:
1188 case OP_TXG:
1189 case OP_TXLQ:
1190 case OP_TEXCSAA:
1191 case OP_TEXPREP:
1192 if (!(mask & JOIN_MASK_TEX))
1193 break;
1194 for (c = 0; insn->srcExists(c) && c != insn->predSrc; ++c)
1195 coalesceValues(insn->getDef(c), insn->getSrc(c), true);
1196 break;
1197 default:
1198 break;
1199 }
1200 }
1201 return true;
1202 }
1203
1204 void
addInterference(RIG_Node *node)1205 GCRA::RIG_Node::addInterference(RIG_Node *node)
1206 {
1207 this->degree += relDegree[node->colors][colors];
1208 node->degree += relDegree[colors][node->colors];
1209
1210 this->attach(node, Graph::Edge::CROSS);
1211 }
1212
1213 void
addRegPreference(RIG_Node *node)1214 GCRA::RIG_Node::addRegPreference(RIG_Node *node)
1215 {
1216 prefRegs.push_back(node);
1217 }
1218
GCRA(Function *fn, SpillCodeInserter& spill, MergedDefs& mergedDefs)1219 GCRA::GCRA(Function *fn, SpillCodeInserter& spill, MergedDefs& mergedDefs) :
1220 nodes(NULL),
1221 nodeCount(0),
1222 func(fn),
1223 regs(fn->getProgram()->getTarget()),
1224 spill(spill),
1225 mergedDefs(mergedDefs)
1226 {
1227 prog = func->getProgram();
1228 }
1229
~GCRA()1230 GCRA::~GCRA()
1231 {
1232 if (nodes)
1233 delete[] nodes;
1234 }
1235
1236 void
checkList(std::list<RIG_Node *>& lst)1237 GCRA::checkList(std::list<RIG_Node *>& lst)
1238 {
1239 GCRA::RIG_Node *prev = NULL;
1240
1241 for (std::list<RIG_Node *>::iterator it = lst.begin();
1242 it != lst.end();
1243 ++it) {
1244 assert((*it)->getValue()->join == (*it)->getValue());
1245 if (prev)
1246 assert(prev->livei.begin() <= (*it)->livei.begin());
1247 prev = *it;
1248 }
1249 }
1250
1251 void
insertOrderedTail(std::list<RIG_Node *>& list, RIG_Node *node)1252 GCRA::insertOrderedTail(std::list<RIG_Node *>& list, RIG_Node *node)
1253 {
1254 if (node->livei.isEmpty())
1255 return;
1256 // only the intervals of joined values don't necessarily arrive in order
1257 std::list<RIG_Node *>::iterator prev, it;
1258 for (it = list.end(); it != list.begin(); it = prev) {
1259 prev = it;
1260 --prev;
1261 if ((*prev)->livei.begin() <= node->livei.begin())
1262 break;
1263 }
1264 list.insert(it, node);
1265 }
1266
1267 void
buildRIG(ArrayList& insns)1268 GCRA::buildRIG(ArrayList& insns)
1269 {
1270 std::list<RIG_Node *> values, active;
1271
1272 for (std::deque<ValueDef>::iterator it = func->ins.begin();
1273 it != func->ins.end(); ++it)
1274 insertOrderedTail(values, getNode(it->get()->asLValue()));
1275
1276 for (int i = 0; i < insns.getSize(); ++i) {
1277 Instruction *insn = reinterpret_cast<Instruction *>(insns.get(i));
1278 for (int d = 0; insn->defExists(d); ++d)
1279 if (insn->getDef(d)->reg.file <= LAST_REGISTER_FILE &&
1280 insn->getDef(d)->rep() == insn->getDef(d))
1281 insertOrderedTail(values, getNode(insn->getDef(d)->asLValue()));
1282 }
1283 checkList(values);
1284
1285 while (!values.empty()) {
1286 RIG_Node *cur = values.front();
1287
1288 for (std::list<RIG_Node *>::iterator it = active.begin();
1289 it != active.end();) {
1290 RIG_Node *node = *it;
1291
1292 if (node->livei.end() <= cur->livei.begin()) {
1293 it = active.erase(it);
1294 } else {
1295 if (node->f == cur->f && node->livei.overlaps(cur->livei))
1296 cur->addInterference(node);
1297 ++it;
1298 }
1299 }
1300 values.pop_front();
1301 active.push_back(cur);
1302 }
1303 }
1304
1305 void
calculateSpillWeights()1306 GCRA::calculateSpillWeights()
1307 {
1308 for (unsigned int i = 0; i < nodeCount; ++i) {
1309 RIG_Node *const n = &nodes[i];
1310 if (!nodes[i].colors || nodes[i].livei.isEmpty())
1311 continue;
1312 if (nodes[i].reg >= 0) {
1313 // update max reg
1314 regs.occupy(n->f, n->reg, n->colors);
1315 continue;
1316 }
1317 LValue *val = nodes[i].getValue();
1318
1319 if (!val->noSpill) {
1320 int rc = 0;
1321 for (ValueDef *def : mergedDefs(val))
1322 rc += def->get()->refCount();
1323
1324 nodes[i].weight =
1325 (float)rc * (float)rc / (float)nodes[i].livei.extent();
1326 }
1327
1328 if (nodes[i].degree < nodes[i].degreeLimit) {
1329 int l = 0;
1330 if (val->reg.size > 4)
1331 l = 1;
1332 DLLIST_ADDHEAD(&lo[l], &nodes[i]);
1333 } else {
1334 DLLIST_ADDHEAD(&hi, &nodes[i]);
1335 }
1336 }
1337 if (prog->dbgFlags & NV50_IR_DEBUG_REG_ALLOC)
1338 printNodeInfo();
1339 }
1340
1341 void
simplifyEdge(RIG_Node *a, RIG_Node *b)1342 GCRA::simplifyEdge(RIG_Node *a, RIG_Node *b)
1343 {
1344 bool move = b->degree >= b->degreeLimit;
1345
1346 INFO_DBG(prog->dbgFlags, REG_ALLOC,
1347 "edge: (%%%i, deg %u/%u) >-< (%%%i, deg %u/%u)\n",
1348 a->getValue()->id, a->degree, a->degreeLimit,
1349 b->getValue()->id, b->degree, b->degreeLimit);
1350
1351 b->degree -= relDegree[a->colors][b->colors];
1352
1353 move = move && b->degree < b->degreeLimit;
1354 if (move && !DLLIST_EMPTY(b)) {
1355 int l = (b->getValue()->reg.size > 4) ? 1 : 0;
1356 DLLIST_DEL(b);
1357 DLLIST_ADDTAIL(&lo[l], b);
1358 }
1359 }
1360
1361 void
simplifyNode(RIG_Node *node)1362 GCRA::simplifyNode(RIG_Node *node)
1363 {
1364 for (Graph::EdgeIterator ei = node->outgoing(); !ei.end(); ei.next())
1365 simplifyEdge(node, RIG_Node::get(ei));
1366
1367 for (Graph::EdgeIterator ei = node->incident(); !ei.end(); ei.next())
1368 simplifyEdge(node, RIG_Node::get(ei));
1369
1370 DLLIST_DEL(node);
1371 stack.push(node->getValue()->id);
1372
1373 INFO_DBG(prog->dbgFlags, REG_ALLOC, "SIMPLIFY: pushed %%%i%s\n",
1374 node->getValue()->id,
1375 (node->degree < node->degreeLimit) ? "" : "(spill)");
1376 }
1377
1378 bool
simplify()1379 GCRA::simplify()
1380 {
1381 for (;;) {
1382 if (!DLLIST_EMPTY(&lo[0])) {
1383 do {
1384 simplifyNode(lo[0].next);
1385 } while (!DLLIST_EMPTY(&lo[0]));
1386 } else
1387 if (!DLLIST_EMPTY(&lo[1])) {
1388 simplifyNode(lo[1].next);
1389 } else
1390 if (!DLLIST_EMPTY(&hi)) {
1391 RIG_Node *best = hi.next;
1392 unsigned bestMaxReg = best->maxReg;
1393 float bestScore = best->weight / (float)best->degree;
1394 // Spill candidate. First go through the ones with the highest max
1395 // register, then the ones with lower. That way the ones with the
1396 // lowest requirement will be allocated first, since it's a stack.
1397 for (RIG_Node *it = best->next; it != &hi; it = it->next) {
1398 float score = it->weight / (float)it->degree;
1399 if (score < bestScore || it->maxReg > bestMaxReg) {
1400 best = it;
1401 bestScore = score;
1402 bestMaxReg = it->maxReg;
1403 }
1404 }
1405 if (isinf(bestScore)) {
1406 ERROR("no viable spill candidates left\n");
1407 return false;
1408 }
1409 simplifyNode(best);
1410 } else {
1411 return true;
1412 }
1413 }
1414 }
1415
1416 void
checkInterference(const RIG_Node *node, Graph::EdgeIterator& ei)1417 GCRA::checkInterference(const RIG_Node *node, Graph::EdgeIterator& ei)
1418 {
1419 const RIG_Node *intf = RIG_Node::get(ei);
1420
1421 if (intf->reg < 0)
1422 return;
1423 LValue *vA = node->getValue();
1424 LValue *vB = intf->getValue();
1425
1426 const uint8_t intfMask = ((1 << intf->colors) - 1) << (intf->reg & 7);
1427
1428 if (vA->compound | vB->compound) {
1429 // NOTE: this only works for >aligned< register tuples !
1430 for (const ValueDef *D : mergedDefs(vA)) {
1431 for (const ValueDef *d : mergedDefs(vB)) {
1432 const LValue *vD = D->get()->asLValue();
1433 const LValue *vd = d->get()->asLValue();
1434
1435 if (!vD->livei.overlaps(vd->livei)) {
1436 INFO_DBG(prog->dbgFlags, REG_ALLOC, "(%%%i) X (%%%i): no overlap\n",
1437 vD->id, vd->id);
1438 continue;
1439 }
1440
1441 uint8_t mask = vD->compound ? vD->compMask : ~0;
1442 if (vd->compound) {
1443 assert(vB->compound);
1444 mask &= vd->compMask & vB->compMask;
1445 } else {
1446 mask &= intfMask;
1447 }
1448
1449 INFO_DBG(prog->dbgFlags, REG_ALLOC,
1450 "(%%%i)%02x X (%%%i)%02x & %02x: $r%i.%02x\n",
1451 vD->id,
1452 vD->compound ? vD->compMask : 0xff,
1453 vd->id,
1454 vd->compound ? vd->compMask : intfMask,
1455 vB->compMask, intf->reg & ~7, mask);
1456 if (mask)
1457 regs.occupyMask(node->f, intf->reg & ~7, mask);
1458 }
1459 }
1460 } else {
1461 INFO_DBG(prog->dbgFlags, REG_ALLOC,
1462 "(%%%i) X (%%%i): $r%i + %u\n",
1463 vA->id, vB->id, intf->reg, intf->colors);
1464 regs.occupy(node->f, intf->reg, intf->colors);
1465 }
1466 }
1467
1468 bool
selectRegisters()1469 GCRA::selectRegisters()
1470 {
1471 INFO_DBG(prog->dbgFlags, REG_ALLOC, "\nSELECT phase\n");
1472
1473 while (!stack.empty()) {
1474 RIG_Node *node = &nodes[stack.top()];
1475 stack.pop();
1476
1477 regs.reset(node->f);
1478
1479 INFO_DBG(prog->dbgFlags, REG_ALLOC, "\nNODE[%%%i, %u colors]\n",
1480 node->getValue()->id, node->colors);
1481
1482 for (Graph::EdgeIterator ei = node->outgoing(); !ei.end(); ei.next())
1483 checkInterference(node, ei);
1484 for (Graph::EdgeIterator ei = node->incident(); !ei.end(); ei.next())
1485 checkInterference(node, ei);
1486
1487 if (!node->prefRegs.empty()) {
1488 for (std::list<RIG_Node *>::const_iterator it = node->prefRegs.begin();
1489 it != node->prefRegs.end();
1490 ++it) {
1491 if ((*it)->reg >= 0 &&
1492 regs.testOccupy(node->f, (*it)->reg, node->colors)) {
1493 node->reg = (*it)->reg;
1494 break;
1495 }
1496 }
1497 }
1498 if (node->reg >= 0)
1499 continue;
1500 LValue *lval = node->getValue();
1501 if (prog->dbgFlags & NV50_IR_DEBUG_REG_ALLOC)
1502 regs.print(node->f);
1503 bool ret = regs.assign(node->reg, node->f, node->colors, node->maxReg);
1504 if (ret) {
1505 INFO_DBG(prog->dbgFlags, REG_ALLOC, "assigned reg %i\n", node->reg);
1506 lval->compMask = node->getCompMask();
1507 } else {
1508 INFO_DBG(prog->dbgFlags, REG_ALLOC, "must spill: %%%i (size %u)\n",
1509 lval->id, lval->reg.size);
1510 Symbol *slot = NULL;
1511 if (lval->reg.file == FILE_GPR)
1512 slot = spill.assignSlot(node->livei, lval->reg.size);
1513 mustSpill.push_back(ValuePair(lval, slot));
1514 }
1515 }
1516 if (!mustSpill.empty())
1517 return false;
1518 for (unsigned int i = 0; i < nodeCount; ++i) {
1519 LValue *lval = nodes[i].getValue();
1520 if (nodes[i].reg >= 0 && nodes[i].colors > 0)
1521 lval->reg.data.id =
1522 regs.unitsToId(nodes[i].f, nodes[i].reg, lval->reg.size);
1523 }
1524 return true;
1525 }
1526
1527 bool
allocateRegisters(ArrayList& insns)1528 GCRA::allocateRegisters(ArrayList& insns)
1529 {
1530 bool ret;
1531
1532 INFO_DBG(prog->dbgFlags, REG_ALLOC,
1533 "allocateRegisters to %u instructions\n", insns.getSize());
1534
1535 nodeCount = func->allLValues.getSize();
1536 nodes = new RIG_Node[nodeCount];
1537 if (!nodes)
1538 return false;
1539 for (unsigned int i = 0; i < nodeCount; ++i) {
1540 LValue *lval = reinterpret_cast<LValue *>(func->allLValues.get(i));
1541 if (lval) {
1542 nodes[i].init(regs, lval);
1543 RIG.insert(&nodes[i]);
1544
1545 if (lval->inFile(FILE_GPR) && lval->getInsn() != NULL) {
1546 Instruction *insn = lval->getInsn();
1547 if (insn->op != OP_MAD && insn->op != OP_FMA && insn->op != OP_SAD)
1548 continue;
1549 // For both of the cases below, we only want to add the preference
1550 // if all arguments are in registers.
1551 if (insn->src(0).getFile() != FILE_GPR ||
1552 insn->src(1).getFile() != FILE_GPR ||
1553 insn->src(2).getFile() != FILE_GPR)
1554 continue;
1555 if (prog->getTarget()->getChipset() < 0xc0) {
1556 // Outputting a flag is not supported with short encodings nor
1557 // with immediate arguments.
1558 // See handleMADforNV50.
1559 if (insn->flagsDef >= 0)
1560 continue;
1561 } else {
1562 // We can only fold immediate arguments if dst == src2. This
1563 // only matters if one of the first two arguments is an
1564 // immediate. This form is also only supported for floats.
1565 // See handleMADforNVC0.
1566 ImmediateValue imm;
1567 if (insn->dType != TYPE_F32)
1568 continue;
1569 if (!insn->src(0).getImmediate(imm) &&
1570 !insn->src(1).getImmediate(imm))
1571 continue;
1572 }
1573
1574 nodes[i].addRegPreference(getNode(insn->getSrc(2)->asLValue()));
1575 }
1576 }
1577 }
1578
1579 // coalesce first, we use only 1 RIG node for a group of joined values
1580 ret = coalesce(insns);
1581 if (!ret)
1582 goto out;
1583
1584 if (func->getProgram()->dbgFlags & NV50_IR_DEBUG_REG_ALLOC)
1585 func->printLiveIntervals();
1586
1587 buildRIG(insns);
1588 calculateSpillWeights();
1589 ret = simplify();
1590 if (!ret)
1591 goto out;
1592
1593 ret = selectRegisters();
1594 if (!ret) {
1595 INFO_DBG(prog->dbgFlags, REG_ALLOC,
1596 "selectRegisters failed, inserting spill code ...\n");
1597 regs.reset(FILE_GPR, true);
1598 spill.run(mustSpill);
1599 if (prog->dbgFlags & NV50_IR_DEBUG_REG_ALLOC)
1600 func->print();
1601 } else {
1602 mergedDefs.merge();
1603 prog->maxGPR = std::max(prog->maxGPR, regs.getMaxAssigned(FILE_GPR));
1604 }
1605
1606 out:
1607 cleanup(ret);
1608 return ret;
1609 }
1610
1611 void
cleanup(const bool success)1612 GCRA::cleanup(const bool success)
1613 {
1614 mustSpill.clear();
1615
1616 for (ArrayList::Iterator it = func->allLValues.iterator();
1617 !it.end(); it.next()) {
1618 LValue *lval = reinterpret_cast<LValue *>(it.get());
1619
1620 lval->livei.clear();
1621
1622 lval->compound = 0;
1623 lval->compMask = 0;
1624
1625 if (lval->join == lval)
1626 continue;
1627
1628 if (success)
1629 lval->reg.data.id = lval->join->reg.data.id;
1630 else
1631 lval->join = lval;
1632 }
1633
1634 if (success)
1635 resolveSplitsAndMerges();
1636 splits.clear(); // avoid duplicate entries on next coalesce pass
1637 merges.clear();
1638
1639 delete[] nodes;
1640 nodes = NULL;
1641 hi.next = hi.prev = &hi;
1642 lo[0].next = lo[0].prev = &lo[0];
1643 lo[1].next = lo[1].prev = &lo[1];
1644 }
1645
1646 Symbol *
assignSlot(const Interval &livei, const unsigned int size)1647 SpillCodeInserter::assignSlot(const Interval &livei, const unsigned int size)
1648 {
1649 SpillSlot slot;
1650 int32_t offsetBase = stackSize;
1651 int32_t offset;
1652 std::list<SpillSlot>::iterator pos = slots.end(), it = slots.begin();
1653
1654 if (!func->stackPtr) {
1655 // Later, we compute the address as (offsetBase + tlsBase)
1656 // tlsBase might not be size-aligned, so we add just enough
1657 // to give the final address the correct alignment
1658 offsetBase = align(offsetBase + func->tlsBase, size) - func->tlsBase;
1659 } else {
1660 offsetBase = align(offsetBase, size);
1661 }
1662
1663 slot.sym = NULL;
1664
1665 for (offset = offsetBase; offset < stackSize; offset += size) {
1666 const int32_t entryEnd = offset + size;
1667 while (it != slots.end() && it->offset < offset)
1668 ++it;
1669 if (it == slots.end()) // no slots left
1670 break;
1671 std::list<SpillSlot>::iterator bgn = it;
1672
1673 while (it != slots.end() && it->offset < entryEnd) {
1674 it->occup.print();
1675 if (it->occup.overlaps(livei))
1676 break;
1677 ++it;
1678 }
1679 if (it == slots.end() || it->offset >= entryEnd) {
1680 // fits
1681 for (; bgn != slots.end() && bgn->offset < entryEnd; ++bgn) {
1682 bgn->occup.insert(livei);
1683 if (bgn->size() == size)
1684 slot.sym = bgn->sym;
1685 }
1686 break;
1687 }
1688 }
1689 if (!slot.sym) {
1690 stackSize = offset + size;
1691 slot.offset = offset;
1692 slot.sym = new_Symbol(func->getProgram(), FILE_MEMORY_LOCAL);
1693 if (!func->stackPtr)
1694 offset += func->tlsBase;
1695 slot.sym->setAddress(NULL, offset);
1696 slot.sym->reg.size = size;
1697 slots.insert(pos, slot)->occup.insert(livei);
1698 }
1699 return slot.sym;
1700 }
1701
1702 Value *
offsetSlot(Value *base, const LValue *lval)1703 SpillCodeInserter::offsetSlot(Value *base, const LValue *lval)
1704 {
1705 if (!lval->compound || (lval->compMask & 0x1))
1706 return base;
1707 Value *slot = cloneShallow(func, base);
1708
1709 slot->reg.data.offset += (ffs(lval->compMask) - 1) * lval->reg.size;
1710 slot->reg.size = lval->reg.size;
1711
1712 return slot;
1713 }
1714
1715 void
spill(Instruction *defi, Value *slot, LValue *lval)1716 SpillCodeInserter::spill(Instruction *defi, Value *slot, LValue *lval)
1717 {
1718 const DataType ty = typeOfSize(lval->reg.size);
1719
1720 slot = offsetSlot(slot, lval);
1721
1722 Instruction *st;
1723 if (slot->reg.file == FILE_MEMORY_LOCAL) {
1724 lval->noSpill = 1;
1725 if (ty != TYPE_B96) {
1726 st = new_Instruction(func, OP_STORE, ty);
1727 st->setSrc(0, slot);
1728 st->setSrc(1, lval);
1729 } else {
1730 st = new_Instruction(func, OP_SPLIT, ty);
1731 st->setSrc(0, lval);
1732 for (int d = 0; d < lval->reg.size / 4; ++d)
1733 st->setDef(d, new_LValue(func, FILE_GPR));
1734
1735 for (int d = lval->reg.size / 4 - 1; d >= 0; --d) {
1736 Value *tmp = cloneShallow(func, slot);
1737 tmp->reg.size = 4;
1738 tmp->reg.data.offset += 4 * d;
1739
1740 Instruction *s = new_Instruction(func, OP_STORE, TYPE_U32);
1741 s->setSrc(0, tmp);
1742 s->setSrc(1, st->getDef(d));
1743 defi->bb->insertAfter(defi, s);
1744 }
1745 }
1746 } else {
1747 st = new_Instruction(func, OP_CVT, ty);
1748 st->setDef(0, slot);
1749 st->setSrc(0, lval);
1750 if (lval->reg.file == FILE_FLAGS)
1751 st->flagsSrc = 0;
1752 }
1753 defi->bb->insertAfter(defi, st);
1754 }
1755
1756 LValue *
unspill(Instruction *usei, LValue *lval, Value *slot)1757 SpillCodeInserter::unspill(Instruction *usei, LValue *lval, Value *slot)
1758 {
1759 const DataType ty = typeOfSize(lval->reg.size);
1760
1761 slot = offsetSlot(slot, lval);
1762 lval = cloneShallow(func, lval);
1763
1764 Instruction *ld;
1765 if (slot->reg.file == FILE_MEMORY_LOCAL) {
1766 lval->noSpill = 1;
1767 if (ty != TYPE_B96) {
1768 ld = new_Instruction(func, OP_LOAD, ty);
1769 } else {
1770 ld = new_Instruction(func, OP_MERGE, ty);
1771 for (int d = 0; d < lval->reg.size / 4; ++d) {
1772 Value *tmp = cloneShallow(func, slot);
1773 LValue *val;
1774 tmp->reg.size = 4;
1775 tmp->reg.data.offset += 4 * d;
1776
1777 Instruction *l = new_Instruction(func, OP_LOAD, TYPE_U32);
1778 l->setDef(0, (val = new_LValue(func, FILE_GPR)));
1779 l->setSrc(0, tmp);
1780 usei->bb->insertBefore(usei, l);
1781 ld->setSrc(d, val);
1782 val->noSpill = 1;
1783 }
1784 ld->setDef(0, lval);
1785 usei->bb->insertBefore(usei, ld);
1786 return lval;
1787 }
1788 } else {
1789 ld = new_Instruction(func, OP_CVT, ty);
1790 }
1791 ld->setDef(0, lval);
1792 ld->setSrc(0, slot);
1793 if (lval->reg.file == FILE_FLAGS)
1794 ld->flagsDef = 0;
1795
1796 usei->bb->insertBefore(usei, ld);
1797 return lval;
1798 }
1799
1800 static bool
value_cmp(ValueRef *a, ValueRef *b)1801 value_cmp(ValueRef *a, ValueRef *b) {
1802 Instruction *ai = a->getInsn(), *bi = b->getInsn();
1803 if (ai->bb != bi->bb)
1804 return ai->bb->getId() < bi->bb->getId();
1805 return ai->serial < bi->serial;
1806 }
1807
1808 // For each value that is to be spilled, go through all its definitions.
1809 // A value can have multiple definitions if it has been coalesced before.
1810 // For each definition, first go through all its uses and insert an unspill
1811 // instruction before it, then replace the use with the temporary register.
1812 // Unspill can be either a load from memory or simply a move to another
1813 // register file.
1814 // For "Pseudo" instructions (like PHI, SPLIT, MERGE) we can erase the use
1815 // if we have spilled to a memory location, or simply with the new register.
1816 // No load or conversion instruction should be needed.
1817 bool
run(const std::list<ValuePair>& lst)1818 SpillCodeInserter::run(const std::list<ValuePair>& lst)
1819 {
1820 for (std::list<ValuePair>::const_iterator it = lst.begin(); it != lst.end();
1821 ++it) {
1822 LValue *lval = it->first->asLValue();
1823 Symbol *mem = it->second ? it->second->asSym() : NULL;
1824
1825 // Keep track of which instructions to delete later. Deleting them
1826 // inside the loop is unsafe since a single instruction may have
1827 // multiple destinations that all need to be spilled (like OP_SPLIT).
1828 std::unordered_set<Instruction *> to_del;
1829
1830 std::list<ValueDef *> &defs = mergedDefs(lval);
1831 for (Value::DefIterator d = defs.begin(); d != defs.end();
1832 ++d) {
1833 Value *slot = mem ?
1834 static_cast<Value *>(mem) : new_LValue(func, FILE_GPR);
1835 Value *tmp = NULL;
1836 Instruction *last = NULL;
1837
1838 LValue *dval = (*d)->get()->asLValue();
1839 Instruction *defi = (*d)->getInsn();
1840
1841 // Sort all the uses by BB/instruction so that we don't unspill
1842 // multiple times in a row, and also remove a source of
1843 // non-determinism.
1844 std::vector<ValueRef *> refs(dval->uses.begin(), dval->uses.end());
1845 std::sort(refs.begin(), refs.end(), value_cmp);
1846
1847 // Unspill at each use *before* inserting spill instructions,
1848 // we don't want to have the spill instructions in the use list here.
1849 for (std::vector<ValueRef*>::const_iterator it = refs.begin();
1850 it != refs.end(); ++it) {
1851 ValueRef *u = *it;
1852 Instruction *usei = u->getInsn();
1853 assert(usei);
1854 if (usei->isPseudo()) {
1855 tmp = (slot->reg.file == FILE_MEMORY_LOCAL) ? NULL : slot;
1856 last = NULL;
1857 } else {
1858 if (!last || (usei != last->next && usei != last))
1859 tmp = unspill(usei, dval, slot);
1860 last = usei;
1861 }
1862 u->set(tmp);
1863 }
1864
1865 assert(defi);
1866 if (defi->isPseudo()) {
1867 d = defs.erase(d);
1868 --d;
1869 if (slot->reg.file == FILE_MEMORY_LOCAL)
1870 to_del.insert(defi);
1871 else
1872 defi->setDef(0, slot);
1873 } else {
1874 spill(defi, slot, dval);
1875 }
1876 }
1877
1878 for (std::unordered_set<Instruction *>::const_iterator it = to_del.begin();
1879 it != to_del.end(); ++it) {
1880 mergedDefs.removeDefsOfInstruction(*it);
1881 delete_Instruction(func->getProgram(), *it);
1882 }
1883 }
1884
1885 // TODO: We're not trying to reuse old slots in a potential next iteration.
1886 // We have to update the slots' livei intervals to be able to do that.
1887 stackBase = stackSize;
1888 slots.clear();
1889 return true;
1890 }
1891
1892 bool
exec()1893 RegAlloc::exec()
1894 {
1895 for (IteratorRef it = prog->calls.iteratorDFS(false);
1896 !it->end(); it->next()) {
1897 func = Function::get(reinterpret_cast<Graph::Node *>(it->get()));
1898
1899 func->tlsBase = prog->tlsSize;
1900 if (!execFunc())
1901 return false;
1902 prog->tlsSize += func->tlsSize;
1903 }
1904 return true;
1905 }
1906
1907 bool
execFunc()1908 RegAlloc::execFunc()
1909 {
1910 MergedDefs mergedDefs;
1911 InsertConstraintsPass insertConstr;
1912 PhiMovesPass insertPhiMoves;
1913 ArgumentMovesPass insertArgMoves;
1914 BuildIntervalsPass buildIntervals;
1915 SpillCodeInserter insertSpills(func, mergedDefs);
1916
1917 GCRA gcra(func, insertSpills, mergedDefs);
1918
1919 unsigned int i, retries;
1920 bool ret;
1921
1922 if (!func->ins.empty()) {
1923 // Insert a nop at the entry so inputs only used by the first instruction
1924 // don't count as having an empty live range.
1925 Instruction *nop = new_Instruction(func, OP_NOP, TYPE_NONE);
1926 BasicBlock::get(func->cfg.getRoot())->insertHead(nop);
1927 }
1928
1929 ret = insertConstr.exec(func);
1930 if (!ret)
1931 goto out;
1932
1933 ret = insertPhiMoves.run(func);
1934 if (!ret)
1935 goto out;
1936
1937 ret = insertArgMoves.run(func);
1938 if (!ret)
1939 goto out;
1940
1941 // TODO: need to fix up spill slot usage ranges to support > 1 retry
1942 for (retries = 0; retries < 3; ++retries) {
1943 if (retries && (prog->dbgFlags & NV50_IR_DEBUG_REG_ALLOC))
1944 INFO("Retry: %i\n", retries);
1945 if (prog->dbgFlags & NV50_IR_DEBUG_REG_ALLOC)
1946 func->print();
1947
1948 // spilling to registers may add live ranges, need to rebuild everything
1949 ret = true;
1950 for (sequence = func->cfg.nextSequence(), i = 0;
1951 ret && i <= func->loopNestingBound;
1952 sequence = func->cfg.nextSequence(), ++i)
1953 ret = buildLiveSets(BasicBlock::get(func->cfg.getRoot()));
1954 // reset marker
1955 for (ArrayList::Iterator bi = func->allBBlocks.iterator();
1956 !bi.end(); bi.next())
1957 BasicBlock::get(bi)->liveSet.marker = false;
1958 if (!ret)
1959 break;
1960 func->orderInstructions(this->insns);
1961
1962 ret = buildIntervals.run(func);
1963 if (!ret)
1964 break;
1965 ret = gcra.allocateRegisters(insns);
1966 if (ret)
1967 break; // success
1968 }
1969 INFO_DBG(prog->dbgFlags, REG_ALLOC, "RegAlloc done: %i\n", ret);
1970
1971 func->tlsSize = insertSpills.getStackSize();
1972 out:
1973 return ret;
1974 }
1975
1976 // TODO: check if modifying Instruction::join here breaks anything
1977 void
resolveSplitsAndMerges()1978 GCRA::resolveSplitsAndMerges()
1979 {
1980 for (std::list<Instruction *>::iterator it = splits.begin();
1981 it != splits.end();
1982 ++it) {
1983 Instruction *split = *it;
1984 unsigned int reg = regs.idToBytes(split->getSrc(0));
1985 for (int d = 0; split->defExists(d); ++d) {
1986 Value *v = split->getDef(d);
1987 v->reg.data.id = regs.bytesToId(v, reg);
1988 v->join = v;
1989 reg += v->reg.size;
1990 }
1991 }
1992 splits.clear();
1993
1994 for (std::list<Instruction *>::iterator it = merges.begin();
1995 it != merges.end();
1996 ++it) {
1997 Instruction *merge = *it;
1998 unsigned int reg = regs.idToBytes(merge->getDef(0));
1999 for (int s = 0; merge->srcExists(s); ++s) {
2000 Value *v = merge->getSrc(s);
2001 v->reg.data.id = regs.bytesToId(v, reg);
2002 v->join = v;
2003 // If the value is defined by a phi/union node, we also need to
2004 // perform the same fixup on that node's sources, since after RA
2005 // their registers should be identical.
2006 if (v->getInsn()->op == OP_PHI || v->getInsn()->op == OP_UNION) {
2007 Instruction *phi = v->getInsn();
2008 for (int phis = 0; phi->srcExists(phis); ++phis) {
2009 phi->getSrc(phis)->join = v;
2010 phi->getSrc(phis)->reg.data.id = v->reg.data.id;
2011 }
2012 }
2013 reg += v->reg.size;
2014 }
2015 }
2016 merges.clear();
2017 }
2018
registerAllocation()2019 bool Program::registerAllocation()
2020 {
2021 RegAlloc ra(this);
2022 return ra.exec();
2023 }
2024
2025 bool
exec(Function *ir)2026 RegAlloc::InsertConstraintsPass::exec(Function *ir)
2027 {
2028 constrList.clear();
2029
2030 bool ret = run(ir, true, true);
2031 if (ret)
2032 ret = insertConstraintMoves();
2033 return ret;
2034 }
2035
2036 // TODO: make part of texture insn
2037 void
textureMask(TexInstruction *tex)2038 RegAlloc::InsertConstraintsPass::textureMask(TexInstruction *tex)
2039 {
2040 Value *def[4];
2041 int c, k, d;
2042 uint8_t mask = 0;
2043
2044 for (d = 0, k = 0, c = 0; c < 4; ++c) {
2045 if (!(tex->tex.mask & (1 << c)))
2046 continue;
2047 if (tex->getDef(k)->refCount()) {
2048 mask |= 1 << c;
2049 def[d++] = tex->getDef(k);
2050 }
2051 ++k;
2052 }
2053 tex->tex.mask = mask;
2054
2055 for (c = 0; c < d; ++c)
2056 tex->setDef(c, def[c]);
2057 for (; c < 4; ++c)
2058 tex->setDef(c, NULL);
2059 }
2060
2061 bool
detectConflict(Instruction *cst, int s)2062 RegAlloc::InsertConstraintsPass::detectConflict(Instruction *cst, int s)
2063 {
2064 Value *v = cst->getSrc(s);
2065
2066 // current register allocation can't handle it if a value participates in
2067 // multiple constraints
2068 for (Value::UseIterator it = v->uses.begin(); it != v->uses.end(); ++it) {
2069 if (cst != (*it)->getInsn())
2070 return true;
2071 }
2072
2073 // can start at s + 1 because detectConflict is called on all sources
2074 for (int c = s + 1; cst->srcExists(c); ++c)
2075 if (v == cst->getSrc(c))
2076 return true;
2077
2078 Instruction *defi = v->getInsn();
2079
2080 return (!defi || defi->constrainedDefs());
2081 }
2082
2083 void
addConstraint(Instruction *i, int s, int n)2084 RegAlloc::InsertConstraintsPass::addConstraint(Instruction *i, int s, int n)
2085 {
2086 Instruction *cst;
2087 int d;
2088
2089 // first, look for an existing identical constraint op
2090 for (std::list<Instruction *>::iterator it = constrList.begin();
2091 it != constrList.end();
2092 ++it) {
2093 cst = (*it);
2094 if (!i->bb->dominatedBy(cst->bb))
2095 break;
2096 for (d = 0; d < n; ++d)
2097 if (cst->getSrc(d) != i->getSrc(d + s))
2098 break;
2099 if (d >= n) {
2100 for (d = 0; d < n; ++d, ++s)
2101 i->setSrc(s, cst->getDef(d));
2102 return;
2103 }
2104 }
2105 cst = new_Instruction(func, OP_CONSTRAINT, i->dType);
2106
2107 for (d = 0; d < n; ++s, ++d) {
2108 cst->setDef(d, new_LValue(func, FILE_GPR));
2109 cst->setSrc(d, i->getSrc(s));
2110 i->setSrc(s, cst->getDef(d));
2111 }
2112 i->bb->insertBefore(i, cst);
2113
2114 constrList.push_back(cst);
2115 }
2116
2117 // Add a dummy use of the pointer source of >= 8 byte loads after the load
2118 // to prevent it from being assigned a register which overlapping the load's
2119 // destination, which would produce random corruptions.
2120 void
addHazard(Instruction *i, const ValueRef *src)2121 RegAlloc::InsertConstraintsPass::addHazard(Instruction *i, const ValueRef *src)
2122 {
2123 Instruction *hzd = new_Instruction(func, OP_NOP, TYPE_NONE);
2124 hzd->setSrc(0, src->get());
2125 i->bb->insertAfter(i, hzd);
2126
2127 }
2128
2129 // b32 { %r0 %r1 %r2 %r3 } -> b128 %r0q
2130 void
condenseDefs(Instruction *insn)2131 RegAlloc::InsertConstraintsPass::condenseDefs(Instruction *insn)
2132 {
2133 int n;
2134 for (n = 0; insn->defExists(n) && insn->def(n).getFile() == FILE_GPR; ++n);
2135 condenseDefs(insn, 0, n - 1);
2136 }
2137
2138 void
condenseDefs(Instruction *insn, const int a, const int b)2139 RegAlloc::InsertConstraintsPass::condenseDefs(Instruction *insn,
2140 const int a, const int b)
2141 {
2142 uint8_t size = 0;
2143 if (a >= b)
2144 return;
2145 for (int s = a; s <= b; ++s)
2146 size += insn->getDef(s)->reg.size;
2147 if (!size)
2148 return;
2149
2150 LValue *lval = new_LValue(func, FILE_GPR);
2151 lval->reg.size = size;
2152
2153 Instruction *split = new_Instruction(func, OP_SPLIT, typeOfSize(size));
2154 split->setSrc(0, lval);
2155 for (int d = a; d <= b; ++d) {
2156 split->setDef(d - a, insn->getDef(d));
2157 insn->setDef(d, NULL);
2158 }
2159 insn->setDef(a, lval);
2160
2161 for (int k = a + 1, d = b + 1; insn->defExists(d); ++d, ++k) {
2162 insn->setDef(k, insn->getDef(d));
2163 insn->setDef(d, NULL);
2164 }
2165 // carry over predicate if any (mainly for OP_UNION uses)
2166 split->setPredicate(insn->cc, insn->getPredicate());
2167
2168 insn->bb->insertAfter(insn, split);
2169 constrList.push_back(split);
2170 }
2171
2172 void
condenseSrcs(Instruction *insn, const int a, const int b)2173 RegAlloc::InsertConstraintsPass::condenseSrcs(Instruction *insn,
2174 const int a, const int b)
2175 {
2176 uint8_t size = 0;
2177 if (a >= b)
2178 return;
2179 for (int s = a; s <= b; ++s)
2180 size += insn->getSrc(s)->reg.size;
2181 if (!size)
2182 return;
2183 LValue *lval = new_LValue(func, FILE_GPR);
2184 lval->reg.size = size;
2185
2186 Value *save[3];
2187 insn->takeExtraSources(0, save);
2188
2189 Instruction *merge = new_Instruction(func, OP_MERGE, typeOfSize(size));
2190 merge->setDef(0, lval);
2191 for (int s = a, i = 0; s <= b; ++s, ++i) {
2192 merge->setSrc(i, insn->getSrc(s));
2193 }
2194 insn->moveSources(b + 1, a - b);
2195 insn->setSrc(a, lval);
2196 insn->bb->insertBefore(insn, merge);
2197
2198 insn->putExtraSources(0, save);
2199
2200 constrList.push_back(merge);
2201 }
2202
2203 bool
isScalarTexGM107(TexInstruction *tex)2204 RegAlloc::InsertConstraintsPass::isScalarTexGM107(TexInstruction *tex)
2205 {
2206 if (tex->tex.sIndirectSrc >= 0 ||
2207 tex->tex.rIndirectSrc >= 0 ||
2208 tex->tex.derivAll)
2209 return false;
2210
2211 if (tex->tex.mask == 5 || tex->tex.mask == 6)
2212 return false;
2213
2214 switch (tex->op) {
2215 case OP_TEX:
2216 case OP_TXF:
2217 case OP_TXG:
2218 case OP_TXL:
2219 break;
2220 default:
2221 return false;
2222 }
2223
2224 // legal variants:
2225 // TEXS.1D.LZ
2226 // TEXS.2D
2227 // TEXS.2D.LZ
2228 // TEXS.2D.LL
2229 // TEXS.2D.DC
2230 // TEXS.2D.LL.DC
2231 // TEXS.2D.LZ.DC
2232 // TEXS.A2D
2233 // TEXS.A2D.LZ
2234 // TEXS.A2D.LZ.DC
2235 // TEXS.3D
2236 // TEXS.3D.LZ
2237 // TEXS.CUBE
2238 // TEXS.CUBE.LL
2239
2240 // TLDS.1D.LZ
2241 // TLDS.1D.LL
2242 // TLDS.2D.LZ
2243 // TLSD.2D.LZ.AOFFI
2244 // TLDS.2D.LZ.MZ
2245 // TLDS.2D.LL
2246 // TLDS.2D.LL.AOFFI
2247 // TLDS.A2D.LZ
2248 // TLDS.3D.LZ
2249
2250 // TLD4S: all 2D/RECT variants and only offset
2251
2252 switch (tex->op) {
2253 case OP_TEX:
2254 if (tex->tex.useOffsets)
2255 return false;
2256
2257 switch (tex->tex.target.getEnum()) {
2258 case TEX_TARGET_1D:
2259 case TEX_TARGET_2D_ARRAY_SHADOW:
2260 return tex->tex.levelZero;
2261 case TEX_TARGET_CUBE:
2262 return !tex->tex.levelZero;
2263 case TEX_TARGET_2D:
2264 case TEX_TARGET_2D_ARRAY:
2265 case TEX_TARGET_2D_SHADOW:
2266 case TEX_TARGET_3D:
2267 case TEX_TARGET_RECT:
2268 case TEX_TARGET_RECT_SHADOW:
2269 return true;
2270 default:
2271 return false;
2272 }
2273
2274 case OP_TXL:
2275 if (tex->tex.useOffsets)
2276 return false;
2277
2278 switch (tex->tex.target.getEnum()) {
2279 case TEX_TARGET_2D:
2280 case TEX_TARGET_2D_SHADOW:
2281 case TEX_TARGET_RECT:
2282 case TEX_TARGET_RECT_SHADOW:
2283 case TEX_TARGET_CUBE:
2284 return true;
2285 default:
2286 return false;
2287 }
2288
2289 case OP_TXF:
2290 switch (tex->tex.target.getEnum()) {
2291 case TEX_TARGET_1D:
2292 return !tex->tex.useOffsets;
2293 case TEX_TARGET_2D:
2294 case TEX_TARGET_RECT:
2295 return true;
2296 case TEX_TARGET_2D_ARRAY:
2297 case TEX_TARGET_2D_MS:
2298 case TEX_TARGET_3D:
2299 return !tex->tex.useOffsets && tex->tex.levelZero;
2300 default:
2301 return false;
2302 }
2303
2304 case OP_TXG:
2305 if (tex->tex.useOffsets > 1)
2306 return false;
2307 if (tex->tex.mask != 0x3 && tex->tex.mask != 0xf)
2308 return false;
2309
2310 switch (tex->tex.target.getEnum()) {
2311 case TEX_TARGET_2D:
2312 case TEX_TARGET_2D_MS:
2313 case TEX_TARGET_2D_SHADOW:
2314 case TEX_TARGET_RECT:
2315 case TEX_TARGET_RECT_SHADOW:
2316 return true;
2317 default:
2318 return false;
2319 }
2320
2321 default:
2322 return false;
2323 }
2324 }
2325
2326 void
handleScalarTexGM107(TexInstruction *tex)2327 RegAlloc::InsertConstraintsPass::handleScalarTexGM107(TexInstruction *tex)
2328 {
2329 int defCount = tex->defCount(0xff);
2330 int srcCount = tex->srcCount(0xff);
2331
2332 tex->tex.scalar = true;
2333
2334 // 1. handle defs
2335 if (defCount > 3)
2336 condenseDefs(tex, 2, 3);
2337 if (defCount > 1)
2338 condenseDefs(tex, 0, 1);
2339
2340 // 2. handle srcs
2341 // special case for TXF.A2D
2342 if (tex->op == OP_TXF && tex->tex.target == TEX_TARGET_2D_ARRAY) {
2343 assert(srcCount >= 3);
2344 condenseSrcs(tex, 1, 2);
2345 } else {
2346 if (srcCount > 3)
2347 condenseSrcs(tex, 2, 3);
2348 // only if we have more than 2 sources
2349 if (srcCount > 2)
2350 condenseSrcs(tex, 0, 1);
2351 }
2352
2353 assert(!tex->defExists(2) && !tex->srcExists(2));
2354 }
2355
2356 void
texConstraintGM107(TexInstruction *tex)2357 RegAlloc::InsertConstraintsPass::texConstraintGM107(TexInstruction *tex)
2358 {
2359 int n, s;
2360
2361 if (isTextureOp(tex->op))
2362 textureMask(tex);
2363
2364 if (targ->getChipset() < NVISA_GV100_CHIPSET) {
2365 if (isScalarTexGM107(tex)) {
2366 handleScalarTexGM107(tex);
2367 return;
2368 }
2369
2370 assert(!tex->tex.scalar);
2371 condenseDefs(tex);
2372 } else {
2373 if (isTextureOp(tex->op)) {
2374 int defCount = tex->defCount(0xff);
2375 if (defCount > 3)
2376 condenseDefs(tex, 2, 3);
2377 if (defCount > 1)
2378 condenseDefs(tex, 0, 1);
2379 } else {
2380 condenseDefs(tex);
2381 }
2382 }
2383
2384 if (isSurfaceOp(tex->op)) {
2385 int s = tex->tex.target.getDim() +
2386 (tex->tex.target.isArray() || tex->tex.target.isCube());
2387 int n = 0;
2388
2389 switch (tex->op) {
2390 case OP_SUSTB:
2391 case OP_SUSTP:
2392 n = 4;
2393 break;
2394 case OP_SUREDB:
2395 case OP_SUREDP:
2396 if (tex->subOp == NV50_IR_SUBOP_ATOM_CAS)
2397 n = 2;
2398 break;
2399 default:
2400 break;
2401 }
2402
2403 if (s > 1)
2404 condenseSrcs(tex, 0, s - 1);
2405 if (n > 1)
2406 condenseSrcs(tex, 1, n); // do not condense the tex handle
2407 } else
2408 if (isTextureOp(tex->op)) {
2409 if (tex->op != OP_TXQ) {
2410 s = tex->tex.target.getArgCount() - tex->tex.target.isMS();
2411 if (tex->op == OP_TXD) {
2412 // Indirect handle belongs in the first arg
2413 if (tex->tex.rIndirectSrc >= 0)
2414 s++;
2415 if (!tex->tex.target.isArray() && tex->tex.useOffsets)
2416 s++;
2417 }
2418 n = tex->srcCount(0xff, true) - s;
2419 // TODO: Is this necessary? Perhaps just has to be aligned to the
2420 // level that the first arg is, not necessarily to 4. This
2421 // requirement has not been rigorously verified, as it has been on
2422 // Kepler.
2423 if (n > 0 && n < 3) {
2424 if (tex->srcExists(n + s)) // move potential predicate out of the way
2425 tex->moveSources(n + s, 3 - n);
2426 while (n < 3)
2427 tex->setSrc(s + n++, new_LValue(func, FILE_GPR));
2428 }
2429 } else {
2430 s = tex->srcCount(0xff, true);
2431 n = 0;
2432 }
2433
2434 if (s > 1)
2435 condenseSrcs(tex, 0, s - 1);
2436 if (n > 1) // NOTE: first call modified positions already
2437 condenseSrcs(tex, 1, n);
2438 }
2439 }
2440
2441 void
texConstraintNVE0(TexInstruction *tex)2442 RegAlloc::InsertConstraintsPass::texConstraintNVE0(TexInstruction *tex)
2443 {
2444 if (isTextureOp(tex->op))
2445 textureMask(tex);
2446 condenseDefs(tex);
2447
2448 if (tex->op == OP_SUSTB || tex->op == OP_SUSTP) {
2449 condenseSrcs(tex, 3, 6);
2450 } else
2451 if (isTextureOp(tex->op)) {
2452 int n = tex->srcCount(0xff, true);
2453 int s = n > 4 ? 4 : n;
2454 if (n > 4 && n < 7) {
2455 if (tex->srcExists(n)) // move potential predicate out of the way
2456 tex->moveSources(n, 7 - n);
2457
2458 while (n < 7)
2459 tex->setSrc(n++, new_LValue(func, FILE_GPR));
2460 }
2461 if (s > 1)
2462 condenseSrcs(tex, 0, s - 1);
2463 if (n > 4)
2464 condenseSrcs(tex, 1, n - s);
2465 }
2466 }
2467
2468 void
texConstraintNVC0(TexInstruction *tex)2469 RegAlloc::InsertConstraintsPass::texConstraintNVC0(TexInstruction *tex)
2470 {
2471 int n, s;
2472
2473 if (isTextureOp(tex->op))
2474 textureMask(tex);
2475
2476 if (tex->op == OP_TXQ) {
2477 s = tex->srcCount(0xff);
2478 n = 0;
2479 } else if (isSurfaceOp(tex->op)) {
2480 s = tex->tex.target.getDim() + (tex->tex.target.isArray() || tex->tex.target.isCube());
2481 if (tex->op == OP_SUSTB || tex->op == OP_SUSTP)
2482 n = 4;
2483 else
2484 n = 0;
2485 } else {
2486 s = tex->tex.target.getArgCount() - tex->tex.target.isMS();
2487 if (!tex->tex.target.isArray() &&
2488 (tex->tex.rIndirectSrc >= 0 || tex->tex.sIndirectSrc >= 0))
2489 ++s;
2490 if (tex->op == OP_TXD && tex->tex.useOffsets)
2491 ++s;
2492 n = tex->srcCount(0xff) - s;
2493 assert(n <= 4);
2494 }
2495
2496 if (s > 1)
2497 condenseSrcs(tex, 0, s - 1);
2498 if (n > 1) // NOTE: first call modified positions already
2499 condenseSrcs(tex, 1, n);
2500
2501 condenseDefs(tex);
2502 }
2503
2504 void
texConstraintNV50(TexInstruction *tex)2505 RegAlloc::InsertConstraintsPass::texConstraintNV50(TexInstruction *tex)
2506 {
2507 Value *pred = tex->getPredicate();
2508 if (pred)
2509 tex->setPredicate(tex->cc, NULL);
2510
2511 textureMask(tex);
2512
2513 assert(tex->defExists(0) && tex->srcExists(0));
2514 // make src and def count match
2515 int c;
2516 for (c = 0; tex->srcExists(c) || tex->defExists(c); ++c) {
2517 if (!tex->srcExists(c))
2518 tex->setSrc(c, new_LValue(func, tex->getSrc(0)->asLValue()));
2519 else
2520 insertConstraintMove(tex, c);
2521 if (!tex->defExists(c))
2522 tex->setDef(c, new_LValue(func, tex->getDef(0)->asLValue()));
2523 }
2524 if (pred)
2525 tex->setPredicate(tex->cc, pred);
2526 condenseDefs(tex);
2527 condenseSrcs(tex, 0, c - 1);
2528 }
2529
2530 // Insert constraint markers for instructions whose multiple sources must be
2531 // located in consecutive registers.
2532 bool
visit(BasicBlock *bb)2533 RegAlloc::InsertConstraintsPass::visit(BasicBlock *bb)
2534 {
2535 TexInstruction *tex;
2536 Instruction *next;
2537 int s, size;
2538
2539 targ = bb->getProgram()->getTarget();
2540
2541 for (Instruction *i = bb->getEntry(); i; i = next) {
2542 next = i->next;
2543
2544 if ((tex = i->asTex())) {
2545 switch (targ->getChipset() & ~0xf) {
2546 case 0x50:
2547 case 0x80:
2548 case 0x90:
2549 case 0xa0:
2550 texConstraintNV50(tex);
2551 break;
2552 case 0xc0:
2553 case 0xd0:
2554 texConstraintNVC0(tex);
2555 break;
2556 case 0xe0:
2557 case 0xf0:
2558 case 0x100:
2559 texConstraintNVE0(tex);
2560 break;
2561 case 0x110:
2562 case 0x120:
2563 case 0x130:
2564 case 0x140:
2565 case 0x160:
2566 case 0x170:
2567 texConstraintGM107(tex);
2568 break;
2569 default:
2570 break;
2571 }
2572 } else
2573 if (i->op == OP_EXPORT || i->op == OP_STORE) {
2574 for (size = typeSizeof(i->dType), s = 1; size > 0; ++s) {
2575 assert(i->srcExists(s));
2576 size -= i->getSrc(s)->reg.size;
2577 }
2578 condenseSrcs(i, 1, s - 1);
2579 } else
2580 if (i->op == OP_LOAD || i->op == OP_VFETCH) {
2581 condenseDefs(i);
2582 if (i->src(0).isIndirect(0) && typeSizeof(i->dType) >= 8)
2583 addHazard(i, i->src(0).getIndirect(0));
2584 if (i->src(0).isIndirect(1) && typeSizeof(i->dType) >= 8)
2585 addHazard(i, i->src(0).getIndirect(1));
2586 if (i->op == OP_LOAD && i->fixed && targ->getChipset() < 0xc0) {
2587 // Add a hazard to make sure we keep the op around. These are used
2588 // for membars.
2589 Instruction *nop = new_Instruction(func, OP_NOP, i->dType);
2590 nop->setSrc(0, i->getDef(0));
2591 i->bb->insertAfter(i, nop);
2592 }
2593 } else
2594 if (i->op == OP_UNION ||
2595 i->op == OP_MERGE ||
2596 i->op == OP_SPLIT) {
2597 constrList.push_back(i);
2598 } else
2599 if (i->op == OP_ATOM && i->subOp == NV50_IR_SUBOP_ATOM_CAS &&
2600 targ->getChipset() < 0xc0) {
2601 // Like a hazard, but for a def.
2602 Instruction *nop = new_Instruction(func, OP_NOP, i->dType);
2603 nop->setSrc(0, i->getDef(0));
2604 i->bb->insertAfter(i, nop);
2605 }
2606 }
2607 return true;
2608 }
2609
2610 void
insertConstraintMove(Instruction *cst, int s)2611 RegAlloc::InsertConstraintsPass::insertConstraintMove(Instruction *cst, int s)
2612 {
2613 const uint8_t size = cst->src(s).getSize();
2614
2615 assert(cst->getSrc(s)->defs.size() == 1); // still SSA
2616
2617 Instruction *defi = cst->getSrc(s)->defs.front()->getInsn();
2618
2619 bool imm = defi->op == OP_MOV &&
2620 defi->src(0).getFile() == FILE_IMMEDIATE;
2621 bool load = defi->op == OP_LOAD &&
2622 defi->src(0).getFile() == FILE_MEMORY_CONST &&
2623 !defi->src(0).isIndirect(0);
2624 // catch some cases where don't really need MOVs
2625 if (cst->getSrc(s)->refCount() == 1 && !defi->constrainedDefs()
2626 && defi->op != OP_MERGE && defi->op != OP_SPLIT) {
2627 if (imm || load) {
2628 // Move the defi right before the cst. No point in expanding
2629 // the range.
2630 defi->bb->remove(defi);
2631 cst->bb->insertBefore(cst, defi);
2632 }
2633 return;
2634 }
2635
2636 LValue *lval = new_LValue(func, cst->src(s).getFile());
2637 lval->reg.size = size;
2638
2639 Instruction *mov = new_Instruction(func, OP_MOV, typeOfSize(size));
2640 mov->setDef(0, lval);
2641 mov->setSrc(0, cst->getSrc(s));
2642
2643 if (load) {
2644 mov->op = OP_LOAD;
2645 mov->setSrc(0, defi->getSrc(0));
2646 } else if (imm) {
2647 mov->setSrc(0, defi->getSrc(0));
2648 }
2649
2650 if (defi->getPredicate())
2651 mov->setPredicate(defi->cc, defi->getPredicate());
2652
2653 cst->setSrc(s, mov->getDef(0));
2654 cst->bb->insertBefore(cst, mov);
2655
2656 cst->getDef(0)->asLValue()->noSpill = 1; // doesn't help
2657 }
2658
2659 // Insert extra moves so that, if multiple register constraints on a value are
2660 // in conflict, these conflicts can be resolved.
2661 bool
insertConstraintMoves()2662 RegAlloc::InsertConstraintsPass::insertConstraintMoves()
2663 {
2664 for (std::list<Instruction *>::iterator it = constrList.begin();
2665 it != constrList.end();
2666 ++it) {
2667 Instruction *cst = *it;
2668 Instruction *mov;
2669
2670 if (cst->op == OP_SPLIT && false) {
2671 // spilling splits is annoying, just make sure they're separate
2672 for (int d = 0; cst->defExists(d); ++d) {
2673 if (!cst->getDef(d)->refCount())
2674 continue;
2675 LValue *lval = new_LValue(func, cst->def(d).getFile());
2676 const uint8_t size = cst->def(d).getSize();
2677 lval->reg.size = size;
2678
2679 mov = new_Instruction(func, OP_MOV, typeOfSize(size));
2680 mov->setSrc(0, lval);
2681 mov->setDef(0, cst->getDef(d));
2682 cst->setDef(d, mov->getSrc(0));
2683 cst->bb->insertAfter(cst, mov);
2684
2685 cst->getSrc(0)->asLValue()->noSpill = 1;
2686 mov->getSrc(0)->asLValue()->noSpill = 1;
2687 }
2688 } else
2689 if (cst->op == OP_MERGE || cst->op == OP_UNION) {
2690 for (int s = 0; cst->srcExists(s); ++s) {
2691 const uint8_t size = cst->src(s).getSize();
2692
2693 if (!cst->getSrc(s)->defs.size()) {
2694 mov = new_Instruction(func, OP_NOP, typeOfSize(size));
2695 mov->setDef(0, cst->getSrc(s));
2696 cst->bb->insertBefore(cst, mov);
2697 continue;
2698 }
2699
2700 insertConstraintMove(cst, s);
2701 }
2702 }
2703 }
2704
2705 return true;
2706 }
2707
2708 } // namespace nv50_ir
2709