summaryrefslogtreecommitdiffstats
path: root/js/src/jit/BacktrackingAllocator.h
diff options
context:
space:
mode:
Diffstat (limited to 'js/src/jit/BacktrackingAllocator.h')
-rw-r--r--js/src/jit/BacktrackingAllocator.h798
1 files changed, 798 insertions, 0 deletions
diff --git a/js/src/jit/BacktrackingAllocator.h b/js/src/jit/BacktrackingAllocator.h
new file mode 100644
index 0000000000..3c3d701f6c
--- /dev/null
+++ b/js/src/jit/BacktrackingAllocator.h
@@ -0,0 +1,798 @@
+/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*-
+ * vim: set ts=8 sts=2 et sw=2 tw=80:
+ * This Source Code Form is subject to the terms of the Mozilla Public
+ * License, v. 2.0. If a copy of the MPL was not distributed with this
+ * file, You can obtain one at http://mozilla.org/MPL/2.0/. */
+
+#ifndef jit_BacktrackingAllocator_h
+#define jit_BacktrackingAllocator_h
+
+#include "mozilla/Array.h"
+#include "mozilla/Atomics.h"
+#include "mozilla/Attributes.h"
+
+#include "ds/AvlTree.h"
+#include "ds/PriorityQueue.h"
+#include "jit/RegisterAllocator.h"
+#include "jit/StackSlotAllocator.h"
+
+// Gives better traces in Nightly/debug builds (could be EARLY_BETA_OR_EARLIER)
+#if defined(NIGHTLY_BUILD) || defined(DEBUG)
+# define AVOID_INLINE_FOR_DEBUGGING MOZ_NEVER_INLINE
+#else
+# define AVOID_INLINE_FOR_DEBUGGING
+#endif
+
+// Backtracking priority queue based register allocator based on that described
+// in the following blog post:
+//
+// http://blog.llvm.org/2011/09/greedy-register-allocation-in-llvm-30.html
+
+namespace js {
+namespace jit {
+
+class Requirement {
+ public:
+ enum Kind { NONE, REGISTER, FIXED };
+
+ Requirement() : kind_(NONE) {}
+
+ explicit Requirement(Kind kind) : kind_(kind) {
+ // FIXED has a dedicated constructor.
+ MOZ_ASSERT(kind != FIXED);
+ }
+
+ explicit Requirement(LAllocation fixed) : kind_(FIXED), allocation_(fixed) {
+ MOZ_ASSERT(!fixed.isBogus() && !fixed.isUse());
+ }
+
+ Kind kind() const { return kind_; }
+
+ LAllocation allocation() const {
+ MOZ_ASSERT(!allocation_.isBogus() && !allocation_.isUse());
+ return allocation_;
+ }
+
+ [[nodiscard]] bool merge(const Requirement& newRequirement) {
+ // Merge newRequirement with any existing requirement, returning false
+ // if the new and old requirements conflict.
+
+ if (newRequirement.kind() == Requirement::FIXED) {
+ if (kind() == Requirement::FIXED) {
+ return newRequirement.allocation() == allocation();
+ }
+ *this = newRequirement;
+ return true;
+ }
+
+ MOZ_ASSERT(newRequirement.kind() == Requirement::REGISTER);
+ if (kind() == Requirement::FIXED) {
+ return allocation().isRegister();
+ }
+
+ *this = newRequirement;
+ return true;
+ }
+
+ private:
+ Kind kind_;
+ LAllocation allocation_;
+};
+
+struct UsePosition : public TempObject,
+ public InlineForwardListNode<UsePosition> {
+ private:
+ // A UsePosition is an LUse* with a CodePosition. UsePosition also has an
+ // optimization that allows access to the associated LUse::Policy without
+ // dereferencing memory: the policy is encoded in the low bits of the LUse*.
+ //
+ // Note however that because LUse* is uintptr_t-aligned, on 32-bit systems
+ // there are only 4 encodable values, for more than 4 use policies; in that
+ // case we allocate the common LUse::ANY, LUse::REGISTER, and LUse::FIXED use
+ // policies to tags, and use tag 0x3 to indicate that dereferencing the LUse
+ // is necessary to get the policy (KEEPALIVE or STACK, in that case).
+ uintptr_t use_;
+ static_assert(LUse::ANY < 0x3,
+ "LUse::ANY can be represented in low tag on 32-bit systems");
+ static_assert(LUse::REGISTER < 0x3,
+ "LUse::REGISTER can be represented in tag on 32-bit systems");
+ static_assert(LUse::FIXED < 0x3,
+ "LUse::FIXED can be represented in tag on 32-bit systems");
+
+ static constexpr uintptr_t PolicyMask = sizeof(uintptr_t) - 1;
+ static constexpr uintptr_t UseMask = ~PolicyMask;
+
+ void setUse(LUse* use) {
+ // RECOVERED_INPUT is used by snapshots and ignored when building the
+ // liveness information. Thus we can safely assume that no such value
+ // would be seen.
+ MOZ_ASSERT(use->policy() != LUse::RECOVERED_INPUT);
+
+ uintptr_t policyBits = use->policy();
+#ifndef JS_64BIT
+ // On a 32-bit machine, LUse::KEEPALIVE and LUse::STACK are accessed by
+ // dereferencing the use pointer.
+ if (policyBits >= PolicyMask) {
+ policyBits = PolicyMask;
+ }
+#endif
+ use_ = uintptr_t(use) | policyBits;
+ MOZ_ASSERT(use->policy() == usePolicy());
+ }
+
+ public:
+ CodePosition pos;
+
+ LUse* use() const { return reinterpret_cast<LUse*>(use_ & UseMask); }
+
+ LUse::Policy usePolicy() const {
+ uintptr_t bits = use_ & PolicyMask;
+#ifndef JS_64BIT
+ // On 32-bit machines, reach out to memory if it's LUse::KEEPALIVE or
+ // LUse::STACK.
+ if (bits == PolicyMask) {
+ return use()->policy();
+ }
+#endif
+ LUse::Policy policy = LUse::Policy(bits);
+ MOZ_ASSERT(use()->policy() == policy);
+ return policy;
+ }
+
+ UsePosition(LUse* use, CodePosition pos) : pos(pos) {
+ // Verify that the usedAtStart() flag is consistent with the
+ // subposition. For now ignore fixed registers, because they
+ // are handled specially around calls.
+ MOZ_ASSERT_IF(!use->isFixedRegister(),
+ pos.subpos() == (use->usedAtStart() ? CodePosition::INPUT
+ : CodePosition::OUTPUT));
+ setUse(use);
+ }
+};
+
+using UsePositionIterator = InlineForwardListIterator<UsePosition>;
+
+// Backtracking allocator data structures overview.
+//
+// LiveRange: A continuous range of positions where a virtual register is live.
+// LiveBundle: A set of LiveRanges which do not overlap.
+// VirtualRegister: A set of all LiveRanges used for some LDefinition.
+//
+// The allocator first performs a liveness ananlysis on the LIR graph which
+// constructs LiveRanges for each VirtualRegister, determining where the
+// registers are live.
+//
+// The ranges are then bundled together according to heuristics, and placed on
+// the allocation queue.
+//
+// As bundles are removed from the allocation queue, we attempt to find a
+// physical register or stack slot allocation for all ranges in the removed
+// bundle, possibly evicting already-allocated bundles. See processBundle()
+// for details.
+//
+// If we are not able to allocate a bundle, it is split according to heuristics
+// into two or more smaller bundles which cover all the ranges of the original.
+// These smaller bundles are then allocated independently.
+
+class LiveBundle;
+class VirtualRegister;
+
+class LiveRange : public TempObject {
+ public:
+ // Linked lists are used to keep track of the ranges in each LiveBundle and
+ // VirtualRegister. Since a LiveRange may be in two lists simultaneously, use
+ // these auxiliary classes to keep things straight.
+ class BundleLink : public InlineForwardListNode<BundleLink> {};
+ class RegisterLink : public InlineForwardListNode<RegisterLink> {};
+
+ using BundleLinkIterator = InlineForwardListIterator<BundleLink>;
+ using RegisterLinkIterator = InlineForwardListIterator<RegisterLink>;
+
+ // Links in the lists in LiveBundle and VirtualRegister.
+ BundleLink bundleLink;
+ RegisterLink registerLink;
+
+ static LiveRange* get(BundleLink* link) {
+ return reinterpret_cast<LiveRange*>(reinterpret_cast<uint8_t*>(link) -
+ offsetof(LiveRange, bundleLink));
+ }
+ static LiveRange* get(RegisterLink* link) {
+ return reinterpret_cast<LiveRange*>(reinterpret_cast<uint8_t*>(link) -
+ offsetof(LiveRange, registerLink));
+ }
+
+ struct Range {
+ // The beginning of this range, inclusive.
+ CodePosition from;
+
+ // The end of this range, exclusive.
+ CodePosition to;
+
+ Range() = default;
+
+ Range(CodePosition from, CodePosition to) : from(from), to(to) {
+ MOZ_ASSERT(!empty());
+ }
+
+ bool empty() {
+ MOZ_ASSERT(from <= to);
+ return from == to;
+ }
+ };
+
+ private:
+ // The virtual register this range is for, or nullptr if this does not have a
+ // virtual register (for example, it is in the callRanges bundle).
+ VirtualRegister* vreg_;
+
+ // The bundle containing this range, null if liveness information is being
+ // constructed and we haven't started allocating bundles yet.
+ LiveBundle* bundle_;
+
+ // The code positions in this range.
+ Range range_;
+
+ // All uses of the virtual register in this range, ordered by location.
+ InlineForwardList<UsePosition> uses_;
+
+ // Total spill weight that calculate from all the uses' policy. Because the
+ // use's policy can't be changed after initialization, we can update the
+ // weight whenever a use is added to or remove from this range. This way, we
+ // don't need to iterate all the uses every time computeSpillWeight() is
+ // called.
+ size_t usesSpillWeight_;
+
+ // Number of uses that have policy LUse::FIXED.
+ uint32_t numFixedUses_;
+
+ // Whether this range contains the virtual register's definition.
+ bool hasDefinition_;
+
+ LiveRange(VirtualRegister* vreg, Range range)
+ : vreg_(vreg),
+ bundle_(nullptr),
+ range_(range),
+ usesSpillWeight_(0),
+ numFixedUses_(0),
+ hasDefinition_(false)
+
+ {
+ MOZ_ASSERT(!range.empty());
+ }
+
+ void noteAddedUse(UsePosition* use);
+ void noteRemovedUse(UsePosition* use);
+
+ public:
+ static LiveRange* FallibleNew(TempAllocator& alloc, VirtualRegister* vreg,
+ CodePosition from, CodePosition to) {
+ return new (alloc.fallible()) LiveRange(vreg, Range(from, to));
+ }
+
+ VirtualRegister& vreg() const {
+ MOZ_ASSERT(hasVreg());
+ return *vreg_;
+ }
+ bool hasVreg() const { return vreg_ != nullptr; }
+
+ LiveBundle* bundle() const { return bundle_; }
+
+ CodePosition from() const { return range_.from; }
+ CodePosition to() const { return range_.to; }
+ bool covers(CodePosition pos) const { return pos >= from() && pos < to(); }
+
+ // Whether this range wholly contains other.
+ bool contains(LiveRange* other) const;
+
+ // Intersect this range with other, returning the subranges of this
+ // that are before, inside, or after other.
+ void intersect(LiveRange* other, Range* pre, Range* inside,
+ Range* post) const;
+
+ // Whether this range has any intersection with other.
+ bool intersects(LiveRange* other) const;
+
+ UsePositionIterator usesBegin() const { return uses_.begin(); }
+ UsePosition* lastUse() const { return uses_.back(); }
+ bool hasUses() const { return !!usesBegin(); }
+ UsePosition* popUse();
+
+ bool hasDefinition() const { return hasDefinition_; }
+
+ void setFrom(CodePosition from) {
+ range_.from = from;
+ MOZ_ASSERT(!range_.empty());
+ }
+ void setTo(CodePosition to) {
+ range_.to = to;
+ MOZ_ASSERT(!range_.empty());
+ }
+
+ void setBundle(LiveBundle* bundle) { bundle_ = bundle; }
+
+ void addUse(UsePosition* use);
+ void tryToMoveDefAndUsesInto(LiveRange* other);
+
+ void setHasDefinition() {
+ MOZ_ASSERT(!hasDefinition_);
+ hasDefinition_ = true;
+ }
+
+ size_t usesSpillWeight() { return usesSpillWeight_; }
+ uint32_t numFixedUses() { return numFixedUses_; }
+
+#ifdef JS_JITSPEW
+ // Return a string describing this range.
+ UniqueChars toString() const;
+#endif
+
+ // Comparator for use in range splay trees.
+ static int compare(LiveRange* v0, LiveRange* v1) {
+ // LiveRange includes 'from' but excludes 'to'.
+ if (v0->to() <= v1->from()) {
+ return -1;
+ }
+ if (v0->from() >= v1->to()) {
+ return 1;
+ }
+ return 0;
+ }
+};
+
+// Tracks information about bundles that should all be spilled to the same
+// physical location. At the beginning of allocation, each bundle has its own
+// spill set. As bundles are split, the new smaller bundles continue to use the
+// same spill set.
+class SpillSet : public TempObject {
+ // All bundles with this spill set which have been spilled. All bundles in
+ // this list will be given the same physical slot.
+ Vector<LiveBundle*, 1, JitAllocPolicy> list_;
+
+ explicit SpillSet(TempAllocator& alloc) : list_(alloc) {}
+
+ public:
+ static SpillSet* New(TempAllocator& alloc) {
+ return new (alloc) SpillSet(alloc);
+ }
+
+ [[nodiscard]] bool addSpilledBundle(LiveBundle* bundle) {
+ return list_.append(bundle);
+ }
+ size_t numSpilledBundles() const { return list_.length(); }
+ LiveBundle* spilledBundle(size_t i) const { return list_[i]; }
+
+ void setAllocation(LAllocation alloc);
+};
+
+#ifdef JS_JITSPEW
+// See comment on LiveBundle::debugId_ just below. This needs to be atomic
+// because TSan automation runs on debug builds will otherwise (correctly)
+// report a race.
+static mozilla::Atomic<uint32_t> LiveBundle_debugIdCounter =
+ mozilla::Atomic<uint32_t>{0};
+#endif
+
+// A set of live ranges which are all pairwise disjoint. The register allocator
+// attempts to find allocations for an entire bundle, and if it fails the
+// bundle will be broken into smaller ones which are allocated independently.
+class LiveBundle : public TempObject {
+ // Set to use if this bundle or one it is split into is spilled.
+ SpillSet* spill_;
+
+ // All the ranges in this set, ordered by location.
+ InlineForwardList<LiveRange::BundleLink> ranges_;
+
+ // Allocation to use for ranges in this set, bogus if unallocated or spilled
+ // and not yet given a physical stack slot.
+ LAllocation alloc_;
+
+ // Bundle which entirely contains this one and has no register uses. This
+ // may or may not be spilled by the allocator, but it can be spilled and
+ // will not be split.
+ LiveBundle* spillParent_;
+
+#ifdef JS_JITSPEW
+ // This is used only for debug-printing bundles. It gives them an
+ // identifiable identity in the debug output, which they otherwise wouldn't
+ // have.
+ uint32_t debugId_;
+#endif
+
+ LiveBundle(SpillSet* spill, LiveBundle* spillParent)
+ : spill_(spill), spillParent_(spillParent) {
+#ifdef JS_JITSPEW
+ debugId_ = LiveBundle_debugIdCounter++;
+#endif
+ }
+
+ public:
+ static LiveBundle* FallibleNew(TempAllocator& alloc, SpillSet* spill,
+ LiveBundle* spillParent) {
+ return new (alloc.fallible()) LiveBundle(spill, spillParent);
+ }
+
+ SpillSet* spillSet() const { return spill_; }
+ void setSpillSet(SpillSet* spill) { spill_ = spill; }
+
+ LiveRange::BundleLinkIterator rangesBegin() const { return ranges_.begin(); }
+ bool hasRanges() const { return !!rangesBegin(); }
+ LiveRange* firstRange() const { return LiveRange::get(*rangesBegin()); }
+ LiveRange* lastRange() const { return LiveRange::get(ranges_.back()); }
+ LiveRange* rangeFor(CodePosition pos) const;
+ void removeRange(LiveRange* range);
+ void removeRangeAndIncrementIterator(LiveRange::BundleLinkIterator& iter) {
+ ranges_.removeAndIncrement(iter);
+ }
+ void addRange(LiveRange* range);
+ [[nodiscard]] bool addRange(TempAllocator& alloc, VirtualRegister* vreg,
+ CodePosition from, CodePosition to);
+ [[nodiscard]] bool addRangeAndDistributeUses(TempAllocator& alloc,
+ LiveRange* oldRange,
+ CodePosition from,
+ CodePosition to);
+ LiveRange* popFirstRange();
+#ifdef DEBUG
+ size_t numRanges() const;
+#endif
+
+ LAllocation allocation() const { return alloc_; }
+ void setAllocation(LAllocation alloc) { alloc_ = alloc; }
+
+ LiveBundle* spillParent() const { return spillParent_; }
+
+#ifdef JS_JITSPEW
+ uint32_t debugId() const { return debugId_; }
+
+ // Return a string describing this bundle.
+ UniqueChars toString() const;
+#endif
+};
+
+// Information about the allocation for a virtual register.
+class VirtualRegister {
+ // Instruction which defines this register.
+ LNode* ins_ = nullptr;
+
+ // Definition in the instruction for this register.
+ LDefinition* def_ = nullptr;
+
+ // All live ranges for this register. These may overlap each other, and are
+ // ordered by their start position.
+ InlineForwardList<LiveRange::RegisterLink> ranges_;
+
+ // Whether def_ is a temp or an output.
+ bool isTemp_ = false;
+
+ // Whether this vreg is an input for some phi. This use is not reflected in
+ // any range on the vreg.
+ bool usedByPhi_ = false;
+
+ // If this register's definition is MUST_REUSE_INPUT, whether a copy must
+ // be introduced before the definition that relaxes the policy.
+ bool mustCopyInput_ = false;
+
+ void operator=(const VirtualRegister&) = delete;
+ VirtualRegister(const VirtualRegister&) = delete;
+
+ public:
+ VirtualRegister() = default;
+
+ void init(LNode* ins, LDefinition* def, bool isTemp) {
+ MOZ_ASSERT(!ins_);
+ ins_ = ins;
+ def_ = def;
+ isTemp_ = isTemp;
+ }
+
+ LNode* ins() const { return ins_; }
+ LDefinition* def() const { return def_; }
+ LDefinition::Type type() const { return def()->type(); }
+ uint32_t vreg() const { return def()->virtualRegister(); }
+ bool isCompatible(const AnyRegister& r) const {
+ return def_->isCompatibleReg(r);
+ }
+ bool isCompatible(const VirtualRegister& vr) const {
+ return def_->isCompatibleDef(*vr.def_);
+ }
+ bool isTemp() const { return isTemp_; }
+
+ void setUsedByPhi() { usedByPhi_ = true; }
+ bool usedByPhi() { return usedByPhi_; }
+
+ void setMustCopyInput() { mustCopyInput_ = true; }
+ bool mustCopyInput() { return mustCopyInput_; }
+
+ LiveRange::RegisterLinkIterator rangesBegin() const {
+ return ranges_.begin();
+ }
+ LiveRange::RegisterLinkIterator rangesBegin(LiveRange* range) const {
+ return ranges_.begin(&range->registerLink);
+ }
+ bool hasRanges() const { return !!rangesBegin(); }
+ LiveRange* firstRange() const { return LiveRange::get(*rangesBegin()); }
+ LiveRange* lastRange() const { return LiveRange::get(ranges_.back()); }
+ LiveRange* rangeFor(CodePosition pos, bool preferRegister = false) const;
+ void removeRange(LiveRange* range);
+ void addRange(LiveRange* range);
+
+ void removeRangeAndIncrement(LiveRange::RegisterLinkIterator& iter) {
+ ranges_.removeAndIncrement(iter);
+ }
+
+ LiveBundle* firstBundle() const { return firstRange()->bundle(); }
+
+ [[nodiscard]] bool addInitialRange(TempAllocator& alloc, CodePosition from,
+ CodePosition to, size_t* numRanges);
+ void addInitialUse(UsePosition* use);
+ void setInitialDefinition(CodePosition from);
+};
+
+// A sequence of code positions, for tellings BacktrackingAllocator::splitAt
+// where to split.
+using SplitPositionVector = js::Vector<CodePosition, 4, SystemAllocPolicy>;
+
+class BacktrackingAllocator : protected RegisterAllocator {
+ friend class JSONSpewer;
+
+ // This flag is set when testing new allocator modifications.
+ bool testbed;
+
+ BitSet* liveIn;
+ FixedList<VirtualRegister> vregs;
+
+ // Allocation state.
+ StackSlotAllocator stackSlotAllocator;
+
+ // Priority queue element: a bundle and the associated priority.
+ struct QueueItem {
+ LiveBundle* bundle;
+
+ QueueItem(LiveBundle* bundle, size_t priority)
+ : bundle(bundle), priority_(priority) {}
+
+ static size_t priority(const QueueItem& v) { return v.priority_; }
+
+ private:
+ size_t priority_;
+ };
+
+ PriorityQueue<QueueItem, QueueItem, 0, SystemAllocPolicy> allocationQueue;
+
+ // This is a set of LiveRanges. They must be non-overlapping. Attempts to
+ // add an overlapping range will cause AvlTree::insert to MOZ_CRASH().
+ using LiveRangeSet = AvlTree<LiveRange*, LiveRange>;
+
+ // Each physical register is associated with the set of ranges over which
+ // that register is currently allocated.
+ struct PhysicalRegister {
+ bool allocatable;
+ AnyRegister reg;
+ LiveRangeSet allocations;
+
+ PhysicalRegister() : allocatable(false) {}
+ };
+ mozilla::Array<PhysicalRegister, AnyRegister::Total> registers;
+
+ // Ranges of code which are considered to be hot, for which good allocation
+ // should be prioritized.
+ LiveRangeSet hotcode;
+
+ struct CallRange : public TempObject, public InlineListNode<CallRange> {
+ LiveRange::Range range;
+
+ CallRange(CodePosition from, CodePosition to) : range(from, to) {}
+
+ // Comparator for use in splay tree.
+ static int compare(CallRange* v0, CallRange* v1) {
+ if (v0->range.to <= v1->range.from) {
+ return -1;
+ }
+ if (v0->range.from >= v1->range.to) {
+ return 1;
+ }
+ return 0;
+ }
+ };
+
+ // Ranges where all registers must be spilled due to call instructions.
+ using CallRangeList = InlineList<CallRange>;
+ CallRangeList callRangesList;
+ AvlTree<CallRange*, CallRange> callRanges;
+
+ // Information about an allocated stack slot.
+ struct SpillSlot : public TempObject,
+ public InlineForwardListNode<SpillSlot> {
+ LStackSlot alloc;
+ LiveRangeSet allocated;
+
+ SpillSlot(uint32_t slot, LifoAlloc* alloc)
+ : alloc(slot), allocated(alloc) {}
+ };
+ using SpillSlotList = InlineForwardList<SpillSlot>;
+
+ // All allocated slots of each width.
+ SpillSlotList normalSlots, doubleSlots, quadSlots;
+
+ Vector<LiveBundle*, 4, SystemAllocPolicy> spilledBundles;
+
+ using LiveBundleVector = Vector<LiveBundle*, 4, SystemAllocPolicy>;
+
+ // Misc accessors
+ bool compilingWasm() { return mir->outerInfo().compilingWasm(); }
+ VirtualRegister& vreg(const LDefinition* def) {
+ return vregs[def->virtualRegister()];
+ }
+ VirtualRegister& vreg(const LAllocation* alloc) {
+ MOZ_ASSERT(alloc->isUse());
+ return vregs[alloc->toUse()->virtualRegister()];
+ }
+
+ // Helpers for creating and adding MoveGroups
+ [[nodiscard]] bool addMove(LMoveGroup* moves, LiveRange* from, LiveRange* to,
+ LDefinition::Type type) {
+ LAllocation fromAlloc = from->bundle()->allocation();
+ LAllocation toAlloc = to->bundle()->allocation();
+ MOZ_ASSERT(fromAlloc != toAlloc);
+ return moves->add(fromAlloc, toAlloc, type);
+ }
+
+ [[nodiscard]] bool moveInput(LInstruction* ins, LiveRange* from,
+ LiveRange* to, LDefinition::Type type) {
+ if (from->bundle()->allocation() == to->bundle()->allocation()) {
+ return true;
+ }
+ LMoveGroup* moves = getInputMoveGroup(ins);
+ return addMove(moves, from, to, type);
+ }
+
+ [[nodiscard]] bool moveAfter(LInstruction* ins, LiveRange* from,
+ LiveRange* to, LDefinition::Type type) {
+ if (from->bundle()->allocation() == to->bundle()->allocation()) {
+ return true;
+ }
+ LMoveGroup* moves = getMoveGroupAfter(ins);
+ return addMove(moves, from, to, type);
+ }
+
+ [[nodiscard]] bool moveAtExit(LBlock* block, LiveRange* from, LiveRange* to,
+ LDefinition::Type type) {
+ if (from->bundle()->allocation() == to->bundle()->allocation()) {
+ return true;
+ }
+ LMoveGroup* moves = block->getExitMoveGroup(alloc());
+ return addMove(moves, from, to, type);
+ }
+
+ [[nodiscard]] bool moveAtEntry(LBlock* block, LiveRange* from, LiveRange* to,
+ LDefinition::Type type) {
+ if (from->bundle()->allocation() == to->bundle()->allocation()) {
+ return true;
+ }
+ LMoveGroup* moves = block->getEntryMoveGroup(alloc());
+ return addMove(moves, from, to, type);
+ }
+
+ // Out-of-line methods, in the same sequence as in BacktrackingAllocator.cpp.
+
+ // Misc helpers: queries about uses
+ bool isReusedInput(LUse* use, LNode* ins, bool considerCopy);
+ bool isRegisterUse(UsePosition* use, LNode* ins, bool considerCopy = false);
+ bool isRegisterDefinition(LiveRange* range);
+
+ // Misc helpers: atomic LIR groups
+ // (these are all in the parent class, RegisterAllocator)
+
+ // Misc helpers: computation of bundle priorities and spill weights
+ size_t computePriority(LiveBundle* bundle);
+ bool minimalDef(LiveRange* range, LNode* ins);
+ bool minimalUse(LiveRange* range, UsePosition* use);
+ bool minimalBundle(LiveBundle* bundle, bool* pfixed = nullptr);
+ size_t computeSpillWeight(LiveBundle* bundle);
+ size_t maximumSpillWeight(const LiveBundleVector& bundles);
+
+ // Initialization of the allocator
+ [[nodiscard]] bool init();
+
+ // Liveness analysis
+ [[nodiscard]] bool addInitialFixedRange(AnyRegister reg, CodePosition from,
+ CodePosition to);
+ [[nodiscard]] bool buildLivenessInfo();
+
+ // Merging and queueing of LiveRange groups
+ [[nodiscard]] bool tryMergeBundles(LiveBundle* bundle0, LiveBundle* bundle1);
+ void allocateStackDefinition(VirtualRegister& reg);
+ [[nodiscard]] bool tryMergeReusedRegister(VirtualRegister& def,
+ VirtualRegister& input);
+ [[nodiscard]] bool mergeAndQueueRegisters();
+
+ // Implementation of splitting decisions, but not the making of those
+ // decisions
+ [[nodiscard]] bool updateVirtualRegisterListsThenRequeueBundles(
+ LiveBundle* bundle, const LiveBundleVector& newBundles);
+
+ // Implementation of splitting decisions, but not the making of those
+ // decisions
+ [[nodiscard]] bool splitAt(LiveBundle* bundle,
+ const SplitPositionVector& splitPositions);
+
+ // Creation of splitting decisions, but not their implementation
+ [[nodiscard]] bool splitAcrossCalls(LiveBundle* bundle);
+ [[nodiscard]] bool trySplitAcrossHotcode(LiveBundle* bundle, bool* success);
+ [[nodiscard]] bool trySplitAfterLastRegisterUse(LiveBundle* bundle,
+ LiveBundle* conflict,
+ bool* success);
+ [[nodiscard]] bool trySplitBeforeFirstRegisterUse(LiveBundle* bundle,
+ LiveBundle* conflict,
+ bool* success);
+
+ // The top level driver for the splitting machinery
+ [[nodiscard]] bool chooseBundleSplit(LiveBundle* bundle, bool fixed,
+ LiveBundle* conflict);
+
+ // Bundle allocation
+ [[nodiscard]] bool computeRequirement(LiveBundle* bundle,
+ Requirement* prequirement,
+ Requirement* phint);
+ [[nodiscard]] bool tryAllocateRegister(PhysicalRegister& r,
+ LiveBundle* bundle, bool* success,
+ bool* pfixed,
+ LiveBundleVector& conflicting);
+ [[nodiscard]] bool tryAllocateAnyRegister(LiveBundle* bundle, bool* success,
+ bool* pfixed,
+ LiveBundleVector& conflicting);
+ [[nodiscard]] bool evictBundle(LiveBundle* bundle);
+ [[nodiscard]] bool tryAllocateFixed(LiveBundle* bundle,
+ Requirement requirement, bool* success,
+ bool* pfixed,
+ LiveBundleVector& conflicting);
+ [[nodiscard]] bool tryAllocateNonFixed(LiveBundle* bundle,
+ Requirement requirement,
+ Requirement hint, bool* success,
+ bool* pfixed,
+ LiveBundleVector& conflicting);
+ [[nodiscard]] bool processBundle(MIRGenerator* mir, LiveBundle* bundle);
+ [[nodiscard]] bool spill(LiveBundle* bundle);
+ [[nodiscard]] AVOID_INLINE_FOR_DEBUGGING bool
+ tryAllocatingRegistersForSpillBundles();
+
+ // Rewriting of the LIR after bundle processing is done
+ [[nodiscard]] bool insertAllRanges(LiveRangeSet& set, LiveBundle* bundle);
+ [[nodiscard]] bool pickStackSlot(SpillSet* spill);
+ [[nodiscard]] AVOID_INLINE_FOR_DEBUGGING bool pickStackSlots();
+ [[nodiscard]] bool moveAtEdge(LBlock* predecessor, LBlock* successor,
+ LiveRange* from, LiveRange* to,
+ LDefinition::Type type);
+ [[nodiscard]] AVOID_INLINE_FOR_DEBUGGING bool deadRange(LiveRange* range);
+ [[nodiscard]] AVOID_INLINE_FOR_DEBUGGING bool
+ createMoveGroupsFromLiveRangeTransitions();
+ size_t findFirstNonCallSafepoint(CodePosition from);
+ void addLiveRegistersForRange(VirtualRegister& reg, LiveRange* range);
+ [[nodiscard]] AVOID_INLINE_FOR_DEBUGGING bool installAllocationsInLIR();
+ size_t findFirstSafepoint(CodePosition pos, size_t startFrom);
+ [[nodiscard]] AVOID_INLINE_FOR_DEBUGGING bool populateSafepoints();
+ [[nodiscard]] AVOID_INLINE_FOR_DEBUGGING bool annotateMoveGroups();
+
+ // Debug-printing support
+#ifdef JS_JITSPEW
+ void dumpLiveRangesByVReg(const char* who);
+ void dumpLiveRangesByBundle(const char* who);
+ void dumpAllocations();
+#endif
+
+ // Top level of the register allocation machinery, and the only externally
+ // visible bit.
+ public:
+ BacktrackingAllocator(MIRGenerator* mir, LIRGenerator* lir, LIRGraph& graph,
+ bool testbed)
+ : RegisterAllocator(mir, lir, graph),
+ testbed(testbed),
+ liveIn(nullptr),
+ callRanges(nullptr) {}
+
+ [[nodiscard]] bool go();
+};
+
+} // namespace jit
+} // namespace js
+
+#endif /* jit_BacktrackingAllocator_h */