diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 19:33:14 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 19:33:14 +0000 |
commit | 36d22d82aa202bb199967e9512281e9a53db42c9 (patch) | |
tree | 105e8c98ddea1c1e4784a60a5a6410fa416be2de /js/src/jit/BaselineJIT.cpp | |
parent | Initial commit. (diff) | |
download | firefox-esr-36d22d82aa202bb199967e9512281e9a53db42c9.tar.xz firefox-esr-36d22d82aa202bb199967e9512281e9a53db42c9.zip |
Adding upstream version 115.7.0esr.upstream/115.7.0esr
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'js/src/jit/BaselineJIT.cpp')
-rw-r--r-- | js/src/jit/BaselineJIT.cpp | 1008 |
1 files changed, 1008 insertions, 0 deletions
diff --git a/js/src/jit/BaselineJIT.cpp b/js/src/jit/BaselineJIT.cpp new file mode 100644 index 0000000000..f4f223040a --- /dev/null +++ b/js/src/jit/BaselineJIT.cpp @@ -0,0 +1,1008 @@ +/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- + * vim: set ts=8 sts=2 et sw=2 tw=80: + * This Source Code Form is subject to the terms of the Mozilla Public + * License, v. 2.0. If a copy of the MPL was not distributed with this + * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ + +#include "jit/BaselineJIT.h" + +#include "mozilla/BinarySearch.h" +#include "mozilla/CheckedInt.h" +#include "mozilla/DebugOnly.h" +#include "mozilla/MemoryReporting.h" + +#include <algorithm> + +#include "debugger/DebugAPI.h" +#include "gc/GCContext.h" +#include "gc/PublicIterators.h" +#include "jit/AutoWritableJitCode.h" +#include "jit/BaselineCodeGen.h" +#include "jit/BaselineIC.h" +#include "jit/CalleeToken.h" +#include "jit/JitCommon.h" +#include "jit/JitRuntime.h" +#include "jit/JitSpewer.h" +#include "jit/MacroAssembler.h" +#include "js/friend/StackLimits.h" // js::AutoCheckRecursionLimit +#include "vm/Interpreter.h" + +#include "debugger/DebugAPI-inl.h" +#include "gc/GC-inl.h" +#include "jit/JitHints-inl.h" +#include "jit/JitScript-inl.h" +#include "vm/GeckoProfiler-inl.h" +#include "vm/JSScript-inl.h" +#include "vm/Stack-inl.h" + +using mozilla::BinarySearchIf; +using mozilla::CheckedInt; +using mozilla::DebugOnly; + +using namespace js; +using namespace js::jit; + +void ICStubSpace::freeAllAfterMinorGC(Zone* zone) { + if (zone->isAtomsZone()) { + MOZ_ASSERT(allocator_.isEmpty()); + } else { + JSRuntime* rt = zone->runtimeFromMainThread(); + rt->gc.queueAllLifoBlocksForFreeAfterMinorGC(&allocator_); + } +} + +static bool CheckFrame(InterpreterFrame* fp) { + if (fp->isDebuggerEvalFrame()) { + // Debugger eval-in-frame. These are likely short-running scripts so + // don't bother compiling them for now. + JitSpew(JitSpew_BaselineAbort, "debugger frame"); + return false; + } + + if (fp->isFunctionFrame() && TooManyActualArguments(fp->numActualArgs())) { + // Fall back to the interpreter to avoid running out of stack space. + JitSpew(JitSpew_BaselineAbort, "Too many arguments (%u)", + fp->numActualArgs()); + return false; + } + + return true; +} + +struct EnterJitData { + explicit EnterJitData(JSContext* cx) + : jitcode(nullptr), + osrFrame(nullptr), + calleeToken(nullptr), + maxArgv(nullptr), + maxArgc(0), + numActualArgs(0), + osrNumStackValues(0), + envChain(cx), + result(cx), + constructing(false) {} + + uint8_t* jitcode; + InterpreterFrame* osrFrame; + + void* calleeToken; + + Value* maxArgv; + unsigned maxArgc; + unsigned numActualArgs; + unsigned osrNumStackValues; + + RootedObject envChain; + RootedValue result; + + bool constructing; +}; + +static JitExecStatus EnterBaseline(JSContext* cx, EnterJitData& data) { + MOZ_ASSERT(data.osrFrame); + + // Check for potential stack overflow before OSR-ing. + uint32_t extra = + BaselineFrame::Size() + (data.osrNumStackValues * sizeof(Value)); + AutoCheckRecursionLimit recursion(cx); + if (!recursion.checkWithExtra(cx, extra)) { + return JitExec_Aborted; + } + +#ifdef DEBUG + // Assert we don't GC before entering JIT code. A GC could discard JIT code + // or move the function stored in the CalleeToken (it won't be traced at + // this point). We use Maybe<> here so we can call reset() to call the + // AutoAssertNoGC destructor before we enter JIT code. + mozilla::Maybe<JS::AutoAssertNoGC> nogc; + nogc.emplace(cx); +#endif + + MOZ_ASSERT(IsBaselineInterpreterEnabled()); + MOZ_ASSERT(CheckFrame(data.osrFrame)); + + EnterJitCode enter = cx->runtime()->jitRuntime()->enterJit(); + + // Caller must construct |this| before invoking the function. + MOZ_ASSERT_IF(data.constructing, + data.maxArgv[0].isObject() || + data.maxArgv[0].isMagic(JS_UNINITIALIZED_LEXICAL)); + + data.result.setInt32(data.numActualArgs); + { + AssertRealmUnchanged aru(cx); + ActivationEntryMonitor entryMonitor(cx, data.calleeToken); + JitActivation activation(cx); + + data.osrFrame->setRunningInJit(); + +#ifdef DEBUG + nogc.reset(); +#endif + // Single transition point from Interpreter to Baseline. + CALL_GENERATED_CODE(enter, data.jitcode, data.maxArgc, data.maxArgv, + data.osrFrame, data.calleeToken, data.envChain.get(), + data.osrNumStackValues, data.result.address()); + + data.osrFrame->clearRunningInJit(); + } + + // Jit callers wrap primitive constructor return, except for derived + // class constructors, which are forced to do it themselves. + if (!data.result.isMagic() && data.constructing && + data.result.isPrimitive()) { + MOZ_ASSERT(data.maxArgv[0].isObject()); + data.result = data.maxArgv[0]; + } + + // Release temporary buffer used for OSR into Ion. + cx->runtime()->jitRuntime()->freeIonOsrTempData(); + + MOZ_ASSERT_IF(data.result.isMagic(), data.result.isMagic(JS_ION_ERROR)); + return data.result.isMagic() ? JitExec_Error : JitExec_Ok; +} + +JitExecStatus jit::EnterBaselineInterpreterAtBranch(JSContext* cx, + InterpreterFrame* fp, + jsbytecode* pc) { + MOZ_ASSERT(JSOp(*pc) == JSOp::LoopHead); + + EnterJitData data(cx); + + // Use the entry point that skips the debug trap because the C++ interpreter + // already handled this for the current op. + const BaselineInterpreter& interp = + cx->runtime()->jitRuntime()->baselineInterpreter(); + data.jitcode = interp.interpretOpNoDebugTrapAddr().value; + + data.osrFrame = fp; + data.osrNumStackValues = + fp->script()->nfixed() + cx->interpreterRegs().stackDepth(); + + if (fp->isFunctionFrame()) { + data.constructing = fp->isConstructing(); + data.numActualArgs = fp->numActualArgs(); + data.maxArgc = std::max(fp->numActualArgs(), fp->numFormalArgs()) + + 1; // +1 = include |this| + data.maxArgv = fp->argv() - 1; // -1 = include |this| + data.envChain = nullptr; + data.calleeToken = CalleeToToken(&fp->callee(), data.constructing); + } else { + data.constructing = false; + data.numActualArgs = 0; + data.maxArgc = 0; + data.maxArgv = nullptr; + data.envChain = fp->environmentChain(); + data.calleeToken = CalleeToToken(fp->script()); + } + + JitExecStatus status = EnterBaseline(cx, data); + if (status != JitExec_Ok) { + return status; + } + + fp->setReturnValue(data.result); + return JitExec_Ok; +} + +MethodStatus jit::BaselineCompile(JSContext* cx, JSScript* script, + bool forceDebugInstrumentation) { + cx->check(script); + MOZ_ASSERT(!script->hasBaselineScript()); + MOZ_ASSERT(script->canBaselineCompile()); + MOZ_ASSERT(IsBaselineJitEnabled(cx)); + AutoGeckoProfilerEntry pseudoFrame( + cx, "Baseline script compilation", + JS::ProfilingCategoryPair::JS_BaselineCompilation); + + TempAllocator temp(&cx->tempLifoAlloc()); + JitContext jctx(cx); + + BaselineCompiler compiler(cx, temp, script); + if (!compiler.init()) { + ReportOutOfMemory(cx); + return Method_Error; + } + + if (forceDebugInstrumentation) { + compiler.setCompileDebugInstrumentation(); + } + + MethodStatus status = compiler.compile(); + + MOZ_ASSERT_IF(status == Method_Compiled, script->hasBaselineScript()); + MOZ_ASSERT_IF(status != Method_Compiled, !script->hasBaselineScript()); + + if (status == Method_CantCompile) { + script->disableBaselineCompile(); + } + + return status; +} + +static MethodStatus CanEnterBaselineJIT(JSContext* cx, HandleScript script, + AbstractFramePtr osrSourceFrame) { + // Skip if the script has been disabled. + if (!script->canBaselineCompile()) { + return Method_Skipped; + } + + if (!IsBaselineJitEnabled(cx)) { + script->disableBaselineCompile(); + return Method_CantCompile; + } + + // This check is needed in the following corner case. Consider a function h, + // + // function h(x) { + // if (!x) + // return; + // h(false); + // for (var i = 0; i < N; i++) + // /* do stuff */ + // } + // + // Suppose h is not yet compiled in baseline and is executing in the + // interpreter. Let this interpreter frame be f_older. The debugger marks + // f_older as isDebuggee. At the point of the recursive call h(false), h is + // compiled in baseline without debug instrumentation, pushing a baseline + // frame f_newer. The debugger never flags f_newer as isDebuggee, and never + // recompiles h. When the recursive call returns and execution proceeds to + // the loop, the interpreter attempts to OSR into baseline. Since h is + // already compiled in baseline, execution jumps directly into baseline + // code. This is incorrect as h's baseline script does not have debug + // instrumentation. + if (osrSourceFrame && osrSourceFrame.isDebuggee() && + !DebugAPI::ensureExecutionObservabilityOfOsrFrame(cx, osrSourceFrame)) { + return Method_Error; + } + + if (script->length() > BaselineMaxScriptLength) { + script->disableBaselineCompile(); + return Method_CantCompile; + } + + if (script->nslots() > BaselineMaxScriptSlots) { + script->disableBaselineCompile(); + return Method_CantCompile; + } + + if (script->hasBaselineScript()) { + return Method_Compiled; + } + + // If a hint is available, skip the warmup count threshold. + bool mightHaveEagerBaselineHint = false; + if (!JitOptions.disableJitHints && !script->noEagerBaselineHint() && + cx->runtime()->jitRuntime()->hasJitHintsMap()) { + JitHintsMap* jitHints = cx->runtime()->jitRuntime()->getJitHintsMap(); + // If this lookup fails, the NoEagerBaselineHint script flag is set + // to true to prevent any further lookups for this script. + if (jitHints->mightHaveEagerBaselineHint(script)) { + mightHaveEagerBaselineHint = true; + } + } + // Check script warm-up counter if no hint. + if (!mightHaveEagerBaselineHint) { + if (script->getWarmUpCount() <= JitOptions.baselineJitWarmUpThreshold) { + return Method_Skipped; + } + } + + // Check this before calling ensureJitRealmExists, so we're less + // likely to report OOM in JSRuntime::createJitRuntime. + if (!CanLikelyAllocateMoreExecutableMemory()) { + return Method_Skipped; + } + + if (!cx->realm()->ensureJitRealmExists(cx)) { + return Method_Error; + } + + if (script->hasForceInterpreterOp()) { + script->disableBaselineCompile(); + return Method_CantCompile; + } + + // Frames can be marked as debuggee frames independently of its underlying + // script being a debuggee script, e.g., when performing + // Debugger.Frame.prototype.eval. + bool forceDebugInstrumentation = + osrSourceFrame && osrSourceFrame.isDebuggee(); + return BaselineCompile(cx, script, forceDebugInstrumentation); +} + +bool jit::CanBaselineInterpretScript(JSScript* script) { + MOZ_ASSERT(IsBaselineInterpreterEnabled()); + + if (script->hasForceInterpreterOp()) { + return false; + } + + if (script->nslots() > BaselineMaxScriptSlots) { + // Avoid overrecursion exceptions when the script has a ton of stack slots + // by forcing such scripts to run in the C++ interpreter with heap-allocated + // stack frames. + return false; + } + + return true; +} + +static bool MaybeCreateBaselineInterpreterEntryScript(JSContext* cx, + JSScript* script) { + MOZ_ASSERT(script->hasJitScript()); + + JitRuntime* jitRuntime = cx->runtime()->jitRuntime(); + if (script->jitCodeRaw() != jitRuntime->baselineInterpreter().codeRaw()) { + // script already has an updated interpreter trampoline. +#ifdef DEBUG + auto p = jitRuntime->getInterpreterEntryMap()->lookup(script); + MOZ_ASSERT(p); + MOZ_ASSERT(p->value().raw() == script->jitCodeRaw()); +#endif + return true; + } + + auto p = jitRuntime->getInterpreterEntryMap()->lookupForAdd(script); + if (!p) { + Rooted<JitCode*> code( + cx, jitRuntime->generateEntryTrampolineForScript(cx, script)); + if (!code) { + return false; + } + + EntryTrampoline entry(cx, code); + if (!jitRuntime->getInterpreterEntryMap()->add(p, script, entry)) { + return false; + } + } + + script->updateJitCodeRaw(cx->runtime()); + return true; +} + +static MethodStatus CanEnterBaselineInterpreter(JSContext* cx, + JSScript* script) { + MOZ_ASSERT(IsBaselineInterpreterEnabled()); + + if (script->hasJitScript()) { + return Method_Compiled; + } + + if (!CanBaselineInterpretScript(script)) { + return Method_CantCompile; + } + + // Check script warm-up counter. + if (script->getWarmUpCount() <= + JitOptions.baselineInterpreterWarmUpThreshold) { + return Method_Skipped; + } + + if (!cx->realm()->ensureJitRealmExists(cx)) { + return Method_Error; + } + + AutoKeepJitScripts keepJitScript(cx); + if (!script->ensureHasJitScript(cx, keepJitScript)) { + return Method_Error; + } + + if (JitOptions.emitInterpreterEntryTrampoline) { + if (!MaybeCreateBaselineInterpreterEntryScript(cx, script)) { + return Method_Error; + } + } + return Method_Compiled; +} + +MethodStatus jit::CanEnterBaselineInterpreterAtBranch(JSContext* cx, + InterpreterFrame* fp) { + if (!CheckFrame(fp)) { + return Method_CantCompile; + } + + // JITs do not respect the debugger's OnNativeCall hook, so JIT execution is + // disabled if this hook might need to be called. + if (cx->insideDebuggerEvaluationWithOnNativeCallHook) { + return Method_CantCompile; + } + + return CanEnterBaselineInterpreter(cx, fp->script()); +} + +template <BaselineTier Tier> +MethodStatus jit::CanEnterBaselineMethod(JSContext* cx, RunState& state) { + if (state.isInvoke()) { + InvokeState& invoke = *state.asInvoke(); + if (TooManyActualArguments(invoke.args().length())) { + JitSpew(JitSpew_BaselineAbort, "Too many arguments (%u)", + invoke.args().length()); + return Method_CantCompile; + } + } else { + if (state.asExecute()->isDebuggerEval()) { + JitSpew(JitSpew_BaselineAbort, "debugger frame"); + return Method_CantCompile; + } + } + + RootedScript script(cx, state.script()); + switch (Tier) { + case BaselineTier::Interpreter: + return CanEnterBaselineInterpreter(cx, script); + + case BaselineTier::Compiler: + return CanEnterBaselineJIT(cx, script, + /* osrSourceFrame = */ NullFramePtr()); + } + + MOZ_CRASH("Unexpected tier"); +} + +template MethodStatus jit::CanEnterBaselineMethod<BaselineTier::Interpreter>( + JSContext* cx, RunState& state); +template MethodStatus jit::CanEnterBaselineMethod<BaselineTier::Compiler>( + JSContext* cx, RunState& state); + +bool jit::BaselineCompileFromBaselineInterpreter(JSContext* cx, + BaselineFrame* frame, + uint8_t** res) { + MOZ_ASSERT(frame->runningInInterpreter()); + + RootedScript script(cx, frame->script()); + jsbytecode* pc = frame->interpreterPC(); + MOZ_ASSERT(pc == script->code() || JSOp(*pc) == JSOp::LoopHead); + + MethodStatus status = CanEnterBaselineJIT(cx, script, + /* osrSourceFrame = */ frame); + switch (status) { + case Method_Error: + return false; + + case Method_CantCompile: + case Method_Skipped: + *res = nullptr; + return true; + + case Method_Compiled: { + if (JSOp(*pc) == JSOp::LoopHead) { + MOZ_ASSERT(pc > script->code(), + "Prologue vs OSR cases must not be ambiguous"); + BaselineScript* baselineScript = script->baselineScript(); + uint32_t pcOffset = script->pcToOffset(pc); + *res = baselineScript->nativeCodeForOSREntry(pcOffset); + } else { + *res = script->baselineScript()->warmUpCheckPrologueAddr(); + } + frame->prepareForBaselineInterpreterToJitOSR(); + return true; + } + } + + MOZ_CRASH("Unexpected status"); +} + +BaselineScript* BaselineScript::New(JSContext* cx, + uint32_t warmUpCheckPrologueOffset, + uint32_t profilerEnterToggleOffset, + uint32_t profilerExitToggleOffset, + size_t retAddrEntries, size_t osrEntries, + size_t debugTrapEntries, + size_t resumeEntries) { + // Compute size including trailing arrays. + CheckedInt<Offset> size = sizeof(BaselineScript); + size += CheckedInt<Offset>(resumeEntries) * sizeof(uintptr_t); + size += CheckedInt<Offset>(retAddrEntries) * sizeof(RetAddrEntry); + size += CheckedInt<Offset>(osrEntries) * sizeof(OSREntry); + size += CheckedInt<Offset>(debugTrapEntries) * sizeof(DebugTrapEntry); + + if (!size.isValid()) { + ReportAllocationOverflow(cx); + return nullptr; + } + + // Allocate contiguous raw buffer. + void* raw = cx->pod_malloc<uint8_t>(size.value()); + MOZ_ASSERT(uintptr_t(raw) % alignof(BaselineScript) == 0); + if (!raw) { + return nullptr; + } + BaselineScript* script = new (raw) + BaselineScript(warmUpCheckPrologueOffset, profilerEnterToggleOffset, + profilerExitToggleOffset); + + Offset cursor = sizeof(BaselineScript); + + MOZ_ASSERT(isAlignedOffset<uintptr_t>(cursor)); + script->resumeEntriesOffset_ = cursor; + cursor += resumeEntries * sizeof(uintptr_t); + + MOZ_ASSERT(isAlignedOffset<RetAddrEntry>(cursor)); + script->retAddrEntriesOffset_ = cursor; + cursor += retAddrEntries * sizeof(RetAddrEntry); + + MOZ_ASSERT(isAlignedOffset<OSREntry>(cursor)); + script->osrEntriesOffset_ = cursor; + cursor += osrEntries * sizeof(OSREntry); + + MOZ_ASSERT(isAlignedOffset<DebugTrapEntry>(cursor)); + script->debugTrapEntriesOffset_ = cursor; + cursor += debugTrapEntries * sizeof(DebugTrapEntry); + + MOZ_ASSERT(isAlignedOffset<uint32_t>(cursor)); + + script->allocBytes_ = cursor; + + MOZ_ASSERT(script->endOffset() == size.value()); + + return script; +} + +void BaselineScript::trace(JSTracer* trc) { + TraceEdge(trc, &method_, "baseline-method"); +} + +void BaselineScript::Destroy(JS::GCContext* gcx, BaselineScript* script) { + MOZ_ASSERT(!script->hasPendingIonCompileTask()); + + // This allocation is tracked by JSScript::setBaselineScriptImpl. + gcx->deleteUntracked(script); +} + +void JS::DeletePolicy<js::jit::BaselineScript>::operator()( + const js::jit::BaselineScript* script) { + BaselineScript::Destroy(rt_->gcContext(), + const_cast<BaselineScript*>(script)); +} + +const RetAddrEntry& BaselineScript::retAddrEntryFromReturnOffset( + CodeOffset returnOffset) { + mozilla::Span<RetAddrEntry> entries = retAddrEntries(); + size_t loc; +#ifdef DEBUG + bool found = +#endif + BinarySearchIf( + entries.data(), 0, entries.size(), + [&returnOffset](const RetAddrEntry& entry) { + size_t roffset = returnOffset.offset(); + size_t entryRoffset = entry.returnOffset().offset(); + if (roffset < entryRoffset) { + return -1; + } + if (entryRoffset < roffset) { + return 1; + } + return 0; + }, + &loc); + + MOZ_ASSERT(found); + MOZ_ASSERT(entries[loc].returnOffset().offset() == returnOffset.offset()); + return entries[loc]; +} + +template <typename Entry> +static bool ComputeBinarySearchMid(mozilla::Span<Entry> entries, + uint32_t pcOffset, size_t* loc) { + return BinarySearchIf( + entries.data(), 0, entries.size(), + [pcOffset](const Entry& entry) { + uint32_t entryOffset = entry.pcOffset(); + if (pcOffset < entryOffset) { + return -1; + } + if (entryOffset < pcOffset) { + return 1; + } + return 0; + }, + loc); +} + +uint8_t* BaselineScript::returnAddressForEntry(const RetAddrEntry& ent) { + return method()->raw() + ent.returnOffset().offset(); +} + +const RetAddrEntry& BaselineScript::retAddrEntryFromPCOffset( + uint32_t pcOffset, RetAddrEntry::Kind kind) { + mozilla::Span<RetAddrEntry> entries = retAddrEntries(); + size_t mid; + MOZ_ALWAYS_TRUE(ComputeBinarySearchMid(entries, pcOffset, &mid)); + MOZ_ASSERT(mid < entries.size()); + + // Search for the first entry for this pc. + size_t first = mid; + while (first > 0 && entries[first - 1].pcOffset() == pcOffset) { + first--; + } + + // Search for the last entry for this pc. + size_t last = mid; + while (last + 1 < entries.size() && + entries[last + 1].pcOffset() == pcOffset) { + last++; + } + + MOZ_ASSERT(first <= last); + MOZ_ASSERT(entries[first].pcOffset() == pcOffset); + MOZ_ASSERT(entries[last].pcOffset() == pcOffset); + + for (size_t i = first; i <= last; i++) { + const RetAddrEntry& entry = entries[i]; + if (entry.kind() != kind) { + continue; + } + +#ifdef DEBUG + // There must be a unique entry for this pcOffset and Kind to ensure our + // return value is well-defined. + for (size_t j = i + 1; j <= last; j++) { + MOZ_ASSERT(entries[j].kind() != kind); + } +#endif + + return entry; + } + + MOZ_CRASH("Didn't find RetAddrEntry."); +} + +const RetAddrEntry& BaselineScript::prologueRetAddrEntry( + RetAddrEntry::Kind kind) { + MOZ_ASSERT(kind == RetAddrEntry::Kind::StackCheck); + + // The prologue entries will always be at a very low offset, so just do a + // linear search from the beginning. + for (const RetAddrEntry& entry : retAddrEntries()) { + if (entry.pcOffset() != 0) { + break; + } + if (entry.kind() == kind) { + return entry; + } + } + MOZ_CRASH("Didn't find prologue RetAddrEntry."); +} + +const RetAddrEntry& BaselineScript::retAddrEntryFromReturnAddress( + const uint8_t* returnAddr) { + MOZ_ASSERT(returnAddr > method_->raw()); + MOZ_ASSERT(returnAddr < method_->raw() + method_->instructionsSize()); + CodeOffset offset(returnAddr - method_->raw()); + return retAddrEntryFromReturnOffset(offset); +} + +uint8_t* BaselineScript::nativeCodeForOSREntry(uint32_t pcOffset) { + mozilla::Span<OSREntry> entries = osrEntries(); + size_t mid; + if (!ComputeBinarySearchMid(entries, pcOffset, &mid)) { + return nullptr; + } + + uint32_t nativeOffset = entries[mid].nativeOffset(); + return method_->raw() + nativeOffset; +} + +void BaselineScript::computeResumeNativeOffsets( + JSScript* script, const ResumeOffsetEntryVector& entries) { + // Translate pcOffset to BaselineScript native address. This may return + // nullptr if compiler decided code was unreachable. + auto computeNative = [this, &entries](uint32_t pcOffset) -> uint8_t* { + mozilla::Span<const ResumeOffsetEntry> entriesSpan = + mozilla::Span(entries.begin(), entries.length()); + size_t mid; + if (!ComputeBinarySearchMid(entriesSpan, pcOffset, &mid)) { + return nullptr; + } + + uint32_t nativeOffset = entries[mid].nativeOffset(); + return method_->raw() + nativeOffset; + }; + + mozilla::Span<const uint32_t> pcOffsets = script->resumeOffsets(); + mozilla::Span<uint8_t*> nativeOffsets = resumeEntryList(); + std::transform(pcOffsets.begin(), pcOffsets.end(), nativeOffsets.begin(), + computeNative); +} + +void BaselineScript::copyRetAddrEntries(const RetAddrEntry* entries) { + std::copy_n(entries, retAddrEntries().size(), retAddrEntries().data()); +} + +void BaselineScript::copyOSREntries(const OSREntry* entries) { + std::copy_n(entries, osrEntries().size(), osrEntries().data()); +} + +void BaselineScript::copyDebugTrapEntries(const DebugTrapEntry* entries) { + std::copy_n(entries, debugTrapEntries().size(), debugTrapEntries().data()); +} + +jsbytecode* BaselineScript::approximatePcForNativeAddress( + JSScript* script, uint8_t* nativeAddress) { + MOZ_ASSERT(script->baselineScript() == this); + MOZ_ASSERT(containsCodeAddress(nativeAddress)); + + uint32_t nativeOffset = nativeAddress - method_->raw(); + + // Use the RetAddrEntry list (sorted on pc and return address) to look for the + // first pc that has a return address >= nativeOffset. This isn't perfect but + // it's a reasonable approximation for the profiler because most non-trivial + // bytecode ops have a RetAddrEntry. + + for (const RetAddrEntry& entry : retAddrEntries()) { + uint32_t retOffset = entry.returnOffset().offset(); + if (retOffset >= nativeOffset) { + return script->offsetToPC(entry.pcOffset()); + } + } + + // Return the last entry's pc. Every BaselineScript has at least one + // RetAddrEntry for the prologue stack overflow check. + MOZ_ASSERT(!retAddrEntries().empty()); + const RetAddrEntry& lastEntry = retAddrEntries()[retAddrEntries().size() - 1]; + return script->offsetToPC(lastEntry.pcOffset()); +} + +void BaselineScript::toggleDebugTraps(JSScript* script, jsbytecode* pc) { + MOZ_ASSERT(script->baselineScript() == this); + + // Only scripts compiled for debug mode have toggled calls. + if (!hasDebugInstrumentation()) { + return; + } + + AutoWritableJitCode awjc(method()); + + for (const DebugTrapEntry& entry : debugTrapEntries()) { + jsbytecode* entryPC = script->offsetToPC(entry.pcOffset()); + + // If the |pc| argument is non-null we can skip all other bytecode ops. + if (pc && pc != entryPC) { + continue; + } + + bool enabled = DebugAPI::stepModeEnabled(script) || + DebugAPI::hasBreakpointsAt(script, entryPC); + + // Patch the trap. + CodeLocationLabel label(method(), CodeOffset(entry.nativeOffset())); + Assembler::ToggleCall(label, enabled); + } +} + +void BaselineScript::setPendingIonCompileTask(JSRuntime* rt, JSScript* script, + IonCompileTask* task) { + MOZ_ASSERT(script->baselineScript() == this); + MOZ_ASSERT(task); + MOZ_ASSERT(!hasPendingIonCompileTask()); + + if (script->isIonCompilingOffThread()) { + script->jitScript()->clearIsIonCompilingOffThread(script); + } + + pendingIonCompileTask_ = task; + script->updateJitCodeRaw(rt); +} + +void BaselineScript::removePendingIonCompileTask(JSRuntime* rt, + JSScript* script) { + MOZ_ASSERT(script->baselineScript() == this); + MOZ_ASSERT(hasPendingIonCompileTask()); + + pendingIonCompileTask_ = nullptr; + script->updateJitCodeRaw(rt); +} + +static void ToggleProfilerInstrumentation(JitCode* code, + uint32_t profilerEnterToggleOffset, + uint32_t profilerExitToggleOffset, + bool enable) { + CodeLocationLabel enterToggleLocation(code, + CodeOffset(profilerEnterToggleOffset)); + CodeLocationLabel exitToggleLocation(code, + CodeOffset(profilerExitToggleOffset)); + if (enable) { + Assembler::ToggleToCmp(enterToggleLocation); + Assembler::ToggleToCmp(exitToggleLocation); + } else { + Assembler::ToggleToJmp(enterToggleLocation); + Assembler::ToggleToJmp(exitToggleLocation); + } +} + +void BaselineScript::toggleProfilerInstrumentation(bool enable) { + if (enable == isProfilerInstrumentationOn()) { + return; + } + + JitSpew(JitSpew_BaselineIC, " toggling profiling %s for BaselineScript %p", + enable ? "on" : "off", this); + + ToggleProfilerInstrumentation(method_, profilerEnterToggleOffset_, + profilerExitToggleOffset_, enable); + + if (enable) { + flags_ |= uint32_t(PROFILER_INSTRUMENTATION_ON); + } else { + flags_ &= ~uint32_t(PROFILER_INSTRUMENTATION_ON); + } +} + +void BaselineInterpreter::toggleProfilerInstrumentation(bool enable) { + if (!IsBaselineInterpreterEnabled()) { + return; + } + + AutoWritableJitCode awjc(code_); + ToggleProfilerInstrumentation(code_, profilerEnterToggleOffset_, + profilerExitToggleOffset_, enable); +} + +void BaselineInterpreter::toggleDebuggerInstrumentation(bool enable) { + if (!IsBaselineInterpreterEnabled()) { + return; + } + + AutoWritableJitCode awjc(code_); + + // Toggle jumps for debugger instrumentation. + for (uint32_t offset : debugInstrumentationOffsets_) { + CodeLocationLabel label(code_, CodeOffset(offset)); + if (enable) { + Assembler::ToggleToCmp(label); + } else { + Assembler::ToggleToJmp(label); + } + } + + // Toggle DebugTrapHandler calls. + + uint8_t* debugTrapHandler = codeAtOffset(debugTrapHandlerOffset_); + + for (uint32_t offset : debugTrapOffsets_) { + uint8_t* trap = codeAtOffset(offset); + if (enable) { + MacroAssembler::patchNopToCall(trap, debugTrapHandler); + } else { + MacroAssembler::patchCallToNop(trap); + } + } +} + +void BaselineInterpreter::toggleCodeCoverageInstrumentationUnchecked( + bool enable) { + if (!IsBaselineInterpreterEnabled()) { + return; + } + + AutoWritableJitCode awjc(code_); + + for (uint32_t offset : codeCoverageOffsets_) { + CodeLocationLabel label(code_, CodeOffset(offset)); + if (enable) { + Assembler::ToggleToCmp(label); + } else { + Assembler::ToggleToJmp(label); + } + } +} + +void BaselineInterpreter::toggleCodeCoverageInstrumentation(bool enable) { + if (coverage::IsLCovEnabled()) { + // Instrumentation is enabled no matter what. + return; + } + + toggleCodeCoverageInstrumentationUnchecked(enable); +} + +void jit::FinishDiscardBaselineScript(JS::GCContext* gcx, JSScript* script) { + MOZ_ASSERT(script->hasBaselineScript()); + MOZ_ASSERT(!script->jitScript()->active()); + + BaselineScript* baseline = + script->jitScript()->clearBaselineScript(gcx, script); + BaselineScript::Destroy(gcx, baseline); +} + +void jit::AddSizeOfBaselineData(JSScript* script, + mozilla::MallocSizeOf mallocSizeOf, + size_t* data) { + if (script->hasBaselineScript()) { + script->baselineScript()->addSizeOfIncludingThis(mallocSizeOf, data); + } +} + +void jit::ToggleBaselineProfiling(JSContext* cx, bool enable) { + JitRuntime* jrt = cx->runtime()->jitRuntime(); + if (!jrt) { + return; + } + + jrt->baselineInterpreter().toggleProfilerInstrumentation(enable); + + for (ZonesIter zone(cx->runtime(), SkipAtoms); !zone.done(); zone.next()) { + for (auto base = zone->cellIter<BaseScript>(); !base.done(); base.next()) { + if (!base->hasJitScript()) { + continue; + } + JSScript* script = base->asJSScript(); + if (enable) { + script->jitScript()->ensureProfileString(cx, script); + } + if (!script->hasBaselineScript()) { + continue; + } + AutoWritableJitCode awjc(script->baselineScript()->method()); + script->baselineScript()->toggleProfilerInstrumentation(enable); + } + } +} + +void BaselineInterpreter::init(JitCode* code, uint32_t interpretOpOffset, + uint32_t interpretOpNoDebugTrapOffset, + uint32_t bailoutPrologueOffset, + uint32_t profilerEnterToggleOffset, + uint32_t profilerExitToggleOffset, + uint32_t debugTrapHandlerOffset, + CodeOffsetVector&& debugInstrumentationOffsets, + CodeOffsetVector&& debugTrapOffsets, + CodeOffsetVector&& codeCoverageOffsets, + ICReturnOffsetVector&& icReturnOffsets, + const CallVMOffsets& callVMOffsets) { + code_ = code; + interpretOpOffset_ = interpretOpOffset; + interpretOpNoDebugTrapOffset_ = interpretOpNoDebugTrapOffset; + bailoutPrologueOffset_ = bailoutPrologueOffset; + profilerEnterToggleOffset_ = profilerEnterToggleOffset; + profilerExitToggleOffset_ = profilerExitToggleOffset; + debugTrapHandlerOffset_ = debugTrapHandlerOffset; + debugInstrumentationOffsets_ = std::move(debugInstrumentationOffsets); + debugTrapOffsets_ = std::move(debugTrapOffsets); + codeCoverageOffsets_ = std::move(codeCoverageOffsets); + icReturnOffsets_ = std::move(icReturnOffsets); + callVMOffsets_ = callVMOffsets; +} + +uint8_t* BaselineInterpreter::retAddrForIC(JSOp op) const { + for (const ICReturnOffset& entry : icReturnOffsets_) { + if (entry.op == op) { + return codeAtOffset(entry.offset); + } + } + MOZ_CRASH("Unexpected op"); +} + +bool jit::GenerateBaselineInterpreter(JSContext* cx, + BaselineInterpreter& interpreter) { + if (IsBaselineInterpreterEnabled()) { + TempAllocator temp(&cx->tempLifoAlloc()); + BaselineInterpreterGenerator generator(cx, temp); + return generator.generate(interpreter); + } + + return true; +} |