From 6bf0a5cb5034a7e684dcc3500e841785237ce2dd Mon Sep 17 00:00:00 2001 From: Daniel Baumann Date: Sun, 7 Apr 2024 19:32:43 +0200 Subject: Adding upstream version 1:115.7.0. Signed-off-by: Daniel Baumann --- js/src/jit/Lowering.cpp | 7172 +++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 7172 insertions(+) create mode 100644 js/src/jit/Lowering.cpp (limited to 'js/src/jit/Lowering.cpp') diff --git a/js/src/jit/Lowering.cpp b/js/src/jit/Lowering.cpp new file mode 100644 index 0000000000..781a536252 --- /dev/null +++ b/js/src/jit/Lowering.cpp @@ -0,0 +1,7172 @@ +/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- + * vim: set ts=8 sts=2 et sw=2 tw=80: + * This Source Code Form is subject to the terms of the Mozilla Public + * License, v. 2.0. If a copy of the MPL was not distributed with this + * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ + +#include "jit/Lowering.h" + +#include "mozilla/DebugOnly.h" +#include "mozilla/EndianUtils.h" +#include "mozilla/FloatingPoint.h" +#include "mozilla/MathAlgorithms.h" + +#include + +#include "jit/ABIArgGenerator.h" +#include "jit/IonOptimizationLevels.h" +#include "jit/JitSpewer.h" +#include "jit/LIR.h" +#include "jit/MacroAssembler.h" +#include "jit/MIR.h" +#include "jit/MIRGraph.h" +#include "jit/SharedICRegisters.h" +#include "js/experimental/JitInfo.h" // JSJitInfo +#include "util/Memory.h" +#include "wasm/WasmCodegenTypes.h" +#include "wasm/WasmInstanceData.h" +#include "wasm/WasmJS.h" // for wasm::ReportSimdAnalysis + +#include "jit/shared/Lowering-shared-inl.h" +#include "vm/BytecodeUtil-inl.h" + +using namespace js; +using namespace jit; + +using JS::GenericNaN; +using mozilla::DebugOnly; + +LBoxAllocation LIRGenerator::useBoxFixedAtStart(MDefinition* mir, + ValueOperand op) { +#if defined(JS_NUNBOX32) + return useBoxFixed(mir, op.typeReg(), op.payloadReg(), true); +#elif defined(JS_PUNBOX64) + return useBoxFixed(mir, op.valueReg(), op.scratchReg(), true); +#endif +} + +LBoxAllocation LIRGenerator::useBoxAtStart(MDefinition* mir, + LUse::Policy policy) { + return useBox(mir, policy, /* useAtStart = */ true); +} + +void LIRGenerator::visitParameter(MParameter* param) { + ptrdiff_t offset; + if (param->index() == MParameter::THIS_SLOT) { + offset = THIS_FRAME_ARGSLOT; + } else { + offset = 1 + param->index(); + } + + LParameter* ins = new (alloc()) LParameter; + defineBox(ins, param, LDefinition::FIXED); + + offset *= sizeof(Value); +#if defined(JS_NUNBOX32) +# if MOZ_BIG_ENDIAN() + ins->getDef(0)->setOutput(LArgument(offset)); + ins->getDef(1)->setOutput(LArgument(offset + 4)); +# else + ins->getDef(0)->setOutput(LArgument(offset + 4)); + ins->getDef(1)->setOutput(LArgument(offset)); +# endif +#elif defined(JS_PUNBOX64) + ins->getDef(0)->setOutput(LArgument(offset)); +#endif +} + +void LIRGenerator::visitCallee(MCallee* ins) { + define(new (alloc()) LCallee(), ins); +} + +void LIRGenerator::visitIsConstructing(MIsConstructing* ins) { + define(new (alloc()) LIsConstructing(), ins); +} + +void LIRGenerator::visitGoto(MGoto* ins) { + add(new (alloc()) LGoto(ins->target())); +} + +void LIRGenerator::visitTableSwitch(MTableSwitch* tableswitch) { + MDefinition* opd = tableswitch->getOperand(0); + + // There should be at least 1 successor. The default case! + MOZ_ASSERT(tableswitch->numSuccessors() > 0); + + // If there are no cases, the default case is always taken. + if (tableswitch->numSuccessors() == 1) { + add(new (alloc()) LGoto(tableswitch->getDefault())); + return; + } + + // If we don't know the type. + if (opd->type() == MIRType::Value) { + LTableSwitchV* lir = newLTableSwitchV(tableswitch); + add(lir); + return; + } + + // Case indices are numeric, so other types will always go to the default + // case. + if (opd->type() != MIRType::Int32 && opd->type() != MIRType::Double) { + add(new (alloc()) LGoto(tableswitch->getDefault())); + return; + } + + // Return an LTableSwitch, capable of handling either an integer or + // floating-point index. + LAllocation index; + LDefinition tempInt; + if (opd->type() == MIRType::Int32) { + index = useRegisterAtStart(opd); + tempInt = tempCopy(opd, 0); + } else { + index = useRegister(opd); + tempInt = temp(LDefinition::GENERAL); + } + add(newLTableSwitch(index, tempInt, tableswitch)); +} + +void LIRGenerator::visitCheckOverRecursed(MCheckOverRecursed* ins) { + LCheckOverRecursed* lir = new (alloc()) LCheckOverRecursed(); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewArray(MNewArray* ins) { + LNewArray* lir = new (alloc()) LNewArray(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewArrayDynamicLength(MNewArrayDynamicLength* ins) { + MDefinition* length = ins->length(); + MOZ_ASSERT(length->type() == MIRType::Int32); + + LNewArrayDynamicLength* lir = + new (alloc()) LNewArrayDynamicLength(useRegister(length), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewIterator(MNewIterator* ins) { + LNewIterator* lir = new (alloc()) LNewIterator(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewTypedArray(MNewTypedArray* ins) { + LNewTypedArray* lir = new (alloc()) LNewTypedArray(temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewTypedArrayDynamicLength( + MNewTypedArrayDynamicLength* ins) { + MDefinition* length = ins->length(); + MOZ_ASSERT(length->type() == MIRType::Int32); + + LNewTypedArrayDynamicLength* lir = + new (alloc()) LNewTypedArrayDynamicLength(useRegister(length), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewTypedArrayFromArray(MNewTypedArrayFromArray* ins) { + MDefinition* array = ins->array(); + MOZ_ASSERT(array->type() == MIRType::Object); + + auto* lir = new (alloc()) LNewTypedArrayFromArray(useRegisterAtStart(array)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewTypedArrayFromArrayBuffer( + MNewTypedArrayFromArrayBuffer* ins) { + MDefinition* arrayBuffer = ins->arrayBuffer(); + MDefinition* byteOffset = ins->byteOffset(); + MDefinition* length = ins->length(); + MOZ_ASSERT(arrayBuffer->type() == MIRType::Object); + MOZ_ASSERT(byteOffset->type() == MIRType::Value); + MOZ_ASSERT(length->type() == MIRType::Value); + + auto* lir = new (alloc()) LNewTypedArrayFromArrayBuffer( + useRegisterAtStart(arrayBuffer), useBoxAtStart(byteOffset), + useBoxAtStart(length)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewObject(MNewObject* ins) { + LNewObject* lir = new (alloc()) LNewObject(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBindFunction(MBindFunction* ins) { + MDefinition* target = ins->target(); + MOZ_ASSERT(target->type() == MIRType::Object); + + if (!lowerCallArguments(ins)) { + abort(AbortReason::Alloc, "OOM: LIRGenerator::visitBindFunction"); + return; + } + + auto* lir = new (alloc()) + LBindFunction(useFixedAtStart(target, CallTempReg0), + tempFixed(CallTempReg1), tempFixed(CallTempReg2)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewBoundFunction(MNewBoundFunction* ins) { + auto* lir = new (alloc()) LNewBoundFunction(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewPlainObject(MNewPlainObject* ins) { + LNewPlainObject* lir = new (alloc()) LNewPlainObject(temp(), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewArrayObject(MNewArrayObject* ins) { + LNewArrayObject* lir = new (alloc()) LNewArrayObject(temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewNamedLambdaObject(MNewNamedLambdaObject* ins) { + LNewNamedLambdaObject* lir = new (alloc()) LNewNamedLambdaObject(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewCallObject(MNewCallObject* ins) { + LNewCallObject* lir = new (alloc()) LNewCallObject(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewStringObject(MNewStringObject* ins) { + MOZ_ASSERT(ins->input()->type() == MIRType::String); + + LNewStringObject* lir = + new (alloc()) LNewStringObject(useRegister(ins->input()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitInitElemGetterSetter(MInitElemGetterSetter* ins) { + LInitElemGetterSetter* lir = new (alloc()) LInitElemGetterSetter( + useRegisterAtStart(ins->object()), useBoxAtStart(ins->id()), + useRegisterAtStart(ins->value())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitMutateProto(MMutateProto* ins) { + LMutateProto* lir = new (alloc()) LMutateProto( + useRegisterAtStart(ins->object()), useBoxAtStart(ins->value())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitInitPropGetterSetter(MInitPropGetterSetter* ins) { + LInitPropGetterSetter* lir = new (alloc()) LInitPropGetterSetter( + useRegisterAtStart(ins->object()), useRegisterAtStart(ins->value())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCreateThis(MCreateThis* ins) { + LCreateThis* lir = + new (alloc()) LCreateThis(useRegisterOrConstantAtStart(ins->callee()), + useRegisterOrConstantAtStart(ins->newTarget())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCreateArgumentsObject(MCreateArgumentsObject* ins) { + LAllocation callObj = useRegisterAtStart(ins->getCallObject()); + LCreateArgumentsObject* lir = new (alloc()) + LCreateArgumentsObject(callObj, tempFixed(CallTempReg0), + tempFixed(CallTempReg1), tempFixed(CallTempReg2)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCreateInlinedArgumentsObject( + MCreateInlinedArgumentsObject* ins) { + LAllocation callObj = useRegisterAtStart(ins->getCallObject()); + LAllocation callee = useRegisterAtStart(ins->getCallee()); + uint32_t numActuals = ins->numActuals(); + uint32_t numOperands = numActuals * BOX_PIECES + + LCreateInlinedArgumentsObject::NumNonArgumentOperands; + + auto* lir = allocateVariadic( + numOperands, tempFixed(CallTempReg0), tempFixed(CallTempReg1)); + if (!lir) { + abort(AbortReason::Alloc, + "OOM: LIRGenerator::visitCreateInlinedArgumentsObject"); + return; + } + + lir->setOperand(LCreateInlinedArgumentsObject::CallObj, callObj); + lir->setOperand(LCreateInlinedArgumentsObject::Callee, callee); + for (uint32_t i = 0; i < numActuals; i++) { + MDefinition* arg = ins->getArg(i); + uint32_t index = LCreateInlinedArgumentsObject::ArgIndex(i); + lir->setBoxOperand(index, useBoxOrTypedOrConstant(arg, + /*useConstant = */ true, + /*useAtStart = */ true)); + } + + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetInlinedArgument(MGetInlinedArgument* ins) { +#if defined(JS_PUNBOX64) + // On 64-bit architectures, we don't support boxing a typed register + // in-place without using a scratch register, so the result register + // can't be the same as any of the inputs. Fortunately, those + // architectures have registers to spare. + const bool useAtStart = false; +#else + const bool useAtStart = true; +#endif + + LAllocation index = + useAtStart ? useRegisterAtStart(ins->index()) : useRegister(ins->index()); + uint32_t numActuals = ins->numActuals(); + uint32_t numOperands = + numActuals * BOX_PIECES + LGetInlinedArgument::NumNonArgumentOperands; + + auto* lir = allocateVariadic(numOperands); + if (!lir) { + abort(AbortReason::Alloc, "OOM: LIRGenerator::visitGetInlinedArgument"); + return; + } + + lir->setOperand(LGetInlinedArgument::Index, index); + for (uint32_t i = 0; i < numActuals; i++) { + MDefinition* arg = ins->getArg(i); + uint32_t index = LGetInlinedArgument::ArgIndex(i); + lir->setBoxOperand( + index, useBoxOrTypedOrConstant(arg, + /*useConstant = */ true, useAtStart)); + } + defineBox(lir, ins); +} + +void LIRGenerator::visitGetInlinedArgumentHole(MGetInlinedArgumentHole* ins) { +#if defined(JS_CODEGEN_X64) || defined(JS_CODEGEN_MIPS64) + // On some 64-bit architectures, we don't support boxing a typed + // register in-place without using a scratch register, so the result + // register can't be the same as any of the inputs. Fortunately, + // those architectures have registers to spare. + const bool useAtStart = false; +#else + const bool useAtStart = true; +#endif + + LAllocation index = + useAtStart ? useRegisterAtStart(ins->index()) : useRegister(ins->index()); + uint32_t numActuals = ins->numActuals(); + uint32_t numOperands = + numActuals * BOX_PIECES + LGetInlinedArgumentHole::NumNonArgumentOperands; + + auto* lir = allocateVariadic(numOperands); + if (!lir) { + abort(AbortReason::Alloc, "OOM: LIRGenerator::visitGetInlinedArgumentHole"); + return; + } + + lir->setOperand(LGetInlinedArgumentHole::Index, index); + for (uint32_t i = 0; i < numActuals; i++) { + MDefinition* arg = ins->getArg(i); + uint32_t index = LGetInlinedArgumentHole::ArgIndex(i); + lir->setBoxOperand( + index, useBoxOrTypedOrConstant(arg, + /*useConstant = */ true, useAtStart)); + } + assignSnapshot(lir, ins->bailoutKind()); + defineBox(lir, ins); +} + +void LIRGenerator::visitGetArgumentsObjectArg(MGetArgumentsObjectArg* ins) { + LAllocation argsObj = useRegister(ins->argsObject()); + LGetArgumentsObjectArg* lir = + new (alloc()) LGetArgumentsObjectArg(argsObj, temp()); + defineBox(lir, ins); +} + +void LIRGenerator::visitSetArgumentsObjectArg(MSetArgumentsObjectArg* ins) { + LAllocation argsObj = useRegister(ins->argsObject()); + LSetArgumentsObjectArg* lir = new (alloc()) + LSetArgumentsObjectArg(argsObj, useBox(ins->value()), temp()); + add(lir, ins); +} + +void LIRGenerator::visitLoadArgumentsObjectArg(MLoadArgumentsObjectArg* ins) { + MDefinition* argsObj = ins->argsObject(); + MOZ_ASSERT(argsObj->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) + LLoadArgumentsObjectArg(useRegister(argsObj), useRegister(index), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineBox(lir, ins); +} + +void LIRGenerator::visitLoadArgumentsObjectArgHole( + MLoadArgumentsObjectArgHole* ins) { + MDefinition* argsObj = ins->argsObject(); + MOZ_ASSERT(argsObj->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) LLoadArgumentsObjectArgHole( + useRegister(argsObj), useRegister(index), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineBox(lir, ins); +} + +void LIRGenerator::visitInArgumentsObjectArg(MInArgumentsObjectArg* ins) { + MDefinition* argsObj = ins->argsObject(); + MOZ_ASSERT(argsObj->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) + LInArgumentsObjectArg(useRegister(argsObj), useRegister(index), temp()); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitArgumentsObjectLength(MArgumentsObjectLength* ins) { + MDefinition* argsObj = ins->argsObject(); + MOZ_ASSERT(argsObj->type() == MIRType::Object); + + auto* lir = new (alloc()) LArgumentsObjectLength(useRegister(argsObj)); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitArrayFromArgumentsObject( + MArrayFromArgumentsObject* ins) { + MDefinition* argsObj = ins->argsObject(); + MOZ_ASSERT(argsObj->type() == MIRType::Object); + + auto* lir = + new (alloc()) LArrayFromArgumentsObject(useRegisterAtStart(argsObj)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGuardArgumentsObjectFlags( + MGuardArgumentsObjectFlags* ins) { + MDefinition* argsObj = ins->argsObject(); + MOZ_ASSERT(argsObj->type() == MIRType::Object); + + auto* lir = + new (alloc()) LGuardArgumentsObjectFlags(useRegister(argsObj), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, argsObj); +} + +void LIRGenerator::visitBoundFunctionNumArgs(MBoundFunctionNumArgs* ins) { + MDefinition* obj = ins->object(); + MOZ_ASSERT(obj->type() == MIRType::Object); + + auto* lir = new (alloc()) LBoundFunctionNumArgs(useRegisterAtStart(obj)); + define(lir, ins); +} + +void LIRGenerator::visitGuardBoundFunctionIsConstructor( + MGuardBoundFunctionIsConstructor* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = new (alloc()) + LGuardBoundFunctionIsConstructor(useRegister(ins->object())); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitReturnFromCtor(MReturnFromCtor* ins) { + LReturnFromCtor* lir = new (alloc()) + LReturnFromCtor(useBox(ins->value()), useRegister(ins->object())); + define(lir, ins); +} + +void LIRGenerator::visitBoxNonStrictThis(MBoxNonStrictThis* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + MOZ_ASSERT(ins->input()->type() == MIRType::Value); + + auto* lir = new (alloc()) LBoxNonStrictThis(useBox(ins->input())); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitImplicitThis(MImplicitThis* ins) { + MDefinition* env = ins->envChain(); + MOZ_ASSERT(env->type() == MIRType::Object); + + LImplicitThis* lir = new (alloc()) LImplicitThis(useRegisterAtStart(env)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +template +bool LIRGenerator::lowerCallArguments(T* call) { + uint32_t argc = call->numStackArgs(); + + // Align the arguments of a call such that the callee would keep the same + // alignment as the caller. + uint32_t baseSlot = 0; + if (JitStackValueAlignment > 1) { + baseSlot = AlignBytes(argc, JitStackValueAlignment); + } else { + baseSlot = argc; + } + + // Save the maximum number of argument, such that we can have one unique + // frame size. + if (baseSlot > maxargslots_) { + maxargslots_ = baseSlot; + } + + for (size_t i = 0; i < argc; i++) { + MDefinition* arg = call->getArg(i); + uint32_t argslot = baseSlot - i; + + // Values take a slow path. + if (arg->type() == MIRType::Value) { + LStackArgV* stack = new (alloc()) LStackArgV(useBox(arg), argslot); + add(stack); + } else { + // Known types can move constant types and/or payloads. + LStackArgT* stack = new (alloc()) + LStackArgT(useRegisterOrConstant(arg), argslot, arg->type()); + add(stack); + } + + if (!alloc().ensureBallast()) { + return false; + } + } + return true; +} + +void LIRGenerator::visitCall(MCall* call) { + MOZ_ASSERT(call->getCallee()->type() == MIRType::Object); + + // In case of oom, skip the rest of the allocations. + if (!lowerCallArguments(call)) { + abort(AbortReason::Alloc, "OOM: LIRGenerator::visitCall"); + return; + } + + WrappedFunction* target = call->getSingleTarget(); + + LInstruction* lir; + + if (call->isCallDOMNative()) { + // Call DOM functions. + MOZ_ASSERT(target && target->isNativeWithoutJitEntry()); + Register cxReg, objReg, privReg, argsReg; + GetTempRegForIntArg(0, 0, &cxReg); + GetTempRegForIntArg(1, 0, &objReg); + GetTempRegForIntArg(2, 0, &privReg); + mozilla::DebugOnly ok = GetTempRegForIntArg(3, 0, &argsReg); + MOZ_ASSERT(ok, "How can we not have four temp registers?"); + lir = new (alloc()) LCallDOMNative(tempFixed(cxReg), tempFixed(objReg), + tempFixed(privReg), tempFixed(argsReg)); + } else if (target) { + // Call known functions. + if (target->isNativeWithoutJitEntry()) { + Register cxReg, numReg, vpReg, tmpReg; + GetTempRegForIntArg(0, 0, &cxReg); + GetTempRegForIntArg(1, 0, &numReg); + GetTempRegForIntArg(2, 0, &vpReg); + + // Even though this is just a temp reg, use the same API to avoid + // register collisions. + mozilla::DebugOnly ok = GetTempRegForIntArg(3, 0, &tmpReg); + MOZ_ASSERT(ok, "How can we not have four temp registers?"); + + lir = new (alloc()) LCallNative(tempFixed(cxReg), tempFixed(numReg), + tempFixed(vpReg), tempFixed(tmpReg)); + } else { + lir = new (alloc()) LCallKnown(useRegisterAtStart(call->getCallee()), + tempFixed(CallTempReg0)); + } + } else { + // Call anything, using the most generic code. + lir = new (alloc()) + LCallGeneric(useRegisterAtStart(call->getCallee()), + tempFixed(CallTempReg0), tempFixed(CallTempReg1)); + } + defineReturn(lir, call); + assignSafepoint(lir, call); +} + +void LIRGenerator::visitCallClassHook(MCallClassHook* call) { + MDefinition* callee = call->getCallee(); + MOZ_ASSERT(callee->type() == MIRType::Object); + + // In case of oom, skip the rest of the allocations. + if (!lowerCallArguments(call)) { + abort(AbortReason::Alloc, "OOM: LIRGenerator::visitCallClassHook"); + return; + } + + Register cxReg, numReg, vpReg, tmpReg; + GetTempRegForIntArg(0, 0, &cxReg); + GetTempRegForIntArg(1, 0, &numReg); + GetTempRegForIntArg(2, 0, &vpReg); + + // Even though this is just a temp reg, use the same API to avoid + // register collisions. + mozilla::DebugOnly ok = GetTempRegForIntArg(3, 0, &tmpReg); + MOZ_ASSERT(ok, "How can we not have four temp registers?"); + + auto* lir = new (alloc()) + LCallClassHook(useRegisterAtStart(callee), tempFixed(cxReg), + tempFixed(numReg), tempFixed(vpReg), tempFixed(tmpReg)); + defineReturn(lir, call); + assignSafepoint(lir, call); +} + +void LIRGenerator::visitApplyArgs(MApplyArgs* apply) { + MOZ_ASSERT(apply->getFunction()->type() == MIRType::Object); + + // Assert if the return value is already erased. + static_assert(CallTempReg2 != JSReturnReg_Type); + static_assert(CallTempReg2 != JSReturnReg_Data); + + LApplyArgsGeneric* lir = new (alloc()) LApplyArgsGeneric( + useFixedAtStart(apply->getFunction(), CallTempReg3), + useFixedAtStart(apply->getArgc(), CallTempReg0), + useBoxFixedAtStart(apply->getThis(), CallTempReg4, CallTempReg5), + tempFixed(CallTempReg1), // object register + tempFixed(CallTempReg2)); // stack counter register + + // Bailout is needed in the case of too many values in the arguments array. + assignSnapshot(lir, apply->bailoutKind()); + + defineReturn(lir, apply); + assignSafepoint(lir, apply); +} + +void LIRGenerator::visitApplyArgsObj(MApplyArgsObj* apply) { + MOZ_ASSERT(apply->getFunction()->type() == MIRType::Object); + + // Assert if the return value is already erased. + static_assert(CallTempReg2 != JSReturnReg_Type); + static_assert(CallTempReg2 != JSReturnReg_Data); + + LApplyArgsObj* lir = new (alloc()) LApplyArgsObj( + useFixedAtStart(apply->getFunction(), CallTempReg3), + useFixedAtStart(apply->getArgsObj(), CallTempReg0), + useBoxFixedAtStart(apply->getThis(), CallTempReg4, CallTempReg5), + tempFixed(CallTempReg1), // object register + tempFixed(CallTempReg2)); // stack counter register + + // Bailout is needed in the case of too many values in the arguments array. + assignSnapshot(lir, apply->bailoutKind()); + + defineReturn(lir, apply); + assignSafepoint(lir, apply); +} + +void LIRGenerator::visitApplyArray(MApplyArray* apply) { + MOZ_ASSERT(apply->getFunction()->type() == MIRType::Object); + + // Assert if the return value is already erased. + static_assert(CallTempReg2 != JSReturnReg_Type); + static_assert(CallTempReg2 != JSReturnReg_Data); + + LApplyArrayGeneric* lir = new (alloc()) LApplyArrayGeneric( + useFixedAtStart(apply->getFunction(), CallTempReg3), + useFixedAtStart(apply->getElements(), CallTempReg0), + useBoxFixedAtStart(apply->getThis(), CallTempReg4, CallTempReg5), + tempFixed(CallTempReg1), // object register + tempFixed(CallTempReg2)); // stack counter register + + // Bailout is needed in the case of too many values in the array, or empty + // space at the end of the array. + assignSnapshot(lir, apply->bailoutKind()); + + defineReturn(lir, apply); + assignSafepoint(lir, apply); +} + +void LIRGenerator::visitConstructArgs(MConstructArgs* mir) { + MOZ_ASSERT(mir->getFunction()->type() == MIRType::Object); + MOZ_ASSERT(mir->getArgc()->type() == MIRType::Int32); + MOZ_ASSERT(mir->getNewTarget()->type() == MIRType::Object); + MOZ_ASSERT(mir->getThis()->type() == MIRType::Value); + + // Assert if the return value is already erased. + static_assert(CallTempReg2 != JSReturnReg_Type); + static_assert(CallTempReg2 != JSReturnReg_Data); + + auto* lir = new (alloc()) LConstructArgsGeneric( + useFixedAtStart(mir->getFunction(), CallTempReg3), + useFixedAtStart(mir->getArgc(), CallTempReg0), + useFixedAtStart(mir->getNewTarget(), CallTempReg1), + useBoxFixedAtStart(mir->getThis(), CallTempReg4, CallTempReg5), + tempFixed(CallTempReg2)); + + // Bailout is needed in the case of too many values in the arguments array. + assignSnapshot(lir, mir->bailoutKind()); + + defineReturn(lir, mir); + assignSafepoint(lir, mir); +} + +void LIRGenerator::visitConstructArray(MConstructArray* mir) { + MOZ_ASSERT(mir->getFunction()->type() == MIRType::Object); + MOZ_ASSERT(mir->getElements()->type() == MIRType::Elements); + MOZ_ASSERT(mir->getNewTarget()->type() == MIRType::Object); + MOZ_ASSERT(mir->getThis()->type() == MIRType::Value); + + // Assert if the return value is already erased. + static_assert(CallTempReg2 != JSReturnReg_Type); + static_assert(CallTempReg2 != JSReturnReg_Data); + + auto* lir = new (alloc()) LConstructArrayGeneric( + useFixedAtStart(mir->getFunction(), CallTempReg3), + useFixedAtStart(mir->getElements(), CallTempReg0), + useFixedAtStart(mir->getNewTarget(), CallTempReg1), + useBoxFixedAtStart(mir->getThis(), CallTempReg4, CallTempReg5), + tempFixed(CallTempReg2)); + + // Bailout is needed in the case of too many values in the array, or empty + // space at the end of the array. + assignSnapshot(lir, mir->bailoutKind()); + + defineReturn(lir, mir); + assignSafepoint(lir, mir); +} + +void LIRGenerator::visitBail(MBail* bail) { + LBail* lir = new (alloc()) LBail(); + assignSnapshot(lir, bail->bailoutKind()); + add(lir, bail); +} + +void LIRGenerator::visitUnreachable(MUnreachable* unreachable) { + LUnreachable* lir = new (alloc()) LUnreachable(); + add(lir, unreachable); +} + +void LIRGenerator::visitEncodeSnapshot(MEncodeSnapshot* mir) { + LEncodeSnapshot* lir = new (alloc()) LEncodeSnapshot(); + assignSnapshot(lir, mir->bailoutKind()); + add(lir, mir); +} + +void LIRGenerator::visitUnreachableResult(MUnreachableResult* mir) { + if (mir->type() == MIRType::Value) { + auto* lir = new (alloc()) LUnreachableResultV(); + defineBox(lir, mir); + } else { + auto* lir = new (alloc()) LUnreachableResultT(); + define(lir, mir); + } +} + +void LIRGenerator::visitAssertFloat32(MAssertFloat32* assertion) { + MIRType type = assertion->input()->type(); + DebugOnly checkIsFloat32 = assertion->mustBeFloat32(); + + if (type != MIRType::Value && !JitOptions.eagerIonCompilation()) { + MOZ_ASSERT_IF(checkIsFloat32, type == MIRType::Float32); + MOZ_ASSERT_IF(!checkIsFloat32, type != MIRType::Float32); + } +} + +void LIRGenerator::visitAssertRecoveredOnBailout( + MAssertRecoveredOnBailout* assertion) { + MOZ_CRASH("AssertRecoveredOnBailout nodes are always recovered on bailouts."); +} + +[[nodiscard]] static JSOp ReorderComparison(JSOp op, MDefinition** lhsp, + MDefinition** rhsp) { + MDefinition* lhs = *lhsp; + MDefinition* rhs = *rhsp; + + if (lhs->maybeConstantValue()) { + *rhsp = lhs; + *lhsp = rhs; + return ReverseCompareOp(op); + } + return op; +} + +void LIRGenerator::visitTest(MTest* test) { + MDefinition* opd = test->getOperand(0); + MBasicBlock* ifTrue = test->ifTrue(); + MBasicBlock* ifFalse = test->ifFalse(); + + // String is converted to length of string in the type analysis phase (see + // TestPolicy). + MOZ_ASSERT(opd->type() != MIRType::String); + + // Testing a constant. + if (MConstant* constant = opd->maybeConstantValue()) { + bool b; + if (constant->valueToBoolean(&b)) { + add(new (alloc()) LGoto(b ? ifTrue : ifFalse)); + return; + } + } + + if (opd->type() == MIRType::Value) { + auto* lir = new (alloc()) LTestVAndBranch( + ifTrue, ifFalse, useBox(opd), tempDouble(), tempToUnbox(), temp()); + add(lir, test); + return; + } + + // Objects are truthy, except if it might emulate undefined. + if (opd->type() == MIRType::Object) { + add(new (alloc()) + LTestOAndBranch(useRegister(opd), ifTrue, ifFalse, temp()), + test); + return; + } + + // These must be explicitly sniffed out since they are constants and have + // no payload. + if (opd->type() == MIRType::Undefined || opd->type() == MIRType::Null) { + add(new (alloc()) LGoto(ifFalse)); + return; + } + + // All symbols are truthy. + if (opd->type() == MIRType::Symbol) { + add(new (alloc()) LGoto(ifTrue)); + return; + } + + // Try to match the pattern + // test=MTest( + // comp=MCompare( + // {EQ,NE} for {Int,UInt}{32,64}, + // bitAnd={MBitAnd,MWasmBinaryBitwise(And{32,64})}(x, y), + // MConstant(0) + // ) + // ) + // and produce a single LBitAndAndBranch node. This requires both `comp` + // and `bitAnd` to be marked emit-at-uses. Since we can't use + // LBitAndAndBranch to represent a 64-bit AND on a 32-bit target, the 64-bit + // case is restricted to 64-bit targets. + if (opd->isCompare() && opd->isEmittedAtUses()) { +#ifdef JS_64BIT + constexpr bool targetIs64 = true; +#else + constexpr bool targetIs64 = false; +#endif + MCompare* comp = opd->toCompare(); + Assembler::Condition compCond = + JSOpToCondition(comp->compareType(), comp->jsop()); + MDefinition* compL = comp->getOperand(0); + MDefinition* compR = comp->getOperand(1); + if ((comp->compareType() == MCompare::Compare_Int32 || + comp->compareType() == MCompare::Compare_UInt32 || + (targetIs64 && comp->compareType() == MCompare::Compare_Int64) || + (targetIs64 && comp->compareType() == MCompare::Compare_UInt64)) && + (compCond == Assembler::Equal || compCond == Assembler::NotEqual) && + compR->isConstant() && + (compR->toConstant()->isInt32(0) || + (targetIs64 && compR->toConstant()->isInt64(0))) && + (compL->isBitAnd() || (compL->isWasmBinaryBitwise() && + compL->toWasmBinaryBitwise()->subOpcode() == + MWasmBinaryBitwise::SubOpcode::And))) { + // The MCompare is OK; now check its first operand (the and-ish node). + MDefinition* bitAnd = compL; + MDefinition* bitAndL = bitAnd->getOperand(0); + MDefinition* bitAndR = bitAnd->getOperand(1); + MIRType bitAndLTy = bitAndL->type(); + MIRType bitAndRTy = bitAndR->type(); + if (bitAnd->isEmittedAtUses() && bitAndLTy == bitAndRTy && + (bitAndLTy == MIRType::Int32 || + (targetIs64 && bitAndLTy == MIRType::Int64))) { + // Pattern match succeeded. + ReorderCommutative(&bitAndL, &bitAndR, test); + if (compCond == Assembler::Equal) { + compCond = Assembler::Zero; + } else if (compCond == Assembler::NotEqual) { + compCond = Assembler::NonZero; + } else { + MOZ_ASSERT_UNREACHABLE("inequality operators cannot be folded"); + } + MOZ_ASSERT_IF(!targetIs64, bitAndLTy == MIRType::Int32); + lowerForBitAndAndBranch( + new (alloc()) LBitAndAndBranch( + ifTrue, ifFalse, bitAndLTy == MIRType::Int64, compCond), + test, bitAndL, bitAndR); + return; + } + } + } + + // Check if the operand for this test is a compare operation. If it is, we + // want to emit an LCompare*AndBranch rather than an LTest*AndBranch, to fuse + // the compare and jump instructions. + if (opd->isCompare() && opd->isEmittedAtUses()) { + MCompare* comp = opd->toCompare(); + MDefinition* left = comp->lhs(); + MDefinition* right = comp->rhs(); + + // Try to fold the comparison so that we don't have to handle all cases. + bool result; + if (comp->tryFold(&result)) { + add(new (alloc()) LGoto(result ? ifTrue : ifFalse)); + return; + } + + // Emit LCompare*AndBranch. + + // Compare and branch null/undefined. + // The second operand has known null/undefined type, + // so just test the first operand. + if (comp->compareType() == MCompare::Compare_Null || + comp->compareType() == MCompare::Compare_Undefined) { + if (left->type() == MIRType::Object) { + auto* lir = new (alloc()) LIsNullOrLikeUndefinedAndBranchT( + comp, useRegister(left), ifTrue, ifFalse, temp()); + add(lir, test); + return; + } + + if (IsLooseEqualityOp(comp->jsop())) { + auto* lir = new (alloc()) LIsNullOrLikeUndefinedAndBranchV( + comp, ifTrue, ifFalse, useBox(left), temp(), tempToUnbox()); + add(lir, test); + return; + } + + if (comp->compareType() == MCompare::Compare_Null) { + auto* lir = + new (alloc()) LIsNullAndBranch(comp, ifTrue, ifFalse, useBox(left)); + add(lir, test); + return; + } + + auto* lir = new (alloc()) + LIsUndefinedAndBranch(comp, ifTrue, ifFalse, useBox(left)); + add(lir, test); + return; + } + + // Compare and branch Int32, Symbol, Object, or RefOrNull pointers. + if (comp->isInt32Comparison() || + comp->compareType() == MCompare::Compare_UInt32 || + comp->compareType() == MCompare::Compare_UIntPtr || + comp->compareType() == MCompare::Compare_Object || + comp->compareType() == MCompare::Compare_Symbol || + comp->compareType() == MCompare::Compare_RefOrNull) { + JSOp op = ReorderComparison(comp->jsop(), &left, &right); + LAllocation lhs = useRegister(left); + LAllocation rhs; + if (comp->isInt32Comparison() || + comp->compareType() == MCompare::Compare_UInt32 || + comp->compareType() == MCompare::Compare_UIntPtr) { + rhs = useAnyOrInt32Constant(right); + } else { + rhs = useAny(right); + } + LCompareAndBranch* lir = + new (alloc()) LCompareAndBranch(comp, op, lhs, rhs, ifTrue, ifFalse); + add(lir, test); + return; + } + + // Compare and branch Int64. + if (comp->compareType() == MCompare::Compare_Int64 || + comp->compareType() == MCompare::Compare_UInt64) { + JSOp op = ReorderComparison(comp->jsop(), &left, &right); + lowerForCompareI64AndBranch(test, comp, op, left, right, ifTrue, ifFalse); + return; + } + + // Compare and branch doubles. + if (comp->isDoubleComparison()) { + LAllocation lhs = useRegister(left); + LAllocation rhs = useRegister(right); + LCompareDAndBranch* lir = + new (alloc()) LCompareDAndBranch(comp, lhs, rhs, ifTrue, ifFalse); + add(lir, test); + return; + } + + // Compare and branch floats. + if (comp->isFloat32Comparison()) { + LAllocation lhs = useRegister(left); + LAllocation rhs = useRegister(right); + LCompareFAndBranch* lir = + new (alloc()) LCompareFAndBranch(comp, lhs, rhs, ifTrue, ifFalse); + add(lir, test); + return; + } + } + + // Check if the operand for this test is a bitand operation. If it is, we want + // to emit an LBitAndAndBranch rather than an LTest*AndBranch. + if (opd->isBitAnd() && opd->isEmittedAtUses()) { + MDefinition* lhs = opd->getOperand(0); + MDefinition* rhs = opd->getOperand(1); + if (lhs->type() == MIRType::Int32 && rhs->type() == MIRType::Int32) { + ReorderCommutative(&lhs, &rhs, test); + lowerForBitAndAndBranch(new (alloc()) LBitAndAndBranch(ifTrue, ifFalse, + /*is64=*/false), + test, lhs, rhs); + return; + } + } + +#if defined(ENABLE_WASM_SIMD) && \ + (defined(JS_CODEGEN_X86) || defined(JS_CODEGEN_X64) || \ + defined(JS_CODEGEN_ARM64)) + // Check if the operand for this test is an any_true/all_true SIMD operation. + // If it is, we want to emit an LWasmReduceAndBranchSimd128 node to avoid + // generating an intermediate boolean result. + if (opd->isWasmReduceSimd128() && opd->isEmittedAtUses()) { + MWasmReduceSimd128* node = opd->toWasmReduceSimd128(); + if (canFoldReduceSimd128AndBranch(node->simdOp())) { +# ifdef DEBUG + js::wasm::ReportSimdAnalysis("simd128-to-scalar-and-branch -> folded"); +# endif + auto* lir = new (alloc()) LWasmReduceAndBranchSimd128( + useRegister(node->input()), node->simdOp(), ifTrue, ifFalse); + add(lir, test); + return; + } + } +#endif + + if (opd->isIsObject() && opd->isEmittedAtUses()) { + MDefinition* input = opd->toIsObject()->input(); + MOZ_ASSERT(input->type() == MIRType::Value); + + LIsObjectAndBranch* lir = + new (alloc()) LIsObjectAndBranch(ifTrue, ifFalse, useBoxAtStart(input)); + add(lir, test); + return; + } + + if (opd->isWasmGcObjectIsSubtypeOfAbstract() && opd->isEmittedAtUses()) { + MWasmGcObjectIsSubtypeOfAbstract* isSubTypeOf = + opd->toWasmGcObjectIsSubtypeOfAbstract(); + LAllocation object = useRegister(isSubTypeOf->object()); + // As in visitWasmGcObjectIsSubtypeOfAbstract, we know we do not need + // scratch2 and superSuperTypeVector because we know this is not a concrete + // type. + LDefinition scratch1 = MacroAssembler::needScratch1ForBranchWasmGcRefType( + isSubTypeOf->destType()) + ? temp() + : LDefinition(); + add(new (alloc()) LWasmGcObjectIsSubtypeOfAbstractAndBranch( + ifTrue, ifFalse, isSubTypeOf->sourceType(), isSubTypeOf->destType(), + object, scratch1), + test); + return; + } + + if (opd->isWasmGcObjectIsSubtypeOfConcrete() && opd->isEmittedAtUses()) { + MWasmGcObjectIsSubtypeOfConcrete* isSubTypeOf = + opd->toWasmGcObjectIsSubtypeOfConcrete(); + LAllocation object = useRegister(isSubTypeOf->object()); + // As in visitWasmGcObjectIsSubtypeOfConcrete, we know we need scratch1 and + // superSuperTypeVector because we know this is a concrete type. + LAllocation superSuperTypeVector = + useRegister(isSubTypeOf->superSuperTypeVector()); + LDefinition scratch1 = temp(); + LDefinition scratch2 = MacroAssembler::needScratch2ForBranchWasmGcRefType( + isSubTypeOf->destType()) + ? temp() + : LDefinition(); + add(new (alloc()) LWasmGcObjectIsSubtypeOfConcreteAndBranch( + ifTrue, ifFalse, isSubTypeOf->sourceType(), isSubTypeOf->destType(), + object, superSuperTypeVector, scratch1, scratch2), + test); + return; + } + + if (opd->isIsNullOrUndefined() && opd->isEmittedAtUses()) { + MIsNullOrUndefined* isNullOrUndefined = opd->toIsNullOrUndefined(); + MDefinition* input = isNullOrUndefined->value(); + MOZ_ASSERT(input->type() == MIRType::Value); + + auto* lir = new (alloc()) LIsNullOrUndefinedAndBranch( + isNullOrUndefined, ifTrue, ifFalse, useBoxAtStart(input)); + add(lir, test); + return; + } + + if (opd->isIsNoIter()) { + MOZ_ASSERT(opd->isEmittedAtUses()); + + MDefinition* input = opd->toIsNoIter()->input(); + MOZ_ASSERT(input->type() == MIRType::Value); + + LIsNoIterAndBranch* lir = + new (alloc()) LIsNoIterAndBranch(ifTrue, ifFalse, useBox(input)); + add(lir, test); + return; + } + + if (opd->isIteratorHasIndices()) { + MOZ_ASSERT(opd->isEmittedAtUses()); + + MDefinition* object = opd->toIteratorHasIndices()->object(); + MDefinition* iterator = opd->toIteratorHasIndices()->iterator(); + LIteratorHasIndicesAndBranch* lir = new (alloc()) + LIteratorHasIndicesAndBranch(ifTrue, ifFalse, useRegister(object), + useRegister(iterator), temp(), temp()); + add(lir, test); + return; + } + + switch (opd->type()) { + case MIRType::Double: + add(new (alloc()) LTestDAndBranch(useRegister(opd), ifTrue, ifFalse)); + break; + case MIRType::Float32: + add(new (alloc()) LTestFAndBranch(useRegister(opd), ifTrue, ifFalse)); + break; + case MIRType::Int32: + case MIRType::Boolean: + add(new (alloc()) LTestIAndBranch(useRegister(opd), ifTrue, ifFalse)); + break; + case MIRType::Int64: + add(new (alloc()) + LTestI64AndBranch(useInt64Register(opd), ifTrue, ifFalse)); + break; + case MIRType::BigInt: + add(new (alloc()) LTestBIAndBranch(useRegister(opd), ifTrue, ifFalse)); + break; + default: + MOZ_CRASH("Bad type"); + } +} + +static inline bool CanEmitCompareAtUses(MInstruction* ins) { + if (!ins->canEmitAtUses()) { + return false; + } + + // If the result is never used, we can usefully defer emission to the use + // point, since that will never happen. + MUseIterator iter(ins->usesBegin()); + if (iter == ins->usesEnd()) { + return true; + } + + // If the first use isn't of the expected form, the answer is No. + MNode* node = iter->consumer(); + if (!node->isDefinition()) { + return false; + } + + MDefinition* use = node->toDefinition(); + if (!use->isTest() && !use->isWasmSelect()) { + return false; + } + + // Emission can be deferred to the first use point, but only if there are no + // other use points. + iter++; + return iter == ins->usesEnd(); +} + +static bool CanCompareCharactersInline(const JSLinearString* linear) { + size_t length = linear->length(); + + // Limit the number of inline instructions used for character comparisons. Use + // the same instruction limit for both encodings, i.e. two-byte uses half the + // limit of Latin-1 strings. + constexpr size_t Latin1StringCompareCutoff = 32; + constexpr size_t TwoByteStringCompareCutoff = 16; + + return length > 0 && + (linear->hasLatin1Chars() ? length <= Latin1StringCompareCutoff + : length <= TwoByteStringCompareCutoff); +} + +void LIRGenerator::visitCompare(MCompare* comp) { + MDefinition* left = comp->lhs(); + MDefinition* right = comp->rhs(); + + // Try to fold the comparison so that we don't have to handle all cases. + bool result; + if (comp->tryFold(&result)) { + define(new (alloc()) LInteger(result), comp); + return; + } + + // Move below the emitAtUses call if we ever implement + // LCompareSAndBranch. Doing this now wouldn't be wrong, but doesn't + // make sense and avoids confusion. + if (comp->compareType() == MCompare::Compare_String) { + if (IsEqualityOp(comp->jsop())) { + MConstant* constant = nullptr; + if (left->isConstant()) { + constant = left->toConstant(); + } else if (right->isConstant()) { + constant = right->toConstant(); + } + + if (constant) { + JSLinearString* linear = &constant->toString()->asLinear(); + + if (CanCompareCharactersInline(linear)) { + MDefinition* input = left->isConstant() ? right : left; + + auto* lir = new (alloc()) LCompareSInline(useRegister(input), linear); + define(lir, comp); + assignSafepoint(lir, comp); + return; + } + } + } + + LCompareS* lir = + new (alloc()) LCompareS(useRegister(left), useRegister(right)); + define(lir, comp); + assignSafepoint(lir, comp); + return; + } + + // Compare two BigInts. + if (comp->compareType() == MCompare::Compare_BigInt) { + auto* lir = new (alloc()) LCompareBigInt( + useRegister(left), useRegister(right), temp(), temp(), temp()); + define(lir, comp); + return; + } + + // Compare BigInt with Int32. + if (comp->compareType() == MCompare::Compare_BigInt_Int32) { + auto* lir = new (alloc()) LCompareBigIntInt32( + useRegister(left), useRegister(right), temp(), temp()); + define(lir, comp); + return; + } + + // Compare BigInt with Double. + if (comp->compareType() == MCompare::Compare_BigInt_Double) { + auto* lir = new (alloc()) LCompareBigIntDouble(useRegisterAtStart(left), + useRegisterAtStart(right)); + defineReturn(lir, comp); + return; + } + + // Compare BigInt with String. + if (comp->compareType() == MCompare::Compare_BigInt_String) { + auto* lir = new (alloc()) LCompareBigIntString(useRegisterAtStart(left), + useRegisterAtStart(right)); + defineReturn(lir, comp); + assignSafepoint(lir, comp); + return; + } + + // Sniff out if the output of this compare is used only for a branching. + // If it is, then we will emit an LCompare*AndBranch instruction in place + // of this compare and any test that uses this compare. Thus, we can + // ignore this Compare. + if (CanEmitCompareAtUses(comp)) { + emitAtUses(comp); + return; + } + + // Compare Null and Undefined. + if (comp->compareType() == MCompare::Compare_Null || + comp->compareType() == MCompare::Compare_Undefined) { + if (left->type() == MIRType::Object) { + define(new (alloc()) LIsNullOrLikeUndefinedT(useRegister(left)), comp); + return; + } + + if (IsLooseEqualityOp(comp->jsop())) { + auto* lir = + new (alloc()) LIsNullOrLikeUndefinedV(useBox(left), tempToUnbox()); + define(lir, comp); + return; + } + + if (comp->compareType() == MCompare::Compare_Null) { + auto* lir = new (alloc()) LIsNull(useBox(left)); + define(lir, comp); + return; + } + + auto* lir = new (alloc()) LIsUndefined(useBox(left)); + define(lir, comp); + return; + } + + // Compare Int32, Symbol, Object or Wasm pointers. + if (comp->isInt32Comparison() || + comp->compareType() == MCompare::Compare_UInt32 || + comp->compareType() == MCompare::Compare_UIntPtr || + comp->compareType() == MCompare::Compare_Object || + comp->compareType() == MCompare::Compare_Symbol || + comp->compareType() == MCompare::Compare_RefOrNull) { + JSOp op = ReorderComparison(comp->jsop(), &left, &right); + LAllocation lhs = useRegister(left); + LAllocation rhs; + if (comp->isInt32Comparison() || + comp->compareType() == MCompare::Compare_UInt32 || + comp->compareType() == MCompare::Compare_UIntPtr) { + rhs = useAnyOrInt32Constant(right); + } else { + rhs = useAny(right); + } + define(new (alloc()) LCompare(op, lhs, rhs), comp); + return; + } + + // Compare Int64. + if (comp->compareType() == MCompare::Compare_Int64 || + comp->compareType() == MCompare::Compare_UInt64) { + JSOp op = ReorderComparison(comp->jsop(), &left, &right); + define(new (alloc()) LCompareI64(op, useInt64Register(left), + useInt64OrConstant(right)), + comp); + return; + } + + // Compare doubles. + if (comp->isDoubleComparison()) { + define(new (alloc()) LCompareD(useRegister(left), useRegister(right)), + comp); + return; + } + + // Compare float32. + if (comp->isFloat32Comparison()) { + define(new (alloc()) LCompareF(useRegister(left), useRegister(right)), + comp); + return; + } + + MOZ_CRASH("Unrecognized compare type."); +} + +void LIRGenerator::visitSameValueDouble(MSameValueDouble* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + + MOZ_ASSERT(lhs->type() == MIRType::Double); + MOZ_ASSERT(rhs->type() == MIRType::Double); + + auto* lir = new (alloc()) + LSameValueDouble(useRegister(lhs), useRegister(rhs), tempDouble()); + define(lir, ins); +} + +void LIRGenerator::visitSameValue(MSameValue* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + + MOZ_ASSERT(lhs->type() == MIRType::Value); + MOZ_ASSERT(rhs->type() == MIRType::Value); + + auto* lir = new (alloc()) LSameValue(useBox(lhs), useBox(rhs)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::lowerBitOp(JSOp op, MBinaryInstruction* ins) { + MDefinition* lhs = ins->getOperand(0); + MDefinition* rhs = ins->getOperand(1); + MOZ_ASSERT(IsIntType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + MOZ_ASSERT(rhs->type() == MIRType::Int32); + ReorderCommutative(&lhs, &rhs, ins); + lowerForALU(new (alloc()) LBitOpI(op), ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + MOZ_ASSERT(rhs->type() == MIRType::Int64); + ReorderCommutative(&lhs, &rhs, ins); + lowerForALUInt64(new (alloc()) LBitOpI64(op), ins, lhs, rhs); + return; + } + + MOZ_CRASH("Unhandled integer specialization"); +} + +void LIRGenerator::visitTypeOf(MTypeOf* ins) { + MDefinition* opd = ins->input(); + + if (opd->type() == MIRType::Object) { + auto* lir = new (alloc()) LTypeOfO(useRegister(opd)); + define(lir, ins); + return; + } + + MOZ_ASSERT(opd->type() == MIRType::Value); + + LTypeOfV* lir = new (alloc()) LTypeOfV(useBox(opd), tempToUnbox()); + define(lir, ins); +} + +void LIRGenerator::visitTypeOfName(MTypeOfName* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::Int32); + + auto* lir = new (alloc()) LTypeOfName(useRegister(input)); + define(lir, ins); +} + +void LIRGenerator::visitTypeOfIs(MTypeOfIs* ins) { + MDefinition* input = ins->input(); + + MOZ_ASSERT(input->type() == MIRType::Object || + input->type() == MIRType::Value); + + switch (ins->jstype()) { + case JSTYPE_UNDEFINED: + case JSTYPE_OBJECT: + case JSTYPE_FUNCTION: { + if (input->type() == MIRType::Object) { + auto* lir = new (alloc()) LTypeOfIsNonPrimitiveO(useRegister(input)); + define(lir, ins); + } else { + auto* lir = + new (alloc()) LTypeOfIsNonPrimitiveV(useBox(input), tempToUnbox()); + define(lir, ins); + } + return; + } + + case JSTYPE_STRING: + case JSTYPE_NUMBER: + case JSTYPE_BOOLEAN: + case JSTYPE_SYMBOL: + case JSTYPE_BIGINT: { + MOZ_ASSERT(input->type() == MIRType::Value); + + auto* lir = new (alloc()) LTypeOfIsPrimitive(useBoxAtStart(input)); + define(lir, ins); + return; + } + +#ifdef ENABLE_RECORD_TUPLE + case JSTYPE_RECORD: + case JSTYPE_TUPLE: +#endif + case JSTYPE_LIMIT: + break; + } + MOZ_CRASH("Unhandled JSType"); +} + +void LIRGenerator::visitToAsyncIter(MToAsyncIter* ins) { + LToAsyncIter* lir = new (alloc()) LToAsyncIter( + useRegisterAtStart(ins->iterator()), useBoxAtStart(ins->nextMethod())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitToPropertyKeyCache(MToPropertyKeyCache* ins) { + MDefinition* input = ins->getOperand(0); + MOZ_ASSERT(ins->type() == MIRType::Value); + + auto* lir = new (alloc()) LToPropertyKeyCache(useBox(input)); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBitNot(MBitNot* ins) { + MDefinition* input = ins->getOperand(0); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(input->type() == MIRType::Int32); + lowerForALU(new (alloc()) LBitNotI(), ins, input); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(input->type() == MIRType::Int64); + lowerForALUInt64(new (alloc()) LBitNotI64(), ins, input); + return; + } + + MOZ_CRASH("Unhandled integer specialization"); +} + +static bool CanEmitBitAndAtUses(MInstruction* ins) { + if (!ins->canEmitAtUses()) { + return false; + } + + MIRType tyL = ins->getOperand(0)->type(); + MIRType tyR = ins->getOperand(1)->type(); + if (tyL != tyR || (tyL != MIRType::Int32 && tyL != MIRType::Int64)) { + return false; + } + + MUseIterator iter(ins->usesBegin()); + if (iter == ins->usesEnd()) { + return false; + } + + MNode* node = iter->consumer(); + if (!node->isDefinition() || !node->toDefinition()->isInstruction()) { + return false; + } + + MInstruction* use = node->toDefinition()->toInstruction(); + if (!use->isTest() && !(use->isCompare() && CanEmitCompareAtUses(use))) { + return false; + } + + iter++; + return iter == ins->usesEnd(); +} + +void LIRGenerator::visitBitAnd(MBitAnd* ins) { + // Sniff out if the output of this bitand is used only for a branching. + // If it is, then we will emit an LBitAndAndBranch instruction in place + // of this bitand and any test that uses this bitand. Thus, we can + // ignore this BitAnd. + if (CanEmitBitAndAtUses(ins)) { + emitAtUses(ins); + } else { + lowerBitOp(JSOp::BitAnd, ins); + } +} + +void LIRGenerator::visitBitOr(MBitOr* ins) { lowerBitOp(JSOp::BitOr, ins); } + +void LIRGenerator::visitBitXor(MBitXor* ins) { lowerBitOp(JSOp::BitXor, ins); } + +void LIRGenerator::visitWasmBinaryBitwise(MWasmBinaryBitwise* ins) { + switch (ins->subOpcode()) { + case MWasmBinaryBitwise::SubOpcode::And: + if (CanEmitBitAndAtUses(ins)) { + emitAtUses(ins); + } else { + lowerBitOp(JSOp::BitAnd, ins); + } + break; + case MWasmBinaryBitwise::SubOpcode::Or: + lowerBitOp(JSOp::BitOr, ins); + break; + case MWasmBinaryBitwise::SubOpcode::Xor: + lowerBitOp(JSOp::BitXor, ins); + break; + default: + MOZ_CRASH(); + } +} + +void LIRGenerator::lowerShiftOp(JSOp op, MShiftInstruction* ins) { + MDefinition* lhs = ins->getOperand(0); + MDefinition* rhs = ins->getOperand(1); + + if (op == JSOp::Ursh && ins->type() == MIRType::Double) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + MOZ_ASSERT(rhs->type() == MIRType::Int32); + lowerUrshD(ins->toUrsh()); + return; + } + + MOZ_ASSERT(IsIntType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + MOZ_ASSERT(rhs->type() == MIRType::Int32); + + LShiftI* lir = new (alloc()) LShiftI(op); + if (op == JSOp::Ursh) { + if (ins->toUrsh()->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + } + lowerForShift(lir, ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + MOZ_ASSERT(rhs->type() == MIRType::Int64); + lowerForShiftInt64(new (alloc()) LShiftI64(op), ins, lhs, rhs); + return; + } + + MOZ_CRASH("Unhandled integer specialization"); +} + +void LIRGenerator::visitLsh(MLsh* ins) { lowerShiftOp(JSOp::Lsh, ins); } + +void LIRGenerator::visitRsh(MRsh* ins) { lowerShiftOp(JSOp::Rsh, ins); } + +void LIRGenerator::visitUrsh(MUrsh* ins) { lowerShiftOp(JSOp::Ursh, ins); } + +void LIRGenerator::visitSignExtendInt32(MSignExtendInt32* ins) { + LInstructionHelper<1, 1, 0>* lir; + + if (ins->mode() == MSignExtendInt32::Byte) { + lir = new (alloc()) + LSignExtendInt32(useByteOpRegisterAtStart(ins->input()), ins->mode()); + } else { + lir = new (alloc()) + LSignExtendInt32(useRegisterAtStart(ins->input()), ins->mode()); + } + + define(lir, ins); +} + +void LIRGenerator::visitRotate(MRotate* ins) { + MDefinition* input = ins->input(); + MDefinition* count = ins->count(); + + if (ins->type() == MIRType::Int32) { + auto* lir = new (alloc()) LRotate(); + lowerForShift(lir, ins, input, count); + } else if (ins->type() == MIRType::Int64) { + auto* lir = new (alloc()) LRotateI64(); + lowerForShiftInt64(lir, ins, input, count); + } else { + MOZ_CRASH("unexpected type in visitRotate"); + } +} + +void LIRGenerator::visitFloor(MFloor* ins) { + MIRType type = ins->input()->type(); + MOZ_ASSERT(IsFloatingPointType(type)); + + LInstructionHelper<1, 1, 0>* lir; + if (type == MIRType::Double) { + lir = new (alloc()) LFloor(useRegister(ins->input())); + } else { + lir = new (alloc()) LFloorF(useRegister(ins->input())); + } + + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitCeil(MCeil* ins) { + MIRType type = ins->input()->type(); + MOZ_ASSERT(IsFloatingPointType(type)); + + LInstructionHelper<1, 1, 0>* lir; + if (type == MIRType::Double) { + lir = new (alloc()) LCeil(useRegister(ins->input())); + } else { + lir = new (alloc()) LCeilF(useRegister(ins->input())); + } + + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitRound(MRound* ins) { + MIRType type = ins->input()->type(); + MOZ_ASSERT(IsFloatingPointType(type)); + + LInstructionHelper<1, 1, 1>* lir; + if (type == MIRType::Double) { + lir = new (alloc()) LRound(useRegister(ins->input()), tempDouble()); + } else { + lir = new (alloc()) LRoundF(useRegister(ins->input()), tempFloat32()); + } + + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitTrunc(MTrunc* ins) { + MIRType type = ins->input()->type(); + MOZ_ASSERT(IsFloatingPointType(type)); + + LInstructionHelper<1, 1, 0>* lir; + if (type == MIRType::Double) { + lir = new (alloc()) LTrunc(useRegister(ins->input())); + } else { + lir = new (alloc()) LTruncF(useRegister(ins->input())); + } + + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitNearbyInt(MNearbyInt* ins) { + MIRType inputType = ins->input()->type(); + MOZ_ASSERT(IsFloatingPointType(inputType)); + MOZ_ASSERT(ins->type() == inputType); + + LInstructionHelper<1, 1, 0>* lir; + if (inputType == MIRType::Double) { + lir = new (alloc()) LNearbyInt(useRegisterAtStart(ins->input())); + } else { + lir = new (alloc()) LNearbyIntF(useRegisterAtStart(ins->input())); + } + + define(lir, ins); +} + +void LIRGenerator::visitMinMax(MMinMax* ins) { + MDefinition* first = ins->getOperand(0); + MDefinition* second = ins->getOperand(1); + + ReorderCommutative(&first, &second, ins); + + LMinMaxBase* lir; + switch (ins->type()) { + case MIRType::Int32: + lir = new (alloc()) + LMinMaxI(useRegisterAtStart(first), useRegisterOrConstant(second)); + break; + case MIRType::Float32: + lir = new (alloc()) + LMinMaxF(useRegisterAtStart(first), useRegister(second)); + break; + case MIRType::Double: + lir = new (alloc()) + LMinMaxD(useRegisterAtStart(first), useRegister(second)); + break; + default: + MOZ_CRASH(); + } + + // Input reuse is OK (for now) even on ARM64: floating min/max are fairly + // expensive due to SNaN -> QNaN conversion, and int min/max is for asm.js. + defineReuseInput(lir, ins, 0); +} + +void LIRGenerator::visitMinMaxArray(MMinMaxArray* ins) { + LInstructionHelper<1, 1, 3>* lir; + if (ins->type() == MIRType::Int32) { + lir = new (alloc()) + LMinMaxArrayI(useRegisterAtStart(ins->array()), temp(), temp(), temp()); + } else { + MOZ_ASSERT(ins->type() == MIRType::Double); + lir = new (alloc()) LMinMaxArrayD(useRegisterAtStart(ins->array()), + tempDouble(), temp(), temp()); + } + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +LInstructionHelper<1, 1, 0>* LIRGenerator::allocateAbs(MAbs* ins, + LAllocation input) { + MDefinition* num = ins->input(); + MOZ_ASSERT(IsNumberType(num->type())); + + LInstructionHelper<1, 1, 0>* lir; + switch (num->type()) { + case MIRType::Int32: + lir = new (alloc()) LAbsI(input); + // needed to handle abs(INT32_MIN) + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + break; + case MIRType::Float32: + lir = new (alloc()) LAbsF(input); + break; + case MIRType::Double: + lir = new (alloc()) LAbsD(input); + break; + default: + MOZ_CRASH(); + } + return lir; +} + +void LIRGenerator::visitClz(MClz* ins) { + MDefinition* num = ins->num(); + + MOZ_ASSERT(IsIntType(ins->type())); + + if (ins->type() == MIRType::Int32) { + LClzI* lir = new (alloc()) LClzI(useRegisterAtStart(num)); + define(lir, ins); + return; + } + + auto* lir = new (alloc()) LClzI64(useInt64RegisterAtStart(num)); + defineInt64(lir, ins); +} + +void LIRGenerator::visitCtz(MCtz* ins) { + MDefinition* num = ins->num(); + + MOZ_ASSERT(IsIntType(ins->type())); + + if (ins->type() == MIRType::Int32) { + LCtzI* lir = new (alloc()) LCtzI(useRegisterAtStart(num)); + define(lir, ins); + return; + } + + auto* lir = new (alloc()) LCtzI64(useInt64RegisterAtStart(num)); + defineInt64(lir, ins); +} + +void LIRGenerator::visitPopcnt(MPopcnt* ins) { + MDefinition* num = ins->num(); + + MOZ_ASSERT(IsIntType(ins->type())); + + if (ins->type() == MIRType::Int32) { + LPopcntI* lir = new (alloc()) LPopcntI(useRegisterAtStart(num), temp()); + define(lir, ins); + return; + } + + auto* lir = new (alloc()) LPopcntI64(useInt64RegisterAtStart(num), temp()); + defineInt64(lir, ins); +} + +void LIRGenerator::visitSqrt(MSqrt* ins) { + MDefinition* num = ins->input(); + MOZ_ASSERT(IsFloatingPointType(num->type())); + + LInstructionHelper<1, 1, 0>* lir; + if (num->type() == MIRType::Double) { + lir = new (alloc()) LSqrtD(useRegisterAtStart(num)); + } else { + lir = new (alloc()) LSqrtF(useRegisterAtStart(num)); + } + define(lir, ins); +} + +void LIRGenerator::visitAtan2(MAtan2* ins) { + MDefinition* y = ins->y(); + MOZ_ASSERT(y->type() == MIRType::Double); + + MDefinition* x = ins->x(); + MOZ_ASSERT(x->type() == MIRType::Double); + + LAtan2D* lir = + new (alloc()) LAtan2D(useRegisterAtStart(y), useRegisterAtStart(x)); + defineReturn(lir, ins); +} + +void LIRGenerator::visitHypot(MHypot* ins) { + LHypot* lir = nullptr; + uint32_t length = ins->numOperands(); + for (uint32_t i = 0; i < length; ++i) { + MOZ_ASSERT(ins->getOperand(i)->type() == MIRType::Double); + } + + switch (length) { + case 2: + lir = new (alloc()) LHypot(useRegisterAtStart(ins->getOperand(0)), + useRegisterAtStart(ins->getOperand(1))); + break; + case 3: + lir = new (alloc()) LHypot(useRegisterAtStart(ins->getOperand(0)), + useRegisterAtStart(ins->getOperand(1)), + useRegisterAtStart(ins->getOperand(2))); + break; + case 4: + lir = new (alloc()) LHypot(useRegisterAtStart(ins->getOperand(0)), + useRegisterAtStart(ins->getOperand(1)), + useRegisterAtStart(ins->getOperand(2)), + useRegisterAtStart(ins->getOperand(3))); + break; + default: + MOZ_CRASH("Unexpected number of arguments to LHypot."); + } + + defineReturn(lir, ins); +} + +void LIRGenerator::visitPow(MPow* ins) { + MDefinition* input = ins->input(); + MDefinition* power = ins->power(); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(input->type() == MIRType::Int32); + MOZ_ASSERT(power->type() == MIRType::Int32); + + if (input->isConstant()) { + // Restrict this optimization to |base <= 256| to avoid generating too + // many consecutive shift instructions. + int32_t base = input->toConstant()->toInt32(); + if (2 <= base && base <= 256 && mozilla::IsPowerOfTwo(uint32_t(base))) { + lowerPowOfTwoI(ins); + return; + } + } + + auto* lir = new (alloc()) + LPowII(useRegister(input), useRegister(power), temp(), temp()); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + return; + } + + MOZ_ASSERT(ins->type() == MIRType::Double); + MOZ_ASSERT(input->type() == MIRType::Double); + MOZ_ASSERT(power->type() == MIRType::Int32 || + power->type() == MIRType::Double); + + LInstruction* lir; + if (power->type() == MIRType::Int32) { + lir = new (alloc()) + LPowI(useRegisterAtStart(input), useRegisterAtStart(power)); + } else { + lir = new (alloc()) + LPowD(useRegisterAtStart(input), useRegisterAtStart(power)); + } + defineReturn(lir, ins); +} + +void LIRGenerator::visitSign(MSign* ins) { + if (ins->type() == ins->input()->type()) { + LInstructionHelper<1, 1, 0>* lir; + if (ins->type() == MIRType::Int32) { + lir = new (alloc()) LSignI(useRegister(ins->input())); + } else { + MOZ_ASSERT(ins->type() == MIRType::Double); + lir = new (alloc()) LSignD(useRegister(ins->input())); + } + define(lir, ins); + } else { + MOZ_ASSERT(ins->type() == MIRType::Int32); + MOZ_ASSERT(ins->input()->type() == MIRType::Double); + + auto* lir = new (alloc()) LSignDI(useRegister(ins->input()), tempDouble()); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + } +} + +void LIRGenerator::visitMathFunction(MMathFunction* ins) { + MOZ_ASSERT(IsFloatingPointType(ins->type())); + MOZ_ASSERT(ins->type() == ins->input()->type()); + + LInstruction* lir; + if (ins->type() == MIRType::Double) { + lir = new (alloc()) LMathFunctionD(useRegisterAtStart(ins->input())); + } else { + lir = new (alloc()) LMathFunctionF(useRegisterAtStart(ins->input())); + } + defineReturn(lir, ins); +} + +void LIRGenerator::visitRandom(MRandom* ins) { + auto* lir = new (alloc()) LRandom(temp(), tempInt64(), tempInt64()); + define(lir, ins); +} + +// Try to mark an add or sub instruction as able to recover its input when +// bailing out. +template +static void MaybeSetRecoversInput(S* mir, T* lir) { + MOZ_ASSERT(lir->mirRaw() == mir); + if (!mir->fallible() || !lir->snapshot()) { + return; + } + + if (lir->output()->policy() != LDefinition::MUST_REUSE_INPUT) { + return; + } + + // The original operands to an add or sub can't be recovered if they both + // use the same register. + if (lir->lhs()->isUse() && lir->rhs()->isUse() && + lir->lhs()->toUse()->virtualRegister() == + lir->rhs()->toUse()->virtualRegister()) { + return; + } + + // Add instructions that are on two different values can recover + // the input they clobbered via MUST_REUSE_INPUT. Thus, a copy + // of that input does not need to be kept alive in the snapshot + // for the instruction. + + lir->setRecoversInput(); + + const LUse* input = lir->getOperand(lir->output()->getReusedInput())->toUse(); + lir->snapshot()->rewriteRecoveredInput(*input); +} + +void LIRGenerator::visitAdd(MAdd* ins) { + MDefinition* lhs = ins->getOperand(0); + MDefinition* rhs = ins->getOperand(1); + + MOZ_ASSERT(lhs->type() == rhs->type()); + MOZ_ASSERT(IsNumberType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + ReorderCommutative(&lhs, &rhs, ins); + LAddI* lir = new (alloc()) LAddI; + + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + + lowerForALU(lir, ins, lhs, rhs); + MaybeSetRecoversInput(ins, lir); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + ReorderCommutative(&lhs, &rhs, ins); + LAddI64* lir = new (alloc()) LAddI64; + lowerForALUInt64(lir, ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Double) { + MOZ_ASSERT(lhs->type() == MIRType::Double); + ReorderCommutative(&lhs, &rhs, ins); + lowerForFPU(new (alloc()) LMathD(JSOp::Add), ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Float32) { + MOZ_ASSERT(lhs->type() == MIRType::Float32); + ReorderCommutative(&lhs, &rhs, ins); + lowerForFPU(new (alloc()) LMathF(JSOp::Add), ins, lhs, rhs); + return; + } + + MOZ_CRASH("Unhandled number specialization"); +} + +void LIRGenerator::visitSub(MSub* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + + MOZ_ASSERT(lhs->type() == rhs->type()); + MOZ_ASSERT(IsNumberType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + + LSubI* lir = new (alloc()) LSubI; + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + + // If our LHS is a constant 0 and we don't have to worry about results that + // can't be represented as an int32, we can optimize to an LNegI. + if (!ins->fallible() && lhs->isConstant() && + lhs->toConstant()->toInt32() == 0) { + lowerNegI(ins, rhs); + return; + } + + lowerForALU(lir, ins, lhs, rhs); + MaybeSetRecoversInput(ins, lir); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + + // If our LHS is a constant 0, we can optimize to an LNegI64. + if (lhs->isConstant() && lhs->toConstant()->toInt64() == 0) { + lowerNegI64(ins, rhs); + return; + } + + LSubI64* lir = new (alloc()) LSubI64; + lowerForALUInt64(lir, ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Double) { + MOZ_ASSERT(lhs->type() == MIRType::Double); + lowerForFPU(new (alloc()) LMathD(JSOp::Sub), ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Float32) { + MOZ_ASSERT(lhs->type() == MIRType::Float32); + lowerForFPU(new (alloc()) LMathF(JSOp::Sub), ins, lhs, rhs); + return; + } + + MOZ_CRASH("Unhandled number specialization"); +} + +void LIRGenerator::visitMul(MMul* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + MOZ_ASSERT(lhs->type() == rhs->type()); + MOZ_ASSERT(IsNumberType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + ReorderCommutative(&lhs, &rhs, ins); + + // If our RHS is a constant -1 and we don't have to worry about results that + // can't be represented as an int32, we can optimize to an LNegI. + if (!ins->fallible() && rhs->isConstant() && + rhs->toConstant()->toInt32() == -1) { + lowerNegI(ins, lhs); + return; + } + + lowerMulI(ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + ReorderCommutative(&lhs, &rhs, ins); + + // If our RHS is a constant -1, we can optimize to an LNegI64. + if (rhs->isConstant() && rhs->toConstant()->toInt64() == -1) { + lowerNegI64(ins, lhs); + return; + } + + LMulI64* lir = new (alloc()) LMulI64; + lowerForMulInt64(lir, ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Double) { + MOZ_ASSERT(lhs->type() == MIRType::Double); + ReorderCommutative(&lhs, &rhs, ins); + + // If our RHS is a constant -1.0, we can optimize to an LNegD. + if (!ins->mustPreserveNaN() && rhs->isConstant() && + rhs->toConstant()->toDouble() == -1.0) { + defineReuseInput(new (alloc()) LNegD(useRegisterAtStart(lhs)), ins, 0); + return; + } + + lowerForFPU(new (alloc()) LMathD(JSOp::Mul), ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Float32) { + MOZ_ASSERT(lhs->type() == MIRType::Float32); + ReorderCommutative(&lhs, &rhs, ins); + + // We apply the same optimizations as for doubles + if (!ins->mustPreserveNaN() && rhs->isConstant() && + rhs->toConstant()->toFloat32() == -1.0f) { + defineReuseInput(new (alloc()) LNegF(useRegisterAtStart(lhs)), ins, 0); + return; + } + + lowerForFPU(new (alloc()) LMathF(JSOp::Mul), ins, lhs, rhs); + return; + } + + MOZ_CRASH("Unhandled number specialization"); +} + +void LIRGenerator::visitDiv(MDiv* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + MOZ_ASSERT(lhs->type() == rhs->type()); + MOZ_ASSERT(IsNumberType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + lowerDivI(ins); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + lowerDivI64(ins); + return; + } + + if (ins->type() == MIRType::Double) { + MOZ_ASSERT(lhs->type() == MIRType::Double); + lowerForFPU(new (alloc()) LMathD(JSOp::Div), ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Float32) { + MOZ_ASSERT(lhs->type() == MIRType::Float32); + lowerForFPU(new (alloc()) LMathF(JSOp::Div), ins, lhs, rhs); + return; + } + + MOZ_CRASH("Unhandled number specialization"); +} + +void LIRGenerator::visitWasmBuiltinDivI64(MWasmBuiltinDivI64* div) { + lowerWasmBuiltinDivI64(div); +} + +void LIRGenerator::visitWasmBuiltinModI64(MWasmBuiltinModI64* mod) { + lowerWasmBuiltinModI64(mod); +} + +void LIRGenerator::visitBuiltinInt64ToFloatingPoint( + MBuiltinInt64ToFloatingPoint* ins) { + lowerBuiltinInt64ToFloatingPoint(ins); +} + +void LIRGenerator::visitWasmBuiltinTruncateToInt64( + MWasmBuiltinTruncateToInt64* ins) { + lowerWasmBuiltinTruncateToInt64(ins); +} + +void LIRGenerator::visitWasmBuiltinModD(MWasmBuiltinModD* ins) { + MOZ_ASSERT(gen->compilingWasm()); + LWasmBuiltinModD* lir = new (alloc()) LWasmBuiltinModD( + useRegisterAtStart(ins->lhs()), useRegisterAtStart(ins->rhs()), + useFixedAtStart(ins->instance(), InstanceReg)); + defineReturn(lir, ins); +} + +void LIRGenerator::visitMod(MMod* ins) { + MOZ_ASSERT(ins->lhs()->type() == ins->rhs()->type()); + MOZ_ASSERT(IsNumberType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(ins->type() == MIRType::Int32); + MOZ_ASSERT(ins->lhs()->type() == MIRType::Int32); + lowerModI(ins); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(ins->type() == MIRType::Int64); + MOZ_ASSERT(ins->lhs()->type() == MIRType::Int64); + lowerModI64(ins); + return; + } + + if (ins->type() == MIRType::Double) { + MOZ_ASSERT(ins->lhs()->type() == MIRType::Double); + MOZ_ASSERT(ins->rhs()->type() == MIRType::Double); + + MOZ_ASSERT(!gen->compilingWasm()); + + if (Assembler::HasRoundInstruction(RoundingMode::TowardsZero)) { + if (ins->rhs()->isConstant()) { + double d = ins->rhs()->toConstant()->toDouble(); + int32_t div; + if (mozilla::NumberIsInt32(d, &div) && div > 0 && + mozilla::IsPowerOfTwo(uint32_t(div))) { + auto* lir = new (alloc()) LModPowTwoD(useRegister(ins->lhs()), div); + define(lir, ins); + return; + } + } + } + + LModD* lir = new (alloc()) + LModD(useRegisterAtStart(ins->lhs()), useRegisterAtStart(ins->rhs())); + defineReturn(lir, ins); + return; + } + + MOZ_CRASH("Unhandled number specialization"); +} + +void LIRGenerator::visitBigIntAdd(MBigIntAdd* ins) { + auto* lir = new (alloc()) LBigIntAdd(useRegister(ins->lhs()), + useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntSub(MBigIntSub* ins) { + auto* lir = new (alloc()) LBigIntSub(useRegister(ins->lhs()), + useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntMul(MBigIntMul* ins) { + auto* lir = new (alloc()) LBigIntMul(useRegister(ins->lhs()), + useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntDiv(MBigIntDiv* ins) { lowerBigIntDiv(ins); } + +void LIRGenerator::visitBigIntMod(MBigIntMod* ins) { lowerBigIntMod(ins); } + +void LIRGenerator::visitBigIntPow(MBigIntPow* ins) { + auto* lir = new (alloc()) LBigIntPow(useRegister(ins->lhs()), + useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntBitAnd(MBigIntBitAnd* ins) { + auto* lir = new (alloc()) LBigIntBitAnd( + useRegister(ins->lhs()), useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntBitOr(MBigIntBitOr* ins) { + auto* lir = new (alloc()) LBigIntBitOr( + useRegister(ins->lhs()), useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntBitXor(MBigIntBitXor* ins) { + auto* lir = new (alloc()) LBigIntBitXor( + useRegister(ins->lhs()), useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntLsh(MBigIntLsh* ins) { lowerBigIntLsh(ins); } + +void LIRGenerator::visitBigIntRsh(MBigIntRsh* ins) { lowerBigIntRsh(ins); } + +void LIRGenerator::visitBigIntIncrement(MBigIntIncrement* ins) { + auto* lir = + new (alloc()) LBigIntIncrement(useRegister(ins->input()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntDecrement(MBigIntDecrement* ins) { + auto* lir = + new (alloc()) LBigIntDecrement(useRegister(ins->input()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntNegate(MBigIntNegate* ins) { + auto* lir = new (alloc()) LBigIntNegate(useRegister(ins->input()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntBitNot(MBigIntBitNot* ins) { + auto* lir = + new (alloc()) LBigIntBitNot(useRegister(ins->input()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitInt32ToStringWithBase(MInt32ToStringWithBase* ins) { + MOZ_ASSERT(ins->input()->type() == MIRType::Int32); + MOZ_ASSERT(ins->base()->type() == MIRType::Int32); + + int32_t baseInt = + ins->base()->isConstant() ? ins->base()->toConstant()->toInt32() : 0; + + LAllocation base; + if (2 <= baseInt && baseInt <= 36) { + base = useRegisterOrConstant(ins->base()); + } else { + base = useRegister(ins->base()); + } + + auto* lir = new (alloc()) + LInt32ToStringWithBase(useRegister(ins->input()), base, temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNumberParseInt(MNumberParseInt* ins) { + MOZ_ASSERT(ins->string()->type() == MIRType::String); + MOZ_ASSERT(ins->radix()->type() == MIRType::Int32); + + auto* lir = new (alloc()) LNumberParseInt(useRegisterAtStart(ins->string()), + useRegisterAtStart(ins->radix()), + tempFixed(CallTempReg0)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitDoubleParseInt(MDoubleParseInt* ins) { + MOZ_ASSERT(ins->number()->type() == MIRType::Double); + + auto* lir = + new (alloc()) LDoubleParseInt(useRegister(ins->number()), tempDouble()); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitConcat(MConcat* ins) { + MDefinition* lhs = ins->getOperand(0); + MDefinition* rhs = ins->getOperand(1); + + MOZ_ASSERT(lhs->type() == MIRType::String); + MOZ_ASSERT(rhs->type() == MIRType::String); + MOZ_ASSERT(ins->type() == MIRType::String); + + LConcat* lir = new (alloc()) LConcat( + useFixedAtStart(lhs, CallTempReg0), useFixedAtStart(rhs, CallTempReg1), + tempFixed(CallTempReg0), tempFixed(CallTempReg1), tempFixed(CallTempReg2), + tempFixed(CallTempReg3), tempFixed(CallTempReg4)); + defineFixed(lir, ins, LAllocation(AnyRegister(CallTempReg5))); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitLinearizeForCharAccess(MLinearizeForCharAccess* ins) { + MDefinition* str = ins->string(); + MDefinition* idx = ins->index(); + + MOZ_ASSERT(str->type() == MIRType::String); + MOZ_ASSERT(idx->type() == MIRType::Int32); + + auto* lir = + new (alloc()) LLinearizeForCharAccess(useRegister(str), useRegister(idx)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCharCodeAt(MCharCodeAt* ins) { + MDefinition* str = ins->string(); + MDefinition* idx = ins->index(); + + MOZ_ASSERT(str->type() == MIRType::String); + MOZ_ASSERT(idx->type() == MIRType::Int32); + + LCharCodeAt* lir = new (alloc()) + LCharCodeAt(useRegister(str), useRegister(idx), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCharCodeAtMaybeOutOfBounds( + MCharCodeAtMaybeOutOfBounds* ins) { + MDefinition* str = ins->string(); + MDefinition* idx = ins->index(); + + MOZ_ASSERT(str->type() == MIRType::String); + MOZ_ASSERT(idx->type() == MIRType::Int32); + + auto* lir = new (alloc()) LCharCodeAtMaybeOutOfBounds( + useRegister(str), useRegister(idx), temp(), temp()); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCharAtMaybeOutOfBounds(MCharAtMaybeOutOfBounds* ins) { + MDefinition* str = ins->string(); + MDefinition* idx = ins->index(); + + MOZ_ASSERT(str->type() == MIRType::String); + MOZ_ASSERT(idx->type() == MIRType::Int32); + + auto* lir = new (alloc()) LCharAtMaybeOutOfBounds( + useRegister(str), useRegister(idx), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitFromCharCode(MFromCharCode* ins) { + MDefinition* code = ins->getOperand(0); + + MOZ_ASSERT(code->type() == MIRType::Int32); + + LFromCharCode* lir = new (alloc()) LFromCharCode(useRegister(code)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitFromCodePoint(MFromCodePoint* ins) { + MDefinition* codePoint = ins->getOperand(0); + + MOZ_ASSERT(codePoint->type() == MIRType::Int32); + + LFromCodePoint* lir = + new (alloc()) LFromCodePoint(useRegister(codePoint), temp(), temp()); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStringIndexOf(MStringIndexOf* ins) { + auto* string = ins->string(); + MOZ_ASSERT(string->type() == MIRType::String); + + auto* searchStr = ins->searchString(); + MOZ_ASSERT(searchStr->type() == MIRType::String); + + auto* lir = new (alloc()) + LStringIndexOf(useRegisterAtStart(string), useRegisterAtStart(searchStr)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStringStartsWith(MStringStartsWith* ins) { + auto* string = ins->string(); + MOZ_ASSERT(string->type() == MIRType::String); + + auto* searchStr = ins->searchString(); + MOZ_ASSERT(searchStr->type() == MIRType::String); + + if (searchStr->isConstant()) { + JSLinearString* linear = &searchStr->toConstant()->toString()->asLinear(); + + if (CanCompareCharactersInline(linear)) { + auto* lir = new (alloc()) + LStringStartsWithInline(useRegister(string), temp(), linear); + define(lir, ins); + assignSafepoint(lir, ins); + return; + } + } + + auto* lir = new (alloc()) LStringStartsWith(useRegisterAtStart(string), + useRegisterAtStart(searchStr)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStringEndsWith(MStringEndsWith* ins) { + auto* string = ins->string(); + MOZ_ASSERT(string->type() == MIRType::String); + + auto* searchStr = ins->searchString(); + MOZ_ASSERT(searchStr->type() == MIRType::String); + + if (searchStr->isConstant()) { + JSLinearString* linear = &searchStr->toConstant()->toString()->asLinear(); + + if (CanCompareCharactersInline(linear)) { + auto* lir = new (alloc()) + LStringEndsWithInline(useRegister(string), temp(), linear); + define(lir, ins); + assignSafepoint(lir, ins); + return; + } + } + + auto* lir = new (alloc()) LStringEndsWith(useRegisterAtStart(string), + useRegisterAtStart(searchStr)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStringConvertCase(MStringConvertCase* ins) { + MOZ_ASSERT(ins->string()->type() == MIRType::String); + + if (ins->mode() == MStringConvertCase::LowerCase) { +#ifdef JS_CODEGEN_X86 + // Due to lack of registers on x86, we reuse the string register as + // temporary. As a result we only need four temporary registers and take a + // bogus temporary as the fifth argument. + LDefinition temp4 = LDefinition::BogusTemp(); +#else + LDefinition temp4 = temp(); +#endif + auto* lir = new (alloc()) + LStringToLowerCase(useRegister(ins->string()), temp(), temp(), temp(), + temp4, tempByteOpRegister()); + define(lir, ins); + assignSafepoint(lir, ins); + } else { + auto* lir = + new (alloc()) LStringToUpperCase(useRegisterAtStart(ins->string())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitStart(MStart* start) {} + +void LIRGenerator::visitNop(MNop* nop) {} + +void LIRGenerator::visitLimitedTruncate(MLimitedTruncate* nop) { + redefine(nop, nop->input()); +} + +void LIRGenerator::visitOsrEntry(MOsrEntry* entry) { + LOsrEntry* lir = new (alloc()) LOsrEntry(temp()); + defineFixed(lir, entry, LAllocation(AnyRegister(OsrFrameReg))); +} + +void LIRGenerator::visitOsrValue(MOsrValue* value) { + LOsrValue* lir = new (alloc()) LOsrValue(useRegister(value->entry())); + defineBox(lir, value); +} + +void LIRGenerator::visitOsrReturnValue(MOsrReturnValue* value) { + LOsrReturnValue* lir = + new (alloc()) LOsrReturnValue(useRegister(value->entry())); + defineBox(lir, value); +} + +void LIRGenerator::visitOsrEnvironmentChain(MOsrEnvironmentChain* object) { + LOsrEnvironmentChain* lir = + new (alloc()) LOsrEnvironmentChain(useRegister(object->entry())); + define(lir, object); +} + +void LIRGenerator::visitOsrArgumentsObject(MOsrArgumentsObject* object) { + LOsrArgumentsObject* lir = + new (alloc()) LOsrArgumentsObject(useRegister(object->entry())); + define(lir, object); +} + +void LIRGenerator::visitToDouble(MToDouble* convert) { + MDefinition* opd = convert->input(); + mozilla::DebugOnly conversion = + convert->conversion(); + + switch (opd->type()) { + case MIRType::Value: { + LValueToDouble* lir = new (alloc()) LValueToDouble(useBox(opd)); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + break; + } + + case MIRType::Null: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + lowerConstantDouble(0, convert); + break; + + case MIRType::Undefined: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + lowerConstantDouble(GenericNaN(), convert); + break; + + case MIRType::Boolean: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + [[fallthrough]]; + + case MIRType::Int32: { + LInt32ToDouble* lir = + new (alloc()) LInt32ToDouble(useRegisterAtStart(opd)); + define(lir, convert); + break; + } + + case MIRType::Float32: { + LFloat32ToDouble* lir = + new (alloc()) LFloat32ToDouble(useRegisterAtStart(opd)); + define(lir, convert); + break; + } + + case MIRType::Double: + redefine(convert, opd); + break; + + default: + // Objects might be effectful. Symbols will throw. + // Strings are complicated - we don't handle them yet. + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitToFloat32(MToFloat32* convert) { + MDefinition* opd = convert->input(); + mozilla::DebugOnly conversion = + convert->conversion(); + + switch (opd->type()) { + case MIRType::Value: { + LValueToFloat32* lir = new (alloc()) LValueToFloat32(useBox(opd)); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + break; + } + + case MIRType::Null: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + lowerConstantFloat32(0, convert); + break; + + case MIRType::Undefined: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + lowerConstantFloat32(GenericNaN(), convert); + break; + + case MIRType::Boolean: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + [[fallthrough]]; + + case MIRType::Int32: { + LInt32ToFloat32* lir = + new (alloc()) LInt32ToFloat32(useRegisterAtStart(opd)); + define(lir, convert); + break; + } + + case MIRType::Double: { + LDoubleToFloat32* lir = + new (alloc()) LDoubleToFloat32(useRegisterAtStart(opd)); + define(lir, convert); + break; + } + + case MIRType::Float32: + redefine(convert, opd); + break; + + default: + // Objects might be effectful. Symbols will throw. + // Strings are complicated - we don't handle them yet. + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitToNumberInt32(MToNumberInt32* convert) { + MDefinition* opd = convert->input(); + + switch (opd->type()) { + case MIRType::Value: { + auto* lir = new (alloc()) LValueToInt32(useBox(opd), tempDouble(), temp(), + LValueToInt32::NORMAL); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + if (lir->mode() == LValueToInt32::TRUNCATE) { + assignSafepoint(lir, convert); + } + break; + } + + case MIRType::Null: + MOZ_ASSERT(convert->conversion() == IntConversionInputKind::Any); + define(new (alloc()) LInteger(0), convert); + break; + + case MIRType::Boolean: + MOZ_ASSERT(convert->conversion() == IntConversionInputKind::Any || + convert->conversion() == + IntConversionInputKind::NumbersOrBoolsOnly); + redefine(convert, opd); + break; + + case MIRType::Int32: + redefine(convert, opd); + break; + + case MIRType::Float32: { + LFloat32ToInt32* lir = new (alloc()) LFloat32ToInt32(useRegister(opd)); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + break; + } + + case MIRType::Double: { + LDoubleToInt32* lir = new (alloc()) LDoubleToInt32(useRegister(opd)); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + break; + } + + case MIRType::String: + case MIRType::Symbol: + case MIRType::BigInt: + case MIRType::Object: + case MIRType::Undefined: + // Objects might be effectful. Symbols and BigInts throw. Undefined + // coerces to NaN, not int32. + MOZ_CRASH("ToInt32 invalid input type"); + + default: + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitBooleanToInt32(MBooleanToInt32* convert) { + MDefinition* opd = convert->input(); + MOZ_ASSERT(opd->type() == MIRType::Boolean); + redefine(convert, opd); +} + +void LIRGenerator::visitTruncateToInt32(MTruncateToInt32* truncate) { + MDefinition* opd = truncate->input(); + + switch (opd->type()) { + case MIRType::Value: { + LValueToInt32* lir = new (alloc()) LValueToInt32( + useBox(opd), tempDouble(), temp(), LValueToInt32::TRUNCATE); + assignSnapshot(lir, truncate->bailoutKind()); + define(lir, truncate); + assignSafepoint(lir, truncate); + break; + } + + case MIRType::Null: + case MIRType::Undefined: + define(new (alloc()) LInteger(0), truncate); + break; + + case MIRType::Int32: + case MIRType::Boolean: + redefine(truncate, opd); + break; + + case MIRType::Double: + // May call into JS::ToInt32() on the slow OOL path. + gen->setNeedsStaticStackAlignment(); + lowerTruncateDToInt32(truncate); + break; + + case MIRType::Float32: + // May call into JS::ToInt32() on the slow OOL path. + gen->setNeedsStaticStackAlignment(); + lowerTruncateFToInt32(truncate); + break; + + default: + // Objects might be effectful. Symbols throw. + // Strings are complicated - we don't handle them yet. + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitInt32ToIntPtr(MInt32ToIntPtr* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::Int32); + MOZ_ASSERT(ins->type() == MIRType::IntPtr); + +#ifdef JS_64BIT + // If the result is only used by instructions that expect a bounds-checked + // index, we must have eliminated or hoisted a bounds check and we can assume + // the index is non-negative. This lets us generate more efficient code. + if (ins->canBeNegative()) { + bool canBeNegative = false; + for (MUseDefIterator iter(ins); iter; iter++) { + if (!iter.def()->isSpectreMaskIndex() && + !iter.def()->isLoadUnboxedScalar() && + !iter.def()->isStoreUnboxedScalar() && + !iter.def()->isLoadDataViewElement() && + !iter.def()->isStoreDataViewElement()) { + canBeNegative = true; + break; + } + } + if (!canBeNegative) { + ins->setCanNotBeNegative(); + } + } + + if (ins->canBeNegative()) { + auto* lir = new (alloc()) LInt32ToIntPtr(useAnyAtStart(input)); + define(lir, ins); + } else { + redefine(ins, input); + } +#else + // On 32-bit platforms this is a no-op. + redefine(ins, input); +#endif +} + +void LIRGenerator::visitNonNegativeIntPtrToInt32( + MNonNegativeIntPtrToInt32* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::IntPtr); + MOZ_ASSERT(ins->type() == MIRType::Int32); + +#ifdef JS_64BIT + auto* lir = + new (alloc()) LNonNegativeIntPtrToInt32(useRegisterAtStart(input)); + assignSnapshot(lir, ins->bailoutKind()); + defineReuseInput(lir, ins, 0); +#else + // On 32-bit platforms this is a no-op. + redefine(ins, input); +#endif +} + +void LIRGenerator::visitWasmExtendU32Index(MWasmExtendU32Index* ins) { +#ifdef JS_64BIT + // Technically this produces an Int64 register and I guess we could clean that + // up, but it's a 64-bit only operation, so it doesn't actually matter. + + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::Int32); + MOZ_ASSERT(ins->type() == MIRType::Int64); + + // Input reuse is OK even on ARM64 because this node *must* reuse its input in + // order not to generate any code at all, as is the intent. + auto* lir = new (alloc()) LWasmExtendU32Index(useRegisterAtStart(input)); + defineReuseInput(lir, ins, 0); +#else + MOZ_CRASH("64-bit only"); +#endif +} + +void LIRGenerator::visitWasmWrapU32Index(MWasmWrapU32Index* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::Int64); + MOZ_ASSERT(ins->type() == MIRType::Int32); + + // Tricky: On 64-bit, this just returns its input (except on MIPS64 there may + // be a sign/zero extension). On 32-bit, it returns the low register of the + // input, and should generate no code. + + // If this assertion does not hold then using "input" unadorned as an alias + // for the low register will not work. +#if defined(JS_NUNBOX32) + static_assert(INT64LOW_INDEX == 0); +#endif + + auto* lir = new (alloc()) LWasmWrapU32Index(useRegisterAtStart(input)); + defineReuseInput(lir, ins, 0); +} + +void LIRGenerator::visitIntPtrToDouble(MIntPtrToDouble* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::IntPtr); + MOZ_ASSERT(ins->type() == MIRType::Double); + + auto* lir = new (alloc()) LIntPtrToDouble(useRegister(input)); + define(lir, ins); +} + +void LIRGenerator::visitAdjustDataViewLength(MAdjustDataViewLength* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::IntPtr); + + auto* lir = new (alloc()) LAdjustDataViewLength(useRegisterAtStart(input)); + assignSnapshot(lir, ins->bailoutKind()); + defineReuseInput(lir, ins, 0); +} + +void LIRGenerator::visitToBigInt(MToBigInt* ins) { + MDefinition* opd = ins->input(); + + switch (opd->type()) { + case MIRType::Value: { + auto* lir = new (alloc()) LValueToBigInt(useBox(opd)); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + assignSafepoint(lir, ins); + break; + } + + case MIRType::BigInt: + redefine(ins, opd); + break; + + default: + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitToInt64(MToInt64* ins) { + MDefinition* opd = ins->input(); + + switch (opd->type()) { + case MIRType::Value: { + auto* lir = new (alloc()) LValueToInt64(useBox(opd), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineInt64(lir, ins); + assignSafepoint(lir, ins); + break; + } + + case MIRType::Boolean: { + auto* lir = new (alloc()) LBooleanToInt64(useRegisterAtStart(opd)); + defineInt64(lir, ins); + break; + } + + case MIRType::String: { + auto* lir = new (alloc()) LStringToInt64(useRegister(opd)); + defineInt64(lir, ins); + assignSafepoint(lir, ins); + break; + } + + // An Int64 may be passed here from a BigInt to Int64 conversion. + case MIRType::Int64: { + redefine(ins, opd); + break; + } + + default: + // Undefined, Null, Number, and Symbol throw. + // Objects may be effectful. + // BigInt operands are eliminated by the type policy. + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitTruncateBigIntToInt64(MTruncateBigIntToInt64* ins) { + MOZ_ASSERT(ins->input()->type() == MIRType::BigInt); + auto* lir = new (alloc()) LTruncateBigIntToInt64(useRegister(ins->input())); + defineInt64(lir, ins); +} + +void LIRGenerator::visitInt64ToBigInt(MInt64ToBigInt* ins) { + MOZ_ASSERT(ins->input()->type() == MIRType::Int64); + auto* lir = + new (alloc()) LInt64ToBigInt(useInt64Register(ins->input()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitWasmTruncateToInt32(MWasmTruncateToInt32* ins) { + MDefinition* input = ins->input(); + switch (input->type()) { + case MIRType::Double: + case MIRType::Float32: { + auto* lir = new (alloc()) LWasmTruncateToInt32(useRegisterAtStart(input)); + define(lir, ins); + break; + } + default: + MOZ_CRASH("unexpected type in WasmTruncateToInt32"); + } +} + +void LIRGenerator::visitWasmBuiltinTruncateToInt32( + MWasmBuiltinTruncateToInt32* truncate) { + mozilla::DebugOnly opd = truncate->input(); + MOZ_ASSERT(opd->type() == MIRType::Double || opd->type() == MIRType::Float32); + + // May call into JS::ToInt32() on the slow OOL path. + gen->setNeedsStaticStackAlignment(); + lowerWasmBuiltinTruncateToInt32(truncate); +} + +void LIRGenerator::visitWasmBoxValue(MWasmBoxValue* ins) { + LWasmBoxValue* lir = new (alloc()) LWasmBoxValue(useBox(ins->input())); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitWasmAnyRefFromJSObject(MWasmAnyRefFromJSObject* ins) { + LWasmAnyRefFromJSObject* lir = + new (alloc()) LWasmAnyRefFromJSObject(useRegisterAtStart(ins->input())); + define(lir, ins); +} + +void LIRGenerator::visitWrapInt64ToInt32(MWrapInt64ToInt32* ins) { + define(new (alloc()) LWrapInt64ToInt32(useInt64AtStart(ins->input())), ins); +} + +void LIRGenerator::visitToString(MToString* ins) { + MDefinition* opd = ins->input(); + + switch (opd->type()) { + case MIRType::Null: { + const JSAtomState& names = gen->runtime->names(); + LPointer* lir = new (alloc()) LPointer(names.null); + define(lir, ins); + break; + } + + case MIRType::Undefined: { + const JSAtomState& names = gen->runtime->names(); + LPointer* lir = new (alloc()) LPointer(names.undefined); + define(lir, ins); + break; + } + + case MIRType::Boolean: { + LBooleanToString* lir = new (alloc()) LBooleanToString(useRegister(opd)); + define(lir, ins); + break; + } + + case MIRType::Double: { + LDoubleToString* lir = + new (alloc()) LDoubleToString(useRegister(opd), temp()); + + define(lir, ins); + assignSafepoint(lir, ins); + break; + } + + case MIRType::Int32: { + LIntToString* lir = new (alloc()) LIntToString(useRegister(opd)); + + define(lir, ins); + assignSafepoint(lir, ins); + break; + } + + case MIRType::String: + redefine(ins, ins->input()); + break; + + case MIRType::Value: { + LValueToString* lir = + new (alloc()) LValueToString(useBox(opd), tempToUnbox()); + if (ins->needsSnapshot()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); + assignSafepoint(lir, ins); + break; + } + + default: + // Float32, symbols, bigint, and objects are not supported. + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitRegExp(MRegExp* ins) { + LRegExp* lir = new (alloc()) LRegExp(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitRegExpMatcher(MRegExpMatcher* ins) { + MOZ_ASSERT(ins->regexp()->type() == MIRType::Object); + MOZ_ASSERT(ins->string()->type() == MIRType::String); + MOZ_ASSERT(ins->lastIndex()->type() == MIRType::Int32); + + LRegExpMatcher* lir = new (alloc()) LRegExpMatcher( + useFixedAtStart(ins->regexp(), RegExpMatcherRegExpReg), + useFixedAtStart(ins->string(), RegExpMatcherStringReg), + useFixedAtStart(ins->lastIndex(), RegExpMatcherLastIndexReg)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitRegExpSearcher(MRegExpSearcher* ins) { + MOZ_ASSERT(ins->regexp()->type() == MIRType::Object); + MOZ_ASSERT(ins->string()->type() == MIRType::String); + MOZ_ASSERT(ins->lastIndex()->type() == MIRType::Int32); + + LRegExpSearcher* lir = new (alloc()) LRegExpSearcher( + useFixedAtStart(ins->regexp(), RegExpSearcherRegExpReg), + useFixedAtStart(ins->string(), RegExpSearcherStringReg), + useFixedAtStart(ins->lastIndex(), RegExpSearcherLastIndexReg)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitRegExpExecMatch(MRegExpExecMatch* ins) { + MOZ_ASSERT(ins->regexp()->type() == MIRType::Object); + MOZ_ASSERT(ins->string()->type() == MIRType::String); + + auto* lir = new (alloc()) + LRegExpExecMatch(useFixedAtStart(ins->regexp(), RegExpMatcherRegExpReg), + useFixedAtStart(ins->string(), RegExpMatcherStringReg)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitRegExpExecTest(MRegExpExecTest* ins) { + MOZ_ASSERT(ins->regexp()->type() == MIRType::Object); + MOZ_ASSERT(ins->string()->type() == MIRType::String); + + auto* lir = new (alloc()) + LRegExpExecTest(useFixedAtStart(ins->regexp(), RegExpExecTestRegExpReg), + useFixedAtStart(ins->string(), RegExpExecTestStringReg)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitRegExpPrototypeOptimizable( + MRegExpPrototypeOptimizable* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + LRegExpPrototypeOptimizable* lir = new (alloc()) + LRegExpPrototypeOptimizable(useRegister(ins->object()), temp()); + define(lir, ins); +} + +void LIRGenerator::visitRegExpInstanceOptimizable( + MRegExpInstanceOptimizable* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->proto()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + LRegExpInstanceOptimizable* lir = new (alloc()) LRegExpInstanceOptimizable( + useRegister(ins->object()), useRegister(ins->proto()), temp()); + define(lir, ins); +} + +void LIRGenerator::visitGetFirstDollarIndex(MGetFirstDollarIndex* ins) { + MOZ_ASSERT(ins->str()->type() == MIRType::String); + MOZ_ASSERT(ins->type() == MIRType::Int32); + LGetFirstDollarIndex* lir = new (alloc()) + LGetFirstDollarIndex(useRegister(ins->str()), temp(), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStringReplace(MStringReplace* ins) { + MOZ_ASSERT(ins->pattern()->type() == MIRType::String); + MOZ_ASSERT(ins->string()->type() == MIRType::String); + MOZ_ASSERT(ins->replacement()->type() == MIRType::String); + + LStringReplace* lir = new (alloc()) + LStringReplace(useRegisterOrConstantAtStart(ins->string()), + useRegisterAtStart(ins->pattern()), + useRegisterOrConstantAtStart(ins->replacement())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBinaryCache(MBinaryCache* ins) { + MDefinition* lhs = ins->getOperand(0); + MDefinition* rhs = ins->getOperand(1); + + MOZ_ASSERT(ins->type() == MIRType::Value || ins->type() == MIRType::Boolean); + LInstruction* lir; + if (ins->type() == MIRType::Value) { + LBinaryValueCache* valueLir = new (alloc()) LBinaryValueCache( + useBox(lhs), useBox(rhs), tempFixed(FloatReg0), tempFixed(FloatReg1)); + defineBox(valueLir, ins); + lir = valueLir; + } else { + MOZ_ASSERT(ins->type() == MIRType::Boolean); + LBinaryBoolCache* boolLir = new (alloc()) LBinaryBoolCache( + useBox(lhs), useBox(rhs), tempFixed(FloatReg0), tempFixed(FloatReg1)); + define(boolLir, ins); + lir = boolLir; + } + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitUnaryCache(MUnaryCache* ins) { + MDefinition* input = ins->getOperand(0); + MOZ_ASSERT(ins->type() == MIRType::Value); + + LUnaryCache* lir = new (alloc()) LUnaryCache(useBox(input)); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitModuleMetadata(MModuleMetadata* ins) { + LModuleMetadata* lir = new (alloc()) LModuleMetadata(); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitDynamicImport(MDynamicImport* ins) { + LDynamicImport* lir = new (alloc()) LDynamicImport( + useBoxAtStart(ins->specifier()), useBoxAtStart(ins->options())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitLambda(MLambda* ins) { + MOZ_ASSERT(ins->environmentChain()->type() == MIRType::Object); + + auto* lir = + new (alloc()) LLambda(useRegister(ins->environmentChain()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitFunctionWithProto(MFunctionWithProto* ins) { + MOZ_ASSERT(ins->environmentChain()->type() == MIRType::Object); + MOZ_ASSERT(ins->prototype()->type() == MIRType::Object); + + auto* lir = new (alloc()) + LFunctionWithProto(useRegisterAtStart(ins->environmentChain()), + useRegisterAtStart(ins->prototype())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitSetFunName(MSetFunName* ins) { + MOZ_ASSERT(ins->fun()->type() == MIRType::Object); + MOZ_ASSERT(ins->name()->type() == MIRType::Value); + + LSetFunName* lir = new (alloc()) + LSetFunName(useRegisterAtStart(ins->fun()), useBoxAtStart(ins->name())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewLexicalEnvironmentObject( + MNewLexicalEnvironmentObject* ins) { + auto* lir = new (alloc()) LNewLexicalEnvironmentObject(temp()); + + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewClassBodyEnvironmentObject( + MNewClassBodyEnvironmentObject* ins) { + auto* lir = new (alloc()) LNewClassBodyEnvironmentObject(temp()); + + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewVarEnvironmentObject(MNewVarEnvironmentObject* ins) { + auto* lir = new (alloc()) LNewVarEnvironmentObject(temp()); + + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitKeepAliveObject(MKeepAliveObject* ins) { + MDefinition* obj = ins->object(); + MOZ_ASSERT(obj->type() == MIRType::Object); + + add(new (alloc()) LKeepAliveObject(useKeepalive(obj)), ins); +} + +void LIRGenerator::visitDebugEnterGCUnsafeRegion( + MDebugEnterGCUnsafeRegion* ins) { + add(new (alloc()) LDebugEnterGCUnsafeRegion(temp()), ins); +} + +void LIRGenerator::visitDebugLeaveGCUnsafeRegion( + MDebugLeaveGCUnsafeRegion* ins) { + add(new (alloc()) LDebugLeaveGCUnsafeRegion(temp()), ins); +} + +void LIRGenerator::visitSlots(MSlots* ins) { + define(new (alloc()) LSlots(useRegisterAtStart(ins->object())), ins); +} + +void LIRGenerator::visitElements(MElements* ins) { + define(new (alloc()) LElements(useRegisterAtStart(ins->object())), ins); +} + +void LIRGenerator::visitLoadDynamicSlot(MLoadDynamicSlot* ins) { + MOZ_ASSERT(ins->type() == MIRType::Value); + defineBox(new (alloc()) LLoadDynamicSlotV(useRegisterAtStart(ins->slots())), + ins); +} + +void LIRGenerator::visitFunctionEnvironment(MFunctionEnvironment* ins) { + define(new (alloc()) + LFunctionEnvironment(useRegisterAtStart(ins->function())), + ins); +} + +void LIRGenerator::visitHomeObject(MHomeObject* ins) { + define(new (alloc()) LHomeObject(useRegisterAtStart(ins->function())), ins); +} + +void LIRGenerator::visitHomeObjectSuperBase(MHomeObjectSuperBase* ins) { + MOZ_ASSERT(ins->homeObject()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Value); + + auto lir = + new (alloc()) LHomeObjectSuperBase(useRegisterAtStart(ins->homeObject())); + defineBox(lir, ins); +} + +void LIRGenerator::visitInterruptCheck(MInterruptCheck* ins) { + LInstruction* lir = new (alloc()) LInterruptCheck(); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitWasmInterruptCheck(MWasmInterruptCheck* ins) { + auto* lir = + new (alloc()) LWasmInterruptCheck(useRegisterAtStart(ins->instance())); + add(lir, ins); + assignWasmSafepoint(lir); +} + +void LIRGenerator::visitWasmTrap(MWasmTrap* ins) { + add(new (alloc()) LWasmTrap, ins); +} + +void LIRGenerator::visitWasmTrapIfNull(MWasmTrapIfNull* ins) { + auto* lir = new (alloc()) LWasmTrapIfNull(useRegister(ins->value())); + add(lir, ins); +} + +void LIRGenerator::visitWasmReinterpret(MWasmReinterpret* ins) { + if (ins->type() == MIRType::Int64) { + defineInt64(new (alloc()) + LWasmReinterpretToI64(useRegisterAtStart(ins->input())), + ins); + } else if (ins->input()->type() == MIRType::Int64) { + define(new (alloc()) + LWasmReinterpretFromI64(useInt64RegisterAtStart(ins->input())), + ins); + } else { + define(new (alloc()) LWasmReinterpret(useRegisterAtStart(ins->input())), + ins); + } +} + +void LIRGenerator::visitStoreDynamicSlot(MStoreDynamicSlot* ins) { + LInstruction* lir; + + switch (ins->value()->type()) { + case MIRType::Value: + lir = new (alloc()) + LStoreDynamicSlotV(useRegister(ins->slots()), useBox(ins->value())); + add(lir, ins); + break; + + case MIRType::Double: + add(new (alloc()) LStoreDynamicSlotT(useRegister(ins->slots()), + useRegister(ins->value())), + ins); + break; + + case MIRType::Float32: + MOZ_CRASH("Float32 shouldn't be stored in a slot."); + + default: + add(new (alloc()) LStoreDynamicSlotT(useRegister(ins->slots()), + useRegisterOrConstant(ins->value())), + ins); + break; + } +} + +// Returns true iff |def| is a constant that's either not a GC thing or is not +// allocated in the nursery. +static bool IsNonNurseryConstant(MDefinition* def) { + if (!def->isConstant()) { + return false; + } + Value v = def->toConstant()->toJSValue(); + return !v.isGCThing() || !IsInsideNursery(v.toGCThing()); +} + +void LIRGenerator::visitPostWriteBarrier(MPostWriteBarrier* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + // LPostWriteBarrier assumes that if it has a constant object then that + // object is tenured, and does not need to be tested for being in the + // nursery. Ensure that assumption holds by lowering constant nursery + // objects to a register. + bool useConstantObject = IsNonNurseryConstant(ins->object()); + + switch (ins->value()->type()) { + case MIRType::Object: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteBarrierO* lir = new (alloc()) + LPostWriteBarrierO(useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::String: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteBarrierS* lir = new (alloc()) + LPostWriteBarrierS(useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::BigInt: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + auto* lir = new (alloc()) + LPostWriteBarrierBI(useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::Value: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteBarrierV* lir = new (alloc()) + LPostWriteBarrierV(useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useBox(ins->value()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + default: + // Currently, only objects and strings can be in the nursery. Other + // instruction types cannot hold nursery pointers. + break; + } +} + +void LIRGenerator::visitPostWriteElementBarrier(MPostWriteElementBarrier* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + // LPostWriteElementBarrier assumes that if it has a constant object then that + // object is tenured, and does not need to be tested for being in the + // nursery. Ensure that assumption holds by lowering constant nursery + // objects to a register. + bool useConstantObject = + ins->object()->isConstant() && + !IsInsideNursery(&ins->object()->toConstant()->toObject()); + + switch (ins->value()->type()) { + case MIRType::Object: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteElementBarrierO* lir = new (alloc()) LPostWriteElementBarrierO( + useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), useRegister(ins->index()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::String: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteElementBarrierS* lir = new (alloc()) LPostWriteElementBarrierS( + useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), useRegister(ins->index()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::BigInt: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + auto* lir = new (alloc()) LPostWriteElementBarrierBI( + useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), useRegister(ins->index()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::Value: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteElementBarrierV* lir = new (alloc()) LPostWriteElementBarrierV( + useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->index()), useBox(ins->value()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + default: + // Currently, only objects, strings, and bigints can be in the nursery. + // Other instruction types cannot hold nursery pointers. + break; + } +} + +void LIRGenerator::visitAssertCanElidePostWriteBarrier( + MAssertCanElidePostWriteBarrier* ins) { + auto* lir = new (alloc()) LAssertCanElidePostWriteBarrier( + useRegister(ins->object()), useBox(ins->value()), temp()); + add(lir, ins); +} + +void LIRGenerator::visitArrayLength(MArrayLength* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + auto* lir = new (alloc()) LArrayLength(useRegisterAtStart(ins->elements())); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitSetArrayLength(MSetArrayLength* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + MOZ_ASSERT(ins->index()->isConstant()); + add(new (alloc()) LSetArrayLength(useRegister(ins->elements()), + useRegisterOrConstant(ins->index())), + ins); +} + +void LIRGenerator::visitFunctionLength(MFunctionLength* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = new (alloc()) LFunctionLength(useRegister(ins->function())); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitFunctionName(MFunctionName* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = new (alloc()) LFunctionName(useRegister(ins->function())); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitGetNextEntryForIterator(MGetNextEntryForIterator* ins) { + MOZ_ASSERT(ins->iter()->type() == MIRType::Object); + MOZ_ASSERT(ins->result()->type() == MIRType::Object); + auto lir = new (alloc()) LGetNextEntryForIterator(useRegister(ins->iter()), + useRegister(ins->result()), + temp(), temp(), temp()); + define(lir, ins); +} + +void LIRGenerator::visitArrayBufferByteLength(MArrayBufferByteLength* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::IntPtr); + + auto* lir = + new (alloc()) LArrayBufferByteLength(useRegisterAtStart(ins->object())); + define(lir, ins); +} + +void LIRGenerator::visitArrayBufferViewLength(MArrayBufferViewLength* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::IntPtr); + + auto* lir = + new (alloc()) LArrayBufferViewLength(useRegisterAtStart(ins->object())); + define(lir, ins); +} + +void LIRGenerator::visitArrayBufferViewByteOffset( + MArrayBufferViewByteOffset* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::IntPtr); + + auto* lir = new (alloc()) + LArrayBufferViewByteOffset(useRegisterAtStart(ins->object())); + define(lir, ins); +} + +void LIRGenerator::visitArrayBufferViewElements(MArrayBufferViewElements* ins) { + MOZ_ASSERT(ins->type() == MIRType::Elements); + define(new (alloc()) + LArrayBufferViewElements(useRegisterAtStart(ins->object())), + ins); +} + +void LIRGenerator::visitTypedArrayElementSize(MTypedArrayElementSize* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + define(new (alloc()) + LTypedArrayElementSize(useRegisterAtStart(ins->object())), + ins); +} + +void LIRGenerator::visitGuardHasAttachedArrayBuffer( + MGuardHasAttachedArrayBuffer* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = new (alloc()) + LGuardHasAttachedArrayBuffer(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardNumberToIntPtrIndex( + MGuardNumberToIntPtrIndex* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::Double); + + auto* lir = new (alloc()) LGuardNumberToIntPtrIndex(useRegister(input)); + if (!ins->supportOOB()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); +} + +void LIRGenerator::visitInitializedLength(MInitializedLength* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + define(new (alloc()) LInitializedLength(useRegisterAtStart(ins->elements())), + ins); +} + +void LIRGenerator::visitSetInitializedLength(MSetInitializedLength* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + MOZ_ASSERT(ins->index()->isConstant()); + add(new (alloc()) LSetInitializedLength(useRegister(ins->elements()), + useRegisterOrConstant(ins->index())), + ins); +} + +void LIRGenerator::visitNot(MNot* ins) { + MDefinition* op = ins->input(); + + // String is converted to length of string in the type analysis phase (see + // TestPolicy). + MOZ_ASSERT(op->type() != MIRType::String); + + // - boolean: x xor 1 + // - int32: LCompare(x, 0) + // - double: LCompare(x, 0) + // - null or undefined: true + // - symbol: false + // - bigint: LNotBI(x) + // - object: false if it never emulates undefined, else LNotO(x) + switch (op->type()) { + case MIRType::Boolean: { + MConstant* cons = MConstant::New(alloc(), Int32Value(1)); + ins->block()->insertBefore(ins, cons); + lowerForALU(new (alloc()) LBitOpI(JSOp::BitXor), ins, op, cons); + break; + } + case MIRType::Int32: + define(new (alloc()) LNotI(useRegisterAtStart(op)), ins); + break; + case MIRType::Int64: + define(new (alloc()) LNotI64(useInt64RegisterAtStart(op)), ins); + break; + case MIRType::Double: + define(new (alloc()) LNotD(useRegister(op)), ins); + break; + case MIRType::Float32: + define(new (alloc()) LNotF(useRegister(op)), ins); + break; + case MIRType::Undefined: + case MIRType::Null: + define(new (alloc()) LInteger(1), ins); + break; + case MIRType::Symbol: + define(new (alloc()) LInteger(0), ins); + break; + case MIRType::BigInt: + define(new (alloc()) LNotBI(useRegisterAtStart(op)), ins); + break; + case MIRType::Object: + define(new (alloc()) LNotO(useRegister(op)), ins); + break; + case MIRType::Value: { + auto* lir = new (alloc()) LNotV(useBox(op), tempDouble(), tempToUnbox()); + define(lir, ins); + break; + } + + default: + MOZ_CRASH("Unexpected MIRType."); + } +} + +void LIRGenerator::visitBoundsCheck(MBoundsCheck* ins) { + MOZ_ASSERT(ins->type() == MIRType::Int32 || ins->type() == MIRType::IntPtr); + MOZ_ASSERT(ins->index()->type() == ins->type()); + MOZ_ASSERT(ins->length()->type() == ins->type()); + + if (!ins->fallible()) { + return; + } + + LInstruction* check; + if (ins->minimum() || ins->maximum()) { + check = new (alloc()) + LBoundsCheckRange(useRegisterOrInt32Constant(ins->index()), + useAny(ins->length()), temp()); + } else { + check = new (alloc()) LBoundsCheck(useRegisterOrInt32Constant(ins->index()), + useAnyOrInt32Constant(ins->length())); + } + assignSnapshot(check, ins->bailoutKind()); + add(check, ins); +} + +void LIRGenerator::visitSpectreMaskIndex(MSpectreMaskIndex* ins) { + MOZ_ASSERT(ins->type() == MIRType::Int32 || ins->type() == MIRType::IntPtr); + MOZ_ASSERT(ins->index()->type() == ins->type()); + MOZ_ASSERT(ins->length()->type() == ins->type()); + + auto* lir = new (alloc()) + LSpectreMaskIndex(useRegister(ins->index()), useAny(ins->length())); + define(lir, ins); +} + +void LIRGenerator::visitBoundsCheckLower(MBoundsCheckLower* ins) { + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + if (!ins->fallible()) { + return; + } + + LInstruction* check = + new (alloc()) LBoundsCheckLower(useRegister(ins->index())); + assignSnapshot(check, ins->bailoutKind()); + add(check, ins); +} + +void LIRGenerator::visitInArray(MInArray* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + MOZ_ASSERT(ins->initLength()->type() == MIRType::Int32); + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + + auto* lir = new (alloc()) LInArray(useRegister(ins->elements()), + useRegisterOrConstant(ins->index()), + useRegister(ins->initLength())); + if (ins->needsNegativeIntCheck()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); +} + +void LIRGenerator::visitGuardElementNotHole(MGuardElementNotHole* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + auto* guard = new (alloc()) + LGuardElementNotHole(useRegisterAtStart(ins->elements()), + useRegisterOrConstantAtStart(ins->index())); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); +} + +void LIRGenerator::visitLoadElement(MLoadElement* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + MOZ_ASSERT(ins->type() == MIRType::Value); + + auto* lir = new (alloc()) LLoadElementV(useRegister(ins->elements()), + useRegisterOrConstant(ins->index())); + assignSnapshot(lir, ins->bailoutKind()); + defineBox(lir, ins); +} + +void LIRGenerator::visitLoadElementHole(MLoadElementHole* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + MOZ_ASSERT(ins->initLength()->type() == MIRType::Int32); + MOZ_ASSERT(ins->type() == MIRType::Value); + + LLoadElementHole* lir = new (alloc()) + LLoadElementHole(useRegister(ins->elements()), useRegister(ins->index()), + useRegister(ins->initLength())); + if (ins->needsNegativeIntCheck()) { + assignSnapshot(lir, ins->bailoutKind()); + } + defineBox(lir, ins); +} + +void LIRGenerator::visitStoreElement(MStoreElement* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + const LUse elements = useRegister(ins->elements()); + const LAllocation index = useRegisterOrConstant(ins->index()); + + switch (ins->value()->type()) { + case MIRType::Value: { + LInstruction* lir = + new (alloc()) LStoreElementV(elements, index, useBox(ins->value())); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + add(lir, ins); + break; + } + + default: { + const LAllocation value = useRegisterOrNonDoubleConstant(ins->value()); + LInstruction* lir = new (alloc()) LStoreElementT(elements, index, value); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + add(lir, ins); + break; + } + } +} + +void LIRGenerator::visitStoreHoleValueElement(MStoreHoleValueElement* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + auto* lir = new (alloc()) LStoreHoleValueElement(useRegister(ins->elements()), + useRegister(ins->index())); + add(lir, ins); +} + +static bool BoundsCheckNeedsSpectreTemp() { + // On x86, spectreBoundsCheck32 can emit better code if it has a scratch + // register and index masking is enabled. +#ifdef JS_CODEGEN_X86 + return JitOptions.spectreIndexMasking; +#else + return false; +#endif +} + +void LIRGenerator::visitStoreElementHole(MStoreElementHole* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + const LUse object = useRegister(ins->object()); + const LUse elements = useRegister(ins->elements()); + const LAllocation index = useRegister(ins->index()); + + LInstruction* lir; + switch (ins->value()->type()) { + case MIRType::Value: + lir = new (alloc()) LStoreElementHoleV(object, elements, index, + useBox(ins->value()), temp()); + break; + + default: { + const LAllocation value = useRegisterOrNonDoubleConstant(ins->value()); + lir = new (alloc()) + LStoreElementHoleT(object, elements, index, value, temp()); + break; + } + } + + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitEffectiveAddress(MEffectiveAddress* ins) { + define(new (alloc()) LEffectiveAddress(useRegister(ins->base()), + useRegister(ins->index())), + ins); +} + +void LIRGenerator::visitArrayPopShift(MArrayPopShift* ins) { + MOZ_ASSERT(ins->type() == MIRType::Value); + + auto* lir = + new (alloc()) LArrayPopShift(useRegister(ins->object()), temp(), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineBox(lir, ins); + + if (ins->mode() == MArrayPopShift::Shift) { + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitArrayPush(MArrayPush* ins) { + MOZ_ASSERT(ins->type() == MIRType::Int32); + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + + LUse object = useRegister(ins->object()); + + LDefinition spectreTemp = + BoundsCheckNeedsSpectreTemp() ? temp() : LDefinition::BogusTemp(); + + auto* lir = new (alloc()) + LArrayPush(object, useBox(ins->value()), temp(), spectreTemp); + // We will bailout before pushing if the length would overflow INT32_MAX. + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitArraySlice(MArraySlice* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->begin()->type() == MIRType::Int32); + MOZ_ASSERT(ins->end()->type() == MIRType::Int32); + + LArraySlice* lir = new (alloc()) LArraySlice( + useRegisterAtStart(ins->object()), useRegisterAtStart(ins->begin()), + useRegisterAtStart(ins->end()), tempFixed(CallTempReg0), + tempFixed(CallTempReg1)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitArgumentsSlice(MArgumentsSlice* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->begin()->type() == MIRType::Int32); + MOZ_ASSERT(ins->end()->type() == MIRType::Int32); + + auto* lir = new (alloc()) LArgumentsSlice( + useRegisterAtStart(ins->object()), useRegisterAtStart(ins->begin()), + useRegisterAtStart(ins->end()), tempFixed(CallTempReg0), + tempFixed(CallTempReg1)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitFrameArgumentsSlice(MFrameArgumentsSlice* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + MOZ_ASSERT(ins->begin()->type() == MIRType::Int32); + MOZ_ASSERT(ins->count()->type() == MIRType::Int32); + + auto* lir = new (alloc()) LFrameArgumentsSlice( + useRegister(ins->begin()), useRegister(ins->count()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitInlineArgumentsSlice(MInlineArgumentsSlice* ins) { + LAllocation begin = useRegisterOrConstant(ins->begin()); + LAllocation count = useRegisterOrConstant(ins->count()); + uint32_t numActuals = ins->numActuals(); + uint32_t numOperands = + numActuals * BOX_PIECES + LInlineArgumentsSlice::NumNonArgumentOperands; + + auto* lir = allocateVariadic(numOperands, temp()); + if (!lir) { + abort(AbortReason::Alloc, "OOM: LIRGenerator::visitInlineArgumentsSlice"); + return; + } + + lir->setOperand(LInlineArgumentsSlice::Begin, begin); + lir->setOperand(LInlineArgumentsSlice::Count, count); + for (uint32_t i = 0; i < numActuals; i++) { + MDefinition* arg = ins->getArg(i); + uint32_t index = LInlineArgumentsSlice::ArgIndex(i); + lir->setBoxOperand(index, + useBoxOrTypedOrConstant(arg, /*useConstant = */ true)); + } + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNormalizeSliceTerm(MNormalizeSliceTerm* ins) { + MOZ_ASSERT(ins->type() == MIRType::Int32); + MOZ_ASSERT(ins->value()->type() == MIRType::Int32); + MOZ_ASSERT(ins->length()->type() == MIRType::Int32); + + auto* lir = new (alloc()) LNormalizeSliceTerm(useRegister(ins->value()), + useRegister(ins->length())); + define(lir, ins); +} + +void LIRGenerator::visitArrayJoin(MArrayJoin* ins) { + MOZ_ASSERT(ins->type() == MIRType::String); + MOZ_ASSERT(ins->array()->type() == MIRType::Object); + MOZ_ASSERT(ins->sep()->type() == MIRType::String); + + auto* lir = new (alloc()) + LArrayJoin(useRegisterAtStart(ins->array()), + useRegisterAtStart(ins->sep()), tempFixed(CallTempReg0)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStringSplit(MStringSplit* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + MOZ_ASSERT(ins->string()->type() == MIRType::String); + MOZ_ASSERT(ins->separator()->type() == MIRType::String); + + LStringSplit* lir = new (alloc()) LStringSplit( + useRegisterAtStart(ins->string()), useRegisterAtStart(ins->separator())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitLoadUnboxedScalar(MLoadUnboxedScalar* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::IntPtr); + MOZ_ASSERT(IsNumericType(ins->type()) || ins->type() == MIRType::Boolean); + + if (Scalar::isBigIntType(ins->storageType()) && + ins->requiresMemoryBarrier()) { + lowerAtomicLoad64(ins); + return; + } + + const LUse elements = useRegister(ins->elements()); + const LAllocation index = useRegisterOrIndexConstant( + ins->index(), ins->storageType(), ins->offsetAdjustment()); + + // NOTE: the generated code must match the assembly code in gen_load in + // GenerateAtomicOperations.py + Synchronization sync = Synchronization::Load(); + if (ins->requiresMemoryBarrier()) { + LMemoryBarrier* fence = new (alloc()) LMemoryBarrier(sync.barrierBefore); + add(fence, ins); + } + + if (!Scalar::isBigIntType(ins->storageType())) { + // We need a temp register for Uint32Array with known double result. + LDefinition tempDef = LDefinition::BogusTemp(); + if (ins->storageType() == Scalar::Uint32 && + IsFloatingPointType(ins->type())) { + tempDef = temp(); + } + + auto* lir = new (alloc()) LLoadUnboxedScalar(elements, index, tempDef); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); + } else { + MOZ_ASSERT(ins->type() == MIRType::BigInt); + + auto* lir = + new (alloc()) LLoadUnboxedBigInt(elements, index, temp(), tempInt64()); + define(lir, ins); + assignSafepoint(lir, ins); + } + + if (ins->requiresMemoryBarrier()) { + LMemoryBarrier* fence = new (alloc()) LMemoryBarrier(sync.barrierAfter); + add(fence, ins); + } +} + +void LIRGenerator::visitLoadDataViewElement(MLoadDataViewElement* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::IntPtr); + + MOZ_ASSERT(IsNumericType(ins->type())); + + const LUse elements = useRegister(ins->elements()); + const LUse index = useRegister(ins->index()); + const LAllocation littleEndian = useRegisterOrConstant(ins->littleEndian()); + + // We need a temp register for: + // - Uint32Array with known double result, + // - Float32Array, + // - and BigInt64Array and BigUint64Array. + LDefinition tempDef = LDefinition::BogusTemp(); + if ((ins->storageType() == Scalar::Uint32 && + IsFloatingPointType(ins->type())) || + ins->storageType() == Scalar::Float32) { + tempDef = temp(); + } + if (Scalar::isBigIntType(ins->storageType())) { +#ifdef JS_CODEGEN_X86 + // There are not enough registers on x86. + if (littleEndian.isConstant()) { + tempDef = temp(); + } +#else + tempDef = temp(); +#endif + } + + // We also need a separate 64-bit temp register for: + // - Float64Array + // - and BigInt64Array and BigUint64Array. + LInt64Definition temp64Def = LInt64Definition::BogusTemp(); + if (Scalar::byteSize(ins->storageType()) == 8) { + temp64Def = tempInt64(); + } + + auto* lir = new (alloc()) + LLoadDataViewElement(elements, index, littleEndian, tempDef, temp64Def); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); + if (Scalar::isBigIntType(ins->storageType())) { + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitClampToUint8(MClampToUint8* ins) { + MDefinition* in = ins->input(); + + switch (in->type()) { + case MIRType::Boolean: + redefine(ins, in); + break; + + case MIRType::Int32: + defineReuseInput(new (alloc()) LClampIToUint8(useRegisterAtStart(in)), + ins, 0); + break; + + case MIRType::Double: + // LClampDToUint8 clobbers its input register. Making it available as + // a temp copy describes this behavior to the register allocator. + define(new (alloc()) + LClampDToUint8(useRegisterAtStart(in), tempCopy(in, 0)), + ins); + break; + + case MIRType::Value: { + LClampVToUint8* lir = + new (alloc()) LClampVToUint8(useBox(in), tempDouble()); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + assignSafepoint(lir, ins); + break; + } + + default: + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitLoadTypedArrayElementHole( + MLoadTypedArrayElementHole* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->index()->type() == MIRType::IntPtr); + + MOZ_ASSERT(ins->type() == MIRType::Value); + + const LUse object = useRegister(ins->object()); + const LAllocation index = useRegister(ins->index()); + + if (!Scalar::isBigIntType(ins->arrayType())) { + auto* lir = new (alloc()) LLoadTypedArrayElementHole(object, index, temp()); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + defineBox(lir, ins); + } else { +#ifdef JS_CODEGEN_X86 + LDefinition tmp = LDefinition::BogusTemp(); +#else + LDefinition tmp = temp(); +#endif + + auto* lir = new (alloc()) + LLoadTypedArrayElementHoleBigInt(object, index, tmp, tempInt64()); + defineBox(lir, ins); + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitStoreUnboxedScalar(MStoreUnboxedScalar* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::IntPtr); + + if (ins->isFloatWrite()) { + MOZ_ASSERT_IF(ins->writeType() == Scalar::Float32, + ins->value()->type() == MIRType::Float32); + MOZ_ASSERT_IF(ins->writeType() == Scalar::Float64, + ins->value()->type() == MIRType::Double); + } else if (ins->isBigIntWrite()) { + MOZ_ASSERT(ins->value()->type() == MIRType::BigInt); + } else { + MOZ_ASSERT(ins->value()->type() == MIRType::Int32); + } + + if (ins->isBigIntWrite() && ins->requiresMemoryBarrier()) { + lowerAtomicStore64(ins); + return; + } + + LUse elements = useRegister(ins->elements()); + LAllocation index = + useRegisterOrIndexConstant(ins->index(), ins->writeType()); + LAllocation value; + + // For byte arrays, the value has to be in a byte register on x86. + if (ins->isByteWrite()) { + value = useByteOpRegisterOrNonDoubleConstant(ins->value()); + } else if (ins->isBigIntWrite()) { + value = useRegister(ins->value()); + } else { + value = useRegisterOrNonDoubleConstant(ins->value()); + } + + // Optimization opportunity for atomics: on some platforms there + // is a store instruction that incorporates the necessary + // barriers, and we could use that instead of separate barrier and + // store instructions. See bug #1077027. + // + // NOTE: the generated code must match the assembly code in gen_store in + // GenerateAtomicOperations.py + Synchronization sync = Synchronization::Store(); + if (ins->requiresMemoryBarrier()) { + LMemoryBarrier* fence = new (alloc()) LMemoryBarrier(sync.barrierBefore); + add(fence, ins); + } + if (!ins->isBigIntWrite()) { + add(new (alloc()) LStoreUnboxedScalar(elements, index, value), ins); + } else { + add(new (alloc()) LStoreUnboxedBigInt(elements, index, value, tempInt64()), + ins); + } + if (ins->requiresMemoryBarrier()) { + LMemoryBarrier* fence = new (alloc()) LMemoryBarrier(sync.barrierAfter); + add(fence, ins); + } +} + +void LIRGenerator::visitStoreDataViewElement(MStoreDataViewElement* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::IntPtr); + MOZ_ASSERT(ins->littleEndian()->type() == MIRType::Boolean); + + if (ins->isFloatWrite()) { + MOZ_ASSERT_IF(ins->writeType() == Scalar::Float32, + ins->value()->type() == MIRType::Float32); + MOZ_ASSERT_IF(ins->writeType() == Scalar::Float64, + ins->value()->type() == MIRType::Double); + } else if (ins->isBigIntWrite()) { + MOZ_ASSERT(ins->value()->type() == MIRType::BigInt); + } else { + MOZ_ASSERT(ins->value()->type() == MIRType::Int32); + } + + LUse elements = useRegister(ins->elements()); + LUse index = useRegister(ins->index()); + LAllocation value; + if (ins->isBigIntWrite()) { + value = useRegister(ins->value()); + } else { + value = useRegisterOrNonDoubleConstant(ins->value()); + } + LAllocation littleEndian = useRegisterOrConstant(ins->littleEndian()); + + LDefinition tempDef = LDefinition::BogusTemp(); + LInt64Definition temp64Def = LInt64Definition::BogusTemp(); + if (Scalar::byteSize(ins->writeType()) < 8) { + tempDef = temp(); + } else { + temp64Def = tempInt64(); + } + + add(new (alloc()) LStoreDataViewElement(elements, index, value, littleEndian, + tempDef, temp64Def), + ins); +} + +void LIRGenerator::visitStoreTypedArrayElementHole( + MStoreTypedArrayElementHole* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::IntPtr); + MOZ_ASSERT(ins->length()->type() == MIRType::IntPtr); + + if (ins->isFloatWrite()) { + MOZ_ASSERT_IF(ins->arrayType() == Scalar::Float32, + ins->value()->type() == MIRType::Float32); + MOZ_ASSERT_IF(ins->arrayType() == Scalar::Float64, + ins->value()->type() == MIRType::Double); + } else if (ins->isBigIntWrite()) { + MOZ_ASSERT(ins->value()->type() == MIRType::BigInt); + } else { + MOZ_ASSERT(ins->value()->type() == MIRType::Int32); + } + + LUse elements = useRegister(ins->elements()); + LAllocation length = useAny(ins->length()); + LAllocation index = useRegister(ins->index()); + + // For byte arrays, the value has to be in a byte register on x86. + LAllocation value; + if (ins->isByteWrite()) { + value = useByteOpRegisterOrNonDoubleConstant(ins->value()); + } else if (ins->isBigIntWrite()) { + value = useRegister(ins->value()); + } else { + value = useRegisterOrNonDoubleConstant(ins->value()); + } + + if (!ins->isBigIntWrite()) { + LDefinition spectreTemp = + BoundsCheckNeedsSpectreTemp() ? temp() : LDefinition::BogusTemp(); + auto* lir = new (alloc()) LStoreTypedArrayElementHole( + elements, length, index, value, spectreTemp); + add(lir, ins); + } else { + auto* lir = new (alloc()) LStoreTypedArrayElementHoleBigInt( + elements, length, index, value, tempInt64()); + add(lir, ins); + } +} + +void LIRGenerator::visitLoadFixedSlot(MLoadFixedSlot* ins) { + MDefinition* obj = ins->object(); + MOZ_ASSERT(obj->type() == MIRType::Object); + + MIRType type = ins->type(); + + if (type == MIRType::Value) { + LLoadFixedSlotV* lir = + new (alloc()) LLoadFixedSlotV(useRegisterAtStart(obj)); + defineBox(lir, ins); + } else { + LLoadFixedSlotT* lir = + new (alloc()) LLoadFixedSlotT(useRegisterForTypedLoad(obj, type)); + define(lir, ins); + } +} + +void LIRGenerator::visitLoadFixedSlotAndUnbox(MLoadFixedSlotAndUnbox* ins) { + MDefinition* obj = ins->object(); + MOZ_ASSERT(obj->type() == MIRType::Object); + + LLoadFixedSlotAndUnbox* lir = + new (alloc()) LLoadFixedSlotAndUnbox(useRegisterAtStart(obj)); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); +} + +void LIRGenerator::visitLoadDynamicSlotAndUnbox(MLoadDynamicSlotAndUnbox* ins) { + MDefinition* slots = ins->slots(); + MOZ_ASSERT(slots->type() == MIRType::Slots); + + auto* lir = new (alloc()) LLoadDynamicSlotAndUnbox(useRegisterAtStart(slots)); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); +} + +void LIRGenerator::visitLoadElementAndUnbox(MLoadElementAndUnbox* ins) { + MDefinition* elements = ins->elements(); + MDefinition* index = ins->index(); + MOZ_ASSERT(elements->type() == MIRType::Elements); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) + LLoadElementAndUnbox(useRegister(elements), useRegisterOrConstant(index)); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); +} + +void LIRGenerator::visitAddAndStoreSlot(MAddAndStoreSlot* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + LDefinition maybeTemp = LDefinition::BogusTemp(); + if (ins->kind() != MAddAndStoreSlot::Kind::FixedSlot) { + maybeTemp = temp(); + } + + auto* lir = new (alloc()) LAddAndStoreSlot(useRegister(ins->object()), + useBox(ins->value()), maybeTemp); + add(lir, ins); +} + +void LIRGenerator::visitAllocateAndStoreSlot(MAllocateAndStoreSlot* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = new (alloc()) LAllocateAndStoreSlot( + useRegisterAtStart(ins->object()), useBoxAtStart(ins->value()), + tempFixed(CallTempReg0), tempFixed(CallTempReg1)); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); +} + +void LIRGenerator::visitAddSlotAndCallAddPropHook( + MAddSlotAndCallAddPropHook* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + + auto* lir = new (alloc()) LAddSlotAndCallAddPropHook( + useRegisterAtStart(ins->object()), useBoxAtStart(ins->value())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStoreFixedSlot(MStoreFixedSlot* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + if (ins->value()->type() == MIRType::Value) { + LStoreFixedSlotV* lir = new (alloc()) + LStoreFixedSlotV(useRegister(ins->object()), useBox(ins->value())); + add(lir, ins); + } else { + LStoreFixedSlotT* lir = new (alloc()) LStoreFixedSlotT( + useRegister(ins->object()), useRegisterOrConstant(ins->value())); + add(lir, ins); + } +} + +void LIRGenerator::visitGetNameCache(MGetNameCache* ins) { + MOZ_ASSERT(ins->envObj()->type() == MIRType::Object); + + // Emit an overrecursed check: this is necessary because the cache can + // attach a scripted getter stub that calls this script recursively. + gen->setNeedsOverrecursedCheck(); + + LGetNameCache* lir = + new (alloc()) LGetNameCache(useRegister(ins->envObj()), temp()); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallGetIntrinsicValue(MCallGetIntrinsicValue* ins) { + LCallGetIntrinsicValue* lir = new (alloc()) LCallGetIntrinsicValue(); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetPropSuperCache(MGetPropSuperCache* ins) { + MDefinition* obj = ins->object(); + MDefinition* receiver = ins->receiver(); + MDefinition* id = ins->idval(); + + gen->setNeedsOverrecursedCheck(); + + bool useConstId = + id->type() == MIRType::String || id->type() == MIRType::Symbol; + + auto* lir = new (alloc()) + LGetPropSuperCache(useRegister(obj), useBoxOrTyped(receiver), + useBoxOrTypedOrConstant(id, useConstId)); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetPropertyCache(MGetPropertyCache* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Object || + value->type() == MIRType::Value); + + MDefinition* id = ins->idval(); + MOZ_ASSERT(id->type() == MIRType::String || id->type() == MIRType::Symbol || + id->type() == MIRType::Int32 || id->type() == MIRType::Value); + + // Emit an overrecursed check: this is necessary because the cache can + // attach a scripted getter stub that calls this script recursively. + gen->setNeedsOverrecursedCheck(); + + // If this is a GetProp, the id is a constant string. Allow passing it as a + // constant to reduce register allocation pressure. + bool useConstId = + id->type() == MIRType::String || id->type() == MIRType::Symbol; + + auto* lir = new (alloc()) LGetPropertyCache( + useBoxOrTyped(value), useBoxOrTypedOrConstant(id, useConstId)); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBindNameCache(MBindNameCache* ins) { + MOZ_ASSERT(ins->envChain()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Object); + + LBindNameCache* lir = + new (alloc()) LBindNameCache(useRegister(ins->envChain()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallBindVar(MCallBindVar* ins) { + MOZ_ASSERT(ins->environmentChain()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Object); + + LCallBindVar* lir = + new (alloc()) LCallBindVar(useRegister(ins->environmentChain())); + define(lir, ins); +} + +void LIRGenerator::visitGuardObjectIdentity(MGuardObjectIdentity* ins) { + LGuardObjectIdentity* guard = new (alloc()) LGuardObjectIdentity( + useRegister(ins->object()), useRegister(ins->expected())); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardSpecificFunction(MGuardSpecificFunction* ins) { + auto* guard = new (alloc()) LGuardSpecificFunction( + useRegister(ins->function()), useRegister(ins->expected())); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, ins->function()); +} + +void LIRGenerator::visitGuardSpecificAtom(MGuardSpecificAtom* ins) { + auto* guard = + new (alloc()) LGuardSpecificAtom(useRegister(ins->str()), temp()); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, ins->str()); + assignSafepoint(guard, ins); +} + +void LIRGenerator::visitGuardSpecificSymbol(MGuardSpecificSymbol* ins) { + auto* guard = new (alloc()) LGuardSpecificSymbol(useRegister(ins->symbol())); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, ins->symbol()); +} + +void LIRGenerator::visitGuardSpecificInt32(MGuardSpecificInt32* ins) { + auto* guard = new (alloc()) LGuardSpecificInt32(useRegister(ins->num())); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, ins->num()); +} + +void LIRGenerator::visitGuardStringToIndex(MGuardStringToIndex* ins) { + MOZ_ASSERT(ins->string()->type() == MIRType::String); + auto* guard = new (alloc()) LGuardStringToIndex(useRegister(ins->string())); + assignSnapshot(guard, ins->bailoutKind()); + define(guard, ins); + assignSafepoint(guard, ins); +} + +void LIRGenerator::visitGuardStringToInt32(MGuardStringToInt32* ins) { + MOZ_ASSERT(ins->string()->type() == MIRType::String); + auto* guard = + new (alloc()) LGuardStringToInt32(useRegister(ins->string()), temp()); + assignSnapshot(guard, ins->bailoutKind()); + define(guard, ins); + assignSafepoint(guard, ins); +} + +void LIRGenerator::visitGuardStringToDouble(MGuardStringToDouble* ins) { + MOZ_ASSERT(ins->string()->type() == MIRType::String); + auto* guard = new (alloc()) + LGuardStringToDouble(useRegister(ins->string()), temp(), temp()); + assignSnapshot(guard, ins->bailoutKind()); + define(guard, ins); + assignSafepoint(guard, ins); +} + +void LIRGenerator::visitGuardNoDenseElements(MGuardNoDenseElements* ins) { + auto* guard = + new (alloc()) LGuardNoDenseElements(useRegister(ins->object()), temp()); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardShape(MGuardShape* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + if (JitOptions.spectreObjectMitigations) { + auto* lir = + new (alloc()) LGuardShape(useRegisterAtStart(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineReuseInput(lir, ins, 0); + } else { + auto* lir = new (alloc()) + LGuardShape(useRegister(ins->object()), LDefinition::BogusTemp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); + } +} + +void LIRGenerator::visitGuardMultipleShapes(MGuardMultipleShapes* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + if (JitOptions.spectreObjectMitigations) { + auto* lir = new (alloc()) LGuardMultipleShapes( + useRegisterAtStart(ins->object()), useRegister(ins->shapeList()), + temp(), temp(), temp(), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineReuseInput(lir, ins, 0); + } else { + auto* lir = new (alloc()) LGuardMultipleShapes( + useRegister(ins->object()), useRegister(ins->shapeList()), temp(), + temp(), temp(), LDefinition::BogusTemp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); + } +} + +void LIRGenerator::visitGuardProto(MGuardProto* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->expected()->type() == MIRType::Object); + + auto* lir = new (alloc()) LGuardProto(useRegister(ins->object()), + useRegister(ins->expected()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardNullProto(MGuardNullProto* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = new (alloc()) LGuardNullProto(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardIsNativeObject(MGuardIsNativeObject* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = + new (alloc()) LGuardIsNativeObject(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardGlobalGeneration(MGuardGlobalGeneration* ins) { + auto* lir = new (alloc()) LGuardGlobalGeneration(temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); +} + +void LIRGenerator::visitGuardIsProxy(MGuardIsProxy* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = new (alloc()) LGuardIsProxy(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardIsNotProxy(MGuardIsNotProxy* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = + new (alloc()) LGuardIsNotProxy(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardIsNotDOMProxy(MGuardIsNotDOMProxy* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + + auto* lir = + new (alloc()) LGuardIsNotDOMProxy(useRegister(ins->proxy()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->proxy()); +} + +void LIRGenerator::visitProxyGet(MProxyGet* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + auto* lir = new (alloc()) + LProxyGet(useRegisterAtStart(ins->proxy()), tempFixed(CallTempReg0)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitProxyGetByValue(MProxyGetByValue* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + MOZ_ASSERT(ins->idVal()->type() == MIRType::Value); + auto* lir = new (alloc()) LProxyGetByValue(useRegisterAtStart(ins->proxy()), + useBoxAtStart(ins->idVal())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitProxyHasProp(MProxyHasProp* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + MOZ_ASSERT(ins->idVal()->type() == MIRType::Value); + auto* lir = new (alloc()) LProxyHasProp(useRegisterAtStart(ins->proxy()), + useBoxAtStart(ins->idVal())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitProxySet(MProxySet* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + MOZ_ASSERT(ins->rhs()->type() == MIRType::Value); + auto* lir = new (alloc()) + LProxySet(useRegisterAtStart(ins->proxy()), useBoxAtStart(ins->rhs()), + tempFixed(CallTempReg0)); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitProxySetByValue(MProxySetByValue* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + MOZ_ASSERT(ins->idVal()->type() == MIRType::Value); + MOZ_ASSERT(ins->rhs()->type() == MIRType::Value); + auto* lir = new (alloc()) + LProxySetByValue(useRegisterAtStart(ins->proxy()), + useBoxAtStart(ins->idVal()), useBoxAtStart(ins->rhs())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallSetArrayLength(MCallSetArrayLength* ins) { + MOZ_ASSERT(ins->obj()->type() == MIRType::Object); + MOZ_ASSERT(ins->rhs()->type() == MIRType::Value); + auto* lir = new (alloc()) LCallSetArrayLength(useRegisterAtStart(ins->obj()), + useBoxAtStart(ins->rhs())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitMegamorphicLoadSlot(MMegamorphicLoadSlot* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + auto* lir = new (alloc()) + LMegamorphicLoadSlot(useRegisterAtStart(ins->object()), + tempFixed(CallTempReg0), tempFixed(CallTempReg1), + tempFixed(CallTempReg2), tempFixed(CallTempReg3)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); +} + +void LIRGenerator::visitMegamorphicLoadSlotByValue( + MMegamorphicLoadSlotByValue* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->idVal()->type() == MIRType::Value); + auto* lir = new (alloc()) LMegamorphicLoadSlotByValue( + useRegisterAtStart(ins->object()), useBoxAtStart(ins->idVal()), + tempFixed(CallTempReg0), tempFixed(CallTempReg1), + tempFixed(CallTempReg2)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); +} + +void LIRGenerator::visitMegamorphicStoreSlot(MMegamorphicStoreSlot* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->rhs()->type() == MIRType::Value); + +#ifdef JS_CODEGEN_X86 + auto* lir = new (alloc()) LMegamorphicStoreSlot( + useFixedAtStart(ins->object(), CallTempReg0), + useBoxFixedAtStart(ins->rhs(), CallTempReg1, CallTempReg2), + tempFixed(CallTempReg5)); +#else + auto* lir = new (alloc()) + LMegamorphicStoreSlot(useRegisterAtStart(ins->object()), + useBoxAtStart(ins->rhs()), tempFixed(CallTempReg0), + tempFixed(CallTempReg1), tempFixed(CallTempReg2)); +#endif + + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitMegamorphicHasProp(MMegamorphicHasProp* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->idVal()->type() == MIRType::Value); + auto* lir = new (alloc()) + LMegamorphicHasProp(useRegisterAtStart(ins->object()), + useBoxAtStart(ins->idVal()), tempFixed(CallTempReg0), + tempFixed(CallTempReg1), tempFixed(CallTempReg2)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); +} + +void LIRGenerator::visitGuardIsNotArrayBufferMaybeShared( + MGuardIsNotArrayBufferMaybeShared* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = new (alloc()) + LGuardIsNotArrayBufferMaybeShared(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardIsTypedArray(MGuardIsTypedArray* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = + new (alloc()) LGuardIsTypedArray(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitNurseryObject(MNurseryObject* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto* lir = new (alloc()) LNurseryObject(); + define(lir, ins); +} + +void LIRGenerator::visitGuardValue(MGuardValue* ins) { + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + auto* lir = new (alloc()) LGuardValue(useBox(ins->value())); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->value()); +} + +void LIRGenerator::visitGuardNullOrUndefined(MGuardNullOrUndefined* ins) { + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + auto* lir = new (alloc()) LGuardNullOrUndefined(useBox(ins->value())); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->value()); +} + +void LIRGenerator::visitGuardIsNotObject(MGuardIsNotObject* ins) { + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + auto* lir = new (alloc()) LGuardIsNotObject(useBox(ins->value())); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->value()); +} + +void LIRGenerator::visitGuardFunctionFlags(MGuardFunctionFlags* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = new (alloc()) LGuardFunctionFlags(useRegister(ins->function())); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->function()); +} + +void LIRGenerator::visitGuardFunctionIsNonBuiltinCtor( + MGuardFunctionIsNonBuiltinCtor* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = new (alloc()) + LGuardFunctionIsNonBuiltinCtor(useRegister(ins->function()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->function()); +} + +void LIRGenerator::visitGuardFunctionKind(MGuardFunctionKind* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = + new (alloc()) LGuardFunctionKind(useRegister(ins->function()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->function()); +} + +void LIRGenerator::visitGuardFunctionScript(MGuardFunctionScript* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = new (alloc()) LGuardFunctionScript(useRegister(ins->function())); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->function()); +} + +void LIRGenerator::visitAssertRange(MAssertRange* ins) { + MDefinition* input = ins->input(); + LInstruction* lir = nullptr; + + switch (input->type()) { + case MIRType::Boolean: + case MIRType::Int32: + case MIRType::IntPtr: + lir = new (alloc()) LAssertRangeI(useRegisterAtStart(input)); + break; + + case MIRType::Double: + lir = new (alloc()) LAssertRangeD(useRegister(input), tempDouble()); + break; + + case MIRType::Float32: + lir = new (alloc()) + LAssertRangeF(useRegister(input), tempDouble(), tempDouble()); + break; + + case MIRType::Value: + lir = new (alloc()) LAssertRangeV(useBox(input), tempToUnbox(), + tempDouble(), tempDouble()); + break; + + default: + MOZ_CRASH("Unexpected Range for MIRType"); + break; + } + + lir->setMir(ins); + add(lir); +} + +void LIRGenerator::visitAssertClass(MAssertClass* ins) { + auto* lir = + new (alloc()) LAssertClass(useRegisterAtStart(ins->input()), temp()); + add(lir, ins); +} + +void LIRGenerator::visitAssertShape(MAssertShape* ins) { + auto* lir = new (alloc()) LAssertShape(useRegisterAtStart(ins->input())); + add(lir, ins); +} + +void LIRGenerator::visitDeleteProperty(MDeleteProperty* ins) { + LCallDeleteProperty* lir = + new (alloc()) LCallDeleteProperty(useBoxAtStart(ins->value())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitDeleteElement(MDeleteElement* ins) { + LCallDeleteElement* lir = new (alloc()) LCallDeleteElement( + useBoxAtStart(ins->value()), useBoxAtStart(ins->index())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitObjectToIterator(MObjectToIterator* ins) { + auto* lir = new (alloc()) + LObjectToIterator(useRegister(ins->object()), temp(), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitValueToIterator(MValueToIterator* ins) { + auto* lir = new (alloc()) LValueToIterator(useBoxAtStart(ins->value())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitLoadSlotByIteratorIndex(MLoadSlotByIteratorIndex* ins) { + auto* lir = new (alloc()) LLoadSlotByIteratorIndex( + useRegisterAtStart(ins->object()), useRegisterAtStart(ins->iterator()), + temp(), temp()); + defineBox(lir, ins); +} + +void LIRGenerator::visitStoreSlotByIteratorIndex( + MStoreSlotByIteratorIndex* ins) { + auto* lir = new (alloc()) LStoreSlotByIteratorIndex( + useRegister(ins->object()), useRegister(ins->iterator()), + useBox(ins->value()), temp(), temp()); + add(lir, ins); +} + +void LIRGenerator::visitIteratorHasIndices(MIteratorHasIndices* ins) { + MOZ_ASSERT(ins->hasOneUse()); + emitAtUses(ins); +} + +void LIRGenerator::visitSetPropertyCache(MSetPropertyCache* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + MDefinition* id = ins->idval(); + MOZ_ASSERT(id->type() == MIRType::String || id->type() == MIRType::Symbol || + id->type() == MIRType::Int32 || id->type() == MIRType::Value); + + // If this is a SetProp, the id is a constant string. Allow passing it as a + // constant to reduce register allocation pressure. + bool useConstId = + id->type() == MIRType::String || id->type() == MIRType::Symbol; + bool useConstValue = IsNonNurseryConstant(ins->value()); + + // Emit an overrecursed check: this is necessary because the cache can + // attach a scripted setter stub that calls this script recursively. + gen->setNeedsOverrecursedCheck(); + + // We need a double temp register for TypedArray or TypedObject stubs. + LDefinition tempD = tempFixed(FloatReg0); + + LInstruction* lir = new (alloc()) LSetPropertyCache( + useRegister(ins->object()), useBoxOrTypedOrConstant(id, useConstId), + useBoxOrTypedOrConstant(ins->value(), useConstValue), temp(), tempD); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitMegamorphicSetElement(MMegamorphicSetElement* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->index()->type() == MIRType::Value); + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + + // See comment in LIROps.yaml (x86 is short on registers) +#ifdef JS_CODEGEN_X86 + auto* lir = new (alloc()) LMegamorphicSetElement( + useFixedAtStart(ins->object(), CallTempReg0), + useBoxFixedAtStart(ins->index(), CallTempReg1, CallTempReg2), + useBoxFixedAtStart(ins->value(), CallTempReg3, CallTempReg4), + tempFixed(CallTempReg5)); +#else + auto* lir = new (alloc()) LMegamorphicSetElement( + useRegisterAtStart(ins->object()), useBoxAtStart(ins->index()), + useBoxAtStart(ins->value()), tempFixed(CallTempReg0), + tempFixed(CallTempReg1), tempFixed(CallTempReg2)); +#endif + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetIteratorCache(MGetIteratorCache* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Object || + value->type() == MIRType::Value); + + LGetIteratorCache* lir = + new (alloc()) LGetIteratorCache(useBoxOrTyped(value), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitOptimizeSpreadCallCache(MOptimizeSpreadCallCache* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Value); + + auto* lir = new (alloc()) LOptimizeSpreadCallCache(useBox(value), temp()); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitIteratorMore(MIteratorMore* ins) { + LIteratorMore* lir = + new (alloc()) LIteratorMore(useRegister(ins->iterator()), temp()); + defineBox(lir, ins); +} + +void LIRGenerator::visitIsNoIter(MIsNoIter* ins) { + MOZ_ASSERT(ins->hasOneUse()); + emitAtUses(ins); +} + +void LIRGenerator::visitIteratorEnd(MIteratorEnd* ins) { + LIteratorEnd* lir = new (alloc()) + LIteratorEnd(useRegister(ins->iterator()), temp(), temp(), temp()); + add(lir, ins); +} + +void LIRGenerator::visitCloseIterCache(MCloseIterCache* ins) { + LCloseIterCache* lir = + new (alloc()) LCloseIterCache(useRegister(ins->iter()), temp()); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStringLength(MStringLength* ins) { + MOZ_ASSERT(ins->string()->type() == MIRType::String); + define(new (alloc()) LStringLength(useRegisterAtStart(ins->string())), ins); +} + +void LIRGenerator::visitArgumentsLength(MArgumentsLength* ins) { + define(new (alloc()) LArgumentsLength(), ins); +} + +void LIRGenerator::visitGetFrameArgument(MGetFrameArgument* ins) { + LGetFrameArgument* lir = + new (alloc()) LGetFrameArgument(useRegisterOrConstant(ins->index())); + defineBox(lir, ins); +} + +void LIRGenerator::visitGetFrameArgumentHole(MGetFrameArgumentHole* ins) { + LDefinition spectreTemp = + BoundsCheckNeedsSpectreTemp() ? temp() : LDefinition::BogusTemp(); + + auto* lir = new (alloc()) LGetFrameArgumentHole( + useRegister(ins->index()), useRegister(ins->length()), spectreTemp); + assignSnapshot(lir, ins->bailoutKind()); + defineBox(lir, ins); +} + +void LIRGenerator::visitNewTarget(MNewTarget* ins) { + LNewTarget* lir = new (alloc()) LNewTarget(); + defineBox(lir, ins); +} + +void LIRGenerator::visitRest(MRest* ins) { + MOZ_ASSERT(ins->numActuals()->type() == MIRType::Int32); + + LRest* lir = new (alloc()) + LRest(useRegisterAtStart(ins->numActuals()), tempFixed(CallTempReg0), + tempFixed(CallTempReg1), tempFixed(CallTempReg2)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitThrow(MThrow* ins) { + MDefinition* value = ins->getOperand(0); + MOZ_ASSERT(value->type() == MIRType::Value); + + LThrow* lir = new (alloc()) LThrow(useBoxAtStart(value)); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitInCache(MInCache* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + + MOZ_ASSERT(lhs->type() == MIRType::String || lhs->type() == MIRType::Symbol || + lhs->type() == MIRType::Int32 || lhs->type() == MIRType::Value); + MOZ_ASSERT(rhs->type() == MIRType::Object); + + LInCache* lir = + new (alloc()) LInCache(useBoxOrTyped(lhs), useRegister(rhs), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitHasOwnCache(MHasOwnCache* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Object || + value->type() == MIRType::Value); + + MDefinition* id = ins->idval(); + MOZ_ASSERT(id->type() == MIRType::String || id->type() == MIRType::Symbol || + id->type() == MIRType::Int32 || id->type() == MIRType::Value); + + // Emit an overrecursed check: this is necessary because the cache can + // attach a scripted getter stub that calls this script recursively. + gen->setNeedsOverrecursedCheck(); + + LHasOwnCache* lir = + new (alloc()) LHasOwnCache(useBoxOrTyped(value), useBoxOrTyped(id)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckPrivateFieldCache(MCheckPrivateFieldCache* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Object || + value->type() == MIRType::Value); + + MDefinition* id = ins->idval(); + MOZ_ASSERT(id->type() == MIRType::String || id->type() == MIRType::Symbol || + id->type() == MIRType::Int32 || id->type() == MIRType::Value); + + LCheckPrivateFieldCache* lir = new (alloc()) + LCheckPrivateFieldCache(useBoxOrTyped(value), useBoxOrTyped(id)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewPrivateName(MNewPrivateName* ins) { + auto* lir = new (alloc()) LNewPrivateName(); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitInstanceOf(MInstanceOf* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + + MOZ_ASSERT(lhs->type() == MIRType::Value || lhs->type() == MIRType::Object); + MOZ_ASSERT(rhs->type() == MIRType::Object); + + if (lhs->type() == MIRType::Object) { + auto* lir = new (alloc()) LInstanceOfO(useRegister(lhs), useRegister(rhs)); + define(lir, ins); + assignSafepoint(lir, ins); + } else { + auto* lir = new (alloc()) LInstanceOfV(useBox(lhs), useRegister(rhs)); + define(lir, ins); + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitInstanceOfCache(MInstanceOfCache* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + + MOZ_ASSERT(lhs->type() == MIRType::Value); + MOZ_ASSERT(rhs->type() == MIRType::Object); + + LInstanceOfCache* lir = + new (alloc()) LInstanceOfCache(useBox(lhs), useRegister(rhs)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitIsArray(MIsArray* ins) { + MOZ_ASSERT(ins->type() == MIRType::Boolean); + + if (ins->value()->type() == MIRType::Object) { + LIsArrayO* lir = new (alloc()) LIsArrayO(useRegister(ins->value())); + define(lir, ins); + assignSafepoint(lir, ins); + } else { + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + LIsArrayV* lir = new (alloc()) LIsArrayV(useBox(ins->value()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitIsTypedArray(MIsTypedArray* ins) { + MOZ_ASSERT(ins->value()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + + auto* lir = new (alloc()) LIsTypedArray(useRegister(ins->value())); + define(lir, ins); + + if (ins->isPossiblyWrapped()) { + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitIsCallable(MIsCallable* ins) { + MOZ_ASSERT(ins->type() == MIRType::Boolean); + + if (ins->object()->type() == MIRType::Object) { + define(new (alloc()) LIsCallableO(useRegister(ins->object())), ins); + } else { + MOZ_ASSERT(ins->object()->type() == MIRType::Value); + define(new (alloc()) LIsCallableV(useBox(ins->object()), temp()), ins); + } +} + +void LIRGenerator::visitIsConstructor(MIsConstructor* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + define(new (alloc()) LIsConstructor(useRegister(ins->object())), ins); +} + +void LIRGenerator::visitIsCrossRealmArrayConstructor( + MIsCrossRealmArrayConstructor* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + define(new (alloc()) + LIsCrossRealmArrayConstructor(useRegister(ins->object())), + ins); +} + +static bool CanEmitAtUseForSingleTest(MInstruction* ins) { + if (!ins->canEmitAtUses()) { + return false; + } + + MUseIterator iter(ins->usesBegin()); + if (iter == ins->usesEnd()) { + return false; + } + + MNode* node = iter->consumer(); + if (!node->isDefinition()) { + return false; + } + + if (!node->toDefinition()->isTest()) { + return false; + } + + iter++; + return iter == ins->usesEnd(); +} + +void LIRGenerator::visitIsObject(MIsObject* ins) { + if (CanEmitAtUseForSingleTest(ins)) { + emitAtUses(ins); + return; + } + + MDefinition* opd = ins->input(); + MOZ_ASSERT(opd->type() == MIRType::Value); + LIsObject* lir = new (alloc()) LIsObject(useBoxAtStart(opd)); + define(lir, ins); +} + +void LIRGenerator::visitIsNullOrUndefined(MIsNullOrUndefined* ins) { + if (CanEmitAtUseForSingleTest(ins)) { + emitAtUses(ins); + return; + } + + MDefinition* opd = ins->input(); + MOZ_ASSERT(opd->type() == MIRType::Value); + LIsNullOrUndefined* lir = + new (alloc()) LIsNullOrUndefined(useBoxAtStart(opd)); + define(lir, ins); +} + +void LIRGenerator::visitHasClass(MHasClass* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + define(new (alloc()) LHasClass(useRegister(ins->object())), ins); +} + +void LIRGenerator::visitGuardToClass(MGuardToClass* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Object); + LGuardToClass* lir = + new (alloc()) LGuardToClass(useRegisterAtStart(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineReuseInput(lir, ins, 0); +} + +void LIRGenerator::visitGuardToFunction(MGuardToFunction* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Object); + LGuardToFunction* lir = + new (alloc()) LGuardToFunction(useRegisterAtStart(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineReuseInput(lir, ins, 0); +} + +void LIRGenerator::visitObjectClassToString(MObjectClassToString* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::String); + auto* lir = new (alloc()) LObjectClassToString( + useRegisterAtStart(ins->object()), tempFixed(CallTempReg0)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); +} + +void LIRGenerator::visitWasmAddOffset(MWasmAddOffset* ins) { + MOZ_ASSERT(ins->offset()); + if (ins->base()->type() == MIRType::Int32) { + MOZ_ASSERT(ins->type() == MIRType::Int32); + MOZ_ASSERT(ins->offset() <= UINT32_MAX); // Because memory32 + define(new (alloc()) LWasmAddOffset(useRegisterAtStart(ins->base())), ins); + } else { + MOZ_ASSERT(ins->type() == MIRType::Int64); +#ifdef JS_64BIT + defineInt64(new (alloc()) + LWasmAddOffset64(useInt64RegisterAtStart(ins->base())), + ins); +#else + // Avoid situation where the input is (a,b) and the output is (b,a). + defineInt64ReuseInput( + new (alloc()) LWasmAddOffset64(useInt64RegisterAtStart(ins->base())), + ins, 0); +#endif + } +} + +void LIRGenerator::visitWasmLoadInstance(MWasmLoadInstance* ins) { + if (ins->type() == MIRType::Int64) { +#ifdef JS_PUNBOX64 + LAllocation instance = useRegisterAtStart(ins->instance()); +#else + // Avoid reusing instance for a 64-bit output pair as the load clobbers the + // first half of that pair before loading the second half. + LAllocation instance = useRegister(ins->instance()); +#endif + auto* lir = new (alloc()) LWasmLoadInstance64(instance); + defineInt64(lir, ins); + } else { + auto* lir = + new (alloc()) LWasmLoadInstance(useRegisterAtStart(ins->instance())); + define(lir, ins); + } +} + +void LIRGenerator::visitWasmStoreInstance(MWasmStoreInstance* ins) { + MDefinition* value = ins->value(); + if (value->type() == MIRType::Int64) { +#ifdef JS_PUNBOX64 + LAllocation instance = useRegisterAtStart(ins->instance()); + LInt64Allocation valueAlloc = useInt64RegisterAtStart(value); +#else + LAllocation instance = useRegister(ins->instance()); + LInt64Allocation valueAlloc = useInt64Register(value); +#endif + add(new (alloc()) LWasmStoreSlotI64(valueAlloc, instance, ins->offset(), + mozilla::Nothing()), + ins); + } else { + MOZ_ASSERT(value->type() != MIRType::RefOrNull); + LAllocation instance = useRegisterAtStart(ins->instance()); + LAllocation valueAlloc = useRegisterAtStart(value); + add(new (alloc()) + LWasmStoreSlot(valueAlloc, instance, ins->offset(), value->type(), + MNarrowingOp::None, mozilla::Nothing()), + ins); + } +} + +void LIRGenerator::visitWasmBoundsCheck(MWasmBoundsCheck* ins) { + MOZ_ASSERT(!ins->isRedundant()); + + MDefinition* index = ins->index(); + MDefinition* boundsCheckLimit = ins->boundsCheckLimit(); + + MOZ_ASSERT(boundsCheckLimit->type() == index->type()); + + if (index->type() == MIRType::Int64) { + if (JitOptions.spectreIndexMasking) { + auto* lir = new (alloc()) LWasmBoundsCheck64( + useInt64RegisterAtStart(index), useInt64Register(boundsCheckLimit)); + defineInt64ReuseInput(lir, ins, 0); + } else { + auto* lir = new (alloc()) + LWasmBoundsCheck64(useInt64RegisterAtStart(index), + useInt64RegisterAtStart(boundsCheckLimit)); + add(lir, ins); + } + } else { + MOZ_ASSERT(index->type() == MIRType::Int32); + + if (JitOptions.spectreIndexMasking) { + auto* lir = new (alloc()) LWasmBoundsCheck(useRegisterAtStart(index), + useRegister(boundsCheckLimit)); + defineReuseInput(lir, ins, 0); + } else { + auto* lir = new (alloc()) LWasmBoundsCheck( + useRegisterAtStart(index), useRegisterAtStart(boundsCheckLimit)); + add(lir, ins); + } + } +} + +void LIRGenerator::visitWasmAlignmentCheck(MWasmAlignmentCheck* ins) { + MDefinition* index = ins->index(); + if (index->type() == MIRType::Int64) { + auto* lir = + new (alloc()) LWasmAlignmentCheck64(useInt64RegisterAtStart(index)); + add(lir, ins); + } else { + auto* lir = new (alloc()) LWasmAlignmentCheck(useRegisterAtStart(index)); + add(lir, ins); + } +} + +void LIRGenerator::visitWasmLoadInstanceDataField( + MWasmLoadInstanceDataField* ins) { + size_t offs = wasm::Instance::offsetInData(ins->instanceDataOffset()); + if (ins->type() == MIRType::Int64) { +#ifdef JS_PUNBOX64 + LAllocation instance = useRegisterAtStart(ins->instance()); +#else + // Avoid reusing instance for the output pair as the load clobbers the first + // half of that pair before loading the second half. + LAllocation instance = useRegister(ins->instance()); +#endif + defineInt64(new (alloc()) + LWasmLoadSlotI64(instance, offs, mozilla::Nothing()), + ins); + } else { + LAllocation instance = useRegisterAtStart(ins->instance()); + define(new (alloc()) LWasmLoadSlot(instance, offs, ins->type(), + MWideningOp::None, mozilla::Nothing()), + ins); + } +} + +void LIRGenerator::visitWasmLoadGlobalCell(MWasmLoadGlobalCell* ins) { + if (ins->type() == MIRType::Int64) { +#ifdef JS_PUNBOX64 + LAllocation cellPtr = useRegisterAtStart(ins->cellPtr()); +#else + // Avoid reusing cellPtr for the output pair as the load clobbers the first + // half of that pair before loading the second half. + LAllocation cellPtr = useRegister(ins->cellPtr()); +#endif + defineInt64(new (alloc()) + LWasmLoadSlotI64(cellPtr, /*offset=*/0, mozilla::Nothing()), + ins); + } else { + LAllocation cellPtr = useRegisterAtStart(ins->cellPtr()); + define(new (alloc()) LWasmLoadSlot(cellPtr, /*offset=*/0, ins->type(), + MWideningOp::None, mozilla::Nothing()), + ins); + } +} + +void LIRGenerator::visitWasmLoadTableElement(MWasmLoadTableElement* ins) { + LAllocation elements = useRegisterAtStart(ins->elements()); + LAllocation index = useRegisterAtStart(ins->index()); + define(new (alloc()) LWasmLoadTableElement(elements, index), ins); +} + +void LIRGenerator::visitWasmStoreInstanceDataField( + MWasmStoreInstanceDataField* ins) { + MDefinition* value = ins->value(); + size_t offs = wasm::Instance::offsetInData(ins->instanceDataOffset()); + if (value->type() == MIRType::Int64) { +#ifdef JS_PUNBOX64 + LAllocation instance = useRegisterAtStart(ins->instance()); + LInt64Allocation valueAlloc = useInt64RegisterAtStart(value); +#else + LAllocation instance = useRegister(ins->instance()); + LInt64Allocation valueAlloc = useInt64Register(value); +#endif + add(new (alloc()) + LWasmStoreSlotI64(valueAlloc, instance, offs, mozilla::Nothing()), + ins); + } else { + MOZ_ASSERT(value->type() != MIRType::RefOrNull); + LAllocation instance = useRegisterAtStart(ins->instance()); + LAllocation valueAlloc = useRegisterAtStart(value); + add(new (alloc()) LWasmStoreSlot(valueAlloc, instance, offs, value->type(), + MNarrowingOp::None, mozilla::Nothing()), + ins); + } +} + +void LIRGenerator::visitWasmStoreGlobalCell(MWasmStoreGlobalCell* ins) { + MDefinition* value = ins->value(); + size_t offs = 0; + if (value->type() == MIRType::Int64) { +#ifdef JS_PUNBOX64 + LAllocation cellPtr = useRegisterAtStart(ins->cellPtr()); + LInt64Allocation valueAlloc = useInt64RegisterAtStart(value); +#else + LAllocation cellPtr = useRegister(ins->cellPtr()); + LInt64Allocation valueAlloc = useInt64Register(value); +#endif + add(new (alloc()) + LWasmStoreSlotI64(valueAlloc, cellPtr, offs, mozilla::Nothing())); + } else { + MOZ_ASSERT(value->type() != MIRType::RefOrNull); + LAllocation cellPtr = useRegisterAtStart(ins->cellPtr()); + LAllocation valueAlloc = useRegisterAtStart(value); + add(new (alloc()) LWasmStoreSlot(valueAlloc, cellPtr, offs, value->type(), + MNarrowingOp::None, mozilla::Nothing())); + } +} + +void LIRGenerator::visitWasmStoreStackResult(MWasmStoreStackResult* ins) { + MDefinition* stackResultArea = ins->stackResultArea(); + MDefinition* value = ins->value(); + size_t offs = ins->offset(); + LInstruction* lir; + if (value->type() == MIRType::Int64) { + lir = new (alloc()) + LWasmStoreSlotI64(useInt64Register(value), useRegister(stackResultArea), + offs, mozilla::Nothing()); + } else { + MOZ_ASSERT(value->type() != MIRType::RefOrNull); + lir = new (alloc()) + LWasmStoreSlot(useRegister(value), useRegister(stackResultArea), offs, + value->type(), MNarrowingOp::None, mozilla::Nothing()); + } + add(lir, ins); +} + +void LIRGenerator::visitWasmDerivedPointer(MWasmDerivedPointer* ins) { + LAllocation base = useRegisterAtStart(ins->base()); + define(new (alloc()) LWasmDerivedPointer(base), ins); +} + +void LIRGenerator::visitWasmDerivedIndexPointer(MWasmDerivedIndexPointer* ins) { + LAllocation base = useRegisterAtStart(ins->base()); + LAllocation index = useRegisterAtStart(ins->index()); + define(new (alloc()) LWasmDerivedIndexPointer(base, index), ins); +} + +void LIRGenerator::visitWasmStoreRef(MWasmStoreRef* ins) { + LAllocation instance = useRegister(ins->instance()); + LAllocation valueBase = useFixed(ins->valueBase(), PreBarrierReg); + LAllocation value = useRegister(ins->value()); + uint32_t valueOffset = ins->offset(); + add(new (alloc()) + LWasmStoreRef(instance, valueBase, value, temp(), valueOffset, + mozilla::Nothing(), ins->preBarrierKind()), + ins); +} + +void LIRGenerator::visitWasmPostWriteBarrier(MWasmPostWriteBarrier* ins) { + LWasmPostWriteBarrier* lir = new (alloc()) LWasmPostWriteBarrier( + useFixed(ins->instance(), InstanceReg), useRegister(ins->object()), + useRegister(ins->valueBase()), useRegister(ins->value()), temp(), + ins->valueOffset()); + add(lir, ins); + assignWasmSafepoint(lir); +} + +void LIRGenerator::visitWasmParameter(MWasmParameter* ins) { + ABIArg abi = ins->abi(); + if (ins->type() == MIRType::StackResults) { + // Functions that return stack results receive an extra incoming parameter + // with type MIRType::StackResults. This value is a pointer to fresh + // memory. Here we treat it as if it were in fact MIRType::Pointer. + auto* lir = new (alloc()) LWasmParameter; + LDefinition def(LDefinition::TypeFrom(MIRType::Pointer), + LDefinition::FIXED); + def.setOutput(abi.argInRegister() ? LAllocation(abi.reg()) + : LArgument(abi.offsetFromArgBase())); + define(lir, ins, def); + return; + } + if (abi.argInRegister()) { +#if defined(JS_NUNBOX32) + if (abi.isGeneralRegPair()) { + defineInt64Fixed( + new (alloc()) LWasmParameterI64, ins, + LInt64Allocation(LAllocation(AnyRegister(abi.gpr64().high)), + LAllocation(AnyRegister(abi.gpr64().low)))); + return; + } +#endif + defineFixed(new (alloc()) LWasmParameter, ins, LAllocation(abi.reg())); + return; + } + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(!abi.argInRegister()); + defineInt64Fixed( + new (alloc()) LWasmParameterI64, ins, +#if defined(JS_NUNBOX32) + LInt64Allocation(LArgument(abi.offsetFromArgBase() + INT64HIGH_OFFSET), + LArgument(abi.offsetFromArgBase() + INT64LOW_OFFSET)) +#else + LInt64Allocation(LArgument(abi.offsetFromArgBase())) +#endif + ); + } else { + MOZ_ASSERT(IsNumberType(ins->type()) || ins->type() == MIRType::RefOrNull +#ifdef ENABLE_WASM_SIMD + || ins->type() == MIRType::Simd128 +#endif + ); + defineFixed(new (alloc()) LWasmParameter, ins, + LArgument(abi.offsetFromArgBase())); + } +} + +void LIRGenerator::visitWasmReturn(MWasmReturn* ins) { + MDefinition* rval = ins->getOperand(0); + MDefinition* instance = ins->getOperand(1); + + if (rval->type() == MIRType::Int64) { + add(new (alloc()) LWasmReturnI64(useInt64Fixed(rval, ReturnReg64), + useFixed(instance, InstanceReg))); + return; + } + + LAllocation returnReg; + if (rval->type() == MIRType::Float32) { + returnReg = useFixed(rval, ReturnFloat32Reg); + } else if (rval->type() == MIRType::Double) { + returnReg = useFixed(rval, ReturnDoubleReg); +#ifdef ENABLE_WASM_SIMD + } else if (rval->type() == MIRType::Simd128) { + returnReg = useFixed(rval, ReturnSimd128Reg); +#endif + } else if (rval->type() == MIRType::Int32 || + rval->type() == MIRType::RefOrNull) { + returnReg = useFixed(rval, ReturnReg); + } else { + MOZ_CRASH("Unexpected wasm return type"); + } + + LWasmReturn* lir = + new (alloc()) LWasmReturn(useFixed(instance, InstanceReg), returnReg); + add(lir); +} + +void LIRGenerator::visitWasmReturnVoid(MWasmReturnVoid* ins) { + MDefinition* instance = ins->getOperand(0); + LWasmReturnVoid* lir = + new (alloc()) LWasmReturnVoid(useFixed(instance, InstanceReg)); + add(lir); +} + +void LIRGenerator::visitWasmStackArg(MWasmStackArg* ins) { + if (ins->arg()->type() == MIRType::Int64) { + add(new (alloc()) + LWasmStackArgI64(useInt64RegisterOrConstantAtStart(ins->arg())), + ins); + } else if (IsFloatingPointType(ins->arg()->type())) { + MOZ_ASSERT(!ins->arg()->isEmittedAtUses()); + add(new (alloc()) LWasmStackArg(useRegisterAtStart(ins->arg())), ins); + } else { + add(new (alloc()) LWasmStackArg(useRegisterOrConstantAtStart(ins->arg())), + ins); + } +} + +void LIRGenerator::visitWasmRegisterResult(MWasmRegisterResult* ins) { + auto* lir = new (alloc()) LWasmRegisterResult(); + uint32_t vreg = getVirtualRegister(); + MOZ_ASSERT(ins->type() != MIRType::Int64); + auto type = LDefinition::TypeFrom(ins->type()); + lir->setDef(0, LDefinition(vreg, type, LGeneralReg(ins->loc()))); + ins->setVirtualRegister(vreg); + add(lir, ins); +} + +void LIRGenerator::visitWasmFloatRegisterResult(MWasmFloatRegisterResult* ins) { + auto* lir = new (alloc()) LWasmRegisterResult(); + uint32_t vreg = getVirtualRegister(); + auto type = LDefinition::TypeFrom(ins->type()); + lir->setDef(0, LDefinition(vreg, type, LFloatReg(ins->loc()))); + ins->setVirtualRegister(vreg); + add(lir, ins); +} + +void LIRGenerator::visitWasmRegister64Result(MWasmRegister64Result* ins) { + MOZ_ASSERT(ins->type() == MIRType::Int64); + uint32_t vreg = getVirtualRegister(); + +#if defined(JS_NUNBOX32) + auto* lir = new (alloc()) LWasmRegisterPairResult(); + lir->setDef(INT64LOW_INDEX, + LDefinition(vreg + INT64LOW_INDEX, LDefinition::GENERAL, + LGeneralReg(ins->loc().low))); + lir->setDef(INT64HIGH_INDEX, + LDefinition(vreg + INT64HIGH_INDEX, LDefinition::GENERAL, + LGeneralReg(ins->loc().high))); + getVirtualRegister(); +#elif defined(JS_PUNBOX64) + auto* lir = new (alloc()) LWasmRegisterResult(); + lir->setDef( + 0, LDefinition(vreg, LDefinition::GENERAL, LGeneralReg(ins->loc().reg))); +#else +# error expected either JS_NUNBOX32 or JS_PUNBOX64 +#endif + + ins->setVirtualRegister(vreg); + add(lir, ins); +} + +void LIRGenerator::visitWasmStackResultArea(MWasmStackResultArea* ins) { + MOZ_ASSERT(ins->type() == MIRType::StackResults); + auto* lir = new (alloc()) LWasmStackResultArea(temp()); + uint32_t vreg = getVirtualRegister(); + lir->setDef(0, + LDefinition(vreg, LDefinition::STACKRESULTS, LDefinition::STACK)); + ins->setVirtualRegister(vreg); + add(lir, ins); +} + +void LIRGenerator::visitWasmStackResult(MWasmStackResult* ins) { + MWasmStackResultArea* area = ins->resultArea()->toWasmStackResultArea(); + LDefinition::Policy pol = LDefinition::STACK; + + if (ins->type() == MIRType::Int64) { + auto* lir = new (alloc()) LWasmStackResult64; + lir->setOperand(0, use(area, LUse(LUse::STACK, /* usedAtStart = */ true))); + uint32_t vreg = getVirtualRegister(); + LDefinition::Type typ = LDefinition::GENERAL; +#if defined(JS_NUNBOX32) + getVirtualRegister(); + lir->setDef(INT64LOW_INDEX, LDefinition(vreg + INT64LOW_INDEX, typ, pol)); + lir->setDef(INT64HIGH_INDEX, LDefinition(vreg + INT64HIGH_INDEX, typ, pol)); +#else + lir->setDef(0, LDefinition(vreg, typ, pol)); +#endif + ins->setVirtualRegister(vreg); + add(lir, ins); + return; + } + + auto* lir = new (alloc()) LWasmStackResult; + lir->setOperand(0, use(area, LUse(LUse::STACK, /* usedAtStart = */ true))); + uint32_t vreg = getVirtualRegister(); + LDefinition::Type typ = LDefinition::TypeFrom(ins->type()); + lir->setDef(0, LDefinition(vreg, typ, pol)); + ins->setVirtualRegister(vreg); + add(lir, ins); +} + +template +void LIRGenerator::visitWasmCall(MWasmCallT ins) { + bool needsBoundsCheck = true; + mozilla::Maybe tableSize; + + if (ins->callee().isTable()) { + MDefinition* index = ins->getOperand(ins->numArgs()); + + if (ins->callee().which() == wasm::CalleeDesc::WasmTable) { + uint32_t minLength = ins->callee().wasmTableMinLength(); + mozilla::Maybe maxLength = ins->callee().wasmTableMaxLength(); + if (index->isConstant() && + uint32_t(index->toConstant()->toInt32()) < minLength) { + needsBoundsCheck = false; + } + if (maxLength.isSome() && *maxLength == minLength) { + tableSize = maxLength; + } + } + } + + auto* lir = allocateVariadic(ins->numOperands(), needsBoundsCheck, + tableSize); + if (!lir) { + abort(AbortReason::Alloc, "OOM: LIRGenerator::lowerWasmCall"); + return; + } + + for (unsigned i = 0; i < ins->numArgs(); i++) { + lir->setOperand( + i, useFixedAtStart(ins->getOperand(i), ins->registerForArg(i))); + } + + if (ins->callee().isTable()) { + MDefinition* index = ins->getOperand(ins->numArgs()); + lir->setOperand(ins->numArgs(), + useFixedAtStart(index, WasmTableCallIndexReg)); + } + if (ins->callee().isFuncRef()) { + MDefinition* ref = ins->getOperand(ins->numArgs()); + lir->setOperand(ins->numArgs(), useFixedAtStart(ref, WasmCallRefReg)); + } + + add(lir, ins); + assignWasmSafepoint(lir); + + // WasmCall with WasmTable has two call instructions, and they both need a + // safepoint associated with them. Create a second safepoint here; the node + // otherwise does nothing, and codegen for it only marks the safepoint at the + // node. + if (ins->callee().which() == wasm::CalleeDesc::WasmTable) { + auto* adjunctSafepoint = new (alloc()) LWasmCallIndirectAdjunctSafepoint(); + add(adjunctSafepoint); + assignWasmSafepoint(adjunctSafepoint); + lir->setAdjunctSafepoint(adjunctSafepoint); + } +} + +void LIRGenerator::visitWasmCallCatchable(MWasmCallCatchable* ins) { + visitWasmCall(ins); +} + +void LIRGenerator::visitWasmCallUncatchable(MWasmCallUncatchable* ins) { + visitWasmCall(ins); +} + +void LIRGenerator::visitWasmCallLandingPrePad(MWasmCallLandingPrePad* ins) { + add(new (alloc()) LWasmCallLandingPrePad, ins); +} + +void LIRGenerator::visitSetDOMProperty(MSetDOMProperty* ins) { + MDefinition* val = ins->value(); + + Register cxReg, objReg, privReg, valueReg; + GetTempRegForIntArg(0, 0, &cxReg); + GetTempRegForIntArg(1, 0, &objReg); + GetTempRegForIntArg(2, 0, &privReg); + GetTempRegForIntArg(3, 0, &valueReg); + + // Keep using GetTempRegForIntArg, since we want to make sure we + // don't clobber registers we're already using. + Register tempReg1, tempReg2; + GetTempRegForIntArg(4, 0, &tempReg1); + mozilla::DebugOnly ok = GetTempRegForIntArg(5, 0, &tempReg2); + MOZ_ASSERT(ok, "How can we not have six temp registers?"); + + LSetDOMProperty* lir = new (alloc()) + LSetDOMProperty(tempFixed(cxReg), useFixedAtStart(ins->object(), objReg), + useBoxFixedAtStart(val, tempReg1, tempReg2), + tempFixed(privReg), tempFixed(valueReg)); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetDOMProperty(MGetDOMProperty* ins) { + Register cxReg, objReg, privReg, valueReg; + GetTempRegForIntArg(0, 0, &cxReg); + GetTempRegForIntArg(1, 0, &objReg); + GetTempRegForIntArg(2, 0, &privReg); + mozilla::DebugOnly ok = GetTempRegForIntArg(3, 0, &valueReg); + MOZ_ASSERT(ok, "How can we not have four temp registers?"); + LGetDOMProperty* lir = new (alloc()) + LGetDOMProperty(tempFixed(cxReg), useFixedAtStart(ins->object(), objReg), + tempFixed(privReg), tempFixed(valueReg)); + + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetDOMMember(MGetDOMMember* ins) { + MOZ_ASSERT(ins->isDomMovable(), "Members had better be movable"); + // We wish we could assert that ins->domAliasSet() == JSJitInfo::AliasNone, + // but some MGetDOMMembers are for [Pure], not [Constant] properties, whose + // value can in fact change as a result of DOM setters and method calls. + MOZ_ASSERT(ins->domAliasSet() != JSJitInfo::AliasEverything, + "Member gets had better not alias the world"); + + MDefinition* obj = ins->object(); + MOZ_ASSERT(obj->type() == MIRType::Object); + + MIRType type = ins->type(); + + if (type == MIRType::Value) { + LGetDOMMemberV* lir = new (alloc()) LGetDOMMemberV(useRegisterAtStart(obj)); + defineBox(lir, ins); + } else { + LGetDOMMemberT* lir = + new (alloc()) LGetDOMMemberT(useRegisterForTypedLoad(obj, type)); + define(lir, ins); + } +} + +void LIRGenerator::visitLoadDOMExpandoValue(MLoadDOMExpandoValue* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + auto* lir = + new (alloc()) LLoadDOMExpandoValue(useRegisterAtStart(ins->proxy())); + defineBox(lir, ins); +} + +void LIRGenerator::visitLoadDOMExpandoValueGuardGeneration( + MLoadDOMExpandoValueGuardGeneration* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + auto* lir = new (alloc()) + LLoadDOMExpandoValueGuardGeneration(useRegisterAtStart(ins->proxy())); + assignSnapshot(lir, ins->bailoutKind()); + defineBox(lir, ins); +} + +void LIRGenerator::visitLoadDOMExpandoValueIgnoreGeneration( + MLoadDOMExpandoValueIgnoreGeneration* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + auto* lir = new (alloc()) + LLoadDOMExpandoValueIgnoreGeneration(useRegisterAtStart(ins->proxy())); + defineBox(lir, ins); +} + +void LIRGenerator::visitGuardDOMExpandoMissingOrGuardShape( + MGuardDOMExpandoMissingOrGuardShape* ins) { + MOZ_ASSERT(ins->expando()->type() == MIRType::Value); + auto* lir = new (alloc()) + LGuardDOMExpandoMissingOrGuardShape(useBox(ins->expando()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->expando()); +} + +void LIRGenerator::visitIncrementWarmUpCounter(MIncrementWarmUpCounter* ins) { + LIncrementWarmUpCounter* lir = new (alloc()) LIncrementWarmUpCounter(temp()); + add(lir, ins); +} + +void LIRGenerator::visitLexicalCheck(MLexicalCheck* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::Value); + LLexicalCheck* lir = new (alloc()) LLexicalCheck(useBox(input)); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, input); +} + +void LIRGenerator::visitThrowRuntimeLexicalError( + MThrowRuntimeLexicalError* ins) { + LThrowRuntimeLexicalError* lir = new (alloc()) LThrowRuntimeLexicalError(); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitThrowMsg(MThrowMsg* ins) { + LThrowMsg* lir = new (alloc()) LThrowMsg(); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGlobalDeclInstantiation(MGlobalDeclInstantiation* ins) { + LGlobalDeclInstantiation* lir = new (alloc()) LGlobalDeclInstantiation(); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitDebugger(MDebugger* ins) { + auto* lir = new (alloc()) LDebugger(tempFixed(CallTempReg0)); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); +} + +void LIRGenerator::visitAtomicIsLockFree(MAtomicIsLockFree* ins) { + define(new (alloc()) LAtomicIsLockFree(useRegister(ins->input())), ins); +} + +void LIRGenerator::visitCheckReturn(MCheckReturn* ins) { + MDefinition* retVal = ins->returnValue(); + MDefinition* thisVal = ins->thisValue(); + MOZ_ASSERT(retVal->type() == MIRType::Value); + MOZ_ASSERT(thisVal->type() == MIRType::Value); + + auto* lir = + new (alloc()) LCheckReturn(useBoxAtStart(retVal), useBoxAtStart(thisVal)); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckIsObj(MCheckIsObj* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::Value); + + LCheckIsObj* lir = new (alloc()) LCheckIsObj(useBox(input)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckObjCoercible(MCheckObjCoercible* ins) { + MDefinition* checkVal = ins->checkValue(); + MOZ_ASSERT(checkVal->type() == MIRType::Value); + + auto* lir = new (alloc()) LCheckObjCoercible(useBoxAtStart(checkVal)); + redefine(ins, checkVal); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckClassHeritage(MCheckClassHeritage* ins) { + MDefinition* heritage = ins->heritage(); + MOZ_ASSERT(heritage->type() == MIRType::Value); + + auto* lir = + new (alloc()) LCheckClassHeritage(useBox(heritage), temp(), temp()); + redefine(ins, heritage); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckThis(MCheckThis* ins) { + MDefinition* thisValue = ins->thisValue(); + MOZ_ASSERT(thisValue->type() == MIRType::Value); + + auto* lir = new (alloc()) LCheckThis(useBoxAtStart(thisValue)); + redefine(ins, thisValue); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckThisReinit(MCheckThisReinit* ins) { + MDefinition* thisValue = ins->thisValue(); + MOZ_ASSERT(thisValue->type() == MIRType::Value); + + auto* lir = new (alloc()) LCheckThisReinit(useBoxAtStart(thisValue)); + redefine(ins, thisValue); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGenerator(MGenerator* ins) { + auto* lir = + new (alloc()) LGenerator(useRegisterAtStart(ins->callee()), + useRegisterAtStart(ins->environmentChain()), + useRegisterAtStart(ins->argsObject())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitAsyncResolve(MAsyncResolve* ins) { + auto* lir = new (alloc()) LAsyncResolve(useRegisterAtStart(ins->generator()), + useBoxAtStart(ins->valueOrReason())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitAsyncAwait(MAsyncAwait* ins) { + MOZ_ASSERT(ins->generator()->type() == MIRType::Object); + auto* lir = new (alloc()) LAsyncAwait(useBoxAtStart(ins->value()), + useRegisterAtStart(ins->generator())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCanSkipAwait(MCanSkipAwait* ins) { + auto* lir = new (alloc()) LCanSkipAwait(useBoxAtStart(ins->value())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitMaybeExtractAwaitValue(MMaybeExtractAwaitValue* ins) { + auto* lir = new (alloc()) LMaybeExtractAwaitValue( + useBoxAtStart(ins->value()), useRegisterAtStart(ins->canSkip())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitDebugCheckSelfHosted(MDebugCheckSelfHosted* ins) { + MDefinition* checkVal = ins->checkValue(); + MOZ_ASSERT(checkVal->type() == MIRType::Value); + + LDebugCheckSelfHosted* lir = + new (alloc()) LDebugCheckSelfHosted(useBoxAtStart(checkVal)); + redefine(ins, checkVal); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitIsPackedArray(MIsPackedArray* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + + auto lir = new (alloc()) LIsPackedArray(useRegister(ins->object()), temp()); + define(lir, ins); +} + +void LIRGenerator::visitGuardArrayIsPacked(MGuardArrayIsPacked* ins) { + MOZ_ASSERT(ins->array()->type() == MIRType::Object); + + auto* lir = new (alloc()) + LGuardArrayIsPacked(useRegister(ins->array()), temp(), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->array()); +} + +void LIRGenerator::visitGetPrototypeOf(MGetPrototypeOf* ins) { + MOZ_ASSERT(ins->target()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Value); + + auto lir = new (alloc()) LGetPrototypeOf(useRegister(ins->target())); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitObjectWithProto(MObjectWithProto* ins) { + MOZ_ASSERT(ins->prototype()->type() == MIRType::Value); + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto* lir = new (alloc()) LObjectWithProto(useBoxAtStart(ins->prototype())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitObjectStaticProto(MObjectStaticProto* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto* lir = + new (alloc()) LObjectStaticProto(useRegisterAtStart(ins->object())); + define(lir, ins); +}; + +void LIRGenerator::visitBuiltinObject(MBuiltinObject* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto* lir = new (alloc()) LBuiltinObject(); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitReturn(MReturn* ret) { + return visitReturnImpl(ret->getOperand(0)); +} + +void LIRGenerator::visitGeneratorReturn(MGeneratorReturn* ret) { + return visitReturnImpl(ret->getOperand(0), true); +} + +void LIRGenerator::visitSuperFunction(MSuperFunction* ins) { + MOZ_ASSERT(ins->callee()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Value); + + auto* lir = new (alloc()) LSuperFunction(useRegister(ins->callee()), temp()); + defineBox(lir, ins); +} + +void LIRGenerator::visitInitHomeObject(MInitHomeObject* ins) { + MDefinition* function = ins->function(); + MOZ_ASSERT(function->type() == MIRType::Object); + + MDefinition* homeObject = ins->homeObject(); + MOZ_ASSERT(homeObject->type() == MIRType::Value); + + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto* lir = new (alloc()) + LInitHomeObject(useRegisterAtStart(function), useBoxAtStart(homeObject)); + redefine(ins, function); + add(lir, ins); +} + +void LIRGenerator::visitIsTypedArrayConstructor(MIsTypedArrayConstructor* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + auto* lir = new (alloc()) LIsTypedArrayConstructor(useRegister(object)); + define(lir, ins); +} + +void LIRGenerator::visitLoadValueTag(MLoadValueTag* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Value); + + define(new (alloc()) LLoadValueTag(useBoxAtStart(value)), ins); +} + +void LIRGenerator::visitGuardTagNotEqual(MGuardTagNotEqual* ins) { + MDefinition* lhs = ins->lhs(); + MOZ_ASSERT(lhs->type() == MIRType::Int32); + + MDefinition* rhs = ins->rhs(); + MOZ_ASSERT(rhs->type() == MIRType::Int32); + + auto* guard = + new (alloc()) LGuardTagNotEqual(useRegister(lhs), useRegister(rhs)); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); +} + +void LIRGenerator::visitLoadWrapperTarget(MLoadWrapperTarget* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + define(new (alloc()) LLoadWrapperTarget(useRegisterAtStart(object)), ins); +} + +void LIRGenerator::visitGuardHasGetterSetter(MGuardHasGetterSetter* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + auto* guard = new (alloc()) + LGuardHasGetterSetter(useRegisterAtStart(object), tempFixed(CallTempReg0), + tempFixed(CallTempReg1), tempFixed(CallTempReg2)); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, object); +} + +void LIRGenerator::visitGuardIsExtensible(MGuardIsExtensible* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + auto* guard = new (alloc()) LGuardIsExtensible(useRegister(object), temp()); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, object); +} + +void LIRGenerator::visitGuardInt32IsNonNegative(MGuardInt32IsNonNegative* ins) { + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* guard = new (alloc()) LGuardInt32IsNonNegative(useRegister(index)); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, index); +} + +void LIRGenerator::visitGuardInt32Range(MGuardInt32Range* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::Int32); + + auto* guard = new (alloc()) LGuardInt32Range(useRegister(input)); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, input); +} + +void LIRGenerator::visitGuardIndexIsNotDenseElement( + MGuardIndexIsNotDenseElement* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + LDefinition spectreTemp = + BoundsCheckNeedsSpectreTemp() ? temp() : LDefinition::BogusTemp(); + + auto* guard = new (alloc()) LGuardIndexIsNotDenseElement( + useRegister(object), useRegister(index), temp(), spectreTemp); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, index); +} + +void LIRGenerator::visitGuardIndexIsValidUpdateOrAdd( + MGuardIndexIsValidUpdateOrAdd* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + LDefinition spectreTemp = + BoundsCheckNeedsSpectreTemp() ? temp() : LDefinition::BogusTemp(); + + auto* guard = new (alloc()) LGuardIndexIsValidUpdateOrAdd( + useRegister(object), useRegister(index), temp(), spectreTemp); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, index); +} + +void LIRGenerator::visitCallAddOrUpdateSparseElement( + MCallAddOrUpdateSparseElement* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Value); + + auto* lir = new (alloc()) LCallAddOrUpdateSparseElement( + useRegisterAtStart(object), useRegisterAtStart(index), + useBoxAtStart(value)); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallGetSparseElement(MCallGetSparseElement* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) LCallGetSparseElement(useRegisterAtStart(object), + useRegisterAtStart(index)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallNativeGetElement(MCallNativeGetElement* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) LCallNativeGetElement(useRegisterAtStart(object), + useRegisterAtStart(index)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallNativeGetElementSuper( + MCallNativeGetElementSuper* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + MDefinition* receiver = ins->receiver(); + + auto* lir = new (alloc()) LCallNativeGetElementSuper( + useRegisterAtStart(object), useRegisterAtStart(index), + useBoxAtStart(receiver)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallObjectHasSparseElement( + MCallObjectHasSparseElement* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) LCallObjectHasSparseElement( + useRegisterAtStart(object), useRegisterAtStart(index), + tempFixed(CallTempReg0), tempFixed(CallTempReg1)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); +} + +void LIRGenerator::visitBigIntAsIntN(MBigIntAsIntN* ins) { + MOZ_ASSERT(ins->bits()->type() == MIRType::Int32); + MOZ_ASSERT(ins->input()->type() == MIRType::BigInt); + + if (ins->bits()->isConstant()) { + int32_t bits = ins->bits()->toConstant()->toInt32(); + if (bits == 64) { + auto* lir = new (alloc()) + LBigIntAsIntN64(useRegister(ins->input()), temp(), tempInt64()); + define(lir, ins); + assignSafepoint(lir, ins); + return; + } + if (bits == 32) { + auto* lir = new (alloc()) + LBigIntAsIntN32(useRegister(ins->input()), temp(), tempInt64()); + define(lir, ins); + assignSafepoint(lir, ins); + return; + } + } + + auto* lir = new (alloc()) LBigIntAsIntN(useRegisterAtStart(ins->bits()), + useRegisterAtStart(ins->input())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntAsUintN(MBigIntAsUintN* ins) { + MOZ_ASSERT(ins->bits()->type() == MIRType::Int32); + MOZ_ASSERT(ins->input()->type() == MIRType::BigInt); + + if (ins->bits()->isConstant()) { + int32_t bits = ins->bits()->toConstant()->toInt32(); + if (bits == 64) { + auto* lir = new (alloc()) + LBigIntAsUintN64(useRegister(ins->input()), temp(), tempInt64()); + define(lir, ins); + assignSafepoint(lir, ins); + return; + } + if (bits == 32) { + auto* lir = new (alloc()) + LBigIntAsUintN32(useRegister(ins->input()), temp(), tempInt64()); + define(lir, ins); + assignSafepoint(lir, ins); + return; + } + } + + auto* lir = new (alloc()) LBigIntAsUintN(useRegisterAtStart(ins->bits()), + useRegisterAtStart(ins->input())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGuardNonGCThing(MGuardNonGCThing* ins) { + MDefinition* input = ins->input(); + + auto* guard = new (alloc()) LGuardNonGCThing(useBox(input)); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, input); +} + +void LIRGenerator::visitToHashableNonGCThing(MToHashableNonGCThing* ins) { + auto* lir = + new (alloc()) LToHashableNonGCThing(useBox(ins->input()), tempDouble()); + defineBox(lir, ins); +} + +void LIRGenerator::visitToHashableString(MToHashableString* ins) { + auto* lir = new (alloc()) LToHashableString(useRegister(ins->input())); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitToHashableValue(MToHashableValue* ins) { + auto* lir = + new (alloc()) LToHashableValue(useBox(ins->input()), tempDouble()); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitHashNonGCThing(MHashNonGCThing* ins) { + auto* lir = new (alloc()) LHashNonGCThing(useBox(ins->input()), temp()); + define(lir, ins); +} + +void LIRGenerator::visitHashString(MHashString* ins) { + auto* lir = new (alloc()) LHashString(useRegister(ins->input()), temp()); + define(lir, ins); +} + +void LIRGenerator::visitHashSymbol(MHashSymbol* ins) { + auto* lir = new (alloc()) LHashSymbol(useRegister(ins->input())); + define(lir, ins); +} + +void LIRGenerator::visitHashBigInt(MHashBigInt* ins) { + auto* lir = new (alloc()) + LHashBigInt(useRegister(ins->input()), temp(), temp(), temp()); + define(lir, ins); +} + +void LIRGenerator::visitHashObject(MHashObject* ins) { + auto* lir = + new (alloc()) LHashObject(useRegister(ins->set()), useBox(ins->input()), + temp(), temp(), temp(), temp()); + define(lir, ins); +} + +void LIRGenerator::visitHashValue(MHashValue* ins) { + auto* lir = + new (alloc()) LHashValue(useRegister(ins->set()), useBox(ins->input()), + temp(), temp(), temp(), temp()); + define(lir, ins); +} + +void LIRGenerator::visitSetObjectHasNonBigInt(MSetObjectHasNonBigInt* ins) { + auto* lir = new (alloc()) + LSetObjectHasNonBigInt(useRegister(ins->set()), useBox(ins->value()), + useRegister(ins->hash()), temp(), temp()); + define(lir, ins); +} + +void LIRGenerator::visitSetObjectHasBigInt(MSetObjectHasBigInt* ins) { + auto* lir = new (alloc()) LSetObjectHasBigInt( + useRegister(ins->set()), useBox(ins->value()), useRegister(ins->hash()), + temp(), temp(), temp(), temp()); + define(lir, ins); +} + +void LIRGenerator::visitSetObjectHasValue(MSetObjectHasValue* ins) { + auto* lir = new (alloc()) LSetObjectHasValue( + useRegister(ins->set()), useBox(ins->value()), useRegister(ins->hash()), + temp(), temp(), temp(), temp()); + define(lir, ins); +} + +void LIRGenerator::visitSetObjectHasValueVMCall(MSetObjectHasValueVMCall* ins) { + auto* lir = new (alloc()) LSetObjectHasValueVMCall( + useRegisterAtStart(ins->set()), useBoxAtStart(ins->value())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitSetObjectSize(MSetObjectSize* ins) { + auto* lir = new (alloc()) LSetObjectSize(useRegisterAtStart(ins->set())); + define(lir, ins); +} + +void LIRGenerator::visitMapObjectHasNonBigInt(MMapObjectHasNonBigInt* ins) { + auto* lir = new (alloc()) + LMapObjectHasNonBigInt(useRegister(ins->map()), useBox(ins->value()), + useRegister(ins->hash()), temp(), temp()); + define(lir, ins); +} + +void LIRGenerator::visitMapObjectHasBigInt(MMapObjectHasBigInt* ins) { + auto* lir = new (alloc()) LMapObjectHasBigInt( + useRegister(ins->map()), useBox(ins->value()), useRegister(ins->hash()), + temp(), temp(), temp(), temp()); + define(lir, ins); +} + +void LIRGenerator::visitMapObjectHasValue(MMapObjectHasValue* ins) { + auto* lir = new (alloc()) LMapObjectHasValue( + useRegister(ins->map()), useBox(ins->value()), useRegister(ins->hash()), + temp(), temp(), temp(), temp()); + define(lir, ins); +} + +void LIRGenerator::visitMapObjectHasValueVMCall(MMapObjectHasValueVMCall* ins) { + auto* lir = new (alloc()) LMapObjectHasValueVMCall( + useRegisterAtStart(ins->map()), useBoxAtStart(ins->value())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitMapObjectGetNonBigInt(MMapObjectGetNonBigInt* ins) { + auto* lir = new (alloc()) + LMapObjectGetNonBigInt(useRegister(ins->map()), useBox(ins->value()), + useRegister(ins->hash()), temp(), temp()); + defineBox(lir, ins); +} + +void LIRGenerator::visitMapObjectGetBigInt(MMapObjectGetBigInt* ins) { + auto* lir = new (alloc()) LMapObjectGetBigInt( + useRegister(ins->map()), useBox(ins->value()), useRegister(ins->hash()), + temp(), temp(), temp(), temp()); + defineBox(lir, ins); +} + +void LIRGenerator::visitMapObjectGetValue(MMapObjectGetValue* ins) { + auto* lir = new (alloc()) LMapObjectGetValue( + useRegister(ins->map()), useBox(ins->value()), useRegister(ins->hash()), + temp(), temp(), temp(), temp()); + defineBox(lir, ins); +} + +void LIRGenerator::visitMapObjectGetValueVMCall(MMapObjectGetValueVMCall* ins) { + auto* lir = new (alloc()) LMapObjectGetValueVMCall( + useRegisterAtStart(ins->map()), useBoxAtStart(ins->value())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitMapObjectSize(MMapObjectSize* ins) { + auto* lir = new (alloc()) LMapObjectSize(useRegisterAtStart(ins->map())); + define(lir, ins); +} + +void LIRGenerator::visitConstant(MConstant* ins) { + if (!IsFloatingPointType(ins->type()) && ins->canEmitAtUses()) { + emitAtUses(ins); + return; + } + + switch (ins->type()) { + case MIRType::Double: + define(new (alloc()) LDouble(ins->toDouble()), ins); + break; + case MIRType::Float32: + define(new (alloc()) LFloat32(ins->toFloat32()), ins); + break; + case MIRType::Boolean: + define(new (alloc()) LInteger(ins->toBoolean()), ins); + break; + case MIRType::Int32: + define(new (alloc()) LInteger(ins->toInt32()), ins); + break; + case MIRType::Int64: + defineInt64(new (alloc()) LInteger64(ins->toInt64()), ins); + break; + case MIRType::IntPtr: +#ifdef JS_64BIT + defineInt64(new (alloc()) LInteger64(ins->toIntPtr()), ins); +#else + define(new (alloc()) LInteger(ins->toIntPtr()), ins); +#endif + break; + case MIRType::String: + define(new (alloc()) LPointer(ins->toString()), ins); + break; + case MIRType::Symbol: + define(new (alloc()) LPointer(ins->toSymbol()), ins); + break; + case MIRType::BigInt: + define(new (alloc()) LPointer(ins->toBigInt()), ins); + break; + case MIRType::Object: + define(new (alloc()) LPointer(&ins->toObject()), ins); + break; + case MIRType::Shape: + MOZ_ASSERT(ins->isEmittedAtUses()); + break; + default: + // Constants of special types (undefined, null) should never flow into + // here directly. Operations blindly consuming them require a Box. + MOZ_CRASH("unexpected constant type"); + } +} + +void LIRGenerator::visitConstantProto(MConstantProto* ins) { + JSObject* obj = &ins->protoObject()->toConstant()->toObject(); + define(new (alloc()) LPointer(obj), ins); +} + +void LIRGenerator::visitWasmNullConstant(MWasmNullConstant* ins) { + define(new (alloc()) LWasmNullConstant(), ins); +} + +void LIRGenerator::visitWasmFloatConstant(MWasmFloatConstant* ins) { + switch (ins->type()) { + case MIRType::Double: + define(new (alloc()) LDouble(ins->toDouble()), ins); + break; + case MIRType::Float32: + define(new (alloc()) LFloat32(ins->toFloat32()), ins); + break; +#ifdef ENABLE_WASM_SIMD + case MIRType::Simd128: + define(new (alloc()) LSimd128(ins->toSimd128()), ins); + break; +#endif + default: + MOZ_CRASH("unexpected constant type"); + } +} + +#ifdef JS_JITSPEW +static void SpewResumePoint(MBasicBlock* block, MInstruction* ins, + MResumePoint* resumePoint) { + Fprinter& out = JitSpewPrinter(); + out.printf("Current resume point %p details:\n", (void*)resumePoint); + out.printf(" frame count: %u\n", resumePoint->frameCount()); + + if (ins) { + out.printf(" taken after: "); + ins->printName(out); + } else { + out.printf(" taken at block %u entry", block->id()); + } + out.printf("\n"); + + out.printf(" pc: %p (script: %p, offset: %d)\n", (void*)resumePoint->pc(), + (void*)resumePoint->block()->info().script(), + int(resumePoint->block()->info().script()->pcToOffset( + resumePoint->pc()))); + + for (size_t i = 0, e = resumePoint->numOperands(); i < e; i++) { + MDefinition* in = resumePoint->getOperand(i); + out.printf(" slot%u: ", (unsigned)i); + in->printName(out); + out.printf("\n"); + } +} +#endif + +void LIRGenerator::visitInstructionDispatch(MInstruction* ins) { +#ifdef JS_CODEGEN_NONE + // Don't compile the switch-statement below so that we don't have to define + // the platform-specific visit* methods for the none-backend. + MOZ_CRASH(); +#else + switch (ins->op()) { +# define MIR_OP(op) \ + case MDefinition::Opcode::op: \ + visit##op(ins->to##op()); \ + break; + MIR_OPCODE_LIST(MIR_OP) +# undef MIR_OP + default: + MOZ_CRASH("Invalid instruction"); + } +#endif +} + +void LIRGeneratorShared::visitEmittedAtUses(MInstruction* ins) { + static_cast(this)->visitInstructionDispatch(ins); +} + +bool LIRGenerator::visitInstruction(MInstruction* ins) { + MOZ_ASSERT(!errored()); + + if (ins->isRecoveredOnBailout()) { + MOZ_ASSERT(!JitOptions.disableRecoverIns); + return true; + } + + if (!gen->ensureBallast()) { + return false; + } + visitInstructionDispatch(ins); + + if (ins->resumePoint()) { + updateResumeState(ins); + } + +#ifdef DEBUG + ins->setInWorklistUnchecked(); +#endif + + // If no safepoint was created, there's no need for an OSI point. + if (LOsiPoint* osiPoint = popOsiPoint()) { + add(osiPoint); + } + + return !errored(); +} + +bool LIRGenerator::definePhis() { + size_t lirIndex = 0; + MBasicBlock* block = current->mir(); + for (MPhiIterator phi(block->phisBegin()); phi != block->phisEnd(); phi++) { + if (phi->type() == MIRType::Value) { + defineUntypedPhi(*phi, lirIndex); + lirIndex += BOX_PIECES; + } else if (phi->type() == MIRType::Int64) { + defineInt64Phi(*phi, lirIndex); + lirIndex += INT64_PIECES; + } else { + defineTypedPhi(*phi, lirIndex); + lirIndex += 1; + } + } + return !errored(); +} + +void LIRGenerator::updateResumeState(MInstruction* ins) { + lastResumePoint_ = ins->resumePoint(); +#ifdef JS_JITSPEW + if (JitSpewEnabled(JitSpew_IonSnapshots) && lastResumePoint_) { + SpewResumePoint(nullptr, ins, lastResumePoint_); + } +#endif +} + +void LIRGenerator::updateResumeState(MBasicBlock* block) { + // Note: RangeAnalysis can flag blocks as unreachable, but they are only + // removed iff GVN (including UCE) is enabled. + MOZ_ASSERT_IF(!mir()->compilingWasm() && !block->unreachable(), + block->entryResumePoint()); + MOZ_ASSERT_IF(block->unreachable(), !mir()->optimizationInfo().gvnEnabled()); + lastResumePoint_ = block->entryResumePoint(); +#ifdef JS_JITSPEW + if (JitSpewEnabled(JitSpew_IonSnapshots) && lastResumePoint_) { + SpewResumePoint(block, nullptr, lastResumePoint_); + } +#endif +} + +bool LIRGenerator::visitBlock(MBasicBlock* block) { + current = block->lir(); + updateResumeState(block); + + if (!definePhis()) { + return false; + } + + MOZ_ASSERT_IF(block->unreachable(), !mir()->optimizationInfo().gvnEnabled()); + for (MInstructionIterator iter = block->begin(); *iter != block->lastIns(); + iter++) { + if (!visitInstruction(*iter)) { + return false; + } + } + + if (block->successorWithPhis()) { + // If we have a successor with phis, lower the phi input now that we + // are approaching the join point. + MBasicBlock* successor = block->successorWithPhis(); + uint32_t position = block->positionInPhiSuccessor(); + size_t lirIndex = 0; + for (MPhiIterator phi(successor->phisBegin()); phi != successor->phisEnd(); + phi++) { + if (!gen->ensureBallast()) { + return false; + } + + MDefinition* opd = phi->getOperand(position); + ensureDefined(opd); + + MOZ_ASSERT(opd->type() == phi->type()); + + if (phi->type() == MIRType::Value) { + lowerUntypedPhiInput(*phi, position, successor->lir(), lirIndex); + lirIndex += BOX_PIECES; + } else if (phi->type() == MIRType::Int64) { + lowerInt64PhiInput(*phi, position, successor->lir(), lirIndex); + lirIndex += INT64_PIECES; + } else { + lowerTypedPhiInput(*phi, position, successor->lir(), lirIndex); + lirIndex += 1; + } + } + } + + // Now emit the last instruction, which is some form of branch. + if (!visitInstruction(block->lastIns())) { + return false; + } + + return true; +} + +void LIRGenerator::visitNaNToZero(MNaNToZero* ins) { + MDefinition* input = ins->input(); + + if (ins->operandIsNeverNaN() && ins->operandIsNeverNegativeZero()) { + redefine(ins, input); + return; + } + LNaNToZero* lir = + new (alloc()) LNaNToZero(useRegisterAtStart(input), tempDouble()); + defineReuseInput(lir, ins, 0); +} + +bool LIRGenerator::generate() { + // Create all blocks and prep all phis beforehand. + for (ReversePostorderIterator block(graph.rpoBegin()); + block != graph.rpoEnd(); block++) { + if (gen->shouldCancel("Lowering (preparation loop)")) { + return false; + } + + if (!lirGraph_.initBlock(*block)) { + return false; + } + } + + for (ReversePostorderIterator block(graph.rpoBegin()); + block != graph.rpoEnd(); block++) { + if (gen->shouldCancel("Lowering (main loop)")) { + return false; + } + + if (!visitBlock(*block)) { + return false; + } + } + + lirGraph_.setArgumentSlotCount(maxargslots_); + return true; +} + +void LIRGenerator::visitPhi(MPhi* phi) { + // Phi nodes are not lowered because they are only meaningful for the register + // allocator. + MOZ_CRASH("Unexpected Phi node during Lowering."); +} + +void LIRGenerator::visitBeta(MBeta* beta) { + // Beta nodes are supposed to be removed before because they are + // only used to carry the range information for Range analysis + MOZ_CRASH("Unexpected Beta node during Lowering."); +} + +void LIRGenerator::visitObjectState(MObjectState* objState) { + // ObjectState nodes are always recovered on bailouts + MOZ_CRASH("Unexpected ObjectState node during Lowering."); +} + +void LIRGenerator::visitArrayState(MArrayState* objState) { + // ArrayState nodes are always recovered on bailouts + MOZ_CRASH("Unexpected ArrayState node during Lowering."); +} + +void LIRGenerator::visitIonToWasmCall(MIonToWasmCall* ins) { + // The instruction needs a temp register: + // - that's not the FramePointer, since wasm is going to use it in the + // function. + // - that's not aliasing an input register. + LDefinition scratch = tempFixed(ABINonArgReg0); + + // Note that since this is a LIR call instruction, regalloc will prevent + // the use*AtStart below from reusing any of the temporaries. + + LInstruction* lir; + if (ins->type() == MIRType::Value) { + lir = allocateVariadic(ins->numOperands(), scratch); + } else if (ins->type() == MIRType::Int64) { + lir = allocateVariadic(ins->numOperands(), scratch); + } else { + lir = allocateVariadic(ins->numOperands(), scratch); + } + if (!lir) { + abort(AbortReason::Alloc, "OOM: LIRGenerator::visitIonToWasmCall"); + return; + } + + ABIArgGenerator abi; + for (unsigned i = 0; i < ins->numOperands(); i++) { + MDefinition* argDef = ins->getOperand(i); + ABIArg arg = abi.next(ToMIRType(argDef->type())); + switch (arg.kind()) { + case ABIArg::GPR: + case ABIArg::FPU: + lir->setOperand(i, useFixedAtStart(argDef, arg.reg())); + break; + case ABIArg::Stack: + lir->setOperand(i, useAtStart(argDef)); + break; +#ifdef JS_CODEGEN_REGISTER_PAIR + case ABIArg::GPR_PAIR: + MOZ_CRASH( + "no way to pass i64, and wasm uses hardfp for function calls"); +#endif + case ABIArg::Uninitialized: + MOZ_CRASH("Uninitialized ABIArg kind"); + } + } + + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitWasmSelect(MWasmSelect* ins) { + MDefinition* condExpr = ins->condExpr(); + + // Pick off specific cases that we can do with LWasmCompareAndSelect to avoid + // generating a boolean that we then have to test again. + if (condExpr->isCompare() && condExpr->isEmittedAtUses()) { + MCompare* comp = condExpr->toCompare(); + MCompare::CompareType compTy = comp->compareType(); + if (canSpecializeWasmCompareAndSelect(compTy, ins->type())) { + JSOp jsop = comp->jsop(); + // We don't currently generate any other JSOPs for the comparison, and if + // that changes, we want to know about it. Hence this assertion. + MOZ_ASSERT(jsop == JSOp::Eq || jsop == JSOp::Ne || jsop == JSOp::Lt || + jsop == JSOp::Gt || jsop == JSOp::Le || jsop == JSOp::Ge); + MDefinition* lhs = comp->lhs(); + MDefinition* rhs = comp->rhs(); + jsop = ReorderComparison(jsop, &lhs, &rhs); + lowerWasmCompareAndSelect(ins, lhs, rhs, compTy, jsop); + return; + } + } + // Fall through to code that generates a boolean and selects on that. + + if (ins->type() == MIRType::Int64) { + lowerWasmSelectI64(ins); + return; + } + + lowerWasmSelectI(ins); +} + +void LIRGenerator::visitWasmFence(MWasmFence* ins) { + add(new (alloc()) LWasmFence, ins); +} + +void LIRGenerator::visitWasmLoadField(MWasmLoadField* ins) { + uint32_t offs = ins->offset(); + LAllocation obj = useRegister(ins->obj()); + MWideningOp wideningOp = ins->wideningOp(); + if (ins->type() == MIRType::Int64) { + MOZ_RELEASE_ASSERT(wideningOp == MWideningOp::None); + defineInt64(new (alloc()) LWasmLoadSlotI64(obj, offs, ins->maybeTrap()), + ins); + } else { + define(new (alloc()) LWasmLoadSlot(obj, offs, ins->type(), wideningOp, + ins->maybeTrap()), + ins); + } +} + +void LIRGenerator::visitWasmLoadFieldKA(MWasmLoadFieldKA* ins) { + uint32_t offs = ins->offset(); + LAllocation obj = useRegister(ins->obj()); + MWideningOp wideningOp = ins->wideningOp(); + if (ins->type() == MIRType::Int64) { + MOZ_RELEASE_ASSERT(wideningOp == MWideningOp::None); + defineInt64(new (alloc()) LWasmLoadSlotI64(obj, offs, ins->maybeTrap()), + ins); + } else { + define(new (alloc()) LWasmLoadSlot(obj, offs, ins->type(), wideningOp, + ins->maybeTrap()), + ins); + } + add(new (alloc()) LKeepAliveObject(useKeepalive(ins->ka())), ins); +} + +void LIRGenerator::visitWasmStoreFieldKA(MWasmStoreFieldKA* ins) { + MDefinition* value = ins->value(); + uint32_t offs = ins->offset(); + MNarrowingOp narrowingOp = ins->narrowingOp(); + LAllocation obj = useRegister(ins->obj()); + LInstruction* lir; + if (value->type() == MIRType::Int64) { + MOZ_RELEASE_ASSERT(narrowingOp == MNarrowingOp::None); + lir = new (alloc()) + LWasmStoreSlotI64(useInt64Register(value), obj, offs, ins->maybeTrap()); + } else { + lir = new (alloc()) + LWasmStoreSlot(useRegister(value), obj, offs, value->type(), + narrowingOp, ins->maybeTrap()); + } + add(lir, ins); + add(new (alloc()) LKeepAliveObject(useKeepalive(ins->ka())), ins); +} + +void LIRGenerator::visitWasmStoreFieldRefKA(MWasmStoreFieldRefKA* ins) { + LAllocation instance = useRegister(ins->instance()); + LAllocation obj = useFixed(ins->obj(), PreBarrierReg); + LAllocation value = useRegister(ins->value()); + uint32_t offset = ins->offset(); + add(new (alloc()) LWasmStoreRef(instance, obj, value, temp(), offset, + ins->maybeTrap(), ins->preBarrierKind()), + ins); + add(new (alloc()) LKeepAliveObject(useKeepalive(ins->ka())), ins); +} + +void LIRGenerator::visitWasmGcObjectIsSubtypeOfAbstract( + MWasmGcObjectIsSubtypeOfAbstract* ins) { + if (CanEmitAtUseForSingleTest(ins)) { + emitAtUses(ins); + return; + } + + // See comment on MacroAssembler::branchWasmGcObjectIsRefType. + // We know we do not need scratch2 and superSuperTypeVector because we know + // this is not a concrete type. + MOZ_ASSERT( + !MacroAssembler::needScratch2ForBranchWasmGcRefType(ins->destType())); + MOZ_ASSERT(!MacroAssembler::needSuperSuperTypeVectorForBranchWasmGcRefType( + ins->destType())); + + LAllocation object = useRegister(ins->object()); + LDefinition scratch1 = + MacroAssembler::needScratch1ForBranchWasmGcRefType(ins->destType()) + ? temp() + : LDefinition(); + define(new (alloc()) LWasmGcObjectIsSubtypeOfAbstract(object, scratch1), ins); +} + +void LIRGenerator::visitWasmGcObjectIsSubtypeOfConcrete( + MWasmGcObjectIsSubtypeOfConcrete* ins) { + if (CanEmitAtUseForSingleTest(ins)) { + emitAtUses(ins); + return; + } + + // See comment on MacroAssembler::branchWasmGcObjectIsRefType. + // We know we need scratch1 and superSuperTypeVector because we know this is a + // concrete type. + MOZ_ASSERT(MacroAssembler::needSuperSuperTypeVectorForBranchWasmGcRefType( + ins->destType())); + MOZ_ASSERT( + MacroAssembler::needScratch1ForBranchWasmGcRefType(ins->destType())); + + LAllocation object = useRegister(ins->object()); + LAllocation superSuperTypeVector = useRegister(ins->superSuperTypeVector()); + LDefinition scratch1 = temp(); + LDefinition scratch2 = + MacroAssembler::needScratch2ForBranchWasmGcRefType(ins->destType()) + ? temp() + : LDefinition(); + define(new (alloc()) LWasmGcObjectIsSubtypeOfConcrete( + object, superSuperTypeVector, scratch1, scratch2), + ins); +} + +#ifdef FUZZING_JS_FUZZILLI +void LIRGenerator::visitFuzzilliHash(MFuzzilliHash* ins) { + MDefinition* value = ins->getOperand(0); + + if (value->type() == MIRType::Undefined || value->type() == MIRType::Null) { + define(new (alloc()) LFuzzilliHashT(LAllocation(), temp(), tempDouble()), + ins); + } else if (value->type() == MIRType::Int32 || + value->type() == MIRType::Double || + value->type() == MIRType::Float32 || + value->type() == MIRType::Boolean || + value->type() == MIRType::BigInt) { + define(new (alloc()) + LFuzzilliHashT(useRegister(value), temp(), tempDouble()), + ins); + } else if (value->type() == MIRType::Object) { + LFuzzilliHashT* lir = + new (alloc()) LFuzzilliHashT(useRegister(value), temp(), tempDouble()); + define(lir, ins); + assignSafepoint(lir, ins); + } else if (value->type() == MIRType::Value) { + LFuzzilliHashV* lir = + new (alloc()) LFuzzilliHashV(useBox(value), temp(), tempDouble()); + define(lir, ins); + assignSafepoint(lir, ins); + } else { + define(new (alloc()) LInteger(0), ins); + } +} + +void LIRGenerator::visitFuzzilliHashStore(MFuzzilliHashStore* ins) { + MDefinition* value = ins->getOperand(0); + MOZ_ASSERT(value->type() == MIRType::Int32); + add(new (alloc()) LFuzzilliHashStore(useRegister(value), temp(), temp()), + ins); +} +#endif + +static_assert(!std::is_polymorphic_v, + "LIRGenerator should not have any virtual methods"); + +#ifdef JS_CODEGEN_NONE +void LIRGenerator::visitReturnImpl(MDefinition*, bool) { MOZ_CRASH(); } +#endif -- cgit v1.2.3