diff options
Diffstat (limited to 'js/src/jit/Lowering.cpp')
-rw-r--r-- | js/src/jit/Lowering.cpp | 5765 |
1 files changed, 5765 insertions, 0 deletions
diff --git a/js/src/jit/Lowering.cpp b/js/src/jit/Lowering.cpp new file mode 100644 index 0000000000..21f31e06f9 --- /dev/null +++ b/js/src/jit/Lowering.cpp @@ -0,0 +1,5765 @@ +/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- + * vim: set ts=8 sts=2 et sw=2 tw=80: + * This Source Code Form is subject to the terms of the Mozilla Public + * License, v. 2.0. If a copy of the MPL was not distributed with this + * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ + +#include "jit/Lowering.h" + +#include "mozilla/DebugOnly.h" +#include "mozilla/EndianUtils.h" +#include "mozilla/FloatingPoint.h" +#include "mozilla/MathAlgorithms.h" + +#include <type_traits> + +#include "jit/IonOptimizationLevels.h" +#include "jit/JitSpewer.h" +#include "jit/LIR.h" +#include "jit/MIR.h" +#include "jit/MIRGraph.h" +#include "jit/SharedICRegisters.h" +#include "js/experimental/JitInfo.h" // JSJitInfo +#include "util/Memory.h" +#include "wasm/WasmTypes.h" + +#include "jit/shared/Lowering-shared-inl.h" +#include "vm/BytecodeUtil-inl.h" +#include "vm/JSObject-inl.h" + +using namespace js; +using namespace jit; + +using JS::GenericNaN; +using mozilla::DebugOnly; + +LBoxAllocation LIRGenerator::useBoxFixedAtStart(MDefinition* mir, + ValueOperand op) { +#if defined(JS_NUNBOX32) + return useBoxFixed(mir, op.typeReg(), op.payloadReg(), true); +#elif defined(JS_PUNBOX64) + return useBoxFixed(mir, op.valueReg(), op.scratchReg(), true); +#endif +} + +LBoxAllocation LIRGenerator::useBoxAtStart(MDefinition* mir, + LUse::Policy policy) { + return useBox(mir, policy, /* useAtStart = */ true); +} + +void LIRGenerator::visitParameter(MParameter* param) { + ptrdiff_t offset; + if (param->index() == MParameter::THIS_SLOT) { + offset = THIS_FRAME_ARGSLOT; + } else { + offset = 1 + param->index(); + } + + LParameter* ins = new (alloc()) LParameter; + defineBox(ins, param, LDefinition::FIXED); + + offset *= sizeof(Value); +#if defined(JS_NUNBOX32) +# if MOZ_BIG_ENDIAN() + ins->getDef(0)->setOutput(LArgument(offset)); + ins->getDef(1)->setOutput(LArgument(offset + 4)); +# else + ins->getDef(0)->setOutput(LArgument(offset + 4)); + ins->getDef(1)->setOutput(LArgument(offset)); +# endif +#elif defined(JS_PUNBOX64) + ins->getDef(0)->setOutput(LArgument(offset)); +#endif +} + +void LIRGenerator::visitCallee(MCallee* ins) { + define(new (alloc()) LCallee(), ins); +} + +void LIRGenerator::visitIsConstructing(MIsConstructing* ins) { + define(new (alloc()) LIsConstructing(), ins); +} + +void LIRGenerator::visitGoto(MGoto* ins) { + add(new (alloc()) LGoto(ins->target())); +} + +void LIRGenerator::visitTableSwitch(MTableSwitch* tableswitch) { + MDefinition* opd = tableswitch->getOperand(0); + + // There should be at least 1 successor. The default case! + MOZ_ASSERT(tableswitch->numSuccessors() > 0); + + // If there are no cases, the default case is always taken. + if (tableswitch->numSuccessors() == 1) { + add(new (alloc()) LGoto(tableswitch->getDefault())); + return; + } + + // If we don't know the type. + if (opd->type() == MIRType::Value) { + LTableSwitchV* lir = newLTableSwitchV(tableswitch); + add(lir); + return; + } + + // Case indices are numeric, so other types will always go to the default + // case. + if (opd->type() != MIRType::Int32 && opd->type() != MIRType::Double) { + add(new (alloc()) LGoto(tableswitch->getDefault())); + return; + } + + // Return an LTableSwitch, capable of handling either an integer or + // floating-point index. + LAllocation index; + LDefinition tempInt; + if (opd->type() == MIRType::Int32) { + index = useRegisterAtStart(opd); + tempInt = tempCopy(opd, 0); + } else { + index = useRegister(opd); + tempInt = temp(LDefinition::GENERAL); + } + add(newLTableSwitch(index, tempInt, tableswitch)); +} + +void LIRGenerator::visitCheckOverRecursed(MCheckOverRecursed* ins) { + LCheckOverRecursed* lir = new (alloc()) LCheckOverRecursed(); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewArray(MNewArray* ins) { + LNewArray* lir = new (alloc()) LNewArray(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewArrayDynamicLength(MNewArrayDynamicLength* ins) { + MDefinition* length = ins->length(); + MOZ_ASSERT(length->type() == MIRType::Int32); + + LNewArrayDynamicLength* lir = + new (alloc()) LNewArrayDynamicLength(useRegister(length), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewIterator(MNewIterator* ins) { + LNewIterator* lir = new (alloc()) LNewIterator(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewTypedArray(MNewTypedArray* ins) { + LNewTypedArray* lir = new (alloc()) LNewTypedArray(temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewTypedArrayDynamicLength( + MNewTypedArrayDynamicLength* ins) { + MDefinition* length = ins->length(); + MOZ_ASSERT(length->type() == MIRType::Int32); + + LNewTypedArrayDynamicLength* lir = + new (alloc()) LNewTypedArrayDynamicLength(useRegister(length), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewTypedArrayFromArray(MNewTypedArrayFromArray* ins) { + MDefinition* array = ins->array(); + MOZ_ASSERT(array->type() == MIRType::Object); + + auto* lir = new (alloc()) LNewTypedArrayFromArray(useRegisterAtStart(array)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewTypedArrayFromArrayBuffer( + MNewTypedArrayFromArrayBuffer* ins) { + MDefinition* arrayBuffer = ins->arrayBuffer(); + MDefinition* byteOffset = ins->byteOffset(); + MDefinition* length = ins->length(); + MOZ_ASSERT(arrayBuffer->type() == MIRType::Object); + MOZ_ASSERT(byteOffset->type() == MIRType::Value); + MOZ_ASSERT(length->type() == MIRType::Value); + + auto* lir = new (alloc()) LNewTypedArrayFromArrayBuffer( + useRegisterAtStart(arrayBuffer), useBoxAtStart(byteOffset), + useBoxAtStart(length)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewObject(MNewObject* ins) { + LNewObject* lir = new (alloc()) LNewObject(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewNamedLambdaObject(MNewNamedLambdaObject* ins) { + LNewNamedLambdaObject* lir = new (alloc()) LNewNamedLambdaObject(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewCallObject(MNewCallObject* ins) { + LNewCallObject* lir = new (alloc()) LNewCallObject(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewStringObject(MNewStringObject* ins) { + MOZ_ASSERT(ins->input()->type() == MIRType::String); + + LNewStringObject* lir = + new (alloc()) LNewStringObject(useRegister(ins->input()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitInitElemGetterSetter(MInitElemGetterSetter* ins) { + LInitElemGetterSetter* lir = new (alloc()) LInitElemGetterSetter( + useRegisterAtStart(ins->object()), useBoxAtStart(ins->idValue()), + useRegisterAtStart(ins->value())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitMutateProto(MMutateProto* ins) { + LMutateProto* lir = new (alloc()) LMutateProto( + useRegisterAtStart(ins->getObject()), useBoxAtStart(ins->getValue())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitInitPropGetterSetter(MInitPropGetterSetter* ins) { + LInitPropGetterSetter* lir = new (alloc()) LInitPropGetterSetter( + useRegisterAtStart(ins->object()), useRegisterAtStart(ins->value())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCreateThisWithTemplate(MCreateThisWithTemplate* ins) { + LCreateThisWithTemplate* lir = new (alloc()) LCreateThisWithTemplate(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCreateThis(MCreateThis* ins) { + LCreateThis* lir = new (alloc()) + LCreateThis(useRegisterOrConstantAtStart(ins->getCallee()), + useRegisterOrConstantAtStart(ins->getNewTarget())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCreateArgumentsObject(MCreateArgumentsObject* ins) { + LAllocation callObj = useRegisterAtStart(ins->getCallObject()); + LCreateArgumentsObject* lir = new (alloc()) + LCreateArgumentsObject(callObj, tempFixed(CallTempReg0), + tempFixed(CallTempReg1), tempFixed(CallTempReg2)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetArgumentsObjectArg(MGetArgumentsObjectArg* ins) { + LAllocation argsObj = useRegister(ins->getArgsObject()); + LGetArgumentsObjectArg* lir = + new (alloc()) LGetArgumentsObjectArg(argsObj, temp()); + defineBox(lir, ins); +} + +void LIRGenerator::visitSetArgumentsObjectArg(MSetArgumentsObjectArg* ins) { + LAllocation argsObj = useRegister(ins->getArgsObject()); + LSetArgumentsObjectArg* lir = new (alloc()) + LSetArgumentsObjectArg(argsObj, useBox(ins->getValue()), temp()); + add(lir, ins); +} + +void LIRGenerator::visitLoadArgumentsObjectArg(MLoadArgumentsObjectArg* ins) { + MDefinition* argsObj = ins->getArgsObject(); + MOZ_ASSERT(argsObj->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) + LLoadArgumentsObjectArg(useRegister(argsObj), useRegister(index), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineBox(lir, ins); +} + +void LIRGenerator::visitArgumentsObjectLength(MArgumentsObjectLength* ins) { + MDefinition* argsObj = ins->getArgsObject(); + MOZ_ASSERT(argsObj->type() == MIRType::Object); + + auto* lir = new (alloc()) LArgumentsObjectLength(useRegister(argsObj)); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitGuardArgumentsObjectNotOverriddenIterator( + MGuardArgumentsObjectNotOverriddenIterator* ins) { + MDefinition* argsObj = ins->getArgsObject(); + MOZ_ASSERT(argsObj->type() == MIRType::Object); + + auto* lir = new (alloc()) + LGuardArgumentsObjectNotOverriddenIterator(useRegister(argsObj), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, argsObj); +} + +void LIRGenerator::visitReturnFromCtor(MReturnFromCtor* ins) { + LReturnFromCtor* lir = new (alloc()) + LReturnFromCtor(useBox(ins->getValue()), useRegister(ins->getObject())); + define(lir, ins); +} + +void LIRGenerator::visitBoxNonStrictThis(MBoxNonStrictThis* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + MOZ_ASSERT(ins->input()->type() == MIRType::Value); + + auto* lir = new (alloc()) LBoxNonStrictThis(useBox(ins->input())); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitImplicitThis(MImplicitThis* ins) { + MDefinition* env = ins->envChain(); + MOZ_ASSERT(env->type() == MIRType::Object); + + LImplicitThis* lir = new (alloc()) LImplicitThis(useRegisterAtStart(env)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitArrowNewTarget(MArrowNewTarget* ins) { + MOZ_ASSERT(ins->type() == MIRType::Value); + MOZ_ASSERT(ins->callee()->type() == MIRType::Object); + + LArrowNewTarget* lir = + new (alloc()) LArrowNewTarget(useRegister(ins->callee())); + defineBox(lir, ins); +} + +bool LIRGenerator::lowerCallArguments(MCall* call) { + uint32_t argc = call->numStackArgs(); + + // Align the arguments of a call such that the callee would keep the same + // alignment as the caller. + uint32_t baseSlot = 0; + if (JitStackValueAlignment > 1) { + baseSlot = AlignBytes(argc, JitStackValueAlignment); + } else { + baseSlot = argc; + } + + // Save the maximum number of argument, such that we can have one unique + // frame size. + if (baseSlot > maxargslots_) { + maxargslots_ = baseSlot; + } + + for (size_t i = 0; i < argc; i++) { + MDefinition* arg = call->getArg(i); + uint32_t argslot = baseSlot - i; + + // Values take a slow path. + if (arg->type() == MIRType::Value) { + LStackArgV* stack = new (alloc()) LStackArgV(argslot, useBox(arg)); + add(stack); + } else { + // Known types can move constant types and/or payloads. + LStackArgT* stack = new (alloc()) + LStackArgT(argslot, arg->type(), useRegisterOrConstant(arg)); + add(stack); + } + + if (!alloc().ensureBallast()) { + return false; + } + } + return true; +} + +void LIRGenerator::visitCall(MCall* call) { + MOZ_ASSERT(call->getCallee()->type() == MIRType::Object); + + // In case of oom, skip the rest of the allocations. + if (!lowerCallArguments(call)) { + abort(AbortReason::Alloc, "OOM: LIRGenerator::visitCall"); + return; + } + + WrappedFunction* target = call->getSingleTarget(); + + LInstruction* lir; + + if (call->isCallDOMNative()) { + // Call DOM functions. + MOZ_ASSERT(target && target->isNativeWithoutJitEntry()); + Register cxReg, objReg, privReg, argsReg; + GetTempRegForIntArg(0, 0, &cxReg); + GetTempRegForIntArg(1, 0, &objReg); + GetTempRegForIntArg(2, 0, &privReg); + mozilla::DebugOnly<bool> ok = GetTempRegForIntArg(3, 0, &argsReg); + MOZ_ASSERT(ok, "How can we not have four temp registers?"); + lir = new (alloc()) LCallDOMNative(tempFixed(cxReg), tempFixed(objReg), + tempFixed(privReg), tempFixed(argsReg)); + } else if (target) { + // Call known functions. + if (target->isNativeWithoutJitEntry()) { + Register cxReg, numReg, vpReg, tmpReg; + GetTempRegForIntArg(0, 0, &cxReg); + GetTempRegForIntArg(1, 0, &numReg); + GetTempRegForIntArg(2, 0, &vpReg); + + // Even though this is just a temp reg, use the same API to avoid + // register collisions. + mozilla::DebugOnly<bool> ok = GetTempRegForIntArg(3, 0, &tmpReg); + MOZ_ASSERT(ok, "How can we not have four temp registers?"); + + lir = new (alloc()) LCallNative(tempFixed(cxReg), tempFixed(numReg), + tempFixed(vpReg), tempFixed(tmpReg)); + } else { + lir = new (alloc()) LCallKnown(useRegisterAtStart(call->getCallee()), + tempFixed(CallTempReg0)); + } + } else { + // Call anything, using the most generic code. + lir = new (alloc()) + LCallGeneric(useRegisterAtStart(call->getCallee()), + tempFixed(CallTempReg0), tempFixed(CallTempReg1)); + } + defineReturn(lir, call); + assignSafepoint(lir, call); +} + +void LIRGenerator::visitApplyArgs(MApplyArgs* apply) { + MOZ_ASSERT(apply->getFunction()->type() == MIRType::Object); + + // Assert if the return value is already erased. + static_assert(CallTempReg2 != JSReturnReg_Type); + static_assert(CallTempReg2 != JSReturnReg_Data); + + LApplyArgsGeneric* lir = new (alloc()) LApplyArgsGeneric( + useFixedAtStart(apply->getFunction(), CallTempReg3), + useFixedAtStart(apply->getArgc(), CallTempReg0), + useBoxFixedAtStart(apply->getThis(), CallTempReg4, CallTempReg5), + tempFixed(CallTempReg1), // object register + tempFixed(CallTempReg2)); // stack counter register + + // Bailout is needed in the case of too many values in the arguments array. + assignSnapshot(lir, apply->bailoutKind()); + + defineReturn(lir, apply); + assignSafepoint(lir, apply); +} + +void LIRGenerator::visitApplyArray(MApplyArray* apply) { + MOZ_ASSERT(apply->getFunction()->type() == MIRType::Object); + + // Assert if the return value is already erased. + static_assert(CallTempReg2 != JSReturnReg_Type); + static_assert(CallTempReg2 != JSReturnReg_Data); + + LApplyArrayGeneric* lir = new (alloc()) LApplyArrayGeneric( + useFixedAtStart(apply->getFunction(), CallTempReg3), + useFixedAtStart(apply->getElements(), CallTempReg0), + useBoxFixedAtStart(apply->getThis(), CallTempReg4, CallTempReg5), + tempFixed(CallTempReg1), // object register + tempFixed(CallTempReg2)); // stack counter register + + // Bailout is needed in the case of too many values in the array, or empty + // space at the end of the array. + assignSnapshot(lir, apply->bailoutKind()); + + defineReturn(lir, apply); + assignSafepoint(lir, apply); +} + +void LIRGenerator::visitConstructArray(MConstructArray* mir) { + MOZ_ASSERT(mir->getFunction()->type() == MIRType::Object); + MOZ_ASSERT(mir->getElements()->type() == MIRType::Elements); + MOZ_ASSERT(mir->getNewTarget()->type() == MIRType::Object); + MOZ_ASSERT(mir->getThis()->type() == MIRType::Value); + + // Assert if the return value is already erased. + static_assert(CallTempReg2 != JSReturnReg_Type); + static_assert(CallTempReg2 != JSReturnReg_Data); + + auto* lir = new (alloc()) LConstructArrayGeneric( + useFixedAtStart(mir->getFunction(), CallTempReg3), + useFixedAtStart(mir->getElements(), CallTempReg0), + useFixedAtStart(mir->getNewTarget(), CallTempReg1), + useBoxFixedAtStart(mir->getThis(), CallTempReg4, CallTempReg5), + tempFixed(CallTempReg2)); + + // Bailout is needed in the case of too many values in the array, or empty + // space at the end of the array. + assignSnapshot(lir, mir->bailoutKind()); + + defineReturn(lir, mir); + assignSafepoint(lir, mir); +} + +void LIRGenerator::visitBail(MBail* bail) { + LBail* lir = new (alloc()) LBail(); + assignSnapshot(lir, bail->bailoutKind()); + add(lir, bail); +} + +void LIRGenerator::visitUnreachable(MUnreachable* unreachable) { + LUnreachable* lir = new (alloc()) LUnreachable(); + add(lir, unreachable); +} + +void LIRGenerator::visitEncodeSnapshot(MEncodeSnapshot* mir) { + LEncodeSnapshot* lir = new (alloc()) LEncodeSnapshot(); + assignSnapshot(lir, mir->bailoutKind()); + add(lir, mir); +} + +void LIRGenerator::visitUnreachableResult(MUnreachableResult* mir) { + if (mir->type() == MIRType::Value) { + auto* lir = new (alloc()) LUnreachableResultV(); + defineBox(lir, mir); + } else { + auto* lir = new (alloc()) LUnreachableResultT(); + define(lir, mir); + } +} + +void LIRGenerator::visitAssertFloat32(MAssertFloat32* assertion) { + MIRType type = assertion->input()->type(); + DebugOnly<bool> checkIsFloat32 = assertion->mustBeFloat32(); + + if (type != MIRType::Value && !JitOptions.eagerIonCompilation()) { + MOZ_ASSERT_IF(checkIsFloat32, type == MIRType::Float32); + MOZ_ASSERT_IF(!checkIsFloat32, type != MIRType::Float32); + } +} + +void LIRGenerator::visitAssertRecoveredOnBailout( + MAssertRecoveredOnBailout* assertion) { + MOZ_CRASH("AssertRecoveredOnBailout nodes are always recovered on bailouts."); +} + +static JSOp ReorderComparison(JSOp op, MDefinition** lhsp, MDefinition** rhsp) { + MDefinition* lhs = *lhsp; + MDefinition* rhs = *rhsp; + + if (lhs->maybeConstantValue()) { + *rhsp = lhs; + *lhsp = rhs; + return ReverseCompareOp(op); + } + return op; +} + +void LIRGenerator::visitTest(MTest* test) { + MDefinition* opd = test->getOperand(0); + MBasicBlock* ifTrue = test->ifTrue(); + MBasicBlock* ifFalse = test->ifFalse(); + + // String is converted to length of string in the type analysis phase (see + // TestPolicy). + MOZ_ASSERT(opd->type() != MIRType::String); + + // BigInt is boxed in type analysis. + MOZ_ASSERT(opd->type() != MIRType::BigInt, + "BigInt should be boxed by TestPolicy"); + + // Testing a constant. + if (MConstant* constant = opd->maybeConstantValue()) { + bool b; + if (constant->valueToBoolean(&b)) { + add(new (alloc()) LGoto(b ? ifTrue : ifFalse)); + return; + } + } + + if (opd->type() == MIRType::Value) { + auto* lir = new (alloc()) LTestVAndBranch(ifTrue, ifFalse, useBox(opd), + tempDouble(), temp(), temp()); + add(lir, test); + return; + } + + // Objects are truthy, except if it might emulate undefined. + if (opd->type() == MIRType::Object) { + add(new (alloc()) + LTestOAndBranch(useRegister(opd), ifTrue, ifFalse, temp()), + test); + return; + } + + // These must be explicitly sniffed out since they are constants and have + // no payload. + if (opd->type() == MIRType::Undefined || opd->type() == MIRType::Null) { + add(new (alloc()) LGoto(ifFalse)); + return; + } + + // All symbols are truthy. + if (opd->type() == MIRType::Symbol) { + add(new (alloc()) LGoto(ifTrue)); + return; + } + + if (opd->isCompare() && opd->isEmittedAtUses()) { + // Emit LBitAndBranch for cases like |if ((x & y) === 0)|. + MCompare* comp = opd->toCompare(); + if ((comp->isInt32Comparison() || + comp->compareType() == MCompare::Compare_UInt32) && + comp->getOperand(1)->isConstant() && + comp->getOperand(1)->toConstant()->isInt32(0) && + comp->getOperand(0)->isBitAnd() && + comp->getOperand(0)->isEmittedAtUses()) { + MDefinition* bitAnd = opd->getOperand(0); + MDefinition* lhs = bitAnd->getOperand(0); + MDefinition* rhs = bitAnd->getOperand(1); + + Assembler::Condition cond = + JSOpToCondition(comp->compareType(), comp->jsop()); + + if (lhs->type() == MIRType::Int32 && rhs->type() == MIRType::Int32 && + (cond == Assembler::Equal || cond == Assembler::NotEqual)) { + ReorderCommutative(&lhs, &rhs, test); + if (cond == Assembler::Equal) { + cond = Assembler::Zero; + } else if (cond == Assembler::NotEqual) { + cond = Assembler::NonZero; + } else { + MOZ_ASSERT_UNREACHABLE("inequality operators cannot be folded"); + } + lowerForBitAndAndBranch(new (alloc()) + LBitAndAndBranch(ifTrue, ifFalse, cond), + test, lhs, rhs); + return; + } + } + } + + // Check if the operand for this test is a compare operation. If it is, we + // want to emit an LCompare*AndBranch rather than an LTest*AndBranch, to fuse + // the compare and jump instructions. + if (opd->isCompare() && opd->isEmittedAtUses()) { + MCompare* comp = opd->toCompare(); + MDefinition* left = comp->lhs(); + MDefinition* right = comp->rhs(); + + // Try to fold the comparison so that we don't have to handle all cases. + bool result; + if (comp->tryFold(&result)) { + add(new (alloc()) LGoto(result ? ifTrue : ifFalse)); + return; + } + + // Emit LCompare*AndBranch. + + // Compare and branch null/undefined. + // The second operand has known null/undefined type, + // so just test the first operand. + if (comp->compareType() == MCompare::Compare_Null || + comp->compareType() == MCompare::Compare_Undefined) { + if (left->type() == MIRType::Object) { + auto* lir = new (alloc()) LIsNullOrLikeUndefinedAndBranchT( + comp, useRegister(left), ifTrue, ifFalse, temp()); + add(lir, test); + return; + } + + auto* lir = new (alloc()) LIsNullOrLikeUndefinedAndBranchV( + comp, ifTrue, ifFalse, useBox(left), temp(), tempToUnbox()); + add(lir, test); + return; + } + + // Compare and branch Int32, Symbol or Object pointers. + if (comp->isInt32Comparison() || + comp->compareType() == MCompare::Compare_UInt32 || + comp->compareType() == MCompare::Compare_Object || + comp->compareType() == MCompare::Compare_Symbol) { + JSOp op = ReorderComparison(comp->jsop(), &left, &right); + LAllocation lhs = useRegister(left); + LAllocation rhs; + if (comp->isInt32Comparison() || + comp->compareType() == MCompare::Compare_UInt32) { + rhs = useAnyOrConstant(right); + } else { + rhs = useRegister(right); + } + LCompareAndBranch* lir = + new (alloc()) LCompareAndBranch(comp, op, lhs, rhs, ifTrue, ifFalse); + add(lir, test); + return; + } + + // Compare and branch Int64. + if (comp->compareType() == MCompare::Compare_Int64 || + comp->compareType() == MCompare::Compare_UInt64) { + JSOp op = ReorderComparison(comp->jsop(), &left, &right); + LCompareI64AndBranch* lir = new (alloc()) + LCompareI64AndBranch(comp, op, useInt64Register(left), + useInt64OrConstant(right), ifTrue, ifFalse); + add(lir, test); + return; + } + + // Compare and branch doubles. + if (comp->isDoubleComparison()) { + LAllocation lhs = useRegister(left); + LAllocation rhs = useRegister(right); + LCompareDAndBranch* lir = + new (alloc()) LCompareDAndBranch(comp, lhs, rhs, ifTrue, ifFalse); + add(lir, test); + return; + } + + // Compare and branch floats. + if (comp->isFloat32Comparison()) { + LAllocation lhs = useRegister(left); + LAllocation rhs = useRegister(right); + LCompareFAndBranch* lir = + new (alloc()) LCompareFAndBranch(comp, lhs, rhs, ifTrue, ifFalse); + add(lir, test); + return; + } + } + + // Check if the operand for this test is a bitand operation. If it is, we want + // to emit an LBitAndAndBranch rather than an LTest*AndBranch. + if (opd->isBitAnd() && opd->isEmittedAtUses()) { + MDefinition* lhs = opd->getOperand(0); + MDefinition* rhs = opd->getOperand(1); + if (lhs->type() == MIRType::Int32 && rhs->type() == MIRType::Int32) { + ReorderCommutative(&lhs, &rhs, test); + lowerForBitAndAndBranch(new (alloc()) LBitAndAndBranch(ifTrue, ifFalse), + test, lhs, rhs); + return; + } + } + +#if defined(ENABLE_WASM_SIMD) && \ + (defined(JS_CODEGEN_X86) || defined(JS_CODEGEN_X64)) + // Check if the operand for this test is an any_true/all_true SIMD operation. + // If it is, we want to emit an LWasmReduceAndBranchSimd128 node to avoid + // generating an intermediate boolean result. + if (opd->isWasmReduceSimd128() && opd->isEmittedAtUses()) { + MWasmReduceSimd128* node = opd->toWasmReduceSimd128(); + if (canFoldReduceSimd128AndBranch(node->simdOp())) { +# ifdef DEBUG + js::wasm::ReportSimdAnalysis("simd128-to-scalar-and-branch -> folded"); +# endif + auto* lir = new (alloc()) LWasmReduceAndBranchSimd128( + useRegister(node->input()), node->simdOp(), ifTrue, ifFalse); + add(lir, test); + return; + } + } +#endif + + if (opd->isIsObject() && opd->isEmittedAtUses()) { + MDefinition* input = opd->toIsObject()->input(); + MOZ_ASSERT(input->type() == MIRType::Value); + + LIsObjectAndBranch* lir = + new (alloc()) LIsObjectAndBranch(ifTrue, ifFalse, useBoxAtStart(input)); + add(lir, test); + return; + } + + if (opd->isIsNullOrUndefined() && opd->isEmittedAtUses()) { + MIsNullOrUndefined* isNullOrUndefined = opd->toIsNullOrUndefined(); + MDefinition* input = isNullOrUndefined->value(); + MOZ_ASSERT(input->type() == MIRType::Value); + + auto* lir = new (alloc()) LIsNullOrUndefinedAndBranch( + isNullOrUndefined, ifTrue, ifFalse, useBoxAtStart(input)); + add(lir, test); + return; + } + + if (opd->isIsNoIter()) { + MOZ_ASSERT(opd->isEmittedAtUses()); + + MDefinition* input = opd->toIsNoIter()->input(); + MOZ_ASSERT(input->type() == MIRType::Value); + + LIsNoIterAndBranch* lir = + new (alloc()) LIsNoIterAndBranch(ifTrue, ifFalse, useBox(input)); + add(lir, test); + return; + } + + switch (opd->type()) { + case MIRType::Double: + add(new (alloc()) LTestDAndBranch(useRegister(opd), ifTrue, ifFalse)); + break; + case MIRType::Float32: + add(new (alloc()) LTestFAndBranch(useRegister(opd), ifTrue, ifFalse)); + break; + case MIRType::Int32: + case MIRType::Boolean: + add(new (alloc()) LTestIAndBranch(useRegister(opd), ifTrue, ifFalse)); + break; + case MIRType::Int64: + add(new (alloc()) + LTestI64AndBranch(useInt64Register(opd), ifTrue, ifFalse)); + break; + default: + MOZ_CRASH("Bad type"); + } +} + +static inline bool CanEmitCompareAtUses(MInstruction* ins) { + if (!ins->canEmitAtUses()) { + return false; + } + + // If the result is never used, we can usefully defer emission to the use + // point, since that will never happen. + MUseIterator iter(ins->usesBegin()); + if (iter == ins->usesEnd()) { + return true; + } + + // If the first use isn't of the expected form, the answer is No. + MNode* node = iter->consumer(); + if (!node->isDefinition()) { + return false; + } + + MDefinition* use = node->toDefinition(); + if (!use->isTest() && !use->isWasmSelect()) { + return false; + } + + // Emission can be deferred to the first use point, but only if there are no + // other use points. + iter++; + return iter == ins->usesEnd(); +} + +void LIRGenerator::visitCompare(MCompare* comp) { + MDefinition* left = comp->lhs(); + MDefinition* right = comp->rhs(); + + // Try to fold the comparison so that we don't have to handle all cases. + bool result; + if (comp->tryFold(&result)) { + define(new (alloc()) LInteger(result), comp); + return; + } + + // Move below the emitAtUses call if we ever implement + // LCompareSAndBranch. Doing this now wouldn't be wrong, but doesn't + // make sense and avoids confusion. + if (comp->compareType() == MCompare::Compare_String) { + LCompareS* lir = + new (alloc()) LCompareS(useRegister(left), useRegister(right)); + define(lir, comp); + assignSafepoint(lir, comp); + return; + } + + // Compare two BigInts. + if (comp->compareType() == MCompare::Compare_BigInt) { + auto* lir = new (alloc()) LCompareBigInt( + useRegister(left), useRegister(right), temp(), temp(), temp()); + define(lir, comp); + return; + } + + // Compare BigInt with Int32. + if (comp->compareType() == MCompare::Compare_BigInt_Int32) { + auto* lir = new (alloc()) LCompareBigIntInt32( + useRegister(left), useRegister(right), temp(), temp()); + define(lir, comp); + return; + } + + // Compare BigInt with Double. + if (comp->compareType() == MCompare::Compare_BigInt_Double) { + auto* lir = new (alloc()) LCompareBigIntDouble(useRegisterAtStart(left), + useRegisterAtStart(right), + tempFixed(CallTempReg0)); + defineReturn(lir, comp); + return; + } + + // Compare BigInt with String. + if (comp->compareType() == MCompare::Compare_BigInt_String) { + auto* lir = new (alloc()) LCompareBigIntString(useRegisterAtStart(left), + useRegisterAtStart(right)); + defineReturn(lir, comp); + assignSafepoint(lir, comp); + return; + } + + // Sniff out if the output of this compare is used only for a branching. + // If it is, then we will emit an LCompare*AndBranch instruction in place + // of this compare and any test that uses this compare. Thus, we can + // ignore this Compare. + if (CanEmitCompareAtUses(comp)) { + emitAtUses(comp); + return; + } + + // Compare Null and Undefined. + if (comp->compareType() == MCompare::Compare_Null || + comp->compareType() == MCompare::Compare_Undefined) { + if (left->type() == MIRType::Object) { + define(new (alloc()) LIsNullOrLikeUndefinedT(useRegister(left)), comp); + return; + } + + auto* lir = new (alloc()) + LIsNullOrLikeUndefinedV(useBox(left), temp(), tempToUnbox()); + define(lir, comp); + return; + } + + // Compare Int32, Symbol, Object or Wasm pointers. + if (comp->isInt32Comparison() || + comp->compareType() == MCompare::Compare_UInt32 || + comp->compareType() == MCompare::Compare_Object || + comp->compareType() == MCompare::Compare_Symbol || + comp->compareType() == MCompare::Compare_RefOrNull) { + JSOp op = ReorderComparison(comp->jsop(), &left, &right); + LAllocation lhs = useRegister(left); + LAllocation rhs; + if (comp->isInt32Comparison() || + comp->compareType() == MCompare::Compare_UInt32) { + rhs = useAnyOrConstant(right); + } else { + rhs = useRegister(right); + } + define(new (alloc()) LCompare(op, lhs, rhs), comp); + return; + } + + // Compare Int64. + if (comp->compareType() == MCompare::Compare_Int64 || + comp->compareType() == MCompare::Compare_UInt64) { + JSOp op = ReorderComparison(comp->jsop(), &left, &right); + define(new (alloc()) LCompareI64(op, useInt64Register(left), + useInt64OrConstant(right)), + comp); + return; + } + + // Compare doubles. + if (comp->isDoubleComparison()) { + define(new (alloc()) LCompareD(useRegister(left), useRegister(right)), + comp); + return; + } + + // Compare float32. + if (comp->isFloat32Comparison()) { + define(new (alloc()) LCompareF(useRegister(left), useRegister(right)), + comp); + return; + } + + MOZ_CRASH("Unrecognized compare type."); +} + +void LIRGenerator::visitSameValue(MSameValue* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + + MOZ_ASSERT(lhs->type() == MIRType::Double); + MOZ_ASSERT(rhs->type() == MIRType::Double); + + auto* lir = new (alloc()) + LSameValueD(useRegister(lhs), useRegister(rhs), tempDouble()); + define(lir, ins); +} + +void LIRGenerator::lowerBitOp(JSOp op, MBinaryBitwiseInstruction* ins) { + MDefinition* lhs = ins->getOperand(0); + MDefinition* rhs = ins->getOperand(1); + MOZ_ASSERT(IsIntType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + MOZ_ASSERT(rhs->type() == MIRType::Int32); + ReorderCommutative(&lhs, &rhs, ins); + lowerForALU(new (alloc()) LBitOpI(op), ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + MOZ_ASSERT(rhs->type() == MIRType::Int64); + ReorderCommutative(&lhs, &rhs, ins); + lowerForALUInt64(new (alloc()) LBitOpI64(op), ins, lhs, rhs); + return; + } + + MOZ_CRASH("Unhandled integer specialization"); +} + +void LIRGenerator::visitTypeOf(MTypeOf* ins) { + MDefinition* opd = ins->input(); + + if (opd->type() == MIRType::Object) { + auto* lir = new (alloc()) LTypeOfO(useRegister(opd)); + define(lir, ins); + return; + } + + MOZ_ASSERT(opd->type() == MIRType::Value); + + LTypeOfV* lir = new (alloc()) LTypeOfV(useBox(opd), tempToUnbox()); + define(lir, ins); +} + +void LIRGenerator::visitToAsyncIter(MToAsyncIter* ins) { + LToAsyncIter* lir = + new (alloc()) LToAsyncIter(useRegisterAtStart(ins->getIterator()), + useBoxAtStart(ins->getNextMethod())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitToPropertyKeyCache(MToPropertyKeyCache* ins) { + MDefinition* input = ins->getOperand(0); + MOZ_ASSERT(ins->type() == MIRType::Value); + + auto* lir = new (alloc()) LToPropertyKeyCache(useBox(input)); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBitNot(MBitNot* ins) { + MDefinition* input = ins->getOperand(0); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(input->type() == MIRType::Int32); + lowerForALU(new (alloc()) LBitNotI(), ins, input); + return; + } + + MOZ_CRASH("Unhandled integer specialization"); +} + +static bool CanEmitBitAndAtUses(MInstruction* ins) { + if (!ins->canEmitAtUses()) { + return false; + } + + if (ins->getOperand(0)->type() != MIRType::Int32 || + ins->getOperand(1)->type() != MIRType::Int32) { + return false; + } + + MUseIterator iter(ins->usesBegin()); + if (iter == ins->usesEnd()) { + return false; + } + + MNode* node = iter->consumer(); + if (!node->isDefinition() || !node->toDefinition()->isInstruction()) { + return false; + } + + MInstruction* use = node->toDefinition()->toInstruction(); + if (!use->isTest() && !(use->isCompare() && CanEmitCompareAtUses(use))) { + return false; + } + + iter++; + return iter == ins->usesEnd(); +} + +void LIRGenerator::visitBitAnd(MBitAnd* ins) { + // Sniff out if the output of this bitand is used only for a branching. + // If it is, then we will emit an LBitAndAndBranch instruction in place + // of this bitand and any test that uses this bitand. Thus, we can + // ignore this BitAnd. + if (CanEmitBitAndAtUses(ins)) { + emitAtUses(ins); + } else { + lowerBitOp(JSOp::BitAnd, ins); + } +} + +void LIRGenerator::visitBitOr(MBitOr* ins) { lowerBitOp(JSOp::BitOr, ins); } + +void LIRGenerator::visitBitXor(MBitXor* ins) { lowerBitOp(JSOp::BitXor, ins); } + +void LIRGenerator::lowerShiftOp(JSOp op, MShiftInstruction* ins) { + MDefinition* lhs = ins->getOperand(0); + MDefinition* rhs = ins->getOperand(1); + + if (op == JSOp::Ursh && ins->type() == MIRType::Double) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + MOZ_ASSERT(rhs->type() == MIRType::Int32); + lowerUrshD(ins->toUrsh()); + return; + } + + MOZ_ASSERT(IsIntType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + MOZ_ASSERT(rhs->type() == MIRType::Int32); + + LShiftI* lir = new (alloc()) LShiftI(op); + if (op == JSOp::Ursh) { + if (ins->toUrsh()->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + } + lowerForShift(lir, ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + MOZ_ASSERT(rhs->type() == MIRType::Int64); + lowerForShiftInt64(new (alloc()) LShiftI64(op), ins, lhs, rhs); + return; + } + + MOZ_CRASH("Unhandled integer specialization"); +} + +void LIRGenerator::visitLsh(MLsh* ins) { lowerShiftOp(JSOp::Lsh, ins); } + +void LIRGenerator::visitRsh(MRsh* ins) { lowerShiftOp(JSOp::Rsh, ins); } + +void LIRGenerator::visitUrsh(MUrsh* ins) { lowerShiftOp(JSOp::Ursh, ins); } + +void LIRGenerator::visitSignExtendInt32(MSignExtendInt32* ins) { + LInstructionHelper<1, 1, 0>* lir; + + if (ins->mode() == MSignExtendInt32::Byte) { + lir = new (alloc()) + LSignExtendInt32(useByteOpRegisterAtStart(ins->input()), ins->mode()); + } else { + lir = new (alloc()) + LSignExtendInt32(useRegisterAtStart(ins->input()), ins->mode()); + } + + define(lir, ins); +} + +void LIRGenerator::visitRotate(MRotate* ins) { + MDefinition* input = ins->input(); + MDefinition* count = ins->count(); + + if (ins->type() == MIRType::Int32) { + auto* lir = new (alloc()) LRotate(); + lowerForShift(lir, ins, input, count); + } else if (ins->type() == MIRType::Int64) { + auto* lir = new (alloc()) LRotateI64(); + lowerForShiftInt64(lir, ins, input, count); + } else { + MOZ_CRASH("unexpected type in visitRotate"); + } +} + +void LIRGenerator::visitFloor(MFloor* ins) { + MIRType type = ins->input()->type(); + MOZ_ASSERT(IsFloatingPointType(type)); + + LInstructionHelper<1, 1, 0>* lir; + if (type == MIRType::Double) { + lir = new (alloc()) LFloor(useRegister(ins->input())); + } else { + lir = new (alloc()) LFloorF(useRegister(ins->input())); + } + + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitCeil(MCeil* ins) { + MIRType type = ins->input()->type(); + MOZ_ASSERT(IsFloatingPointType(type)); + + LInstructionHelper<1, 1, 0>* lir; + if (type == MIRType::Double) { + lir = new (alloc()) LCeil(useRegister(ins->input())); + } else { + lir = new (alloc()) LCeilF(useRegister(ins->input())); + } + + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitRound(MRound* ins) { + MIRType type = ins->input()->type(); + MOZ_ASSERT(IsFloatingPointType(type)); + + LInstructionHelper<1, 1, 1>* lir; + if (type == MIRType::Double) { + lir = new (alloc()) LRound(useRegister(ins->input()), tempDouble()); + } else { + lir = new (alloc()) LRoundF(useRegister(ins->input()), tempFloat32()); + } + + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitTrunc(MTrunc* ins) { + MIRType type = ins->input()->type(); + MOZ_ASSERT(IsFloatingPointType(type)); + + LInstructionHelper<1, 1, 0>* lir; + if (type == MIRType::Double) { + lir = new (alloc()) LTrunc(useRegister(ins->input())); + } else { + lir = new (alloc()) LTruncF(useRegister(ins->input())); + } + + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitNearbyInt(MNearbyInt* ins) { + MIRType inputType = ins->input()->type(); + MOZ_ASSERT(IsFloatingPointType(inputType)); + MOZ_ASSERT(ins->type() == inputType); + + LInstructionHelper<1, 1, 0>* lir; + if (inputType == MIRType::Double) { + lir = new (alloc()) LNearbyInt(useRegisterAtStart(ins->input())); + } else { + lir = new (alloc()) LNearbyIntF(useRegisterAtStart(ins->input())); + } + + define(lir, ins); +} + +void LIRGenerator::visitMinMax(MMinMax* ins) { + MDefinition* first = ins->getOperand(0); + MDefinition* second = ins->getOperand(1); + + ReorderCommutative(&first, &second, ins); + + LMinMaxBase* lir; + switch (ins->type()) { + case MIRType::Int32: + lir = new (alloc()) + LMinMaxI(useRegisterAtStart(first), useRegisterOrConstant(second)); + break; + case MIRType::Float32: + lir = new (alloc()) + LMinMaxF(useRegisterAtStart(first), useRegister(second)); + break; + case MIRType::Double: + lir = new (alloc()) + LMinMaxD(useRegisterAtStart(first), useRegister(second)); + break; + default: + MOZ_CRASH(); + } + + defineReuseInput(lir, ins, 0); +} + +void LIRGenerator::visitAbs(MAbs* ins) { + MDefinition* num = ins->input(); + MOZ_ASSERT(IsNumberType(num->type())); + + LInstructionHelper<1, 1, 0>* lir; + switch (num->type()) { + case MIRType::Int32: + lir = new (alloc()) LAbsI(useRegisterAtStart(num)); + // needed to handle abs(INT32_MIN) + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + break; + case MIRType::Float32: + lir = new (alloc()) LAbsF(useRegisterAtStart(num)); + break; + case MIRType::Double: + lir = new (alloc()) LAbsD(useRegisterAtStart(num)); + break; + default: + MOZ_CRASH(); + } + defineReuseInput(lir, ins, 0); +} + +void LIRGenerator::visitClz(MClz* ins) { + MDefinition* num = ins->num(); + + MOZ_ASSERT(IsIntType(ins->type())); + + if (ins->type() == MIRType::Int32) { + LClzI* lir = new (alloc()) LClzI(useRegisterAtStart(num)); + define(lir, ins); + return; + } + + auto* lir = new (alloc()) LClzI64(useInt64RegisterAtStart(num)); + defineInt64(lir, ins); +} + +void LIRGenerator::visitCtz(MCtz* ins) { + MDefinition* num = ins->num(); + + MOZ_ASSERT(IsIntType(ins->type())); + + if (ins->type() == MIRType::Int32) { + LCtzI* lir = new (alloc()) LCtzI(useRegisterAtStart(num)); + define(lir, ins); + return; + } + + auto* lir = new (alloc()) LCtzI64(useInt64RegisterAtStart(num)); + defineInt64(lir, ins); +} + +void LIRGenerator::visitPopcnt(MPopcnt* ins) { + MDefinition* num = ins->num(); + + MOZ_ASSERT(IsIntType(ins->type())); + + if (ins->type() == MIRType::Int32) { + LPopcntI* lir = new (alloc()) LPopcntI(useRegisterAtStart(num), temp()); + define(lir, ins); + return; + } + + auto* lir = new (alloc()) LPopcntI64(useInt64RegisterAtStart(num), temp()); + defineInt64(lir, ins); +} + +void LIRGenerator::visitSqrt(MSqrt* ins) { + MDefinition* num = ins->input(); + MOZ_ASSERT(IsFloatingPointType(num->type())); + + LInstructionHelper<1, 1, 0>* lir; + if (num->type() == MIRType::Double) { + lir = new (alloc()) LSqrtD(useRegisterAtStart(num)); + } else { + lir = new (alloc()) LSqrtF(useRegisterAtStart(num)); + } + define(lir, ins); +} + +void LIRGenerator::visitAtan2(MAtan2* ins) { + MDefinition* y = ins->y(); + MOZ_ASSERT(y->type() == MIRType::Double); + + MDefinition* x = ins->x(); + MOZ_ASSERT(x->type() == MIRType::Double); + + LAtan2D* lir = new (alloc()) LAtan2D( + useRegisterAtStart(y), useRegisterAtStart(x), tempFixed(CallTempReg0)); + defineReturn(lir, ins); +} + +void LIRGenerator::visitHypot(MHypot* ins) { + LHypot* lir = nullptr; + uint32_t length = ins->numOperands(); + for (uint32_t i = 0; i < length; ++i) { + MOZ_ASSERT(ins->getOperand(i)->type() == MIRType::Double); + } + + switch (length) { + case 2: + lir = new (alloc()) LHypot(useRegisterAtStart(ins->getOperand(0)), + useRegisterAtStart(ins->getOperand(1)), + tempFixed(CallTempReg0)); + break; + case 3: + lir = new (alloc()) LHypot(useRegisterAtStart(ins->getOperand(0)), + useRegisterAtStart(ins->getOperand(1)), + useRegisterAtStart(ins->getOperand(2)), + tempFixed(CallTempReg0)); + break; + case 4: + lir = new (alloc()) LHypot(useRegisterAtStart(ins->getOperand(0)), + useRegisterAtStart(ins->getOperand(1)), + useRegisterAtStart(ins->getOperand(2)), + useRegisterAtStart(ins->getOperand(3)), + tempFixed(CallTempReg0)); + break; + default: + MOZ_CRASH("Unexpected number of arguments to LHypot."); + } + + defineReturn(lir, ins); +} + +void LIRGenerator::visitPow(MPow* ins) { + MDefinition* input = ins->input(); + MDefinition* power = ins->power(); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(input->type() == MIRType::Int32); + MOZ_ASSERT(power->type() == MIRType::Int32); + + if (input->isConstant()) { + // Restrict this optimization to |base <= 256| to avoid generating too + // many consecutive shift instructions. + int32_t base = input->toConstant()->toInt32(); + if (2 <= base && base <= 256 && mozilla::IsPowerOfTwo(uint32_t(base))) { + lowerPowOfTwoI(ins); + return; + } + } + + auto* lir = new (alloc()) + LPowII(useRegister(input), useRegister(power), temp(), temp()); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + return; + } + + MOZ_ASSERT(ins->type() == MIRType::Double); + MOZ_ASSERT(input->type() == MIRType::Double); + MOZ_ASSERT(power->type() == MIRType::Int32 || + power->type() == MIRType::Double); + + LInstruction* lir; + if (power->type() == MIRType::Int32) { + lir = + new (alloc()) LPowI(useRegisterAtStart(input), + useRegisterAtStart(power), tempFixed(CallTempReg0)); + } else { + lir = + new (alloc()) LPowD(useRegisterAtStart(input), + useRegisterAtStart(power), tempFixed(CallTempReg0)); + } + defineReturn(lir, ins); +} + +void LIRGenerator::visitSign(MSign* ins) { + if (ins->type() == ins->input()->type()) { + LInstructionHelper<1, 1, 0>* lir; + if (ins->type() == MIRType::Int32) { + lir = new (alloc()) LSignI(useRegister(ins->input())); + } else { + MOZ_ASSERT(ins->type() == MIRType::Double); + lir = new (alloc()) LSignD(useRegister(ins->input())); + } + define(lir, ins); + } else { + MOZ_ASSERT(ins->type() == MIRType::Int32); + MOZ_ASSERT(ins->input()->type() == MIRType::Double); + + auto* lir = new (alloc()) LSignDI(useRegister(ins->input()), tempDouble()); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + } +} + +void LIRGenerator::visitMathFunction(MMathFunction* ins) { + MOZ_ASSERT(IsFloatingPointType(ins->type())); + MOZ_ASSERT(ins->type() == ins->input()->type()); + + LInstruction* lir; + if (ins->type() == MIRType::Double) { + lir = new (alloc()) LMathFunctionD(useRegisterAtStart(ins->input()), + tempFixed(CallTempReg0)); + } else { + lir = new (alloc()) LMathFunctionF(useRegisterAtStart(ins->input()), + tempFixed(CallTempReg0)); + } + defineReturn(lir, ins); +} + +void LIRGenerator::visitRandom(MRandom* ins) { + auto* lir = new (alloc()) LRandom(temp(), tempInt64(), tempInt64()); + define(lir, ins); +} + +// Try to mark an add or sub instruction as able to recover its input when +// bailing out. +template <typename S, typename T> +static void MaybeSetRecoversInput(S* mir, T* lir) { + MOZ_ASSERT(lir->mirRaw() == mir); + if (!mir->fallible() || !lir->snapshot()) { + return; + } + + if (lir->output()->policy() != LDefinition::MUST_REUSE_INPUT) { + return; + } + + // The original operands to an add or sub can't be recovered if they both + // use the same register. + if (lir->lhs()->isUse() && lir->rhs()->isUse() && + lir->lhs()->toUse()->virtualRegister() == + lir->rhs()->toUse()->virtualRegister()) { + return; + } + + // Add instructions that are on two different values can recover + // the input they clobbered via MUST_REUSE_INPUT. Thus, a copy + // of that input does not need to be kept alive in the snapshot + // for the instruction. + + lir->setRecoversInput(); + + const LUse* input = lir->getOperand(lir->output()->getReusedInput())->toUse(); + lir->snapshot()->rewriteRecoveredInput(*input); +} + +void LIRGenerator::visitAdd(MAdd* ins) { + MDefinition* lhs = ins->getOperand(0); + MDefinition* rhs = ins->getOperand(1); + + MOZ_ASSERT(lhs->type() == rhs->type()); + MOZ_ASSERT(IsNumberType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + ReorderCommutative(&lhs, &rhs, ins); + LAddI* lir = new (alloc()) LAddI; + + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + + lowerForALU(lir, ins, lhs, rhs); + MaybeSetRecoversInput(ins, lir); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + ReorderCommutative(&lhs, &rhs, ins); + LAddI64* lir = new (alloc()) LAddI64; + lowerForALUInt64(lir, ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Double) { + MOZ_ASSERT(lhs->type() == MIRType::Double); + ReorderCommutative(&lhs, &rhs, ins); + lowerForFPU(new (alloc()) LMathD(JSOp::Add), ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Float32) { + MOZ_ASSERT(lhs->type() == MIRType::Float32); + ReorderCommutative(&lhs, &rhs, ins); + lowerForFPU(new (alloc()) LMathF(JSOp::Add), ins, lhs, rhs); + return; + } + + MOZ_CRASH("Unhandled number specialization"); +} + +void LIRGenerator::visitSub(MSub* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + + MOZ_ASSERT(lhs->type() == rhs->type()); + MOZ_ASSERT(IsNumberType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + + LSubI* lir = new (alloc()) LSubI; + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + + lowerForALU(lir, ins, lhs, rhs); + MaybeSetRecoversInput(ins, lir); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + LSubI64* lir = new (alloc()) LSubI64; + lowerForALUInt64(lir, ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Double) { + MOZ_ASSERT(lhs->type() == MIRType::Double); + lowerForFPU(new (alloc()) LMathD(JSOp::Sub), ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Float32) { + MOZ_ASSERT(lhs->type() == MIRType::Float32); + lowerForFPU(new (alloc()) LMathF(JSOp::Sub), ins, lhs, rhs); + return; + } + + MOZ_CRASH("Unhandled number specialization"); +} + +void LIRGenerator::visitMul(MMul* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + MOZ_ASSERT(lhs->type() == rhs->type()); + MOZ_ASSERT(IsNumberType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + ReorderCommutative(&lhs, &rhs, ins); + + // If our RHS is a constant -1 and we don't have to worry about + // overflow, we can optimize to an LNegI. + if (!ins->fallible() && rhs->isConstant() && + rhs->toConstant()->toInt32() == -1) { + defineReuseInput(new (alloc()) LNegI(useRegisterAtStart(lhs)), ins, 0); + } else { + lowerMulI(ins, lhs, rhs); + } + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + ReorderCommutative(&lhs, &rhs, ins); + LMulI64* lir = new (alloc()) LMulI64; + lowerForMulInt64(lir, ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Double) { + MOZ_ASSERT(lhs->type() == MIRType::Double); + ReorderCommutative(&lhs, &rhs, ins); + + // If our RHS is a constant -1.0, we can optimize to an LNegD. + if (!ins->mustPreserveNaN() && rhs->isConstant() && + rhs->toConstant()->toDouble() == -1.0) { + defineReuseInput(new (alloc()) LNegD(useRegisterAtStart(lhs)), ins, 0); + } else { + lowerForFPU(new (alloc()) LMathD(JSOp::Mul), ins, lhs, rhs); + } + return; + } + + if (ins->type() == MIRType::Float32) { + MOZ_ASSERT(lhs->type() == MIRType::Float32); + ReorderCommutative(&lhs, &rhs, ins); + + // We apply the same optimizations as for doubles + if (!ins->mustPreserveNaN() && rhs->isConstant() && + rhs->toConstant()->toFloat32() == -1.0f) { + defineReuseInput(new (alloc()) LNegF(useRegisterAtStart(lhs)), ins, 0); + } else { + lowerForFPU(new (alloc()) LMathF(JSOp::Mul), ins, lhs, rhs); + } + return; + } + + MOZ_CRASH("Unhandled number specialization"); +} + +void LIRGenerator::visitDiv(MDiv* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + MOZ_ASSERT(lhs->type() == rhs->type()); + MOZ_ASSERT(IsNumberType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(lhs->type() == MIRType::Int32); + lowerDivI(ins); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(lhs->type() == MIRType::Int64); + lowerDivI64(ins); + return; + } + + if (ins->type() == MIRType::Double) { + MOZ_ASSERT(lhs->type() == MIRType::Double); + lowerForFPU(new (alloc()) LMathD(JSOp::Div), ins, lhs, rhs); + return; + } + + if (ins->type() == MIRType::Float32) { + MOZ_ASSERT(lhs->type() == MIRType::Float32); + lowerForFPU(new (alloc()) LMathF(JSOp::Div), ins, lhs, rhs); + return; + } + + MOZ_CRASH("Unhandled number specialization"); +} + +void LIRGenerator::visitWasmBuiltinDivI64(MWasmBuiltinDivI64* div) { + lowerWasmBuiltinDivI64(div); +} + +void LIRGenerator::visitWasmBuiltinModI64(MWasmBuiltinModI64* mod) { + lowerWasmBuiltinModI64(mod); +} + +void LIRGenerator::visitBuiltinInt64ToFloatingPoint( + MBuiltinInt64ToFloatingPoint* ins) { + lowerBuiltinInt64ToFloatingPoint(ins); +} + +void LIRGenerator::visitWasmBuiltinTruncateToInt64( + MWasmBuiltinTruncateToInt64* ins) { + lowerWasmBuiltinTruncateToInt64(ins); +} + +void LIRGenerator::visitWasmBuiltinModD(MWasmBuiltinModD* ins) { + MOZ_ASSERT(gen->compilingWasm()); + LWasmBuiltinModD* lir = new (alloc()) LWasmBuiltinModD( + useRegisterAtStart(ins->lhs()), useRegisterAtStart(ins->rhs()), + useFixedAtStart(ins->tls(), WasmTlsReg)); + defineReturn(lir, ins); +} + +void LIRGenerator::visitMod(MMod* ins) { + MOZ_ASSERT(ins->lhs()->type() == ins->rhs()->type()); + MOZ_ASSERT(IsNumberType(ins->type())); + + if (ins->type() == MIRType::Int32) { + MOZ_ASSERT(ins->type() == MIRType::Int32); + MOZ_ASSERT(ins->lhs()->type() == MIRType::Int32); + lowerModI(ins); + return; + } + + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(ins->type() == MIRType::Int64); + MOZ_ASSERT(ins->lhs()->type() == MIRType::Int64); + lowerModI64(ins); + return; + } + + if (ins->type() == MIRType::Double) { + MOZ_ASSERT(ins->lhs()->type() == MIRType::Double); + MOZ_ASSERT(ins->rhs()->type() == MIRType::Double); + + MOZ_ASSERT(!gen->compilingWasm()); + + if (Assembler::HasRoundInstruction(RoundingMode::TowardsZero)) { + if (ins->rhs()->isConstant()) { + double d = ins->rhs()->toConstant()->toDouble(); + int32_t div; + if (mozilla::NumberIsInt32(d, &div) && div > 0 && + mozilla::IsPowerOfTwo(uint32_t(div))) { + auto* lir = new (alloc()) LModPowTwoD(useRegister(ins->lhs()), div); + define(lir, ins); + return; + } + } + } + + // Ion does an unaligned ABI call and thus needs a temp register. + // Note: useRegisterAtStart is safe here, the temp is not a FP register. + LModD* lir = new (alloc()) + LModD(useRegisterAtStart(ins->lhs()), useRegisterAtStart(ins->rhs()), + tempFixed(CallTempReg0)); + defineReturn(lir, ins); + return; + } + + MOZ_CRASH("Unhandled number specialization"); +} + +void LIRGenerator::visitBigIntAdd(MBigIntAdd* ins) { + auto* lir = new (alloc()) LBigIntAdd(useRegister(ins->lhs()), + useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntSub(MBigIntSub* ins) { + auto* lir = new (alloc()) LBigIntSub(useRegister(ins->lhs()), + useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntMul(MBigIntMul* ins) { + auto* lir = new (alloc()) LBigIntMul(useRegister(ins->lhs()), + useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntDiv(MBigIntDiv* ins) { lowerBigIntDiv(ins); } + +void LIRGenerator::visitBigIntMod(MBigIntMod* ins) { lowerBigIntMod(ins); } + +void LIRGenerator::visitBigIntPow(MBigIntPow* ins) { + auto* lir = new (alloc()) LBigIntPow(useRegister(ins->lhs()), + useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntBitAnd(MBigIntBitAnd* ins) { + auto* lir = new (alloc()) LBigIntBitAnd( + useRegister(ins->lhs()), useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntBitOr(MBigIntBitOr* ins) { + auto* lir = new (alloc()) LBigIntBitOr( + useRegister(ins->lhs()), useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntBitXor(MBigIntBitXor* ins) { + auto* lir = new (alloc()) LBigIntBitXor( + useRegister(ins->lhs()), useRegister(ins->rhs()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntLsh(MBigIntLsh* ins) { lowerBigIntLsh(ins); } + +void LIRGenerator::visitBigIntRsh(MBigIntRsh* ins) { lowerBigIntRsh(ins); } + +void LIRGenerator::visitBigIntIncrement(MBigIntIncrement* ins) { + auto* lir = + new (alloc()) LBigIntIncrement(useRegister(ins->input()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntDecrement(MBigIntDecrement* ins) { + auto* lir = + new (alloc()) LBigIntDecrement(useRegister(ins->input()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntNegate(MBigIntNegate* ins) { + auto* lir = new (alloc()) LBigIntNegate(useRegister(ins->input()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntBitNot(MBigIntBitNot* ins) { + auto* lir = + new (alloc()) LBigIntBitNot(useRegister(ins->input()), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitConcat(MConcat* ins) { + MDefinition* lhs = ins->getOperand(0); + MDefinition* rhs = ins->getOperand(1); + + MOZ_ASSERT(lhs->type() == MIRType::String); + MOZ_ASSERT(rhs->type() == MIRType::String); + MOZ_ASSERT(ins->type() == MIRType::String); + + LConcat* lir = new (alloc()) LConcat( + useFixedAtStart(lhs, CallTempReg0), useFixedAtStart(rhs, CallTempReg1), + tempFixed(CallTempReg0), tempFixed(CallTempReg1), tempFixed(CallTempReg2), + tempFixed(CallTempReg3), tempFixed(CallTempReg4)); + defineFixed(lir, ins, LAllocation(AnyRegister(CallTempReg5))); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCharCodeAt(MCharCodeAt* ins) { + MDefinition* str = ins->string(); + MDefinition* idx = ins->index(); + + MOZ_ASSERT(str->type() == MIRType::String); + MOZ_ASSERT(idx->type() == MIRType::Int32); + + LCharCodeAt* lir = + new (alloc()) LCharCodeAt(useRegister(str), useRegister(idx), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitFromCharCode(MFromCharCode* ins) { + MDefinition* code = ins->getOperand(0); + + MOZ_ASSERT(code->type() == MIRType::Int32); + + LFromCharCode* lir = new (alloc()) LFromCharCode(useRegister(code)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitFromCodePoint(MFromCodePoint* ins) { + MDefinition* codePoint = ins->getOperand(0); + + MOZ_ASSERT(codePoint->type() == MIRType::Int32); + + LFromCodePoint* lir = + new (alloc()) LFromCodePoint(useRegister(codePoint), temp(), temp()); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStringConvertCase(MStringConvertCase* ins) { + MOZ_ASSERT(ins->string()->type() == MIRType::String); + + auto* lir = + new (alloc()) LStringConvertCase(useRegisterAtStart(ins->string())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStart(MStart* start) {} + +void LIRGenerator::visitNop(MNop* nop) {} + +void LIRGenerator::visitLimitedTruncate(MLimitedTruncate* nop) { + redefine(nop, nop->input()); +} + +void LIRGenerator::visitOsrEntry(MOsrEntry* entry) { + LOsrEntry* lir = new (alloc()) LOsrEntry(temp()); + defineFixed(lir, entry, LAllocation(AnyRegister(OsrFrameReg))); +} + +void LIRGenerator::visitOsrValue(MOsrValue* value) { + LOsrValue* lir = new (alloc()) LOsrValue(useRegister(value->entry())); + defineBox(lir, value); +} + +void LIRGenerator::visitOsrReturnValue(MOsrReturnValue* value) { + LOsrReturnValue* lir = + new (alloc()) LOsrReturnValue(useRegister(value->entry())); + defineBox(lir, value); +} + +void LIRGenerator::visitOsrEnvironmentChain(MOsrEnvironmentChain* object) { + LOsrEnvironmentChain* lir = + new (alloc()) LOsrEnvironmentChain(useRegister(object->entry())); + define(lir, object); +} + +void LIRGenerator::visitOsrArgumentsObject(MOsrArgumentsObject* object) { + LOsrArgumentsObject* lir = + new (alloc()) LOsrArgumentsObject(useRegister(object->entry())); + define(lir, object); +} + +void LIRGenerator::visitToDouble(MToDouble* convert) { + MDefinition* opd = convert->input(); + mozilla::DebugOnly<MToFPInstruction::ConversionKind> conversion = + convert->conversion(); + + switch (opd->type()) { + case MIRType::Value: { + LValueToDouble* lir = new (alloc()) LValueToDouble(useBox(opd)); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + break; + } + + case MIRType::Null: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + lowerConstantDouble(0, convert); + break; + + case MIRType::Undefined: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + lowerConstantDouble(GenericNaN(), convert); + break; + + case MIRType::Boolean: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + [[fallthrough]]; + + case MIRType::Int32: { + LInt32ToDouble* lir = + new (alloc()) LInt32ToDouble(useRegisterAtStart(opd)); + define(lir, convert); + break; + } + + case MIRType::Float32: { + LFloat32ToDouble* lir = + new (alloc()) LFloat32ToDouble(useRegisterAtStart(opd)); + define(lir, convert); + break; + } + + case MIRType::Double: + redefine(convert, opd); + break; + + default: + // Objects might be effectful. Symbols will throw. + // Strings are complicated - we don't handle them yet. + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitToFloat32(MToFloat32* convert) { + MDefinition* opd = convert->input(); + mozilla::DebugOnly<MToFloat32::ConversionKind> conversion = + convert->conversion(); + + switch (opd->type()) { + case MIRType::Value: { + LValueToFloat32* lir = new (alloc()) LValueToFloat32(useBox(opd)); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + break; + } + + case MIRType::Null: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + lowerConstantFloat32(0, convert); + break; + + case MIRType::Undefined: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + lowerConstantFloat32(GenericNaN(), convert); + break; + + case MIRType::Boolean: + MOZ_ASSERT(conversion == MToFPInstruction::NonStringPrimitives); + [[fallthrough]]; + + case MIRType::Int32: { + LInt32ToFloat32* lir = + new (alloc()) LInt32ToFloat32(useRegisterAtStart(opd)); + define(lir, convert); + break; + } + + case MIRType::Double: { + LDoubleToFloat32* lir = + new (alloc()) LDoubleToFloat32(useRegisterAtStart(opd)); + define(lir, convert); + break; + } + + case MIRType::Float32: + redefine(convert, opd); + break; + + default: + // Objects might be effectful. Symbols will throw. + // Strings are complicated - we don't handle them yet. + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitToNumberInt32(MToNumberInt32* convert) { + MDefinition* opd = convert->input(); + + switch (opd->type()) { + case MIRType::Value: { + auto* lir = new (alloc()) LValueToInt32(useBox(opd), tempDouble(), temp(), + LValueToInt32::NORMAL); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + if (lir->mode() == LValueToInt32::TRUNCATE || + lir->mode() == LValueToInt32::TRUNCATE_NOWRAP) { + assignSafepoint(lir, convert); + } + break; + } + + case MIRType::Null: + MOZ_ASSERT(convert->conversion() == IntConversionInputKind::Any); + define(new (alloc()) LInteger(0), convert); + break; + + case MIRType::Boolean: + MOZ_ASSERT(convert->conversion() == IntConversionInputKind::Any || + convert->conversion() == + IntConversionInputKind::NumbersOrBoolsOnly); + redefine(convert, opd); + break; + + case MIRType::Int32: + redefine(convert, opd); + break; + + case MIRType::Float32: { + LFloat32ToInt32* lir = new (alloc()) LFloat32ToInt32(useRegister(opd)); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + break; + } + + case MIRType::Double: { + LDoubleToInt32* lir = new (alloc()) LDoubleToInt32(useRegister(opd)); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + break; + } + + case MIRType::String: + case MIRType::Symbol: + case MIRType::BigInt: + case MIRType::Object: + case MIRType::Undefined: + // Objects might be effectful. Symbols and BigInts throw. Undefined + // coerces to NaN, not int32. + MOZ_CRASH("ToInt32 invalid input type"); + + default: + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitToIntegerInt32(MToIntegerInt32* convert) { + MDefinition* opd = convert->input(); + + switch (opd->type()) { + case MIRType::Value: { + auto* lir = new (alloc()) LValueToInt32(useBox(opd), tempDouble(), temp(), + LValueToInt32::TRUNCATE_NOWRAP); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + assignSafepoint(lir, convert); + break; + } + + case MIRType::Undefined: + case MIRType::Null: + define(new (alloc()) LInteger(0), convert); + break; + + case MIRType::Boolean: + case MIRType::Int32: + redefine(convert, opd); + break; + + case MIRType::Float32: { + auto* lir = new (alloc()) LFloat32ToIntegerInt32(useRegister(opd)); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + break; + } + + case MIRType::Double: { + auto* lir = new (alloc()) LDoubleToIntegerInt32(useRegister(opd)); + assignSnapshot(lir, convert->bailoutKind()); + define(lir, convert); + break; + } + + case MIRType::String: + case MIRType::Symbol: + case MIRType::BigInt: + case MIRType::Object: + // Objects might be effectful. Symbols and BigInts throw. + // Strings are complicated - we don't handle them yet. + MOZ_CRASH("ToIntegerInt32 invalid input type"); + + default: + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitTruncateToInt32(MTruncateToInt32* truncate) { + MDefinition* opd = truncate->input(); + + switch (opd->type()) { + case MIRType::Value: { + LValueToInt32* lir = new (alloc()) LValueToInt32( + useBox(opd), tempDouble(), temp(), LValueToInt32::TRUNCATE); + assignSnapshot(lir, truncate->bailoutKind()); + define(lir, truncate); + assignSafepoint(lir, truncate); + break; + } + + case MIRType::Null: + case MIRType::Undefined: + define(new (alloc()) LInteger(0), truncate); + break; + + case MIRType::Int32: + case MIRType::Boolean: + redefine(truncate, opd); + break; + + case MIRType::Double: + // May call into JS::ToInt32() on the slow OOL path. + gen->setNeedsStaticStackAlignment(); + lowerTruncateDToInt32(truncate); + break; + + case MIRType::Float32: + // May call into JS::ToInt32() on the slow OOL path. + gen->setNeedsStaticStackAlignment(); + lowerTruncateFToInt32(truncate); + break; + + default: + // Objects might be effectful. Symbols throw. + // Strings are complicated - we don't handle them yet. + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitToBigInt(MToBigInt* ins) { + MDefinition* opd = ins->input(); + + switch (opd->type()) { + case MIRType::Value: { + auto* lir = new (alloc()) LValueToBigInt(useBox(opd)); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + assignSafepoint(lir, ins); + break; + } + + case MIRType::BigInt: + redefine(ins, opd); + break; + + default: + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitToInt64(MToInt64* ins) { + MDefinition* opd = ins->input(); + + switch (opd->type()) { + case MIRType::Value: { + auto* lir = new (alloc()) LValueToInt64(useBox(opd), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineInt64(lir, ins); + assignSafepoint(lir, ins); + break; + } + + case MIRType::Boolean: { + auto* lir = new (alloc()) LBooleanToInt64(useRegisterAtStart(opd)); + defineInt64(lir, ins); + break; + } + + case MIRType::String: { + auto* lir = new (alloc()) LStringToInt64(useRegister(opd)); + defineInt64(lir, ins); + assignSafepoint(lir, ins); + break; + } + + // An Int64 may be passed here from a BigInt to Int64 conversion. + case MIRType::Int64: { + redefine(ins, opd); + break; + } + + default: + // Undefined, Null, Number, and Symbol throw. + // Objects may be effectful. + // BigInt operands are eliminated by the type policy. + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitTruncateBigIntToInt64(MTruncateBigIntToInt64* ins) { + MOZ_ASSERT(ins->input()->type() == MIRType::BigInt); + auto* lir = new (alloc()) LTruncateBigIntToInt64(useRegister(ins->input())); + defineInt64(lir, ins); +} + +void LIRGenerator::visitInt64ToBigInt(MInt64ToBigInt* ins) { + MOZ_ASSERT(ins->input()->type() == MIRType::Int64); + auto* lir = + new (alloc()) LInt64ToBigInt(useInt64Register(ins->input()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitWasmTruncateToInt32(MWasmTruncateToInt32* ins) { + MDefinition* input = ins->input(); + switch (input->type()) { + case MIRType::Double: + case MIRType::Float32: { + auto* lir = new (alloc()) LWasmTruncateToInt32(useRegisterAtStart(input)); + define(lir, ins); + break; + } + default: + MOZ_CRASH("unexpected type in WasmTruncateToInt32"); + } +} + +void LIRGenerator::visitWasmBuiltinTruncateToInt32( + MWasmBuiltinTruncateToInt32* truncate) { + mozilla::DebugOnly<MDefinition*> opd = truncate->input(); + MOZ_ASSERT(opd->type() == MIRType::Double || opd->type() == MIRType::Float32); + + // May call into JS::ToInt32() on the slow OOL path. + gen->setNeedsStaticStackAlignment(); + lowerWasmBuiltinTruncateToInt32(truncate); +} + +void LIRGenerator::visitWasmBoxValue(MWasmBoxValue* ins) { + LWasmBoxValue* lir = new (alloc()) LWasmBoxValue(useBox(ins->input())); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitWasmAnyRefFromJSObject(MWasmAnyRefFromJSObject* ins) { + LWasmAnyRefFromJSObject* lir = + new (alloc()) LWasmAnyRefFromJSObject(useRegisterAtStart(ins->input())); + define(lir, ins); +} + +void LIRGenerator::visitWrapInt64ToInt32(MWrapInt64ToInt32* ins) { + define(new (alloc()) LWrapInt64ToInt32(useInt64AtStart(ins->input())), ins); +} + +void LIRGenerator::visitToString(MToString* ins) { + MDefinition* opd = ins->input(); + + switch (opd->type()) { + case MIRType::Null: { + const JSAtomState& names = gen->runtime->names(); + LPointer* lir = new (alloc()) LPointer(names.null); + define(lir, ins); + break; + } + + case MIRType::Undefined: { + const JSAtomState& names = gen->runtime->names(); + LPointer* lir = new (alloc()) LPointer(names.undefined); + define(lir, ins); + break; + } + + case MIRType::Boolean: { + LBooleanToString* lir = new (alloc()) LBooleanToString(useRegister(opd)); + define(lir, ins); + break; + } + + case MIRType::Double: { + LDoubleToString* lir = + new (alloc()) LDoubleToString(useRegister(opd), temp()); + + define(lir, ins); + assignSafepoint(lir, ins); + break; + } + + case MIRType::Int32: { + LIntToString* lir = new (alloc()) LIntToString(useRegister(opd)); + + define(lir, ins); + assignSafepoint(lir, ins); + break; + } + + case MIRType::String: + redefine(ins, ins->input()); + break; + + case MIRType::Value: { + LValueToString* lir = + new (alloc()) LValueToString(useBox(opd), tempToUnbox()); + if (ins->needsSnapshot()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); + assignSafepoint(lir, ins); + break; + } + + default: + // Float32, symbols, bigint, and objects are not supported. + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitRegExp(MRegExp* ins) { + LRegExp* lir = new (alloc()) LRegExp(temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitRegExpMatcher(MRegExpMatcher* ins) { + MOZ_ASSERT(ins->regexp()->type() == MIRType::Object); + MOZ_ASSERT(ins->string()->type() == MIRType::String); + MOZ_ASSERT(ins->lastIndex()->type() == MIRType::Int32); + + LRegExpMatcher* lir = new (alloc()) LRegExpMatcher( + useFixedAtStart(ins->regexp(), RegExpMatcherRegExpReg), + useFixedAtStart(ins->string(), RegExpMatcherStringReg), + useFixedAtStart(ins->lastIndex(), RegExpMatcherLastIndexReg)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitRegExpSearcher(MRegExpSearcher* ins) { + MOZ_ASSERT(ins->regexp()->type() == MIRType::Object); + MOZ_ASSERT(ins->string()->type() == MIRType::String); + MOZ_ASSERT(ins->lastIndex()->type() == MIRType::Int32); + + LRegExpSearcher* lir = new (alloc()) LRegExpSearcher( + useFixedAtStart(ins->regexp(), RegExpTesterRegExpReg), + useFixedAtStart(ins->string(), RegExpTesterStringReg), + useFixedAtStart(ins->lastIndex(), RegExpTesterLastIndexReg)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitRegExpTester(MRegExpTester* ins) { + MOZ_ASSERT(ins->regexp()->type() == MIRType::Object); + MOZ_ASSERT(ins->string()->type() == MIRType::String); + MOZ_ASSERT(ins->lastIndex()->type() == MIRType::Int32); + + LRegExpTester* lir = new (alloc()) LRegExpTester( + useFixedAtStart(ins->regexp(), RegExpTesterRegExpReg), + useFixedAtStart(ins->string(), RegExpTesterStringReg), + useFixedAtStart(ins->lastIndex(), RegExpTesterLastIndexReg)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitRegExpPrototypeOptimizable( + MRegExpPrototypeOptimizable* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + LRegExpPrototypeOptimizable* lir = new (alloc()) + LRegExpPrototypeOptimizable(useRegister(ins->object()), temp()); + define(lir, ins); +} + +void LIRGenerator::visitRegExpInstanceOptimizable( + MRegExpInstanceOptimizable* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->proto()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + LRegExpInstanceOptimizable* lir = new (alloc()) LRegExpInstanceOptimizable( + useRegister(ins->object()), useRegister(ins->proto()), temp()); + define(lir, ins); +} + +void LIRGenerator::visitGetFirstDollarIndex(MGetFirstDollarIndex* ins) { + MOZ_ASSERT(ins->str()->type() == MIRType::String); + MOZ_ASSERT(ins->type() == MIRType::Int32); + LGetFirstDollarIndex* lir = new (alloc()) + LGetFirstDollarIndex(useRegister(ins->str()), temp(), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStringReplace(MStringReplace* ins) { + MOZ_ASSERT(ins->pattern()->type() == MIRType::String); + MOZ_ASSERT(ins->string()->type() == MIRType::String); + MOZ_ASSERT(ins->replacement()->type() == MIRType::String); + + LStringReplace* lir = new (alloc()) + LStringReplace(useRegisterOrConstantAtStart(ins->string()), + useRegisterAtStart(ins->pattern()), + useRegisterOrConstantAtStart(ins->replacement())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBinaryCache(MBinaryCache* ins) { + MDefinition* lhs = ins->getOperand(0); + MDefinition* rhs = ins->getOperand(1); + + MOZ_ASSERT(ins->type() == MIRType::Value || ins->type() == MIRType::Boolean); + LInstruction* lir; + if (ins->type() == MIRType::Value) { + LBinaryValueCache* valueLir = new (alloc()) LBinaryValueCache( + useBox(lhs), useBox(rhs), tempFixed(FloatReg0), tempFixed(FloatReg1)); + defineBox(valueLir, ins); + lir = valueLir; + } else { + MOZ_ASSERT(ins->type() == MIRType::Boolean); + LBinaryBoolCache* boolLir = new (alloc()) LBinaryBoolCache( + useBox(lhs), useBox(rhs), tempFixed(FloatReg0), tempFixed(FloatReg1)); + define(boolLir, ins); + lir = boolLir; + } + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitUnaryCache(MUnaryCache* ins) { + MDefinition* input = ins->getOperand(0); + MOZ_ASSERT(ins->type() == MIRType::Value); + + LUnaryCache* lir = new (alloc()) LUnaryCache(useBox(input)); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitClassConstructor(MClassConstructor* ins) { + LClassConstructor* lir = new (alloc()) LClassConstructor(); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitDerivedClassConstructor(MDerivedClassConstructor* ins) { + MDefinition* proto = ins->prototype(); + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto* lir = new (alloc()) LDerivedClassConstructor(useRegisterAtStart(proto)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitModuleMetadata(MModuleMetadata* ins) { + LModuleMetadata* lir = new (alloc()) LModuleMetadata(); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitDynamicImport(MDynamicImport* ins) { + LDynamicImport* lir = + new (alloc()) LDynamicImport(useBoxAtStart(ins->specifier())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitLambda(MLambda* ins) { + auto* lir = + new (alloc()) LLambda(useRegister(ins->environmentChain()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitLambdaArrow(MLambdaArrow* ins) { + MOZ_ASSERT(ins->environmentChain()->type() == MIRType::Object); + MOZ_ASSERT(ins->newTargetDef()->type() == MIRType::Value); + + LLambdaArrow* lir = + new (alloc()) LLambdaArrow(useRegister(ins->environmentChain()), + useBox(ins->newTargetDef()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitFunctionWithProto(MFunctionWithProto* ins) { + MOZ_ASSERT(ins->environmentChain()->type() == MIRType::Object); + MOZ_ASSERT(ins->prototype()->type() == MIRType::Object); + + auto* lir = new (alloc()) + LFunctionWithProto(useRegisterAtStart(ins->environmentChain()), + useRegisterAtStart(ins->prototype())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitSetFunName(MSetFunName* ins) { + MOZ_ASSERT(ins->fun()->type() == MIRType::Object); + MOZ_ASSERT(ins->name()->type() == MIRType::Value); + + LSetFunName* lir = new (alloc()) + LSetFunName(useRegisterAtStart(ins->fun()), useBoxAtStart(ins->name())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitNewLexicalEnvironmentObject( + MNewLexicalEnvironmentObject* ins) { + MDefinition* enclosing = ins->enclosing(); + MOZ_ASSERT(enclosing->type() == MIRType::Object); + + LNewLexicalEnvironmentObject* lir = + new (alloc()) LNewLexicalEnvironmentObject(useRegisterAtStart(enclosing)); + + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCopyLexicalEnvironmentObject( + MCopyLexicalEnvironmentObject* ins) { + MDefinition* env = ins->env(); + MOZ_ASSERT(env->type() == MIRType::Object); + + LCopyLexicalEnvironmentObject* lir = + new (alloc()) LCopyLexicalEnvironmentObject(useRegisterAtStart(env)); + + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitKeepAliveObject(MKeepAliveObject* ins) { + MDefinition* obj = ins->object(); + MOZ_ASSERT(obj->type() == MIRType::Object); + + add(new (alloc()) LKeepAliveObject(useKeepalive(obj)), ins); +} + +void LIRGenerator::visitSlots(MSlots* ins) { + define(new (alloc()) LSlots(useRegisterAtStart(ins->object())), ins); +} + +void LIRGenerator::visitElements(MElements* ins) { + define(new (alloc()) LElements(useRegisterAtStart(ins->object())), ins); +} + +void LIRGenerator::visitLoadDynamicSlot(MLoadDynamicSlot* ins) { + MOZ_ASSERT(ins->type() == MIRType::Value); + defineBox(new (alloc()) LLoadDynamicSlotV(useRegisterAtStart(ins->slots())), + ins); +} + +void LIRGenerator::visitFunctionEnvironment(MFunctionEnvironment* ins) { + define(new (alloc()) + LFunctionEnvironment(useRegisterAtStart(ins->function())), + ins); +} + +void LIRGenerator::visitHomeObject(MHomeObject* ins) { + define(new (alloc()) LHomeObject(useRegisterAtStart(ins->function())), ins); +} + +void LIRGenerator::visitHomeObjectSuperBase(MHomeObjectSuperBase* ins) { + MOZ_ASSERT(ins->homeObject()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto lir = + new (alloc()) LHomeObjectSuperBase(useRegisterAtStart(ins->homeObject())); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitInterruptCheck(MInterruptCheck* ins) { + LInstruction* lir = new (alloc()) LInterruptCheck(); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitWasmInterruptCheck(MWasmInterruptCheck* ins) { + auto* lir = + new (alloc()) LWasmInterruptCheck(useRegisterAtStart(ins->tlsPtr())); + add(lir, ins); + + assignWasmSafepoint(lir, ins); +} + +void LIRGenerator::visitWasmTrap(MWasmTrap* ins) { + add(new (alloc()) LWasmTrap, ins); +} + +void LIRGenerator::visitWasmReinterpret(MWasmReinterpret* ins) { + if (ins->type() == MIRType::Int64) { + defineInt64(new (alloc()) + LWasmReinterpretToI64(useRegisterAtStart(ins->input())), + ins); + } else if (ins->input()->type() == MIRType::Int64) { + define(new (alloc()) + LWasmReinterpretFromI64(useInt64RegisterAtStart(ins->input())), + ins); + } else { + define(new (alloc()) LWasmReinterpret(useRegisterAtStart(ins->input())), + ins); + } +} + +void LIRGenerator::visitStoreDynamicSlot(MStoreDynamicSlot* ins) { + LInstruction* lir; + + switch (ins->value()->type()) { + case MIRType::Value: + lir = new (alloc()) + LStoreDynamicSlotV(useRegister(ins->slots()), useBox(ins->value())); + add(lir, ins); + break; + + case MIRType::Double: + add(new (alloc()) LStoreDynamicSlotT(useRegister(ins->slots()), + useRegister(ins->value())), + ins); + break; + + case MIRType::Float32: + MOZ_CRASH("Float32 shouldn't be stored in a slot."); + + default: + add(new (alloc()) LStoreDynamicSlotT(useRegister(ins->slots()), + useRegisterOrConstant(ins->value())), + ins); + break; + } +} + +// Returns true iff |def| is a constant that's either not a GC thing or is not +// allocated in the nursery. +static bool IsNonNurseryConstant(MDefinition* def) { + if (!def->isConstant()) { + return false; + } + Value v = def->toConstant()->toJSValue(); + return !v.isGCThing() || !IsInsideNursery(v.toGCThing()); +} + +void LIRGenerator::visitPostWriteBarrier(MPostWriteBarrier* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + // LPostWriteBarrier assumes that if it has a constant object then that + // object is tenured, and does not need to be tested for being in the + // nursery. Ensure that assumption holds by lowering constant nursery + // objects to a register. + bool useConstantObject = IsNonNurseryConstant(ins->object()); + + switch (ins->value()->type()) { + case MIRType::Object: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteBarrierO* lir = new (alloc()) + LPostWriteBarrierO(useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::String: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteBarrierS* lir = new (alloc()) + LPostWriteBarrierS(useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::BigInt: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + auto* lir = new (alloc()) + LPostWriteBarrierBI(useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::Value: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteBarrierV* lir = new (alloc()) + LPostWriteBarrierV(useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useBox(ins->value()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + default: + // Currently, only objects and strings can be in the nursery. Other + // instruction types cannot hold nursery pointers. + break; + } +} + +void LIRGenerator::visitPostWriteElementBarrier(MPostWriteElementBarrier* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + // LPostWriteElementBarrier assumes that if it has a constant object then that + // object is tenured, and does not need to be tested for being in the + // nursery. Ensure that assumption holds by lowering constant nursery + // objects to a register. + bool useConstantObject = + ins->object()->isConstant() && + !IsInsideNursery(&ins->object()->toConstant()->toObject()); + + switch (ins->value()->type()) { + case MIRType::Object: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteElementBarrierO* lir = new (alloc()) LPostWriteElementBarrierO( + useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), useRegister(ins->index()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::String: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteElementBarrierS* lir = new (alloc()) LPostWriteElementBarrierS( + useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), useRegister(ins->index()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::BigInt: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + auto* lir = new (alloc()) LPostWriteElementBarrierBI( + useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->value()), useRegister(ins->index()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + case MIRType::Value: { + LDefinition tmp = + needTempForPostBarrier() ? temp() : LDefinition::BogusTemp(); + LPostWriteElementBarrierV* lir = new (alloc()) LPostWriteElementBarrierV( + useConstantObject ? useOrConstant(ins->object()) + : useRegister(ins->object()), + useRegister(ins->index()), useBox(ins->value()), tmp); + add(lir, ins); + assignSafepoint(lir, ins); + break; + } + default: + // Currently, only objects, strings, and bigints can be in the nursery. + // Other instruction types cannot hold nursery pointers. + break; + } +} + +void LIRGenerator::visitArrayLength(MArrayLength* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + auto* lir = new (alloc()) LArrayLength(useRegisterAtStart(ins->elements())); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitSetArrayLength(MSetArrayLength* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + MOZ_ASSERT(ins->index()->isConstant()); + add(new (alloc()) LSetArrayLength(useRegister(ins->elements()), + useRegisterOrConstant(ins->index())), + ins); +} + +void LIRGenerator::visitFunctionLength(MFunctionLength* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = new (alloc()) LFunctionLength(useRegister(ins->function())); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitFunctionName(MFunctionName* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = new (alloc()) LFunctionName(useRegister(ins->function())); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); +} + +void LIRGenerator::visitGetNextEntryForIterator(MGetNextEntryForIterator* ins) { + MOZ_ASSERT(ins->iter()->type() == MIRType::Object); + MOZ_ASSERT(ins->result()->type() == MIRType::Object); + auto lir = new (alloc()) LGetNextEntryForIterator(useRegister(ins->iter()), + useRegister(ins->result()), + temp(), temp(), temp()); + define(lir, ins); +} + +void LIRGenerator::visitArrayBufferByteLengthInt32( + MArrayBufferByteLengthInt32* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + auto* lir = new (alloc()) + LArrayBufferByteLengthInt32(useRegisterAtStart(ins->object())); + define(lir, ins); +} + +void LIRGenerator::visitArrayBufferViewLength(MArrayBufferViewLength* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + define(new (alloc()) + LArrayBufferViewLength(useRegisterAtStart(ins->object())), + ins); +} + +void LIRGenerator::visitArrayBufferViewByteOffset( + MArrayBufferViewByteOffset* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + define(new (alloc()) + LArrayBufferViewByteOffset(useRegisterAtStart(ins->object())), + ins); +} + +void LIRGenerator::visitArrayBufferViewElements(MArrayBufferViewElements* ins) { + MOZ_ASSERT(ins->type() == MIRType::Elements); + define(new (alloc()) + LArrayBufferViewElements(useRegisterAtStart(ins->object())), + ins); +} + +void LIRGenerator::visitTypedArrayElementShift(MTypedArrayElementShift* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + define(new (alloc()) + LTypedArrayElementShift(useRegisterAtStart(ins->object())), + ins); +} + +void LIRGenerator::visitTypedArrayIndexToInt32(MTypedArrayIndexToInt32* ins) { + MDefinition* input = ins->input(); + if (input->type() == MIRType::Int32) { + redefine(ins, input); + } else { + MOZ_ASSERT(input->type() == MIRType::Double); + define(new (alloc()) LTypedArrayIndexToInt32(useRegister(input)), ins); + } +} + +void LIRGenerator::visitInitializedLength(MInitializedLength* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + define(new (alloc()) LInitializedLength(useRegisterAtStart(ins->elements())), + ins); +} + +void LIRGenerator::visitSetInitializedLength(MSetInitializedLength* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + MOZ_ASSERT(ins->index()->isConstant()); + add(new (alloc()) LSetInitializedLength(useRegister(ins->elements()), + useRegisterOrConstant(ins->index())), + ins); +} + +void LIRGenerator::visitNot(MNot* ins) { + MDefinition* op = ins->input(); + + // String is converted to length of string in the type analysis phase (see + // TestPolicy). + MOZ_ASSERT(op->type() != MIRType::String); + MOZ_ASSERT(op->type() != MIRType::BigInt, + "BigInt should be boxed by TestPolicy"); + + // - boolean: x xor 1 + // - int32: LCompare(x, 0) + // - double: LCompare(x, 0) + // - null or undefined: true + // - object: false if it never emulates undefined, else LNotO(x) + switch (op->type()) { + case MIRType::Boolean: { + MConstant* cons = MConstant::New(alloc(), Int32Value(1)); + ins->block()->insertBefore(ins, cons); + lowerForALU(new (alloc()) LBitOpI(JSOp::BitXor), ins, op, cons); + break; + } + case MIRType::Int32: + define(new (alloc()) LNotI(useRegisterAtStart(op)), ins); + break; + case MIRType::Int64: + define(new (alloc()) LNotI64(useInt64RegisterAtStart(op)), ins); + break; + case MIRType::Double: + define(new (alloc()) LNotD(useRegister(op)), ins); + break; + case MIRType::Float32: + define(new (alloc()) LNotF(useRegister(op)), ins); + break; + case MIRType::Undefined: + case MIRType::Null: + define(new (alloc()) LInteger(1), ins); + break; + case MIRType::Symbol: + define(new (alloc()) LInteger(0), ins); + break; + case MIRType::Object: + define(new (alloc()) LNotO(useRegister(op)), ins); + break; + case MIRType::Value: { + auto* lir = new (alloc()) LNotV(useBox(op), tempDouble(), temp(), temp()); + define(lir, ins); + break; + } + + default: + MOZ_CRASH("Unexpected MIRType."); + } +} + +void LIRGenerator::visitBoundsCheck(MBoundsCheck* ins) { + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + MOZ_ASSERT(ins->length()->type() == MIRType::Int32); + MOZ_ASSERT(ins->type() == MIRType::Int32); + + if (!ins->fallible()) { + return; + } + + LInstruction* check; + if (ins->minimum() || ins->maximum()) { + check = new (alloc()) LBoundsCheckRange(useRegisterOrConstant(ins->index()), + useAny(ins->length()), temp()); + } else { + check = new (alloc()) LBoundsCheck(useRegisterOrConstant(ins->index()), + useAnyOrConstant(ins->length())); + } + assignSnapshot(check, ins->bailoutKind()); + add(check, ins); +} + +void LIRGenerator::visitSpectreMaskIndex(MSpectreMaskIndex* ins) { + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + MOZ_ASSERT(ins->length()->type() == MIRType::Int32); + MOZ_ASSERT(ins->type() == MIRType::Int32); + + auto* lir = new (alloc()) + LSpectreMaskIndex(useRegister(ins->index()), useAny(ins->length())); + define(lir, ins); +} + +void LIRGenerator::visitBoundsCheckLower(MBoundsCheckLower* ins) { + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + if (!ins->fallible()) { + return; + } + + LInstruction* check = + new (alloc()) LBoundsCheckLower(useRegister(ins->index())); + assignSnapshot(check, ins->bailoutKind()); + add(check, ins); +} + +void LIRGenerator::visitInArray(MInArray* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + MOZ_ASSERT(ins->initLength()->type() == MIRType::Int32); + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + + auto* lir = new (alloc()) LInArray(useRegister(ins->elements()), + useRegisterOrConstant(ins->index()), + useRegister(ins->initLength())); + if (ins->needsNegativeIntCheck()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); +} + +void LIRGenerator::visitGuardElementNotHole(MGuardElementNotHole* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + auto* guard = new (alloc()) + LGuardElementNotHole(useRegisterAtStart(ins->elements()), + useRegisterOrConstantAtStart(ins->index())); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); +} + +void LIRGenerator::visitLoadElement(MLoadElement* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + MOZ_ASSERT(ins->type() == MIRType::Value); + + auto* lir = new (alloc()) LLoadElementV(useRegister(ins->elements()), + useRegisterOrConstant(ins->index())); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + defineBox(lir, ins); +} + +void LIRGenerator::visitLoadElementHole(MLoadElementHole* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + MOZ_ASSERT(ins->initLength()->type() == MIRType::Int32); + MOZ_ASSERT(ins->type() == MIRType::Value); + + LLoadElementHole* lir = new (alloc()) + LLoadElementHole(useRegister(ins->elements()), useRegister(ins->index()), + useRegister(ins->initLength())); + if (ins->needsNegativeIntCheck()) { + assignSnapshot(lir, ins->bailoutKind()); + } + defineBox(lir, ins); +} + +void LIRGenerator::visitStoreElement(MStoreElement* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + const LUse elements = useRegister(ins->elements()); + const LAllocation index = useRegisterOrConstant(ins->index()); + + switch (ins->value()->type()) { + case MIRType::Value: { + LInstruction* lir = + new (alloc()) LStoreElementV(elements, index, useBox(ins->value())); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + add(lir, ins); + break; + } + + default: { + const LAllocation value = useRegisterOrNonDoubleConstant(ins->value()); + LInstruction* lir = new (alloc()) LStoreElementT(elements, index, value); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + add(lir, ins); + break; + } + } +} + +void LIRGenerator::visitStoreHoleValueElement(MStoreHoleValueElement* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + auto* lir = new (alloc()) LStoreHoleValueElement(useRegister(ins->elements()), + useRegister(ins->index())); + add(lir, ins); +} + +static bool BoundsCheckNeedsSpectreTemp() { + // On x86, spectreBoundsCheck32 can emit better code if it has a scratch + // register and index masking is enabled. +#ifdef JS_CODEGEN_X86 + return JitOptions.spectreIndexMasking; +#else + return false; +#endif +} + +void LIRGenerator::visitStoreElementHole(MStoreElementHole* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + const LUse object = useRegister(ins->object()); + const LUse elements = useRegister(ins->elements()); + const LAllocation index = useRegister(ins->index()); + + LDefinition spectreTemp = + BoundsCheckNeedsSpectreTemp() ? temp() : LDefinition::BogusTemp(); + + LInstruction* lir; + switch (ins->value()->type()) { + case MIRType::Value: + lir = new (alloc()) LStoreElementHoleV(object, elements, index, + useBox(ins->value()), spectreTemp); + break; + + default: { + const LAllocation value = useRegisterOrNonDoubleConstant(ins->value()); + lir = new (alloc()) + LStoreElementHoleT(object, elements, index, value, spectreTemp); + break; + } + } + + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitEffectiveAddress(MEffectiveAddress* ins) { + define(new (alloc()) LEffectiveAddress(useRegister(ins->base()), + useRegister(ins->index())), + ins); +} + +void LIRGenerator::visitArrayPopShift(MArrayPopShift* ins) { + MOZ_ASSERT(ins->type() == MIRType::Value); + + auto* lir = + new (alloc()) LArrayPopShift(useRegister(ins->object()), temp(), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineBox(lir, ins); + + if (ins->mode() == MArrayPopShift::Shift) { + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitArrayPush(MArrayPush* ins) { + MOZ_ASSERT(ins->type() == MIRType::Int32); + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + + LUse object = useRegister(ins->object()); + + LDefinition spectreTemp = + BoundsCheckNeedsSpectreTemp() ? temp() : LDefinition::BogusTemp(); + + auto* lir = new (alloc()) + LArrayPush(object, useBox(ins->value()), temp(), spectreTemp); + // We will bailout before pushing if the length would overflow INT32_MAX. + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitArraySlice(MArraySlice* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->begin()->type() == MIRType::Int32); + MOZ_ASSERT(ins->end()->type() == MIRType::Int32); + + LArraySlice* lir = new (alloc()) LArraySlice( + useRegisterAtStart(ins->object()), useRegisterAtStart(ins->begin()), + useRegisterAtStart(ins->end()), tempFixed(CallTempReg0), + tempFixed(CallTempReg1)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitArrayJoin(MArrayJoin* ins) { + MOZ_ASSERT(ins->type() == MIRType::String); + MOZ_ASSERT(ins->array()->type() == MIRType::Object); + MOZ_ASSERT(ins->sep()->type() == MIRType::String); + + auto* lir = new (alloc()) + LArrayJoin(useRegisterAtStart(ins->array()), + useRegisterAtStart(ins->sep()), tempFixed(CallTempReg0)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitStringSplit(MStringSplit* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + MOZ_ASSERT(ins->string()->type() == MIRType::String); + MOZ_ASSERT(ins->separator()->type() == MIRType::String); + + LStringSplit* lir = new (alloc()) LStringSplit( + useRegisterAtStart(ins->string()), useRegisterAtStart(ins->separator())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitLoadUnboxedScalar(MLoadUnboxedScalar* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + const LUse elements = useRegister(ins->elements()); + const LAllocation index = useRegisterOrConstant(ins->index()); + + MOZ_ASSERT(IsNumericType(ins->type()) || ins->type() == MIRType::Boolean); + + Synchronization sync = Synchronization::Load(); + if (ins->requiresMemoryBarrier()) { + LMemoryBarrier* fence = new (alloc()) LMemoryBarrier(sync.barrierBefore); + add(fence, ins); + } + + if (!Scalar::isBigIntType(ins->storageType())) { + // We need a temp register for Uint32Array with known double result. + LDefinition tempDef = LDefinition::BogusTemp(); + if (ins->storageType() == Scalar::Uint32 && + IsFloatingPointType(ins->type())) { + tempDef = temp(); + } + + auto* lir = new (alloc()) LLoadUnboxedScalar(elements, index, tempDef); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); + } else { + MOZ_ASSERT(ins->type() == MIRType::BigInt); + + auto* lir = + new (alloc()) LLoadUnboxedBigInt(elements, index, temp(), tempInt64()); + define(lir, ins); + assignSafepoint(lir, ins); + } + + if (ins->requiresMemoryBarrier()) { + LMemoryBarrier* fence = new (alloc()) LMemoryBarrier(sync.barrierAfter); + add(fence, ins); + } +} + +void LIRGenerator::visitLoadDataViewElement(MLoadDataViewElement* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + MOZ_ASSERT(IsNumericType(ins->type())); + + const LUse elements = useRegister(ins->elements()); + const LUse index = useRegister(ins->index()); + const LAllocation littleEndian = useRegisterOrConstant(ins->littleEndian()); + + // We need a temp register for: + // - Uint32Array with known double result, + // - Float32Array, + // - and BigInt64Array and BigUint64Array. + LDefinition tempDef = LDefinition::BogusTemp(); + if ((ins->storageType() == Scalar::Uint32 && + IsFloatingPointType(ins->type())) || + ins->storageType() == Scalar::Float32) { + tempDef = temp(); + } + if (Scalar::isBigIntType(ins->storageType())) { +#ifdef JS_CODEGEN_X86 + // There are not enough registers on x86. + if (littleEndian.isConstant()) { + tempDef = temp(); + } +#else + tempDef = temp(); +#endif + } + + // We also need a separate 64-bit temp register for: + // - Float64Array + // - and BigInt64Array and BigUint64Array. + LInt64Definition temp64Def = LInt64Definition::BogusTemp(); + if (Scalar::byteSize(ins->storageType()) == 8) { + temp64Def = tempInt64(); + } + + auto* lir = new (alloc()) + LLoadDataViewElement(elements, index, littleEndian, tempDef, temp64Def); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); + if (Scalar::isBigIntType(ins->storageType())) { + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitClampToUint8(MClampToUint8* ins) { + MDefinition* in = ins->input(); + + switch (in->type()) { + case MIRType::Boolean: + redefine(ins, in); + break; + + case MIRType::Int32: + defineReuseInput(new (alloc()) LClampIToUint8(useRegisterAtStart(in)), + ins, 0); + break; + + case MIRType::Double: + // LClampDToUint8 clobbers its input register. Making it available as + // a temp copy describes this behavior to the register allocator. + define(new (alloc()) + LClampDToUint8(useRegisterAtStart(in), tempCopy(in, 0)), + ins); + break; + + case MIRType::Value: { + LClampVToUint8* lir = + new (alloc()) LClampVToUint8(useBox(in), tempDouble()); + assignSnapshot(lir, ins->bailoutKind()); + define(lir, ins); + assignSafepoint(lir, ins); + break; + } + + default: + MOZ_CRASH("unexpected type"); + } +} + +void LIRGenerator::visitLoadTypedArrayElementHole( + MLoadTypedArrayElementHole* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + MOZ_ASSERT(ins->type() == MIRType::Value); + + const LUse object = useRegister(ins->object()); + const LAllocation index = useRegister(ins->index()); + + if (!Scalar::isBigIntType(ins->arrayType())) { + auto* lir = new (alloc()) LLoadTypedArrayElementHole(object, index, temp()); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + defineBox(lir, ins); + } else { +#ifdef JS_CODEGEN_X86 + LDefinition tmp = LDefinition::BogusTemp(); +#else + LDefinition tmp = temp(); +#endif + + auto* lir = new (alloc()) + LLoadTypedArrayElementHoleBigInt(object, index, tmp, tempInt64()); + defineBox(lir, ins); + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitStoreUnboxedScalar(MStoreUnboxedScalar* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + + if (ins->isFloatWrite()) { + MOZ_ASSERT_IF(ins->writeType() == Scalar::Float32, + ins->value()->type() == MIRType::Float32); + MOZ_ASSERT_IF(ins->writeType() == Scalar::Float64, + ins->value()->type() == MIRType::Double); + } else if (ins->isBigIntWrite()) { + MOZ_ASSERT(ins->value()->type() == MIRType::BigInt); + } else { + MOZ_ASSERT(ins->value()->type() == MIRType::Int32); + } + + LUse elements = useRegister(ins->elements()); + LAllocation index = useRegisterOrConstant(ins->index()); + LAllocation value; + + // For byte arrays, the value has to be in a byte register on x86. + if (ins->isByteWrite()) { + value = useByteOpRegisterOrNonDoubleConstant(ins->value()); + } else if (ins->isBigIntWrite()) { + value = useRegister(ins->value()); + } else { + value = useRegisterOrNonDoubleConstant(ins->value()); + } + + // Optimization opportunity for atomics: on some platforms there + // is a store instruction that incorporates the necessary + // barriers, and we could use that instead of separate barrier and + // store instructions. See bug #1077027. + Synchronization sync = Synchronization::Store(); + if (ins->requiresMemoryBarrier()) { + LMemoryBarrier* fence = new (alloc()) LMemoryBarrier(sync.barrierBefore); + add(fence, ins); + } + if (!ins->isBigIntWrite()) { + add(new (alloc()) LStoreUnboxedScalar(elements, index, value), ins); + } else { + add(new (alloc()) LStoreUnboxedBigInt(elements, index, value, tempInt64()), + ins); + } + if (ins->requiresMemoryBarrier()) { + LMemoryBarrier* fence = new (alloc()) LMemoryBarrier(sync.barrierAfter); + add(fence, ins); + } +} + +void LIRGenerator::visitStoreDataViewElement(MStoreDataViewElement* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + MOZ_ASSERT(ins->littleEndian()->type() == MIRType::Boolean); + + if (ins->isFloatWrite()) { + MOZ_ASSERT_IF(ins->writeType() == Scalar::Float32, + ins->value()->type() == MIRType::Float32); + MOZ_ASSERT_IF(ins->writeType() == Scalar::Float64, + ins->value()->type() == MIRType::Double); + } else if (ins->isBigIntWrite()) { + MOZ_ASSERT(ins->value()->type() == MIRType::BigInt); + } else { + MOZ_ASSERT(ins->value()->type() == MIRType::Int32); + } + + LUse elements = useRegister(ins->elements()); + LUse index = useRegister(ins->index()); + LAllocation value; + if (ins->isBigIntWrite()) { + value = useRegister(ins->value()); + } else { + value = useRegisterOrNonDoubleConstant(ins->value()); + } + LAllocation littleEndian = useRegisterOrConstant(ins->littleEndian()); + + LDefinition tempDef = LDefinition::BogusTemp(); + LInt64Definition temp64Def = LInt64Definition::BogusTemp(); + if (Scalar::byteSize(ins->writeType()) < 8) { + tempDef = temp(); + } else { + temp64Def = tempInt64(); + } + + add(new (alloc()) LStoreDataViewElement(elements, index, value, littleEndian, + tempDef, temp64Def), + ins); +} + +void LIRGenerator::visitStoreTypedArrayElementHole( + MStoreTypedArrayElementHole* ins) { + MOZ_ASSERT(ins->elements()->type() == MIRType::Elements); + MOZ_ASSERT(ins->index()->type() == MIRType::Int32); + MOZ_ASSERT(ins->length()->type() == MIRType::Int32); + + if (ins->isFloatWrite()) { + MOZ_ASSERT_IF(ins->arrayType() == Scalar::Float32, + ins->value()->type() == MIRType::Float32); + MOZ_ASSERT_IF(ins->arrayType() == Scalar::Float64, + ins->value()->type() == MIRType::Double); + } else if (ins->isBigIntWrite()) { + MOZ_ASSERT(ins->value()->type() == MIRType::BigInt); + } else { + MOZ_ASSERT(ins->value()->type() == MIRType::Int32); + } + + LUse elements = useRegister(ins->elements()); + LAllocation length = useAny(ins->length()); + LAllocation index = useRegister(ins->index()); + + // For byte arrays, the value has to be in a byte register on x86. + LAllocation value; + if (ins->isByteWrite()) { + value = useByteOpRegisterOrNonDoubleConstant(ins->value()); + } else if (ins->isBigIntWrite()) { + value = useRegister(ins->value()); + } else { + value = useRegisterOrNonDoubleConstant(ins->value()); + } + + if (!ins->isBigIntWrite()) { + LDefinition spectreTemp = + BoundsCheckNeedsSpectreTemp() ? temp() : LDefinition::BogusTemp(); + auto* lir = new (alloc()) LStoreTypedArrayElementHole( + elements, length, index, value, spectreTemp); + add(lir, ins); + } else { + auto* lir = new (alloc()) LStoreTypedArrayElementHoleBigInt( + elements, length, index, value, tempInt64()); + add(lir, ins); + } +} + +void LIRGenerator::visitLoadFixedSlot(MLoadFixedSlot* ins) { + MDefinition* obj = ins->object(); + MOZ_ASSERT(obj->type() == MIRType::Object); + + MIRType type = ins->type(); + + if (type == MIRType::Value) { + LLoadFixedSlotV* lir = + new (alloc()) LLoadFixedSlotV(useRegisterAtStart(obj)); + defineBox(lir, ins); + } else { + LLoadFixedSlotT* lir = + new (alloc()) LLoadFixedSlotT(useRegisterForTypedLoad(obj, type)); + define(lir, ins); + } +} + +void LIRGenerator::visitLoadFixedSlotAndUnbox(MLoadFixedSlotAndUnbox* ins) { + MDefinition* obj = ins->object(); + MOZ_ASSERT(obj->type() == MIRType::Object); + + LLoadFixedSlotAndUnbox* lir = + new (alloc()) LLoadFixedSlotAndUnbox(useRegisterAtStart(obj)); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); +} + +void LIRGenerator::visitLoadDynamicSlotAndUnbox(MLoadDynamicSlotAndUnbox* ins) { + MDefinition* slots = ins->slots(); + MOZ_ASSERT(slots->type() == MIRType::Slots); + + auto* lir = new (alloc()) LLoadDynamicSlotAndUnbox(useRegisterAtStart(slots)); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); +} + +void LIRGenerator::visitLoadElementAndUnbox(MLoadElementAndUnbox* ins) { + MDefinition* elements = ins->elements(); + MDefinition* index = ins->index(); + MOZ_ASSERT(elements->type() == MIRType::Elements); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) + LLoadElementAndUnbox(useRegister(elements), useRegisterOrConstant(index)); + if (ins->fallible()) { + assignSnapshot(lir, ins->bailoutKind()); + } + define(lir, ins); +} + +void LIRGenerator::visitAddAndStoreSlot(MAddAndStoreSlot* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + LDefinition maybeTemp = LDefinition::BogusTemp(); + if (ins->kind() != MAddAndStoreSlot::Kind::FixedSlot) { + maybeTemp = temp(); + } + + auto* lir = new (alloc()) LAddAndStoreSlot(useRegister(ins->object()), + useBox(ins->value()), maybeTemp); + add(lir, ins); +} + +void LIRGenerator::visitAllocateAndStoreSlot(MAllocateAndStoreSlot* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = new (alloc()) LAllocateAndStoreSlot( + useRegisterAtStart(ins->object()), useBoxAtStart(ins->value()), + tempFixed(CallTempReg0), tempFixed(CallTempReg1)); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); +} + +void LIRGenerator::visitStoreFixedSlot(MStoreFixedSlot* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + if (ins->value()->type() == MIRType::Value) { + LStoreFixedSlotV* lir = new (alloc()) + LStoreFixedSlotV(useRegister(ins->object()), useBox(ins->value())); + add(lir, ins); + } else { + LStoreFixedSlotT* lir = new (alloc()) LStoreFixedSlotT( + useRegister(ins->object()), useRegisterOrConstant(ins->value())); + add(lir, ins); + } +} + +void LIRGenerator::visitGetNameCache(MGetNameCache* ins) { + MOZ_ASSERT(ins->envObj()->type() == MIRType::Object); + + // Emit an overrecursed check: this is necessary because the cache can + // attach a scripted getter stub that calls this script recursively. + gen->setNeedsOverrecursedCheck(); + + LGetNameCache* lir = + new (alloc()) LGetNameCache(useRegister(ins->envObj()), temp()); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallGetIntrinsicValue(MCallGetIntrinsicValue* ins) { + LCallGetIntrinsicValue* lir = new (alloc()) LCallGetIntrinsicValue(); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetPropSuperCache(MGetPropSuperCache* ins) { + MDefinition* obj = ins->object(); + MDefinition* receiver = ins->receiver(); + MDefinition* id = ins->idval(); + + gen->setNeedsOverrecursedCheck(); + + bool useConstId = + id->type() == MIRType::String || id->type() == MIRType::Symbol; + + auto* lir = new (alloc()) + LGetPropSuperCache(useRegister(obj), useBoxOrTyped(receiver), + useBoxOrTypedOrConstant(id, useConstId)); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetPropertyCache(MGetPropertyCache* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Object || + value->type() == MIRType::Value); + + MDefinition* id = ins->idval(); + MOZ_ASSERT(id->type() == MIRType::String || id->type() == MIRType::Symbol || + id->type() == MIRType::Int32 || id->type() == MIRType::Value); + + // Emit an overrecursed check: this is necessary because the cache can + // attach a scripted getter stub that calls this script recursively. + gen->setNeedsOverrecursedCheck(); + + // If this is a GetProp, the id is a constant string. Allow passing it as a + // constant to reduce register allocation pressure. + bool useConstId = + id->type() == MIRType::String || id->type() == MIRType::Symbol; + + auto* lir = new (alloc()) LGetPropertyCache( + useBoxOrTyped(value), useBoxOrTypedOrConstant(id, useConstId)); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBindNameCache(MBindNameCache* ins) { + MOZ_ASSERT(ins->environmentChain()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Object); + + LBindNameCache* lir = new (alloc()) + LBindNameCache(useRegister(ins->environmentChain()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallBindVar(MCallBindVar* ins) { + MOZ_ASSERT(ins->environmentChain()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Object); + + LCallBindVar* lir = + new (alloc()) LCallBindVar(useRegister(ins->environmentChain())); + define(lir, ins); +} + +void LIRGenerator::visitGuardObjectIdentity(MGuardObjectIdentity* ins) { + LGuardObjectIdentity* guard = new (alloc()) LGuardObjectIdentity( + useRegister(ins->object()), useRegister(ins->expected())); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardSpecificFunction(MGuardSpecificFunction* ins) { + auto* guard = new (alloc()) LGuardSpecificFunction( + useRegister(ins->function()), useRegister(ins->expected())); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, ins->function()); +} + +void LIRGenerator::visitGuardSpecificAtom(MGuardSpecificAtom* ins) { + auto* guard = + new (alloc()) LGuardSpecificAtom(useRegister(ins->str()), temp()); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, ins->str()); + assignSafepoint(guard, ins); +} + +void LIRGenerator::visitGuardSpecificSymbol(MGuardSpecificSymbol* ins) { + auto* guard = new (alloc()) LGuardSpecificSymbol(useRegister(ins->symbol())); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, ins->symbol()); +} + +void LIRGenerator::visitGuardStringToIndex(MGuardStringToIndex* ins) { + MOZ_ASSERT(ins->string()->type() == MIRType::String); + auto* guard = new (alloc()) LGuardStringToIndex(useRegister(ins->string())); + assignSnapshot(guard, ins->bailoutKind()); + define(guard, ins); + assignSafepoint(guard, ins); +} + +void LIRGenerator::visitGuardStringToInt32(MGuardStringToInt32* ins) { + MOZ_ASSERT(ins->string()->type() == MIRType::String); + auto* guard = + new (alloc()) LGuardStringToInt32(useRegister(ins->string()), temp()); + assignSnapshot(guard, ins->bailoutKind()); + define(guard, ins); + assignSafepoint(guard, ins); +} + +void LIRGenerator::visitGuardStringToDouble(MGuardStringToDouble* ins) { + MOZ_ASSERT(ins->string()->type() == MIRType::String); + auto* guard = new (alloc()) + LGuardStringToDouble(useRegister(ins->string()), temp(), temp()); + assignSnapshot(guard, ins->bailoutKind()); + define(guard, ins); + assignSafepoint(guard, ins); +} + +void LIRGenerator::visitGuardNoDenseElements(MGuardNoDenseElements* ins) { + auto* guard = + new (alloc()) LGuardNoDenseElements(useRegister(ins->object()), temp()); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardShape(MGuardShape* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + if (JitOptions.spectreObjectMitigationsMisc) { + auto* lir = + new (alloc()) LGuardShape(useRegisterAtStart(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineReuseInput(lir, ins, 0); + } else { + auto* lir = new (alloc()) + LGuardShape(useRegister(ins->object()), LDefinition::BogusTemp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); + } +} + +void LIRGenerator::visitGuardProto(MGuardProto* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->expected()->type() == MIRType::Object); + + auto* lir = new (alloc()) LGuardProto(useRegister(ins->object()), + useRegister(ins->expected()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardNullProto(MGuardNullProto* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = new (alloc()) LGuardNullProto(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardIsNativeObject(MGuardIsNativeObject* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = + new (alloc()) LGuardIsNativeObject(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardIsProxy(MGuardIsProxy* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = new (alloc()) LGuardIsProxy(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardIsNotProxy(MGuardIsNotProxy* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = + new (alloc()) LGuardIsNotProxy(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardIsNotDOMProxy(MGuardIsNotDOMProxy* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + + auto* lir = + new (alloc()) LGuardIsNotDOMProxy(useRegister(ins->proxy()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->proxy()); +} + +void LIRGenerator::visitProxyGet(MProxyGet* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + auto* lir = new (alloc()) + LProxyGet(useRegisterAtStart(ins->proxy()), tempFixed(CallTempReg0)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitProxyGetByValue(MProxyGetByValue* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + MOZ_ASSERT(ins->idVal()->type() == MIRType::Value); + auto* lir = new (alloc()) LProxyGetByValue(useRegisterAtStart(ins->proxy()), + useBoxAtStart(ins->idVal())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitProxyHasProp(MProxyHasProp* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + MOZ_ASSERT(ins->idVal()->type() == MIRType::Value); + auto* lir = new (alloc()) LProxyHasProp(useRegisterAtStart(ins->proxy()), + useBoxAtStart(ins->idVal())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitProxySet(MProxySet* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + MOZ_ASSERT(ins->rhs()->type() == MIRType::Value); + auto* lir = new (alloc()) + LProxySet(useRegisterAtStart(ins->proxy()), useBoxAtStart(ins->rhs()), + tempFixed(CallTempReg0)); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitProxySetByValue(MProxySetByValue* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + MOZ_ASSERT(ins->idVal()->type() == MIRType::Value); + MOZ_ASSERT(ins->rhs()->type() == MIRType::Value); + auto* lir = new (alloc()) + LProxySetByValue(useRegisterAtStart(ins->proxy()), + useBoxAtStart(ins->idVal()), useBoxAtStart(ins->rhs())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallSetArrayLength(MCallSetArrayLength* ins) { + MOZ_ASSERT(ins->obj()->type() == MIRType::Object); + MOZ_ASSERT(ins->rhs()->type() == MIRType::Value); + auto* lir = new (alloc()) LCallSetArrayLength(useRegisterAtStart(ins->obj()), + useBoxAtStart(ins->rhs())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitMegamorphicLoadSlot(MMegamorphicLoadSlot* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + auto* lir = new (alloc()) LMegamorphicLoadSlot( + useRegisterAtStart(ins->object()), tempFixed(CallTempReg0), + tempFixed(CallTempReg1), tempFixed(CallTempReg2)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); +} + +void LIRGenerator::visitMegamorphicLoadSlotByValue( + MMegamorphicLoadSlotByValue* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->idVal()->type() == MIRType::Value); + auto* lir = new (alloc()) LMegamorphicLoadSlotByValue( + useRegisterAtStart(ins->object()), useBoxAtStart(ins->idVal()), + tempFixed(CallTempReg0), tempFixed(CallTempReg1)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); +} + +void LIRGenerator::visitMegamorphicStoreSlot(MMegamorphicStoreSlot* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->rhs()->type() == MIRType::Value); + auto* lir = new (alloc()) + LMegamorphicStoreSlot(useRegisterAtStart(ins->object()), + useBoxAtStart(ins->rhs()), tempFixed(CallTempReg0), + tempFixed(CallTempReg1), tempFixed(CallTempReg2)); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); +} + +void LIRGenerator::visitMegamorphicHasProp(MMegamorphicHasProp* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->idVal()->type() == MIRType::Value); + auto* lir = new (alloc()) LMegamorphicHasProp( + useRegisterAtStart(ins->object()), useBoxAtStart(ins->idVal()), + tempFixed(CallTempReg0), tempFixed(CallTempReg1)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); +} + +void LIRGenerator::visitGuardIsNotArrayBufferMaybeShared( + MGuardIsNotArrayBufferMaybeShared* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = new (alloc()) + LGuardIsNotArrayBufferMaybeShared(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitGuardIsTypedArray(MGuardIsTypedArray* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + auto* lir = + new (alloc()) LGuardIsTypedArray(useRegister(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); +} + +void LIRGenerator::visitNurseryObject(MNurseryObject* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto* lir = new (alloc()) LNurseryObject(); + define(lir, ins); +} + +void LIRGenerator::visitGuardObjectGroup(MGuardObjectGroup* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + if (JitOptions.spectreObjectMitigationsMisc) { + auto* lir = new (alloc()) + LGuardObjectGroup(useRegisterAtStart(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineReuseInput(lir, ins, 0); + } else { + auto* lir = new (alloc()) + LGuardObjectGroup(useRegister(ins->object()), LDefinition::BogusTemp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->object()); + } +} + +void LIRGenerator::visitGuardValue(MGuardValue* ins) { + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + auto* lir = new (alloc()) LGuardValue(useBox(ins->value())); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->value()); +} + +void LIRGenerator::visitGuardNotOptimizedArguments( + MGuardNotOptimizedArguments* ins) { + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + auto* lir = new (alloc()) LGuardNotOptimizedArguments(useBox(ins->value())); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->value()); +} + +void LIRGenerator::visitGuardNullOrUndefined(MGuardNullOrUndefined* ins) { + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + auto* lir = new (alloc()) LGuardNullOrUndefined(useBox(ins->value())); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->value()); +} + +void LIRGenerator::visitGuardFunctionFlags(MGuardFunctionFlags* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = new (alloc()) LGuardFunctionFlags(useRegister(ins->function())); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->function()); +} + +void LIRGenerator::visitGuardFunctionIsNonBuiltinCtor( + MGuardFunctionIsNonBuiltinCtor* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = new (alloc()) + LGuardFunctionIsNonBuiltinCtor(useRegister(ins->function()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->function()); +} + +void LIRGenerator::visitGuardFunctionKind(MGuardFunctionKind* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = + new (alloc()) LGuardFunctionKind(useRegister(ins->function()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->function()); +} + +void LIRGenerator::visitGuardFunctionScript(MGuardFunctionScript* ins) { + MOZ_ASSERT(ins->function()->type() == MIRType::Object); + + auto* lir = new (alloc()) LGuardFunctionScript(useRegister(ins->function())); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->function()); +} + +void LIRGenerator::visitAssertRange(MAssertRange* ins) { + MDefinition* input = ins->input(); + LInstruction* lir = nullptr; + + switch (input->type()) { + case MIRType::Boolean: + case MIRType::Int32: + lir = new (alloc()) LAssertRangeI(useRegisterAtStart(input)); + break; + + case MIRType::Double: + lir = new (alloc()) LAssertRangeD(useRegister(input), tempDouble()); + break; + + case MIRType::Float32: + lir = new (alloc()) + LAssertRangeF(useRegister(input), tempDouble(), tempDouble()); + break; + + case MIRType::Value: + lir = new (alloc()) LAssertRangeV(useBox(input), tempToUnbox(), + tempDouble(), tempDouble()); + break; + + default: + MOZ_CRASH("Unexpected Range for MIRType"); + break; + } + + lir->setMir(ins); + add(lir); +} + +void LIRGenerator::visitAssertClass(MAssertClass* ins) { + auto* lir = + new (alloc()) LAssertClass(useRegisterAtStart(ins->input()), temp()); + add(lir, ins); +} + +void LIRGenerator::visitAssertShape(MAssertShape* ins) { + auto* lir = new (alloc()) LAssertShape(useRegisterAtStart(ins->input())); + add(lir, ins); +} + +void LIRGenerator::visitDeleteProperty(MDeleteProperty* ins) { + LCallDeleteProperty* lir = + new (alloc()) LCallDeleteProperty(useBoxAtStart(ins->value())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitDeleteElement(MDeleteElement* ins) { + LCallDeleteElement* lir = new (alloc()) LCallDeleteElement( + useBoxAtStart(ins->value()), useBoxAtStart(ins->index())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitSetPropertyCache(MSetPropertyCache* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + + MDefinition* id = ins->idval(); + MOZ_ASSERT(id->type() == MIRType::String || id->type() == MIRType::Symbol || + id->type() == MIRType::Int32 || id->type() == MIRType::Value); + + // If this is a SetProp, the id is a constant string. Allow passing it as a + // constant to reduce register allocation pressure. + bool useConstId = + id->type() == MIRType::String || id->type() == MIRType::Symbol; + bool useConstValue = IsNonNurseryConstant(ins->value()); + + // Emit an overrecursed check: this is necessary because the cache can + // attach a scripted setter stub that calls this script recursively. + gen->setNeedsOverrecursedCheck(); + + // We need a double temp register for TypedArray or TypedObject stubs. + LDefinition tempD = tempFixed(FloatReg0); + + LInstruction* lir = new (alloc()) LSetPropertyCache( + useRegister(ins->object()), useBoxOrTypedOrConstant(id, useConstId), + useBoxOrTypedOrConstant(ins->value(), useConstValue), temp(), tempD); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallSetElement(MCallSetElement* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->index()->type() == MIRType::Value); + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + + LCallSetElement* lir = new (alloc()) + LCallSetElement(useRegisterAtStart(ins->object()), + useBoxAtStart(ins->index()), useBoxAtStart(ins->value())); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetIteratorCache(MGetIteratorCache* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Object || + value->type() == MIRType::Value); + + LGetIteratorCache* lir = + new (alloc()) LGetIteratorCache(useBoxOrTyped(value), temp(), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitOptimizeSpreadCallCache(MOptimizeSpreadCallCache* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Value); + + auto* lir = new (alloc()) LOptimizeSpreadCallCache(useBox(value), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitIteratorMore(MIteratorMore* ins) { + LIteratorMore* lir = + new (alloc()) LIteratorMore(useRegister(ins->iterator()), temp()); + defineBox(lir, ins); +} + +void LIRGenerator::visitIsNoIter(MIsNoIter* ins) { + MOZ_ASSERT(ins->hasOneUse()); + emitAtUses(ins); +} + +void LIRGenerator::visitIteratorEnd(MIteratorEnd* ins) { + LIteratorEnd* lir = new (alloc()) + LIteratorEnd(useRegister(ins->iterator()), temp(), temp(), temp()); + add(lir, ins); +} + +void LIRGenerator::visitStringLength(MStringLength* ins) { + MOZ_ASSERT(ins->string()->type() == MIRType::String); + define(new (alloc()) LStringLength(useRegisterAtStart(ins->string())), ins); +} + +void LIRGenerator::visitArgumentsLength(MArgumentsLength* ins) { + define(new (alloc()) LArgumentsLength(), ins); +} + +void LIRGenerator::visitGetFrameArgument(MGetFrameArgument* ins) { + LGetFrameArgument* lir = + new (alloc()) LGetFrameArgument(useRegisterOrConstant(ins->index())); + defineBox(lir, ins); +} + +void LIRGenerator::visitNewTarget(MNewTarget* ins) { + LNewTarget* lir = new (alloc()) LNewTarget(); + defineBox(lir, ins); +} + +void LIRGenerator::visitRest(MRest* ins) { + MOZ_ASSERT(ins->numActuals()->type() == MIRType::Int32); + + LRest* lir = new (alloc()) + LRest(useRegisterAtStart(ins->numActuals()), tempFixed(CallTempReg0), + tempFixed(CallTempReg1), tempFixed(CallTempReg2)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitThrow(MThrow* ins) { + MDefinition* value = ins->getOperand(0); + MOZ_ASSERT(value->type() == MIRType::Value); + + LThrow* lir = new (alloc()) LThrow(useBoxAtStart(value)); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitInCache(MInCache* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + + MOZ_ASSERT(lhs->type() == MIRType::String || lhs->type() == MIRType::Symbol || + lhs->type() == MIRType::Int32 || lhs->type() == MIRType::Value); + MOZ_ASSERT(rhs->type() == MIRType::Object); + + LInCache* lir = + new (alloc()) LInCache(useBoxOrTyped(lhs), useRegister(rhs), temp()); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitHasOwnCache(MHasOwnCache* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Object || + value->type() == MIRType::Value); + + MDefinition* id = ins->idval(); + MOZ_ASSERT(id->type() == MIRType::String || id->type() == MIRType::Symbol || + id->type() == MIRType::Int32 || id->type() == MIRType::Value); + + // Emit an overrecursed check: this is necessary because the cache can + // attach a scripted getter stub that calls this script recursively. + gen->setNeedsOverrecursedCheck(); + + LHasOwnCache* lir = + new (alloc()) LHasOwnCache(useBoxOrTyped(value), useBoxOrTyped(id)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckPrivateFieldCache(MCheckPrivateFieldCache* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Object || + value->type() == MIRType::Value); + + MDefinition* id = ins->idval(); + MOZ_ASSERT(id->type() == MIRType::String || id->type() == MIRType::Symbol || + id->type() == MIRType::Int32 || id->type() == MIRType::Value); + + LCheckPrivateFieldCache* lir = new (alloc()) + LCheckPrivateFieldCache(useBoxOrTyped(value), useBoxOrTyped(id)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitInstanceOf(MInstanceOf* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + + MOZ_ASSERT(lhs->type() == MIRType::Value || lhs->type() == MIRType::Object); + MOZ_ASSERT(rhs->type() == MIRType::Object); + + if (lhs->type() == MIRType::Object) { + auto* lir = new (alloc()) + LInstanceOfO(useRegister(lhs), useRegisterOrConstant(rhs)); + define(lir, ins); + assignSafepoint(lir, ins); + } else { + auto* lir = + new (alloc()) LInstanceOfV(useBox(lhs), useRegisterOrConstant(rhs)); + define(lir, ins); + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitInstanceOfCache(MInstanceOfCache* ins) { + MDefinition* lhs = ins->lhs(); + MDefinition* rhs = ins->rhs(); + + MOZ_ASSERT(lhs->type() == MIRType::Value); + MOZ_ASSERT(rhs->type() == MIRType::Object); + + LInstanceOfCache* lir = + new (alloc()) LInstanceOfCache(useBox(lhs), useRegister(rhs)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitIsArray(MIsArray* ins) { + MOZ_ASSERT(ins->type() == MIRType::Boolean); + + if (ins->value()->type() == MIRType::Object) { + LIsArrayO* lir = new (alloc()) LIsArrayO(useRegister(ins->value())); + define(lir, ins); + assignSafepoint(lir, ins); + } else { + MOZ_ASSERT(ins->value()->type() == MIRType::Value); + LIsArrayV* lir = new (alloc()) LIsArrayV(useBox(ins->value()), temp()); + define(lir, ins); + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitIsTypedArray(MIsTypedArray* ins) { + MOZ_ASSERT(ins->value()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + + auto* lir = new (alloc()) LIsTypedArray(useRegister(ins->value())); + define(lir, ins); + + if (ins->isPossiblyWrapped()) { + assignSafepoint(lir, ins); + } +} + +void LIRGenerator::visitIsCallable(MIsCallable* ins) { + MOZ_ASSERT(ins->type() == MIRType::Boolean); + + if (ins->object()->type() == MIRType::Object) { + define(new (alloc()) LIsCallableO(useRegister(ins->object())), ins); + } else { + MOZ_ASSERT(ins->object()->type() == MIRType::Value); + define(new (alloc()) LIsCallableV(useBox(ins->object()), temp()), ins); + } +} + +void LIRGenerator::visitIsConstructor(MIsConstructor* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + define(new (alloc()) LIsConstructor(useRegister(ins->object())), ins); +} + +void LIRGenerator::visitIsCrossRealmArrayConstructor( + MIsCrossRealmArrayConstructor* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + define(new (alloc()) + LIsCrossRealmArrayConstructor(useRegister(ins->object())), + ins); +} + +static bool CanEmitIsObjectOrIsNullOrUndefinedAtUses(MInstruction* ins) { + if (!ins->canEmitAtUses()) { + return false; + } + + MUseIterator iter(ins->usesBegin()); + if (iter == ins->usesEnd()) { + return false; + } + + MNode* node = iter->consumer(); + if (!node->isDefinition()) { + return false; + } + + if (!node->toDefinition()->isTest()) { + return false; + } + + iter++; + return iter == ins->usesEnd(); +} + +void LIRGenerator::visitIsObject(MIsObject* ins) { + if (CanEmitIsObjectOrIsNullOrUndefinedAtUses(ins)) { + emitAtUses(ins); + return; + } + + MDefinition* opd = ins->input(); + MOZ_ASSERT(opd->type() == MIRType::Value); + LIsObject* lir = new (alloc()) LIsObject(useBoxAtStart(opd)); + define(lir, ins); +} + +void LIRGenerator::visitIsNullOrUndefined(MIsNullOrUndefined* ins) { + if (CanEmitIsObjectOrIsNullOrUndefinedAtUses(ins)) { + emitAtUses(ins); + return; + } + + MDefinition* opd = ins->input(); + MOZ_ASSERT(opd->type() == MIRType::Value); + LIsNullOrUndefined* lir = + new (alloc()) LIsNullOrUndefined(useBoxAtStart(opd)); + define(lir, ins); +} + +void LIRGenerator::visitHasClass(MHasClass* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + define(new (alloc()) LHasClass(useRegister(ins->object())), ins); +} + +void LIRGenerator::visitGuardToClass(MGuardToClass* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Object); + LGuardToClass* lir = + new (alloc()) LGuardToClass(useRegisterAtStart(ins->object()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + defineReuseInput(lir, ins, 0); +} + +void LIRGenerator::visitObjectClassToString(MObjectClassToString* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::String); + auto* lir = new (alloc()) LObjectClassToString( + useRegisterAtStart(ins->object()), tempFixed(CallTempReg0)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); +} + +void LIRGenerator::visitWasmAddOffset(MWasmAddOffset* ins) { + MOZ_ASSERT(ins->base()->type() == MIRType::Int32); + MOZ_ASSERT(ins->type() == MIRType::Int32); + MOZ_ASSERT(ins->offset()); + define(new (alloc()) LWasmAddOffset(useRegisterAtStart(ins->base())), ins); +} + +void LIRGenerator::visitWasmLoadTls(MWasmLoadTls* ins) { + auto* lir = new (alloc()) LWasmLoadTls(useRegisterAtStart(ins->tlsPtr())); + define(lir, ins); +} + +void LIRGenerator::visitWasmBoundsCheck(MWasmBoundsCheck* ins) { + MOZ_ASSERT(!ins->isRedundant()); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + MDefinition* boundsCheckLimit = ins->boundsCheckLimit(); + MOZ_ASSERT(boundsCheckLimit->type() == MIRType::Int32); + + if (JitOptions.spectreIndexMasking) { + auto* lir = new (alloc()) LWasmBoundsCheck(useRegisterAtStart(index), + useRegister(boundsCheckLimit)); + defineReuseInput(lir, ins, 0); + } else { + auto* lir = new (alloc()) LWasmBoundsCheck( + useRegisterAtStart(index), useRegisterAtStart(boundsCheckLimit)); + add(lir, ins); + } +} + +void LIRGenerator::visitWasmAlignmentCheck(MWasmAlignmentCheck* ins) { + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) LWasmAlignmentCheck(useRegisterAtStart(index)); + add(lir, ins); +} + +void LIRGenerator::visitWasmLoadGlobalVar(MWasmLoadGlobalVar* ins) { + size_t offs = offsetof(wasm::TlsData, globalArea) + ins->globalDataOffset(); + if (ins->type() == MIRType::Int64) { +#ifdef JS_PUNBOX64 + LAllocation tlsPtr = useRegisterAtStart(ins->tlsPtr()); +#else + LAllocation tlsPtr = useRegister(ins->tlsPtr()); +#endif + defineInt64(new (alloc()) LWasmLoadSlotI64(tlsPtr, offs), ins); + } else { + LAllocation tlsPtr = useRegisterAtStart(ins->tlsPtr()); + define(new (alloc()) LWasmLoadSlot(tlsPtr, offs, ins->type()), ins); + } +} + +void LIRGenerator::visitWasmLoadGlobalCell(MWasmLoadGlobalCell* ins) { + if (ins->type() == MIRType::Int64) { +#ifdef JS_PUNBOX64 + LAllocation cellPtr = useRegisterAtStart(ins->cellPtr()); +#else + LAllocation cellPtr = useRegister(ins->cellPtr()); +#endif + defineInt64(new (alloc()) LWasmLoadSlotI64(cellPtr, /*offs=*/0), ins); + } else { + LAllocation cellPtr = useRegisterAtStart(ins->cellPtr()); + define(new (alloc()) LWasmLoadSlot(cellPtr, /*offs=*/0, ins->type()), ins); + } +} + +void LIRGenerator::visitWasmStoreGlobalVar(MWasmStoreGlobalVar* ins) { + MDefinition* value = ins->value(); + size_t offs = offsetof(wasm::TlsData, globalArea) + ins->globalDataOffset(); + if (value->type() == MIRType::Int64) { +#ifdef JS_PUNBOX64 + LAllocation tlsPtr = useRegisterAtStart(ins->tlsPtr()); + LInt64Allocation valueAlloc = useInt64RegisterAtStart(value); +#else + LAllocation tlsPtr = useRegister(ins->tlsPtr()); + LInt64Allocation valueAlloc = useInt64Register(value); +#endif + add(new (alloc()) LWasmStoreSlotI64(valueAlloc, tlsPtr, offs), ins); + } else { + MOZ_ASSERT(value->type() != MIRType::RefOrNull); + LAllocation tlsPtr = useRegisterAtStart(ins->tlsPtr()); + LAllocation valueAlloc = useRegisterAtStart(value); + add(new (alloc()) LWasmStoreSlot(valueAlloc, tlsPtr, offs, value->type()), + ins); + } +} + +void LIRGenerator::visitWasmStoreGlobalCell(MWasmStoreGlobalCell* ins) { + MDefinition* value = ins->value(); + size_t offs = 0; + if (value->type() == MIRType::Int64) { +#ifdef JS_PUNBOX64 + LAllocation cellPtr = useRegisterAtStart(ins->cellPtr()); + LInt64Allocation valueAlloc = useInt64RegisterAtStart(value); +#else + LAllocation cellPtr = useRegister(ins->cellPtr()); + LInt64Allocation valueAlloc = useInt64Register(value); +#endif + add(new (alloc()) LWasmStoreSlotI64(valueAlloc, cellPtr, offs)); + } else { + MOZ_ASSERT(value->type() != MIRType::RefOrNull); + LAllocation cellPtr = useRegisterAtStart(ins->cellPtr()); + LAllocation valueAlloc = useRegisterAtStart(value); + add(new (alloc()) LWasmStoreSlot(valueAlloc, cellPtr, offs, value->type())); + } +} + +void LIRGenerator::visitWasmStoreStackResult(MWasmStoreStackResult* ins) { + MDefinition* stackResultArea = ins->stackResultArea(); + MDefinition* value = ins->value(); + size_t offs = ins->offset(); + LInstruction* lir; + if (value->type() == MIRType::Int64) { + lir = new (alloc()) LWasmStoreSlotI64(useInt64Register(value), + useRegister(stackResultArea), offs); + } else { + MOZ_ASSERT(value->type() != MIRType::RefOrNull); + lir = new (alloc()) LWasmStoreSlot( + useRegister(value), useRegister(stackResultArea), offs, value->type()); + } + add(lir, ins); +} + +void LIRGenerator::visitWasmDerivedPointer(MWasmDerivedPointer* ins) { + LAllocation base = useRegisterAtStart(ins->base()); + define(new (alloc()) LWasmDerivedPointer(base), ins); +} + +void LIRGenerator::visitWasmStoreRef(MWasmStoreRef* ins) { + LAllocation tls = useRegister(ins->tls()); + LAllocation valueAddr = useFixed(ins->valueAddr(), PreBarrierReg); + LAllocation value = useRegister(ins->value()); + add(new (alloc()) LWasmStoreRef(tls, valueAddr, value, temp()), ins); +} + +void LIRGenerator::visitWasmParameter(MWasmParameter* ins) { + ABIArg abi = ins->abi(); + if (ins->type() == MIRType::StackResults) { + // Functions that return stack results receive an extra incoming parameter + // with type MIRType::StackResults. This value is a pointer to fresh + // memory. Here we treat it as if it were in fact MIRType::Pointer. + auto* lir = new (alloc()) LWasmParameter; + LDefinition def(LDefinition::TypeFrom(MIRType::Pointer), + LDefinition::FIXED); + def.setOutput(abi.argInRegister() ? LAllocation(abi.reg()) + : LArgument(abi.offsetFromArgBase())); + define(lir, ins, def); + return; + } + if (abi.argInRegister()) { +#if defined(JS_NUNBOX32) + if (abi.isGeneralRegPair()) { + defineInt64Fixed( + new (alloc()) LWasmParameterI64, ins, + LInt64Allocation(LAllocation(AnyRegister(abi.gpr64().high)), + LAllocation(AnyRegister(abi.gpr64().low)))); + return; + } +#endif + defineFixed(new (alloc()) LWasmParameter, ins, LAllocation(abi.reg())); + return; + } + if (ins->type() == MIRType::Int64) { + MOZ_ASSERT(!abi.argInRegister()); + defineInt64Fixed( + new (alloc()) LWasmParameterI64, ins, +#if defined(JS_NUNBOX32) + LInt64Allocation(LArgument(abi.offsetFromArgBase() + INT64HIGH_OFFSET), + LArgument(abi.offsetFromArgBase() + INT64LOW_OFFSET)) +#else + LInt64Allocation(LArgument(abi.offsetFromArgBase())) +#endif + ); + } else { + MOZ_ASSERT(IsNumberType(ins->type()) || ins->type() == MIRType::RefOrNull +#ifdef ENABLE_WASM_SIMD + || ins->type() == MIRType::Simd128 +#endif + ); + defineFixed(new (alloc()) LWasmParameter, ins, + LArgument(abi.offsetFromArgBase())); + } +} + +void LIRGenerator::visitWasmReturn(MWasmReturn* ins) { + MDefinition* rval = ins->getOperand(0); + MDefinition* tlsParam = ins->getOperand(1); + + if (rval->type() == MIRType::Int64) { + add(new (alloc()) LWasmReturnI64(useInt64Fixed(rval, ReturnReg64), + useFixed(tlsParam, WasmTlsReg))); + return; + } + + LWasmReturn* lir = new (alloc()) LWasmReturn; + if (rval->type() == MIRType::Float32) { + lir->setOperand(0, useFixed(rval, ReturnFloat32Reg)); + } else if (rval->type() == MIRType::Double) { + lir->setOperand(0, useFixed(rval, ReturnDoubleReg)); +#ifdef ENABLE_WASM_SIMD + } else if (rval->type() == MIRType::Simd128) { + lir->setOperand(0, useFixed(rval, ReturnSimd128Reg)); +#endif + } else if (rval->type() == MIRType::Int32 || + rval->type() == MIRType::RefOrNull) { + lir->setOperand(0, useFixed(rval, ReturnReg)); + } else { + MOZ_CRASH("Unexpected wasm return type"); + } + + lir->setOperand(1, useFixed(tlsParam, WasmTlsReg)); + + add(lir); +} + +void LIRGenerator::visitWasmReturnVoid(MWasmReturnVoid* ins) { + MDefinition* tlsParam = ins->getOperand(0); + LWasmReturnVoid* lir = new (alloc()) LWasmReturnVoid; + lir->setOperand(0, useFixed(tlsParam, WasmTlsReg)); + add(lir); +} + +void LIRGenerator::visitWasmStackArg(MWasmStackArg* ins) { + if (ins->arg()->type() == MIRType::Int64) { + add(new (alloc()) + LWasmStackArgI64(useInt64RegisterOrConstantAtStart(ins->arg())), + ins); + } else if (IsFloatingPointType(ins->arg()->type())) { + MOZ_ASSERT(!ins->arg()->isEmittedAtUses()); + add(new (alloc()) LWasmStackArg(useRegisterAtStart(ins->arg())), ins); + } else { + add(new (alloc()) LWasmStackArg(useRegisterOrConstantAtStart(ins->arg())), + ins); + } +} + +void LIRGenerator::visitWasmRegisterResult(MWasmRegisterResult* ins) { + auto* lir = new (alloc()) LWasmRegisterResult(); + uint32_t vreg = getVirtualRegister(); + MOZ_ASSERT(ins->type() != MIRType::Int64); + auto type = LDefinition::TypeFrom(ins->type()); + lir->setDef(0, LDefinition(vreg, type, LGeneralReg(ins->loc()))); + ins->setVirtualRegister(vreg); + add(lir, ins); +} + +void LIRGenerator::visitWasmFloatRegisterResult(MWasmFloatRegisterResult* ins) { + auto* lir = new (alloc()) LWasmRegisterResult(); + uint32_t vreg = getVirtualRegister(); + auto type = LDefinition::TypeFrom(ins->type()); + lir->setDef(0, LDefinition(vreg, type, LFloatReg(ins->loc()))); + ins->setVirtualRegister(vreg); + add(lir, ins); +} + +void LIRGenerator::visitWasmRegister64Result(MWasmRegister64Result* ins) { + MOZ_ASSERT(ins->type() == MIRType::Int64); + uint32_t vreg = getVirtualRegister(); + +#if defined(JS_NUNBOX32) + auto* lir = new (alloc()) LWasmRegisterPairResult(); + lir->setDef(INT64LOW_INDEX, + LDefinition(vreg + INT64LOW_INDEX, LDefinition::GENERAL, + LGeneralReg(ins->loc().low))); + lir->setDef(INT64HIGH_INDEX, + LDefinition(vreg + INT64HIGH_INDEX, LDefinition::GENERAL, + LGeneralReg(ins->loc().high))); + getVirtualRegister(); +#elif defined(JS_PUNBOX64) + auto* lir = new (alloc()) LWasmRegisterResult(); + lir->setDef( + 0, LDefinition(vreg, LDefinition::GENERAL, LGeneralReg(ins->loc().reg))); +#else +# error expected either JS_NUNBOX32 or JS_PUNBOX64 +#endif + + ins->setVirtualRegister(vreg); + add(lir, ins); +} + +void LIRGenerator::visitWasmStackResultArea(MWasmStackResultArea* ins) { + MOZ_ASSERT(ins->type() == MIRType::StackResults); + auto* lir = new (alloc()) LWasmStackResultArea(temp()); + uint32_t vreg = getVirtualRegister(); + lir->setDef(0, + LDefinition(vreg, LDefinition::STACKRESULTS, LDefinition::STACK)); + ins->setVirtualRegister(vreg); + add(lir, ins); +} + +void LIRGenerator::visitWasmStackResult(MWasmStackResult* ins) { + MWasmStackResultArea* area = ins->resultArea()->toWasmStackResultArea(); + LDefinition::Policy pol = LDefinition::STACK; + + if (ins->type() == MIRType::Int64) { + auto* lir = new (alloc()) LWasmStackResult64; + lir->setOperand(0, use(area, LUse(LUse::STACK, /* usedAtStart = */ true))); + uint32_t vreg = getVirtualRegister(); + LDefinition::Type typ = LDefinition::GENERAL; +#if defined(JS_NUNBOX32) + getVirtualRegister(); + lir->setDef(INT64LOW_INDEX, LDefinition(vreg + INT64LOW_INDEX, typ, pol)); + lir->setDef(INT64HIGH_INDEX, LDefinition(vreg + INT64HIGH_INDEX, typ, pol)); +#else + lir->setDef(0, LDefinition(vreg, typ, pol)); +#endif + ins->setVirtualRegister(vreg); + add(lir, ins); + return; + } + + auto* lir = new (alloc()) LWasmStackResult; + lir->setOperand(0, use(area, LUse(LUse::STACK, /* usedAtStart = */ true))); + uint32_t vreg = getVirtualRegister(); + LDefinition::Type typ = LDefinition::TypeFrom(ins->type()); + lir->setDef(0, LDefinition(vreg, typ, pol)); + ins->setVirtualRegister(vreg); + add(lir, ins); +} + +void LIRGenerator::visitWasmCall(MWasmCall* ins) { + bool needsBoundsCheck = true; + if (ins->callee().isTable()) { + MDefinition* index = ins->getOperand(ins->numArgs()); + + if (ins->callee().which() == wasm::CalleeDesc::WasmTable && + index->isConstant()) { + if (uint32_t(index->toConstant()->toInt32()) < + ins->callee().wasmTableMinLength()) { + needsBoundsCheck = false; + } + } + } + + auto* lir = allocateVariadic<LWasmCall>(ins->numOperands(), needsBoundsCheck); + if (!lir) { + abort(AbortReason::Alloc, "OOM: LIRGenerator::lowerWasmCall"); + return; + } + + for (unsigned i = 0; i < ins->numArgs(); i++) { + lir->setOperand( + i, useFixedAtStart(ins->getOperand(i), ins->registerForArg(i))); + } + + if (ins->callee().isTable()) { + MDefinition* index = ins->getOperand(ins->numArgs()); + lir->setOperand(ins->numArgs(), + useFixedAtStart(index, WasmTableCallIndexReg)); + } + + add(lir, ins); + + assignWasmSafepoint(lir, ins); +} + +void LIRGenerator::visitSetDOMProperty(MSetDOMProperty* ins) { + MDefinition* val = ins->value(); + + Register cxReg, objReg, privReg, valueReg; + GetTempRegForIntArg(0, 0, &cxReg); + GetTempRegForIntArg(1, 0, &objReg); + GetTempRegForIntArg(2, 0, &privReg); + GetTempRegForIntArg(3, 0, &valueReg); + + // Keep using GetTempRegForIntArg, since we want to make sure we + // don't clobber registers we're already using. + Register tempReg1, tempReg2; + GetTempRegForIntArg(4, 0, &tempReg1); + mozilla::DebugOnly<bool> ok = GetTempRegForIntArg(5, 0, &tempReg2); + MOZ_ASSERT(ok, "How can we not have six temp registers?"); + + LSetDOMProperty* lir = new (alloc()) + LSetDOMProperty(tempFixed(cxReg), useFixedAtStart(ins->object(), objReg), + useBoxFixedAtStart(val, tempReg1, tempReg2), + tempFixed(privReg), tempFixed(valueReg)); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetDOMProperty(MGetDOMProperty* ins) { + Register cxReg, objReg, privReg, valueReg; + GetTempRegForIntArg(0, 0, &cxReg); + GetTempRegForIntArg(1, 0, &objReg); + GetTempRegForIntArg(2, 0, &privReg); + mozilla::DebugOnly<bool> ok = GetTempRegForIntArg(3, 0, &valueReg); + MOZ_ASSERT(ok, "How can we not have four temp registers?"); + LGetDOMProperty* lir = new (alloc()) + LGetDOMProperty(tempFixed(cxReg), useFixedAtStart(ins->object(), objReg), + tempFixed(privReg), tempFixed(valueReg)); + + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGetDOMMember(MGetDOMMember* ins) { + MOZ_ASSERT(ins->isDomMovable(), "Members had better be movable"); + // We wish we could assert that ins->domAliasSet() == JSJitInfo::AliasNone, + // but some MGetDOMMembers are for [Pure], not [Constant] properties, whose + // value can in fact change as a result of DOM setters and method calls. + MOZ_ASSERT(ins->domAliasSet() != JSJitInfo::AliasEverything, + "Member gets had better not alias the world"); + + MDefinition* obj = ins->object(); + MOZ_ASSERT(obj->type() == MIRType::Object); + + MIRType type = ins->type(); + + if (type == MIRType::Value) { + LGetDOMMemberV* lir = new (alloc()) LGetDOMMemberV(useRegisterAtStart(obj)); + defineBox(lir, ins); + } else { + LGetDOMMemberT* lir = + new (alloc()) LGetDOMMemberT(useRegisterForTypedLoad(obj, type)); + define(lir, ins); + } +} + +void LIRGenerator::visitLoadDOMExpandoValue(MLoadDOMExpandoValue* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + auto* lir = + new (alloc()) LLoadDOMExpandoValue(useRegisterAtStart(ins->proxy())); + defineBox(lir, ins); +} + +void LIRGenerator::visitLoadDOMExpandoValueGuardGeneration( + MLoadDOMExpandoValueGuardGeneration* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + auto* lir = new (alloc()) + LLoadDOMExpandoValueGuardGeneration(useRegisterAtStart(ins->proxy())); + assignSnapshot(lir, ins->bailoutKind()); + defineBox(lir, ins); +} + +void LIRGenerator::visitLoadDOMExpandoValueIgnoreGeneration( + MLoadDOMExpandoValueIgnoreGeneration* ins) { + MOZ_ASSERT(ins->proxy()->type() == MIRType::Object); + auto* lir = new (alloc()) + LLoadDOMExpandoValueIgnoreGeneration(useRegisterAtStart(ins->proxy())); + defineBox(lir, ins); +} + +void LIRGenerator::visitGuardDOMExpandoMissingOrGuardShape( + MGuardDOMExpandoMissingOrGuardShape* ins) { + MOZ_ASSERT(ins->expando()->type() == MIRType::Value); + auto* lir = new (alloc()) + LGuardDOMExpandoMissingOrGuardShape(useBox(ins->expando()), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->expando()); +} + +void LIRGenerator::visitIncrementWarmUpCounter(MIncrementWarmUpCounter* ins) { + LIncrementWarmUpCounter* lir = new (alloc()) LIncrementWarmUpCounter(temp()); + add(lir, ins); +} + +void LIRGenerator::visitLexicalCheck(MLexicalCheck* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::Value); + LLexicalCheck* lir = new (alloc()) LLexicalCheck(useBox(input)); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, input); +} + +void LIRGenerator::visitThrowRuntimeLexicalError( + MThrowRuntimeLexicalError* ins) { + LThrowRuntimeLexicalError* lir = new (alloc()) LThrowRuntimeLexicalError(); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitThrowMsg(MThrowMsg* ins) { + LThrowMsg* lir = new (alloc()) LThrowMsg(); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGlobalDeclInstantiation(MGlobalDeclInstantiation* ins) { + LGlobalDeclInstantiation* lir = new (alloc()) LGlobalDeclInstantiation(); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitDebugger(MDebugger* ins) { + LDebugger* lir = + new (alloc()) LDebugger(tempFixed(CallTempReg0), tempFixed(CallTempReg1)); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); +} + +void LIRGenerator::visitAtomicIsLockFree(MAtomicIsLockFree* ins) { + define(new (alloc()) LAtomicIsLockFree(useRegister(ins->input())), ins); +} + +void LIRGenerator::visitCheckReturn(MCheckReturn* ins) { + MDefinition* retVal = ins->returnValue(); + MDefinition* thisVal = ins->thisValue(); + MOZ_ASSERT(retVal->type() == MIRType::Value); + MOZ_ASSERT(thisVal->type() == MIRType::Value); + + auto* lir = + new (alloc()) LCheckReturn(useBoxAtStart(retVal), useBoxAtStart(thisVal)); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckIsObj(MCheckIsObj* ins) { + MDefinition* input = ins->input(); + MOZ_ASSERT(input->type() == MIRType::Value); + + LCheckIsObj* lir = new (alloc()) LCheckIsObj(useBox(input)); + define(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckObjCoercible(MCheckObjCoercible* ins) { + MDefinition* checkVal = ins->checkValue(); + MOZ_ASSERT(checkVal->type() == MIRType::Value); + + auto* lir = new (alloc()) LCheckObjCoercible(useBoxAtStart(checkVal)); + redefine(ins, checkVal); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckClassHeritage(MCheckClassHeritage* ins) { + MDefinition* heritage = ins->heritage(); + MOZ_ASSERT(heritage->type() == MIRType::Value); + + auto* lir = new (alloc()) LCheckClassHeritage(useBox(heritage), temp()); + redefine(ins, heritage); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckThis(MCheckThis* ins) { + MDefinition* thisValue = ins->thisValue(); + MOZ_ASSERT(thisValue->type() == MIRType::Value); + + auto* lir = new (alloc()) LCheckThis(useBoxAtStart(thisValue)); + redefine(ins, thisValue); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCheckThisReinit(MCheckThisReinit* ins) { + MDefinition* thisValue = ins->thisValue(); + MOZ_ASSERT(thisValue->type() == MIRType::Value); + + auto* lir = new (alloc()) LCheckThisReinit(useBoxAtStart(thisValue)); + redefine(ins, thisValue); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitGenerator(MGenerator* ins) { + auto* lir = + new (alloc()) LGenerator(useRegisterAtStart(ins->callee()), + useRegisterAtStart(ins->environmentChain()), + useRegisterAtStart(ins->argsObject())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitAsyncResolve(MAsyncResolve* ins) { + auto* lir = new (alloc()) LAsyncResolve(useRegisterAtStart(ins->generator()), + useBoxAtStart(ins->valueOrReason())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitAsyncAwait(MAsyncAwait* ins) { + MOZ_ASSERT(ins->generator()->type() == MIRType::Object); + auto* lir = new (alloc()) LAsyncAwait(useBoxAtStart(ins->value()), + useRegisterAtStart(ins->generator())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCanSkipAwait(MCanSkipAwait* ins) { + auto* lir = new (alloc()) LCanSkipAwait(useBoxAtStart(ins->value())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitMaybeExtractAwaitValue(MMaybeExtractAwaitValue* ins) { + auto* lir = new (alloc()) LMaybeExtractAwaitValue( + useBoxAtStart(ins->value()), useRegisterAtStart(ins->canSkip())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitDebugCheckSelfHosted(MDebugCheckSelfHosted* ins) { + MDefinition* checkVal = ins->checkValue(); + MOZ_ASSERT(checkVal->type() == MIRType::Value); + + LDebugCheckSelfHosted* lir = + new (alloc()) LDebugCheckSelfHosted(useBoxAtStart(checkVal)); + redefine(ins, checkVal); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitFinishBoundFunctionInit(MFinishBoundFunctionInit* ins) { + auto lir = new (alloc()) LFinishBoundFunctionInit( + useRegister(ins->bound()), useRegister(ins->target()), + useRegister(ins->argCount()), temp(), temp()); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitIsPackedArray(MIsPackedArray* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Boolean); + + auto lir = new (alloc()) LIsPackedArray(useRegister(ins->object()), temp()); + define(lir, ins); +} + +void LIRGenerator::visitGuardArrayIsPacked(MGuardArrayIsPacked* ins) { + MOZ_ASSERT(ins->array()->type() == MIRType::Object); + + auto* lir = new (alloc()) + LGuardArrayIsPacked(useRegister(ins->array()), temp(), temp()); + assignSnapshot(lir, ins->bailoutKind()); + add(lir, ins); + redefine(ins, ins->array()); +} + +void LIRGenerator::visitGetPrototypeOf(MGetPrototypeOf* ins) { + MOZ_ASSERT(ins->target()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Value); + + auto lir = new (alloc()) LGetPrototypeOf(useRegister(ins->target())); + defineBox(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitObjectWithProto(MObjectWithProto* ins) { + MOZ_ASSERT(ins->prototype()->type() == MIRType::Value); + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto* lir = new (alloc()) LObjectWithProto(useBoxAtStart(ins->prototype())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitObjectStaticProto(MObjectStaticProto* ins) { + MOZ_ASSERT(ins->object()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto* lir = + new (alloc()) LObjectStaticProto(useRegisterAtStart(ins->object())); + define(lir, ins); +}; + +void LIRGenerator::visitBuiltinObject(MBuiltinObject* ins) { + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto* lir = new (alloc()) LBuiltinObject(); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitReturn(MReturn* ret) { + return visitReturnImpl(ret->getOperand(0)); +} + +void LIRGenerator::visitGeneratorReturn(MGeneratorReturn* ret) { + return visitReturnImpl(ret->getOperand(0), true); +} + +void LIRGenerator::visitSuperFunction(MSuperFunction* ins) { + MOZ_ASSERT(ins->callee()->type() == MIRType::Object); + MOZ_ASSERT(ins->type() == MIRType::Value); + + auto* lir = new (alloc()) LSuperFunction(useRegister(ins->callee()), temp()); + defineBox(lir, ins); +} + +void LIRGenerator::visitInitHomeObject(MInitHomeObject* ins) { + MDefinition* function = ins->function(); + MOZ_ASSERT(function->type() == MIRType::Object); + + MDefinition* homeObject = ins->homeObject(); + MOZ_ASSERT(homeObject->type() == MIRType::Value); + + MOZ_ASSERT(ins->type() == MIRType::Object); + + auto* lir = new (alloc()) + LInitHomeObject(useRegisterAtStart(function), useBoxAtStart(homeObject)); + redefine(ins, function); + add(lir, ins); +} + +void LIRGenerator::visitIsTypedArrayConstructor(MIsTypedArrayConstructor* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + auto* lir = new (alloc()) LIsTypedArrayConstructor(useRegister(object)); + define(lir, ins); +} + +void LIRGenerator::visitLoadValueTag(MLoadValueTag* ins) { + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Value); + + define(new (alloc()) LLoadValueTag(useBoxAtStart(value)), ins); +} + +void LIRGenerator::visitGuardTagNotEqual(MGuardTagNotEqual* ins) { + MDefinition* lhs = ins->lhs(); + MOZ_ASSERT(lhs->type() == MIRType::Int32); + + MDefinition* rhs = ins->rhs(); + MOZ_ASSERT(rhs->type() == MIRType::Int32); + + auto* guard = + new (alloc()) LGuardTagNotEqual(useRegister(lhs), useRegister(rhs)); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); +} + +void LIRGenerator::visitLoadWrapperTarget(MLoadWrapperTarget* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + define(new (alloc()) LLoadWrapperTarget(useRegisterAtStart(object)), ins); +} + +void LIRGenerator::visitGuardHasGetterSetter(MGuardHasGetterSetter* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + auto* guard = new (alloc()) + LGuardHasGetterSetter(useRegisterAtStart(object), tempFixed(CallTempReg0), + tempFixed(CallTempReg1)); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, object); +} + +void LIRGenerator::visitGuardIsExtensible(MGuardIsExtensible* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + auto* guard = new (alloc()) LGuardIsExtensible(useRegister(object), temp()); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, object); +} + +void LIRGenerator::visitGuardInt32IsNonNegative(MGuardInt32IsNonNegative* ins) { + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* guard = new (alloc()) LGuardInt32IsNonNegative(useRegister(index)); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, index); +} + +void LIRGenerator::visitGuardIndexGreaterThanDenseInitLength( + MGuardIndexGreaterThanDenseInitLength* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + LDefinition spectreTemp = + BoundsCheckNeedsSpectreTemp() ? temp() : LDefinition::BogusTemp(); + + auto* guard = new (alloc()) LGuardIndexGreaterThanDenseInitLength( + useRegister(object), useRegister(index), temp(), spectreTemp); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, index); +} + +void LIRGenerator::visitGuardIndexIsValidUpdateOrAdd( + MGuardIndexIsValidUpdateOrAdd* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + LDefinition spectreTemp = + BoundsCheckNeedsSpectreTemp() ? temp() : LDefinition::BogusTemp(); + + auto* guard = new (alloc()) LGuardIndexIsValidUpdateOrAdd( + useRegister(object), useRegister(index), temp(), spectreTemp); + assignSnapshot(guard, ins->bailoutKind()); + add(guard, ins); + redefine(ins, index); +} + +void LIRGenerator::visitCallAddOrUpdateSparseElement( + MCallAddOrUpdateSparseElement* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + MDefinition* value = ins->value(); + MOZ_ASSERT(value->type() == MIRType::Value); + + auto* lir = new (alloc()) LCallAddOrUpdateSparseElement( + useRegisterAtStart(object), useRegisterAtStart(index), + useBoxAtStart(value)); + add(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallGetSparseElement(MCallGetSparseElement* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) LCallGetSparseElement(useRegisterAtStart(object), + useRegisterAtStart(index)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallNativeGetElement(MCallNativeGetElement* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) LCallNativeGetElement(useRegisterAtStart(object), + useRegisterAtStart(index)); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitCallObjectHasSparseElement( + MCallObjectHasSparseElement* ins) { + MDefinition* object = ins->object(); + MOZ_ASSERT(object->type() == MIRType::Object); + + MDefinition* index = ins->index(); + MOZ_ASSERT(index->type() == MIRType::Int32); + + auto* lir = new (alloc()) LCallObjectHasSparseElement( + useRegisterAtStart(object), useRegisterAtStart(index), + tempFixed(CallTempReg0), tempFixed(CallTempReg1)); + assignSnapshot(lir, ins->bailoutKind()); + defineReturn(lir, ins); +} + +void LIRGenerator::visitBigIntAsIntN(MBigIntAsIntN* ins) { + MOZ_ASSERT(ins->bits()->type() == MIRType::Int32); + MOZ_ASSERT(ins->input()->type() == MIRType::BigInt); + + if (ins->bits()->isConstant()) { + int32_t bits = ins->bits()->toConstant()->toInt32(); + if (bits == 64) { + auto* lir = new (alloc()) + LBigIntAsIntN64(useRegister(ins->input()), temp(), tempInt64()); + define(lir, ins); + assignSafepoint(lir, ins); + return; + } + if (bits == 32) { + auto* lir = new (alloc()) + LBigIntAsIntN32(useRegister(ins->input()), temp(), tempInt64()); + define(lir, ins); + assignSafepoint(lir, ins); + return; + } + } + + auto* lir = new (alloc()) LBigIntAsIntN(useRegisterAtStart(ins->bits()), + useRegisterAtStart(ins->input())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitBigIntAsUintN(MBigIntAsUintN* ins) { + MOZ_ASSERT(ins->bits()->type() == MIRType::Int32); + MOZ_ASSERT(ins->input()->type() == MIRType::BigInt); + + if (ins->bits()->isConstant()) { + int32_t bits = ins->bits()->toConstant()->toInt32(); + if (bits == 64) { + auto* lir = new (alloc()) + LBigIntAsUintN64(useRegister(ins->input()), temp(), tempInt64()); + define(lir, ins); + assignSafepoint(lir, ins); + return; + } + if (bits == 32) { + auto* lir = new (alloc()) + LBigIntAsUintN32(useRegister(ins->input()), temp(), tempInt64()); + define(lir, ins); + assignSafepoint(lir, ins); + return; + } + } + + auto* lir = new (alloc()) LBigIntAsUintN(useRegisterAtStart(ins->bits()), + useRegisterAtStart(ins->input())); + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitConstant(MConstant* ins) { + if (!IsFloatingPointType(ins->type()) && ins->canEmitAtUses()) { + emitAtUses(ins); + return; + } + + switch (ins->type()) { + case MIRType::Double: + define(new (alloc()) LDouble(ins->toDouble()), ins); + break; + case MIRType::Float32: + define(new (alloc()) LFloat32(ins->toFloat32()), ins); + break; + case MIRType::Boolean: + define(new (alloc()) LInteger(ins->toBoolean()), ins); + break; + case MIRType::Int32: + define(new (alloc()) LInteger(ins->toInt32()), ins); + break; + case MIRType::Int64: + defineInt64(new (alloc()) LInteger64(ins->toInt64()), ins); + break; + case MIRType::String: + define(new (alloc()) LPointer(ins->toString()), ins); + break; + case MIRType::Symbol: + define(new (alloc()) LPointer(ins->toSymbol()), ins); + break; + case MIRType::BigInt: + define(new (alloc()) LPointer(ins->toBigInt()), ins); + break; + case MIRType::Object: + define(new (alloc()) LPointer(&ins->toObject()), ins); + break; + default: + // Constants of special types (undefined, null) should never flow into + // here directly. Operations blindly consuming them require a Box. + MOZ_CRASH("unexpected constant type"); + } +} + +void LIRGenerator::visitWasmNullConstant(MWasmNullConstant* ins) { + define(new (alloc()) LWasmNullConstant(), ins); +} + +void LIRGenerator::visitWasmFloatConstant(MWasmFloatConstant* ins) { + switch (ins->type()) { + case MIRType::Double: + define(new (alloc()) LDouble(ins->toDouble()), ins); + break; + case MIRType::Float32: + define(new (alloc()) LFloat32(ins->toFloat32()), ins); + break; +#ifdef ENABLE_WASM_SIMD + case MIRType::Simd128: + define(new (alloc()) LSimd128(ins->toSimd128()), ins); + break; +#endif + default: + MOZ_CRASH("unexpected constant type"); + } +} + +#ifdef JS_JITSPEW +static void SpewResumePoint(MBasicBlock* block, MInstruction* ins, + MResumePoint* resumePoint) { + Fprinter& out = JitSpewPrinter(); + out.printf("Current resume point %p details:\n", (void*)resumePoint); + out.printf(" frame count: %u\n", resumePoint->frameCount()); + + if (ins) { + out.printf(" taken after: "); + ins->printName(out); + } else { + out.printf(" taken at block %u entry", block->id()); + } + out.printf("\n"); + + out.printf(" pc: %p (script: %p, offset: %d)\n", (void*)resumePoint->pc(), + (void*)resumePoint->block()->info().script(), + int(resumePoint->block()->info().script()->pcToOffset( + resumePoint->pc()))); + + for (size_t i = 0, e = resumePoint->numOperands(); i < e; i++) { + MDefinition* in = resumePoint->getOperand(i); + out.printf(" slot%u: ", (unsigned)i); + in->printName(out); + out.printf("\n"); + } +} +#endif + +void LIRGenerator::visitInstructionDispatch(MInstruction* ins) { +#ifdef JS_CODEGEN_NONE + // Don't compile the switch-statement below so that we don't have to define + // the platform-specific visit* methods for the none-backend. + MOZ_CRASH(); +#else + switch (ins->op()) { +# define MIR_OP(op) \ + case MDefinition::Opcode::op: \ + visit##op(ins->to##op()); \ + break; + MIR_OPCODE_LIST(MIR_OP) +# undef MIR_OP + default: + MOZ_CRASH("Invalid instruction"); + } +#endif +} + +void LIRGeneratorShared::visitEmittedAtUses(MInstruction* ins) { + static_cast<LIRGenerator*>(this)->visitInstructionDispatch(ins); +} + +bool LIRGenerator::visitInstruction(MInstruction* ins) { + MOZ_ASSERT(!errored()); + + if (ins->isRecoveredOnBailout()) { + MOZ_ASSERT(!JitOptions.disableRecoverIns); + return true; + } + + if (!gen->ensureBallast()) { + return false; + } + visitInstructionDispatch(ins); + + if (ins->resumePoint()) { + updateResumeState(ins); + } + +#ifdef DEBUG + ins->setInWorklistUnchecked(); +#endif + + // If no safepoint was created, there's no need for an OSI point. + if (LOsiPoint* osiPoint = popOsiPoint()) { + add(osiPoint); + } + + return !errored(); +} + +void LIRGenerator::definePhis() { + size_t lirIndex = 0; + MBasicBlock* block = current->mir(); + for (MPhiIterator phi(block->phisBegin()); phi != block->phisEnd(); phi++) { + if (phi->type() == MIRType::Value) { + defineUntypedPhi(*phi, lirIndex); + lirIndex += BOX_PIECES; + } else if (phi->type() == MIRType::Int64) { + defineInt64Phi(*phi, lirIndex); + lirIndex += INT64_PIECES; + } else { + defineTypedPhi(*phi, lirIndex); + lirIndex += 1; + } + } +} + +void LIRGenerator::updateResumeState(MInstruction* ins) { + lastResumePoint_ = ins->resumePoint(); +#ifdef JS_JITSPEW + if (JitSpewEnabled(JitSpew_IonSnapshots) && lastResumePoint_) { + SpewResumePoint(nullptr, ins, lastResumePoint_); + } +#endif +} + +void LIRGenerator::updateResumeState(MBasicBlock* block) { + // As Value Numbering phase can remove edges from the entry basic block to a + // code paths reachable from the OSR entry point, we have to add fixup + // blocks to keep the dominator tree organized the same way. These fixup + // blocks are flaged as unreachable, and should only exist iff the graph has + // an OSR block. + // + // Note: RangeAnalysis can flag blocks as unreachable, but they are only + // removed iff GVN (including UCE) is enabled. + MOZ_ASSERT_IF(!mir()->compilingWasm() && !block->unreachable(), + block->entryResumePoint()); + MOZ_ASSERT_IF( + block->unreachable(), + block->graph().osrBlock() || !mir()->optimizationInfo().gvnEnabled()); + lastResumePoint_ = block->entryResumePoint(); +#ifdef JS_JITSPEW + if (JitSpewEnabled(JitSpew_IonSnapshots) && lastResumePoint_) { + SpewResumePoint(block, nullptr, lastResumePoint_); + } +#endif +} + +bool LIRGenerator::visitBlock(MBasicBlock* block) { + current = block->lir(); + updateResumeState(block); + + definePhis(); + + // See fixup blocks added by Value Numbering, to keep the dominator relation + // modified by the presence of the OSR block. + MOZ_ASSERT_IF(block->unreachable(), + *block->begin() == block->lastIns() || + !mir()->optimizationInfo().gvnEnabled()); + MOZ_ASSERT_IF( + block->unreachable(), + block->graph().osrBlock() || !mir()->optimizationInfo().gvnEnabled()); + for (MInstructionIterator iter = block->begin(); *iter != block->lastIns(); + iter++) { + if (!visitInstruction(*iter)) { + return false; + } + } + + if (block->successorWithPhis()) { + // If we have a successor with phis, lower the phi input now that we + // are approaching the join point. + MBasicBlock* successor = block->successorWithPhis(); + uint32_t position = block->positionInPhiSuccessor(); + size_t lirIndex = 0; + for (MPhiIterator phi(successor->phisBegin()); phi != successor->phisEnd(); + phi++) { + if (!gen->ensureBallast()) { + return false; + } + + MDefinition* opd = phi->getOperand(position); + ensureDefined(opd); + + MOZ_ASSERT(opd->type() == phi->type()); + + if (phi->type() == MIRType::Value) { + lowerUntypedPhiInput(*phi, position, successor->lir(), lirIndex); + lirIndex += BOX_PIECES; + } else if (phi->type() == MIRType::Int64) { + lowerInt64PhiInput(*phi, position, successor->lir(), lirIndex); + lirIndex += INT64_PIECES; + } else { + lowerTypedPhiInput(*phi, position, successor->lir(), lirIndex); + lirIndex += 1; + } + } + } + + // Now emit the last instruction, which is some form of branch. + if (!visitInstruction(block->lastIns())) { + return false; + } + + return true; +} + +void LIRGenerator::visitNaNToZero(MNaNToZero* ins) { + MDefinition* input = ins->input(); + + if (ins->operandIsNeverNaN() && ins->operandIsNeverNegativeZero()) { + redefine(ins, input); + return; + } + LNaNToZero* lir = + new (alloc()) LNaNToZero(useRegisterAtStart(input), tempDouble()); + defineReuseInput(lir, ins, 0); +} + +bool LIRGenerator::generate() { + // Create all blocks and prep all phis beforehand. + for (ReversePostorderIterator block(graph.rpoBegin()); + block != graph.rpoEnd(); block++) { + if (gen->shouldCancel("Lowering (preparation loop)")) { + return false; + } + + if (!lirGraph_.initBlock(*block)) { + return false; + } + } + + for (ReversePostorderIterator block(graph.rpoBegin()); + block != graph.rpoEnd(); block++) { + if (gen->shouldCancel("Lowering (main loop)")) { + return false; + } + + if (!visitBlock(*block)) { + return false; + } + } + + lirGraph_.setArgumentSlotCount(maxargslots_); + return true; +} + +void LIRGenerator::visitPhi(MPhi* phi) { + // Phi nodes are not lowered because they are only meaningful for the register + // allocator. + MOZ_CRASH("Unexpected Phi node during Lowering."); +} + +void LIRGenerator::visitBeta(MBeta* beta) { + // Beta nodes are supposed to be removed before because they are + // only used to carry the range information for Range analysis + MOZ_CRASH("Unexpected Beta node during Lowering."); +} + +void LIRGenerator::visitObjectState(MObjectState* objState) { + // ObjectState nodes are always recovered on bailouts + MOZ_CRASH("Unexpected ObjectState node during Lowering."); +} + +void LIRGenerator::visitArrayState(MArrayState* objState) { + // ArrayState nodes are always recovered on bailouts + MOZ_CRASH("Unexpected ArrayState node during Lowering."); +} + +void LIRGenerator::visitIonToWasmCall(MIonToWasmCall* ins) { + // The instruction needs a temp register: + // - that's not the FramePointer, since wasm is going to use it in the + // function. + // - that's not aliasing an input register. + LDefinition scratch = tempFixed(ABINonArgReg0); + + // Also prevent register allocation from using wasm's FramePointer, in + // non-profiling mode. + LDefinition fp = gen->isProfilerInstrumentationEnabled() + ? LDefinition::BogusTemp() + : tempFixed(FramePointer); + + // Note that since this is a LIR call instruction, regalloc will prevent + // the use*AtStart below from reusing any of the temporaries. + + LInstruction* lir; + if (ins->type() == MIRType::Value) { + lir = allocateVariadic<LIonToWasmCallV>(ins->numOperands(), scratch, fp); + } else if (ins->type() == MIRType::Int64) { + lir = allocateVariadic<LIonToWasmCallI64>(ins->numOperands(), scratch, fp); + } else { + lir = allocateVariadic<LIonToWasmCall>(ins->numOperands(), scratch, fp); + } + if (!lir) { + abort(AbortReason::Alloc, "OOM: LIRGenerator::visitIonToWasmCall"); + return; + } + + ABIArgGenerator abi; + for (unsigned i = 0; i < ins->numOperands(); i++) { + MDefinition* argDef = ins->getOperand(i); + ABIArg arg = abi.next(ToMIRType(argDef->type())); + switch (arg.kind()) { + case ABIArg::GPR: + case ABIArg::FPU: + lir->setOperand(i, useFixedAtStart(argDef, arg.reg())); + break; + case ABIArg::Stack: + lir->setOperand(i, useAtStart(argDef)); + break; +#ifdef JS_CODEGEN_REGISTER_PAIR + case ABIArg::GPR_PAIR: + MOZ_CRASH( + "no way to pass i64, and wasm uses hardfp for function calls"); +#endif + case ABIArg::Uninitialized: + MOZ_CRASH("Uninitialized ABIArg kind"); + } + } + + defineReturn(lir, ins); + assignSafepoint(lir, ins); +} + +void LIRGenerator::visitWasmSelect(MWasmSelect* ins) { + MDefinition* condExpr = ins->condExpr(); + + // Pick off specific cases that we can do with LWasmCompareAndSelect. + // Currently only{U,}Int32 selection driven by a comparison of {U,}Int32 + // values. + if (condExpr->isCompare() && condExpr->isEmittedAtUses() && + ins->type() == MIRType::Int32) { + MCompare* comp = condExpr->toCompare(); + JSOp jsop = comp->jsop(); + MCompare::CompareType compTy = comp->compareType(); + // We don't currently generate any other JSOPs for the comparison, and if + // that changes, we want to know about it. Hence this assertion. + MOZ_ASSERT(jsop == JSOp::Eq || jsop == JSOp::Ne || jsop == JSOp::Lt || + jsop == JSOp::Gt || jsop == JSOp::Le || jsop == JSOp::Ge); + if (compTy == MCompare::Compare_Int32 || + compTy == MCompare::Compare_UInt32) { + auto* lir = new (alloc()) LWasmCompareAndSelect( + useRegister(comp->lhs()), useAny(comp->rhs()), compTy, jsop, + useRegisterAtStart(ins->trueExpr()), useAny(ins->falseExpr())); + + defineReuseInput(lir, ins, LWasmCompareAndSelect::IfTrueExprIndex); + return; + } + // Otherwise fall through to normal handling, which appears to emit the + // condexpr itself anyway. + } + + if (ins->type() == MIRType::Int64) { + auto* lir = new (alloc()) LWasmSelectI64( + useInt64RegisterAtStart(ins->trueExpr()), useInt64(ins->falseExpr()), + useRegister(ins->condExpr())); + + defineInt64ReuseInput(lir, ins, LWasmSelectI64::TrueExprIndex); + return; + } + + auto* lir = new (alloc()) + LWasmSelect(useRegisterAtStart(ins->trueExpr()), useAny(ins->falseExpr()), + useRegister(ins->condExpr())); + + defineReuseInput(lir, ins, LWasmSelect::TrueExprIndex); +} + +void LIRGenerator::visitWasmFence(MWasmFence* ins) { + add(new (alloc()) LWasmFence, ins); +} + +static_assert(!std::is_polymorphic_v<LIRGenerator>, + "LIRGenerator should not have any virtual methods"); + +#ifdef JS_CODEGEN_NONE +void LIRGenerator::visitReturnImpl(MDefinition*, bool) { MOZ_CRASH(); } +#endif |