diff options
Diffstat (limited to 'js/src/wasm/WasmCraneliftCompile.cpp')
-rw-r--r-- | js/src/wasm/WasmCraneliftCompile.cpp | 768 |
1 files changed, 768 insertions, 0 deletions
diff --git a/js/src/wasm/WasmCraneliftCompile.cpp b/js/src/wasm/WasmCraneliftCompile.cpp new file mode 100644 index 0000000000..561d675503 --- /dev/null +++ b/js/src/wasm/WasmCraneliftCompile.cpp @@ -0,0 +1,768 @@ +/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- + * vim: set ts=8 sts=2 et sw=2 tw=80: + * + * Copyright 2018 Mozilla Foundation + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +#include "wasm/WasmCraneliftCompile.h" + +#include "mozilla/CheckedInt.h" +#include "mozilla/ScopeExit.h" + +#include "jit/Disassemble.h" +#include "js/Printf.h" +#include "vm/JSContext.h" + +#include "wasm/cranelift/baldrapi.h" +#include "wasm/cranelift/clifapi.h" +#include "wasm/WasmFrameIter.h" // js::wasm::GenerateFunction{Pro,Epi}logue +#include "wasm/WasmGC.h" +#include "wasm/WasmGenerator.h" +#include "wasm/WasmStubs.h" + +#include "jit/MacroAssembler-inl.h" + +using namespace js; +using namespace js::jit; +using namespace js::wasm; + +using mozilla::CheckedInt; + +bool wasm::CraneliftPlatformSupport() { return cranelift_supports_platform(); } + +static inline SymbolicAddress ToSymbolicAddress(BD_SymbolicAddress bd) { + switch (bd) { + case BD_SymbolicAddress::RefFunc: + return SymbolicAddress::RefFunc; + case BD_SymbolicAddress::MemoryGrow: + return SymbolicAddress::MemoryGrow; + case BD_SymbolicAddress::MemorySize: + return SymbolicAddress::MemorySize; + case BD_SymbolicAddress::MemoryCopy: + return SymbolicAddress::MemCopy; + case BD_SymbolicAddress::MemoryCopyShared: + return SymbolicAddress::MemCopyShared; + case BD_SymbolicAddress::DataDrop: + return SymbolicAddress::DataDrop; + case BD_SymbolicAddress::MemoryFill: + return SymbolicAddress::MemFill; + case BD_SymbolicAddress::MemoryFillShared: + return SymbolicAddress::MemFillShared; + case BD_SymbolicAddress::MemoryInit: + return SymbolicAddress::MemInit; + case BD_SymbolicAddress::TableCopy: + return SymbolicAddress::TableCopy; + case BD_SymbolicAddress::ElemDrop: + return SymbolicAddress::ElemDrop; + case BD_SymbolicAddress::TableFill: + return SymbolicAddress::TableFill; + case BD_SymbolicAddress::TableGet: + return SymbolicAddress::TableGet; + case BD_SymbolicAddress::TableGrow: + return SymbolicAddress::TableGrow; + case BD_SymbolicAddress::TableInit: + return SymbolicAddress::TableInit; + case BD_SymbolicAddress::TableSet: + return SymbolicAddress::TableSet; + case BD_SymbolicAddress::TableSize: + return SymbolicAddress::TableSize; + case BD_SymbolicAddress::FloorF32: + return SymbolicAddress::FloorF; + case BD_SymbolicAddress::FloorF64: + return SymbolicAddress::FloorD; + case BD_SymbolicAddress::CeilF32: + return SymbolicAddress::CeilF; + case BD_SymbolicAddress::CeilF64: + return SymbolicAddress::CeilD; + case BD_SymbolicAddress::NearestF32: + return SymbolicAddress::NearbyIntF; + case BD_SymbolicAddress::NearestF64: + return SymbolicAddress::NearbyIntD; + case BD_SymbolicAddress::TruncF32: + return SymbolicAddress::TruncF; + case BD_SymbolicAddress::TruncF64: + return SymbolicAddress::TruncD; + case BD_SymbolicAddress::PreBarrier: + return SymbolicAddress::PreBarrierFiltering; + case BD_SymbolicAddress::PostBarrier: + return SymbolicAddress::PostBarrierFiltering; + case BD_SymbolicAddress::WaitI32: + return SymbolicAddress::WaitI32; + case BD_SymbolicAddress::WaitI64: + return SymbolicAddress::WaitI64; + case BD_SymbolicAddress::Wake: + return SymbolicAddress::Wake; + case BD_SymbolicAddress::Limit: + break; + } + MOZ_CRASH("unknown baldrdash symbolic address"); +} + +static bool GenerateCraneliftCode( + WasmMacroAssembler& masm, const CraneliftCompiledFunc& func, + const FuncType& funcType, const TypeIdDesc& funcTypeId, + uint32_t lineOrBytecode, uint32_t funcBytecodeSize, StackMaps* stackMaps, + size_t stackMapsOffset, size_t stackMapsCount, FuncOffsets* offsets) { + wasm::GenerateFunctionPrologue(masm, funcTypeId, mozilla::Nothing(), offsets); + + // Omit the check when framePushed is small and we know there's no + // recursion. + if (func.frame_pushed < MAX_UNCHECKED_LEAF_FRAME_SIZE && + !func.contains_calls) { + masm.reserveStack(func.frame_pushed); + } else { + std::pair<CodeOffset, uint32_t> pair = masm.wasmReserveStackChecked( + func.frame_pushed, BytecodeOffset(lineOrBytecode)); + CodeOffset trapInsnOffset = pair.first; + size_t nBytesReservedBeforeTrap = pair.second; + + MachineState trapExitLayout; + size_t trapExitLayoutNumWords; + GenerateTrapExitMachineState(&trapExitLayout, &trapExitLayoutNumWords); + + size_t nInboundStackArgBytes = StackArgAreaSizeUnaligned(funcType.args()); + + ArgTypeVector args(funcType); + wasm::StackMap* functionEntryStackMap = nullptr; + if (!CreateStackMapForFunctionEntryTrap( + args, trapExitLayout, trapExitLayoutNumWords, + nBytesReservedBeforeTrap, nInboundStackArgBytes, + &functionEntryStackMap)) { + return false; + } + + // In debug builds, we'll always have a stack map, even if there are no + // refs to track. + MOZ_ASSERT(functionEntryStackMap); + + if (functionEntryStackMap && + !stackMaps->add((uint8_t*)(uintptr_t)trapInsnOffset.offset(), + functionEntryStackMap)) { + functionEntryStackMap->destroy(); + return false; + } + } + MOZ_ASSERT(masm.framePushed() == func.frame_pushed); + + // Copy the machine code; handle jump tables and other read-only data below. + uint32_t funcBase = masm.currentOffset(); + if (func.code_size && !masm.appendRawCode(func.code, func.code_size)) { + return false; + } +#if defined(JS_CODEGEN_X64) || defined(JS_CODEGEN_X86) + uint32_t codeEnd = masm.currentOffset(); +#endif + + wasm::GenerateFunctionEpilogue(masm, func.frame_pushed, offsets); + + if (func.num_rodata_relocs > 0) { +#if defined(JS_CODEGEN_X64) || defined(JS_CODEGEN_X86) + constexpr size_t jumptableElementSize = 4; + + MOZ_ASSERT(func.jumptables_size % jumptableElementSize == 0); + + // Align the jump tables properly. + masm.haltingAlign(jumptableElementSize); + + // Copy over the tables and read-only data. + uint32_t rodataBase = masm.currentOffset(); + if (!masm.appendRawCode(func.code + func.code_size, + func.total_size - func.code_size)) { + return false; + } + + uint32_t numElem = func.jumptables_size / jumptableElementSize; + uint32_t bias = rodataBase - codeEnd; + + // Bias the jump table(s). The table values are negative values + // representing backward jumps. By shifting the table down we increase the + // distance and so we add a negative value to reflect the larger distance. + // + // Note addToPCRel4() works from the end of the instruction, hence the loop + // bounds. + for (uint32_t i = 1; i <= numElem; i++) { + masm.addToPCRel4(rodataBase + (i * jumptableElementSize), -bias); + } + + // Patch up the code locations. These represent forward distances that also + // become greater, so we add a positive value. + for (uint32_t i = 0; i < func.num_rodata_relocs; i++) { + MOZ_ASSERT(func.rodata_relocs[i] < func.code_size); + masm.addToPCRel4(funcBase + func.rodata_relocs[i], bias); + } +#else + MOZ_CRASH("No jump table support on this platform"); +#endif + } + + masm.flush(); + if (masm.oom()) { + return false; + } + offsets->end = masm.currentOffset(); + + for (size_t i = 0; i < stackMapsCount; i++) { + auto* maplet = stackMaps->getRef(stackMapsOffset + i); + maplet->offsetBy(funcBase); + } + + for (size_t i = 0; i < func.num_metadata; i++) { + const CraneliftMetadataEntry& metadata = func.metadatas[i]; + + CheckedInt<size_t> offset = funcBase; + offset += metadata.code_offset; + if (!offset.isValid()) { + return false; + } + +#ifdef DEBUG + // Check code offsets. + MOZ_ASSERT(offset.value() >= offsets->uncheckedCallEntry); + MOZ_ASSERT(offset.value() < offsets->ret); + MOZ_ASSERT(metadata.module_bytecode_offset != 0); + + // Check bytecode offsets. + if (lineOrBytecode > 0) { + MOZ_ASSERT(metadata.module_bytecode_offset >= lineOrBytecode); + MOZ_ASSERT(metadata.module_bytecode_offset < + lineOrBytecode + funcBytecodeSize); + } +#endif + uint32_t bytecodeOffset = metadata.module_bytecode_offset; + + switch (metadata.which) { + case CraneliftMetadataEntry::Which::DirectCall: { + CallSiteDesc desc(bytecodeOffset, CallSiteDesc::Func); + masm.append(desc, CodeOffset(offset.value()), metadata.extra); + break; + } + case CraneliftMetadataEntry::Which::IndirectCall: { + CallSiteDesc desc(bytecodeOffset, CallSiteDesc::Dynamic); + masm.append(desc, CodeOffset(offset.value())); + break; + } + case CraneliftMetadataEntry::Which::Trap: { + Trap trap = (Trap)metadata.extra; + BytecodeOffset trapOffset(bytecodeOffset); + masm.append(trap, wasm::TrapSite(offset.value(), trapOffset)); + break; + } + case CraneliftMetadataEntry::Which::SymbolicAccess: { + CodeOffset raOffset(offset.value()); + CallSiteDesc desc(bytecodeOffset, CallSiteDesc::Symbolic); + masm.append(desc, raOffset); + + SymbolicAddress sym = + ToSymbolicAddress(BD_SymbolicAddress(metadata.extra)); + masm.append(SymbolicAccess(raOffset, sym)); + break; + } + default: { + MOZ_CRASH("unknown cranelift metadata kind"); + } + } + } + + return true; +} + +// In Rust, a BatchCompiler variable has a lifetime constrained by those of its +// associated StaticEnvironment and ModuleEnvironment. This RAII class ties +// them together, as well as makes sure that the compiler is properly destroyed +// when it exits scope. + +class CraneliftContext { + CraneliftStaticEnvironment staticEnv_; + CraneliftModuleEnvironment moduleEnv_; + CraneliftCompiler* compiler_; + + public: + explicit CraneliftContext(const ModuleEnvironment& moduleEnv) + : moduleEnv_(moduleEnv), compiler_(nullptr) { + staticEnv_.ref_types_enabled = moduleEnv.refTypesEnabled(); + staticEnv_.threads_enabled = true; + staticEnv_.v128_enabled = moduleEnv.v128Enabled(); +#ifdef WASM_SUPPORTS_HUGE_MEMORY + if (moduleEnv.hugeMemoryEnabled()) { + // In the huge memory configuration, we always reserve the full 4 GB + // index space for a heap. + staticEnv_.static_memory_bound = HugeIndexRange; + staticEnv_.memory_guard_size = HugeOffsetGuardLimit; + } else { + staticEnv_.memory_guard_size = OffsetGuardLimit; + } +#endif + // Otherwise, heap bounds are stored in the `boundsCheckLimit32` field + // of TlsData. + } + bool init() { + compiler_ = cranelift_compiler_create(&staticEnv_, &moduleEnv_); + return !!compiler_; + } + ~CraneliftContext() { + if (compiler_) { + cranelift_compiler_destroy(compiler_); + } + } + operator CraneliftCompiler*() { return compiler_; } +}; + +CraneliftFuncCompileInput::CraneliftFuncCompileInput( + const FuncCompileInput& func) + : bytecode(func.begin), + bytecode_size(func.end - func.begin), + index(func.index), + offset_in_module(func.lineOrBytecode) {} + +static_assert(offsetof(TlsData, boundsCheckLimit32) == sizeof(void*), + "fix make_heap() in wasm2clif.rs"); + +CraneliftStaticEnvironment::CraneliftStaticEnvironment() + : +#ifdef JS_CODEGEN_X64 + has_sse2(Assembler::HasSSE2()), + has_sse3(Assembler::HasSSE3()), + has_sse41(Assembler::HasSSE41()), + has_sse42(Assembler::HasSSE42()), + has_popcnt(Assembler::HasPOPCNT()), + has_avx(Assembler::HasAVX()), + has_bmi1(Assembler::HasBMI1()), + has_bmi2(Assembler::HasBMI2()), + has_lzcnt(Assembler::HasLZCNT()), +#else + has_sse2(false), + has_sse3(false), + has_sse41(false), + has_sse42(false), + has_popcnt(false), + has_avx(false), + has_bmi1(false), + has_bmi2(false), + has_lzcnt(false), +#endif +#if defined(XP_WIN) + platform_is_windows(true), +#else + platform_is_windows(false), +#endif + ref_types_enabled(false), + threads_enabled(false), + v128_enabled(false), + static_memory_bound(0), + memory_guard_size(0), + memory_base_tls_offset(offsetof(TlsData, memoryBase)), + instance_tls_offset(offsetof(TlsData, instance)), + interrupt_tls_offset(offsetof(TlsData, interrupt)), + cx_tls_offset(offsetof(TlsData, cx)), + realm_cx_offset(JSContext::offsetOfRealm()), + realm_tls_offset(offsetof(TlsData, realm)), + realm_func_import_tls_offset(offsetof(FuncImportTls, realm)), + size_of_wasm_frame(sizeof(wasm::Frame)) { +} + +// Most of BaldrMonkey's data structures refer to a "global offset" which is a +// byte offset into the `globalArea` field of the `TlsData` struct. +// +// Cranelift represents global variables with their byte offset from the "VM +// context pointer" which is the `WasmTlsReg` pointing to the `TlsData` +// struct. +// +// This function translates between the two. + +static size_t globalToTlsOffset(size_t globalOffset) { + return offsetof(wasm::TlsData, globalArea) + globalOffset; +} + +CraneliftModuleEnvironment::CraneliftModuleEnvironment( + const ModuleEnvironment& env) + : env(&env) { + // env.minMemoryLength is in bytes. Convert it to wasm pages. + static_assert(sizeof(env.minMemoryLength) == 8); + MOZ_RELEASE_ASSERT(env.minMemoryLength <= (((uint64_t)1) << 32)); + MOZ_RELEASE_ASSERT((env.minMemoryLength & wasm::PageMask) == 0); + min_memory_length = (uint32_t)(env.minMemoryLength >> wasm::PageBits); +} + +TypeCode env_unpack(BD_ValType valType) { + return UnpackTypeCodeType(PackedTypeCode(valType.packed)); +} + +size_t env_num_datas(const CraneliftModuleEnvironment* env) { + return env->env->dataCount.valueOr(0); +} + +size_t env_num_elems(const CraneliftModuleEnvironment* env) { + return env->env->elemSegments.length(); +} +TypeCode env_elem_typecode(const CraneliftModuleEnvironment* env, + uint32_t index) { + return UnpackTypeCodeType(env->env->elemSegments[index]->elemType.packed()); +} + +uint32_t env_max_memory(const CraneliftModuleEnvironment* env) { + // env.maxMemoryLength is in bytes. Convert it to wasm pages. + if (env->env->maxMemoryLength.isSome()) { + // We use |auto| here rather than |uint64_t| so that the static_assert will + // fail if |maxMemoryLength| is changed to some other size. + auto inBytes = *(env->env->maxMemoryLength); + static_assert(sizeof(inBytes) == 8); + MOZ_RELEASE_ASSERT(inBytes <= (((uint64_t)1) << 32)); + MOZ_RELEASE_ASSERT((inBytes & wasm::PageMask) == 0); + return (uint32_t)(inBytes >> wasm::PageBits); + } else { + return UINT32_MAX; + } +} + +bool env_uses_shared_memory(const CraneliftModuleEnvironment* env) { + return env->env->usesSharedMemory(); +} + +bool env_has_memory(const CraneliftModuleEnvironment* env) { + return env->env->usesMemory(); +} + +size_t env_num_types(const CraneliftModuleEnvironment* env) { + return env->env->types.length(); +} +const FuncType* env_type(const CraneliftModuleEnvironment* env, + size_t typeIndex) { + return &env->env->types[typeIndex].funcType(); +} + +size_t env_num_funcs(const CraneliftModuleEnvironment* env) { + return env->env->funcs.length(); +} +const FuncType* env_func_sig(const CraneliftModuleEnvironment* env, + size_t funcIndex) { + return env->env->funcs[funcIndex].type; +} +const TypeIdDesc* env_func_sig_id(const CraneliftModuleEnvironment* env, + size_t funcIndex) { + return env->env->funcs[funcIndex].typeId; +} +size_t env_func_sig_index(const CraneliftModuleEnvironment* env, + size_t funcIndex) { + return env->env->funcs[funcIndex].typeIndex; +} +bool env_is_func_valid_for_ref(const CraneliftModuleEnvironment* env, + uint32_t index) { + return env->env->validForRefFunc.getBit(index); +} + +size_t env_func_import_tls_offset(const CraneliftModuleEnvironment* env, + size_t funcIndex) { + return globalToTlsOffset(env->env->funcImportGlobalDataOffsets[funcIndex]); +} + +bool env_func_is_import(const CraneliftModuleEnvironment* env, + size_t funcIndex) { + return env->env->funcIsImport(funcIndex); +} + +const FuncType* env_signature(const CraneliftModuleEnvironment* env, + size_t funcTypeIndex) { + return &env->env->types[funcTypeIndex].funcType(); +} + +const TypeIdDesc* env_signature_id(const CraneliftModuleEnvironment* env, + size_t funcTypeIndex) { + return &env->env->typeIds[funcTypeIndex]; +} + +size_t env_num_tables(const CraneliftModuleEnvironment* env) { + return env->env->tables.length(); +} +const TableDesc* env_table(const CraneliftModuleEnvironment* env, + size_t tableIndex) { + return &env->env->tables[tableIndex]; +} + +size_t env_num_globals(const CraneliftModuleEnvironment* env) { + return env->env->globals.length(); +} +const GlobalDesc* env_global(const CraneliftModuleEnvironment* env, + size_t globalIndex) { + return &env->env->globals[globalIndex]; +} + +bool wasm::CraneliftCompileFunctions(const ModuleEnvironment& moduleEnv, + const CompilerEnvironment& compilerEnv, + LifoAlloc& lifo, + const FuncCompileInputVector& inputs, + CompiledCode* code, UniqueChars* error) { + MOZ_RELEASE_ASSERT(CraneliftPlatformSupport()); + + MOZ_ASSERT(compilerEnv.tier() == Tier::Optimized); + MOZ_ASSERT(compilerEnv.debug() == DebugEnabled::False); + MOZ_ASSERT(compilerEnv.optimizedBackend() == OptimizedBackend::Cranelift); + MOZ_ASSERT(!moduleEnv.isAsmJS()); + + TempAllocator alloc(&lifo); + JitContext jitContext(&alloc); + WasmMacroAssembler masm(alloc, moduleEnv); + MOZ_ASSERT(IsCompilingWasm()); + + // Swap in already-allocated empty vectors to avoid malloc/free. + MOZ_ASSERT(code->empty()); + + CraneliftReusableData reusableContext; + if (!code->swapCranelift(masm, reusableContext)) { + return false; + } + + if (!reusableContext) { + auto context = MakeUnique<CraneliftContext>(moduleEnv); + if (!context || !context->init()) { + return false; + } + reusableContext.reset((void**)context.release()); + } + + CraneliftContext* compiler = (CraneliftContext*)reusableContext.get(); + + // Disable instruction spew if we're going to disassemble after code + // generation, or the output will be a mess. + + bool jitSpew = JitSpewEnabled(js::jit::JitSpew_Codegen); + if (jitSpew) { + DisableChannel(js::jit::JitSpew_Codegen); + } + auto reenableSpew = mozilla::MakeScopeExit([&] { + if (jitSpew) { + EnableChannel(js::jit::JitSpew_Codegen); + } + }); + + for (const FuncCompileInput& func : inputs) { + Decoder d(func.begin, func.end, func.lineOrBytecode, error); + + size_t funcBytecodeSize = func.end - func.begin; + + size_t previousStackmapCount = code->stackMaps.length(); + + CraneliftFuncCompileInput clifInput(func); + clifInput.stackmaps = (BD_Stackmaps*)&code->stackMaps; + + CraneliftCompiledFunc clifFunc; + + char* clifError = nullptr; + if (!cranelift_compile_function(*compiler, &clifInput, &clifFunc, + &clifError)) { + *error = JS_smprintf("%s", clifError); + cranelift_compiler_free_error(clifError); + return false; + } + + uint32_t lineOrBytecode = func.lineOrBytecode; + const FuncType& funcType = *moduleEnv.funcs[clifInput.index].type; + const TypeIdDesc& funcTypeId = *moduleEnv.funcs[clifInput.index].typeId; + + FuncOffsets offsets; + if (!GenerateCraneliftCode( + masm, clifFunc, funcType, funcTypeId, lineOrBytecode, + funcBytecodeSize, &code->stackMaps, previousStackmapCount, + code->stackMaps.length() - previousStackmapCount, &offsets)) { + return false; + } + + if (!code->codeRanges.emplaceBack(func.index, lineOrBytecode, offsets)) { + return false; + } + } + + masm.finish(); + if (masm.oom()) { + return false; + } + + if (jitSpew) { + // The disassembler uses the jitspew for output, so re-enable now. + EnableChannel(js::jit::JitSpew_Codegen); + + uint32_t totalCodeSize = masm.currentOffset(); + uint8_t* codeBuf = (uint8_t*)js_malloc(totalCodeSize); + if (codeBuf) { + masm.executableCopy(codeBuf); + + const CodeRangeVector& codeRanges = code->codeRanges; + MOZ_ASSERT(codeRanges.length() >= inputs.length()); + + // Within the current batch, functions' code ranges have been added in + // the same order as the inputs. + size_t firstCodeRangeIndex = codeRanges.length() - inputs.length(); + + for (size_t i = 0; i < inputs.length(); i++) { + int funcIndex = inputs[i].index; + mozilla::Unused << funcIndex; + + JitSpew(JitSpew_Codegen, "# ========================================"); + JitSpew(JitSpew_Codegen, "# Start of wasm cranelift code for index %d", + funcIndex); + + size_t codeRangeIndex = firstCodeRangeIndex + i; + uint32_t codeStart = codeRanges[codeRangeIndex].begin(); + uint32_t codeEnd = codeRanges[codeRangeIndex].end(); + + jit::Disassemble( + codeBuf + codeStart, codeEnd - codeStart, + [](const char* text) { JitSpew(JitSpew_Codegen, "%s", text); }); + + JitSpew(JitSpew_Codegen, "# End of wasm cranelift code for index %d", + funcIndex); + } + js_free(codeBuf); + } + } + + return code->swapCranelift(masm, reusableContext); +} + +void wasm::CraneliftFreeReusableData(void* ptr) { + CraneliftContext* compiler = (CraneliftContext*)ptr; + if (compiler) { + js_delete(compiler); + } +} + +//////////////////////////////////////////////////////////////////////////////// +// +// Callbacks from Rust to C++. + +// Offsets assumed by the `make_heap()` function. +static_assert(offsetof(wasm::TlsData, memoryBase) == 0, "memory base moved"); + +// The translate_call() function in wasm2clif.rs depends on these offsets. +static_assert(offsetof(wasm::FuncImportTls, code) == 0, + "Import code field moved"); +static_assert(offsetof(wasm::FuncImportTls, tls) == sizeof(void*), + "Import tls moved"); + +// Global + +bool global_isConstant(const GlobalDesc* global) { + return global->isConstant(); +} + +bool global_isIndirect(const GlobalDesc* global) { + return global->isIndirect(); +} + +BD_ConstantValue global_constantValue(const GlobalDesc* global) { + Val value(global->constantValue()); + BD_ConstantValue v; + v.t = TypeCode(value.type().kind()); + switch (v.t) { + case TypeCode::I32: + v.u.i32 = value.i32(); + break; + case TypeCode::I64: + v.u.i64 = value.i64(); + break; + case TypeCode::F32: + v.u.f32 = value.f32(); + break; + case TypeCode::F64: + v.u.f64 = value.f64(); + break; + case TypeCode::V128: + memcpy(&v.u.v128, &value.v128(), sizeof(v.u.v128)); + break; + case AbstractReferenceTypeCode: + v.u.r = value.ref().forCompiledCode(); + break; + default: + MOZ_CRASH("Bad type"); + } + return v; +} + +TypeCode global_type(const GlobalDesc* global) { + return UnpackTypeCodeType(global->type().packed()); +} + +size_t global_tlsOffset(const GlobalDesc* global) { + return globalToTlsOffset(global->offset()); +} + +// TableDesc + +size_t table_tlsOffset(const TableDesc* table) { + return globalToTlsOffset(table->globalDataOffset); +} + +uint32_t table_initialLimit(const TableDesc* table) { + return table->initialLength; +} +uint32_t table_maximumLimit(const TableDesc* table) { + return table->maximumLength.valueOr(UINT32_MAX); +} +TypeCode table_elementTypeCode(const TableDesc* table) { + return UnpackTypeCodeType(table->elemType.packed()); +} + +// Sig + +size_t funcType_numArgs(const FuncType* funcType) { + return funcType->args().length(); +} + +const BD_ValType* funcType_args(const FuncType* funcType) { + static_assert(sizeof(BD_ValType) == sizeof(ValType), "update BD_ValType"); + return (const BD_ValType*)funcType->args().begin(); +} + +size_t funcType_numResults(const FuncType* funcType) { + return funcType->results().length(); +} + +const BD_ValType* funcType_results(const FuncType* funcType) { + static_assert(sizeof(BD_ValType) == sizeof(ValType), "update BD_ValType"); + return (const BD_ValType*)funcType->results().begin(); +} + +TypeIdDescKind funcType_idKind(const TypeIdDesc* funcTypeId) { + return funcTypeId->kind(); +} + +size_t funcType_idImmediate(const TypeIdDesc* funcTypeId) { + return funcTypeId->immediate(); +} + +size_t funcType_idTlsOffset(const TypeIdDesc* funcTypeId) { + return globalToTlsOffset(funcTypeId->globalDataOffset()); +} + +void stackmaps_add(BD_Stackmaps* sink, const uint32_t* bitMap, + size_t mappedWords, size_t argsSize, size_t codeOffset) { + const uint32_t BitElemSize = sizeof(uint32_t) * 8; + + StackMaps* maps = (StackMaps*)sink; + StackMap* map = StackMap::create(mappedWords); + MOZ_ALWAYS_TRUE(map); + + // Copy the cranelift stackmap into our spidermonkey one + // TODO: Take ownership of the cranelift stackmap and avoid a copy + for (uint32_t i = 0; i < mappedWords; i++) { + uint32_t bit = (bitMap[i / BitElemSize] >> (i % BitElemSize)) & 0x1; + if (bit) { + map->setBit(i); + } + } + + map->setFrameOffsetFromTop((argsSize + sizeof(wasm::Frame)) / + sizeof(uintptr_t)); + MOZ_ALWAYS_TRUE(maps->add((uint8_t*)codeOffset, map)); +} |