summaryrefslogtreecommitdiffstats
path: root/src/VBox/Storage/VMDK.cpp
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-27 14:19:18 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-27 14:19:18 +0000
commit4035b1bfb1e5843a539a8b624d21952b756974d1 (patch)
treef1e9cd5bf548cbc57ff2fddfb2b4aa9ae95587e2 /src/VBox/Storage/VMDK.cpp
parentInitial commit. (diff)
downloadvirtualbox-upstream.tar.xz
virtualbox-upstream.zip
Adding upstream version 6.1.22-dfsg.upstream/6.1.22-dfsgupstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'src/VBox/Storage/VMDK.cpp')
-rw-r--r--src/VBox/Storage/VMDK.cpp6607
1 files changed, 6607 insertions, 0 deletions
diff --git a/src/VBox/Storage/VMDK.cpp b/src/VBox/Storage/VMDK.cpp
new file mode 100644
index 00000000..d6b6b751
--- /dev/null
+++ b/src/VBox/Storage/VMDK.cpp
@@ -0,0 +1,6607 @@
+/* $Id: VMDK.cpp $ */
+/** @file
+ * VMDK disk image, core code.
+ */
+
+/*
+ * Copyright (C) 2006-2020 Oracle Corporation
+ *
+ * This file is part of VirtualBox Open Source Edition (OSE), as
+ * available from http://www.virtualbox.org. This file is free software;
+ * you can redistribute it and/or modify it under the terms of the GNU
+ * General Public License (GPL) as published by the Free Software
+ * Foundation, in version 2 as it comes in the "COPYING" file of the
+ * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
+ * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
+ */
+
+
+/*********************************************************************************************************************************
+* Header Files *
+*********************************************************************************************************************************/
+#define LOG_GROUP LOG_GROUP_VD_VMDK
+#include <VBox/vd-plugin.h>
+#include <VBox/err.h>
+
+#include <VBox/log.h>
+#include <iprt/assert.h>
+#include <iprt/alloc.h>
+#include <iprt/uuid.h>
+#include <iprt/path.h>
+#include <iprt/string.h>
+#include <iprt/rand.h>
+#include <iprt/zip.h>
+#include <iprt/asm.h>
+
+#include "VDBackends.h"
+
+
+/*********************************************************************************************************************************
+* Constants And Macros, Structures and Typedefs *
+*********************************************************************************************************************************/
+
+/** Maximum encoded string size (including NUL) we allow for VMDK images.
+ * Deliberately not set high to avoid running out of descriptor space. */
+#define VMDK_ENCODED_COMMENT_MAX 1024
+
+/** VMDK descriptor DDB entry for PCHS cylinders. */
+#define VMDK_DDB_GEO_PCHS_CYLINDERS "ddb.geometry.cylinders"
+
+/** VMDK descriptor DDB entry for PCHS heads. */
+#define VMDK_DDB_GEO_PCHS_HEADS "ddb.geometry.heads"
+
+/** VMDK descriptor DDB entry for PCHS sectors. */
+#define VMDK_DDB_GEO_PCHS_SECTORS "ddb.geometry.sectors"
+
+/** VMDK descriptor DDB entry for LCHS cylinders. */
+#define VMDK_DDB_GEO_LCHS_CYLINDERS "ddb.geometry.biosCylinders"
+
+/** VMDK descriptor DDB entry for LCHS heads. */
+#define VMDK_DDB_GEO_LCHS_HEADS "ddb.geometry.biosHeads"
+
+/** VMDK descriptor DDB entry for LCHS sectors. */
+#define VMDK_DDB_GEO_LCHS_SECTORS "ddb.geometry.biosSectors"
+
+/** VMDK descriptor DDB entry for image UUID. */
+#define VMDK_DDB_IMAGE_UUID "ddb.uuid.image"
+
+/** VMDK descriptor DDB entry for image modification UUID. */
+#define VMDK_DDB_MODIFICATION_UUID "ddb.uuid.modification"
+
+/** VMDK descriptor DDB entry for parent image UUID. */
+#define VMDK_DDB_PARENT_UUID "ddb.uuid.parent"
+
+/** VMDK descriptor DDB entry for parent image modification UUID. */
+#define VMDK_DDB_PARENT_MODIFICATION_UUID "ddb.uuid.parentmodification"
+
+/** No compression for streamOptimized files. */
+#define VMDK_COMPRESSION_NONE 0
+
+/** Deflate compression for streamOptimized files. */
+#define VMDK_COMPRESSION_DEFLATE 1
+
+/** Marker that the actual GD value is stored in the footer. */
+#define VMDK_GD_AT_END 0xffffffffffffffffULL
+
+/** Marker for end-of-stream in streamOptimized images. */
+#define VMDK_MARKER_EOS 0
+
+/** Marker for grain table block in streamOptimized images. */
+#define VMDK_MARKER_GT 1
+
+/** Marker for grain directory block in streamOptimized images. */
+#define VMDK_MARKER_GD 2
+
+/** Marker for footer in streamOptimized images. */
+#define VMDK_MARKER_FOOTER 3
+
+/** Marker for unknown purpose in streamOptimized images.
+ * Shows up in very recent images created by vSphere, but only sporadically.
+ * They "forgot" to document that one in the VMDK specification. */
+#define VMDK_MARKER_UNSPECIFIED 4
+
+/** Dummy marker for "don't check the marker value". */
+#define VMDK_MARKER_IGNORE 0xffffffffU
+
+/**
+ * Magic number for hosted images created by VMware Workstation 4, VMware
+ * Workstation 5, VMware Server or VMware Player. Not necessarily sparse.
+ */
+#define VMDK_SPARSE_MAGICNUMBER 0x564d444b /* 'V' 'M' 'D' 'K' */
+
+/**
+ * VMDK hosted binary extent header. The "Sparse" is a total misnomer, as
+ * this header is also used for monolithic flat images.
+ */
+#pragma pack(1)
+typedef struct SparseExtentHeader
+{
+ uint32_t magicNumber;
+ uint32_t version;
+ uint32_t flags;
+ uint64_t capacity;
+ uint64_t grainSize;
+ uint64_t descriptorOffset;
+ uint64_t descriptorSize;
+ uint32_t numGTEsPerGT;
+ uint64_t rgdOffset;
+ uint64_t gdOffset;
+ uint64_t overHead;
+ bool uncleanShutdown;
+ char singleEndLineChar;
+ char nonEndLineChar;
+ char doubleEndLineChar1;
+ char doubleEndLineChar2;
+ uint16_t compressAlgorithm;
+ uint8_t pad[433];
+} SparseExtentHeader;
+#pragma pack()
+
+/** The maximum allowed descriptor size in the extent header in sectors. */
+#define VMDK_SPARSE_DESCRIPTOR_SIZE_MAX UINT64_C(20480) /* 10MB */
+
+/** VMDK capacity for a single chunk when 2G splitting is turned on. Should be
+ * divisible by the default grain size (64K) */
+#define VMDK_2G_SPLIT_SIZE (2047 * 1024 * 1024)
+
+/** VMDK streamOptimized file format marker. The type field may or may not
+ * be actually valid, but there's always data to read there. */
+#pragma pack(1)
+typedef struct VMDKMARKER
+{
+ uint64_t uSector;
+ uint32_t cbSize;
+ uint32_t uType;
+} VMDKMARKER, *PVMDKMARKER;
+#pragma pack()
+
+
+/** Convert sector number/size to byte offset/size. */
+#define VMDK_SECTOR2BYTE(u) ((uint64_t)(u) << 9)
+
+/** Convert byte offset/size to sector number/size. */
+#define VMDK_BYTE2SECTOR(u) ((u) >> 9)
+
+/**
+ * VMDK extent type.
+ */
+typedef enum VMDKETYPE
+{
+ /** Hosted sparse extent. */
+ VMDKETYPE_HOSTED_SPARSE = 1,
+ /** Flat extent. */
+ VMDKETYPE_FLAT,
+ /** Zero extent. */
+ VMDKETYPE_ZERO,
+ /** VMFS extent, used by ESX. */
+ VMDKETYPE_VMFS
+} VMDKETYPE, *PVMDKETYPE;
+
+/**
+ * VMDK access type for a extent.
+ */
+typedef enum VMDKACCESS
+{
+ /** No access allowed. */
+ VMDKACCESS_NOACCESS = 0,
+ /** Read-only access. */
+ VMDKACCESS_READONLY,
+ /** Read-write access. */
+ VMDKACCESS_READWRITE
+} VMDKACCESS, *PVMDKACCESS;
+
+/** Forward declaration for PVMDKIMAGE. */
+typedef struct VMDKIMAGE *PVMDKIMAGE;
+
+/**
+ * Extents files entry. Used for opening a particular file only once.
+ */
+typedef struct VMDKFILE
+{
+ /** Pointer to file path. Local copy. */
+ const char *pszFilename;
+ /** Pointer to base name. Local copy. */
+ const char *pszBasename;
+ /** File open flags for consistency checking. */
+ unsigned fOpen;
+ /** Handle for sync/async file abstraction.*/
+ PVDIOSTORAGE pStorage;
+ /** Reference counter. */
+ unsigned uReferences;
+ /** Flag whether the file should be deleted on last close. */
+ bool fDelete;
+ /** Pointer to the image we belong to (for debugging purposes). */
+ PVMDKIMAGE pImage;
+ /** Pointer to next file descriptor. */
+ struct VMDKFILE *pNext;
+ /** Pointer to the previous file descriptor. */
+ struct VMDKFILE *pPrev;
+} VMDKFILE, *PVMDKFILE;
+
+/**
+ * VMDK extent data structure.
+ */
+typedef struct VMDKEXTENT
+{
+ /** File handle. */
+ PVMDKFILE pFile;
+ /** Base name of the image extent. */
+ const char *pszBasename;
+ /** Full name of the image extent. */
+ const char *pszFullname;
+ /** Number of sectors in this extent. */
+ uint64_t cSectors;
+ /** Number of sectors per block (grain in VMDK speak). */
+ uint64_t cSectorsPerGrain;
+ /** Starting sector number of descriptor. */
+ uint64_t uDescriptorSector;
+ /** Size of descriptor in sectors. */
+ uint64_t cDescriptorSectors;
+ /** Starting sector number of grain directory. */
+ uint64_t uSectorGD;
+ /** Starting sector number of redundant grain directory. */
+ uint64_t uSectorRGD;
+ /** Total number of metadata sectors. */
+ uint64_t cOverheadSectors;
+ /** Nominal size (i.e. as described by the descriptor) of this extent. */
+ uint64_t cNominalSectors;
+ /** Sector offset (i.e. as described by the descriptor) of this extent. */
+ uint64_t uSectorOffset;
+ /** Number of entries in a grain table. */
+ uint32_t cGTEntries;
+ /** Number of sectors reachable via a grain directory entry. */
+ uint32_t cSectorsPerGDE;
+ /** Number of entries in the grain directory. */
+ uint32_t cGDEntries;
+ /** Pointer to the next free sector. Legacy information. Do not use. */
+ uint32_t uFreeSector;
+ /** Number of this extent in the list of images. */
+ uint32_t uExtent;
+ /** Pointer to the descriptor (NULL if no descriptor in this extent). */
+ char *pDescData;
+ /** Pointer to the grain directory. */
+ uint32_t *pGD;
+ /** Pointer to the redundant grain directory. */
+ uint32_t *pRGD;
+ /** VMDK version of this extent. 1=1.0/1.1 */
+ uint32_t uVersion;
+ /** Type of this extent. */
+ VMDKETYPE enmType;
+ /** Access to this extent. */
+ VMDKACCESS enmAccess;
+ /** Flag whether this extent is marked as unclean. */
+ bool fUncleanShutdown;
+ /** Flag whether the metadata in the extent header needs to be updated. */
+ bool fMetaDirty;
+ /** Flag whether there is a footer in this extent. */
+ bool fFooter;
+ /** Compression type for this extent. */
+ uint16_t uCompression;
+ /** Append position for writing new grain. Only for sparse extents. */
+ uint64_t uAppendPosition;
+ /** Last grain which was accessed. Only for streamOptimized extents. */
+ uint32_t uLastGrainAccess;
+ /** Starting sector corresponding to the grain buffer. */
+ uint32_t uGrainSectorAbs;
+ /** Grain number corresponding to the grain buffer. */
+ uint32_t uGrain;
+ /** Actual size of the compressed data, only valid for reading. */
+ uint32_t cbGrainStreamRead;
+ /** Size of compressed grain buffer for streamOptimized extents. */
+ size_t cbCompGrain;
+ /** Compressed grain buffer for streamOptimized extents, with marker. */
+ void *pvCompGrain;
+ /** Decompressed grain buffer for streamOptimized extents. */
+ void *pvGrain;
+ /** Reference to the image in which this extent is used. Do not use this
+ * on a regular basis to avoid passing pImage references to functions
+ * explicitly. */
+ struct VMDKIMAGE *pImage;
+} VMDKEXTENT, *PVMDKEXTENT;
+
+/**
+ * Grain table cache size. Allocated per image.
+ */
+#define VMDK_GT_CACHE_SIZE 256
+
+/**
+ * Grain table block size. Smaller than an actual grain table block to allow
+ * more grain table blocks to be cached without having to allocate excessive
+ * amounts of memory for the cache.
+ */
+#define VMDK_GT_CACHELINE_SIZE 128
+
+
+/**
+ * Maximum number of lines in a descriptor file. Not worth the effort of
+ * making it variable. Descriptor files are generally very short (~20 lines),
+ * with the exception of sparse files split in 2G chunks, which need for the
+ * maximum size (almost 2T) exactly 1025 lines for the disk database.
+ */
+#define VMDK_DESCRIPTOR_LINES_MAX 1100U
+
+/**
+ * Parsed descriptor information. Allows easy access and update of the
+ * descriptor (whether separate file or not). Free form text files suck.
+ */
+typedef struct VMDKDESCRIPTOR
+{
+ /** Line number of first entry of the disk descriptor. */
+ unsigned uFirstDesc;
+ /** Line number of first entry in the extent description. */
+ unsigned uFirstExtent;
+ /** Line number of first disk database entry. */
+ unsigned uFirstDDB;
+ /** Total number of lines. */
+ unsigned cLines;
+ /** Total amount of memory available for the descriptor. */
+ size_t cbDescAlloc;
+ /** Set if descriptor has been changed and not yet written to disk. */
+ bool fDirty;
+ /** Array of pointers to the data in the descriptor. */
+ char *aLines[VMDK_DESCRIPTOR_LINES_MAX];
+ /** Array of line indices pointing to the next non-comment line. */
+ unsigned aNextLines[VMDK_DESCRIPTOR_LINES_MAX];
+} VMDKDESCRIPTOR, *PVMDKDESCRIPTOR;
+
+
+/**
+ * Cache entry for translating extent/sector to a sector number in that
+ * extent.
+ */
+typedef struct VMDKGTCACHEENTRY
+{
+ /** Extent number for which this entry is valid. */
+ uint32_t uExtent;
+ /** GT data block number. */
+ uint64_t uGTBlock;
+ /** Data part of the cache entry. */
+ uint32_t aGTData[VMDK_GT_CACHELINE_SIZE];
+} VMDKGTCACHEENTRY, *PVMDKGTCACHEENTRY;
+
+/**
+ * Cache data structure for blocks of grain table entries. For now this is a
+ * fixed size direct mapping cache, but this should be adapted to the size of
+ * the sparse image and maybe converted to a set-associative cache. The
+ * implementation below implements a write-through cache with write allocate.
+ */
+typedef struct VMDKGTCACHE
+{
+ /** Cache entries. */
+ VMDKGTCACHEENTRY aGTCache[VMDK_GT_CACHE_SIZE];
+ /** Number of cache entries (currently unused). */
+ unsigned cEntries;
+} VMDKGTCACHE, *PVMDKGTCACHE;
+
+/**
+ * Complete VMDK image data structure. Mainly a collection of extents and a few
+ * extra global data fields.
+ */
+typedef struct VMDKIMAGE
+{
+ /** Image name. */
+ const char *pszFilename;
+ /** Descriptor file if applicable. */
+ PVMDKFILE pFile;
+
+ /** Pointer to the per-disk VD interface list. */
+ PVDINTERFACE pVDIfsDisk;
+ /** Pointer to the per-image VD interface list. */
+ PVDINTERFACE pVDIfsImage;
+
+ /** Error interface. */
+ PVDINTERFACEERROR pIfError;
+ /** I/O interface. */
+ PVDINTERFACEIOINT pIfIo;
+
+
+ /** Pointer to the image extents. */
+ PVMDKEXTENT pExtents;
+ /** Number of image extents. */
+ unsigned cExtents;
+ /** Pointer to the files list, for opening a file referenced multiple
+ * times only once (happens mainly with raw partition access). */
+ PVMDKFILE pFiles;
+
+ /**
+ * Pointer to an array of segment entries for async I/O.
+ * This is an optimization because the task number to submit is not known
+ * and allocating/freeing an array in the read/write functions every time
+ * is too expensive.
+ */
+ PPDMDATASEG paSegments;
+ /** Entries available in the segments array. */
+ unsigned cSegments;
+
+ /** Open flags passed by VBoxHD layer. */
+ unsigned uOpenFlags;
+ /** Image flags defined during creation or determined during open. */
+ unsigned uImageFlags;
+ /** Total size of the image. */
+ uint64_t cbSize;
+ /** Physical geometry of this image. */
+ VDGEOMETRY PCHSGeometry;
+ /** Logical geometry of this image. */
+ VDGEOMETRY LCHSGeometry;
+ /** Image UUID. */
+ RTUUID ImageUuid;
+ /** Image modification UUID. */
+ RTUUID ModificationUuid;
+ /** Parent image UUID. */
+ RTUUID ParentUuid;
+ /** Parent image modification UUID. */
+ RTUUID ParentModificationUuid;
+
+ /** Pointer to grain table cache, if this image contains sparse extents. */
+ PVMDKGTCACHE pGTCache;
+ /** Pointer to the descriptor (NULL if no separate descriptor file). */
+ char *pDescData;
+ /** Allocation size of the descriptor file. */
+ size_t cbDescAlloc;
+ /** Parsed descriptor file content. */
+ VMDKDESCRIPTOR Descriptor;
+ /** The static region list. */
+ VDREGIONLIST RegionList;
+} VMDKIMAGE;
+
+
+/** State for the input/output callout of the inflate reader/deflate writer. */
+typedef struct VMDKCOMPRESSIO
+{
+ /* Image this operation relates to. */
+ PVMDKIMAGE pImage;
+ /* Current read position. */
+ ssize_t iOffset;
+ /* Size of the compressed grain buffer (available data). */
+ size_t cbCompGrain;
+ /* Pointer to the compressed grain buffer. */
+ void *pvCompGrain;
+} VMDKCOMPRESSIO;
+
+
+/** Tracks async grain allocation. */
+typedef struct VMDKGRAINALLOCASYNC
+{
+ /** Flag whether the allocation failed. */
+ bool fIoErr;
+ /** Current number of transfers pending.
+ * If reached 0 and there is an error the old state is restored. */
+ unsigned cIoXfersPending;
+ /** Sector number */
+ uint64_t uSector;
+ /** Flag whether the grain table needs to be updated. */
+ bool fGTUpdateNeeded;
+ /** Extent the allocation happens. */
+ PVMDKEXTENT pExtent;
+ /** Position of the new grain, required for the grain table update. */
+ uint64_t uGrainOffset;
+ /** Grain table sector. */
+ uint64_t uGTSector;
+ /** Backup grain table sector. */
+ uint64_t uRGTSector;
+} VMDKGRAINALLOCASYNC, *PVMDKGRAINALLOCASYNC;
+
+/**
+ * State information for vmdkRename() and helpers.
+ */
+typedef struct VMDKRENAMESTATE
+{
+ /** Array of old filenames. */
+ char **apszOldName;
+ /** Array of new filenames. */
+ char **apszNewName;
+ /** Array of new lines in the extent descriptor. */
+ char **apszNewLines;
+ /** Name of the old descriptor file if not a sparse image. */
+ char *pszOldDescName;
+ /** Flag whether we called vmdkFreeImage(). */
+ bool fImageFreed;
+ /** Flag whther the descriptor is embedded in the image (sparse) or
+ * in a separate file. */
+ bool fEmbeddedDesc;
+ /** Number of extents in the image. */
+ unsigned cExtents;
+ /** New base filename. */
+ char *pszNewBaseName;
+ /** The old base filename. */
+ char *pszOldBaseName;
+ /** New full filename. */
+ char *pszNewFullName;
+ /** Old full filename. */
+ char *pszOldFullName;
+ /** The old image name. */
+ const char *pszOldImageName;
+ /** Copy of the original VMDK descriptor. */
+ VMDKDESCRIPTOR DescriptorCopy;
+ /** Copy of the extent state for sparse images. */
+ VMDKEXTENT ExtentCopy;
+} VMDKRENAMESTATE;
+/** Pointer to a VMDK rename state. */
+typedef VMDKRENAMESTATE *PVMDKRENAMESTATE;
+
+
+/*********************************************************************************************************************************
+* Static Variables *
+*********************************************************************************************************************************/
+
+/** NULL-terminated array of supported file extensions. */
+static const VDFILEEXTENSION s_aVmdkFileExtensions[] =
+{
+ {"vmdk", VDTYPE_HDD},
+ {NULL, VDTYPE_INVALID}
+};
+
+
+/*********************************************************************************************************************************
+* Internal Functions *
+*********************************************************************************************************************************/
+
+static void vmdkFreeStreamBuffers(PVMDKEXTENT pExtent);
+static int vmdkFreeExtentData(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
+ bool fDelete);
+
+static int vmdkCreateExtents(PVMDKIMAGE pImage, unsigned cExtents);
+static int vmdkFlushImage(PVMDKIMAGE pImage, PVDIOCTX pIoCtx);
+static int vmdkSetImageComment(PVMDKIMAGE pImage, const char *pszComment);
+static int vmdkFreeImage(PVMDKIMAGE pImage, bool fDelete, bool fFlush);
+
+static DECLCALLBACK(int) vmdkAllocGrainComplete(void *pBackendData, PVDIOCTX pIoCtx,
+ void *pvUser, int rcReq);
+
+/**
+ * Internal: open a file (using a file descriptor cache to ensure each file
+ * is only opened once - anything else can cause locking problems).
+ */
+static int vmdkFileOpen(PVMDKIMAGE pImage, PVMDKFILE *ppVmdkFile,
+ const char *pszBasename, const char *pszFilename, uint32_t fOpen)
+{
+ int rc = VINF_SUCCESS;
+ PVMDKFILE pVmdkFile;
+
+ for (pVmdkFile = pImage->pFiles;
+ pVmdkFile != NULL;
+ pVmdkFile = pVmdkFile->pNext)
+ {
+ if (!strcmp(pszFilename, pVmdkFile->pszFilename))
+ {
+ Assert(fOpen == pVmdkFile->fOpen);
+ pVmdkFile->uReferences++;
+
+ *ppVmdkFile = pVmdkFile;
+
+ return rc;
+ }
+ }
+
+ /* If we get here, there's no matching entry in the cache. */
+ pVmdkFile = (PVMDKFILE)RTMemAllocZ(sizeof(VMDKFILE));
+ if (!pVmdkFile)
+ {
+ *ppVmdkFile = NULL;
+ return VERR_NO_MEMORY;
+ }
+
+ pVmdkFile->pszFilename = RTStrDup(pszFilename);
+ if (!pVmdkFile->pszFilename)
+ {
+ RTMemFree(pVmdkFile);
+ *ppVmdkFile = NULL;
+ return VERR_NO_MEMORY;
+ }
+
+ if (pszBasename)
+ {
+ pVmdkFile->pszBasename = RTStrDup(pszBasename);
+ if (!pVmdkFile->pszBasename)
+ {
+ RTStrFree((char *)(void *)pVmdkFile->pszFilename);
+ RTMemFree(pVmdkFile);
+ *ppVmdkFile = NULL;
+ return VERR_NO_MEMORY;
+ }
+ }
+
+ pVmdkFile->fOpen = fOpen;
+
+ rc = vdIfIoIntFileOpen(pImage->pIfIo, pszFilename, fOpen,
+ &pVmdkFile->pStorage);
+ if (RT_SUCCESS(rc))
+ {
+ pVmdkFile->uReferences = 1;
+ pVmdkFile->pImage = pImage;
+ pVmdkFile->pNext = pImage->pFiles;
+ if (pImage->pFiles)
+ pImage->pFiles->pPrev = pVmdkFile;
+ pImage->pFiles = pVmdkFile;
+ *ppVmdkFile = pVmdkFile;
+ }
+ else
+ {
+ RTStrFree((char *)(void *)pVmdkFile->pszFilename);
+ RTMemFree(pVmdkFile);
+ *ppVmdkFile = NULL;
+ }
+
+ return rc;
+}
+
+/**
+ * Internal: close a file, updating the file descriptor cache.
+ */
+static int vmdkFileClose(PVMDKIMAGE pImage, PVMDKFILE *ppVmdkFile, bool fDelete)
+{
+ int rc = VINF_SUCCESS;
+ PVMDKFILE pVmdkFile = *ppVmdkFile;
+
+ AssertPtr(pVmdkFile);
+
+ pVmdkFile->fDelete |= fDelete;
+ Assert(pVmdkFile->uReferences);
+ pVmdkFile->uReferences--;
+ if (pVmdkFile->uReferences == 0)
+ {
+ PVMDKFILE pPrev;
+ PVMDKFILE pNext;
+
+ /* Unchain the element from the list. */
+ pPrev = pVmdkFile->pPrev;
+ pNext = pVmdkFile->pNext;
+
+ if (pNext)
+ pNext->pPrev = pPrev;
+ if (pPrev)
+ pPrev->pNext = pNext;
+ else
+ pImage->pFiles = pNext;
+
+ rc = vdIfIoIntFileClose(pImage->pIfIo, pVmdkFile->pStorage);
+
+ bool fFileDel = pVmdkFile->fDelete;
+ if ( pVmdkFile->pszBasename
+ && fFileDel)
+ {
+ const char *pszSuffix = RTPathSuffix(pVmdkFile->pszBasename);
+ if ( RTPathHasPath(pVmdkFile->pszBasename)
+ || !pszSuffix
+ || ( strcmp(pszSuffix, ".vmdk")
+ && strcmp(pszSuffix, ".bin")
+ && strcmp(pszSuffix, ".img")))
+ fFileDel = false;
+ }
+
+ if (fFileDel)
+ {
+ int rc2 = vdIfIoIntFileDelete(pImage->pIfIo, pVmdkFile->pszFilename);
+ if (RT_SUCCESS(rc))
+ rc = rc2;
+ }
+ else if (pVmdkFile->fDelete)
+ LogRel(("VMDK: Denying deletion of %s\n", pVmdkFile->pszBasename));
+ RTStrFree((char *)(void *)pVmdkFile->pszFilename);
+ if (pVmdkFile->pszBasename)
+ RTStrFree((char *)(void *)pVmdkFile->pszBasename);
+ RTMemFree(pVmdkFile);
+ }
+
+ *ppVmdkFile = NULL;
+ return rc;
+}
+
+/*#define VMDK_USE_BLOCK_DECOMP_API - test and enable */
+#ifndef VMDK_USE_BLOCK_DECOMP_API
+static DECLCALLBACK(int) vmdkFileInflateHelper(void *pvUser, void *pvBuf, size_t cbBuf, size_t *pcbBuf)
+{
+ VMDKCOMPRESSIO *pInflateState = (VMDKCOMPRESSIO *)pvUser;
+ size_t cbInjected = 0;
+
+ Assert(cbBuf);
+ if (pInflateState->iOffset < 0)
+ {
+ *(uint8_t *)pvBuf = RTZIPTYPE_ZLIB;
+ pvBuf = (uint8_t *)pvBuf + 1;
+ cbBuf--;
+ cbInjected = 1;
+ pInflateState->iOffset = RT_UOFFSETOF(VMDKMARKER, uType);
+ }
+ if (!cbBuf)
+ {
+ if (pcbBuf)
+ *pcbBuf = cbInjected;
+ return VINF_SUCCESS;
+ }
+ cbBuf = RT_MIN(cbBuf, pInflateState->cbCompGrain - pInflateState->iOffset);
+ memcpy(pvBuf,
+ (uint8_t *)pInflateState->pvCompGrain + pInflateState->iOffset,
+ cbBuf);
+ pInflateState->iOffset += cbBuf;
+ Assert(pcbBuf);
+ *pcbBuf = cbBuf + cbInjected;
+ return VINF_SUCCESS;
+}
+#endif
+
+/**
+ * Internal: read from a file and inflate the compressed data,
+ * distinguishing between async and normal operation
+ */
+DECLINLINE(int) vmdkFileInflateSync(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
+ uint64_t uOffset, void *pvBuf,
+ size_t cbToRead, const void *pcvMarker,
+ uint64_t *puLBA, uint32_t *pcbMarkerData)
+{
+ int rc;
+#ifndef VMDK_USE_BLOCK_DECOMP_API
+ PRTZIPDECOMP pZip = NULL;
+#endif
+ VMDKMARKER *pMarker = (VMDKMARKER *)pExtent->pvCompGrain;
+ size_t cbCompSize, cbActuallyRead;
+
+ if (!pcvMarker)
+ {
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ uOffset, pMarker, RT_UOFFSETOF(VMDKMARKER, uType));
+ if (RT_FAILURE(rc))
+ return rc;
+ }
+ else
+ {
+ memcpy(pMarker, pcvMarker, RT_UOFFSETOF(VMDKMARKER, uType));
+ /* pcvMarker endianness has already been partially transformed, fix it */
+ pMarker->uSector = RT_H2LE_U64(pMarker->uSector);
+ pMarker->cbSize = RT_H2LE_U32(pMarker->cbSize);
+ }
+
+ cbCompSize = RT_LE2H_U32(pMarker->cbSize);
+ if (cbCompSize == 0)
+ {
+ AssertMsgFailed(("VMDK: corrupted marker\n"));
+ return VERR_VD_VMDK_INVALID_FORMAT;
+ }
+
+ /* Sanity check - the expansion ratio should be much less than 2. */
+ Assert(cbCompSize < 2 * cbToRead);
+ if (cbCompSize >= 2 * cbToRead)
+ return VERR_VD_VMDK_INVALID_FORMAT;
+
+ /* Compressed grain marker. Data follows immediately. */
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ uOffset + RT_UOFFSETOF(VMDKMARKER, uType),
+ (uint8_t *)pExtent->pvCompGrain
+ + RT_UOFFSETOF(VMDKMARKER, uType),
+ RT_ALIGN_Z( cbCompSize
+ + RT_UOFFSETOF(VMDKMARKER, uType),
+ 512)
+ - RT_UOFFSETOF(VMDKMARKER, uType));
+
+ if (puLBA)
+ *puLBA = RT_LE2H_U64(pMarker->uSector);
+ if (pcbMarkerData)
+ *pcbMarkerData = RT_ALIGN( cbCompSize
+ + RT_UOFFSETOF(VMDKMARKER, uType),
+ 512);
+
+#ifdef VMDK_USE_BLOCK_DECOMP_API
+ rc = RTZipBlockDecompress(RTZIPTYPE_ZLIB, 0 /*fFlags*/,
+ pExtent->pvCompGrain, cbCompSize + RT_UOFFSETOF(VMDKMARKER, uType), NULL,
+ pvBuf, cbToRead, &cbActuallyRead);
+#else
+ VMDKCOMPRESSIO InflateState;
+ InflateState.pImage = pImage;
+ InflateState.iOffset = -1;
+ InflateState.cbCompGrain = cbCompSize + RT_UOFFSETOF(VMDKMARKER, uType);
+ InflateState.pvCompGrain = pExtent->pvCompGrain;
+
+ rc = RTZipDecompCreate(&pZip, &InflateState, vmdkFileInflateHelper);
+ if (RT_FAILURE(rc))
+ return rc;
+ rc = RTZipDecompress(pZip, pvBuf, cbToRead, &cbActuallyRead);
+ RTZipDecompDestroy(pZip);
+#endif /* !VMDK_USE_BLOCK_DECOMP_API */
+ if (RT_FAILURE(rc))
+ {
+ if (rc == VERR_ZIP_CORRUPTED)
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: Compressed image is corrupted '%s'"), pExtent->pszFullname);
+ return rc;
+ }
+ if (cbActuallyRead != cbToRead)
+ rc = VERR_VD_VMDK_INVALID_FORMAT;
+ return rc;
+}
+
+static DECLCALLBACK(int) vmdkFileDeflateHelper(void *pvUser, const void *pvBuf, size_t cbBuf)
+{
+ VMDKCOMPRESSIO *pDeflateState = (VMDKCOMPRESSIO *)pvUser;
+
+ Assert(cbBuf);
+ if (pDeflateState->iOffset < 0)
+ {
+ pvBuf = (const uint8_t *)pvBuf + 1;
+ cbBuf--;
+ pDeflateState->iOffset = RT_UOFFSETOF(VMDKMARKER, uType);
+ }
+ if (!cbBuf)
+ return VINF_SUCCESS;
+ if (pDeflateState->iOffset + cbBuf > pDeflateState->cbCompGrain)
+ return VERR_BUFFER_OVERFLOW;
+ memcpy((uint8_t *)pDeflateState->pvCompGrain + pDeflateState->iOffset,
+ pvBuf, cbBuf);
+ pDeflateState->iOffset += cbBuf;
+ return VINF_SUCCESS;
+}
+
+/**
+ * Internal: deflate the uncompressed data and write to a file,
+ * distinguishing between async and normal operation
+ */
+DECLINLINE(int) vmdkFileDeflateSync(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
+ uint64_t uOffset, const void *pvBuf,
+ size_t cbToWrite, uint64_t uLBA,
+ uint32_t *pcbMarkerData)
+{
+ int rc;
+ PRTZIPCOMP pZip = NULL;
+ VMDKCOMPRESSIO DeflateState;
+
+ DeflateState.pImage = pImage;
+ DeflateState.iOffset = -1;
+ DeflateState.cbCompGrain = pExtent->cbCompGrain;
+ DeflateState.pvCompGrain = pExtent->pvCompGrain;
+
+ rc = RTZipCompCreate(&pZip, &DeflateState, vmdkFileDeflateHelper,
+ RTZIPTYPE_ZLIB, RTZIPLEVEL_DEFAULT);
+ if (RT_FAILURE(rc))
+ return rc;
+ rc = RTZipCompress(pZip, pvBuf, cbToWrite);
+ if (RT_SUCCESS(rc))
+ rc = RTZipCompFinish(pZip);
+ RTZipCompDestroy(pZip);
+ if (RT_SUCCESS(rc))
+ {
+ Assert( DeflateState.iOffset > 0
+ && (size_t)DeflateState.iOffset <= DeflateState.cbCompGrain);
+
+ /* pad with zeroes to get to a full sector size */
+ uint32_t uSize = DeflateState.iOffset;
+ if (uSize % 512)
+ {
+ uint32_t uSizeAlign = RT_ALIGN(uSize, 512);
+ memset((uint8_t *)pExtent->pvCompGrain + uSize, '\0',
+ uSizeAlign - uSize);
+ uSize = uSizeAlign;
+ }
+
+ if (pcbMarkerData)
+ *pcbMarkerData = uSize;
+
+ /* Compressed grain marker. Data follows immediately. */
+ VMDKMARKER *pMarker = (VMDKMARKER *)pExtent->pvCompGrain;
+ pMarker->uSector = RT_H2LE_U64(uLBA);
+ pMarker->cbSize = RT_H2LE_U32( DeflateState.iOffset
+ - RT_UOFFSETOF(VMDKMARKER, uType));
+ rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ uOffset, pMarker, uSize);
+ if (RT_FAILURE(rc))
+ return rc;
+ }
+ return rc;
+}
+
+
+/**
+ * Internal: check if all files are closed, prevent leaking resources.
+ */
+static int vmdkFileCheckAllClose(PVMDKIMAGE pImage)
+{
+ int rc = VINF_SUCCESS, rc2;
+ PVMDKFILE pVmdkFile;
+
+ Assert(pImage->pFiles == NULL);
+ for (pVmdkFile = pImage->pFiles;
+ pVmdkFile != NULL;
+ pVmdkFile = pVmdkFile->pNext)
+ {
+ LogRel(("VMDK: leaking reference to file \"%s\"\n",
+ pVmdkFile->pszFilename));
+ pImage->pFiles = pVmdkFile->pNext;
+
+ rc2 = vmdkFileClose(pImage, &pVmdkFile, pVmdkFile->fDelete);
+
+ if (RT_SUCCESS(rc))
+ rc = rc2;
+ }
+ return rc;
+}
+
+/**
+ * Internal: truncate a string (at a UTF8 code point boundary) and encode the
+ * critical non-ASCII characters.
+ */
+static char *vmdkEncodeString(const char *psz)
+{
+ char szEnc[VMDK_ENCODED_COMMENT_MAX + 3];
+ char *pszDst = szEnc;
+
+ AssertPtr(psz);
+
+ for (; *psz; psz = RTStrNextCp(psz))
+ {
+ char *pszDstPrev = pszDst;
+ RTUNICP Cp = RTStrGetCp(psz);
+ if (Cp == '\\')
+ {
+ pszDst = RTStrPutCp(pszDst, Cp);
+ pszDst = RTStrPutCp(pszDst, Cp);
+ }
+ else if (Cp == '\n')
+ {
+ pszDst = RTStrPutCp(pszDst, '\\');
+ pszDst = RTStrPutCp(pszDst, 'n');
+ }
+ else if (Cp == '\r')
+ {
+ pszDst = RTStrPutCp(pszDst, '\\');
+ pszDst = RTStrPutCp(pszDst, 'r');
+ }
+ else
+ pszDst = RTStrPutCp(pszDst, Cp);
+ if (pszDst - szEnc >= VMDK_ENCODED_COMMENT_MAX - 1)
+ {
+ pszDst = pszDstPrev;
+ break;
+ }
+ }
+ *pszDst = '\0';
+ return RTStrDup(szEnc);
+}
+
+/**
+ * Internal: decode a string and store it into the specified string.
+ */
+static int vmdkDecodeString(const char *pszEncoded, char *psz, size_t cb)
+{
+ int rc = VINF_SUCCESS;
+ char szBuf[4];
+
+ if (!cb)
+ return VERR_BUFFER_OVERFLOW;
+
+ AssertPtr(psz);
+
+ for (; *pszEncoded; pszEncoded = RTStrNextCp(pszEncoded))
+ {
+ char *pszDst = szBuf;
+ RTUNICP Cp = RTStrGetCp(pszEncoded);
+ if (Cp == '\\')
+ {
+ pszEncoded = RTStrNextCp(pszEncoded);
+ RTUNICP CpQ = RTStrGetCp(pszEncoded);
+ if (CpQ == 'n')
+ RTStrPutCp(pszDst, '\n');
+ else if (CpQ == 'r')
+ RTStrPutCp(pszDst, '\r');
+ else if (CpQ == '\0')
+ {
+ rc = VERR_VD_VMDK_INVALID_HEADER;
+ break;
+ }
+ else
+ RTStrPutCp(pszDst, CpQ);
+ }
+ else
+ pszDst = RTStrPutCp(pszDst, Cp);
+
+ /* Need to leave space for terminating NUL. */
+ if ((size_t)(pszDst - szBuf) + 1 >= cb)
+ {
+ rc = VERR_BUFFER_OVERFLOW;
+ break;
+ }
+ memcpy(psz, szBuf, pszDst - szBuf);
+ psz += pszDst - szBuf;
+ }
+ *psz = '\0';
+ return rc;
+}
+
+/**
+ * Internal: free all buffers associated with grain directories.
+ */
+static void vmdkFreeGrainDirectory(PVMDKEXTENT pExtent)
+{
+ if (pExtent->pGD)
+ {
+ RTMemFree(pExtent->pGD);
+ pExtent->pGD = NULL;
+ }
+ if (pExtent->pRGD)
+ {
+ RTMemFree(pExtent->pRGD);
+ pExtent->pRGD = NULL;
+ }
+}
+
+/**
+ * Internal: allocate the compressed/uncompressed buffers for streamOptimized
+ * images.
+ */
+static int vmdkAllocStreamBuffers(PVMDKIMAGE pImage, PVMDKEXTENT pExtent)
+{
+ int rc = VINF_SUCCESS;
+
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ /* streamOptimized extents need a compressed grain buffer, which must
+ * be big enough to hold uncompressible data (which needs ~8 bytes
+ * more than the uncompressed data), the marker and padding. */
+ pExtent->cbCompGrain = RT_ALIGN_Z( VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain)
+ + 8 + sizeof(VMDKMARKER), 512);
+ pExtent->pvCompGrain = RTMemAlloc(pExtent->cbCompGrain);
+ if (RT_LIKELY(pExtent->pvCompGrain))
+ {
+ /* streamOptimized extents need a decompressed grain buffer. */
+ pExtent->pvGrain = RTMemAlloc(VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain));
+ if (!pExtent->pvGrain)
+ rc = VERR_NO_MEMORY;
+ }
+ else
+ rc = VERR_NO_MEMORY;
+ }
+
+ if (RT_FAILURE(rc))
+ vmdkFreeStreamBuffers(pExtent);
+ return rc;
+}
+
+/**
+ * Internal: allocate all buffers associated with grain directories.
+ */
+static int vmdkAllocGrainDirectory(PVMDKIMAGE pImage, PVMDKEXTENT pExtent)
+{
+ RT_NOREF1(pImage);
+ int rc = VINF_SUCCESS;
+ size_t cbGD = pExtent->cGDEntries * sizeof(uint32_t);
+
+ pExtent->pGD = (uint32_t *)RTMemAllocZ(cbGD);
+ if (RT_LIKELY(pExtent->pGD))
+ {
+ if (pExtent->uSectorRGD)
+ {
+ pExtent->pRGD = (uint32_t *)RTMemAllocZ(cbGD);
+ if (RT_UNLIKELY(!pExtent->pRGD))
+ rc = VERR_NO_MEMORY;
+ }
+ }
+ else
+ rc = VERR_NO_MEMORY;
+
+ if (RT_FAILURE(rc))
+ vmdkFreeGrainDirectory(pExtent);
+ return rc;
+}
+
+/**
+ * Converts the grain directory from little to host endianess.
+ *
+ * @returns nothing.
+ * @param pGD The grain directory.
+ * @param cGDEntries Number of entries in the grain directory to convert.
+ */
+DECLINLINE(void) vmdkGrainDirectoryConvToHost(uint32_t *pGD, uint32_t cGDEntries)
+{
+ uint32_t *pGDTmp = pGD;
+
+ for (uint32_t i = 0; i < cGDEntries; i++, pGDTmp++)
+ *pGDTmp = RT_LE2H_U32(*pGDTmp);
+}
+
+/**
+ * Read the grain directory and allocated grain tables verifying them against
+ * their back up copies if available.
+ *
+ * @returns VBox status code.
+ * @param pImage Image instance data.
+ * @param pExtent The VMDK extent.
+ */
+static int vmdkReadGrainDirectory(PVMDKIMAGE pImage, PVMDKEXTENT pExtent)
+{
+ int rc = VINF_SUCCESS;
+ size_t cbGD = pExtent->cGDEntries * sizeof(uint32_t);
+
+ AssertReturn(( pExtent->enmType == VMDKETYPE_HOSTED_SPARSE
+ && pExtent->uSectorGD != VMDK_GD_AT_END
+ && pExtent->uSectorRGD != VMDK_GD_AT_END), VERR_INTERNAL_ERROR);
+
+ rc = vmdkAllocGrainDirectory(pImage, pExtent);
+ if (RT_SUCCESS(rc))
+ {
+ /* The VMDK 1.1 spec seems to talk about compressed grain directories,
+ * but in reality they are not compressed. */
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(pExtent->uSectorGD),
+ pExtent->pGD, cbGD);
+ if (RT_SUCCESS(rc))
+ {
+ vmdkGrainDirectoryConvToHost(pExtent->pGD, pExtent->cGDEntries);
+
+ if ( pExtent->uSectorRGD
+ && !(pImage->uOpenFlags & VD_OPEN_FLAGS_SKIP_CONSISTENCY_CHECKS))
+ {
+ /* The VMDK 1.1 spec seems to talk about compressed grain directories,
+ * but in reality they are not compressed. */
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(pExtent->uSectorRGD),
+ pExtent->pRGD, cbGD);
+ if (RT_SUCCESS(rc))
+ {
+ vmdkGrainDirectoryConvToHost(pExtent->pRGD, pExtent->cGDEntries);
+
+ /* Check grain table and redundant grain table for consistency. */
+ size_t cbGT = pExtent->cGTEntries * sizeof(uint32_t);
+ size_t cbGTBuffers = cbGT; /* Start with space for one GT. */
+ size_t cbGTBuffersMax = _1M;
+
+ uint32_t *pTmpGT1 = (uint32_t *)RTMemAlloc(cbGTBuffers);
+ uint32_t *pTmpGT2 = (uint32_t *)RTMemAlloc(cbGTBuffers);
+
+ if ( !pTmpGT1
+ || !pTmpGT2)
+ rc = VERR_NO_MEMORY;
+
+ size_t i = 0;
+ uint32_t *pGDTmp = pExtent->pGD;
+ uint32_t *pRGDTmp = pExtent->pRGD;
+
+ /* Loop through all entries. */
+ while (i < pExtent->cGDEntries)
+ {
+ uint32_t uGTStart = *pGDTmp;
+ uint32_t uRGTStart = *pRGDTmp;
+ size_t cbGTRead = cbGT;
+
+ /* If no grain table is allocated skip the entry. */
+ if (*pGDTmp == 0 && *pRGDTmp == 0)
+ {
+ i++;
+ continue;
+ }
+
+ if (*pGDTmp == 0 || *pRGDTmp == 0 || *pGDTmp == *pRGDTmp)
+ {
+ /* Just one grain directory entry refers to a not yet allocated
+ * grain table or both grain directory copies refer to the same
+ * grain table. Not allowed. */
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: inconsistent references to grain directory in '%s'"), pExtent->pszFullname);
+ break;
+ }
+
+ i++;
+ pGDTmp++;
+ pRGDTmp++;
+
+ /*
+ * Read a few tables at once if adjacent to decrease the number
+ * of I/O requests. Read at maximum 1MB at once.
+ */
+ while ( i < pExtent->cGDEntries
+ && cbGTRead < cbGTBuffersMax)
+ {
+ /* If no grain table is allocated skip the entry. */
+ if (*pGDTmp == 0 && *pRGDTmp == 0)
+ {
+ i++;
+ continue;
+ }
+
+ if (*pGDTmp == 0 || *pRGDTmp == 0 || *pGDTmp == *pRGDTmp)
+ {
+ /* Just one grain directory entry refers to a not yet allocated
+ * grain table or both grain directory copies refer to the same
+ * grain table. Not allowed. */
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: inconsistent references to grain directory in '%s'"), pExtent->pszFullname);
+ break;
+ }
+
+ /* Check that the start offsets are adjacent.*/
+ if ( VMDK_SECTOR2BYTE(uGTStart) + cbGTRead != VMDK_SECTOR2BYTE(*pGDTmp)
+ || VMDK_SECTOR2BYTE(uRGTStart) + cbGTRead != VMDK_SECTOR2BYTE(*pRGDTmp))
+ break;
+
+ i++;
+ pGDTmp++;
+ pRGDTmp++;
+ cbGTRead += cbGT;
+ }
+
+ /* Increase buffers if required. */
+ if ( RT_SUCCESS(rc)
+ && cbGTBuffers < cbGTRead)
+ {
+ uint32_t *pTmp;
+ pTmp = (uint32_t *)RTMemRealloc(pTmpGT1, cbGTRead);
+ if (pTmp)
+ {
+ pTmpGT1 = pTmp;
+ pTmp = (uint32_t *)RTMemRealloc(pTmpGT2, cbGTRead);
+ if (pTmp)
+ pTmpGT2 = pTmp;
+ else
+ rc = VERR_NO_MEMORY;
+ }
+ else
+ rc = VERR_NO_MEMORY;
+
+ if (rc == VERR_NO_MEMORY)
+ {
+ /* Reset to the old values. */
+ rc = VINF_SUCCESS;
+ i -= cbGTRead / cbGT;
+ cbGTRead = cbGT;
+
+ /* Don't try to increase the buffer again in the next run. */
+ cbGTBuffersMax = cbGTBuffers;
+ }
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ /* The VMDK 1.1 spec seems to talk about compressed grain tables,
+ * but in reality they are not compressed. */
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uGTStart),
+ pTmpGT1, cbGTRead);
+ if (RT_FAILURE(rc))
+ {
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
+ N_("VMDK: error reading grain table in '%s'"), pExtent->pszFullname);
+ break;
+ }
+ /* The VMDK 1.1 spec seems to talk about compressed grain tables,
+ * but in reality they are not compressed. */
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uRGTStart),
+ pTmpGT2, cbGTRead);
+ if (RT_FAILURE(rc))
+ {
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
+ N_("VMDK: error reading backup grain table in '%s'"), pExtent->pszFullname);
+ break;
+ }
+ if (memcmp(pTmpGT1, pTmpGT2, cbGTRead))
+ {
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: inconsistency between grain table and backup grain table in '%s'"), pExtent->pszFullname);
+ break;
+ }
+ }
+ } /* while (i < pExtent->cGDEntries) */
+
+ /** @todo figure out what to do for unclean VMDKs. */
+ if (pTmpGT1)
+ RTMemFree(pTmpGT1);
+ if (pTmpGT2)
+ RTMemFree(pTmpGT2);
+ }
+ else
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
+ N_("VMDK: could not read redundant grain directory in '%s'"), pExtent->pszFullname);
+ }
+ }
+ else
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
+ N_("VMDK: could not read grain directory in '%s': %Rrc"), pExtent->pszFullname, rc);
+ }
+
+ if (RT_FAILURE(rc))
+ vmdkFreeGrainDirectory(pExtent);
+ return rc;
+}
+
+/**
+ * Creates a new grain directory for the given extent at the given start sector.
+ *
+ * @returns VBox status code.
+ * @param pImage Image instance data.
+ * @param pExtent The VMDK extent.
+ * @param uStartSector Where the grain directory should be stored in the image.
+ * @param fPreAlloc Flag whether to pre allocate the grain tables at this point.
+ */
+static int vmdkCreateGrainDirectory(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
+ uint64_t uStartSector, bool fPreAlloc)
+{
+ int rc = VINF_SUCCESS;
+ unsigned i;
+ size_t cbGD = pExtent->cGDEntries * sizeof(uint32_t);
+ size_t cbGDRounded = RT_ALIGN_64(cbGD, 512);
+ size_t cbGTRounded;
+ uint64_t cbOverhead;
+
+ if (fPreAlloc)
+ {
+ cbGTRounded = RT_ALIGN_64(pExtent->cGDEntries * pExtent->cGTEntries * sizeof(uint32_t), 512);
+ cbOverhead = VMDK_SECTOR2BYTE(uStartSector) + cbGDRounded + cbGTRounded;
+ }
+ else
+ {
+ /* Use a dummy start sector for layout computation. */
+ if (uStartSector == VMDK_GD_AT_END)
+ uStartSector = 1;
+ cbGTRounded = 0;
+ cbOverhead = VMDK_SECTOR2BYTE(uStartSector) + cbGDRounded;
+ }
+
+ /* For streamOptimized extents there is only one grain directory,
+ * and for all others take redundant grain directory into account. */
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ cbOverhead = RT_ALIGN_64(cbOverhead,
+ VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain));
+ }
+ else
+ {
+ cbOverhead += cbGDRounded + cbGTRounded;
+ cbOverhead = RT_ALIGN_64(cbOverhead,
+ VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain));
+ rc = vdIfIoIntFileSetSize(pImage->pIfIo, pExtent->pFile->pStorage, cbOverhead);
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ pExtent->uAppendPosition = cbOverhead;
+ pExtent->cOverheadSectors = VMDK_BYTE2SECTOR(cbOverhead);
+
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ pExtent->uSectorRGD = 0;
+ pExtent->uSectorGD = uStartSector;
+ }
+ else
+ {
+ pExtent->uSectorRGD = uStartSector;
+ pExtent->uSectorGD = uStartSector + VMDK_BYTE2SECTOR(cbGDRounded + cbGTRounded);
+ }
+
+ rc = vmdkAllocStreamBuffers(pImage, pExtent);
+ if (RT_SUCCESS(rc))
+ {
+ rc = vmdkAllocGrainDirectory(pImage, pExtent);
+ if ( RT_SUCCESS(rc)
+ && fPreAlloc)
+ {
+ uint32_t uGTSectorLE;
+ uint64_t uOffsetSectors;
+
+ if (pExtent->pRGD)
+ {
+ uOffsetSectors = pExtent->uSectorRGD + VMDK_BYTE2SECTOR(cbGDRounded);
+ for (i = 0; i < pExtent->cGDEntries; i++)
+ {
+ pExtent->pRGD[i] = uOffsetSectors;
+ uGTSectorLE = RT_H2LE_U64(uOffsetSectors);
+ /* Write the redundant grain directory entry to disk. */
+ rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(pExtent->uSectorRGD) + i * sizeof(uGTSectorLE),
+ &uGTSectorLE, sizeof(uGTSectorLE));
+ if (RT_FAILURE(rc))
+ {
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write new redundant grain directory entry in '%s'"), pExtent->pszFullname);
+ break;
+ }
+ uOffsetSectors += VMDK_BYTE2SECTOR(pExtent->cGTEntries * sizeof(uint32_t));
+ }
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ uOffsetSectors = pExtent->uSectorGD + VMDK_BYTE2SECTOR(cbGDRounded);
+ for (i = 0; i < pExtent->cGDEntries; i++)
+ {
+ pExtent->pGD[i] = uOffsetSectors;
+ uGTSectorLE = RT_H2LE_U64(uOffsetSectors);
+ /* Write the grain directory entry to disk. */
+ rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(pExtent->uSectorGD) + i * sizeof(uGTSectorLE),
+ &uGTSectorLE, sizeof(uGTSectorLE));
+ if (RT_FAILURE(rc))
+ {
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write new grain directory entry in '%s'"), pExtent->pszFullname);
+ break;
+ }
+ uOffsetSectors += VMDK_BYTE2SECTOR(pExtent->cGTEntries * sizeof(uint32_t));
+ }
+ }
+ }
+ }
+ }
+
+ if (RT_FAILURE(rc))
+ vmdkFreeGrainDirectory(pExtent);
+ return rc;
+}
+
+/**
+ * Unquotes the given string returning the result in a separate buffer.
+ *
+ * @returns VBox status code.
+ * @param pImage The VMDK image state.
+ * @param pszStr The string to unquote.
+ * @param ppszUnquoted Where to store the return value, use RTMemTmpFree to
+ * free.
+ * @param ppszNext Where to store the pointer to any character following
+ * the quoted value, optional.
+ */
+static int vmdkStringUnquote(PVMDKIMAGE pImage, const char *pszStr,
+ char **ppszUnquoted, char **ppszNext)
+{
+ const char *pszStart = pszStr;
+ char *pszQ;
+ char *pszUnquoted;
+
+ /* Skip over whitespace. */
+ while (*pszStr == ' ' || *pszStr == '\t')
+ pszStr++;
+
+ if (*pszStr != '"')
+ {
+ pszQ = (char *)pszStr;
+ while (*pszQ && *pszQ != ' ' && *pszQ != '\t')
+ pszQ++;
+ }
+ else
+ {
+ pszStr++;
+ pszQ = (char *)strchr(pszStr, '"');
+ if (pszQ == NULL)
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: incorrectly quoted value in descriptor in '%s' (raw value %s)"),
+ pImage->pszFilename, pszStart);
+ }
+
+ pszUnquoted = (char *)RTMemTmpAlloc(pszQ - pszStr + 1);
+ if (!pszUnquoted)
+ return VERR_NO_MEMORY;
+ memcpy(pszUnquoted, pszStr, pszQ - pszStr);
+ pszUnquoted[pszQ - pszStr] = '\0';
+ *ppszUnquoted = pszUnquoted;
+ if (ppszNext)
+ *ppszNext = pszQ + 1;
+ return VINF_SUCCESS;
+}
+
+static int vmdkDescInitStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
+ const char *pszLine)
+{
+ char *pEnd = pDescriptor->aLines[pDescriptor->cLines];
+ ssize_t cbDiff = strlen(pszLine) + 1;
+
+ if ( pDescriptor->cLines >= VMDK_DESCRIPTOR_LINES_MAX - 1
+ && pEnd - pDescriptor->aLines[0] > (ptrdiff_t)pDescriptor->cbDescAlloc - cbDiff)
+ return vdIfError(pImage->pIfError, VERR_BUFFER_OVERFLOW, RT_SRC_POS, N_("VMDK: descriptor too big in '%s'"), pImage->pszFilename);
+
+ memcpy(pEnd, pszLine, cbDiff);
+ pDescriptor->cLines++;
+ pDescriptor->aLines[pDescriptor->cLines] = pEnd + cbDiff;
+ pDescriptor->fDirty = true;
+
+ return VINF_SUCCESS;
+}
+
+static bool vmdkDescGetStr(PVMDKDESCRIPTOR pDescriptor, unsigned uStart,
+ const char *pszKey, const char **ppszValue)
+{
+ size_t cbKey = strlen(pszKey);
+ const char *pszValue;
+
+ while (uStart != 0)
+ {
+ if (!strncmp(pDescriptor->aLines[uStart], pszKey, cbKey))
+ {
+ /* Key matches, check for a '=' (preceded by whitespace). */
+ pszValue = pDescriptor->aLines[uStart] + cbKey;
+ while (*pszValue == ' ' || *pszValue == '\t')
+ pszValue++;
+ if (*pszValue == '=')
+ {
+ *ppszValue = pszValue + 1;
+ break;
+ }
+ }
+ uStart = pDescriptor->aNextLines[uStart];
+ }
+ return !!uStart;
+}
+
+static int vmdkDescSetStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
+ unsigned uStart,
+ const char *pszKey, const char *pszValue)
+{
+ char *pszTmp = NULL; /* (MSC naturally cannot figure this isn't used uninitialized) */
+ size_t cbKey = strlen(pszKey);
+ unsigned uLast = 0;
+
+ while (uStart != 0)
+ {
+ if (!strncmp(pDescriptor->aLines[uStart], pszKey, cbKey))
+ {
+ /* Key matches, check for a '=' (preceded by whitespace). */
+ pszTmp = pDescriptor->aLines[uStart] + cbKey;
+ while (*pszTmp == ' ' || *pszTmp == '\t')
+ pszTmp++;
+ if (*pszTmp == '=')
+ {
+ pszTmp++;
+ /** @todo r=bird: Doesn't skipping trailing blanks here just cause unecessary
+ * bloat and potentially out of space error? */
+ while (*pszTmp == ' ' || *pszTmp == '\t')
+ pszTmp++;
+ break;
+ }
+ }
+ if (!pDescriptor->aNextLines[uStart])
+ uLast = uStart;
+ uStart = pDescriptor->aNextLines[uStart];
+ }
+ if (uStart)
+ {
+ if (pszValue)
+ {
+ /* Key already exists, replace existing value. */
+ size_t cbOldVal = strlen(pszTmp);
+ size_t cbNewVal = strlen(pszValue);
+ ssize_t cbDiff = cbNewVal - cbOldVal;
+ /* Check for buffer overflow. */
+ if ( pDescriptor->aLines[pDescriptor->cLines] - pDescriptor->aLines[0]
+ > (ptrdiff_t)pDescriptor->cbDescAlloc - cbDiff)
+ return vdIfError(pImage->pIfError, VERR_BUFFER_OVERFLOW, RT_SRC_POS, N_("VMDK: descriptor too big in '%s'"), pImage->pszFilename);
+
+ memmove(pszTmp + cbNewVal, pszTmp + cbOldVal,
+ pDescriptor->aLines[pDescriptor->cLines] - pszTmp - cbOldVal);
+ memcpy(pszTmp, pszValue, cbNewVal + 1);
+ for (unsigned i = uStart + 1; i <= pDescriptor->cLines; i++)
+ pDescriptor->aLines[i] += cbDiff;
+ }
+ else
+ {
+ memmove(pDescriptor->aLines[uStart], pDescriptor->aLines[uStart+1],
+ pDescriptor->aLines[pDescriptor->cLines] - pDescriptor->aLines[uStart+1] + 1);
+ for (unsigned i = uStart + 1; i <= pDescriptor->cLines; i++)
+ {
+ pDescriptor->aLines[i-1] = pDescriptor->aLines[i];
+ if (pDescriptor->aNextLines[i])
+ pDescriptor->aNextLines[i-1] = pDescriptor->aNextLines[i] - 1;
+ else
+ pDescriptor->aNextLines[i-1] = 0;
+ }
+ pDescriptor->cLines--;
+ /* Adjust starting line numbers of following descriptor sections. */
+ if (uStart < pDescriptor->uFirstExtent)
+ pDescriptor->uFirstExtent--;
+ if (uStart < pDescriptor->uFirstDDB)
+ pDescriptor->uFirstDDB--;
+ }
+ }
+ else
+ {
+ /* Key doesn't exist, append after the last entry in this category. */
+ if (!pszValue)
+ {
+ /* Key doesn't exist, and it should be removed. Simply a no-op. */
+ return VINF_SUCCESS;
+ }
+ cbKey = strlen(pszKey);
+ size_t cbValue = strlen(pszValue);
+ ssize_t cbDiff = cbKey + 1 + cbValue + 1;
+ /* Check for buffer overflow. */
+ if ( (pDescriptor->cLines >= VMDK_DESCRIPTOR_LINES_MAX - 1)
+ || ( pDescriptor->aLines[pDescriptor->cLines]
+ - pDescriptor->aLines[0] > (ptrdiff_t)pDescriptor->cbDescAlloc - cbDiff))
+ return vdIfError(pImage->pIfError, VERR_BUFFER_OVERFLOW, RT_SRC_POS, N_("VMDK: descriptor too big in '%s'"), pImage->pszFilename);
+ for (unsigned i = pDescriptor->cLines + 1; i > uLast + 1; i--)
+ {
+ pDescriptor->aLines[i] = pDescriptor->aLines[i - 1];
+ if (pDescriptor->aNextLines[i - 1])
+ pDescriptor->aNextLines[i] = pDescriptor->aNextLines[i - 1] + 1;
+ else
+ pDescriptor->aNextLines[i] = 0;
+ }
+ uStart = uLast + 1;
+ pDescriptor->aNextLines[uLast] = uStart;
+ pDescriptor->aNextLines[uStart] = 0;
+ pDescriptor->cLines++;
+ pszTmp = pDescriptor->aLines[uStart];
+ memmove(pszTmp + cbDiff, pszTmp,
+ pDescriptor->aLines[pDescriptor->cLines] - pszTmp);
+ memcpy(pDescriptor->aLines[uStart], pszKey, cbKey);
+ pDescriptor->aLines[uStart][cbKey] = '=';
+ memcpy(pDescriptor->aLines[uStart] + cbKey + 1, pszValue, cbValue + 1);
+ for (unsigned i = uStart + 1; i <= pDescriptor->cLines; i++)
+ pDescriptor->aLines[i] += cbDiff;
+
+ /* Adjust starting line numbers of following descriptor sections. */
+ if (uStart <= pDescriptor->uFirstExtent)
+ pDescriptor->uFirstExtent++;
+ if (uStart <= pDescriptor->uFirstDDB)
+ pDescriptor->uFirstDDB++;
+ }
+ pDescriptor->fDirty = true;
+ return VINF_SUCCESS;
+}
+
+static int vmdkDescBaseGetU32(PVMDKDESCRIPTOR pDescriptor, const char *pszKey,
+ uint32_t *puValue)
+{
+ const char *pszValue;
+
+ if (!vmdkDescGetStr(pDescriptor, pDescriptor->uFirstDesc, pszKey,
+ &pszValue))
+ return VERR_VD_VMDK_VALUE_NOT_FOUND;
+ return RTStrToUInt32Ex(pszValue, NULL, 10, puValue);
+}
+
+/**
+ * Returns the value of the given key as a string allocating the necessary memory.
+ *
+ * @returns VBox status code.
+ * @retval VERR_VD_VMDK_VALUE_NOT_FOUND if the value could not be found.
+ * @param pImage The VMDK image state.
+ * @param pDescriptor The descriptor to fetch the value from.
+ * @param pszKey The key to get the value from.
+ * @param ppszValue Where to store the return value, use RTMemTmpFree to
+ * free.
+ */
+static int vmdkDescBaseGetStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
+ const char *pszKey, char **ppszValue)
+{
+ const char *pszValue;
+ char *pszValueUnquoted;
+
+ if (!vmdkDescGetStr(pDescriptor, pDescriptor->uFirstDesc, pszKey,
+ &pszValue))
+ return VERR_VD_VMDK_VALUE_NOT_FOUND;
+ int rc = vmdkStringUnquote(pImage, pszValue, &pszValueUnquoted, NULL);
+ if (RT_FAILURE(rc))
+ return rc;
+ *ppszValue = pszValueUnquoted;
+ return rc;
+}
+
+static int vmdkDescBaseSetStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
+ const char *pszKey, const char *pszValue)
+{
+ char *pszValueQuoted;
+
+ RTStrAPrintf(&pszValueQuoted, "\"%s\"", pszValue);
+ if (!pszValueQuoted)
+ return VERR_NO_STR_MEMORY;
+ int rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDesc, pszKey,
+ pszValueQuoted);
+ RTStrFree(pszValueQuoted);
+ return rc;
+}
+
+static void vmdkDescExtRemoveDummy(PVMDKIMAGE pImage,
+ PVMDKDESCRIPTOR pDescriptor)
+{
+ RT_NOREF1(pImage);
+ unsigned uEntry = pDescriptor->uFirstExtent;
+ ssize_t cbDiff;
+
+ if (!uEntry)
+ return;
+
+ cbDiff = strlen(pDescriptor->aLines[uEntry]) + 1;
+ /* Move everything including \0 in the entry marking the end of buffer. */
+ memmove(pDescriptor->aLines[uEntry], pDescriptor->aLines[uEntry + 1],
+ pDescriptor->aLines[pDescriptor->cLines] - pDescriptor->aLines[uEntry + 1] + 1);
+ for (unsigned i = uEntry + 1; i <= pDescriptor->cLines; i++)
+ {
+ pDescriptor->aLines[i - 1] = pDescriptor->aLines[i] - cbDiff;
+ if (pDescriptor->aNextLines[i])
+ pDescriptor->aNextLines[i - 1] = pDescriptor->aNextLines[i] - 1;
+ else
+ pDescriptor->aNextLines[i - 1] = 0;
+ }
+ pDescriptor->cLines--;
+ if (pDescriptor->uFirstDDB)
+ pDescriptor->uFirstDDB--;
+
+ return;
+}
+
+static int vmdkDescExtInsert(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
+ VMDKACCESS enmAccess, uint64_t cNominalSectors,
+ VMDKETYPE enmType, const char *pszBasename,
+ uint64_t uSectorOffset)
+{
+ static const char *apszAccess[] = { "NOACCESS", "RDONLY", "RW" };
+ static const char *apszType[] = { "", "SPARSE", "FLAT", "ZERO", "VMFS" };
+ char *pszTmp;
+ unsigned uStart = pDescriptor->uFirstExtent, uLast = 0;
+ char szExt[1024];
+ ssize_t cbDiff;
+
+ Assert((unsigned)enmAccess < RT_ELEMENTS(apszAccess));
+ Assert((unsigned)enmType < RT_ELEMENTS(apszType));
+
+ /* Find last entry in extent description. */
+ while (uStart)
+ {
+ if (!pDescriptor->aNextLines[uStart])
+ uLast = uStart;
+ uStart = pDescriptor->aNextLines[uStart];
+ }
+
+ if (enmType == VMDKETYPE_ZERO)
+ {
+ RTStrPrintf(szExt, sizeof(szExt), "%s %llu %s ", apszAccess[enmAccess],
+ cNominalSectors, apszType[enmType]);
+ }
+ else if (enmType == VMDKETYPE_FLAT)
+ {
+ RTStrPrintf(szExt, sizeof(szExt), "%s %llu %s \"%s\" %llu",
+ apszAccess[enmAccess], cNominalSectors,
+ apszType[enmType], pszBasename, uSectorOffset);
+ }
+ else
+ {
+ RTStrPrintf(szExt, sizeof(szExt), "%s %llu %s \"%s\"",
+ apszAccess[enmAccess], cNominalSectors,
+ apszType[enmType], pszBasename);
+ }
+ cbDiff = strlen(szExt) + 1;
+
+ /* Check for buffer overflow. */
+ if ( (pDescriptor->cLines >= VMDK_DESCRIPTOR_LINES_MAX - 1)
+ || ( pDescriptor->aLines[pDescriptor->cLines]
+ - pDescriptor->aLines[0] > (ptrdiff_t)pDescriptor->cbDescAlloc - cbDiff))
+ return vdIfError(pImage->pIfError, VERR_BUFFER_OVERFLOW, RT_SRC_POS, N_("VMDK: descriptor too big in '%s'"), pImage->pszFilename);
+
+ for (unsigned i = pDescriptor->cLines + 1; i > uLast + 1; i--)
+ {
+ pDescriptor->aLines[i] = pDescriptor->aLines[i - 1];
+ if (pDescriptor->aNextLines[i - 1])
+ pDescriptor->aNextLines[i] = pDescriptor->aNextLines[i - 1] + 1;
+ else
+ pDescriptor->aNextLines[i] = 0;
+ }
+ uStart = uLast + 1;
+ pDescriptor->aNextLines[uLast] = uStart;
+ pDescriptor->aNextLines[uStart] = 0;
+ pDescriptor->cLines++;
+ pszTmp = pDescriptor->aLines[uStart];
+ memmove(pszTmp + cbDiff, pszTmp,
+ pDescriptor->aLines[pDescriptor->cLines] - pszTmp);
+ memcpy(pDescriptor->aLines[uStart], szExt, cbDiff);
+ for (unsigned i = uStart + 1; i <= pDescriptor->cLines; i++)
+ pDescriptor->aLines[i] += cbDiff;
+
+ /* Adjust starting line numbers of following descriptor sections. */
+ if (uStart <= pDescriptor->uFirstDDB)
+ pDescriptor->uFirstDDB++;
+
+ pDescriptor->fDirty = true;
+ return VINF_SUCCESS;
+}
+
+/**
+ * Returns the value of the given key from the DDB as a string allocating
+ * the necessary memory.
+ *
+ * @returns VBox status code.
+ * @retval VERR_VD_VMDK_VALUE_NOT_FOUND if the value could not be found.
+ * @param pImage The VMDK image state.
+ * @param pDescriptor The descriptor to fetch the value from.
+ * @param pszKey The key to get the value from.
+ * @param ppszValue Where to store the return value, use RTMemTmpFree to
+ * free.
+ */
+static int vmdkDescDDBGetStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
+ const char *pszKey, char **ppszValue)
+{
+ const char *pszValue;
+ char *pszValueUnquoted;
+
+ if (!vmdkDescGetStr(pDescriptor, pDescriptor->uFirstDDB, pszKey,
+ &pszValue))
+ return VERR_VD_VMDK_VALUE_NOT_FOUND;
+ int rc = vmdkStringUnquote(pImage, pszValue, &pszValueUnquoted, NULL);
+ if (RT_FAILURE(rc))
+ return rc;
+ *ppszValue = pszValueUnquoted;
+ return rc;
+}
+
+static int vmdkDescDDBGetU32(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
+ const char *pszKey, uint32_t *puValue)
+{
+ const char *pszValue;
+ char *pszValueUnquoted;
+
+ if (!vmdkDescGetStr(pDescriptor, pDescriptor->uFirstDDB, pszKey,
+ &pszValue))
+ return VERR_VD_VMDK_VALUE_NOT_FOUND;
+ int rc = vmdkStringUnquote(pImage, pszValue, &pszValueUnquoted, NULL);
+ if (RT_FAILURE(rc))
+ return rc;
+ rc = RTStrToUInt32Ex(pszValueUnquoted, NULL, 10, puValue);
+ RTMemTmpFree(pszValueUnquoted);
+ return rc;
+}
+
+static int vmdkDescDDBGetUuid(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
+ const char *pszKey, PRTUUID pUuid)
+{
+ const char *pszValue;
+ char *pszValueUnquoted;
+
+ if (!vmdkDescGetStr(pDescriptor, pDescriptor->uFirstDDB, pszKey,
+ &pszValue))
+ return VERR_VD_VMDK_VALUE_NOT_FOUND;
+ int rc = vmdkStringUnquote(pImage, pszValue, &pszValueUnquoted, NULL);
+ if (RT_FAILURE(rc))
+ return rc;
+ rc = RTUuidFromStr(pUuid, pszValueUnquoted);
+ RTMemTmpFree(pszValueUnquoted);
+ return rc;
+}
+
+static int vmdkDescDDBSetStr(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
+ const char *pszKey, const char *pszVal)
+{
+ int rc;
+ char *pszValQuoted;
+
+ if (pszVal)
+ {
+ RTStrAPrintf(&pszValQuoted, "\"%s\"", pszVal);
+ if (!pszValQuoted)
+ return VERR_NO_STR_MEMORY;
+ }
+ else
+ pszValQuoted = NULL;
+ rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDDB, pszKey,
+ pszValQuoted);
+ if (pszValQuoted)
+ RTStrFree(pszValQuoted);
+ return rc;
+}
+
+static int vmdkDescDDBSetUuid(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
+ const char *pszKey, PCRTUUID pUuid)
+{
+ char *pszUuid;
+
+ RTStrAPrintf(&pszUuid, "\"%RTuuid\"", pUuid);
+ if (!pszUuid)
+ return VERR_NO_STR_MEMORY;
+ int rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDDB, pszKey,
+ pszUuid);
+ RTStrFree(pszUuid);
+ return rc;
+}
+
+static int vmdkDescDDBSetU32(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDescriptor,
+ const char *pszKey, uint32_t uValue)
+{
+ char *pszValue;
+
+ RTStrAPrintf(&pszValue, "\"%d\"", uValue);
+ if (!pszValue)
+ return VERR_NO_STR_MEMORY;
+ int rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDDB, pszKey,
+ pszValue);
+ RTStrFree(pszValue);
+ return rc;
+}
+
+/**
+ * Splits the descriptor data into individual lines checking for correct line
+ * endings and descriptor size.
+ *
+ * @returns VBox status code.
+ * @param pImage The image instance.
+ * @param pDesc The descriptor.
+ * @param pszTmp The raw descriptor data from the image.
+ */
+static int vmdkDescSplitLines(PVMDKIMAGE pImage, PVMDKDESCRIPTOR pDesc, char *pszTmp)
+{
+ unsigned cLine = 0;
+ int rc = VINF_SUCCESS;
+
+ while ( RT_SUCCESS(rc)
+ && *pszTmp != '\0')
+ {
+ pDesc->aLines[cLine++] = pszTmp;
+ if (cLine >= VMDK_DESCRIPTOR_LINES_MAX)
+ {
+ vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: descriptor too big in '%s'"), pImage->pszFilename);
+ rc = VERR_VD_VMDK_INVALID_HEADER;
+ break;
+ }
+
+ while (*pszTmp != '\0' && *pszTmp != '\n')
+ {
+ if (*pszTmp == '\r')
+ {
+ if (*(pszTmp + 1) != '\n')
+ {
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: unsupported end of line in descriptor in '%s'"), pImage->pszFilename);
+ break;
+ }
+ else
+ {
+ /* Get rid of CR character. */
+ *pszTmp = '\0';
+ }
+ }
+ pszTmp++;
+ }
+
+ if (RT_FAILURE(rc))
+ break;
+
+ /* Get rid of LF character. */
+ if (*pszTmp == '\n')
+ {
+ *pszTmp = '\0';
+ pszTmp++;
+ }
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ pDesc->cLines = cLine;
+ /* Pointer right after the end of the used part of the buffer. */
+ pDesc->aLines[cLine] = pszTmp;
+ }
+
+ return rc;
+}
+
+static int vmdkPreprocessDescriptor(PVMDKIMAGE pImage, char *pDescData,
+ size_t cbDescData, PVMDKDESCRIPTOR pDescriptor)
+{
+ pDescriptor->cbDescAlloc = cbDescData;
+ int rc = vmdkDescSplitLines(pImage, pDescriptor, pDescData);
+ if (RT_SUCCESS(rc))
+ {
+ if ( strcmp(pDescriptor->aLines[0], "# Disk DescriptorFile")
+ && strcmp(pDescriptor->aLines[0], "# Disk Descriptor File")
+ && strcmp(pDescriptor->aLines[0], "#Disk Descriptor File")
+ && strcmp(pDescriptor->aLines[0], "#Disk DescriptorFile"))
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: descriptor does not start as expected in '%s'"), pImage->pszFilename);
+ else
+ {
+ unsigned uLastNonEmptyLine = 0;
+
+ /* Initialize those, because we need to be able to reopen an image. */
+ pDescriptor->uFirstDesc = 0;
+ pDescriptor->uFirstExtent = 0;
+ pDescriptor->uFirstDDB = 0;
+ for (unsigned i = 0; i < pDescriptor->cLines; i++)
+ {
+ if (*pDescriptor->aLines[i] != '#' && *pDescriptor->aLines[i] != '\0')
+ {
+ if ( !strncmp(pDescriptor->aLines[i], "RW", 2)
+ || !strncmp(pDescriptor->aLines[i], "RDONLY", 6)
+ || !strncmp(pDescriptor->aLines[i], "NOACCESS", 8) )
+ {
+ /* An extent descriptor. */
+ if (!pDescriptor->uFirstDesc || pDescriptor->uFirstDDB)
+ {
+ /* Incorrect ordering of entries. */
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: incorrect ordering of entries in descriptor in '%s'"), pImage->pszFilename);
+ break;
+ }
+ if (!pDescriptor->uFirstExtent)
+ {
+ pDescriptor->uFirstExtent = i;
+ uLastNonEmptyLine = 0;
+ }
+ }
+ else if (!strncmp(pDescriptor->aLines[i], "ddb.", 4))
+ {
+ /* A disk database entry. */
+ if (!pDescriptor->uFirstDesc || !pDescriptor->uFirstExtent)
+ {
+ /* Incorrect ordering of entries. */
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: incorrect ordering of entries in descriptor in '%s'"), pImage->pszFilename);
+ break;
+ }
+ if (!pDescriptor->uFirstDDB)
+ {
+ pDescriptor->uFirstDDB = i;
+ uLastNonEmptyLine = 0;
+ }
+ }
+ else
+ {
+ /* A normal entry. */
+ if (pDescriptor->uFirstExtent || pDescriptor->uFirstDDB)
+ {
+ /* Incorrect ordering of entries. */
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: incorrect ordering of entries in descriptor in '%s'"), pImage->pszFilename);
+ break;
+ }
+ if (!pDescriptor->uFirstDesc)
+ {
+ pDescriptor->uFirstDesc = i;
+ uLastNonEmptyLine = 0;
+ }
+ }
+ if (uLastNonEmptyLine)
+ pDescriptor->aNextLines[uLastNonEmptyLine] = i;
+ uLastNonEmptyLine = i;
+ }
+ }
+ }
+ }
+
+ return rc;
+}
+
+static int vmdkDescSetPCHSGeometry(PVMDKIMAGE pImage,
+ PCVDGEOMETRY pPCHSGeometry)
+{
+ int rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_PCHS_CYLINDERS,
+ pPCHSGeometry->cCylinders);
+ if (RT_FAILURE(rc))
+ return rc;
+ rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_PCHS_HEADS,
+ pPCHSGeometry->cHeads);
+ if (RT_FAILURE(rc))
+ return rc;
+ rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_PCHS_SECTORS,
+ pPCHSGeometry->cSectors);
+ return rc;
+}
+
+static int vmdkDescSetLCHSGeometry(PVMDKIMAGE pImage,
+ PCVDGEOMETRY pLCHSGeometry)
+{
+ int rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_LCHS_CYLINDERS,
+ pLCHSGeometry->cCylinders);
+ if (RT_FAILURE(rc))
+ return rc;
+ rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_LCHS_HEADS,
+
+ pLCHSGeometry->cHeads);
+ if (RT_FAILURE(rc))
+ return rc;
+ rc = vmdkDescDDBSetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_LCHS_SECTORS,
+ pLCHSGeometry->cSectors);
+ return rc;
+}
+
+static int vmdkCreateDescriptor(PVMDKIMAGE pImage, char *pDescData,
+ size_t cbDescData, PVMDKDESCRIPTOR pDescriptor)
+{
+ pDescriptor->uFirstDesc = 0;
+ pDescriptor->uFirstExtent = 0;
+ pDescriptor->uFirstDDB = 0;
+ pDescriptor->cLines = 0;
+ pDescriptor->cbDescAlloc = cbDescData;
+ pDescriptor->fDirty = false;
+ pDescriptor->aLines[pDescriptor->cLines] = pDescData;
+ memset(pDescriptor->aNextLines, '\0', sizeof(pDescriptor->aNextLines));
+
+ int rc = vmdkDescInitStr(pImage, pDescriptor, "# Disk DescriptorFile");
+ if (RT_SUCCESS(rc))
+ rc = vmdkDescInitStr(pImage, pDescriptor, "version=1");
+ if (RT_SUCCESS(rc))
+ {
+ pDescriptor->uFirstDesc = pDescriptor->cLines - 1;
+ rc = vmdkDescInitStr(pImage, pDescriptor, "");
+ }
+ if (RT_SUCCESS(rc))
+ rc = vmdkDescInitStr(pImage, pDescriptor, "# Extent description");
+ if (RT_SUCCESS(rc))
+ rc = vmdkDescInitStr(pImage, pDescriptor, "NOACCESS 0 ZERO ");
+ if (RT_SUCCESS(rc))
+ {
+ pDescriptor->uFirstExtent = pDescriptor->cLines - 1;
+ rc = vmdkDescInitStr(pImage, pDescriptor, "");
+ }
+ if (RT_SUCCESS(rc))
+ {
+ /* The trailing space is created by VMware, too. */
+ rc = vmdkDescInitStr(pImage, pDescriptor, "# The disk Data Base ");
+ }
+ if (RT_SUCCESS(rc))
+ rc = vmdkDescInitStr(pImage, pDescriptor, "#DDB");
+ if (RT_SUCCESS(rc))
+ rc = vmdkDescInitStr(pImage, pDescriptor, "");
+ if (RT_SUCCESS(rc))
+ rc = vmdkDescInitStr(pImage, pDescriptor, "ddb.virtualHWVersion = \"4\"");
+ if (RT_SUCCESS(rc))
+ {
+ pDescriptor->uFirstDDB = pDescriptor->cLines - 1;
+
+ /* Now that the framework is in place, use the normal functions to insert
+ * the remaining keys. */
+ char szBuf[9];
+ RTStrPrintf(szBuf, sizeof(szBuf), "%08x", RTRandU32());
+ rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDesc,
+ "CID", szBuf);
+ }
+ if (RT_SUCCESS(rc))
+ rc = vmdkDescSetStr(pImage, pDescriptor, pDescriptor->uFirstDesc,
+ "parentCID", "ffffffff");
+ if (RT_SUCCESS(rc))
+ rc = vmdkDescDDBSetStr(pImage, pDescriptor, "ddb.adapterType", "ide");
+
+ return rc;
+}
+
+static int vmdkParseDescriptor(PVMDKIMAGE pImage, char *pDescData, size_t cbDescData)
+{
+ int rc;
+ unsigned cExtents;
+ unsigned uLine;
+ unsigned i;
+
+ rc = vmdkPreprocessDescriptor(pImage, pDescData, cbDescData,
+ &pImage->Descriptor);
+ if (RT_FAILURE(rc))
+ return rc;
+
+ /* Check version, must be 1. */
+ uint32_t uVersion;
+ rc = vmdkDescBaseGetU32(&pImage->Descriptor, "version", &uVersion);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error finding key 'version' in descriptor in '%s'"), pImage->pszFilename);
+ if (uVersion != 1)
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_UNSUPPORTED_VERSION, RT_SRC_POS, N_("VMDK: unsupported format version in descriptor in '%s'"), pImage->pszFilename);
+
+ /* Get image creation type and determine image flags. */
+ char *pszCreateType = NULL; /* initialized to make gcc shut up */
+ rc = vmdkDescBaseGetStr(pImage, &pImage->Descriptor, "createType",
+ &pszCreateType);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot get image type from descriptor in '%s'"), pImage->pszFilename);
+ if ( !strcmp(pszCreateType, "twoGbMaxExtentSparse")
+ || !strcmp(pszCreateType, "twoGbMaxExtentFlat"))
+ pImage->uImageFlags |= VD_VMDK_IMAGE_FLAGS_SPLIT_2G;
+ else if ( !strcmp(pszCreateType, "partitionedDevice")
+ || !strcmp(pszCreateType, "fullDevice"))
+ pImage->uImageFlags |= VD_VMDK_IMAGE_FLAGS_RAWDISK;
+ else if (!strcmp(pszCreateType, "streamOptimized"))
+ pImage->uImageFlags |= VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED;
+ else if (!strcmp(pszCreateType, "vmfs"))
+ pImage->uImageFlags |= VD_IMAGE_FLAGS_FIXED | VD_VMDK_IMAGE_FLAGS_ESX;
+ RTMemTmpFree(pszCreateType);
+
+ /* Count the number of extent config entries. */
+ for (uLine = pImage->Descriptor.uFirstExtent, cExtents = 0;
+ uLine != 0;
+ uLine = pImage->Descriptor.aNextLines[uLine], cExtents++)
+ /* nothing */;
+
+ if (!pImage->pDescData && cExtents != 1)
+ {
+ /* Monolithic image, must have only one extent (already opened). */
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: monolithic image may only have one extent in '%s'"), pImage->pszFilename);
+ }
+
+ if (pImage->pDescData)
+ {
+ /* Non-monolithic image, extents need to be allocated. */
+ rc = vmdkCreateExtents(pImage, cExtents);
+ if (RT_FAILURE(rc))
+ return rc;
+ }
+
+ for (i = 0, uLine = pImage->Descriptor.uFirstExtent;
+ i < cExtents; i++, uLine = pImage->Descriptor.aNextLines[uLine])
+ {
+ char *pszLine = pImage->Descriptor.aLines[uLine];
+
+ /* Access type of the extent. */
+ if (!strncmp(pszLine, "RW", 2))
+ {
+ pImage->pExtents[i].enmAccess = VMDKACCESS_READWRITE;
+ pszLine += 2;
+ }
+ else if (!strncmp(pszLine, "RDONLY", 6))
+ {
+ pImage->pExtents[i].enmAccess = VMDKACCESS_READONLY;
+ pszLine += 6;
+ }
+ else if (!strncmp(pszLine, "NOACCESS", 8))
+ {
+ pImage->pExtents[i].enmAccess = VMDKACCESS_NOACCESS;
+ pszLine += 8;
+ }
+ else
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
+ if (*pszLine++ != ' ')
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
+
+ /* Nominal size of the extent. */
+ rc = RTStrToUInt64Ex(pszLine, &pszLine, 10,
+ &pImage->pExtents[i].cNominalSectors);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
+ if (*pszLine++ != ' ')
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
+
+ /* Type of the extent. */
+ if (!strncmp(pszLine, "SPARSE", 6))
+ {
+ pImage->pExtents[i].enmType = VMDKETYPE_HOSTED_SPARSE;
+ pszLine += 6;
+ }
+ else if (!strncmp(pszLine, "FLAT", 4))
+ {
+ pImage->pExtents[i].enmType = VMDKETYPE_FLAT;
+ pszLine += 4;
+ }
+ else if (!strncmp(pszLine, "ZERO", 4))
+ {
+ pImage->pExtents[i].enmType = VMDKETYPE_ZERO;
+ pszLine += 4;
+ }
+ else if (!strncmp(pszLine, "VMFS", 4))
+ {
+ pImage->pExtents[i].enmType = VMDKETYPE_VMFS;
+ pszLine += 4;
+ }
+ else
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
+
+ if (pImage->pExtents[i].enmType == VMDKETYPE_ZERO)
+ {
+ /* This one has no basename or offset. */
+ if (*pszLine == ' ')
+ pszLine++;
+ if (*pszLine != '\0')
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
+ pImage->pExtents[i].pszBasename = NULL;
+ }
+ else
+ {
+ /* All other extent types have basename and optional offset. */
+ if (*pszLine++ != ' ')
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
+
+ /* Basename of the image. Surrounded by quotes. */
+ char *pszBasename;
+ rc = vmdkStringUnquote(pImage, pszLine, &pszBasename, &pszLine);
+ if (RT_FAILURE(rc))
+ return rc;
+ pImage->pExtents[i].pszBasename = pszBasename;
+ if (*pszLine == ' ')
+ {
+ pszLine++;
+ if (*pszLine != '\0')
+ {
+ /* Optional offset in extent specified. */
+ rc = RTStrToUInt64Ex(pszLine, &pszLine, 10,
+ &pImage->pExtents[i].uSectorOffset);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
+ }
+ }
+
+ if (*pszLine != '\0')
+ return vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: parse error in extent description in '%s'"), pImage->pszFilename);
+ }
+ }
+
+ /* Determine PCHS geometry (autogenerate if necessary). */
+ rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_PCHS_CYLINDERS,
+ &pImage->PCHSGeometry.cCylinders);
+ if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
+ pImage->PCHSGeometry.cCylinders = 0;
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting PCHS geometry from extent description in '%s'"), pImage->pszFilename);
+ rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_PCHS_HEADS,
+ &pImage->PCHSGeometry.cHeads);
+ if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
+ pImage->PCHSGeometry.cHeads = 0;
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting PCHS geometry from extent description in '%s'"), pImage->pszFilename);
+ rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_PCHS_SECTORS,
+ &pImage->PCHSGeometry.cSectors);
+ if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
+ pImage->PCHSGeometry.cSectors = 0;
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting PCHS geometry from extent description in '%s'"), pImage->pszFilename);
+ if ( pImage->PCHSGeometry.cCylinders == 0
+ || pImage->PCHSGeometry.cHeads == 0
+ || pImage->PCHSGeometry.cHeads > 16
+ || pImage->PCHSGeometry.cSectors == 0
+ || pImage->PCHSGeometry.cSectors > 63)
+ {
+ /* Mark PCHS geometry as not yet valid (can't do the calculation here
+ * as the total image size isn't known yet). */
+ pImage->PCHSGeometry.cCylinders = 0;
+ pImage->PCHSGeometry.cHeads = 16;
+ pImage->PCHSGeometry.cSectors = 63;
+ }
+
+ /* Determine LCHS geometry (set to 0 if not specified). */
+ rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_LCHS_CYLINDERS,
+ &pImage->LCHSGeometry.cCylinders);
+ if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
+ pImage->LCHSGeometry.cCylinders = 0;
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting LCHS geometry from extent description in '%s'"), pImage->pszFilename);
+ rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_LCHS_HEADS,
+ &pImage->LCHSGeometry.cHeads);
+ if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
+ pImage->LCHSGeometry.cHeads = 0;
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting LCHS geometry from extent description in '%s'"), pImage->pszFilename);
+ rc = vmdkDescDDBGetU32(pImage, &pImage->Descriptor,
+ VMDK_DDB_GEO_LCHS_SECTORS,
+ &pImage->LCHSGeometry.cSectors);
+ if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
+ pImage->LCHSGeometry.cSectors = 0;
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting LCHS geometry from extent description in '%s'"), pImage->pszFilename);
+ if ( pImage->LCHSGeometry.cCylinders == 0
+ || pImage->LCHSGeometry.cHeads == 0
+ || pImage->LCHSGeometry.cSectors == 0)
+ {
+ pImage->LCHSGeometry.cCylinders = 0;
+ pImage->LCHSGeometry.cHeads = 0;
+ pImage->LCHSGeometry.cSectors = 0;
+ }
+
+ /* Get image UUID. */
+ rc = vmdkDescDDBGetUuid(pImage, &pImage->Descriptor, VMDK_DDB_IMAGE_UUID,
+ &pImage->ImageUuid);
+ if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
+ {
+ /* Image without UUID. Probably created by VMware and not yet used
+ * by VirtualBox. Can only be added for images opened in read/write
+ * mode, so don't bother producing a sensible UUID otherwise. */
+ if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ RTUuidClear(&pImage->ImageUuid);
+ else
+ {
+ rc = RTUuidCreate(&pImage->ImageUuid);
+ if (RT_FAILURE(rc))
+ return rc;
+ rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
+ VMDK_DDB_IMAGE_UUID, &pImage->ImageUuid);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing image UUID in descriptor in '%s'"), pImage->pszFilename);
+ }
+ }
+ else if (RT_FAILURE(rc))
+ return rc;
+
+ /* Get image modification UUID. */
+ rc = vmdkDescDDBGetUuid(pImage, &pImage->Descriptor,
+ VMDK_DDB_MODIFICATION_UUID,
+ &pImage->ModificationUuid);
+ if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
+ {
+ /* Image without UUID. Probably created by VMware and not yet used
+ * by VirtualBox. Can only be added for images opened in read/write
+ * mode, so don't bother producing a sensible UUID otherwise. */
+ if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ RTUuidClear(&pImage->ModificationUuid);
+ else
+ {
+ rc = RTUuidCreate(&pImage->ModificationUuid);
+ if (RT_FAILURE(rc))
+ return rc;
+ rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
+ VMDK_DDB_MODIFICATION_UUID,
+ &pImage->ModificationUuid);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing image modification UUID in descriptor in '%s'"), pImage->pszFilename);
+ }
+ }
+ else if (RT_FAILURE(rc))
+ return rc;
+
+ /* Get UUID of parent image. */
+ rc = vmdkDescDDBGetUuid(pImage, &pImage->Descriptor, VMDK_DDB_PARENT_UUID,
+ &pImage->ParentUuid);
+ if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
+ {
+ /* Image without UUID. Probably created by VMware and not yet used
+ * by VirtualBox. Can only be added for images opened in read/write
+ * mode, so don't bother producing a sensible UUID otherwise. */
+ if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ RTUuidClear(&pImage->ParentUuid);
+ else
+ {
+ rc = RTUuidClear(&pImage->ParentUuid);
+ if (RT_FAILURE(rc))
+ return rc;
+ rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
+ VMDK_DDB_PARENT_UUID, &pImage->ParentUuid);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing parent UUID in descriptor in '%s'"), pImage->pszFilename);
+ }
+ }
+ else if (RT_FAILURE(rc))
+ return rc;
+
+ /* Get parent image modification UUID. */
+ rc = vmdkDescDDBGetUuid(pImage, &pImage->Descriptor,
+ VMDK_DDB_PARENT_MODIFICATION_UUID,
+ &pImage->ParentModificationUuid);
+ if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
+ {
+ /* Image without UUID. Probably created by VMware and not yet used
+ * by VirtualBox. Can only be added for images opened in read/write
+ * mode, so don't bother producing a sensible UUID otherwise. */
+ if (pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ RTUuidClear(&pImage->ParentModificationUuid);
+ else
+ {
+ RTUuidClear(&pImage->ParentModificationUuid);
+ rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
+ VMDK_DDB_PARENT_MODIFICATION_UUID,
+ &pImage->ParentModificationUuid);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing parent modification UUID in descriptor in '%s'"), pImage->pszFilename);
+ }
+ }
+ else if (RT_FAILURE(rc))
+ return rc;
+
+ return VINF_SUCCESS;
+}
+
+/**
+ * Internal : Prepares the descriptor to write to the image.
+ */
+static int vmdkDescriptorPrepare(PVMDKIMAGE pImage, uint64_t cbLimit,
+ void **ppvData, size_t *pcbData)
+{
+ int rc = VINF_SUCCESS;
+
+ /*
+ * Allocate temporary descriptor buffer.
+ * In case there is no limit allocate a default
+ * and increase if required.
+ */
+ size_t cbDescriptor = cbLimit ? cbLimit : 4 * _1K;
+ char *pszDescriptor = (char *)RTMemAllocZ(cbDescriptor);
+ size_t offDescriptor = 0;
+
+ if (!pszDescriptor)
+ return VERR_NO_MEMORY;
+
+ for (unsigned i = 0; i < pImage->Descriptor.cLines; i++)
+ {
+ const char *psz = pImage->Descriptor.aLines[i];
+ size_t cb = strlen(psz);
+
+ /*
+ * Increase the descriptor if there is no limit and
+ * there is not enough room left for this line.
+ */
+ if (offDescriptor + cb + 1 > cbDescriptor)
+ {
+ if (cbLimit)
+ {
+ rc = vdIfError(pImage->pIfError, VERR_BUFFER_OVERFLOW, RT_SRC_POS, N_("VMDK: descriptor too long in '%s'"), pImage->pszFilename);
+ break;
+ }
+ else
+ {
+ char *pszDescriptorNew = NULL;
+ LogFlow(("Increasing descriptor cache\n"));
+
+ pszDescriptorNew = (char *)RTMemRealloc(pszDescriptor, cbDescriptor + cb + 4 * _1K);
+ if (!pszDescriptorNew)
+ {
+ rc = VERR_NO_MEMORY;
+ break;
+ }
+ pszDescriptor = pszDescriptorNew;
+ cbDescriptor += cb + 4 * _1K;
+ }
+ }
+
+ if (cb > 0)
+ {
+ memcpy(pszDescriptor + offDescriptor, psz, cb);
+ offDescriptor += cb;
+ }
+
+ memcpy(pszDescriptor + offDescriptor, "\n", 1);
+ offDescriptor++;
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ *ppvData = pszDescriptor;
+ *pcbData = offDescriptor;
+ }
+ else if (pszDescriptor)
+ RTMemFree(pszDescriptor);
+
+ return rc;
+}
+
+/**
+ * Internal: write/update the descriptor part of the image.
+ */
+static int vmdkWriteDescriptor(PVMDKIMAGE pImage, PVDIOCTX pIoCtx)
+{
+ int rc = VINF_SUCCESS;
+ uint64_t cbLimit;
+ uint64_t uOffset;
+ PVMDKFILE pDescFile;
+ void *pvDescriptor = NULL;
+ size_t cbDescriptor;
+
+ if (pImage->pDescData)
+ {
+ /* Separate descriptor file. */
+ uOffset = 0;
+ cbLimit = 0;
+ pDescFile = pImage->pFile;
+ }
+ else
+ {
+ /* Embedded descriptor file. */
+ uOffset = VMDK_SECTOR2BYTE(pImage->pExtents[0].uDescriptorSector);
+ cbLimit = VMDK_SECTOR2BYTE(pImage->pExtents[0].cDescriptorSectors);
+ pDescFile = pImage->pExtents[0].pFile;
+ }
+ /* Bail out if there is no file to write to. */
+ if (pDescFile == NULL)
+ return VERR_INVALID_PARAMETER;
+
+ rc = vmdkDescriptorPrepare(pImage, cbLimit, &pvDescriptor, &cbDescriptor);
+ if (RT_SUCCESS(rc))
+ {
+ rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pDescFile->pStorage,
+ uOffset, pvDescriptor,
+ cbLimit ? cbLimit : cbDescriptor,
+ pIoCtx, NULL, NULL);
+ if ( RT_FAILURE(rc)
+ && rc != VERR_VD_ASYNC_IO_IN_PROGRESS)
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error writing descriptor in '%s'"), pImage->pszFilename);
+ }
+
+ if (RT_SUCCESS(rc) && !cbLimit)
+ {
+ rc = vdIfIoIntFileSetSize(pImage->pIfIo, pDescFile->pStorage, cbDescriptor);
+ if (RT_FAILURE(rc))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error truncating descriptor in '%s'"), pImage->pszFilename);
+ }
+
+ if (RT_SUCCESS(rc))
+ pImage->Descriptor.fDirty = false;
+
+ if (pvDescriptor)
+ RTMemFree(pvDescriptor);
+ return rc;
+
+}
+
+/**
+ * Internal: validate the consistency check values in a binary header.
+ */
+static int vmdkValidateHeader(PVMDKIMAGE pImage, PVMDKEXTENT pExtent, const SparseExtentHeader *pHeader)
+{
+ int rc = VINF_SUCCESS;
+ if (RT_LE2H_U32(pHeader->magicNumber) != VMDK_SPARSE_MAGICNUMBER)
+ {
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: incorrect magic in sparse extent header in '%s'"), pExtent->pszFullname);
+ return rc;
+ }
+ if (RT_LE2H_U32(pHeader->version) != 1 && RT_LE2H_U32(pHeader->version) != 3)
+ {
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_UNSUPPORTED_VERSION, RT_SRC_POS, N_("VMDK: incorrect version in sparse extent header in '%s', not a VMDK 1.0/1.1 conforming file"), pExtent->pszFullname);
+ return rc;
+ }
+ if ( (RT_LE2H_U32(pHeader->flags) & 1)
+ && ( pHeader->singleEndLineChar != '\n'
+ || pHeader->nonEndLineChar != ' '
+ || pHeader->doubleEndLineChar1 != '\r'
+ || pHeader->doubleEndLineChar2 != '\n') )
+ {
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: corrupted by CR/LF translation in '%s'"), pExtent->pszFullname);
+ return rc;
+ }
+ if (RT_LE2H_U64(pHeader->descriptorSize) > VMDK_SPARSE_DESCRIPTOR_SIZE_MAX)
+ {
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: descriptor size out of bounds (%llu vs %llu) '%s'"),
+ pExtent->pszFullname, RT_LE2H_U64(pHeader->descriptorSize), VMDK_SPARSE_DESCRIPTOR_SIZE_MAX);
+ return rc;
+ }
+ return rc;
+}
+
+/**
+ * Internal: read metadata belonging to an extent with binary header, i.e.
+ * as found in monolithic files.
+ */
+static int vmdkReadBinaryMetaExtent(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
+ bool fMagicAlreadyRead)
+{
+ SparseExtentHeader Header;
+ int rc;
+
+ if (!fMagicAlreadyRead)
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage, 0,
+ &Header, sizeof(Header));
+ else
+ {
+ Header.magicNumber = RT_H2LE_U32(VMDK_SPARSE_MAGICNUMBER);
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ RT_UOFFSETOF(SparseExtentHeader, version),
+ &Header.version,
+ sizeof(Header)
+ - RT_UOFFSETOF(SparseExtentHeader, version));
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ rc = vmdkValidateHeader(pImage, pExtent, &Header);
+ if (RT_SUCCESS(rc))
+ {
+ uint64_t cbFile = 0;
+
+ if ( (RT_LE2H_U32(Header.flags) & RT_BIT(17))
+ && RT_LE2H_U64(Header.gdOffset) == VMDK_GD_AT_END)
+ pExtent->fFooter = true;
+
+ if ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ || ( pExtent->fFooter
+ && !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL)))
+ {
+ rc = vdIfIoIntFileGetSize(pImage->pIfIo, pExtent->pFile->pStorage, &cbFile);
+ if (RT_FAILURE(rc))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot get size of '%s'"), pExtent->pszFullname);
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ pExtent->uAppendPosition = RT_ALIGN_64(cbFile, 512);
+
+ if ( pExtent->fFooter
+ && ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ || !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL)))
+ {
+ /* Read the footer, which comes before the end-of-stream marker. */
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ cbFile - 2*512, &Header,
+ sizeof(Header));
+ if (RT_FAILURE(rc))
+ {
+ vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error reading extent footer in '%s'"), pExtent->pszFullname);
+ rc = VERR_VD_VMDK_INVALID_HEADER;
+ }
+
+ if (RT_SUCCESS(rc))
+ rc = vmdkValidateHeader(pImage, pExtent, &Header);
+ /* Prohibit any writes to this extent. */
+ pExtent->uAppendPosition = 0;
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ pExtent->uVersion = RT_LE2H_U32(Header.version);
+ pExtent->enmType = VMDKETYPE_HOSTED_SPARSE; /* Just dummy value, changed later. */
+ pExtent->cSectors = RT_LE2H_U64(Header.capacity);
+ pExtent->cSectorsPerGrain = RT_LE2H_U64(Header.grainSize);
+ pExtent->uDescriptorSector = RT_LE2H_U64(Header.descriptorOffset);
+ pExtent->cDescriptorSectors = RT_LE2H_U64(Header.descriptorSize);
+ pExtent->cGTEntries = RT_LE2H_U32(Header.numGTEsPerGT);
+ pExtent->cOverheadSectors = RT_LE2H_U64(Header.overHead);
+ pExtent->fUncleanShutdown = !!Header.uncleanShutdown;
+ pExtent->uCompression = RT_LE2H_U16(Header.compressAlgorithm);
+ if (RT_LE2H_U32(Header.flags) & RT_BIT(1))
+ {
+ pExtent->uSectorRGD = RT_LE2H_U64(Header.rgdOffset);
+ pExtent->uSectorGD = RT_LE2H_U64(Header.gdOffset);
+ }
+ else
+ {
+ pExtent->uSectorGD = RT_LE2H_U64(Header.gdOffset);
+ pExtent->uSectorRGD = 0;
+ }
+
+ if (pExtent->uDescriptorSector && !pExtent->cDescriptorSectors)
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: inconsistent embedded descriptor config in '%s'"), pExtent->pszFullname);
+
+ if ( RT_SUCCESS(rc)
+ && ( pExtent->uSectorGD == VMDK_GD_AT_END
+ || pExtent->uSectorRGD == VMDK_GD_AT_END)
+ && ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ || !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL)))
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: cannot resolve grain directory offset in '%s'"), pExtent->pszFullname);
+
+ if (RT_SUCCESS(rc))
+ {
+ uint64_t cSectorsPerGDE = pExtent->cGTEntries * pExtent->cSectorsPerGrain;
+ if (!cSectorsPerGDE || cSectorsPerGDE > UINT32_MAX)
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: incorrect grain directory size in '%s'"), pExtent->pszFullname);
+ else
+ {
+ pExtent->cSectorsPerGDE = cSectorsPerGDE;
+ pExtent->cGDEntries = (pExtent->cSectors + cSectorsPerGDE - 1) / cSectorsPerGDE;
+
+ /* Fix up the number of descriptor sectors, as some flat images have
+ * really just one, and this causes failures when inserting the UUID
+ * values and other extra information. */
+ if (pExtent->cDescriptorSectors != 0 && pExtent->cDescriptorSectors < 4)
+ {
+ /* Do it the easy way - just fix it for flat images which have no
+ * other complicated metadata which needs space too. */
+ if ( pExtent->uDescriptorSector + 4 < pExtent->cOverheadSectors
+ && pExtent->cGTEntries * pExtent->cGDEntries == 0)
+ pExtent->cDescriptorSectors = 4;
+ }
+ }
+ }
+ }
+ }
+ }
+ }
+ else
+ {
+ vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error reading extent header in '%s'"), pExtent->pszFullname);
+ rc = VERR_VD_VMDK_INVALID_HEADER;
+ }
+
+ if (RT_FAILURE(rc))
+ vmdkFreeExtentData(pImage, pExtent, false);
+
+ return rc;
+}
+
+/**
+ * Internal: read additional metadata belonging to an extent. For those
+ * extents which have no additional metadata just verify the information.
+ */
+static int vmdkReadMetaExtent(PVMDKIMAGE pImage, PVMDKEXTENT pExtent)
+{
+ int rc = VINF_SUCCESS;
+
+/* disabled the check as there are too many truncated vmdk images out there */
+#ifdef VBOX_WITH_VMDK_STRICT_SIZE_CHECK
+ uint64_t cbExtentSize;
+ /* The image must be a multiple of a sector in size and contain the data
+ * area (flat images only). If not, it means the image is at least
+ * truncated, or even seriously garbled. */
+ rc = vdIfIoIntFileGetSize(pImage->pIfIo, pExtent->pFile->pStorage, &cbExtentSize);
+ if (RT_FAILURE(rc))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error getting size in '%s'"), pExtent->pszFullname);
+ else if ( cbExtentSize != RT_ALIGN_64(cbExtentSize, 512)
+ && (pExtent->enmType != VMDKETYPE_FLAT || pExtent->cNominalSectors + pExtent->uSectorOffset > VMDK_BYTE2SECTOR(cbExtentSize)))
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: file size is not a multiple of 512 in '%s', file is truncated or otherwise garbled"), pExtent->pszFullname);
+#endif /* VBOX_WITH_VMDK_STRICT_SIZE_CHECK */
+ if ( RT_SUCCESS(rc)
+ && pExtent->enmType == VMDKETYPE_HOSTED_SPARSE)
+ {
+ /* The spec says that this must be a power of two and greater than 8,
+ * but probably they meant not less than 8. */
+ if ( (pExtent->cSectorsPerGrain & (pExtent->cSectorsPerGrain - 1))
+ || pExtent->cSectorsPerGrain < 8)
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: invalid extent grain size %u in '%s'"), pExtent->cSectorsPerGrain, pExtent->pszFullname);
+ else
+ {
+ /* This code requires that a grain table must hold a power of two multiple
+ * of the number of entries per GT cache entry. */
+ if ( (pExtent->cGTEntries & (pExtent->cGTEntries - 1))
+ || pExtent->cGTEntries < VMDK_GT_CACHELINE_SIZE)
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS,
+ N_("VMDK: grain table cache size problem in '%s'"), pExtent->pszFullname);
+ else
+ {
+ rc = vmdkAllocStreamBuffers(pImage, pExtent);
+ if (RT_SUCCESS(rc))
+ {
+ /* Prohibit any writes to this streamOptimized extent. */
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ pExtent->uAppendPosition = 0;
+
+ if ( !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ || !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ || !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL))
+ rc = vmdkReadGrainDirectory(pImage, pExtent);
+ else
+ {
+ pExtent->uGrainSectorAbs = pExtent->cOverheadSectors;
+ pExtent->cbGrainStreamRead = 0;
+ }
+ }
+ }
+ }
+ }
+
+ if (RT_FAILURE(rc))
+ vmdkFreeExtentData(pImage, pExtent, false);
+
+ return rc;
+}
+
+/**
+ * Internal: write/update the metadata for a sparse extent.
+ */
+static int vmdkWriteMetaSparseExtent(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
+ uint64_t uOffset, PVDIOCTX pIoCtx)
+{
+ SparseExtentHeader Header;
+
+ memset(&Header, '\0', sizeof(Header));
+ Header.magicNumber = RT_H2LE_U32(VMDK_SPARSE_MAGICNUMBER);
+ Header.version = RT_H2LE_U32(pExtent->uVersion);
+ Header.flags = RT_H2LE_U32(RT_BIT(0));
+ if (pExtent->pRGD)
+ Header.flags |= RT_H2LE_U32(RT_BIT(1));
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ Header.flags |= RT_H2LE_U32(RT_BIT(16) | RT_BIT(17));
+ Header.capacity = RT_H2LE_U64(pExtent->cSectors);
+ Header.grainSize = RT_H2LE_U64(pExtent->cSectorsPerGrain);
+ Header.descriptorOffset = RT_H2LE_U64(pExtent->uDescriptorSector);
+ Header.descriptorSize = RT_H2LE_U64(pExtent->cDescriptorSectors);
+ Header.numGTEsPerGT = RT_H2LE_U32(pExtent->cGTEntries);
+ if (pExtent->fFooter && uOffset == 0)
+ {
+ if (pExtent->pRGD)
+ {
+ Assert(pExtent->uSectorRGD);
+ Header.rgdOffset = RT_H2LE_U64(VMDK_GD_AT_END);
+ Header.gdOffset = RT_H2LE_U64(VMDK_GD_AT_END);
+ }
+ else
+ Header.gdOffset = RT_H2LE_U64(VMDK_GD_AT_END);
+ }
+ else
+ {
+ if (pExtent->pRGD)
+ {
+ Assert(pExtent->uSectorRGD);
+ Header.rgdOffset = RT_H2LE_U64(pExtent->uSectorRGD);
+ Header.gdOffset = RT_H2LE_U64(pExtent->uSectorGD);
+ }
+ else
+ Header.gdOffset = RT_H2LE_U64(pExtent->uSectorGD);
+ }
+ Header.overHead = RT_H2LE_U64(pExtent->cOverheadSectors);
+ Header.uncleanShutdown = pExtent->fUncleanShutdown;
+ Header.singleEndLineChar = '\n';
+ Header.nonEndLineChar = ' ';
+ Header.doubleEndLineChar1 = '\r';
+ Header.doubleEndLineChar2 = '\n';
+ Header.compressAlgorithm = RT_H2LE_U16(pExtent->uCompression);
+
+ int rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
+ uOffset, &Header, sizeof(Header),
+ pIoCtx, NULL, NULL);
+ if (RT_FAILURE(rc) && (rc != VERR_VD_ASYNC_IO_IN_PROGRESS))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error writing extent header in '%s'"), pExtent->pszFullname);
+ return rc;
+}
+
+/**
+ * Internal: free the buffers used for streamOptimized images.
+ */
+static void vmdkFreeStreamBuffers(PVMDKEXTENT pExtent)
+{
+ if (pExtent->pvCompGrain)
+ {
+ RTMemFree(pExtent->pvCompGrain);
+ pExtent->pvCompGrain = NULL;
+ }
+ if (pExtent->pvGrain)
+ {
+ RTMemFree(pExtent->pvGrain);
+ pExtent->pvGrain = NULL;
+ }
+}
+
+/**
+ * Internal: free the memory used by the extent data structure, optionally
+ * deleting the referenced files.
+ *
+ * @returns VBox status code.
+ * @param pImage Pointer to the image instance data.
+ * @param pExtent The extent to free.
+ * @param fDelete Flag whether to delete the backing storage.
+ */
+static int vmdkFreeExtentData(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
+ bool fDelete)
+{
+ int rc = VINF_SUCCESS;
+
+ vmdkFreeGrainDirectory(pExtent);
+ if (pExtent->pDescData)
+ {
+ RTMemFree(pExtent->pDescData);
+ pExtent->pDescData = NULL;
+ }
+ if (pExtent->pFile != NULL)
+ {
+ /* Do not delete raw extents, these have full and base names equal. */
+ rc = vmdkFileClose(pImage, &pExtent->pFile,
+ fDelete
+ && pExtent->pszFullname
+ && pExtent->pszBasename
+ && strcmp(pExtent->pszFullname, pExtent->pszBasename));
+ }
+ if (pExtent->pszBasename)
+ {
+ RTMemTmpFree((void *)pExtent->pszBasename);
+ pExtent->pszBasename = NULL;
+ }
+ if (pExtent->pszFullname)
+ {
+ RTStrFree((char *)(void *)pExtent->pszFullname);
+ pExtent->pszFullname = NULL;
+ }
+ vmdkFreeStreamBuffers(pExtent);
+
+ return rc;
+}
+
+/**
+ * Internal: allocate grain table cache if necessary for this image.
+ */
+static int vmdkAllocateGrainTableCache(PVMDKIMAGE pImage)
+{
+ PVMDKEXTENT pExtent;
+
+ /* Allocate grain table cache if any sparse extent is present. */
+ for (unsigned i = 0; i < pImage->cExtents; i++)
+ {
+ pExtent = &pImage->pExtents[i];
+ if (pExtent->enmType == VMDKETYPE_HOSTED_SPARSE)
+ {
+ /* Allocate grain table cache. */
+ pImage->pGTCache = (PVMDKGTCACHE)RTMemAllocZ(sizeof(VMDKGTCACHE));
+ if (!pImage->pGTCache)
+ return VERR_NO_MEMORY;
+ for (unsigned j = 0; j < VMDK_GT_CACHE_SIZE; j++)
+ {
+ PVMDKGTCACHEENTRY pGCE = &pImage->pGTCache->aGTCache[j];
+ pGCE->uExtent = UINT32_MAX;
+ }
+ pImage->pGTCache->cEntries = VMDK_GT_CACHE_SIZE;
+ break;
+ }
+ }
+
+ return VINF_SUCCESS;
+}
+
+/**
+ * Internal: allocate the given number of extents.
+ */
+static int vmdkCreateExtents(PVMDKIMAGE pImage, unsigned cExtents)
+{
+ int rc = VINF_SUCCESS;
+ PVMDKEXTENT pExtents = (PVMDKEXTENT)RTMemAllocZ(cExtents * sizeof(VMDKEXTENT));
+ if (pExtents)
+ {
+ for (unsigned i = 0; i < cExtents; i++)
+ {
+ pExtents[i].pFile = NULL;
+ pExtents[i].pszBasename = NULL;
+ pExtents[i].pszFullname = NULL;
+ pExtents[i].pGD = NULL;
+ pExtents[i].pRGD = NULL;
+ pExtents[i].pDescData = NULL;
+ pExtents[i].uVersion = 1;
+ pExtents[i].uCompression = VMDK_COMPRESSION_NONE;
+ pExtents[i].uExtent = i;
+ pExtents[i].pImage = pImage;
+ }
+ pImage->pExtents = pExtents;
+ pImage->cExtents = cExtents;
+ }
+ else
+ rc = VERR_NO_MEMORY;
+
+ return rc;
+}
+
+/**
+ * Reads and processes the descriptor embedded in sparse images.
+ *
+ * @returns VBox status code.
+ * @param pImage VMDK image instance.
+ * @param pFile The sparse file handle.
+ */
+static int vmdkDescriptorReadSparse(PVMDKIMAGE pImage, PVMDKFILE pFile)
+{
+ /* It's a hosted single-extent image. */
+ int rc = vmdkCreateExtents(pImage, 1);
+ if (RT_SUCCESS(rc))
+ {
+ /* The opened file is passed to the extent. No separate descriptor
+ * file, so no need to keep anything open for the image. */
+ PVMDKEXTENT pExtent = &pImage->pExtents[0];
+ pExtent->pFile = pFile;
+ pImage->pFile = NULL;
+ pExtent->pszFullname = RTPathAbsDup(pImage->pszFilename);
+ if (RT_LIKELY(pExtent->pszFullname))
+ {
+ /* As we're dealing with a monolithic image here, there must
+ * be a descriptor embedded in the image file. */
+ rc = vmdkReadBinaryMetaExtent(pImage, pExtent, true /* fMagicAlreadyRead */);
+ if ( RT_SUCCESS(rc)
+ && pExtent->uDescriptorSector
+ && pExtent->cDescriptorSectors)
+ {
+ /* HACK: extend the descriptor if it is unusually small and it fits in
+ * the unused space after the image header. Allows opening VMDK files
+ * with extremely small descriptor in read/write mode.
+ *
+ * The previous version introduced a possible regression for VMDK stream
+ * optimized images from VMware which tend to have only a single sector sized
+ * descriptor. Increasing the descriptor size resulted in adding the various uuid
+ * entries required to make it work with VBox but for stream optimized images
+ * the updated binary header wasn't written to the disk creating a mismatch
+ * between advertised and real descriptor size.
+ *
+ * The descriptor size will be increased even if opened readonly now if there
+ * enough room but the new value will not be written back to the image.
+ */
+ if ( pExtent->cDescriptorSectors < 3
+ && (int64_t)pExtent->uSectorGD - pExtent->uDescriptorSector >= 4
+ && (!pExtent->uSectorRGD || (int64_t)pExtent->uSectorRGD - pExtent->uDescriptorSector >= 4))
+ {
+ uint64_t cDescriptorSectorsOld = pExtent->cDescriptorSectors;
+
+ pExtent->cDescriptorSectors = 4;
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ {
+ /*
+ * Update the on disk number now to make sure we don't introduce inconsistencies
+ * in case of stream optimized images from VMware where the descriptor is just
+ * one sector big (the binary header is not written to disk for complete
+ * stream optimized images in vmdkFlushImage()).
+ */
+ uint64_t u64DescSizeNew = RT_H2LE_U64(pExtent->cDescriptorSectors);
+ rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pFile->pStorage,
+ RT_UOFFSETOF(SparseExtentHeader, descriptorSize),
+ &u64DescSizeNew, sizeof(u64DescSizeNew));
+ if (RT_FAILURE(rc))
+ {
+ LogFlowFunc(("Increasing the descriptor size failed with %Rrc\n", rc));
+ /* Restore the old size and carry on. */
+ pExtent->cDescriptorSectors = cDescriptorSectorsOld;
+ }
+ }
+ }
+ /* Read the descriptor from the extent. */
+ pExtent->pDescData = (char *)RTMemAllocZ(VMDK_SECTOR2BYTE(pExtent->cDescriptorSectors));
+ if (RT_LIKELY(pExtent->pDescData))
+ {
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(pExtent->uDescriptorSector),
+ pExtent->pDescData,
+ VMDK_SECTOR2BYTE(pExtent->cDescriptorSectors));
+ if (RT_SUCCESS(rc))
+ {
+ rc = vmdkParseDescriptor(pImage, pExtent->pDescData,
+ VMDK_SECTOR2BYTE(pExtent->cDescriptorSectors));
+ if ( RT_SUCCESS(rc)
+ && ( !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ || !(pImage->uOpenFlags & VD_OPEN_FLAGS_ASYNC_IO)))
+ {
+ rc = vmdkReadMetaExtent(pImage, pExtent);
+ if (RT_SUCCESS(rc))
+ {
+ /* Mark the extent as unclean if opened in read-write mode. */
+ if ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ && !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
+ {
+ pExtent->fUncleanShutdown = true;
+ pExtent->fMetaDirty = true;
+ }
+ }
+ }
+ else if (RT_SUCCESS(rc))
+ rc = VERR_NOT_SUPPORTED;
+ }
+ else
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: read error for descriptor in '%s'"), pExtent->pszFullname);
+ }
+ else
+ rc = VERR_NO_MEMORY;
+ }
+ else if (RT_SUCCESS(rc))
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: monolithic image without descriptor in '%s'"), pImage->pszFilename);
+ }
+ else
+ rc = VERR_NO_MEMORY;
+ }
+
+ return rc;
+}
+
+/**
+ * Reads the descriptor from a pure text file.
+ *
+ * @returns VBox status code.
+ * @param pImage VMDK image instance.
+ * @param pFile The descriptor file handle.
+ */
+static int vmdkDescriptorReadAscii(PVMDKIMAGE pImage, PVMDKFILE pFile)
+{
+ /* Allocate at least 10K, and make sure that there is 5K free space
+ * in case new entries need to be added to the descriptor. Never
+ * allocate more than 128K, because that's no valid descriptor file
+ * and will result in the correct "truncated read" error handling. */
+ uint64_t cbFileSize;
+ int rc = vdIfIoIntFileGetSize(pImage->pIfIo, pFile->pStorage, &cbFileSize);
+ if ( RT_SUCCESS(rc)
+ && cbFileSize >= 50)
+ {
+ uint64_t cbSize = cbFileSize;
+ if (cbSize % VMDK_SECTOR2BYTE(10))
+ cbSize += VMDK_SECTOR2BYTE(20) - cbSize % VMDK_SECTOR2BYTE(10);
+ else
+ cbSize += VMDK_SECTOR2BYTE(10);
+ cbSize = RT_MIN(cbSize, _128K);
+ pImage->cbDescAlloc = RT_MAX(VMDK_SECTOR2BYTE(20), cbSize);
+ pImage->pDescData = (char *)RTMemAllocZ(pImage->cbDescAlloc);
+ if (RT_LIKELY(pImage->pDescData))
+ {
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pFile->pStorage, 0, pImage->pDescData,
+ RT_MIN(pImage->cbDescAlloc, cbFileSize));
+ if (RT_SUCCESS(rc))
+ {
+#if 0 /** @todo Revisit */
+ cbRead += sizeof(u32Magic);
+ if (cbRead == pImage->cbDescAlloc)
+ {
+ /* Likely the read is truncated. Better fail a bit too early
+ * (normally the descriptor is much smaller than our buffer). */
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: cannot read descriptor in '%s'"), pImage->pszFilename);
+ goto out;
+ }
+#endif
+ rc = vmdkParseDescriptor(pImage, pImage->pDescData,
+ pImage->cbDescAlloc);
+ if (RT_SUCCESS(rc))
+ {
+ for (unsigned i = 0; i < pImage->cExtents && RT_SUCCESS(rc); i++)
+ {
+ PVMDKEXTENT pExtent = &pImage->pExtents[i];
+ if (pExtent->pszBasename)
+ {
+ /* Hack to figure out whether the specified name in the
+ * extent descriptor is absolute. Doesn't always work, but
+ * should be good enough for now. */
+ char *pszFullname;
+ /** @todo implement proper path absolute check. */
+ if (pExtent->pszBasename[0] == RTPATH_SLASH)
+ {
+ pszFullname = RTStrDup(pExtent->pszBasename);
+ if (!pszFullname)
+ {
+ rc = VERR_NO_MEMORY;
+ break;
+ }
+ }
+ else
+ {
+ char *pszDirname = RTStrDup(pImage->pszFilename);
+ if (!pszDirname)
+ {
+ rc = VERR_NO_MEMORY;
+ break;
+ }
+ RTPathStripFilename(pszDirname);
+ pszFullname = RTPathJoinA(pszDirname, pExtent->pszBasename);
+ RTStrFree(pszDirname);
+ if (!pszFullname)
+ {
+ rc = VERR_NO_STR_MEMORY;
+ break;
+ }
+ }
+ pExtent->pszFullname = pszFullname;
+ }
+ else
+ pExtent->pszFullname = NULL;
+
+ unsigned uOpenFlags = pImage->uOpenFlags | ((pExtent->enmAccess == VMDKACCESS_READONLY) ? VD_OPEN_FLAGS_READONLY : 0);
+ switch (pExtent->enmType)
+ {
+ case VMDKETYPE_HOSTED_SPARSE:
+ rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszBasename, pExtent->pszFullname,
+ VDOpenFlagsToFileOpenFlags(uOpenFlags, false /* fCreate */));
+ if (RT_FAILURE(rc))
+ {
+ /* Do NOT signal an appropriate error here, as the VD
+ * layer has the choice of retrying the open if it
+ * failed. */
+ break;
+ }
+ rc = vmdkReadBinaryMetaExtent(pImage, pExtent,
+ false /* fMagicAlreadyRead */);
+ if (RT_FAILURE(rc))
+ break;
+ rc = vmdkReadMetaExtent(pImage, pExtent);
+ if (RT_FAILURE(rc))
+ break;
+
+ /* Mark extent as unclean if opened in read-write mode. */
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ {
+ pExtent->fUncleanShutdown = true;
+ pExtent->fMetaDirty = true;
+ }
+ break;
+ case VMDKETYPE_VMFS:
+ case VMDKETYPE_FLAT:
+ rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszBasename, pExtent->pszFullname,
+ VDOpenFlagsToFileOpenFlags(uOpenFlags, false /* fCreate */));
+ if (RT_FAILURE(rc))
+ {
+ /* Do NOT signal an appropriate error here, as the VD
+ * layer has the choice of retrying the open if it
+ * failed. */
+ break;
+ }
+ break;
+ case VMDKETYPE_ZERO:
+ /* Nothing to do. */
+ break;
+ default:
+ AssertMsgFailed(("unknown vmdk extent type %d\n", pExtent->enmType));
+ }
+ }
+ }
+ }
+ else
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: read error for descriptor in '%s'"), pImage->pszFilename);
+ }
+ else
+ rc = VERR_NO_MEMORY;
+ }
+ else if (RT_SUCCESS(rc))
+ rc = vdIfError(pImage->pIfError, VERR_VD_VMDK_INVALID_HEADER, RT_SRC_POS, N_("VMDK: descriptor in '%s' is too short"), pImage->pszFilename);
+
+ return rc;
+}
+
+/**
+ * Read and process the descriptor based on the image type.
+ *
+ * @returns VBox status code.
+ * @param pImage VMDK image instance.
+ * @param pFile VMDK file handle.
+ */
+static int vmdkDescriptorRead(PVMDKIMAGE pImage, PVMDKFILE pFile)
+{
+ uint32_t u32Magic;
+
+ /* Read magic (if present). */
+ int rc = vdIfIoIntFileReadSync(pImage->pIfIo, pFile->pStorage, 0,
+ &u32Magic, sizeof(u32Magic));
+ if (RT_SUCCESS(rc))
+ {
+ /* Handle the file according to its magic number. */
+ if (RT_LE2H_U32(u32Magic) == VMDK_SPARSE_MAGICNUMBER)
+ rc = vmdkDescriptorReadSparse(pImage, pFile);
+ else
+ rc = vmdkDescriptorReadAscii(pImage, pFile);
+ }
+ else
+ {
+ vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error reading the magic number in '%s'"), pImage->pszFilename);
+ rc = VERR_VD_VMDK_INVALID_HEADER;
+ }
+
+ return rc;
+}
+
+/**
+ * Internal: Open an image, constructing all necessary data structures.
+ */
+static int vmdkOpenImage(PVMDKIMAGE pImage, unsigned uOpenFlags)
+{
+ pImage->uOpenFlags = uOpenFlags;
+ pImage->pIfError = VDIfErrorGet(pImage->pVDIfsDisk);
+ pImage->pIfIo = VDIfIoIntGet(pImage->pVDIfsImage);
+ AssertPtrReturn(pImage->pIfIo, VERR_INVALID_PARAMETER);
+
+ /*
+ * Open the image.
+ * We don't have to check for asynchronous access because
+ * we only support raw access and the opened file is a description
+ * file were no data is stored.
+ */
+ PVMDKFILE pFile;
+ int rc = vmdkFileOpen(pImage, &pFile, NULL, pImage->pszFilename,
+ VDOpenFlagsToFileOpenFlags(uOpenFlags, false /* fCreate */));
+ if (RT_SUCCESS(rc))
+ {
+ pImage->pFile = pFile;
+
+ rc = vmdkDescriptorRead(pImage, pFile);
+ if (RT_SUCCESS(rc))
+ {
+ /* Determine PCHS geometry if not set. */
+ if (pImage->PCHSGeometry.cCylinders == 0)
+ {
+ uint64_t cCylinders = VMDK_BYTE2SECTOR(pImage->cbSize)
+ / pImage->PCHSGeometry.cHeads
+ / pImage->PCHSGeometry.cSectors;
+ pImage->PCHSGeometry.cCylinders = (unsigned)RT_MIN(cCylinders, 16383);
+ if ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ && !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
+ {
+ rc = vmdkDescSetPCHSGeometry(pImage, &pImage->PCHSGeometry);
+ AssertRC(rc);
+ }
+ }
+
+ /* Update the image metadata now in case has changed. */
+ rc = vmdkFlushImage(pImage, NULL);
+ if (RT_SUCCESS(rc))
+ {
+ /* Figure out a few per-image constants from the extents. */
+ pImage->cbSize = 0;
+ for (unsigned i = 0; i < pImage->cExtents; i++)
+ {
+ PVMDKEXTENT pExtent = &pImage->pExtents[i];
+ if (pExtent->enmType == VMDKETYPE_HOSTED_SPARSE)
+ {
+ /* Here used to be a check whether the nominal size of an extent
+ * is a multiple of the grain size. The spec says that this is
+ * always the case, but unfortunately some files out there in the
+ * wild violate the spec (e.g. ReactOS 0.3.1). */
+ }
+ else if ( pExtent->enmType == VMDKETYPE_FLAT
+ || pExtent->enmType == VMDKETYPE_ZERO)
+ pImage->uImageFlags |= VD_IMAGE_FLAGS_FIXED;
+
+ pImage->cbSize += VMDK_SECTOR2BYTE(pExtent->cNominalSectors);
+ }
+
+ if ( !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ || !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ || !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL))
+ rc = vmdkAllocateGrainTableCache(pImage);
+ }
+ }
+ }
+ /* else: Do NOT signal an appropriate error here, as the VD layer has the
+ * choice of retrying the open if it failed. */
+
+ if (RT_SUCCESS(rc))
+ {
+ PVDREGIONDESC pRegion = &pImage->RegionList.aRegions[0];
+ pImage->RegionList.fFlags = 0;
+ pImage->RegionList.cRegions = 1;
+
+ pRegion->offRegion = 0; /* Disk start. */
+ pRegion->cbBlock = 512;
+ pRegion->enmDataForm = VDREGIONDATAFORM_RAW;
+ pRegion->enmMetadataForm = VDREGIONMETADATAFORM_NONE;
+ pRegion->cbData = 512;
+ pRegion->cbMetadata = 0;
+ pRegion->cRegionBlocksOrBytes = pImage->cbSize;
+ }
+ else
+ vmdkFreeImage(pImage, false, false /*fFlush*/); /* Don't try to flush anything if opening failed. */
+ return rc;
+}
+
+/**
+ * Internal: create VMDK images for raw disk/partition access.
+ */
+static int vmdkCreateRawImage(PVMDKIMAGE pImage, const PVDISKRAW pRaw,
+ uint64_t cbSize)
+{
+ int rc = VINF_SUCCESS;
+ PVMDKEXTENT pExtent;
+
+ if (pRaw->uFlags & VDISKRAW_DISK)
+ {
+ /* Full raw disk access. This requires setting up a descriptor
+ * file and open the (flat) raw disk. */
+ rc = vmdkCreateExtents(pImage, 1);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new extent list in '%s'"), pImage->pszFilename);
+ pExtent = &pImage->pExtents[0];
+ /* Create raw disk descriptor file. */
+ rc = vmdkFileOpen(pImage, &pImage->pFile, NULL, pImage->pszFilename,
+ VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags,
+ true /* fCreate */));
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new file '%s'"), pImage->pszFilename);
+
+ /* Set up basename for extent description. Cannot use StrDup. */
+ size_t cbBasename = strlen(pRaw->pszRawDisk) + 1;
+ char *pszBasename = (char *)RTMemTmpAlloc(cbBasename);
+ if (!pszBasename)
+ return VERR_NO_MEMORY;
+ memcpy(pszBasename, pRaw->pszRawDisk, cbBasename);
+ pExtent->pszBasename = pszBasename;
+ /* For raw disks the full name is identical to the base name. */
+ pExtent->pszFullname = RTStrDup(pszBasename);
+ if (!pExtent->pszFullname)
+ return VERR_NO_MEMORY;
+ pExtent->enmType = VMDKETYPE_FLAT;
+ pExtent->cNominalSectors = VMDK_BYTE2SECTOR(cbSize);
+ pExtent->uSectorOffset = 0;
+ pExtent->enmAccess = (pRaw->uFlags & VDISKRAW_READONLY) ? VMDKACCESS_READONLY : VMDKACCESS_READWRITE;
+ pExtent->fMetaDirty = false;
+
+ /* Open flat image, the raw disk. */
+ rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszBasename, pExtent->pszFullname,
+ VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags | ((pExtent->enmAccess == VMDKACCESS_READONLY) ? VD_OPEN_FLAGS_READONLY : 0),
+ false /* fCreate */));
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not open raw disk file '%s'"), pExtent->pszFullname);
+ }
+ else
+ {
+ /* Raw partition access. This requires setting up a descriptor
+ * file, write the partition information to a flat extent and
+ * open all the (flat) raw disk partitions. */
+
+ /* First pass over the partition data areas to determine how many
+ * extents we need. One data area can require up to 2 extents, as
+ * it might be necessary to skip over unpartitioned space. */
+ unsigned cExtents = 0;
+ uint64_t uStart = 0;
+ for (unsigned i = 0; i < pRaw->cPartDescs; i++)
+ {
+ PVDISKRAWPARTDESC pPart = &pRaw->pPartDescs[i];
+ if (uStart > pPart->uStart)
+ return vdIfError(pImage->pIfError, VERR_INVALID_PARAMETER, RT_SRC_POS,
+ N_("VMDK: incorrect partition data area ordering set up by the caller in '%s'"), pImage->pszFilename);
+
+ if (uStart < pPart->uStart)
+ cExtents++;
+ uStart = pPart->uStart + pPart->cbData;
+ cExtents++;
+ }
+ /* Another extent for filling up the rest of the image. */
+ if (uStart != cbSize)
+ cExtents++;
+
+ rc = vmdkCreateExtents(pImage, cExtents);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new extent list in '%s'"), pImage->pszFilename);
+
+ /* Create raw partition descriptor file. */
+ rc = vmdkFileOpen(pImage, &pImage->pFile, NULL, pImage->pszFilename,
+ VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags,
+ true /* fCreate */));
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new file '%s'"), pImage->pszFilename);
+
+ /* Create base filename for the partition table extent. */
+ /** @todo remove fixed buffer without creating memory leaks. */
+ char pszPartition[1024];
+ const char *pszBase = RTPathFilename(pImage->pszFilename);
+ const char *pszSuff = RTPathSuffix(pszBase);
+ if (pszSuff == NULL)
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: invalid filename '%s'"), pImage->pszFilename);
+ char *pszBaseBase = RTStrDup(pszBase);
+ if (!pszBaseBase)
+ return VERR_NO_MEMORY;
+ RTPathStripSuffix(pszBaseBase);
+ RTStrPrintf(pszPartition, sizeof(pszPartition), "%s-pt%s",
+ pszBaseBase, pszSuff);
+ RTStrFree(pszBaseBase);
+
+ /* Second pass over the partitions, now define all extents. */
+ uint64_t uPartOffset = 0;
+ cExtents = 0;
+ uStart = 0;
+ for (unsigned i = 0; i < pRaw->cPartDescs; i++)
+ {
+ PVDISKRAWPARTDESC pPart = &pRaw->pPartDescs[i];
+ pExtent = &pImage->pExtents[cExtents++];
+
+ if (uStart < pPart->uStart)
+ {
+ pExtent->pszBasename = NULL;
+ pExtent->pszFullname = NULL;
+ pExtent->enmType = VMDKETYPE_ZERO;
+ pExtent->cNominalSectors = VMDK_BYTE2SECTOR(pPart->uStart - uStart);
+ pExtent->uSectorOffset = 0;
+ pExtent->enmAccess = VMDKACCESS_READWRITE;
+ pExtent->fMetaDirty = false;
+ /* go to next extent */
+ pExtent = &pImage->pExtents[cExtents++];
+ }
+ uStart = pPart->uStart + pPart->cbData;
+
+ if (pPart->pvPartitionData)
+ {
+ /* Set up basename for extent description. Can't use StrDup. */
+ size_t cbBasename = strlen(pszPartition) + 1;
+ char *pszBasename = (char *)RTMemTmpAlloc(cbBasename);
+ if (!pszBasename)
+ return VERR_NO_MEMORY;
+ memcpy(pszBasename, pszPartition, cbBasename);
+ pExtent->pszBasename = pszBasename;
+
+ /* Set up full name for partition extent. */
+ char *pszDirname = RTStrDup(pImage->pszFilename);
+ if (!pszDirname)
+ return VERR_NO_STR_MEMORY;
+ RTPathStripFilename(pszDirname);
+ char *pszFullname = RTPathJoinA(pszDirname, pExtent->pszBasename);
+ RTStrFree(pszDirname);
+ if (!pszFullname)
+ return VERR_NO_STR_MEMORY;
+ pExtent->pszFullname = pszFullname;
+ pExtent->enmType = VMDKETYPE_FLAT;
+ pExtent->cNominalSectors = VMDK_BYTE2SECTOR(pPart->cbData);
+ pExtent->uSectorOffset = uPartOffset;
+ pExtent->enmAccess = VMDKACCESS_READWRITE;
+ pExtent->fMetaDirty = false;
+
+ /* Create partition table flat image. */
+ rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszBasename, pExtent->pszFullname,
+ VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags | ((pExtent->enmAccess == VMDKACCESS_READONLY) ? VD_OPEN_FLAGS_READONLY : 0),
+ true /* fCreate */));
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new partition data file '%s'"), pExtent->pszFullname);
+ rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uPartOffset),
+ pPart->pvPartitionData,
+ pPart->cbData);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not write partition data to '%s'"), pExtent->pszFullname);
+ uPartOffset += VMDK_BYTE2SECTOR(pPart->cbData);
+ }
+ else
+ {
+ if (pPart->pszRawDevice)
+ {
+ /* Set up basename for extent descr. Can't use StrDup. */
+ size_t cbBasename = strlen(pPart->pszRawDevice) + 1;
+ char *pszBasename = (char *)RTMemTmpAlloc(cbBasename);
+ if (!pszBasename)
+ return VERR_NO_MEMORY;
+ memcpy(pszBasename, pPart->pszRawDevice, cbBasename);
+ pExtent->pszBasename = pszBasename;
+ /* For raw disks full name is identical to base name. */
+ pExtent->pszFullname = RTStrDup(pszBasename);
+ if (!pExtent->pszFullname)
+ return VERR_NO_MEMORY;
+ pExtent->enmType = VMDKETYPE_FLAT;
+ pExtent->cNominalSectors = VMDK_BYTE2SECTOR(pPart->cbData);
+ pExtent->uSectorOffset = VMDK_BYTE2SECTOR(pPart->uStartOffset);
+ pExtent->enmAccess = (pPart->uFlags & VDISKRAW_READONLY) ? VMDKACCESS_READONLY : VMDKACCESS_READWRITE;
+ pExtent->fMetaDirty = false;
+
+ /* Open flat image, the raw partition. */
+ rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszBasename, pExtent->pszFullname,
+ VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags | ((pExtent->enmAccess == VMDKACCESS_READONLY) ? VD_OPEN_FLAGS_READONLY : 0),
+ false /* fCreate */));
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not open raw partition file '%s'"), pExtent->pszFullname);
+ }
+ else
+ {
+ pExtent->pszBasename = NULL;
+ pExtent->pszFullname = NULL;
+ pExtent->enmType = VMDKETYPE_ZERO;
+ pExtent->cNominalSectors = VMDK_BYTE2SECTOR(pPart->cbData);
+ pExtent->uSectorOffset = 0;
+ pExtent->enmAccess = VMDKACCESS_READWRITE;
+ pExtent->fMetaDirty = false;
+ }
+ }
+ }
+ /* Another extent for filling up the rest of the image. */
+ if (uStart != cbSize)
+ {
+ pExtent = &pImage->pExtents[cExtents++];
+ pExtent->pszBasename = NULL;
+ pExtent->pszFullname = NULL;
+ pExtent->enmType = VMDKETYPE_ZERO;
+ pExtent->cNominalSectors = VMDK_BYTE2SECTOR(cbSize - uStart);
+ pExtent->uSectorOffset = 0;
+ pExtent->enmAccess = VMDKACCESS_READWRITE;
+ pExtent->fMetaDirty = false;
+ }
+ }
+
+ rc = vmdkDescBaseSetStr(pImage, &pImage->Descriptor, "createType",
+ (pRaw->uFlags & VDISKRAW_DISK) ?
+ "fullDevice" : "partitionedDevice");
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not set the image type in '%s'"), pImage->pszFilename);
+ return rc;
+}
+
+/**
+ * Internal: create a regular (i.e. file-backed) VMDK image.
+ */
+static int vmdkCreateRegularImage(PVMDKIMAGE pImage, uint64_t cbSize,
+ unsigned uImageFlags, PVDINTERFACEPROGRESS pIfProgress,
+ unsigned uPercentStart, unsigned uPercentSpan)
+{
+ int rc = VINF_SUCCESS;
+ unsigned cExtents = 1;
+ uint64_t cbOffset = 0;
+ uint64_t cbRemaining = cbSize;
+
+ if (uImageFlags & VD_VMDK_IMAGE_FLAGS_SPLIT_2G)
+ {
+ cExtents = cbSize / VMDK_2G_SPLIT_SIZE;
+ /* Do proper extent computation: need one smaller extent if the total
+ * size isn't evenly divisible by the split size. */
+ if (cbSize % VMDK_2G_SPLIT_SIZE)
+ cExtents++;
+ }
+ rc = vmdkCreateExtents(pImage, cExtents);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new extent list in '%s'"), pImage->pszFilename);
+
+ /* Basename strings needed for constructing the extent names. */
+ char *pszBasenameSubstr = RTPathFilename(pImage->pszFilename);
+ AssertPtr(pszBasenameSubstr);
+ size_t cbBasenameSubstr = strlen(pszBasenameSubstr) + 1;
+
+ /* Create separate descriptor file if necessary. */
+ if (cExtents != 1 || (uImageFlags & VD_IMAGE_FLAGS_FIXED))
+ {
+ rc = vmdkFileOpen(pImage, &pImage->pFile, NULL, pImage->pszFilename,
+ VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags,
+ true /* fCreate */));
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new sparse descriptor file '%s'"), pImage->pszFilename);
+ }
+ else
+ pImage->pFile = NULL;
+
+ /* Set up all extents. */
+ for (unsigned i = 0; i < cExtents; i++)
+ {
+ PVMDKEXTENT pExtent = &pImage->pExtents[i];
+ uint64_t cbExtent = cbRemaining;
+
+ /* Set up fullname/basename for extent description. Cannot use StrDup
+ * for basename, as it is not guaranteed that the memory can be freed
+ * with RTMemTmpFree, which must be used as in other code paths
+ * StrDup is not usable. */
+ if (cExtents == 1 && !(uImageFlags & VD_IMAGE_FLAGS_FIXED))
+ {
+ char *pszBasename = (char *)RTMemTmpAlloc(cbBasenameSubstr);
+ if (!pszBasename)
+ return VERR_NO_MEMORY;
+ memcpy(pszBasename, pszBasenameSubstr, cbBasenameSubstr);
+ pExtent->pszBasename = pszBasename;
+ }
+ else
+ {
+ char *pszBasenameSuff = RTPathSuffix(pszBasenameSubstr);
+ char *pszBasenameBase = RTStrDup(pszBasenameSubstr);
+ RTPathStripSuffix(pszBasenameBase);
+ char *pszTmp;
+ size_t cbTmp;
+ if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
+ {
+ if (cExtents == 1)
+ RTStrAPrintf(&pszTmp, "%s-flat%s", pszBasenameBase,
+ pszBasenameSuff);
+ else
+ RTStrAPrintf(&pszTmp, "%s-f%03d%s", pszBasenameBase,
+ i+1, pszBasenameSuff);
+ }
+ else
+ RTStrAPrintf(&pszTmp, "%s-s%03d%s", pszBasenameBase, i+1,
+ pszBasenameSuff);
+ RTStrFree(pszBasenameBase);
+ if (!pszTmp)
+ return VERR_NO_STR_MEMORY;
+ cbTmp = strlen(pszTmp) + 1;
+ char *pszBasename = (char *)RTMemTmpAlloc(cbTmp);
+ if (!pszBasename)
+ {
+ RTStrFree(pszTmp);
+ return VERR_NO_MEMORY;
+ }
+ memcpy(pszBasename, pszTmp, cbTmp);
+ RTStrFree(pszTmp);
+ pExtent->pszBasename = pszBasename;
+ if (uImageFlags & VD_VMDK_IMAGE_FLAGS_SPLIT_2G)
+ cbExtent = RT_MIN(cbRemaining, VMDK_2G_SPLIT_SIZE);
+ }
+ char *pszBasedirectory = RTStrDup(pImage->pszFilename);
+ if (!pszBasedirectory)
+ return VERR_NO_STR_MEMORY;
+ RTPathStripFilename(pszBasedirectory);
+ char *pszFullname = RTPathJoinA(pszBasedirectory, pExtent->pszBasename);
+ RTStrFree(pszBasedirectory);
+ if (!pszFullname)
+ return VERR_NO_STR_MEMORY;
+ pExtent->pszFullname = pszFullname;
+
+ /* Create file for extent. */
+ rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszBasename, pExtent->pszFullname,
+ VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags,
+ true /* fCreate */));
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new file '%s'"), pExtent->pszFullname);
+ if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
+ {
+ rc = vdIfIoIntFileSetAllocationSize(pImage->pIfIo, pExtent->pFile->pStorage, cbExtent,
+ 0 /* fFlags */, pIfProgress,
+ uPercentStart + cbOffset * uPercentSpan / cbSize,
+ cbExtent * uPercentSpan / cbSize);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not set size of new file '%s'"), pExtent->pszFullname);
+ }
+
+ /* Place descriptor file information (where integrated). */
+ if (cExtents == 1 && !(uImageFlags & VD_IMAGE_FLAGS_FIXED))
+ {
+ pExtent->uDescriptorSector = 1;
+ pExtent->cDescriptorSectors = VMDK_BYTE2SECTOR(pImage->cbDescAlloc);
+ /* The descriptor is part of the (only) extent. */
+ pExtent->pDescData = pImage->pDescData;
+ pImage->pDescData = NULL;
+ }
+
+ if (!(uImageFlags & VD_IMAGE_FLAGS_FIXED))
+ {
+ uint64_t cSectorsPerGDE, cSectorsPerGD;
+ pExtent->enmType = VMDKETYPE_HOSTED_SPARSE;
+ pExtent->cSectors = VMDK_BYTE2SECTOR(RT_ALIGN_64(cbExtent, _64K));
+ pExtent->cSectorsPerGrain = VMDK_BYTE2SECTOR(_64K);
+ pExtent->cGTEntries = 512;
+ cSectorsPerGDE = pExtent->cGTEntries * pExtent->cSectorsPerGrain;
+ pExtent->cSectorsPerGDE = cSectorsPerGDE;
+ pExtent->cGDEntries = (pExtent->cSectors + cSectorsPerGDE - 1) / cSectorsPerGDE;
+ cSectorsPerGD = (pExtent->cGDEntries + (512 / sizeof(uint32_t) - 1)) / (512 / sizeof(uint32_t));
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ /* The spec says version is 1 for all VMDKs, but the vast
+ * majority of streamOptimized VMDKs actually contain
+ * version 3 - so go with the majority. Both are accepted. */
+ pExtent->uVersion = 3;
+ pExtent->uCompression = VMDK_COMPRESSION_DEFLATE;
+ }
+ }
+ else
+ {
+ if (uImageFlags & VD_VMDK_IMAGE_FLAGS_ESX)
+ pExtent->enmType = VMDKETYPE_VMFS;
+ else
+ pExtent->enmType = VMDKETYPE_FLAT;
+ }
+
+ pExtent->enmAccess = VMDKACCESS_READWRITE;
+ pExtent->fUncleanShutdown = true;
+ pExtent->cNominalSectors = VMDK_BYTE2SECTOR(cbExtent);
+ pExtent->uSectorOffset = 0;
+ pExtent->fMetaDirty = true;
+
+ if (!(uImageFlags & VD_IMAGE_FLAGS_FIXED))
+ {
+ /* fPreAlloc should never be false because VMware can't use such images. */
+ rc = vmdkCreateGrainDirectory(pImage, pExtent,
+ RT_MAX( pExtent->uDescriptorSector
+ + pExtent->cDescriptorSectors,
+ 1),
+ true /* fPreAlloc */);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new grain directory in '%s'"), pExtent->pszFullname);
+ }
+
+ cbOffset += cbExtent;
+
+ if (RT_SUCCESS(rc))
+ vdIfProgress(pIfProgress, uPercentStart + cbOffset * uPercentSpan / cbSize);
+
+ cbRemaining -= cbExtent;
+ }
+
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_ESX)
+ {
+ /* VirtualBox doesn't care, but VMWare ESX freaks out if the wrong
+ * controller type is set in an image. */
+ rc = vmdkDescDDBSetStr(pImage, &pImage->Descriptor, "ddb.adapterType", "lsilogic");
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not set controller type to lsilogic in '%s'"), pImage->pszFilename);
+ }
+
+ const char *pszDescType = NULL;
+ if (uImageFlags & VD_IMAGE_FLAGS_FIXED)
+ {
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_ESX)
+ pszDescType = "vmfs";
+ else
+ pszDescType = (cExtents == 1)
+ ? "monolithicFlat" : "twoGbMaxExtentFlat";
+ }
+ else
+ {
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ pszDescType = "streamOptimized";
+ else
+ {
+ pszDescType = (cExtents == 1)
+ ? "monolithicSparse" : "twoGbMaxExtentSparse";
+ }
+ }
+ rc = vmdkDescBaseSetStr(pImage, &pImage->Descriptor, "createType",
+ pszDescType);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not set the image type in '%s'"), pImage->pszFilename);
+ return rc;
+}
+
+/**
+ * Internal: Create a real stream optimized VMDK using only linear writes.
+ */
+static int vmdkCreateStreamImage(PVMDKIMAGE pImage, uint64_t cbSize)
+{
+ int rc = vmdkCreateExtents(pImage, 1);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new extent list in '%s'"), pImage->pszFilename);
+
+ /* Basename strings needed for constructing the extent names. */
+ const char *pszBasenameSubstr = RTPathFilename(pImage->pszFilename);
+ AssertPtr(pszBasenameSubstr);
+ size_t cbBasenameSubstr = strlen(pszBasenameSubstr) + 1;
+
+ /* No separate descriptor file. */
+ pImage->pFile = NULL;
+
+ /* Set up all extents. */
+ PVMDKEXTENT pExtent = &pImage->pExtents[0];
+
+ /* Set up fullname/basename for extent description. Cannot use StrDup
+ * for basename, as it is not guaranteed that the memory can be freed
+ * with RTMemTmpFree, which must be used as in other code paths
+ * StrDup is not usable. */
+ char *pszBasename = (char *)RTMemTmpAlloc(cbBasenameSubstr);
+ if (!pszBasename)
+ return VERR_NO_MEMORY;
+ memcpy(pszBasename, pszBasenameSubstr, cbBasenameSubstr);
+ pExtent->pszBasename = pszBasename;
+
+ char *pszBasedirectory = RTStrDup(pImage->pszFilename);
+ RTPathStripFilename(pszBasedirectory);
+ char *pszFullname = RTPathJoinA(pszBasedirectory, pExtent->pszBasename);
+ RTStrFree(pszBasedirectory);
+ if (!pszFullname)
+ return VERR_NO_STR_MEMORY;
+ pExtent->pszFullname = pszFullname;
+
+ /* Create file for extent. Make it write only, no reading allowed. */
+ rc = vmdkFileOpen(pImage, &pExtent->pFile, pExtent->pszBasename, pExtent->pszFullname,
+ VDOpenFlagsToFileOpenFlags(pImage->uOpenFlags,
+ true /* fCreate */)
+ & ~RTFILE_O_READ);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new file '%s'"), pExtent->pszFullname);
+
+ /* Place descriptor file information. */
+ pExtent->uDescriptorSector = 1;
+ pExtent->cDescriptorSectors = VMDK_BYTE2SECTOR(pImage->cbDescAlloc);
+ /* The descriptor is part of the (only) extent. */
+ pExtent->pDescData = pImage->pDescData;
+ pImage->pDescData = NULL;
+
+ uint64_t cSectorsPerGDE, cSectorsPerGD;
+ pExtent->enmType = VMDKETYPE_HOSTED_SPARSE;
+ pExtent->cSectors = VMDK_BYTE2SECTOR(RT_ALIGN_64(cbSize, _64K));
+ pExtent->cSectorsPerGrain = VMDK_BYTE2SECTOR(_64K);
+ pExtent->cGTEntries = 512;
+ cSectorsPerGDE = pExtent->cGTEntries * pExtent->cSectorsPerGrain;
+ pExtent->cSectorsPerGDE = cSectorsPerGDE;
+ pExtent->cGDEntries = (pExtent->cSectors + cSectorsPerGDE - 1) / cSectorsPerGDE;
+ cSectorsPerGD = (pExtent->cGDEntries + (512 / sizeof(uint32_t) - 1)) / (512 / sizeof(uint32_t));
+
+ /* The spec says version is 1 for all VMDKs, but the vast
+ * majority of streamOptimized VMDKs actually contain
+ * version 3 - so go with the majority. Both are accepted. */
+ pExtent->uVersion = 3;
+ pExtent->uCompression = VMDK_COMPRESSION_DEFLATE;
+ pExtent->fFooter = true;
+
+ pExtent->enmAccess = VMDKACCESS_READONLY;
+ pExtent->fUncleanShutdown = false;
+ pExtent->cNominalSectors = VMDK_BYTE2SECTOR(cbSize);
+ pExtent->uSectorOffset = 0;
+ pExtent->fMetaDirty = true;
+
+ /* Create grain directory, without preallocating it straight away. It will
+ * be constructed on the fly when writing out the data and written when
+ * closing the image. The end effect is that the full grain directory is
+ * allocated, which is a requirement of the VMDK specs. */
+ rc = vmdkCreateGrainDirectory(pImage, pExtent, VMDK_GD_AT_END,
+ false /* fPreAlloc */);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new grain directory in '%s'"), pExtent->pszFullname);
+
+ rc = vmdkDescBaseSetStr(pImage, &pImage->Descriptor, "createType",
+ "streamOptimized");
+ if (RT_FAILURE(rc))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not set the image type in '%s'"), pImage->pszFilename);
+
+ return rc;
+}
+
+/**
+ * Initializes the UUID fields in the DDB.
+ *
+ * @returns VBox status code.
+ * @param pImage The VMDK image instance.
+ */
+static int vmdkCreateImageDdbUuidsInit(PVMDKIMAGE pImage)
+{
+ int rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor, VMDK_DDB_IMAGE_UUID, &pImage->ImageUuid);
+ if (RT_SUCCESS(rc))
+ {
+ rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor, VMDK_DDB_PARENT_UUID, &pImage->ParentUuid);
+ if (RT_SUCCESS(rc))
+ {
+ rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor, VMDK_DDB_MODIFICATION_UUID,
+ &pImage->ModificationUuid);
+ if (RT_SUCCESS(rc))
+ {
+ rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor, VMDK_DDB_PARENT_MODIFICATION_UUID,
+ &pImage->ParentModificationUuid);
+ if (RT_FAILURE(rc))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
+ N_("VMDK: error storing parent modification UUID in new descriptor in '%s'"), pImage->pszFilename);
+ }
+ else
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
+ N_("VMDK: error storing modification UUID in new descriptor in '%s'"), pImage->pszFilename);
+ }
+ else
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
+ N_("VMDK: error storing parent image UUID in new descriptor in '%s'"), pImage->pszFilename);
+ }
+ else
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
+ N_("VMDK: error storing image UUID in new descriptor in '%s'"), pImage->pszFilename);
+
+ return rc;
+}
+
+/**
+ * Internal: The actual code for creating any VMDK variant currently in
+ * existence on hosted environments.
+ */
+static int vmdkCreateImage(PVMDKIMAGE pImage, uint64_t cbSize,
+ unsigned uImageFlags, const char *pszComment,
+ PCVDGEOMETRY pPCHSGeometry,
+ PCVDGEOMETRY pLCHSGeometry, PCRTUUID pUuid,
+ PVDINTERFACEPROGRESS pIfProgress,
+ unsigned uPercentStart, unsigned uPercentSpan)
+{
+ pImage->uImageFlags = uImageFlags;
+
+ pImage->pIfError = VDIfErrorGet(pImage->pVDIfsDisk);
+ pImage->pIfIo = VDIfIoIntGet(pImage->pVDIfsImage);
+ AssertPtrReturn(pImage->pIfIo, VERR_INVALID_PARAMETER);
+
+ int rc = vmdkCreateDescriptor(pImage, pImage->pDescData, pImage->cbDescAlloc,
+ &pImage->Descriptor);
+ if (RT_SUCCESS(rc))
+ {
+ if ( (uImageFlags & VD_IMAGE_FLAGS_FIXED)
+ && (uImageFlags & VD_VMDK_IMAGE_FLAGS_RAWDISK))
+ {
+ /* Raw disk image (includes raw partition). */
+ const PVDISKRAW pRaw = (const PVDISKRAW)pszComment;
+ /* As the comment is misused, zap it so that no garbage comment
+ * is set below. */
+ pszComment = NULL;
+ rc = vmdkCreateRawImage(pImage, pRaw, cbSize);
+ }
+ else if (uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ /* Stream optimized sparse image (monolithic). */
+ rc = vmdkCreateStreamImage(pImage, cbSize);
+ }
+ else
+ {
+ /* Regular fixed or sparse image (monolithic or split). */
+ rc = vmdkCreateRegularImage(pImage, cbSize, uImageFlags,
+ pIfProgress, uPercentStart,
+ uPercentSpan * 95 / 100);
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ vdIfProgress(pIfProgress, uPercentStart + uPercentSpan * 98 / 100);
+
+ pImage->cbSize = cbSize;
+
+ for (unsigned i = 0; i < pImage->cExtents; i++)
+ {
+ PVMDKEXTENT pExtent = &pImage->pExtents[i];
+
+ rc = vmdkDescExtInsert(pImage, &pImage->Descriptor, pExtent->enmAccess,
+ pExtent->cNominalSectors, pExtent->enmType,
+ pExtent->pszBasename, pExtent->uSectorOffset);
+ if (RT_FAILURE(rc))
+ {
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not insert the extent list into descriptor in '%s'"), pImage->pszFilename);
+ break;
+ }
+ }
+
+ if (RT_SUCCESS(rc))
+ vmdkDescExtRemoveDummy(pImage, &pImage->Descriptor);
+
+ if ( RT_SUCCESS(rc)
+ && pPCHSGeometry->cCylinders != 0
+ && pPCHSGeometry->cHeads != 0
+ && pPCHSGeometry->cSectors != 0)
+ rc = vmdkDescSetPCHSGeometry(pImage, pPCHSGeometry);
+
+ if ( RT_SUCCESS(rc)
+ && pLCHSGeometry->cCylinders != 0
+ && pLCHSGeometry->cHeads != 0
+ && pLCHSGeometry->cSectors != 0)
+ rc = vmdkDescSetLCHSGeometry(pImage, pLCHSGeometry);
+
+ pImage->LCHSGeometry = *pLCHSGeometry;
+ pImage->PCHSGeometry = *pPCHSGeometry;
+
+ pImage->ImageUuid = *pUuid;
+ RTUuidClear(&pImage->ParentUuid);
+ RTUuidClear(&pImage->ModificationUuid);
+ RTUuidClear(&pImage->ParentModificationUuid);
+
+ if (RT_SUCCESS(rc))
+ rc = vmdkCreateImageDdbUuidsInit(pImage);
+
+ if (RT_SUCCESS(rc))
+ rc = vmdkAllocateGrainTableCache(pImage);
+
+ if (RT_SUCCESS(rc))
+ {
+ rc = vmdkSetImageComment(pImage, pszComment);
+ if (RT_FAILURE(rc))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot set image comment in '%s'"), pImage->pszFilename);
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ vdIfProgress(pIfProgress, uPercentStart + uPercentSpan * 99 / 100);
+
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ /* streamOptimized is a bit special, we cannot trigger the flush
+ * until all data has been written. So we write the necessary
+ * information explicitly. */
+ pImage->pExtents[0].cDescriptorSectors = VMDK_BYTE2SECTOR(RT_ALIGN_64( pImage->Descriptor.aLines[pImage->Descriptor.cLines]
+ - pImage->Descriptor.aLines[0], 512));
+ rc = vmdkWriteMetaSparseExtent(pImage, &pImage->pExtents[0], 0, NULL);
+ if (RT_SUCCESS(rc))
+ {
+ rc = vmdkWriteDescriptor(pImage, NULL);
+ if (RT_FAILURE(rc))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write VMDK descriptor in '%s'"), pImage->pszFilename);
+ }
+ else
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write VMDK header in '%s'"), pImage->pszFilename);
+ }
+ else
+ rc = vmdkFlushImage(pImage, NULL);
+ }
+ }
+ }
+ else
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: could not create new descriptor in '%s'"), pImage->pszFilename);
+
+
+ if (RT_SUCCESS(rc))
+ {
+ PVDREGIONDESC pRegion = &pImage->RegionList.aRegions[0];
+ pImage->RegionList.fFlags = 0;
+ pImage->RegionList.cRegions = 1;
+
+ pRegion->offRegion = 0; /* Disk start. */
+ pRegion->cbBlock = 512;
+ pRegion->enmDataForm = VDREGIONDATAFORM_RAW;
+ pRegion->enmMetadataForm = VDREGIONMETADATAFORM_NONE;
+ pRegion->cbData = 512;
+ pRegion->cbMetadata = 0;
+ pRegion->cRegionBlocksOrBytes = pImage->cbSize;
+
+ vdIfProgress(pIfProgress, uPercentStart + uPercentSpan);
+ }
+ else
+ vmdkFreeImage(pImage, rc != VERR_ALREADY_EXISTS, false /*fFlush*/);
+ return rc;
+}
+
+/**
+ * Internal: Update image comment.
+ */
+static int vmdkSetImageComment(PVMDKIMAGE pImage, const char *pszComment)
+{
+ char *pszCommentEncoded = NULL;
+ if (pszComment)
+ {
+ pszCommentEncoded = vmdkEncodeString(pszComment);
+ if (!pszCommentEncoded)
+ return VERR_NO_MEMORY;
+ }
+
+ int rc = vmdkDescDDBSetStr(pImage, &pImage->Descriptor,
+ "ddb.comment", pszCommentEncoded);
+ if (pszCommentEncoded)
+ RTStrFree(pszCommentEncoded);
+ if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing image comment in descriptor in '%s'"), pImage->pszFilename);
+ return VINF_SUCCESS;
+}
+
+/**
+ * Internal. Clear the grain table buffer for real stream optimized writing.
+ */
+static void vmdkStreamClearGT(PVMDKIMAGE pImage, PVMDKEXTENT pExtent)
+{
+ uint32_t cCacheLines = RT_ALIGN(pExtent->cGTEntries, VMDK_GT_CACHELINE_SIZE) / VMDK_GT_CACHELINE_SIZE;
+ for (uint32_t i = 0; i < cCacheLines; i++)
+ memset(&pImage->pGTCache->aGTCache[i].aGTData[0], '\0',
+ VMDK_GT_CACHELINE_SIZE * sizeof(uint32_t));
+}
+
+/**
+ * Internal. Flush the grain table buffer for real stream optimized writing.
+ */
+static int vmdkStreamFlushGT(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
+ uint32_t uGDEntry)
+{
+ int rc = VINF_SUCCESS;
+ uint32_t cCacheLines = RT_ALIGN(pExtent->cGTEntries, VMDK_GT_CACHELINE_SIZE) / VMDK_GT_CACHELINE_SIZE;
+
+ /* VMware does not write out completely empty grain tables in the case
+ * of streamOptimized images, which according to my interpretation of
+ * the VMDK 1.1 spec is bending the rules. Since they do it and we can
+ * handle it without problems do it the same way and save some bytes. */
+ bool fAllZero = true;
+ for (uint32_t i = 0; i < cCacheLines; i++)
+ {
+ /* Convert the grain table to little endian in place, as it will not
+ * be used at all after this function has been called. */
+ uint32_t *pGTTmp = &pImage->pGTCache->aGTCache[i].aGTData[0];
+ for (uint32_t j = 0; j < VMDK_GT_CACHELINE_SIZE; j++, pGTTmp++)
+ if (*pGTTmp)
+ {
+ fAllZero = false;
+ break;
+ }
+ if (!fAllZero)
+ break;
+ }
+ if (fAllZero)
+ return VINF_SUCCESS;
+
+ uint64_t uFileOffset = pExtent->uAppendPosition;
+ if (!uFileOffset)
+ return VERR_INTERNAL_ERROR;
+ /* Align to sector, as the previous write could have been any size. */
+ uFileOffset = RT_ALIGN_64(uFileOffset, 512);
+
+ /* Grain table marker. */
+ uint8_t aMarker[512];
+ PVMDKMARKER pMarker = (PVMDKMARKER)&aMarker[0];
+ memset(pMarker, '\0', sizeof(aMarker));
+ pMarker->uSector = RT_H2LE_U64(VMDK_BYTE2SECTOR((uint64_t)pExtent->cGTEntries * sizeof(uint32_t)));
+ pMarker->uType = RT_H2LE_U32(VMDK_MARKER_GT);
+ rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage, uFileOffset,
+ aMarker, sizeof(aMarker));
+ AssertRC(rc);
+ uFileOffset += 512;
+
+ if (!pExtent->pGD || pExtent->pGD[uGDEntry])
+ return VERR_INTERNAL_ERROR;
+
+ pExtent->pGD[uGDEntry] = VMDK_BYTE2SECTOR(uFileOffset);
+
+ for (uint32_t i = 0; i < cCacheLines; i++)
+ {
+ /* Convert the grain table to little endian in place, as it will not
+ * be used at all after this function has been called. */
+ uint32_t *pGTTmp = &pImage->pGTCache->aGTCache[i].aGTData[0];
+ for (uint32_t j = 0; j < VMDK_GT_CACHELINE_SIZE; j++, pGTTmp++)
+ *pGTTmp = RT_H2LE_U32(*pGTTmp);
+
+ rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage, uFileOffset,
+ &pImage->pGTCache->aGTCache[i].aGTData[0],
+ VMDK_GT_CACHELINE_SIZE * sizeof(uint32_t));
+ uFileOffset += VMDK_GT_CACHELINE_SIZE * sizeof(uint32_t);
+ if (RT_FAILURE(rc))
+ break;
+ }
+ Assert(!(uFileOffset % 512));
+ pExtent->uAppendPosition = RT_ALIGN_64(uFileOffset, 512);
+ return rc;
+}
+
+/**
+ * Internal. Free all allocated space for representing an image, and optionally
+ * delete the image from disk.
+ */
+static int vmdkFreeImage(PVMDKIMAGE pImage, bool fDelete, bool fFlush)
+{
+ int rc = VINF_SUCCESS;
+
+ /* Freeing a never allocated image (e.g. because the open failed) is
+ * not signalled as an error. After all nothing bad happens. */
+ if (pImage)
+ {
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ {
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ /* Check if all extents are clean. */
+ for (unsigned i = 0; i < pImage->cExtents; i++)
+ {
+ Assert(!pImage->pExtents[i].fUncleanShutdown);
+ }
+ }
+ else
+ {
+ /* Mark all extents as clean. */
+ for (unsigned i = 0; i < pImage->cExtents; i++)
+ {
+ if ( pImage->pExtents[i].enmType == VMDKETYPE_HOSTED_SPARSE
+ && pImage->pExtents[i].fUncleanShutdown)
+ {
+ pImage->pExtents[i].fUncleanShutdown = false;
+ pImage->pExtents[i].fMetaDirty = true;
+ }
+
+ /* From now on it's not safe to append any more data. */
+ pImage->pExtents[i].uAppendPosition = 0;
+ }
+ }
+ }
+
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ /* No need to write any pending data if the file will be deleted
+ * or if the new file wasn't successfully created. */
+ if ( !fDelete && pImage->pExtents
+ && pImage->pExtents[0].cGTEntries
+ && pImage->pExtents[0].uAppendPosition)
+ {
+ PVMDKEXTENT pExtent = &pImage->pExtents[0];
+ uint32_t uLastGDEntry = pExtent->uLastGrainAccess / pExtent->cGTEntries;
+ rc = vmdkStreamFlushGT(pImage, pExtent, uLastGDEntry);
+ AssertRC(rc);
+ vmdkStreamClearGT(pImage, pExtent);
+ for (uint32_t i = uLastGDEntry + 1; i < pExtent->cGDEntries; i++)
+ {
+ rc = vmdkStreamFlushGT(pImage, pExtent, i);
+ AssertRC(rc);
+ }
+
+ uint64_t uFileOffset = pExtent->uAppendPosition;
+ if (!uFileOffset)
+ return VERR_INTERNAL_ERROR;
+ uFileOffset = RT_ALIGN_64(uFileOffset, 512);
+
+ /* From now on it's not safe to append any more data. */
+ pExtent->uAppendPosition = 0;
+
+ /* Grain directory marker. */
+ uint8_t aMarker[512];
+ PVMDKMARKER pMarker = (PVMDKMARKER)&aMarker[0];
+ memset(pMarker, '\0', sizeof(aMarker));
+ pMarker->uSector = VMDK_BYTE2SECTOR(RT_ALIGN_64(RT_H2LE_U64((uint64_t)pExtent->cGDEntries * sizeof(uint32_t)), 512));
+ pMarker->uType = RT_H2LE_U32(VMDK_MARKER_GD);
+ rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage, uFileOffset,
+ aMarker, sizeof(aMarker));
+ AssertRC(rc);
+ uFileOffset += 512;
+
+ /* Write grain directory in little endian style. The array will
+ * not be used after this, so convert in place. */
+ uint32_t *pGDTmp = pExtent->pGD;
+ for (uint32_t i = 0; i < pExtent->cGDEntries; i++, pGDTmp++)
+ *pGDTmp = RT_H2LE_U32(*pGDTmp);
+ rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ uFileOffset, pExtent->pGD,
+ pExtent->cGDEntries * sizeof(uint32_t));
+ AssertRC(rc);
+
+ pExtent->uSectorGD = VMDK_BYTE2SECTOR(uFileOffset);
+ pExtent->uSectorRGD = VMDK_BYTE2SECTOR(uFileOffset);
+ uFileOffset = RT_ALIGN_64( uFileOffset
+ + pExtent->cGDEntries * sizeof(uint32_t),
+ 512);
+
+ /* Footer marker. */
+ memset(pMarker, '\0', sizeof(aMarker));
+ pMarker->uSector = VMDK_BYTE2SECTOR(512);
+ pMarker->uType = RT_H2LE_U32(VMDK_MARKER_FOOTER);
+ rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ uFileOffset, aMarker, sizeof(aMarker));
+ AssertRC(rc);
+
+ uFileOffset += 512;
+ rc = vmdkWriteMetaSparseExtent(pImage, pExtent, uFileOffset, NULL);
+ AssertRC(rc);
+
+ uFileOffset += 512;
+ /* End-of-stream marker. */
+ memset(pMarker, '\0', sizeof(aMarker));
+ rc = vdIfIoIntFileWriteSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ uFileOffset, aMarker, sizeof(aMarker));
+ AssertRC(rc);
+ }
+ }
+ else if (!fDelete && fFlush)
+ vmdkFlushImage(pImage, NULL);
+
+ if (pImage->pExtents != NULL)
+ {
+ for (unsigned i = 0 ; i < pImage->cExtents; i++)
+ {
+ int rc2 = vmdkFreeExtentData(pImage, &pImage->pExtents[i], fDelete);
+ if (RT_SUCCESS(rc))
+ rc = rc2; /* Propogate any error when closing the file. */
+ }
+ RTMemFree(pImage->pExtents);
+ pImage->pExtents = NULL;
+ }
+ pImage->cExtents = 0;
+ if (pImage->pFile != NULL)
+ {
+ int rc2 = vmdkFileClose(pImage, &pImage->pFile, fDelete);
+ if (RT_SUCCESS(rc))
+ rc = rc2; /* Propogate any error when closing the file. */
+ }
+ int rc2 = vmdkFileCheckAllClose(pImage);
+ if (RT_SUCCESS(rc))
+ rc = rc2; /* Propogate any error when closing the file. */
+
+ if (pImage->pGTCache)
+ {
+ RTMemFree(pImage->pGTCache);
+ pImage->pGTCache = NULL;
+ }
+ if (pImage->pDescData)
+ {
+ RTMemFree(pImage->pDescData);
+ pImage->pDescData = NULL;
+ }
+ }
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/**
+ * Internal. Flush image data (and metadata) to disk.
+ */
+static int vmdkFlushImage(PVMDKIMAGE pImage, PVDIOCTX pIoCtx)
+{
+ PVMDKEXTENT pExtent;
+ int rc = VINF_SUCCESS;
+
+ /* Update descriptor if changed. */
+ if (pImage->Descriptor.fDirty)
+ rc = vmdkWriteDescriptor(pImage, pIoCtx);
+
+ if (RT_SUCCESS(rc))
+ {
+ for (unsigned i = 0; i < pImage->cExtents; i++)
+ {
+ pExtent = &pImage->pExtents[i];
+ if (pExtent->pFile != NULL && pExtent->fMetaDirty)
+ {
+ switch (pExtent->enmType)
+ {
+ case VMDKETYPE_HOSTED_SPARSE:
+ if (!pExtent->fFooter)
+ rc = vmdkWriteMetaSparseExtent(pImage, pExtent, 0, pIoCtx);
+ else
+ {
+ uint64_t uFileOffset = pExtent->uAppendPosition;
+ /* Simply skip writing anything if the streamOptimized
+ * image hasn't been just created. */
+ if (!uFileOffset)
+ break;
+ uFileOffset = RT_ALIGN_64(uFileOffset, 512);
+ rc = vmdkWriteMetaSparseExtent(pImage, pExtent,
+ uFileOffset, pIoCtx);
+ }
+ break;
+ case VMDKETYPE_VMFS:
+ case VMDKETYPE_FLAT:
+ /* Nothing to do. */
+ break;
+ case VMDKETYPE_ZERO:
+ default:
+ AssertMsgFailed(("extent with type %d marked as dirty\n",
+ pExtent->enmType));
+ break;
+ }
+ }
+
+ if (RT_FAILURE(rc))
+ break;
+
+ switch (pExtent->enmType)
+ {
+ case VMDKETYPE_HOSTED_SPARSE:
+ case VMDKETYPE_VMFS:
+ case VMDKETYPE_FLAT:
+ /** @todo implement proper path absolute check. */
+ if ( pExtent->pFile != NULL
+ && !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ && !(pExtent->pszBasename[0] == RTPATH_SLASH))
+ rc = vdIfIoIntFileFlush(pImage->pIfIo, pExtent->pFile->pStorage, pIoCtx,
+ NULL, NULL);
+ break;
+ case VMDKETYPE_ZERO:
+ /* No need to do anything for this extent. */
+ break;
+ default:
+ AssertMsgFailed(("unknown extent type %d\n", pExtent->enmType));
+ break;
+ }
+ }
+ }
+
+ return rc;
+}
+
+/**
+ * Internal. Find extent corresponding to the sector number in the disk.
+ */
+static int vmdkFindExtent(PVMDKIMAGE pImage, uint64_t offSector,
+ PVMDKEXTENT *ppExtent, uint64_t *puSectorInExtent)
+{
+ PVMDKEXTENT pExtent = NULL;
+ int rc = VINF_SUCCESS;
+
+ for (unsigned i = 0; i < pImage->cExtents; i++)
+ {
+ if (offSector < pImage->pExtents[i].cNominalSectors)
+ {
+ pExtent = &pImage->pExtents[i];
+ *puSectorInExtent = offSector + pImage->pExtents[i].uSectorOffset;
+ break;
+ }
+ offSector -= pImage->pExtents[i].cNominalSectors;
+ }
+
+ if (pExtent)
+ *ppExtent = pExtent;
+ else
+ rc = VERR_IO_SECTOR_NOT_FOUND;
+
+ return rc;
+}
+
+/**
+ * Internal. Hash function for placing the grain table hash entries.
+ */
+static uint32_t vmdkGTCacheHash(PVMDKGTCACHE pCache, uint64_t uSector,
+ unsigned uExtent)
+{
+ /** @todo this hash function is quite simple, maybe use a better one which
+ * scrambles the bits better. */
+ return (uSector + uExtent) % pCache->cEntries;
+}
+
+/**
+ * Internal. Get sector number in the extent file from the relative sector
+ * number in the extent.
+ */
+static int vmdkGetSector(PVMDKIMAGE pImage, PVDIOCTX pIoCtx,
+ PVMDKEXTENT pExtent, uint64_t uSector,
+ uint64_t *puExtentSector)
+{
+ PVMDKGTCACHE pCache = pImage->pGTCache;
+ uint64_t uGDIndex, uGTSector, uGTBlock;
+ uint32_t uGTHash, uGTBlockIndex;
+ PVMDKGTCACHEENTRY pGTCacheEntry;
+ uint32_t aGTDataTmp[VMDK_GT_CACHELINE_SIZE];
+ int rc;
+
+ /* For newly created and readonly/sequentially opened streamOptimized
+ * images this must be a no-op, as the grain directory is not there. */
+ if ( ( pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED
+ && pExtent->uAppendPosition)
+ || ( pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED
+ && pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY
+ && pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL))
+ {
+ *puExtentSector = 0;
+ return VINF_SUCCESS;
+ }
+
+ uGDIndex = uSector / pExtent->cSectorsPerGDE;
+ if (uGDIndex >= pExtent->cGDEntries)
+ return VERR_OUT_OF_RANGE;
+ uGTSector = pExtent->pGD[uGDIndex];
+ if (!uGTSector)
+ {
+ /* There is no grain table referenced by this grain directory
+ * entry. So there is absolutely no data in this area. */
+ *puExtentSector = 0;
+ return VINF_SUCCESS;
+ }
+
+ uGTBlock = uSector / (pExtent->cSectorsPerGrain * VMDK_GT_CACHELINE_SIZE);
+ uGTHash = vmdkGTCacheHash(pCache, uGTBlock, pExtent->uExtent);
+ pGTCacheEntry = &pCache->aGTCache[uGTHash];
+ if ( pGTCacheEntry->uExtent != pExtent->uExtent
+ || pGTCacheEntry->uGTBlock != uGTBlock)
+ {
+ /* Cache miss, fetch data from disk. */
+ PVDMETAXFER pMetaXfer;
+ rc = vdIfIoIntFileReadMeta(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uGTSector) + (uGTBlock % (pExtent->cGTEntries / VMDK_GT_CACHELINE_SIZE)) * sizeof(aGTDataTmp),
+ aGTDataTmp, sizeof(aGTDataTmp), pIoCtx, &pMetaXfer, NULL, NULL);
+ if (RT_FAILURE(rc))
+ return rc;
+ /* We can release the metadata transfer immediately. */
+ vdIfIoIntMetaXferRelease(pImage->pIfIo, pMetaXfer);
+ pGTCacheEntry->uExtent = pExtent->uExtent;
+ pGTCacheEntry->uGTBlock = uGTBlock;
+ for (unsigned i = 0; i < VMDK_GT_CACHELINE_SIZE; i++)
+ pGTCacheEntry->aGTData[i] = RT_LE2H_U32(aGTDataTmp[i]);
+ }
+ uGTBlockIndex = (uSector / pExtent->cSectorsPerGrain) % VMDK_GT_CACHELINE_SIZE;
+ uint32_t uGrainSector = pGTCacheEntry->aGTData[uGTBlockIndex];
+ if (uGrainSector)
+ *puExtentSector = uGrainSector + uSector % pExtent->cSectorsPerGrain;
+ else
+ *puExtentSector = 0;
+ return VINF_SUCCESS;
+}
+
+/**
+ * Internal. Writes the grain and also if necessary the grain tables.
+ * Uses the grain table cache as a true grain table.
+ */
+static int vmdkStreamAllocGrain(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
+ uint64_t uSector, PVDIOCTX pIoCtx,
+ uint64_t cbWrite)
+{
+ uint32_t uGrain;
+ uint32_t uGDEntry, uLastGDEntry;
+ uint32_t cbGrain = 0;
+ uint32_t uCacheLine, uCacheEntry;
+ const void *pData;
+ int rc;
+
+ /* Very strict requirements: always write at least one full grain, with
+ * proper alignment. Everything else would require reading of already
+ * written data, which we don't support for obvious reasons. The only
+ * exception is the last grain, and only if the image size specifies
+ * that only some portion holds data. In any case the write must be
+ * within the image limits, no "overshoot" allowed. */
+ if ( cbWrite == 0
+ || ( cbWrite < VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain)
+ && pExtent->cNominalSectors - uSector >= pExtent->cSectorsPerGrain)
+ || uSector % pExtent->cSectorsPerGrain
+ || uSector + VMDK_BYTE2SECTOR(cbWrite) > pExtent->cNominalSectors)
+ return VERR_INVALID_PARAMETER;
+
+ /* Clip write range to at most the rest of the grain. */
+ cbWrite = RT_MIN(cbWrite, VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain - uSector % pExtent->cSectorsPerGrain));
+
+ /* Do not allow to go back. */
+ uGrain = uSector / pExtent->cSectorsPerGrain;
+ uCacheLine = uGrain % pExtent->cGTEntries / VMDK_GT_CACHELINE_SIZE;
+ uCacheEntry = uGrain % VMDK_GT_CACHELINE_SIZE;
+ uGDEntry = uGrain / pExtent->cGTEntries;
+ uLastGDEntry = pExtent->uLastGrainAccess / pExtent->cGTEntries;
+ if (uGrain < pExtent->uLastGrainAccess)
+ return VERR_VD_VMDK_INVALID_WRITE;
+
+ /* Zero byte write optimization. Since we don't tell VBoxHDD that we need
+ * to allocate something, we also need to detect the situation ourself. */
+ if ( !(pImage->uOpenFlags & VD_OPEN_FLAGS_HONOR_ZEROES)
+ && vdIfIoIntIoCtxIsZero(pImage->pIfIo, pIoCtx, cbWrite, true /* fAdvance */))
+ return VINF_SUCCESS;
+
+ if (uGDEntry != uLastGDEntry)
+ {
+ rc = vmdkStreamFlushGT(pImage, pExtent, uLastGDEntry);
+ if (RT_FAILURE(rc))
+ return rc;
+ vmdkStreamClearGT(pImage, pExtent);
+ for (uint32_t i = uLastGDEntry + 1; i < uGDEntry; i++)
+ {
+ rc = vmdkStreamFlushGT(pImage, pExtent, i);
+ if (RT_FAILURE(rc))
+ return rc;
+ }
+ }
+
+ uint64_t uFileOffset;
+ uFileOffset = pExtent->uAppendPosition;
+ if (!uFileOffset)
+ return VERR_INTERNAL_ERROR;
+ /* Align to sector, as the previous write could have been any size. */
+ uFileOffset = RT_ALIGN_64(uFileOffset, 512);
+
+ /* Paranoia check: extent type, grain table buffer presence and
+ * grain table buffer space. Also grain table entry must be clear. */
+ if ( pExtent->enmType != VMDKETYPE_HOSTED_SPARSE
+ || !pImage->pGTCache
+ || pExtent->cGTEntries > VMDK_GT_CACHE_SIZE * VMDK_GT_CACHELINE_SIZE
+ || pImage->pGTCache->aGTCache[uCacheLine].aGTData[uCacheEntry])
+ return VERR_INTERNAL_ERROR;
+
+ /* Update grain table entry. */
+ pImage->pGTCache->aGTCache[uCacheLine].aGTData[uCacheEntry] = VMDK_BYTE2SECTOR(uFileOffset);
+
+ if (cbWrite != VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain))
+ {
+ vdIfIoIntIoCtxCopyFrom(pImage->pIfIo, pIoCtx, pExtent->pvGrain, cbWrite);
+ memset((char *)pExtent->pvGrain + cbWrite, '\0',
+ VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain) - cbWrite);
+ pData = pExtent->pvGrain;
+ }
+ else
+ {
+ RTSGSEG Segment;
+ unsigned cSegments = 1;
+ size_t cbSeg = 0;
+
+ cbSeg = vdIfIoIntIoCtxSegArrayCreate(pImage->pIfIo, pIoCtx, &Segment,
+ &cSegments, VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain));
+ Assert(cbSeg == VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain));
+ pData = Segment.pvSeg;
+ }
+ rc = vmdkFileDeflateSync(pImage, pExtent, uFileOffset, pData,
+ VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain),
+ uSector, &cbGrain);
+ if (RT_FAILURE(rc))
+ {
+ pExtent->uGrainSectorAbs = 0;
+ AssertRC(rc);
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write compressed data block in '%s'"), pExtent->pszFullname);
+ }
+ pExtent->uLastGrainAccess = uGrain;
+ pExtent->uAppendPosition += cbGrain;
+
+ return rc;
+}
+
+/**
+ * Internal: Updates the grain table during grain allocation.
+ */
+static int vmdkAllocGrainGTUpdate(PVMDKIMAGE pImage, PVMDKEXTENT pExtent, PVDIOCTX pIoCtx,
+ PVMDKGRAINALLOCASYNC pGrainAlloc)
+{
+ int rc = VINF_SUCCESS;
+ PVMDKGTCACHE pCache = pImage->pGTCache;
+ uint32_t aGTDataTmp[VMDK_GT_CACHELINE_SIZE];
+ uint32_t uGTHash, uGTBlockIndex;
+ uint64_t uGTSector, uRGTSector, uGTBlock;
+ uint64_t uSector = pGrainAlloc->uSector;
+ PVMDKGTCACHEENTRY pGTCacheEntry;
+
+ LogFlowFunc(("pImage=%#p pExtent=%#p pCache=%#p pIoCtx=%#p pGrainAlloc=%#p\n",
+ pImage, pExtent, pCache, pIoCtx, pGrainAlloc));
+
+ uGTSector = pGrainAlloc->uGTSector;
+ uRGTSector = pGrainAlloc->uRGTSector;
+ LogFlow(("uGTSector=%llu uRGTSector=%llu\n", uGTSector, uRGTSector));
+
+ /* Update the grain table (and the cache). */
+ uGTBlock = uSector / (pExtent->cSectorsPerGrain * VMDK_GT_CACHELINE_SIZE);
+ uGTHash = vmdkGTCacheHash(pCache, uGTBlock, pExtent->uExtent);
+ pGTCacheEntry = &pCache->aGTCache[uGTHash];
+ if ( pGTCacheEntry->uExtent != pExtent->uExtent
+ || pGTCacheEntry->uGTBlock != uGTBlock)
+ {
+ /* Cache miss, fetch data from disk. */
+ LogFlow(("Cache miss, fetch data from disk\n"));
+ PVDMETAXFER pMetaXfer = NULL;
+ rc = vdIfIoIntFileReadMeta(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uGTSector) + (uGTBlock % (pExtent->cGTEntries / VMDK_GT_CACHELINE_SIZE)) * sizeof(aGTDataTmp),
+ aGTDataTmp, sizeof(aGTDataTmp), pIoCtx,
+ &pMetaXfer, vmdkAllocGrainComplete, pGrainAlloc);
+ if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
+ {
+ pGrainAlloc->cIoXfersPending++;
+ pGrainAlloc->fGTUpdateNeeded = true;
+ /* Leave early, we will be called again after the read completed. */
+ LogFlowFunc(("Metadata read in progress, leaving\n"));
+ return rc;
+ }
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot read allocated grain table entry in '%s'"), pExtent->pszFullname);
+ vdIfIoIntMetaXferRelease(pImage->pIfIo, pMetaXfer);
+ pGTCacheEntry->uExtent = pExtent->uExtent;
+ pGTCacheEntry->uGTBlock = uGTBlock;
+ for (unsigned i = 0; i < VMDK_GT_CACHELINE_SIZE; i++)
+ pGTCacheEntry->aGTData[i] = RT_LE2H_U32(aGTDataTmp[i]);
+ }
+ else
+ {
+ /* Cache hit. Convert grain table block back to disk format, otherwise
+ * the code below will write garbage for all but the updated entry. */
+ for (unsigned i = 0; i < VMDK_GT_CACHELINE_SIZE; i++)
+ aGTDataTmp[i] = RT_H2LE_U32(pGTCacheEntry->aGTData[i]);
+ }
+ pGrainAlloc->fGTUpdateNeeded = false;
+ uGTBlockIndex = (uSector / pExtent->cSectorsPerGrain) % VMDK_GT_CACHELINE_SIZE;
+ aGTDataTmp[uGTBlockIndex] = RT_H2LE_U32(VMDK_BYTE2SECTOR(pGrainAlloc->uGrainOffset));
+ pGTCacheEntry->aGTData[uGTBlockIndex] = VMDK_BYTE2SECTOR(pGrainAlloc->uGrainOffset);
+ /* Update grain table on disk. */
+ rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uGTSector) + (uGTBlock % (pExtent->cGTEntries / VMDK_GT_CACHELINE_SIZE)) * sizeof(aGTDataTmp),
+ aGTDataTmp, sizeof(aGTDataTmp), pIoCtx,
+ vmdkAllocGrainComplete, pGrainAlloc);
+ if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
+ pGrainAlloc->cIoXfersPending++;
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write updated grain table in '%s'"), pExtent->pszFullname);
+ if (pExtent->pRGD)
+ {
+ /* Update backup grain table on disk. */
+ rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uRGTSector) + (uGTBlock % (pExtent->cGTEntries / VMDK_GT_CACHELINE_SIZE)) * sizeof(aGTDataTmp),
+ aGTDataTmp, sizeof(aGTDataTmp), pIoCtx,
+ vmdkAllocGrainComplete, pGrainAlloc);
+ if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
+ pGrainAlloc->cIoXfersPending++;
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write updated backup grain table in '%s'"), pExtent->pszFullname);
+ }
+
+ LogFlowFunc(("leaving rc=%Rrc\n", rc));
+ return rc;
+}
+
+/**
+ * Internal - complete the grain allocation by updating disk grain table if required.
+ */
+static int vmdkAllocGrainComplete(void *pBackendData, PVDIOCTX pIoCtx, void *pvUser, int rcReq)
+{
+ RT_NOREF1(rcReq);
+ int rc = VINF_SUCCESS;
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ PVMDKGRAINALLOCASYNC pGrainAlloc = (PVMDKGRAINALLOCASYNC)pvUser;
+
+ LogFlowFunc(("pBackendData=%#p pIoCtx=%#p pvUser=%#p rcReq=%Rrc\n",
+ pBackendData, pIoCtx, pvUser, rcReq));
+
+ pGrainAlloc->cIoXfersPending--;
+ if (!pGrainAlloc->cIoXfersPending && pGrainAlloc->fGTUpdateNeeded)
+ rc = vmdkAllocGrainGTUpdate(pImage, pGrainAlloc->pExtent, pIoCtx, pGrainAlloc);
+
+ if (!pGrainAlloc->cIoXfersPending)
+ {
+ /* Grain allocation completed. */
+ RTMemFree(pGrainAlloc);
+ }
+
+ LogFlowFunc(("Leaving rc=%Rrc\n", rc));
+ return rc;
+}
+
+/**
+ * Internal. Allocates a new grain table (if necessary).
+ */
+static int vmdkAllocGrain(PVMDKIMAGE pImage, PVMDKEXTENT pExtent, PVDIOCTX pIoCtx,
+ uint64_t uSector, uint64_t cbWrite)
+{
+ PVMDKGTCACHE pCache = pImage->pGTCache; NOREF(pCache);
+ uint64_t uGDIndex, uGTSector, uRGTSector;
+ uint64_t uFileOffset;
+ PVMDKGRAINALLOCASYNC pGrainAlloc = NULL;
+ int rc;
+
+ LogFlowFunc(("pCache=%#p pExtent=%#p pIoCtx=%#p uSector=%llu cbWrite=%llu\n",
+ pCache, pExtent, pIoCtx, uSector, cbWrite));
+
+ pGrainAlloc = (PVMDKGRAINALLOCASYNC)RTMemAllocZ(sizeof(VMDKGRAINALLOCASYNC));
+ if (!pGrainAlloc)
+ return VERR_NO_MEMORY;
+
+ pGrainAlloc->pExtent = pExtent;
+ pGrainAlloc->uSector = uSector;
+
+ uGDIndex = uSector / pExtent->cSectorsPerGDE;
+ if (uGDIndex >= pExtent->cGDEntries)
+ {
+ RTMemFree(pGrainAlloc);
+ return VERR_OUT_OF_RANGE;
+ }
+ uGTSector = pExtent->pGD[uGDIndex];
+ if (pExtent->pRGD)
+ uRGTSector = pExtent->pRGD[uGDIndex];
+ else
+ uRGTSector = 0; /**< avoid compiler warning */
+ if (!uGTSector)
+ {
+ LogFlow(("Allocating new grain table\n"));
+
+ /* There is no grain table referenced by this grain directory
+ * entry. So there is absolutely no data in this area. Allocate
+ * a new grain table and put the reference to it in the GDs. */
+ uFileOffset = pExtent->uAppendPosition;
+ if (!uFileOffset)
+ {
+ RTMemFree(pGrainAlloc);
+ return VERR_INTERNAL_ERROR;
+ }
+ Assert(!(uFileOffset % 512));
+
+ uFileOffset = RT_ALIGN_64(uFileOffset, 512);
+ uGTSector = VMDK_BYTE2SECTOR(uFileOffset);
+
+ /* Normally the grain table is preallocated for hosted sparse extents
+ * that support more than 32 bit sector numbers. So this shouldn't
+ * ever happen on a valid extent. */
+ if (uGTSector > UINT32_MAX)
+ {
+ RTMemFree(pGrainAlloc);
+ return VERR_VD_VMDK_INVALID_HEADER;
+ }
+
+ /* Write grain table by writing the required number of grain table
+ * cache chunks. Allocate memory dynamically here or we flood the
+ * metadata cache with very small entries. */
+ size_t cbGTDataTmp = pExtent->cGTEntries * sizeof(uint32_t);
+ uint32_t *paGTDataTmp = (uint32_t *)RTMemTmpAllocZ(cbGTDataTmp);
+
+ if (!paGTDataTmp)
+ {
+ RTMemFree(pGrainAlloc);
+ return VERR_NO_MEMORY;
+ }
+
+ memset(paGTDataTmp, '\0', cbGTDataTmp);
+ rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uGTSector),
+ paGTDataTmp, cbGTDataTmp, pIoCtx,
+ vmdkAllocGrainComplete, pGrainAlloc);
+ if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
+ pGrainAlloc->cIoXfersPending++;
+ else if (RT_FAILURE(rc))
+ {
+ RTMemTmpFree(paGTDataTmp);
+ RTMemFree(pGrainAlloc);
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write grain table allocation in '%s'"), pExtent->pszFullname);
+ }
+ pExtent->uAppendPosition = RT_ALIGN_64( pExtent->uAppendPosition
+ + cbGTDataTmp, 512);
+
+ if (pExtent->pRGD)
+ {
+ AssertReturn(!uRGTSector, VERR_VD_VMDK_INVALID_HEADER);
+ uFileOffset = pExtent->uAppendPosition;
+ if (!uFileOffset)
+ return VERR_INTERNAL_ERROR;
+ Assert(!(uFileOffset % 512));
+ uRGTSector = VMDK_BYTE2SECTOR(uFileOffset);
+
+ /* Normally the redundant grain table is preallocated for hosted
+ * sparse extents that support more than 32 bit sector numbers. So
+ * this shouldn't ever happen on a valid extent. */
+ if (uRGTSector > UINT32_MAX)
+ {
+ RTMemTmpFree(paGTDataTmp);
+ return VERR_VD_VMDK_INVALID_HEADER;
+ }
+
+ /* Write grain table by writing the required number of grain table
+ * cache chunks. Allocate memory dynamically here or we flood the
+ * metadata cache with very small entries. */
+ rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uRGTSector),
+ paGTDataTmp, cbGTDataTmp, pIoCtx,
+ vmdkAllocGrainComplete, pGrainAlloc);
+ if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
+ pGrainAlloc->cIoXfersPending++;
+ else if (RT_FAILURE(rc))
+ {
+ RTMemTmpFree(paGTDataTmp);
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write backup grain table allocation in '%s'"), pExtent->pszFullname);
+ }
+
+ pExtent->uAppendPosition = pExtent->uAppendPosition + cbGTDataTmp;
+ }
+
+ RTMemTmpFree(paGTDataTmp);
+
+ /* Update the grain directory on disk (doing it before writing the
+ * grain table will result in a garbled extent if the operation is
+ * aborted for some reason. Otherwise the worst that can happen is
+ * some unused sectors in the extent. */
+ uint32_t uGTSectorLE = RT_H2LE_U64(uGTSector);
+ rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(pExtent->uSectorGD) + uGDIndex * sizeof(uGTSectorLE),
+ &uGTSectorLE, sizeof(uGTSectorLE), pIoCtx,
+ vmdkAllocGrainComplete, pGrainAlloc);
+ if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
+ pGrainAlloc->cIoXfersPending++;
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write grain directory entry in '%s'"), pExtent->pszFullname);
+ if (pExtent->pRGD)
+ {
+ uint32_t uRGTSectorLE = RT_H2LE_U64(uRGTSector);
+ rc = vdIfIoIntFileWriteMeta(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(pExtent->uSectorRGD) + uGDIndex * sizeof(uGTSectorLE),
+ &uRGTSectorLE, sizeof(uRGTSectorLE), pIoCtx,
+ vmdkAllocGrainComplete, pGrainAlloc);
+ if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
+ pGrainAlloc->cIoXfersPending++;
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write backup grain directory entry in '%s'"), pExtent->pszFullname);
+ }
+
+ /* As the final step update the in-memory copy of the GDs. */
+ pExtent->pGD[uGDIndex] = uGTSector;
+ if (pExtent->pRGD)
+ pExtent->pRGD[uGDIndex] = uRGTSector;
+ }
+
+ LogFlow(("uGTSector=%llu uRGTSector=%llu\n", uGTSector, uRGTSector));
+ pGrainAlloc->uGTSector = uGTSector;
+ pGrainAlloc->uRGTSector = uRGTSector;
+
+ uFileOffset = pExtent->uAppendPosition;
+ if (!uFileOffset)
+ return VERR_INTERNAL_ERROR;
+ Assert(!(uFileOffset % 512));
+
+ pGrainAlloc->uGrainOffset = uFileOffset;
+
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ AssertMsgReturn(vdIfIoIntIoCtxIsSynchronous(pImage->pIfIo, pIoCtx),
+ ("Accesses to stream optimized images must be synchronous\n"),
+ VERR_INVALID_STATE);
+
+ if (cbWrite != VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain))
+ return vdIfError(pImage->pIfError, VERR_INTERNAL_ERROR, RT_SRC_POS, N_("VMDK: not enough data for a compressed data block in '%s'"), pExtent->pszFullname);
+
+ /* Invalidate cache, just in case some code incorrectly allows mixing
+ * of reads and writes. Normally shouldn't be needed. */
+ pExtent->uGrainSectorAbs = 0;
+
+ /* Write compressed data block and the markers. */
+ uint32_t cbGrain = 0;
+ size_t cbSeg = 0;
+ RTSGSEG Segment;
+ unsigned cSegments = 1;
+
+ cbSeg = vdIfIoIntIoCtxSegArrayCreate(pImage->pIfIo, pIoCtx, &Segment,
+ &cSegments, cbWrite);
+ Assert(cbSeg == cbWrite);
+
+ rc = vmdkFileDeflateSync(pImage, pExtent, uFileOffset,
+ Segment.pvSeg, cbWrite, uSector, &cbGrain);
+ if (RT_FAILURE(rc))
+ {
+ AssertRC(rc);
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write allocated compressed data block in '%s'"), pExtent->pszFullname);
+ }
+ pExtent->uLastGrainAccess = uSector / pExtent->cSectorsPerGrain;
+ pExtent->uAppendPosition += cbGrain;
+ }
+ else
+ {
+ /* Write the data. Always a full grain, or we're in big trouble. */
+ rc = vdIfIoIntFileWriteUser(pImage->pIfIo, pExtent->pFile->pStorage,
+ uFileOffset, pIoCtx, cbWrite,
+ vmdkAllocGrainComplete, pGrainAlloc);
+ if (rc == VERR_VD_ASYNC_IO_IN_PROGRESS)
+ pGrainAlloc->cIoXfersPending++;
+ else if (RT_FAILURE(rc))
+ return vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: cannot write allocated data block in '%s'"), pExtent->pszFullname);
+
+ pExtent->uAppendPosition += cbWrite;
+ }
+
+ rc = vmdkAllocGrainGTUpdate(pImage, pExtent, pIoCtx, pGrainAlloc);
+
+ if (!pGrainAlloc->cIoXfersPending)
+ {
+ /* Grain allocation completed. */
+ RTMemFree(pGrainAlloc);
+ }
+
+ LogFlowFunc(("leaving rc=%Rrc\n", rc));
+
+ return rc;
+}
+
+/**
+ * Internal. Reads the contents by sequentially going over the compressed
+ * grains (hoping that they are in sequence).
+ */
+static int vmdkStreamReadSequential(PVMDKIMAGE pImage, PVMDKEXTENT pExtent,
+ uint64_t uSector, PVDIOCTX pIoCtx,
+ uint64_t cbRead)
+{
+ int rc;
+
+ LogFlowFunc(("pImage=%#p pExtent=%#p uSector=%llu pIoCtx=%#p cbRead=%llu\n",
+ pImage, pExtent, uSector, pIoCtx, cbRead));
+
+ AssertMsgReturn(vdIfIoIntIoCtxIsSynchronous(pImage->pIfIo, pIoCtx),
+ ("Async I/O not supported for sequential stream optimized images\n"),
+ VERR_INVALID_STATE);
+
+ /* Do not allow to go back. */
+ uint32_t uGrain = uSector / pExtent->cSectorsPerGrain;
+ if (uGrain < pExtent->uLastGrainAccess)
+ return VERR_VD_VMDK_INVALID_STATE;
+ pExtent->uLastGrainAccess = uGrain;
+
+ /* After a previous error do not attempt to recover, as it would need
+ * seeking (in the general case backwards which is forbidden). */
+ if (!pExtent->uGrainSectorAbs)
+ return VERR_VD_VMDK_INVALID_STATE;
+
+ /* Check if we need to read something from the image or if what we have
+ * in the buffer is good to fulfill the request. */
+ if (!pExtent->cbGrainStreamRead || uGrain > pExtent->uGrain)
+ {
+ uint32_t uGrainSectorAbs = pExtent->uGrainSectorAbs
+ + VMDK_BYTE2SECTOR(pExtent->cbGrainStreamRead);
+
+ /* Get the marker from the next data block - and skip everything which
+ * is not a compressed grain. If it's a compressed grain which is for
+ * the requested sector (or after), read it. */
+ VMDKMARKER Marker;
+ do
+ {
+ RT_ZERO(Marker);
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uGrainSectorAbs),
+ &Marker, RT_UOFFSETOF(VMDKMARKER, uType));
+ if (RT_FAILURE(rc))
+ return rc;
+ Marker.uSector = RT_LE2H_U64(Marker.uSector);
+ Marker.cbSize = RT_LE2H_U32(Marker.cbSize);
+
+ if (Marker.cbSize == 0)
+ {
+ /* A marker for something else than a compressed grain. */
+ rc = vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uGrainSectorAbs)
+ + RT_UOFFSETOF(VMDKMARKER, uType),
+ &Marker.uType, sizeof(Marker.uType));
+ if (RT_FAILURE(rc))
+ return rc;
+ Marker.uType = RT_LE2H_U32(Marker.uType);
+ switch (Marker.uType)
+ {
+ case VMDK_MARKER_EOS:
+ uGrainSectorAbs++;
+ /* Read (or mostly skip) to the end of file. Uses the
+ * Marker (LBA sector) as it is unused anyway. This
+ * makes sure that really everything is read in the
+ * success case. If this read fails it means the image
+ * is truncated, but this is harmless so ignore. */
+ vdIfIoIntFileReadSync(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uGrainSectorAbs)
+ + 511,
+ &Marker.uSector, 1);
+ break;
+ case VMDK_MARKER_GT:
+ uGrainSectorAbs += 1 + VMDK_BYTE2SECTOR(pExtent->cGTEntries * sizeof(uint32_t));
+ break;
+ case VMDK_MARKER_GD:
+ uGrainSectorAbs += 1 + VMDK_BYTE2SECTOR(RT_ALIGN(pExtent->cGDEntries * sizeof(uint32_t), 512));
+ break;
+ case VMDK_MARKER_FOOTER:
+ uGrainSectorAbs += 2;
+ break;
+ case VMDK_MARKER_UNSPECIFIED:
+ /* Skip over the contents of the unspecified marker
+ * type 4 which exists in some vSphere created files. */
+ /** @todo figure out what the payload means. */
+ uGrainSectorAbs += 1;
+ break;
+ default:
+ AssertMsgFailed(("VMDK: corrupted marker, type=%#x\n", Marker.uType));
+ pExtent->uGrainSectorAbs = 0;
+ return VERR_VD_VMDK_INVALID_STATE;
+ }
+ pExtent->cbGrainStreamRead = 0;
+ }
+ else
+ {
+ /* A compressed grain marker. If it is at/after what we're
+ * interested in read and decompress data. */
+ if (uSector > Marker.uSector + pExtent->cSectorsPerGrain)
+ {
+ uGrainSectorAbs += VMDK_BYTE2SECTOR(RT_ALIGN(Marker.cbSize + RT_UOFFSETOF(VMDKMARKER, uType), 512));
+ continue;
+ }
+ uint64_t uLBA = 0;
+ uint32_t cbGrainStreamRead = 0;
+ rc = vmdkFileInflateSync(pImage, pExtent,
+ VMDK_SECTOR2BYTE(uGrainSectorAbs),
+ pExtent->pvGrain,
+ VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain),
+ &Marker, &uLBA, &cbGrainStreamRead);
+ if (RT_FAILURE(rc))
+ {
+ pExtent->uGrainSectorAbs = 0;
+ return rc;
+ }
+ if ( pExtent->uGrain
+ && uLBA / pExtent->cSectorsPerGrain <= pExtent->uGrain)
+ {
+ pExtent->uGrainSectorAbs = 0;
+ return VERR_VD_VMDK_INVALID_STATE;
+ }
+ pExtent->uGrain = uLBA / pExtent->cSectorsPerGrain;
+ pExtent->cbGrainStreamRead = cbGrainStreamRead;
+ break;
+ }
+ } while (Marker.uType != VMDK_MARKER_EOS);
+
+ pExtent->uGrainSectorAbs = uGrainSectorAbs;
+
+ if (!pExtent->cbGrainStreamRead && Marker.uType == VMDK_MARKER_EOS)
+ {
+ pExtent->uGrain = UINT32_MAX;
+ /* Must set a non-zero value for pExtent->cbGrainStreamRead or
+ * the next read would try to get more data, and we're at EOF. */
+ pExtent->cbGrainStreamRead = 1;
+ }
+ }
+
+ if (pExtent->uGrain > uSector / pExtent->cSectorsPerGrain)
+ {
+ /* The next data block we have is not for this area, so just return
+ * that there is no data. */
+ LogFlowFunc(("returns VERR_VD_BLOCK_FREE\n"));
+ return VERR_VD_BLOCK_FREE;
+ }
+
+ uint32_t uSectorInGrain = uSector % pExtent->cSectorsPerGrain;
+ vdIfIoIntIoCtxCopyTo(pImage->pIfIo, pIoCtx,
+ (uint8_t *)pExtent->pvGrain + VMDK_SECTOR2BYTE(uSectorInGrain),
+ cbRead);
+ LogFlowFunc(("returns VINF_SUCCESS\n"));
+ return VINF_SUCCESS;
+}
+
+/**
+ * Replaces a fragment of a string with the specified string.
+ *
+ * @returns Pointer to the allocated UTF-8 string.
+ * @param pszWhere UTF-8 string to search in.
+ * @param pszWhat UTF-8 string to search for.
+ * @param pszByWhat UTF-8 string to replace the found string with.
+ *
+ * @note r=bird: This is only used by vmdkRenameWorker(). The first use is
+ * for updating the base name in the descriptor, the second is for
+ * generating new filenames for extents. This code borked when
+ * RTPathAbs started correcting the driver letter case on windows,
+ * when strstr failed because the pExtent->pszFullname was not
+ * subjected to RTPathAbs but while pExtent->pszFullname was. I fixed
+ * this by apply RTPathAbs to the places it wasn't applied.
+ *
+ * However, this highlights some undocumented ASSUMPTIONS as well as
+ * terrible short commings of the approach.
+ *
+ * Given the right filename, it may also screw up the descriptor. Take
+ * the descriptor text 'RW 2048 SPARSE "Test0.vmdk"' for instance,
+ * we'll be asked to replace "Test0" with something, no problem. No,
+ * imagine 'RW 2048 SPARSE "SPARSE.vmdk"', 'RW 2048 SPARSE "RW.vmdk"'
+ * or 'RW 2048 SPARSE "2048.vmdk"', and the strstr approach falls on
+ * its bum. The descriptor string must be parsed and reconstructed,
+ * the lazy strstr approach doesn't cut it.
+ *
+ * I'm also curious as to what would be the correct escaping of '"' in
+ * the file name and how that is supposed to be handled, because it
+ * needs to be or such names must be rejected in several places (maybe
+ * they are, I didn't check).
+ *
+ * When this function is used to replace the start of a path, I think
+ * the assumption from the prep/setup code is that we kind of knows
+ * what we're working on (I could be wrong). However, using strstr
+ * instead of strncmp/RTStrNICmp makes no sense and isn't future proof.
+ * Especially on unix systems, weird stuff could happen if someone
+ * unwittingly tinkers with the prep/setup code. What should really be
+ * done here is using a new RTPathStartEx function that (via flags)
+ * allows matching partial final component and returns the length of
+ * what it matched up (in case it skipped slashes and '.' components).
+ *
+ */
+static char *vmdkStrReplace(const char *pszWhere, const char *pszWhat,
+ const char *pszByWhat)
+{
+ AssertPtr(pszWhere);
+ AssertPtr(pszWhat);
+ AssertPtr(pszByWhat);
+ const char *pszFoundStr = strstr(pszWhere, pszWhat);
+ if (!pszFoundStr)
+ {
+ LogFlowFunc(("Failed to find '%s' in '%s'!\n", pszWhat, pszWhere));
+ return NULL;
+ }
+ size_t cbFinal = strlen(pszWhere) + 1 + strlen(pszByWhat) - strlen(pszWhat);
+ char *pszNewStr = (char *)RTMemAlloc(cbFinal);
+ if (pszNewStr)
+ {
+ char *pszTmp = pszNewStr;
+ memcpy(pszTmp, pszWhere, pszFoundStr - pszWhere);
+ pszTmp += pszFoundStr - pszWhere;
+ memcpy(pszTmp, pszByWhat, strlen(pszByWhat));
+ pszTmp += strlen(pszByWhat);
+ strcpy(pszTmp, pszFoundStr + strlen(pszWhat));
+ }
+ return pszNewStr;
+}
+
+
+/** @copydoc VDIMAGEBACKEND::pfnProbe */
+static DECLCALLBACK(int) vmdkProbe(const char *pszFilename, PVDINTERFACE pVDIfsDisk,
+ PVDINTERFACE pVDIfsImage, VDTYPE enmDesiredType, VDTYPE *penmType)
+{
+ RT_NOREF(enmDesiredType);
+ LogFlowFunc(("pszFilename=\"%s\" pVDIfsDisk=%#p pVDIfsImage=%#p penmType=%#p\n",
+ pszFilename, pVDIfsDisk, pVDIfsImage, penmType));
+
+ AssertReturn((VALID_PTR(pszFilename) && *pszFilename), VERR_INVALID_PARAMETER);
+
+ int rc = VINF_SUCCESS;
+ PVMDKIMAGE pImage = (PVMDKIMAGE)RTMemAllocZ(RT_UOFFSETOF(VMDKIMAGE, RegionList.aRegions[1]));
+ if (RT_LIKELY(pImage))
+ {
+ pImage->pszFilename = pszFilename;
+ pImage->pFile = NULL;
+ pImage->pExtents = NULL;
+ pImage->pFiles = NULL;
+ pImage->pGTCache = NULL;
+ pImage->pDescData = NULL;
+ pImage->pVDIfsDisk = pVDIfsDisk;
+ pImage->pVDIfsImage = pVDIfsImage;
+ /** @todo speed up this test open (VD_OPEN_FLAGS_INFO) by skipping as
+ * much as possible in vmdkOpenImage. */
+ rc = vmdkOpenImage(pImage, VD_OPEN_FLAGS_INFO | VD_OPEN_FLAGS_READONLY);
+ vmdkFreeImage(pImage, false, false /*fFlush*/);
+ RTMemFree(pImage);
+
+ if (RT_SUCCESS(rc))
+ *penmType = VDTYPE_HDD;
+ }
+ else
+ rc = VERR_NO_MEMORY;
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnOpen */
+static DECLCALLBACK(int) vmdkOpen(const char *pszFilename, unsigned uOpenFlags,
+ PVDINTERFACE pVDIfsDisk, PVDINTERFACE pVDIfsImage,
+ VDTYPE enmType, void **ppBackendData)
+{
+ RT_NOREF1(enmType); /**< @todo r=klaus make use of the type info. */
+
+ LogFlowFunc(("pszFilename=\"%s\" uOpenFlags=%#x pVDIfsDisk=%#p pVDIfsImage=%#p enmType=%u ppBackendData=%#p\n",
+ pszFilename, uOpenFlags, pVDIfsDisk, pVDIfsImage, enmType, ppBackendData));
+ int rc;
+
+ /* Check open flags. All valid flags are supported. */
+ AssertReturn(!(uOpenFlags & ~VD_OPEN_FLAGS_MASK), VERR_INVALID_PARAMETER);
+ AssertReturn((VALID_PTR(pszFilename) && *pszFilename), VERR_INVALID_PARAMETER);
+
+ PVMDKIMAGE pImage = (PVMDKIMAGE)RTMemAllocZ(RT_UOFFSETOF(VMDKIMAGE, RegionList.aRegions[1]));
+ if (RT_LIKELY(pImage))
+ {
+ pImage->pszFilename = pszFilename;
+ pImage->pFile = NULL;
+ pImage->pExtents = NULL;
+ pImage->pFiles = NULL;
+ pImage->pGTCache = NULL;
+ pImage->pDescData = NULL;
+ pImage->pVDIfsDisk = pVDIfsDisk;
+ pImage->pVDIfsImage = pVDIfsImage;
+
+ rc = vmdkOpenImage(pImage, uOpenFlags);
+ if (RT_SUCCESS(rc))
+ *ppBackendData = pImage;
+ else
+ RTMemFree(pImage);
+ }
+ else
+ rc = VERR_NO_MEMORY;
+
+ LogFlowFunc(("returns %Rrc (pBackendData=%#p)\n", rc, *ppBackendData));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnCreate */
+static DECLCALLBACK(int) vmdkCreate(const char *pszFilename, uint64_t cbSize,
+ unsigned uImageFlags, const char *pszComment,
+ PCVDGEOMETRY pPCHSGeometry, PCVDGEOMETRY pLCHSGeometry,
+ PCRTUUID pUuid, unsigned uOpenFlags,
+ unsigned uPercentStart, unsigned uPercentSpan,
+ PVDINTERFACE pVDIfsDisk, PVDINTERFACE pVDIfsImage,
+ PVDINTERFACE pVDIfsOperation, VDTYPE enmType,
+ void **ppBackendData)
+{
+ LogFlowFunc(("pszFilename=\"%s\" cbSize=%llu uImageFlags=%#x pszComment=\"%s\" pPCHSGeometry=%#p pLCHSGeometry=%#p Uuid=%RTuuid uOpenFlags=%#x uPercentStart=%u uPercentSpan=%u pVDIfsDisk=%#p pVDIfsImage=%#p pVDIfsOperation=%#p enmType=%u ppBackendData=%#p\n",
+ pszFilename, cbSize, uImageFlags, pszComment, pPCHSGeometry, pLCHSGeometry, pUuid, uOpenFlags, uPercentStart, uPercentSpan, pVDIfsDisk, pVDIfsImage, pVDIfsOperation, enmType, ppBackendData));
+ int rc;
+
+ /* Check the VD container type and image flags. */
+ if ( enmType != VDTYPE_HDD
+ || (uImageFlags & ~VD_VMDK_IMAGE_FLAGS_MASK) != 0)
+ return VERR_VD_INVALID_TYPE;
+
+ /* Check size. Maximum 256TB-64K for sparse images, otherwise unlimited. */
+ if ( !cbSize
+ || (!(uImageFlags & VD_IMAGE_FLAGS_FIXED) && cbSize >= _1T * 256 - _64K))
+ return VERR_VD_INVALID_SIZE;
+
+ /* Check image flags for invalid combinations. */
+ if ( (uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ && (uImageFlags & ~(VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED | VD_IMAGE_FLAGS_DIFF)))
+ return VERR_INVALID_PARAMETER;
+
+ /* Check open flags. All valid flags are supported. */
+ AssertReturn(!(uOpenFlags & ~VD_OPEN_FLAGS_MASK), VERR_INVALID_PARAMETER);
+ AssertReturn( VALID_PTR(pszFilename)
+ && *pszFilename
+ && VALID_PTR(pPCHSGeometry)
+ && VALID_PTR(pLCHSGeometry)
+ && !( uImageFlags & VD_VMDK_IMAGE_FLAGS_ESX
+ && !(uImageFlags & VD_IMAGE_FLAGS_FIXED)),
+ VERR_INVALID_PARAMETER);
+
+ PVMDKIMAGE pImage = (PVMDKIMAGE)RTMemAllocZ(RT_UOFFSETOF(VMDKIMAGE, RegionList.aRegions[1]));
+ if (RT_LIKELY(pImage))
+ {
+ PVDINTERFACEPROGRESS pIfProgress = VDIfProgressGet(pVDIfsOperation);
+
+ pImage->pszFilename = pszFilename;
+ pImage->pFile = NULL;
+ pImage->pExtents = NULL;
+ pImage->pFiles = NULL;
+ pImage->pGTCache = NULL;
+ pImage->pDescData = NULL;
+ pImage->pVDIfsDisk = pVDIfsDisk;
+ pImage->pVDIfsImage = pVDIfsImage;
+ /* Descriptors for split images can be pretty large, especially if the
+ * filename is long. So prepare for the worst, and allocate quite some
+ * memory for the descriptor in this case. */
+ if (uImageFlags & VD_VMDK_IMAGE_FLAGS_SPLIT_2G)
+ pImage->cbDescAlloc = VMDK_SECTOR2BYTE(200);
+ else
+ pImage->cbDescAlloc = VMDK_SECTOR2BYTE(20);
+ pImage->pDescData = (char *)RTMemAllocZ(pImage->cbDescAlloc);
+ if (RT_LIKELY(pImage->pDescData))
+ {
+ rc = vmdkCreateImage(pImage, cbSize, uImageFlags, pszComment,
+ pPCHSGeometry, pLCHSGeometry, pUuid,
+ pIfProgress, uPercentStart, uPercentSpan);
+ if (RT_SUCCESS(rc))
+ {
+ /* So far the image is opened in read/write mode. Make sure the
+ * image is opened in read-only mode if the caller requested that. */
+ if (uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ {
+ vmdkFreeImage(pImage, false, true /*fFlush*/);
+ rc = vmdkOpenImage(pImage, uOpenFlags);
+ }
+
+ if (RT_SUCCESS(rc))
+ *ppBackendData = pImage;
+ }
+
+ if (RT_FAILURE(rc))
+ RTMemFree(pImage->pDescData);
+ }
+ else
+ rc = VERR_NO_MEMORY;
+
+ if (RT_FAILURE(rc))
+ RTMemFree(pImage);
+ }
+ else
+ rc = VERR_NO_MEMORY;
+
+ LogFlowFunc(("returns %Rrc (pBackendData=%#p)\n", rc, *ppBackendData));
+ return rc;
+}
+
+/**
+ * Prepares the state for renaming a VMDK image, setting up the state and allocating
+ * memory.
+ *
+ * @returns VBox status code.
+ * @param pImage VMDK image instance.
+ * @param pRenameState The state to initialize.
+ * @param pszFilename The new filename.
+ */
+static int vmdkRenameStatePrepare(PVMDKIMAGE pImage, PVMDKRENAMESTATE pRenameState, const char *pszFilename)
+{
+ AssertReturn(RTPathFilename(pszFilename) != NULL, VERR_INVALID_PARAMETER);
+
+ int rc = VINF_SUCCESS;
+
+ memset(&pRenameState->DescriptorCopy, 0, sizeof(pRenameState->DescriptorCopy));
+
+ /*
+ * Allocate an array to store both old and new names of renamed files
+ * in case we have to roll back the changes. Arrays are initialized
+ * with zeros. We actually save stuff when and if we change it.
+ */
+ pRenameState->cExtents = pImage->cExtents;
+ pRenameState->apszOldName = (char **)RTMemTmpAllocZ((pRenameState->cExtents + 1) * sizeof(char *));
+ pRenameState->apszNewName = (char **)RTMemTmpAllocZ((pRenameState->cExtents + 1) * sizeof(char *));
+ pRenameState->apszNewLines = (char **)RTMemTmpAllocZ(pRenameState->cExtents * sizeof(char *));
+ if ( pRenameState->apszOldName
+ && pRenameState->apszNewName
+ && pRenameState->apszNewLines)
+ {
+ /* Save the descriptor size and position. */
+ if (pImage->pDescData)
+ {
+ /* Separate descriptor file. */
+ pRenameState->fEmbeddedDesc = false;
+ }
+ else
+ {
+ /* Embedded descriptor file. */
+ pRenameState->ExtentCopy = pImage->pExtents[0];
+ pRenameState->fEmbeddedDesc = true;
+ }
+
+ /* Save the descriptor content. */
+ pRenameState->DescriptorCopy.cLines = pImage->Descriptor.cLines;
+ for (unsigned i = 0; i < pRenameState->DescriptorCopy.cLines; i++)
+ {
+ pRenameState->DescriptorCopy.aLines[i] = RTStrDup(pImage->Descriptor.aLines[i]);
+ if (!pRenameState->DescriptorCopy.aLines[i])
+ {
+ rc = VERR_NO_MEMORY;
+ break;
+ }
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ /* Prepare both old and new base names used for string replacement. */
+ pRenameState->pszNewBaseName = RTStrDup(RTPathFilename(pszFilename));
+ AssertReturn(pRenameState->pszNewBaseName, VERR_NO_STR_MEMORY);
+ RTPathStripSuffix(pRenameState->pszNewBaseName);
+
+ pRenameState->pszOldBaseName = RTStrDup(RTPathFilename(pImage->pszFilename));
+ AssertReturn(pRenameState->pszOldBaseName, VERR_NO_STR_MEMORY);
+ RTPathStripSuffix(pRenameState->pszOldBaseName);
+
+ /* Prepare both old and new full names used for string replacement.
+ Note! Must abspath the stuff here, so the strstr weirdness later in
+ the renaming process get a match against abspath'ed extent paths.
+ See RTPathAbsDup call in vmdkDescriptorReadSparse(). */
+ pRenameState->pszNewFullName = RTPathAbsDup(pszFilename);
+ AssertReturn(pRenameState->pszNewFullName, VERR_NO_STR_MEMORY);
+ RTPathStripSuffix(pRenameState->pszNewFullName);
+
+ pRenameState->pszOldFullName = RTPathAbsDup(pImage->pszFilename);
+ AssertReturn(pRenameState->pszOldFullName, VERR_NO_STR_MEMORY);
+ RTPathStripSuffix(pRenameState->pszOldFullName);
+
+ /* Save the old name for easy access to the old descriptor file. */
+ pRenameState->pszOldDescName = RTStrDup(pImage->pszFilename);
+ AssertReturn(pRenameState->pszOldDescName, VERR_NO_STR_MEMORY);
+
+ /* Save old image name. */
+ pRenameState->pszOldImageName = pImage->pszFilename;
+ }
+ }
+ else
+ rc = VERR_NO_TMP_MEMORY;
+
+ return rc;
+}
+
+/**
+ * Destroys the given rename state, freeing all allocated memory.
+ *
+ * @returns nothing.
+ * @param pRenameState The rename state to destroy.
+ */
+static void vmdkRenameStateDestroy(PVMDKRENAMESTATE pRenameState)
+{
+ for (unsigned i = 0; i < pRenameState->DescriptorCopy.cLines; i++)
+ if (pRenameState->DescriptorCopy.aLines[i])
+ RTStrFree(pRenameState->DescriptorCopy.aLines[i]);
+ if (pRenameState->apszOldName)
+ {
+ for (unsigned i = 0; i <= pRenameState->cExtents; i++)
+ if (pRenameState->apszOldName[i])
+ RTStrFree(pRenameState->apszOldName[i]);
+ RTMemTmpFree(pRenameState->apszOldName);
+ }
+ if (pRenameState->apszNewName)
+ {
+ for (unsigned i = 0; i <= pRenameState->cExtents; i++)
+ if (pRenameState->apszNewName[i])
+ RTStrFree(pRenameState->apszNewName[i]);
+ RTMemTmpFree(pRenameState->apszNewName);
+ }
+ if (pRenameState->apszNewLines)
+ {
+ for (unsigned i = 0; i < pRenameState->cExtents; i++)
+ if (pRenameState->apszNewLines[i])
+ RTStrFree(pRenameState->apszNewLines[i]);
+ RTMemTmpFree(pRenameState->apszNewLines);
+ }
+ if (pRenameState->pszOldDescName)
+ RTStrFree(pRenameState->pszOldDescName);
+ if (pRenameState->pszOldBaseName)
+ RTStrFree(pRenameState->pszOldBaseName);
+ if (pRenameState->pszNewBaseName)
+ RTStrFree(pRenameState->pszNewBaseName);
+ if (pRenameState->pszOldFullName)
+ RTStrFree(pRenameState->pszOldFullName);
+ if (pRenameState->pszNewFullName)
+ RTStrFree(pRenameState->pszNewFullName);
+}
+
+/**
+ * Rolls back the rename operation to the original state.
+ *
+ * @returns VBox status code.
+ * @param pImage VMDK image instance.
+ * @param pRenameState The rename state.
+ */
+static int vmdkRenameRollback(PVMDKIMAGE pImage, PVMDKRENAMESTATE pRenameState)
+{
+ int rc = VINF_SUCCESS;
+
+ if (!pRenameState->fImageFreed)
+ {
+ /*
+ * Some extents may have been closed, close the rest. We will
+ * re-open the whole thing later.
+ */
+ vmdkFreeImage(pImage, false, true /*fFlush*/);
+ }
+
+ /* Rename files back. */
+ for (unsigned i = 0; i <= pRenameState->cExtents; i++)
+ {
+ if (pRenameState->apszOldName[i])
+ {
+ rc = vdIfIoIntFileMove(pImage->pIfIo, pRenameState->apszNewName[i], pRenameState->apszOldName[i], 0);
+ AssertRC(rc);
+ }
+ }
+ /* Restore the old descriptor. */
+ PVMDKFILE pFile;
+ rc = vmdkFileOpen(pImage, &pFile, NULL, pRenameState->pszOldDescName,
+ VDOpenFlagsToFileOpenFlags(VD_OPEN_FLAGS_NORMAL,
+ false /* fCreate */));
+ AssertRC(rc);
+ if (pRenameState->fEmbeddedDesc)
+ {
+ pRenameState->ExtentCopy.pFile = pFile;
+ pImage->pExtents = &pRenameState->ExtentCopy;
+ }
+ else
+ {
+ /* Shouldn't be null for separate descriptor.
+ * There will be no access to the actual content.
+ */
+ pImage->pDescData = pRenameState->pszOldDescName;
+ pImage->pFile = pFile;
+ }
+ pImage->Descriptor = pRenameState->DescriptorCopy;
+ vmdkWriteDescriptor(pImage, NULL);
+ vmdkFileClose(pImage, &pFile, false);
+ /* Get rid of the stuff we implanted. */
+ pImage->pExtents = NULL;
+ pImage->pFile = NULL;
+ pImage->pDescData = NULL;
+ /* Re-open the image back. */
+ pImage->pszFilename = pRenameState->pszOldImageName;
+ rc = vmdkOpenImage(pImage, pImage->uOpenFlags);
+
+ return rc;
+}
+
+/**
+ * Rename worker doing the real work.
+ *
+ * @returns VBox status code.
+ * @param pImage VMDK image instance.
+ * @param pRenameState The rename state.
+ * @param pszFilename The new filename.
+ */
+static int vmdkRenameWorker(PVMDKIMAGE pImage, PVMDKRENAMESTATE pRenameState, const char *pszFilename)
+{
+ int rc = VINF_SUCCESS;
+ unsigned i, line;
+
+ /* Update the descriptor with modified extent names. */
+ for (i = 0, line = pImage->Descriptor.uFirstExtent;
+ i < pRenameState->cExtents;
+ i++, line = pImage->Descriptor.aNextLines[line])
+ {
+ /* Update the descriptor. */
+ pRenameState->apszNewLines[i] = vmdkStrReplace(pImage->Descriptor.aLines[line],
+ pRenameState->pszOldBaseName,
+ pRenameState->pszNewBaseName);
+ if (!pRenameState->apszNewLines[i])
+ {
+ rc = VERR_NO_MEMORY;
+ break;
+ }
+ pImage->Descriptor.aLines[line] = pRenameState->apszNewLines[i];
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ /* Make sure the descriptor gets written back. */
+ pImage->Descriptor.fDirty = true;
+ /* Flush the descriptor now, in case it is embedded. */
+ vmdkFlushImage(pImage, NULL);
+
+ /* Close and rename/move extents. */
+ for (i = 0; i < pRenameState->cExtents; i++)
+ {
+ PVMDKEXTENT pExtent = &pImage->pExtents[i];
+ /* Compose new name for the extent. */
+ pRenameState->apszNewName[i] = vmdkStrReplace(pExtent->pszFullname,
+ pRenameState->pszOldFullName,
+ pRenameState->pszNewFullName);
+ if (!pRenameState->apszNewName[i])
+ {
+ rc = VERR_NO_MEMORY;
+ break;
+ }
+ /* Close the extent file. */
+ rc = vmdkFileClose(pImage, &pExtent->pFile, false);
+ if (RT_FAILURE(rc))
+ break;;
+
+ /* Rename the extent file. */
+ rc = vdIfIoIntFileMove(pImage->pIfIo, pExtent->pszFullname, pRenameState->apszNewName[i], 0);
+ if (RT_FAILURE(rc))
+ break;
+ /* Remember the old name. */
+ pRenameState->apszOldName[i] = RTStrDup(pExtent->pszFullname);
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ /* Release all old stuff. */
+ rc = vmdkFreeImage(pImage, false, true /*fFlush*/);
+ if (RT_SUCCESS(rc))
+ {
+ pRenameState->fImageFreed = true;
+
+ /* Last elements of new/old name arrays are intended for
+ * storing descriptor's names.
+ */
+ pRenameState->apszNewName[pRenameState->cExtents] = RTStrDup(pszFilename);
+ /* Rename the descriptor file if it's separate. */
+ if (!pRenameState->fEmbeddedDesc)
+ {
+ rc = vdIfIoIntFileMove(pImage->pIfIo, pImage->pszFilename, pRenameState->apszNewName[pRenameState->cExtents], 0);
+ if (RT_SUCCESS(rc))
+ {
+ /* Save old name only if we may need to change it back. */
+ pRenameState->apszOldName[pRenameState->cExtents] = RTStrDup(pszFilename);
+ }
+ }
+
+ /* Update pImage with the new information. */
+ pImage->pszFilename = pszFilename;
+
+ /* Open the new image. */
+ rc = vmdkOpenImage(pImage, pImage->uOpenFlags);
+ }
+ }
+ }
+
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnRename */
+static DECLCALLBACK(int) vmdkRename(void *pBackendData, const char *pszFilename)
+{
+ LogFlowFunc(("pBackendData=%#p pszFilename=%#p\n", pBackendData, pszFilename));
+
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ VMDKRENAMESTATE RenameState;
+
+ memset(&RenameState, 0, sizeof(RenameState));
+
+ /* Check arguments. */
+ AssertReturn(( pImage
+ && VALID_PTR(pszFilename)
+ && *pszFilename
+ && !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_RAWDISK)), VERR_INVALID_PARAMETER);
+
+ int rc = vmdkRenameStatePrepare(pImage, &RenameState, pszFilename);
+ if (RT_SUCCESS(rc))
+ {
+ /* --- Up to this point we have not done any damage yet. --- */
+
+ rc = vmdkRenameWorker(pImage, &RenameState, pszFilename);
+ /* Roll back all changes in case of failure. */
+ if (RT_FAILURE(rc))
+ {
+ int rrc = vmdkRenameRollback(pImage, &RenameState);
+ AssertRC(rrc);
+ }
+ }
+
+ vmdkRenameStateDestroy(&RenameState);
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnClose */
+static DECLCALLBACK(int) vmdkClose(void *pBackendData, bool fDelete)
+{
+ LogFlowFunc(("pBackendData=%#p fDelete=%d\n", pBackendData, fDelete));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ int rc = vmdkFreeImage(pImage, fDelete, true /*fFlush*/);
+ RTMemFree(pImage);
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnRead */
+static DECLCALLBACK(int) vmdkRead(void *pBackendData, uint64_t uOffset, size_t cbToRead,
+ PVDIOCTX pIoCtx, size_t *pcbActuallyRead)
+{
+ LogFlowFunc(("pBackendData=%#p uOffset=%llu pIoCtx=%#p cbToRead=%zu pcbActuallyRead=%#p\n",
+ pBackendData, uOffset, pIoCtx, cbToRead, pcbActuallyRead));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ AssertPtr(pImage);
+ Assert(uOffset % 512 == 0);
+ Assert(cbToRead % 512 == 0);
+ AssertReturn((VALID_PTR(pIoCtx) && cbToRead), VERR_INVALID_PARAMETER);
+ AssertReturn(uOffset + cbToRead <= pImage->cbSize, VERR_INVALID_PARAMETER);
+
+ /* Find the extent and check access permissions as defined in the extent descriptor. */
+ PVMDKEXTENT pExtent;
+ uint64_t uSectorExtentRel;
+ int rc = vmdkFindExtent(pImage, VMDK_BYTE2SECTOR(uOffset),
+ &pExtent, &uSectorExtentRel);
+ if ( RT_SUCCESS(rc)
+ && pExtent->enmAccess != VMDKACCESS_NOACCESS)
+ {
+ /* Clip read range to remain in this extent. */
+ cbToRead = RT_MIN(cbToRead, VMDK_SECTOR2BYTE(pExtent->uSectorOffset + pExtent->cNominalSectors - uSectorExtentRel));
+
+ /* Handle the read according to the current extent type. */
+ switch (pExtent->enmType)
+ {
+ case VMDKETYPE_HOSTED_SPARSE:
+ {
+ uint64_t uSectorExtentAbs;
+
+ rc = vmdkGetSector(pImage, pIoCtx, pExtent, uSectorExtentRel, &uSectorExtentAbs);
+ if (RT_FAILURE(rc))
+ break;
+ /* Clip read range to at most the rest of the grain. */
+ cbToRead = RT_MIN(cbToRead, VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain - uSectorExtentRel % pExtent->cSectorsPerGrain));
+ Assert(!(cbToRead % 512));
+ if (uSectorExtentAbs == 0)
+ {
+ if ( !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ || !(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY)
+ || !(pImage->uOpenFlags & VD_OPEN_FLAGS_SEQUENTIAL))
+ rc = VERR_VD_BLOCK_FREE;
+ else
+ rc = vmdkStreamReadSequential(pImage, pExtent,
+ uSectorExtentRel,
+ pIoCtx, cbToRead);
+ }
+ else
+ {
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ AssertMsg(vdIfIoIntIoCtxIsSynchronous(pImage->pIfIo, pIoCtx),
+ ("Async I/O is not supported for stream optimized VMDK's\n"));
+
+ uint32_t uSectorInGrain = uSectorExtentRel % pExtent->cSectorsPerGrain;
+ uSectorExtentAbs -= uSectorInGrain;
+ if (pExtent->uGrainSectorAbs != uSectorExtentAbs)
+ {
+ uint64_t uLBA = 0; /* gcc maybe uninitialized */
+ rc = vmdkFileInflateSync(pImage, pExtent,
+ VMDK_SECTOR2BYTE(uSectorExtentAbs),
+ pExtent->pvGrain,
+ VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain),
+ NULL, &uLBA, NULL);
+ if (RT_FAILURE(rc))
+ {
+ pExtent->uGrainSectorAbs = 0;
+ break;
+ }
+ pExtent->uGrainSectorAbs = uSectorExtentAbs;
+ pExtent->uGrain = uSectorExtentRel / pExtent->cSectorsPerGrain;
+ Assert(uLBA == uSectorExtentRel);
+ }
+ vdIfIoIntIoCtxCopyTo(pImage->pIfIo, pIoCtx,
+ (uint8_t *)pExtent->pvGrain
+ + VMDK_SECTOR2BYTE(uSectorInGrain),
+ cbToRead);
+ }
+ else
+ rc = vdIfIoIntFileReadUser(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uSectorExtentAbs),
+ pIoCtx, cbToRead);
+ }
+ break;
+ }
+ case VMDKETYPE_VMFS:
+ case VMDKETYPE_FLAT:
+ rc = vdIfIoIntFileReadUser(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uSectorExtentRel),
+ pIoCtx, cbToRead);
+ break;
+ case VMDKETYPE_ZERO:
+ {
+ size_t cbSet;
+
+ cbSet = vdIfIoIntIoCtxSet(pImage->pIfIo, pIoCtx, 0, cbToRead);
+ Assert(cbSet == cbToRead);
+ break;
+ }
+ }
+ if (pcbActuallyRead)
+ *pcbActuallyRead = cbToRead;
+ }
+ else if (RT_SUCCESS(rc))
+ rc = VERR_VD_VMDK_INVALID_STATE;
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnWrite */
+static DECLCALLBACK(int) vmdkWrite(void *pBackendData, uint64_t uOffset, size_t cbToWrite,
+ PVDIOCTX pIoCtx, size_t *pcbWriteProcess, size_t *pcbPreRead,
+ size_t *pcbPostRead, unsigned fWrite)
+{
+ LogFlowFunc(("pBackendData=%#p uOffset=%llu pIoCtx=%#p cbToWrite=%zu pcbWriteProcess=%#p pcbPreRead=%#p pcbPostRead=%#p\n",
+ pBackendData, uOffset, pIoCtx, cbToWrite, pcbWriteProcess, pcbPreRead, pcbPostRead));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ int rc;
+
+ AssertPtr(pImage);
+ Assert(uOffset % 512 == 0);
+ Assert(cbToWrite % 512 == 0);
+ AssertReturn((VALID_PTR(pIoCtx) && cbToWrite), VERR_INVALID_PARAMETER);
+
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ {
+ PVMDKEXTENT pExtent;
+ uint64_t uSectorExtentRel;
+ uint64_t uSectorExtentAbs;
+
+ /* No size check here, will do that later when the extent is located.
+ * There are sparse images out there which according to the spec are
+ * invalid, because the total size is not a multiple of the grain size.
+ * Also for sparse images which are stitched together in odd ways (not at
+ * grain boundaries, and with the nominal size not being a multiple of the
+ * grain size), this would prevent writing to the last grain. */
+
+ rc = vmdkFindExtent(pImage, VMDK_BYTE2SECTOR(uOffset),
+ &pExtent, &uSectorExtentRel);
+ if (RT_SUCCESS(rc))
+ {
+ if ( pExtent->enmAccess != VMDKACCESS_READWRITE
+ && ( !(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ && !pImage->pExtents[0].uAppendPosition
+ && pExtent->enmAccess != VMDKACCESS_READONLY))
+ rc = VERR_VD_VMDK_INVALID_STATE;
+ else
+ {
+ /* Handle the write according to the current extent type. */
+ switch (pExtent->enmType)
+ {
+ case VMDKETYPE_HOSTED_SPARSE:
+ rc = vmdkGetSector(pImage, pIoCtx, pExtent, uSectorExtentRel, &uSectorExtentAbs);
+ if (RT_SUCCESS(rc))
+ {
+ if ( pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED
+ && uSectorExtentRel < (uint64_t)pExtent->uLastGrainAccess * pExtent->cSectorsPerGrain)
+ rc = VERR_VD_VMDK_INVALID_WRITE;
+ else
+ {
+ /* Clip write range to at most the rest of the grain. */
+ cbToWrite = RT_MIN(cbToWrite,
+ VMDK_SECTOR2BYTE( pExtent->cSectorsPerGrain
+ - uSectorExtentRel % pExtent->cSectorsPerGrain));
+ if (uSectorExtentAbs == 0)
+ {
+ if (!(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
+ {
+ if (cbToWrite == VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain))
+ {
+ /* Full block write to a previously unallocated block.
+ * Check if the caller wants to avoid the automatic alloc. */
+ if (!(fWrite & VD_WRITE_NO_ALLOC))
+ {
+ /* Allocate GT and find out where to store the grain. */
+ rc = vmdkAllocGrain(pImage, pExtent, pIoCtx,
+ uSectorExtentRel, cbToWrite);
+ }
+ else
+ rc = VERR_VD_BLOCK_FREE;
+ *pcbPreRead = 0;
+ *pcbPostRead = 0;
+ }
+ else
+ {
+ /* Clip write range to remain in this extent. */
+ cbToWrite = RT_MIN(cbToWrite,
+ VMDK_SECTOR2BYTE( pExtent->uSectorOffset
+ + pExtent->cNominalSectors - uSectorExtentRel));
+ *pcbPreRead = VMDK_SECTOR2BYTE(uSectorExtentRel % pExtent->cSectorsPerGrain);
+ *pcbPostRead = VMDK_SECTOR2BYTE(pExtent->cSectorsPerGrain) - cbToWrite - *pcbPreRead;
+ rc = VERR_VD_BLOCK_FREE;
+ }
+ }
+ else
+ rc = vmdkStreamAllocGrain(pImage, pExtent, uSectorExtentRel,
+ pIoCtx, cbToWrite);
+ }
+ else
+ {
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ /* A partial write to a streamOptimized image is simply
+ * invalid. It requires rewriting already compressed data
+ * which is somewhere between expensive and impossible. */
+ rc = VERR_VD_VMDK_INVALID_STATE;
+ pExtent->uGrainSectorAbs = 0;
+ AssertRC(rc);
+ }
+ else
+ {
+ Assert(!(pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED));
+ rc = vdIfIoIntFileWriteUser(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uSectorExtentAbs),
+ pIoCtx, cbToWrite, NULL, NULL);
+ }
+ }
+ }
+ }
+ break;
+ case VMDKETYPE_VMFS:
+ case VMDKETYPE_FLAT:
+ /* Clip write range to remain in this extent. */
+ cbToWrite = RT_MIN(cbToWrite, VMDK_SECTOR2BYTE(pExtent->uSectorOffset + pExtent->cNominalSectors - uSectorExtentRel));
+ rc = vdIfIoIntFileWriteUser(pImage->pIfIo, pExtent->pFile->pStorage,
+ VMDK_SECTOR2BYTE(uSectorExtentRel),
+ pIoCtx, cbToWrite, NULL, NULL);
+ break;
+ case VMDKETYPE_ZERO:
+ /* Clip write range to remain in this extent. */
+ cbToWrite = RT_MIN(cbToWrite, VMDK_SECTOR2BYTE(pExtent->uSectorOffset + pExtent->cNominalSectors - uSectorExtentRel));
+ break;
+ }
+ }
+
+ if (pcbWriteProcess)
+ *pcbWriteProcess = cbToWrite;
+ }
+ }
+ else
+ rc = VERR_VD_IMAGE_READ_ONLY;
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnFlush */
+static DECLCALLBACK(int) vmdkFlush(void *pBackendData, PVDIOCTX pIoCtx)
+{
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ return vmdkFlushImage(pImage, pIoCtx);
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnGetVersion */
+static DECLCALLBACK(unsigned) vmdkGetVersion(void *pBackendData)
+{
+ LogFlowFunc(("pBackendData=%#p\n", pBackendData));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ AssertPtrReturn(pImage, 0);
+
+ return VMDK_IMAGE_VERSION;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnGetFileSize */
+static DECLCALLBACK(uint64_t) vmdkGetFileSize(void *pBackendData)
+{
+ LogFlowFunc(("pBackendData=%#p\n", pBackendData));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ uint64_t cb = 0;
+
+ AssertPtrReturn(pImage, 0);
+
+ if (pImage->pFile != NULL)
+ {
+ uint64_t cbFile;
+ int rc = vdIfIoIntFileGetSize(pImage->pIfIo, pImage->pFile->pStorage, &cbFile);
+ if (RT_SUCCESS(rc))
+ cb += cbFile;
+ }
+ for (unsigned i = 0; i < pImage->cExtents; i++)
+ {
+ if (pImage->pExtents[i].pFile != NULL)
+ {
+ uint64_t cbFile;
+ int rc = vdIfIoIntFileGetSize(pImage->pIfIo, pImage->pExtents[i].pFile->pStorage, &cbFile);
+ if (RT_SUCCESS(rc))
+ cb += cbFile;
+ }
+ }
+
+ LogFlowFunc(("returns %lld\n", cb));
+ return cb;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnGetPCHSGeometry */
+static DECLCALLBACK(int) vmdkGetPCHSGeometry(void *pBackendData, PVDGEOMETRY pPCHSGeometry)
+{
+ LogFlowFunc(("pBackendData=%#p pPCHSGeometry=%#p\n", pBackendData, pPCHSGeometry));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ int rc = VINF_SUCCESS;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ if (pImage->PCHSGeometry.cCylinders)
+ *pPCHSGeometry = pImage->PCHSGeometry;
+ else
+ rc = VERR_VD_GEOMETRY_NOT_SET;
+
+ LogFlowFunc(("returns %Rrc (PCHS=%u/%u/%u)\n", rc, pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads, pPCHSGeometry->cSectors));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnSetPCHSGeometry */
+static DECLCALLBACK(int) vmdkSetPCHSGeometry(void *pBackendData, PCVDGEOMETRY pPCHSGeometry)
+{
+ LogFlowFunc(("pBackendData=%#p pPCHSGeometry=%#p PCHS=%u/%u/%u\n",
+ pBackendData, pPCHSGeometry, pPCHSGeometry->cCylinders, pPCHSGeometry->cHeads, pPCHSGeometry->cSectors));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ int rc = VINF_SUCCESS;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ {
+ if (!(pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
+ {
+ rc = vmdkDescSetPCHSGeometry(pImage, pPCHSGeometry);
+ if (RT_SUCCESS(rc))
+ pImage->PCHSGeometry = *pPCHSGeometry;
+ }
+ else
+ rc = VERR_NOT_SUPPORTED;
+ }
+ else
+ rc = VERR_VD_IMAGE_READ_ONLY;
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnGetLCHSGeometry */
+static DECLCALLBACK(int) vmdkGetLCHSGeometry(void *pBackendData, PVDGEOMETRY pLCHSGeometry)
+{
+ LogFlowFunc(("pBackendData=%#p pLCHSGeometry=%#p\n", pBackendData, pLCHSGeometry));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ int rc = VINF_SUCCESS;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ if (pImage->LCHSGeometry.cCylinders)
+ *pLCHSGeometry = pImage->LCHSGeometry;
+ else
+ rc = VERR_VD_GEOMETRY_NOT_SET;
+
+ LogFlowFunc(("returns %Rrc (LCHS=%u/%u/%u)\n", rc, pLCHSGeometry->cCylinders, pLCHSGeometry->cHeads, pLCHSGeometry->cSectors));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnSetLCHSGeometry */
+static DECLCALLBACK(int) vmdkSetLCHSGeometry(void *pBackendData, PCVDGEOMETRY pLCHSGeometry)
+{
+ LogFlowFunc(("pBackendData=%#p pLCHSGeometry=%#p LCHS=%u/%u/%u\n",
+ pBackendData, pLCHSGeometry, pLCHSGeometry->cCylinders, pLCHSGeometry->cHeads, pLCHSGeometry->cSectors));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ int rc = VINF_SUCCESS;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ {
+ if (!(pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
+ {
+ rc = vmdkDescSetLCHSGeometry(pImage, pLCHSGeometry);
+ if (RT_SUCCESS(rc))
+ pImage->LCHSGeometry = *pLCHSGeometry;
+ }
+ else
+ rc = VERR_NOT_SUPPORTED;
+ }
+ else
+ rc = VERR_VD_IMAGE_READ_ONLY;
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnQueryRegions */
+static DECLCALLBACK(int) vmdkQueryRegions(void *pBackendData, PCVDREGIONLIST *ppRegionList)
+{
+ LogFlowFunc(("pBackendData=%#p ppRegionList=%#p\n", pBackendData, ppRegionList));
+ PVMDKIMAGE pThis = (PVMDKIMAGE)pBackendData;
+
+ AssertPtrReturn(pThis, VERR_VD_NOT_OPENED);
+
+ *ppRegionList = &pThis->RegionList;
+ LogFlowFunc(("returns %Rrc\n", VINF_SUCCESS));
+ return VINF_SUCCESS;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnRegionListRelease */
+static DECLCALLBACK(void) vmdkRegionListRelease(void *pBackendData, PCVDREGIONLIST pRegionList)
+{
+ RT_NOREF1(pRegionList);
+ LogFlowFunc(("pBackendData=%#p pRegionList=%#p\n", pBackendData, pRegionList));
+ PVMDKIMAGE pThis = (PVMDKIMAGE)pBackendData;
+ AssertPtr(pThis); RT_NOREF(pThis);
+
+ /* Nothing to do here. */
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnGetImageFlags */
+static DECLCALLBACK(unsigned) vmdkGetImageFlags(void *pBackendData)
+{
+ LogFlowFunc(("pBackendData=%#p\n", pBackendData));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ AssertPtrReturn(pImage, 0);
+
+ LogFlowFunc(("returns %#x\n", pImage->uImageFlags));
+ return pImage->uImageFlags;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnGetOpenFlags */
+static DECLCALLBACK(unsigned) vmdkGetOpenFlags(void *pBackendData)
+{
+ LogFlowFunc(("pBackendData=%#p\n", pBackendData));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ AssertPtrReturn(pImage, 0);
+
+ LogFlowFunc(("returns %#x\n", pImage->uOpenFlags));
+ return pImage->uOpenFlags;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnSetOpenFlags */
+static DECLCALLBACK(int) vmdkSetOpenFlags(void *pBackendData, unsigned uOpenFlags)
+{
+ LogFlowFunc(("pBackendData=%#p uOpenFlags=%#x\n", pBackendData, uOpenFlags));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ int rc;
+
+ /* Image must be opened and the new flags must be valid. */
+ if (!pImage || (uOpenFlags & ~( VD_OPEN_FLAGS_READONLY | VD_OPEN_FLAGS_INFO
+ | VD_OPEN_FLAGS_ASYNC_IO | VD_OPEN_FLAGS_SHAREABLE
+ | VD_OPEN_FLAGS_SEQUENTIAL | VD_OPEN_FLAGS_SKIP_CONSISTENCY_CHECKS)))
+ rc = VERR_INVALID_PARAMETER;
+ else
+ {
+ /* StreamOptimized images need special treatment: reopen is prohibited. */
+ if (pImage->uImageFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED)
+ {
+ if (pImage->uOpenFlags == uOpenFlags)
+ rc = VINF_SUCCESS;
+ else
+ rc = VERR_INVALID_PARAMETER;
+ }
+ else
+ {
+ /* Implement this operation via reopening the image. */
+ vmdkFreeImage(pImage, false, true /*fFlush*/);
+ rc = vmdkOpenImage(pImage, uOpenFlags);
+ }
+ }
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnGetComment */
+static DECLCALLBACK(int) vmdkGetComment(void *pBackendData, char *pszComment, size_t cbComment)
+{
+ LogFlowFunc(("pBackendData=%#p pszComment=%#p cbComment=%zu\n", pBackendData, pszComment, cbComment));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ char *pszCommentEncoded = NULL;
+ int rc = vmdkDescDDBGetStr(pImage, &pImage->Descriptor,
+ "ddb.comment", &pszCommentEncoded);
+ if (rc == VERR_VD_VMDK_VALUE_NOT_FOUND)
+ {
+ pszCommentEncoded = NULL;
+ rc = VINF_SUCCESS;
+ }
+
+ if (RT_SUCCESS(rc))
+ {
+ if (pszComment && pszCommentEncoded)
+ rc = vmdkDecodeString(pszCommentEncoded, pszComment, cbComment);
+ else if (pszComment)
+ *pszComment = '\0';
+
+ if (pszCommentEncoded)
+ RTMemTmpFree(pszCommentEncoded);
+ }
+
+ LogFlowFunc(("returns %Rrc comment='%s'\n", rc, pszComment));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnSetComment */
+static DECLCALLBACK(int) vmdkSetComment(void *pBackendData, const char *pszComment)
+{
+ LogFlowFunc(("pBackendData=%#p pszComment=\"%s\"\n", pBackendData, pszComment));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ int rc;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ {
+ if (!(pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
+ rc = vmdkSetImageComment(pImage, pszComment);
+ else
+ rc = VERR_NOT_SUPPORTED;
+ }
+ else
+ rc = VERR_VD_IMAGE_READ_ONLY;
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnGetUuid */
+static DECLCALLBACK(int) vmdkGetUuid(void *pBackendData, PRTUUID pUuid)
+{
+ LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ *pUuid = pImage->ImageUuid;
+
+ LogFlowFunc(("returns %Rrc (%RTuuid)\n", VINF_SUCCESS, pUuid));
+ return VINF_SUCCESS;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnSetUuid */
+static DECLCALLBACK(int) vmdkSetUuid(void *pBackendData, PCRTUUID pUuid)
+{
+ LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ int rc = VINF_SUCCESS;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ {
+ if (!(pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
+ {
+ pImage->ImageUuid = *pUuid;
+ rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
+ VMDK_DDB_IMAGE_UUID, pUuid);
+ if (RT_FAILURE(rc))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
+ N_("VMDK: error storing image UUID in descriptor in '%s'"), pImage->pszFilename);
+ }
+ else
+ rc = VERR_NOT_SUPPORTED;
+ }
+ else
+ rc = VERR_VD_IMAGE_READ_ONLY;
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnGetModificationUuid */
+static DECLCALLBACK(int) vmdkGetModificationUuid(void *pBackendData, PRTUUID pUuid)
+{
+ LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ *pUuid = pImage->ModificationUuid;
+
+ LogFlowFunc(("returns %Rrc (%RTuuid)\n", VINF_SUCCESS, pUuid));
+ return VINF_SUCCESS;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnSetModificationUuid */
+static DECLCALLBACK(int) vmdkSetModificationUuid(void *pBackendData, PCRTUUID pUuid)
+{
+ LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ int rc = VINF_SUCCESS;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ {
+ if (!(pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
+ {
+ /* Only touch the modification uuid if it changed. */
+ if (RTUuidCompare(&pImage->ModificationUuid, pUuid))
+ {
+ pImage->ModificationUuid = *pUuid;
+ rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
+ VMDK_DDB_MODIFICATION_UUID, pUuid);
+ if (RT_FAILURE(rc))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing modification UUID in descriptor in '%s'"), pImage->pszFilename);
+ }
+ }
+ else
+ rc = VERR_NOT_SUPPORTED;
+ }
+ else
+ rc = VERR_VD_IMAGE_READ_ONLY;
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnGetParentUuid */
+static DECLCALLBACK(int) vmdkGetParentUuid(void *pBackendData, PRTUUID pUuid)
+{
+ LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ *pUuid = pImage->ParentUuid;
+
+ LogFlowFunc(("returns %Rrc (%RTuuid)\n", VINF_SUCCESS, pUuid));
+ return VINF_SUCCESS;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnSetParentUuid */
+static DECLCALLBACK(int) vmdkSetParentUuid(void *pBackendData, PCRTUUID pUuid)
+{
+ LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ int rc = VINF_SUCCESS;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ {
+ if (!(pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
+ {
+ pImage->ParentUuid = *pUuid;
+ rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
+ VMDK_DDB_PARENT_UUID, pUuid);
+ if (RT_FAILURE(rc))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS,
+ N_("VMDK: error storing parent image UUID in descriptor in '%s'"), pImage->pszFilename);
+ }
+ else
+ rc = VERR_NOT_SUPPORTED;
+ }
+ else
+ rc = VERR_VD_IMAGE_READ_ONLY;
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnGetParentModificationUuid */
+static DECLCALLBACK(int) vmdkGetParentModificationUuid(void *pBackendData, PRTUUID pUuid)
+{
+ LogFlowFunc(("pBackendData=%#p pUuid=%#p\n", pBackendData, pUuid));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ *pUuid = pImage->ParentModificationUuid;
+
+ LogFlowFunc(("returns %Rrc (%RTuuid)\n", VINF_SUCCESS, pUuid));
+ return VINF_SUCCESS;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnSetParentModificationUuid */
+static DECLCALLBACK(int) vmdkSetParentModificationUuid(void *pBackendData, PCRTUUID pUuid)
+{
+ LogFlowFunc(("pBackendData=%#p Uuid=%RTuuid\n", pBackendData, pUuid));
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+ int rc = VINF_SUCCESS;
+
+ AssertPtrReturn(pImage, VERR_VD_NOT_OPENED);
+
+ if (!(pImage->uOpenFlags & VD_OPEN_FLAGS_READONLY))
+ {
+ if (!(pImage->uOpenFlags & VD_VMDK_IMAGE_FLAGS_STREAM_OPTIMIZED))
+ {
+ pImage->ParentModificationUuid = *pUuid;
+ rc = vmdkDescDDBSetUuid(pImage, &pImage->Descriptor,
+ VMDK_DDB_PARENT_MODIFICATION_UUID, pUuid);
+ if (RT_FAILURE(rc))
+ rc = vdIfError(pImage->pIfError, rc, RT_SRC_POS, N_("VMDK: error storing parent image UUID in descriptor in '%s'"), pImage->pszFilename);
+ }
+ else
+ rc = VERR_NOT_SUPPORTED;
+ }
+ else
+ rc = VERR_VD_IMAGE_READ_ONLY;
+
+ LogFlowFunc(("returns %Rrc\n", rc));
+ return rc;
+}
+
+/** @copydoc VDIMAGEBACKEND::pfnDump */
+static DECLCALLBACK(void) vmdkDump(void *pBackendData)
+{
+ PVMDKIMAGE pImage = (PVMDKIMAGE)pBackendData;
+
+ AssertPtrReturnVoid(pImage);
+ vdIfErrorMessage(pImage->pIfError, "Header: Geometry PCHS=%u/%u/%u LCHS=%u/%u/%u cbSector=%llu\n",
+ pImage->PCHSGeometry.cCylinders, pImage->PCHSGeometry.cHeads, pImage->PCHSGeometry.cSectors,
+ pImage->LCHSGeometry.cCylinders, pImage->LCHSGeometry.cHeads, pImage->LCHSGeometry.cSectors,
+ VMDK_BYTE2SECTOR(pImage->cbSize));
+ vdIfErrorMessage(pImage->pIfError, "Header: uuidCreation={%RTuuid}\n", &pImage->ImageUuid);
+ vdIfErrorMessage(pImage->pIfError, "Header: uuidModification={%RTuuid}\n", &pImage->ModificationUuid);
+ vdIfErrorMessage(pImage->pIfError, "Header: uuidParent={%RTuuid}\n", &pImage->ParentUuid);
+ vdIfErrorMessage(pImage->pIfError, "Header: uuidParentModification={%RTuuid}\n", &pImage->ParentModificationUuid);
+}
+
+
+
+const VDIMAGEBACKEND g_VmdkBackend =
+{
+ /* u32Version */
+ VD_IMGBACKEND_VERSION,
+ /* pszBackendName */
+ "VMDK",
+ /* uBackendCaps */
+ VD_CAP_UUID | VD_CAP_CREATE_FIXED | VD_CAP_CREATE_DYNAMIC
+ | VD_CAP_CREATE_SPLIT_2G | VD_CAP_DIFF | VD_CAP_FILE | VD_CAP_ASYNC
+ | VD_CAP_VFS | VD_CAP_PREFERRED,
+ /* paFileExtensions */
+ s_aVmdkFileExtensions,
+ /* paConfigInfo */
+ NULL,
+ /* pfnProbe */
+ vmdkProbe,
+ /* pfnOpen */
+ vmdkOpen,
+ /* pfnCreate */
+ vmdkCreate,
+ /* pfnRename */
+ vmdkRename,
+ /* pfnClose */
+ vmdkClose,
+ /* pfnRead */
+ vmdkRead,
+ /* pfnWrite */
+ vmdkWrite,
+ /* pfnFlush */
+ vmdkFlush,
+ /* pfnDiscard */
+ NULL,
+ /* pfnGetVersion */
+ vmdkGetVersion,
+ /* pfnGetFileSize */
+ vmdkGetFileSize,
+ /* pfnGetPCHSGeometry */
+ vmdkGetPCHSGeometry,
+ /* pfnSetPCHSGeometry */
+ vmdkSetPCHSGeometry,
+ /* pfnGetLCHSGeometry */
+ vmdkGetLCHSGeometry,
+ /* pfnSetLCHSGeometry */
+ vmdkSetLCHSGeometry,
+ /* pfnQueryRegions */
+ vmdkQueryRegions,
+ /* pfnRegionListRelease */
+ vmdkRegionListRelease,
+ /* pfnGetImageFlags */
+ vmdkGetImageFlags,
+ /* pfnGetOpenFlags */
+ vmdkGetOpenFlags,
+ /* pfnSetOpenFlags */
+ vmdkSetOpenFlags,
+ /* pfnGetComment */
+ vmdkGetComment,
+ /* pfnSetComment */
+ vmdkSetComment,
+ /* pfnGetUuid */
+ vmdkGetUuid,
+ /* pfnSetUuid */
+ vmdkSetUuid,
+ /* pfnGetModificationUuid */
+ vmdkGetModificationUuid,
+ /* pfnSetModificationUuid */
+ vmdkSetModificationUuid,
+ /* pfnGetParentUuid */
+ vmdkGetParentUuid,
+ /* pfnSetParentUuid */
+ vmdkSetParentUuid,
+ /* pfnGetParentModificationUuid */
+ vmdkGetParentModificationUuid,
+ /* pfnSetParentModificationUuid */
+ vmdkSetParentModificationUuid,
+ /* pfnDump */
+ vmdkDump,
+ /* pfnGetTimestamp */
+ NULL,
+ /* pfnGetParentTimestamp */
+ NULL,
+ /* pfnSetParentTimestamp */
+ NULL,
+ /* pfnGetParentFilename */
+ NULL,
+ /* pfnSetParentFilename */
+ NULL,
+ /* pfnComposeLocation */
+ genericFileComposeLocation,
+ /* pfnComposeName */
+ genericFileComposeName,
+ /* pfnCompact */
+ NULL,
+ /* pfnResize */
+ NULL,
+ /* pfnRepair */
+ NULL,
+ /* pfnTraverseMetadata */
+ NULL,
+ /* u32VersionEnd */
+ VD_IMGBACKEND_VERSION
+};