/** @file * VMM - The Virtual Machine Monitor. */ /* * Copyright (C) 2006-2010 Oracle Corporation * * This file is part of VirtualBox Open Source Edition (OSE), as * available from http://www.virtualbox.org. This file is free software; * you can redistribute it and/or modify it under the terms of the GNU * General Public License (GPL) as published by the Free Software * Foundation, in version 2 as it comes in the "COPYING" file of the * VirtualBox OSE distribution. VirtualBox OSE is distributed in the * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind. * * The contents of this file may alternatively be used under the terms * of the Common Development and Distribution License Version 1.0 * (CDDL) only, as it comes in the "COPYING.CDDL" file of the * VirtualBox OSE distribution, in which case the provisions of the * CDDL are applicable instead of those of the GPL. * * You may elect to license modified versions of this file under the * terms and conditions of either the GPL or the CDDL or both. */ #ifndef ___VBox_vmm_vmm_h #define ___VBox_vmm_vmm_h #include #include #include #include #include RT_C_DECLS_BEGIN /** @defgroup grp_vmm The Virtual Machine Monitor API * @{ */ /** * World switcher identifiers. */ typedef enum VMMSWITCHER { /** The usual invalid 0. */ VMMSWITCHER_INVALID = 0, /** Switcher for 32-bit host to 32-bit shadow paging. */ VMMSWITCHER_32_TO_32, /** Switcher for 32-bit host paging to PAE shadow paging. */ VMMSWITCHER_32_TO_PAE, /** Switcher for 32-bit host paging to AMD64 shadow paging. */ VMMSWITCHER_32_TO_AMD64, /** Switcher for PAE host to 32-bit shadow paging. */ VMMSWITCHER_PAE_TO_32, /** Switcher for PAE host to PAE shadow paging. */ VMMSWITCHER_PAE_TO_PAE, /** Switcher for PAE host paging to AMD64 shadow paging. */ VMMSWITCHER_PAE_TO_AMD64, /** Switcher for AMD64 host paging to 32-bit shadow paging. */ VMMSWITCHER_AMD64_TO_32, /** Switcher for AMD64 host paging to PAE shadow paging. */ VMMSWITCHER_AMD64_TO_PAE, /** Switcher for AMD64 host paging to AMD64 shadow paging. */ VMMSWITCHER_AMD64_TO_AMD64, /** Used to make a count for array declarations and suchlike. */ VMMSWITCHER_MAX, /** The usual 32-bit paranoia. */ VMMSWITCHER_32BIT_HACK = 0x7fffffff } VMMSWITCHER; /** * VMMRZCallRing3 operations. */ typedef enum VMMCALLRING3 { /** Invalid operation. */ VMMCALLRING3_INVALID = 0, /** Acquire the PDM lock. */ VMMCALLRING3_PDM_LOCK, /** Acquire the PGM lock. */ VMMCALLRING3_PGM_LOCK, /** Grow the PGM shadow page pool. */ VMMCALLRING3_PGM_POOL_GROW, /** Maps a chunk into ring-3. */ VMMCALLRING3_PGM_MAP_CHUNK, /** Allocates more handy pages. */ VMMCALLRING3_PGM_ALLOCATE_HANDY_PAGES, /** Allocates a large (2MB) page. */ VMMCALLRING3_PGM_ALLOCATE_LARGE_HANDY_PAGE, /** Acquire the MM hypervisor heap lock. */ VMMCALLRING3_MMHYPER_LOCK, /** Replay the REM handler notifications. */ VMMCALLRING3_REM_REPLAY_HANDLER_NOTIFICATIONS, /** Flush the GC/R0 logger. */ VMMCALLRING3_VMM_LOGGER_FLUSH, /** Set the VM error message. */ VMMCALLRING3_VM_SET_ERROR, /** Set the VM runtime error message. */ VMMCALLRING3_VM_SET_RUNTIME_ERROR, /** Signal a ring 0 assertion. */ VMMCALLRING3_VM_R0_ASSERTION, /** Ring switch to force preemption. */ VMMCALLRING3_VM_R0_PREEMPT, /** Sync the FTM state with the standby node. */ VMMCALLRING3_FTM_SET_CHECKPOINT, /** The usual 32-bit hack. */ VMMCALLRING3_32BIT_HACK = 0x7fffffff } VMMCALLRING3; /** * VMMR3AtomicExecuteHandler callback function. * * @returns VBox status code. * @param pVM Pointer to the shared VM structure. * @param pvUser User specified argument * * @todo missing prefix. */ typedef DECLCALLBACK(int) FNATOMICHANDLER(PVM pVM, void *pvUser); /** Pointer to a FNMMATOMICHANDLER(). */ typedef FNATOMICHANDLER *PFNATOMICHANDLER; /** * Rendezvous callback. * * @returns VBox strict status code - EM scheduling. Do not return * informational status code other than the ones used by EM for * scheduling. * * @param pVM The VM handle. * @param pVCpu The handle of the calling virtual CPU. * @param pvUser The user argument. */ typedef DECLCALLBACK(VBOXSTRICTRC) FNVMMEMTRENDEZVOUS(PVM pVM, PVMCPU pVCpu, void *pvUser); /** Pointer to a rendezvous callback function. */ typedef FNVMMEMTRENDEZVOUS *PFNVMMEMTRENDEZVOUS; /** * Method table that the VMM uses to call back the user of the VMM. */ typedef struct VMM2USERMETHODS { /** Magic value (VMM2USERMETHODS_MAGIC). */ uint32_t u32Magic; /** Structure version (VMM2USERMETHODS_VERSION). */ uint32_t u32Version; /** * Save the VM state. * * @returns VBox status code. * @param pThis Pointer to the callback method table. * @param pUVM The user mode VM handle. * * @remarks This member shall be set to NULL if the operation is not * supported. */ DECLR3CALLBACKMEMBER(int, pfnSaveState,(PCVMM2USERMETHODS pThis, PUVM pUVM)); /** @todo Move pfnVMAtError and pfnCFGMConstructor here? */ /** * EMT initialization notification callback. * * This is intended for doing per-thread initialization for EMTs (like COM * init). * * @param pThis Pointer to the callback method table. * @param pUVM The user mode VM handle. * @param pUVCpu The user mode virtual CPU handle. * * @remarks This is optional and shall be set to NULL if not wanted. */ DECLR3CALLBACKMEMBER(void, pfnNotifyEmtInit,(PCVMM2USERMETHODS pThis, PUVM pUVM, PUVMCPU pUVCpu)); /** * EMT termination notification callback. * * This is intended for doing per-thread cleanups for EMTs (like COM). * * @param pThis Pointer to the callback method table. * @param pUVM The user mode VM handle. * @param pUVCpu The user mode virtual CPU handle. * * @remarks This is optional and shall be set to NULL if not wanted. */ DECLR3CALLBACKMEMBER(void, pfnNotifyEmtTerm,(PCVMM2USERMETHODS pThis, PUVM pUVM, PUVMCPU pUVCpu)); /** * PDM thread initialization notification callback. * * This is intended for doing per-thread initialization (like COM init). * * @param pThis Pointer to the callback method table. * @param pUVM The user mode VM handle. * * @remarks This is optional and shall be set to NULL if not wanted. */ DECLR3CALLBACKMEMBER(void, pfnNotifyPdmtInit,(PCVMM2USERMETHODS pThis, PUVM pUVM)); /** * EMT termination notification callback. * * This is intended for doing per-thread cleanups for EMTs (like COM). * * @param pThis Pointer to the callback method table. * @param pUVM The user mode VM handle. * * @remarks This is optional and shall be set to NULL if not wanted. */ DECLR3CALLBACKMEMBER(void, pfnNotifyPdmtTerm,(PCVMM2USERMETHODS pThis, PUVM pUVM)); /** Magic value (VMM2USERMETHODS_MAGIC) marking the end of the structure. */ uint32_t u32EndMagic; } VMM2USERMETHODS; /** Magic value of the VMM2USERMETHODS (Franz Kafka). */ #define VMM2USERMETHODS_MAGIC UINT32_C(0x18830703) /** The VMM2USERMETHODS structure version. */ #define VMM2USERMETHODS_VERSION UINT32_C(0x00020000) VMMDECL(RTRCPTR) VMMGetStackRC(PVMCPU pVCpu); VMMDECL(VMCPUID) VMMGetCpuId(PVM pVM); VMMDECL(PVMCPU) VMMGetCpu(PVM pVM); VMMDECL(PVMCPU) VMMGetCpu0(PVM pVM); VMMDECL(PVMCPU) VMMGetCpuById(PVM pVM, VMCPUID idCpu); VMMDECL(uint32_t) VMMGetSvnRev(void); VMMDECL(VMMSWITCHER) VMMGetSwitcher(PVM pVM); VMMDECL(void) VMMTrashVolatileXMMRegs(void); /** @def VMMIsHwVirtExtForced * Checks if forced to use the hardware assisted virtualization extensions. * * This is intended for making setup decisions where we can save resources when * using hardware assisted virtualization. * * @returns true / false. * @param pVM Pointer to the shared VM structure. */ #define VMMIsHwVirtExtForced(pVM) ((pVM)->fHwVirtExtForced) #ifdef IN_RING3 /** @defgroup grp_vmm_r3 The VMM Host Context Ring 3 API * @ingroup grp_vmm * @{ */ VMMR3_INT_DECL(int) VMMR3Init(PVM pVM); VMMR3_INT_DECL(int) VMMR3InitR0(PVM pVM); VMMR3_INT_DECL(int) VMMR3InitRC(PVM pVM); VMMR3_INT_DECL(int) VMMR3InitCompleted(PVM pVM, VMINITCOMPLETED enmWhat); VMMR3_INT_DECL(int) VMMR3Term(PVM pVM); VMMR3_INT_DECL(void) VMMR3Relocate(PVM pVM, RTGCINTPTR offDelta); VMMR3_INT_DECL(int) VMMR3UpdateLoggers(PVM pVM); VMMR3DECL(const char *) VMMR3GetRZAssertMsg1(PVM pVM); VMMR3DECL(const char *) VMMR3GetRZAssertMsg2(PVM pVM); VMMR3_INT_DECL(int) VMMR3GetImportRC(PVM pVM, const char *pszSymbol, PRTRCPTR pRCPtrValue); VMMR3_INT_DECL(int) VMMR3SelectSwitcher(PVM pVM, VMMSWITCHER enmSwitcher); VMMR3_INT_DECL(int) VMMR3DisableSwitcher(PVM pVM); VMMR3_INT_DECL(RTR0PTR) VMMR3GetHostToGuestSwitcher(PVM pVM, VMMSWITCHER enmSwitcher); VMMR3_INT_DECL(int) VMMR3RawRunGC(PVM pVM, PVMCPU pVCpu); VMMR3_INT_DECL(int) VMMR3HwAccRunGC(PVM pVM, PVMCPU pVCpu); VMMR3DECL(int) VMMR3CallRC(PVM pVM, RTRCPTR RCPtrEntry, unsigned cArgs, ...); VMMR3DECL(int) VMMR3CallRCV(PVM pVM, RTRCPTR RCPtrEntry, unsigned cArgs, va_list args); VMMR3DECL(int) VMMR3CallR0(PVM pVM, uint32_t uOperation, uint64_t u64Arg, PSUPVMMR0REQHDR pReqHdr); VMMR3DECL(int) VMMR3ResumeHyper(PVM pVM, PVMCPU pVCpu); VMMR3DECL(void) VMMR3FatalDump(PVM pVM, PVMCPU pVCpu, int rcErr); VMMR3_INT_DECL(void) VMMR3YieldSuspend(PVM pVM); VMMR3_INT_DECL(void) VMMR3YieldStop(PVM pVM); VMMR3_INT_DECL(void) VMMR3YieldResume(PVM pVM); VMMR3_INT_DECL(void) VMMR3SendSipi(PVM pVM, VMCPUID idCpu, uint32_t uVector); VMMR3_INT_DECL(void) VMMR3SendInitIpi(PVM pVM, VMCPUID idCpu); VMMR3DECL(int) VMMR3RegisterPatchMemory(PVM pVM, RTGCPTR pPatchMem, unsigned cbPatchMem); VMMR3DECL(int) VMMR3DeregisterPatchMemory(PVM pVM, RTGCPTR pPatchMem, unsigned cbPatchMem); VMMR3DECL(int) VMMR3AtomicExecuteHandler(PVM pVM, PFNATOMICHANDLER pfnHandler, void *pvUser); VMMR3DECL(int) VMMR3EmtRendezvous(PVM pVM, uint32_t fFlags, PFNVMMEMTRENDEZVOUS pfnRendezvous, void *pvUser); /** @defgroup grp_VMMR3EmtRendezvous_fFlags VMMR3EmtRendezvous flags * @{ */ /** Execution type mask. */ #define VMMEMTRENDEZVOUS_FLAGS_TYPE_MASK UINT32_C(0x00000007) /** Invalid execution type. */ #define VMMEMTRENDEZVOUS_FLAGS_TYPE_INVALID UINT32_C(0) /** Let the EMTs execute the callback one by one (in no particular order). */ #define VMMEMTRENDEZVOUS_FLAGS_TYPE_ONE_BY_ONE UINT32_C(1) /** Let all the EMTs execute the callback at the same time. */ #define VMMEMTRENDEZVOUS_FLAGS_TYPE_ALL_AT_ONCE UINT32_C(2) /** Only execute the callback on one EMT (no particular one). */ #define VMMEMTRENDEZVOUS_FLAGS_TYPE_ONCE UINT32_C(3) /** Let the EMTs execute the callback one by one in ascending order. */ #define VMMEMTRENDEZVOUS_FLAGS_TYPE_ASCENDING UINT32_C(4) /** Let the EMTs execute the callback one by one in descending order. */ #define VMMEMTRENDEZVOUS_FLAGS_TYPE_DESCENDING UINT32_C(5) /** Stop after the first error. * This is not valid for any execution type where more than one EMT is active * at a time. */ #define VMMEMTRENDEZVOUS_FLAGS_STOP_ON_ERROR UINT32_C(0x00000008) /** The valid flags. */ #define VMMEMTRENDEZVOUS_FLAGS_VALID_MASK UINT32_C(0x0000000f) /** @} */ VMMR3_INT_DECL(int) VMMR3EmtRendezvousFF(PVM pVM, PVMCPU pVCpu); VMMR3_INT_DECL(int) VMMR3ReadR0Stack(PVM pVM, VMCPUID idCpu, RTHCUINTPTR R0Addr, void *pvBuf, size_t cbRead); /** @} */ #endif /* IN_RING3 */ /** @defgroup grp_vmm_r0 The VMM Host Context Ring 0 API * @ingroup grp_vmm * @{ */ /** * The VMMR0Entry() codes. */ typedef enum VMMR0OPERATION { /** Run guest context. */ VMMR0_DO_RAW_RUN = SUP_VMMR0_DO_RAW_RUN, /** Run guest code using the available hardware acceleration technology. */ VMMR0_DO_HWACC_RUN = SUP_VMMR0_DO_HWACC_RUN, /** Official NOP that we use for profiling. */ VMMR0_DO_NOP = SUP_VMMR0_DO_NOP, /** Official slow iocl NOP that we use for profiling. */ VMMR0_DO_SLOW_NOP, /** Ask the GVMM to create a new VM. */ VMMR0_DO_GVMM_CREATE_VM, /** Ask the GVMM to destroy the VM. */ VMMR0_DO_GVMM_DESTROY_VM, /** Call GVMMR0SchedHalt(). */ VMMR0_DO_GVMM_SCHED_HALT, /** Call GVMMR0SchedWakeUp(). */ VMMR0_DO_GVMM_SCHED_WAKE_UP, /** Call GVMMR0SchedPoke(). */ VMMR0_DO_GVMM_SCHED_POKE, /** Call GVMMR0SchedWakeUpAndPokeCpus(). */ VMMR0_DO_GVMM_SCHED_WAKE_UP_AND_POKE_CPUS, /** Call GVMMR0SchedPoll(). */ VMMR0_DO_GVMM_SCHED_POLL, /** Call GVMMR0QueryStatistics(). */ VMMR0_DO_GVMM_QUERY_STATISTICS, /** Call GVMMR0ResetStatistics(). */ VMMR0_DO_GVMM_RESET_STATISTICS, /** Call GVMMR0RegisterVCpu(). */ VMMR0_DO_GVMM_REGISTER_VMCPU, /** Call VMMR0 Per VM Init. */ VMMR0_DO_VMMR0_INIT, /** Call VMMR0 Per VM Termination. */ VMMR0_DO_VMMR0_TERM, /** Setup the hardware accelerated raw-mode session. */ VMMR0_DO_HWACC_SETUP_VM, /** Attempt to enable or disable hardware accelerated raw-mode. */ VMMR0_DO_HWACC_ENABLE, /** Calls function in the hypervisor. * The caller must setup the hypervisor context so the call will be performed. * The difference between VMMR0_DO_RUN_GC and this one is the handling of * the return GC code. The return code will not be interpreted by this operation. */ VMMR0_DO_CALL_HYPERVISOR, /** Call PGMR0PhysAllocateHandyPages(). */ VMMR0_DO_PGM_ALLOCATE_HANDY_PAGES, /** Call PGMR0AllocateLargePage(). */ VMMR0_DO_PGM_ALLOCATE_LARGE_HANDY_PAGE, /** Call PGMR0PhysSetupIommu(). */ VMMR0_DO_PGM_PHYS_SETUP_IOMMU, /** Call GMMR0InitialReservation(). */ VMMR0_DO_GMM_INITIAL_RESERVATION, /** Call GMMR0UpdateReservation(). */ VMMR0_DO_GMM_UPDATE_RESERVATION, /** Call GMMR0AllocatePages(). */ VMMR0_DO_GMM_ALLOCATE_PAGES, /** Call GMMR0FreePages(). */ VMMR0_DO_GMM_FREE_PAGES, /** Call GMMR0FreeLargePage(). */ VMMR0_DO_GMM_FREE_LARGE_PAGE, /** Call GMMR0QueryHypervisorMemoryStatsReq(). */ VMMR0_DO_GMM_QUERY_HYPERVISOR_MEM_STATS, /** Call GMMR0QueryMemoryStatsReq(). */ VMMR0_DO_GMM_QUERY_MEM_STATS, /** Call GMMR0BalloonedPages(). */ VMMR0_DO_GMM_BALLOONED_PAGES, /** Call GMMR0MapUnmapChunk(). */ VMMR0_DO_GMM_MAP_UNMAP_CHUNK, /** Call GMMR0SeedChunk(). */ VMMR0_DO_GMM_SEED_CHUNK, /** Call GMMR0RegisterSharedModule. */ VMMR0_DO_GMM_REGISTER_SHARED_MODULE, /** Call GMMR0UnregisterSharedModule. */ VMMR0_DO_GMM_UNREGISTER_SHARED_MODULE, /** Call GMMR0ResetSharedModules. */ VMMR0_DO_GMM_RESET_SHARED_MODULES, /** Call GMMR0CheckSharedModules. */ VMMR0_DO_GMM_CHECK_SHARED_MODULES, /** Call GMMR0FindDuplicatePage. */ VMMR0_DO_GMM_FIND_DUPLICATE_PAGE, /** Set a GVMM or GMM configuration value. */ VMMR0_DO_GCFGM_SET_VALUE, /** Query a GVMM or GMM configuration value. */ VMMR0_DO_GCFGM_QUERY_VALUE, /** Call PDMR0DriverCallReqHandler. */ VMMR0_DO_PDM_DRIVER_CALL_REQ_HANDLER, /** Call PDMR0DeviceCallReqHandler. */ VMMR0_DO_PDM_DEVICE_CALL_REQ_HANDLER, /** The start of the R0 service operations. */ VMMR0_DO_SRV_START, /** Call IntNetR0Open(). */ VMMR0_DO_INTNET_OPEN, /** Call IntNetR0IfClose(). */ VMMR0_DO_INTNET_IF_CLOSE, /** Call IntNetR0IfGetBufferPtrs(). */ VMMR0_DO_INTNET_IF_GET_BUFFER_PTRS, /** Call IntNetR0IfSetPromiscuousMode(). */ VMMR0_DO_INTNET_IF_SET_PROMISCUOUS_MODE, /** Call IntNetR0IfSetMacAddress(). */ VMMR0_DO_INTNET_IF_SET_MAC_ADDRESS, /** Call IntNetR0IfSetActive(). */ VMMR0_DO_INTNET_IF_SET_ACTIVE, /** Call IntNetR0IfSend(). */ VMMR0_DO_INTNET_IF_SEND, /** Call IntNetR0IfWait(). */ VMMR0_DO_INTNET_IF_WAIT, /** Call IntNetR0IfAbortWait(). */ VMMR0_DO_INTNET_IF_ABORT_WAIT, /** Forward call to the PCI driver */ VMMR0_DO_PCIRAW_REQ, /** The end of the R0 service operations. */ VMMR0_DO_SRV_END, /** Official call we use for testing Ring-0 APIs. */ VMMR0_DO_TESTS, /** Test the 32->64 bits switcher. */ VMMR0_DO_TEST_SWITCHER3264, /** The usual 32-bit type blow up. */ VMMR0_DO_32BIT_HACK = 0x7fffffff } VMMR0OPERATION; /** * Request buffer for VMMR0_DO_GCFGM_SET_VALUE and VMMR0_DO_GCFGM_QUERY_VALUE. * @todo Move got GCFGM.h when it's implemented. */ typedef struct GCFGMVALUEREQ { /** The request header.*/ SUPVMMR0REQHDR Hdr; /** The support driver session handle. */ PSUPDRVSESSION pSession; /** The value. * This is input for the set request and output for the query. */ uint64_t u64Value; /** The variable name. * This is fixed sized just to make things simple for the mock-up. */ char szName[48]; } GCFGMVALUEREQ; /** Pointer to a VMMR0_DO_GCFGM_SET_VALUE and VMMR0_DO_GCFGM_QUERY_VALUE request buffer. * @todo Move got GCFGM.h when it's implemented. */ typedef GCFGMVALUEREQ *PGCFGMVALUEREQ; VMMR0DECL(int) VMMR0EntryInt(PVM pVM, VMMR0OPERATION enmOperation, void *pvArg); VMMR0DECL(void) VMMR0EntryFast(PVM pVM, VMCPUID idCpu, VMMR0OPERATION enmOperation); VMMR0DECL(int) VMMR0EntryEx(PVM pVM, VMCPUID idCpu, VMMR0OPERATION enmOperation, PSUPVMMR0REQHDR pReq, uint64_t u64Arg, PSUPDRVSESSION); VMMR0DECL(int) VMMR0TermVM(PVM pVM, PGVM pGVM); #ifdef LOG_ENABLED VMMR0DECL(void) VMMR0LogFlushDisable(PVMCPU pVCpu); VMMR0DECL(void) VMMR0LogFlushEnable(PVMCPU pVCpu); #else #define VMMR0LogFlushDisable(pVCpu) do { } while(0) #define VMMR0LogFlushEnable(pVCpu) do { } while(0) #endif /** @} */ #ifdef IN_RC /** @defgroup grp_vmm_rc The VMM Raw-Mode Context API * @ingroup grp_vmm * @{ */ VMMRCDECL(int) VMMGCEntry(PVM pVM, unsigned uOperation, unsigned uArg, ...); VMMRCDECL(void) VMMGCGuestToHost(PVM pVM, int rc); VMMRCDECL(void) VMMGCLogFlushIfFull(PVM pVM); /** @} */ #endif /* IN_RC */ #if defined(IN_RC) || defined(IN_RING0) /** @defgroup grp_vmm_rz The VMM Raw-Mode and Ring-0 Context API * @ingroup grp_vmm * @{ */ VMMRZDECL(int) VMMRZCallRing3(PVM pVM, PVMCPU pVCpu, VMMCALLRING3 enmOperation, uint64_t uArg); VMMRZDECL(int) VMMRZCallRing3NoCpu(PVM pVM, VMMCALLRING3 enmOperation, uint64_t uArg); VMMRZDECL(void) VMMRZCallRing3Disable(PVMCPU pVCpu); VMMRZDECL(void) VMMRZCallRing3Enable(PVMCPU pVCpu); VMMRZDECL(bool) VMMRZCallRing3IsEnabled(PVMCPU pVCpu); /** @} */ #endif /** @} */ RT_C_DECLS_END #endif