X-Git-Url: http://v3vee.org/palacios/gitweb/gitweb.cgi?a=blobdiff_plain;f=palacios%2Fsrc%2Fpalacios%2Fvmm_direct_paging.c;h=6bf987ecb2372ccfbaeafd87bc10dc4b463eb7b3;hb=774bac9fbb03ef8bf7c2ca2c79a8b87c9bc4c526;hp=55bf30d94c0b50d82420d3a4d6a288a9946fee91;hpb=496b928649b2a63c3e21e099ce914ec00b3d7227;p=palacios.git diff --git a/palacios/src/palacios/vmm_direct_paging.c b/palacios/src/palacios/vmm_direct_paging.c index 55bf30d..6bf987e 100644 --- a/palacios/src/palacios/vmm_direct_paging.c +++ b/palacios/src/palacios/vmm_direct_paging.c @@ -1,50 +1,262 @@ -#include - -// Inline handler functions for each cpu mode -#include "vmm_direct_paging_32.h" +/* + * This file is part of the Palacios Virtual Machine Monitor developed + * by the V3VEE Project with funding from the United States National + * Science Foundation and the Department of Energy. + * + * The V3VEE Project is a joint project between Northwestern University + * and the University of New Mexico. You can find out more at + * http://www.v3vee.org + * + * Copyright (c) 2008, Steven Jaconette + * Copyright (c) 2008, Jack Lange + * Copyright (c) 2008, The V3VEE Project + * All rights reserved. + * + * Author: Steven Jaconette + * + * This is free software. You are permitted to use, + * redistribute, and modify it as specified in the file "V3VEE_LICENSE". + */ +#include #include #include #include #include +#include + + +#ifndef V3_CONFIG_DEBUG_NESTED_PAGING +#undef PrintDebug +#define PrintDebug(fmt, args...) +#endif + + +static addr_t create_generic_pt_page(struct guest_info *core) { + void * page = 0; + void *temp; + + temp = V3_AllocPagesExtended(1, PAGE_SIZE_4KB, -1, 0); // no constraints + + if (!temp) { + PrintError(VM_NONE, VCORE_NONE,"Cannot allocate page\n"); + return 0; + } + + page = V3_VAddr(temp); + memset(page, 0, PAGE_SIZE); + + return (addr_t)page; +} + +// Inline handler functions for each cpu mode +#include "vmm_direct_paging_32.h" +#include "vmm_direct_paging_32pae.h" +#include "vmm_direct_paging_64.h" + +int v3_init_passthrough_pts(struct guest_info * info) { + info->direct_map_pt = (addr_t)V3_PAddr((void *)create_generic_pt_page(info)); + return 0; +} + + +int v3_free_passthrough_pts(struct guest_info * core) { + v3_cpu_mode_t mode = v3_get_vm_cpu_mode(core); + + // Delete the old direct map page tables + switch(mode) { + case REAL: + case PROTECTED: + // Intentional fallthrough here + // There are *only* PAE tables + case PROTECTED_PAE: + case LONG: + case LONG_32_COMPAT: + // Long mode will only use 32PAE page tables... + delete_page_tables_32pae((pdpe32pae_t *)V3_VAddr((void *)(core->direct_map_pt))); + break; + default: + PrintError(core->vm_info, core, "Unknown CPU Mode\n"); + return -1; + break; + } + + return 0; +} + + +int v3_reset_passthrough_pts(struct guest_info * core) { + + v3_free_passthrough_pts(core); + + // create new direct map page table + v3_init_passthrough_pts(core); + + return 0; +} + + + +int v3_activate_passthrough_pt(struct guest_info * info) { + // For now... But we need to change this.... + // As soon as shadow paging becomes active the passthrough tables are hosed + // So this will cause chaos if it is called at that time + struct cr3_32_PAE * shadow_cr3 = (struct cr3_32_PAE *) &(info->ctrl_regs.cr3); + struct cr4_32 * shadow_cr4 = (struct cr4_32 *) &(info->ctrl_regs.cr4); + addr_t shadow_pt_addr = *(addr_t*)&(info->direct_map_pt); + // Passthrough PTs will only be PAE page tables. + shadow_cr3->pdpt_base_addr = shadow_pt_addr >> 5; + shadow_cr4->pae = 1; + PrintDebug(info->vm_info, info, "Activated Passthrough Page tables\n"); + return 0; +} + + +int v3_handle_passthrough_pagefault(struct guest_info * info, addr_t fault_addr, pf_error_t error_code) { + v3_cpu_mode_t mode = v3_get_vm_cpu_mode(info); + + switch(mode) { + case REAL: + case PROTECTED: + // Note intentional fallthrough here + // There are only PAE page tables now + case PROTECTED_PAE: + case LONG: + case LONG_32_COMPAT: + // Long mode will only use 32PAE page tables... + return handle_passthrough_pagefault_32pae(info, fault_addr, error_code); + + default: + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); + break; + } + return -1; +} + + + +int v3_handle_nested_pagefault(struct guest_info * info, addr_t fault_addr, pf_error_t error_code) { + v3_cpu_mode_t mode = v3_get_host_cpu_mode(); + + + PrintDebug(info->vm_info, info, "Nested PageFault: fault_addr=%p, error_code=%u\n", (void *)fault_addr, *(uint_t *)&error_code); + + switch(mode) { + case REAL: + case PROTECTED: + return handle_passthrough_pagefault_32(info, fault_addr, error_code); + + case PROTECTED_PAE: + return handle_passthrough_pagefault_32pae(info, fault_addr, error_code); + + case LONG: + case LONG_32_COMPAT: + return handle_passthrough_pagefault_64(info, fault_addr, error_code); + + default: + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); + break; + } + return -1; +} + +int v3_invalidate_passthrough_addr(struct guest_info * info, addr_t inv_addr) { + v3_cpu_mode_t mode = v3_get_vm_cpu_mode(info); + + switch(mode) { + case REAL: + case PROTECTED: + // Intentional fallthrough - there + // are only PAE page tables now + case PROTECTED_PAE: + case LONG: + case LONG_32_COMPAT: + // Long mode will only use 32PAE page tables... + return invalidate_addr_32pae(info, inv_addr); + + default: + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); + break; + } + return -1; +} + + +int v3_invalidate_passthrough_addr_range(struct guest_info * info, + addr_t inv_addr_start, addr_t inv_addr_end) { + v3_cpu_mode_t mode = v3_get_vm_cpu_mode(info); + + switch(mode) { + case REAL: + case PROTECTED: + // Intentional fallthrough + // There are only PAE PTs now + case PROTECTED_PAE: + case LONG: + case LONG_32_COMPAT: + // Long mode will only use 32PAE page tables... + return invalidate_addr_32pae_range(info, inv_addr_start, inv_addr_end); + + default: + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); + break; + } + return -1; +} + +int v3_invalidate_nested_addr(struct guest_info * info, addr_t inv_addr) { + +#ifdef __V3_64BIT__ + v3_cpu_mode_t mode = LONG; +#else + v3_cpu_mode_t mode = PROTECTED; +#endif + + switch(mode) { + case REAL: + case PROTECTED: + return invalidate_addr_32(info, inv_addr); + + case PROTECTED_PAE: + return invalidate_addr_32pae(info, inv_addr); + + case LONG: + case LONG_32_COMPAT: + return invalidate_addr_64(info, inv_addr); + + default: + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); + break; + } + + return -1; +} + +int v3_invalidate_nested_addr_range(struct guest_info * info, + addr_t inv_addr_start, addr_t inv_addr_end) { + +#ifdef __V3_64BIT__ + v3_cpu_mode_t mode = LONG; +#else + v3_cpu_mode_t mode = PROTECTED; +#endif + + switch(mode) { + case REAL: + case PROTECTED: + return invalidate_addr_32_range(info, inv_addr_start, inv_addr_end); + + case PROTECTED_PAE: + return invalidate_addr_32pae_range(info, inv_addr_start, inv_addr_end); + + case LONG: + case LONG_32_COMPAT: + return invalidate_addr_64_range(info, inv_addr_start, inv_addr_end); + + default: + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); + break; + } -pde32_t * v3_create_direct_passthrough_pts(struct guest_info * info) { - v3_vm_cpu_mode_t mode = v3_get_cpu_mode(info); - switch(mode) { - case REAL: - //break; - case PROTECTED: - PrintError("ABC\n"); - return v3_create_direct_passthrough_pts_32(info); - case PROTECTED_PAE: - break; - case LONG: - break; - case LONG_32_COMPAT: - break; - default: - PrintError("Unknown CPU Mode\n"); - break; - } - return NULL; -} - -int v3_handle_shadow_pagefault_physical_mode(struct guest_info * info, addr_t fault_addr, pf_error_t error_code) { - v3_vm_cpu_mode_t mode = v3_get_cpu_mode(info); - switch(mode) { - case REAL: - // break; - case PROTECTED: - return v3_handle_shadow_pagefault_physical_mode_32(info, fault_addr, error_code); - case PROTECTED_PAE: - break; - case LONG: - break; - case LONG_32_COMPAT: - break; - default: - PrintError("Unknown CPU Mode\n"); - break; - } - return -1; + return -1; }