X-Git-Url: http://v3vee.org/palacios/gitweb/gitweb.cgi?a=blobdiff_plain;f=palacios%2Fsrc%2Fpalacios%2Fvmm_direct_paging.c;h=5f5be4c559f4e31e0e5060c86e49085417c45dfb;hb=ed72a7fb6a7ccbf370dbfa7900d2548ab34f3d63;hp=e50a458f0d048b873e664ff2cf70709e5fabd87b;hpb=de3fce7c94e36342005ae5f79fcf22937acde586;p=palacios.git diff --git a/palacios/src/palacios/vmm_direct_paging.c b/palacios/src/palacios/vmm_direct_paging.c index e50a458..5f5be4c 100644 --- a/palacios/src/palacios/vmm_direct_paging.c +++ b/palacios/src/palacios/vmm_direct_paging.c @@ -25,15 +25,24 @@ #include -#ifndef DEBUG_NESTED_PAGING +#ifndef V3_CONFIG_DEBUG_NESTED_PAGING #undef PrintDebug #define PrintDebug(fmt, args...) #endif -static addr_t create_generic_pt_page() { +static addr_t create_generic_pt_page(struct guest_info *core) { void * page = 0; - page = V3_VAddr(V3_AllocPages(1)); + void *temp; + + temp = V3_AllocPagesExtended(1, PAGE_SIZE_4KB, -1, + core->shdw_pg_mode==SHADOW_PAGING ? V3_ALLOC_PAGES_CONSTRAINT_4GB : 0); + if (!temp) { + PrintError(VM_NONE, VCORE_NONE,"Cannot allocate page\n"); + return 0; + } + + page = V3_VAddr(temp); memset(page, 0, PAGE_SIZE); return (addr_t)page; @@ -45,44 +54,55 @@ static addr_t create_generic_pt_page() { #include "vmm_direct_paging_64.h" int v3_init_passthrough_pts(struct guest_info * info) { - info->direct_map_pt = (addr_t)V3_PAddr((void *)create_generic_pt_page()); + info->direct_map_pt = (addr_t)V3_PAddr((void *)create_generic_pt_page(info)); return 0; } -int v3_reset_passthrough_pts(struct guest_info * info) { - v3_cpu_mode_t mode = v3_get_vm_cpu_mode(info); + +int v3_free_passthrough_pts(struct guest_info * core) { + v3_cpu_mode_t mode = v3_get_vm_cpu_mode(core); // Delete the old direct map page tables switch(mode) { case REAL: case PROTECTED: - delete_page_tables_32((pde32_t *)V3_VAddr((void *)(info->direct_map_pt))); + delete_page_tables_32((pde32_t *)V3_VAddr((void *)(core->direct_map_pt))); break; case PROTECTED_PAE: case LONG: case LONG_32_COMPAT: // Long mode will only use 32PAE page tables... - delete_page_tables_32pae((pdpe32pae_t *)V3_VAddr((void *)(info->direct_map_pt))); + delete_page_tables_32pae((pdpe32pae_t *)V3_VAddr((void *)(core->direct_map_pt))); break; default: - PrintError("Unknown CPU Mode\n"); + PrintError(core->vm_info, core, "Unknown CPU Mode\n"); + return -1; break; } - + + return 0; +} + + +int v3_reset_passthrough_pts(struct guest_info * core) { + + v3_free_passthrough_pts(core); + // create new direct map page table - v3_init_passthrough_pts(info); + v3_init_passthrough_pts(core); return 0; } + int v3_activate_passthrough_pt(struct guest_info * info) { // For now... But we need to change this.... // As soon as shadow paging becomes active the passthrough tables are hosed // So this will cause chaos if it is called at that time info->ctrl_regs.cr3 = *(addr_t*)&(info->direct_map_pt); - //PrintError("Activate Passthrough Page tables not implemented\n"); + //PrintError(info->vm_info, info, "Activate Passthrough Page tables not implemented\n"); return 0; } @@ -102,7 +122,7 @@ int v3_handle_passthrough_pagefault(struct guest_info * info, addr_t fault_addr, return handle_passthrough_pagefault_32pae(info, fault_addr, error_code); default: - PrintError("Unknown CPU Mode\n"); + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); break; } return -1; @@ -111,10 +131,10 @@ int v3_handle_passthrough_pagefault(struct guest_info * info, addr_t fault_addr, int v3_handle_nested_pagefault(struct guest_info * info, addr_t fault_addr, pf_error_t error_code) { - v3_cpu_mode_t mode = v3_get_host_cpu_mode() + v3_cpu_mode_t mode = v3_get_host_cpu_mode(); - PrintDebug("Nested PageFault: fault_addr=%p, error_code=%u\n",(void*)fault_addr, *(uint_t *)&error_code); + PrintDebug(info->vm_info, info, "Nested PageFault: fault_addr=%p, error_code=%u\n", (void *)fault_addr, *(uint_t *)&error_code); switch(mode) { case REAL: @@ -129,7 +149,7 @@ int v3_handle_nested_pagefault(struct guest_info * info, addr_t fault_addr, pf_e return handle_passthrough_pagefault_64(info, fault_addr, error_code); default: - PrintError("Unknown CPU Mode\n"); + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); break; } return -1; @@ -150,15 +170,42 @@ int v3_invalidate_passthrough_addr(struct guest_info * info, addr_t inv_addr) { return invalidate_addr_32pae(info, inv_addr); default: - PrintError("Unknown CPU Mode\n"); + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); break; } return -1; } +int v3_invalidate_passthrough_addr_range(struct guest_info * info, + addr_t inv_addr_start, addr_t inv_addr_end) { + v3_cpu_mode_t mode = v3_get_vm_cpu_mode(info); + + switch(mode) { + case REAL: + case PROTECTED: + return invalidate_addr_32_range(info, inv_addr_start, inv_addr_end); + + case PROTECTED_PAE: + case LONG: + case LONG_32_COMPAT: + // Long mode will only use 32PAE page tables... + return invalidate_addr_32pae_range(info, inv_addr_start, inv_addr_end); + + default: + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); + break; + } + return -1; +} + int v3_invalidate_nested_addr(struct guest_info * info, addr_t inv_addr) { + +#ifdef __V3_64BIT__ v3_cpu_mode_t mode = LONG; +#else + v3_cpu_mode_t mode = PROTECTED; +#endif switch(mode) { case REAL: @@ -173,7 +220,36 @@ int v3_invalidate_nested_addr(struct guest_info * info, addr_t inv_addr) { return invalidate_addr_64(info, inv_addr); default: - PrintError("Unknown CPU Mode\n"); + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); + break; + } + + return -1; +} + +int v3_invalidate_nested_addr_range(struct guest_info * info, + addr_t inv_addr_start, addr_t inv_addr_end) { + +#ifdef __V3_64BIT__ + v3_cpu_mode_t mode = LONG; +#else + v3_cpu_mode_t mode = PROTECTED; +#endif + + switch(mode) { + case REAL: + case PROTECTED: + return invalidate_addr_32_range(info, inv_addr_start, inv_addr_end); + + case PROTECTED_PAE: + return invalidate_addr_32pae_range(info, inv_addr_start, inv_addr_end); + + case LONG: + case LONG_32_COMPAT: + return invalidate_addr_64_range(info, inv_addr_start, inv_addr_end); + + default: + PrintError(info->vm_info, info, "Unknown CPU Mode\n"); break; }