X-Git-Url: http://v3vee.org/palacios/gitweb/gitweb.cgi?a=blobdiff_plain;f=palacios%2Fsrc%2Fpalacios%2Fvmm_ctrl_regs.c;h=4cb40d8b848a48027fae87bdc164f4d097bb9779;hb=9b31f917eae9b397cb21ff78d81084301b289e43;hp=31d00001cb697ecc55eb73bb3316279845a77eef;hpb=a2b7cc4f2d739213d1edefb85ff941c41c86907b;p=palacios.git diff --git a/palacios/src/palacios/vmm_ctrl_regs.c b/palacios/src/palacios/vmm_ctrl_regs.c index 31d0000..4cb40d8 100644 --- a/palacios/src/palacios/vmm_ctrl_regs.c +++ b/palacios/src/palacios/vmm_ctrl_regs.c @@ -39,6 +39,23 @@ #endif +static int handle_lmsw(struct guest_info * info, struct x86_instr * dec_instr); +static int handle_clts(struct guest_info * info, struct x86_instr * dec_instr); + +static int handle_mov_to_cr0(struct guest_info * info, struct x86_instr * dec_instr); +static int handle_mov_to_cr0_32(struct guest_info * info, struct x86_instr * dec_instr); +static int handle_mov_to_cr0_32pae(struct guest_info * info, struct x86_instr * dec_instr); +static int handle_mov_to_cr0_64(struct guest_info * info, struct x86_instr * dec_instr); +static int handle_mov_to_cr0_64compat(struct guest_info * info, struct x86_instr * dec_instr); + +static int handle_mov_to_cr3_32(struct guest_info * info, struct x86_instr * dec_instr); +static int handle_mov_to_cr3_32pae(struct guest_info * info, struct x86_instr * dec_instr); +static int handle_mov_to_cr3_64(struct guest_info * info, struct x86_instr * dec_instr); +static int handle_mov_to_cr3_64compat(struct guest_info * info, struct x86_instr * dec_instr); + + + + // First Attempt = 494 lines // current = 106 lines int v3_handle_cr0_write(struct guest_info * info) { @@ -52,96 +69,31 @@ int v3_handle_cr0_write(struct guest_info * info) { ret = read_guest_va_memory(info, get_addr_linear(info, info->rip, &(info->segments.cs)), 15, instr); } - /* The IFetch will already have faulted in the necessary bytes for the full instruction - if (ret != 15) { - // I think we should inject a GPF into the guest - PrintError("Could not read instruction (ret=%d)\n", ret); - return -1; - } - */ + if (v3_decode(info, (addr_t)instr, &dec_instr) == -1) { PrintError("Could not decode instruction\n"); return -1; } - if (v3_opcode_cmp(V3_OPCODE_LMSW, (const uchar_t *)(dec_instr.opcode)) == 0) { - struct cr0_real *real_cr0 = (struct cr0_real*)&(info->ctrl_regs.cr0); - struct cr0_real *new_cr0 = (struct cr0_real *)(dec_instr.src_operand.operand); - uchar_t new_cr0_val; - - PrintDebug("LMSW\n"); - - new_cr0_val = (*(char*)(new_cr0)) & 0x0f; - - PrintDebug("OperandVal = %x\n", new_cr0_val); - - PrintDebug("Old CR0=%x\n", *real_cr0); - *(uchar_t*)real_cr0 &= 0xf0; - *(uchar_t*)real_cr0 |= new_cr0_val; - PrintDebug("New CR0=%x\n", *real_cr0); - - if (info->shdw_pg_mode == SHADOW_PAGING) { - struct cr0_real * shadow_cr0 = (struct cr0_real*)&(info->shdw_pg_state.guest_cr0); - - PrintDebug(" Old Shadow CR0=%x\n", *shadow_cr0); - *(uchar_t*)shadow_cr0 &= 0xf0; - *(uchar_t*)shadow_cr0 |= new_cr0_val; - PrintDebug("New Shadow CR0=%x\n", *shadow_cr0); + if (handle_lmsw(info, &dec_instr) == -1) { + return -1; } + } else if (v3_opcode_cmp(V3_OPCODE_MOV2CR, (const uchar_t *)(dec_instr.opcode)) == 0) { - PrintDebug("MOV2CR0\n"); - if (info->cpu_mode == LONG) { - // 64 bit registers - } else { - // 32 bit registers - struct cr0_32 *real_cr0 = (struct cr0_32*)&(info->ctrl_regs.cr0); - struct cr0_32 *new_cr0= (struct cr0_32 *)(dec_instr.src_operand.operand); - - PrintDebug("OperandVal = %x, length=%d\n", *new_cr0, dec_instr.src_operand.size); - - - PrintDebug("Old CR0=%x\n", *real_cr0); - *real_cr0 = *new_cr0; - - - if (info->shdw_pg_mode == SHADOW_PAGING) { - struct cr0_32 * shadow_cr0 = (struct cr0_32 *)&(info->shdw_pg_state.guest_cr0); - - PrintDebug("Old Shadow CR0=%x\n", *shadow_cr0); - - real_cr0->et = 1; - - *shadow_cr0 = *new_cr0; - shadow_cr0->et = 1; - - if (v3_get_mem_mode(info) == VIRTUAL_MEM) { - struct cr3_32 * shadow_cr3 = (struct cr3_32 *)&(info->shdw_pg_state.shadow_cr3); - - info->ctrl_regs.cr3 = *(addr_t*)shadow_cr3; - } else { - info->ctrl_regs.cr3 = *(addr_t*)&(info->direct_map_pt); - real_cr0->pg = 1; - } - - PrintDebug("New Shadow CR0=%x\n",*shadow_cr0); - } - PrintDebug("New CR0=%x\n", *real_cr0); + if (handle_mov_to_cr0(info, &dec_instr) == -1) { + return -1; } } else if (v3_opcode_cmp(V3_OPCODE_CLTS, (const uchar_t *)(dec_instr.opcode)) == 0) { - // CLTS - struct cr0_32 *real_cr0 = (struct cr0_32*)&(info->ctrl_regs.cr0); - - real_cr0->ts = 0; - if (info->shdw_pg_mode == SHADOW_PAGING) { - struct cr0_32 * shadow_cr0 = (struct cr0_32 *)&(info->shdw_pg_state.guest_cr0); - shadow_cr0->ts = 0; + if (handle_clts(info, &dec_instr) == -1) { + return -1; } + } else { PrintError("Unhandled opcode in handle_cr0_write\n"); return -1; @@ -153,6 +105,143 @@ int v3_handle_cr0_write(struct guest_info * info) { } + + +static int handle_mov_to_cr0(struct guest_info * info, struct x86_instr * dec_instr) { + PrintDebug("MOV2CR0\n"); + + switch (info->cpu_mode) { + case REAL: + case PROTECTED: + return handle_mov_to_cr0_32(info, dec_instr); + case PROTECTED_PAE: + return handle_mov_to_cr0_32pae(info, dec_instr); + case LONG: + return handle_mov_to_cr0_64(info, dec_instr); + case LONG_32_COMPAT: + return handle_mov_to_cr0_64compat(info, dec_instr); + default: + PrintError("Invalid CPU Operating Mode: %d\n", info->cpu_mode); + return -1; + + } +} + +static int handle_mov_to_cr0_32pae(struct guest_info * info, struct x86_instr * dec_instr) { + PrintError("32 bit PAE mov to CR0 not implemented\n"); + return -1; +} + +static int handle_mov_to_cr0_64(struct guest_info * info, struct x86_instr * dec_instr) { + PrintError("64 bit mov to CR0 not implemented\n"); + return -1; +} + +static int handle_mov_to_cr0_64compat(struct guest_info * info, struct x86_instr * dec_instr) { + PrintError("64 bit compatibility mode move to CR0 not implemented\n"); + return -1; +} + + +static int handle_mov_to_cr0_32(struct guest_info * info, struct x86_instr * dec_instr) { + // 32 bit registers + struct cr0_32 * shadow_cr0 = (struct cr0_32 *)&(info->ctrl_regs.cr0); + struct cr0_32 * new_cr0 = (struct cr0_32 *)(dec_instr->src_operand.operand); + struct cr0_32 * guest_cr0 = (struct cr0_32 *)&(info->shdw_pg_state.guest_cr0); + + PrintDebug("OperandVal = %x, length=%d\n", *(uint_t *)new_cr0, dec_instr->src_operand.size); + + PrintDebug("Old CR0=%x\n", *(uint_t *)shadow_cr0); + PrintDebug("Old Guest CR0=%x\n", *(uint_t *)guest_cr0); + + // Guest always sees the value they wrote + *guest_cr0 = *new_cr0; + + // This value must always be set to 1 + guest_cr0->et = 1; + + // Set the shadow register to catch non-virtualized flags + *shadow_cr0 = *guest_cr0; + + + if (v3_get_mem_mode(info) == VIRTUAL_MEM) { + /* struct cr3_32 * guest_cr3 = (struct cr3_32 *)&(info->shdw_pg_state.guest_cr3); + info->ctrl_regs.cr3 = *(addr_t*)guest_cr3; + */ + PrintDebug("Activating Shadow Page Tables\n"); + + if (v3_activate_shadow_pt(info) == -1) { + PrintError("Failed to activate shadow page tables\n"); + return -1; + } + } else { + + if (v3_activate_passthrough_pt(info) == -1) { + PrintError("Failed to activate passthrough page tables\n"); + return -1; + } + shadow_cr0->pg = 1; + } + + PrintDebug("New Guest CR0=%x\n",*(uint_t *)guest_cr0); + + PrintDebug("New CR0=%x\n", *(uint_t *)shadow_cr0); + + return 0; +} + + + + +static int handle_clts(struct guest_info * info, struct x86_instr * dec_instr) { + // CLTS + struct cr0_32 * real_cr0 = (struct cr0_32*)&(info->ctrl_regs.cr0); + + real_cr0->ts = 0; + + if (info->shdw_pg_mode == SHADOW_PAGING) { + struct cr0_32 * guest_cr0 = (struct cr0_32 *)&(info->shdw_pg_state.guest_cr0); + guest_cr0->ts = 0; + } + return 0; +} + + +static int handle_lmsw(struct guest_info * info, struct x86_instr * dec_instr) { + struct cr0_real * real_cr0 = (struct cr0_real*)&(info->ctrl_regs.cr0); + struct cr0_real * new_cr0 = (struct cr0_real *)(dec_instr->src_operand.operand); + uchar_t new_cr0_val; + + PrintDebug("LMSW\n"); + + new_cr0_val = (*(char*)(new_cr0)) & 0x0f; + + PrintDebug("OperandVal = %x\n", new_cr0_val); + + // We can just copy the new value through + // we don't need to virtualize the lower 4 bits + PrintDebug("Old CR0=%x\n", *(uint_t *)real_cr0); + *(uchar_t*)real_cr0 &= 0xf0; + *(uchar_t*)real_cr0 |= new_cr0_val; + PrintDebug("New CR0=%x\n", *(uint_t *)real_cr0); + + + // If Shadow paging is enabled we push the changes to the virtualized copy of cr0 + if (info->shdw_pg_mode == SHADOW_PAGING) { + struct cr0_real * guest_cr0 = (struct cr0_real*)&(info->shdw_pg_state.guest_cr0); + + PrintDebug("Old Guest CR0=%x\n", *(uint_t *)guest_cr0); + *(uchar_t*)guest_cr0 &= 0xf0; + *(uchar_t*)guest_cr0 |= new_cr0_val; + PrintDebug("New Guest CR0=%x\n", *(uint_t *)guest_cr0); + } + return 0; +} + + + + + // First attempt = 253 lines // current = 51 lines int v3_handle_cr0_read(struct guest_info * info) { @@ -166,13 +255,6 @@ int v3_handle_cr0_read(struct guest_info * info) { ret = read_guest_va_memory(info, get_addr_linear(info, info->rip, &(info->segments.cs)), 15, instr); } - /* The IFetch will already have faulted in the necessary bytes for the full instruction - if (ret != 15) { - // I think we should inject a GPF into the guest - PrintError("Could not read instruction (ret=%d)\n", ret); - return -1; - } - */ if (v3_decode(info, (addr_t)instr, &dec_instr) == -1) { PrintError("Could not decode instruction\n"); @@ -180,31 +262,31 @@ int v3_handle_cr0_read(struct guest_info * info) { } if (v3_opcode_cmp(V3_OPCODE_MOVCR2, (const uchar_t *)(dec_instr.opcode)) == 0) { - struct cr0_32 * virt_cr0 = (struct cr0_32 *)(dec_instr.dst_operand.operand); - struct cr0_32 * real_cr0 = (struct cr0_32 *)&(info->ctrl_regs.cr0); - + struct cr0_32 * dst_reg = (struct cr0_32 *)(dec_instr.dst_operand.operand); + struct cr0_32 * shadow_cr0 = (struct cr0_32 *)&(info->ctrl_regs.cr0); + PrintDebug("MOVCR2\n"); - PrintDebug("CR0 at 0x%x\n", real_cr0); if (info->shdw_pg_mode == SHADOW_PAGING) { - *virt_cr0 = *(struct cr0_32 *)&(info->shdw_pg_state.guest_cr0); + struct cr0_32 * guest_cr0 = (struct cr0_32 *)&(info->shdw_pg_state.guest_cr0); + *dst_reg = *guest_cr0; } else { - *virt_cr0 = *real_cr0; + *dst_reg = *shadow_cr0; } - - PrintDebug("real CR0: %x\n", *(uint_t*)real_cr0); - PrintDebug("returned CR0: %x\n", *(uint_t*)virt_cr0); + + PrintDebug("Shadow CR0: %x\n", *(uint_t*)shadow_cr0); + PrintDebug("returned CR0: %x\n", *(uint_t*)dst_reg); } else if (v3_opcode_cmp(V3_OPCODE_SMSW, (const uchar_t *)(dec_instr.opcode)) == 0) { - struct cr0_real *real_cr0= (struct cr0_real*)&(info->ctrl_regs.cr0); - struct cr0_real *virt_cr0 = (struct cr0_real *)(dec_instr.dst_operand.operand); - char cr0_val = *(char*)real_cr0 & 0x0f; + struct cr0_real * shadow_cr0 = (struct cr0_real *)&(info->ctrl_regs.cr0); + struct cr0_real * dst_reg = (struct cr0_real *)(dec_instr.dst_operand.operand); + char cr0_val = *(char*)shadow_cr0 & 0x0f; PrintDebug("SMSW\n"); - PrintDebug("CR0 at 0x%x\n", real_cr0); - - *(char *)virt_cr0 &= 0xf0; - *(char *)virt_cr0 |= cr0_val; + // The lower 4 bits of the guest/shadow CR0 are mapped through + // We can treat nested and shadow paging the same here + *(char *)dst_reg &= 0xf0; + *(char *)dst_reg |= cr0_val; } else { PrintError("Unhandled opcode in handle_cr0_read\n"); @@ -218,6 +300,7 @@ int v3_handle_cr0_read(struct guest_info * info) { + // First Attempt = 256 lines // current = 65 lines int v3_handle_cr3_write(struct guest_info * info) { @@ -231,67 +314,53 @@ int v3_handle_cr3_write(struct guest_info * info) { ret = read_guest_va_memory(info, get_addr_linear(info, info->rip, &(info->segments.cs)), 15, instr); } - /* The IFetch will already have faulted in the necessary bytes for the full instruction - if (ret != 15) { - // I think we should inject a GPF into the guest - PrintError("Could not read instruction (ret=%d)\n", ret); - return -1; - } - */ - if (v3_decode(info, (addr_t)instr, &dec_instr) == -1) { PrintError("Could not decode instruction\n"); return -1; } if (v3_opcode_cmp(V3_OPCODE_MOV2CR, (const uchar_t *)(dec_instr.opcode)) == 0) { - PrintDebug("MOV2CR3\n"); - PrintDebug("CR3 at 0x%x\n", &(info->ctrl_regs.cr3)); - - if (info->shdw_pg_mode == SHADOW_PAGING) { - struct cr3_32 * new_cr3 = (struct cr3_32 *)(dec_instr.src_operand.operand); - struct cr3_32 * guest_cr3 = (struct cr3_32 *)&(info->shdw_pg_state.guest_cr3); - struct cr3_32 * shadow_cr3 = (struct cr3_32 *)&(info->shdw_pg_state.shadow_cr3); - int cached = 0; - - - PrintDebug("Old Shadow CR3=%x; Old Guest CR3=%x\n", - *(uint_t*)shadow_cr3, *(uint_t*)guest_cr3); - - - cached = v3_cache_page_tables32(info, (addr_t)CR3_TO_PDE32((void *)*(addr_t *)new_cr3)); + if (info->mem_mode == PHYSICAL_MEM) { + // All we do is update the guest CR3 - if (cached == -1) { - PrintError("CR3 Cache failed\n"); - return -1; - } else if (cached == 0) { - addr_t shadow_pt; - - PrintDebug("New CR3 is different - flushing shadow page table\n"); - - delete_page_tables_pde32((pde32_t *)CR3_TO_PDE32(*(uint_t*)shadow_cr3)); - - shadow_pt = v3_create_new_shadow_pt32(); - - shadow_cr3->pdt_base_addr = (addr_t)V3_PAddr((void *)(addr_t)PD32_BASE_ADDR(shadow_pt)); + if (info->cpu_mode == LONG) { + struct cr3_64 * new_cr3 = (struct cr3_64 *)(dec_instr.src_operand.operand); + struct cr3_64 * guest_cr3 = (struct cr3_64 *)&(info->shdw_pg_state.guest_cr3); + *guest_cr3 = *new_cr3; } else { - PrintDebug("Reusing cached shadow Page table\n"); + struct cr3_32 * new_cr3 = (struct cr3_32 *)(dec_instr.src_operand.operand); + struct cr3_32 * guest_cr3 = (struct cr3_32 *)&(info->shdw_pg_state.guest_cr3); + *guest_cr3 = *new_cr3; } - - shadow_cr3->pwt = new_cr3->pwt; - shadow_cr3->pcd = new_cr3->pcd; - - // What the hell... - *guest_cr3 = *new_cr3; - - PrintDebug("New Shadow CR3=%x; New Guest CR3=%x\n", - *(uint_t*)shadow_cr3, *(uint_t*)guest_cr3); - if (info->mem_mode == VIRTUAL_MEM) { - // If we aren't in paged mode then we have to preserve the identity mapped CR3 - info->ctrl_regs.cr3 = *(addr_t*)shadow_cr3; + } else { + + switch (info->cpu_mode) { + case PROTECTED: + if (handle_mov_to_cr3_32(info, &dec_instr) == -1) { + return -1; + } + break; + case PROTECTED_PAE: + if (handle_mov_to_cr3_32pae(info, &dec_instr) == -1) { + return -1; + } + break; + case LONG: + if (handle_mov_to_cr3_64(info, &dec_instr) == -1) { + return -1; + } + break; + case LONG_32_COMPAT: + if (handle_mov_to_cr3_64compat(info, &dec_instr) == -1) { + return -1; + } + break; + default: + PrintError("Unhandled CPU mode: %d\n", info->cpu_mode); + return -1; } } } else { @@ -306,6 +375,56 @@ int v3_handle_cr3_write(struct guest_info * info) { + + + + +static int handle_mov_to_cr3_32(struct guest_info * info, struct x86_instr * dec_instr) { + PrintDebug("CR3 at 0x%p\n", &(info->ctrl_regs.cr3)); + + if (info->shdw_pg_mode == SHADOW_PAGING) { + struct cr3_32 * new_cr3 = (struct cr3_32 *)(dec_instr->src_operand.operand); + struct cr3_32 * guest_cr3 = (struct cr3_32 *)&(info->shdw_pg_state.guest_cr3); +#ifdef DEBUG_CTRL_REGS + struct cr3_32 * shadow_cr3 = (struct cr3_32 *)&(info->ctrl_regs.cr3); +#endif + + PrintDebug("Old Shadow CR3=%x; Old Guest CR3=%x\n", + *(uint_t*)shadow_cr3, *(uint_t*)guest_cr3); + + + // Store the write value to virtualize CR3 + *guest_cr3 = *new_cr3; + + if (v3_activate_shadow_pt(info) == -1) { + PrintError("Failed to activate 32 bit shadow page table\n"); + return -1; + } + + PrintDebug("New Shadow CR3=%x; New Guest CR3=%x\n", + *(uint_t*)shadow_cr3, *(uint_t*)guest_cr3); + } + return 0; +} + + +static int handle_mov_to_cr3_32pae(struct guest_info * info, struct x86_instr * dec_instr) { + PrintError("32 Bit PAE mode Mov to CR3 not implemented\n"); + return -1; +} + +static int handle_mov_to_cr3_64(struct guest_info * info, struct x86_instr * dec_instr) { + PrintError("Long mode Mov to CR3 not implemented\n"); + return -1; +} + +static int handle_mov_to_cr3_64compat(struct guest_info * info, struct x86_instr * dec_instr) { + PrintError("Long compatiblity mode move to CR3 not implemented\n"); + return -1; +} + + + // first attempt = 156 lines // current = 36 lines int v3_handle_cr3_read(struct guest_info * info) { @@ -319,13 +438,7 @@ int v3_handle_cr3_read(struct guest_info * info) { ret = read_guest_va_memory(info, get_addr_linear(info, info->rip, &(info->segments.cs)), 15, instr); } - /* The IFetch will already have faulted in the necessary bytes for the full instruction - if (ret != 15) { - // I think we should inject a GPF into the guest - PrintError("Could not read instruction (ret=%d)\n", ret); - return -1; - } - */ + if (v3_decode(info, (addr_t)instr, &dec_instr) == -1) { PrintError("Could not decode instruction\n"); @@ -334,14 +447,14 @@ int v3_handle_cr3_read(struct guest_info * info) { if (v3_opcode_cmp(V3_OPCODE_MOVCR2, (const uchar_t *)(dec_instr.opcode)) == 0) { PrintDebug("MOVCR32\n"); - struct cr3_32 * virt_cr3 = (struct cr3_32 *)(dec_instr.dst_operand.operand); + struct cr3_32 * dst_reg = (struct cr3_32 *)(dec_instr.dst_operand.operand); - PrintDebug("CR3 at 0x%x\n", &(info->ctrl_regs.cr3)); + PrintDebug("CR3 at 0x%p\n", &(info->ctrl_regs.cr3)); if (info->shdw_pg_mode == SHADOW_PAGING) { - *virt_cr3 = *(struct cr3_32 *)&(info->shdw_pg_state.guest_cr3); + *dst_reg = *(struct cr3_32 *)&(info->shdw_pg_state.guest_cr3); } else { - *virt_cr3 = *(struct cr3_32 *)&(info->ctrl_regs.cr3); + *dst_reg = *(struct cr3_32 *)&(info->ctrl_regs.cr3); } } else { PrintError("Unhandled opcode in handle_cr3_read\n"); @@ -352,3 +465,149 @@ int v3_handle_cr3_read(struct guest_info * info) { return 0; } + + +// We don't need to virtualize CR4, all we need is to detect the activation of PAE +int v3_handle_cr4_read(struct guest_info * info) { + // PrintError("CR4 Read not handled\n"); + // Do nothing... + return 0; +} + +int v3_handle_cr4_write(struct guest_info * info) { + uchar_t instr[15]; + int ret; + struct x86_instr dec_instr; + + if (info->mem_mode == PHYSICAL_MEM) { + ret = read_guest_pa_memory(info, get_addr_linear(info, info->rip, &(info->segments.cs)), 15, instr); + } else { + ret = read_guest_va_memory(info, get_addr_linear(info, info->rip, &(info->segments.cs)), 15, instr); + } + + if (v3_decode(info, (addr_t)instr, &dec_instr) == -1) { + PrintError("Could not decode instruction\n"); + return -1; + } + + if (v3_opcode_cmp(V3_OPCODE_MOV2CR, (const uchar_t *)(dec_instr.opcode)) != 0) { + PrintError("Invalid opcode in write to CR4\n"); + return -1; + } + + if ((info->cpu_mode == PROTECTED) || (info->cpu_mode == PROTECTED_PAE)) { + struct cr4_32 * new_cr4 = (struct cr4_32 *)(dec_instr.src_operand.operand); + struct cr4_32 * cr4 = (struct cr4_32 *)&(info->ctrl_regs.cr4); + + PrintDebug("OperandVal = %x, length = %d\n", *(uint_t *)new_cr4, dec_instr.src_operand.size); + PrintDebug("Old CR4=%x\n", *(uint_t *)cr4); + + if ((info->shdw_pg_mode == SHADOW_PAGING) && + (v3_get_mem_mode(info) == PHYSICAL_MEM)) { + + if ((cr4->pae == 0) && (new_cr4->pae == 1)) { + PrintDebug("Creating PAE passthrough tables\n"); + + // Delete the old 32 bit direct map page tables + delete_page_tables_32((pde32_t *)V3_VAddr((void *)(info->direct_map_pt))); + + // create 32 bit PAE direct map page table + info->direct_map_pt = (addr_t)V3_PAddr(create_passthrough_pts_32PAE(info)); + + // reset cr3 to new page tables + info->ctrl_regs.cr3 = *(addr_t*)&(info->direct_map_pt); + + } else if ((cr4->pae == 1) && (new_cr4->pae == 0)) { + // Create passthrough standard 32bit pagetables + return -1; + } + } + + *cr4 = *new_cr4; + PrintDebug("New CR4=%x\n", *(uint_t *)cr4); + + } else { + PrintError("CR4 write not supported in CPU_MODE: %d\n", info->cpu_mode); + return -1; + } + + info->rip += dec_instr.instr_length; + return 0; +} + + +int v3_handle_efer_read(uint_t msr, struct v3_msr * dst, void * priv_data) { + struct guest_info * info = (struct guest_info *)(priv_data); + PrintDebug("EFER Read\n"); + + dst->value = info->guest_efer.value; + + info->rip += 2; // WRMSR/RDMSR are two byte operands + return 0; +} + + +int v3_handle_efer_write(uint_t msr, struct v3_msr src, void * priv_data) { + struct guest_info * info = (struct guest_info *)(priv_data); + struct efer_64 * new_efer = (struct efer_64 *)&(src.value); + struct efer_64 * shadow_efer = (struct efer_64 *)&(info->ctrl_regs.efer); + struct v3_msr * guest_efer = &(info->guest_efer); + + PrintDebug("EFER Write\n"); + PrintDebug("Old EFER=%p\n", (void *)*(addr_t*)(shadow_efer)); + + // We virtualize the guests efer to hide the SVME and LMA bits + guest_efer->value = src.value; + + + if ((info->shdw_pg_mode == SHADOW_PAGING) && + (v3_get_mem_mode(info) == PHYSICAL_MEM)) { + + if ((shadow_efer->lme == 0) && (new_efer->lme == 1)) { + PrintDebug("Transition to longmode\n"); + PrintDebug("Creating Passthrough 64 bit page tables\n"); + + // Delete the old 32 bit direct map page tables + /* + * JRL BUG? + * Will these page tables always be in PAE format?? + */ + PrintDebug("Deleting old PAE Page tables\n"); + PrintError("JRL BUG?: Will the old page tables always be in PAE format??\n"); + delete_page_tables_32PAE((pdpe32pae_t *)V3_VAddr((void *)(info->direct_map_pt))); + + // create 64 bit direct map page table + info->direct_map_pt = (addr_t)V3_PAddr(create_passthrough_pts_64(info)); + + // reset cr3 to new page tables + info->ctrl_regs.cr3 = *(addr_t*)&(info->direct_map_pt); + + // We mark the Long Mode active because we have paging enabled + // We do this in new_efer because we copy the msr in full below + new_efer->lma = 1; + + } else if ((shadow_efer->lme == 1) && (new_efer->lme == 0)) { + // transition out of long mode + //((struct efer_64 *)&(info->guest_efer.value))->lme = 0; + //((struct efer_64 *)&(info->guest_efer.value))->lma = 0; + + return -1; + } + + // accept all changes to the efer, but make sure that the SVME bit is set... (SVM specific) + *shadow_efer = *new_efer; + shadow_efer->svme = 1; + + + + PrintDebug("New EFER=%p\n", (void *)*(addr_t *)(shadow_efer)); + } else { + PrintError("Write to EFER in NESTED_PAGING or VIRTUAL_MEM mode not supported\n"); + // Should probably just check for a long mode transition, and bomb out if it is + return -1; + } + + info->rip += 2; // WRMSR/RDMSR are two byte operands + + return 0; +}