#include <palacios/vmm_lowlevel.h>
#include <palacios/svm_msr.h>
+#include <palacios/vmm_rbtree.h>
+#include <palacios/vmm_profiler.h>
+
+#include <palacios/vmm_direct_paging.h>
extern void v3_stgi();
extern void v3_clgi();
+//extern int v3_svm_launch(vmcb_t * vmcb, struct v3_gprs * vm_regs, uint64_t * fs, uint64_t * gs);
extern int v3_svm_launch(vmcb_t * vmcb, struct v3_gprs * vm_regs);
//ctrl_area->instrs.instrs.CR0 = 1;
ctrl_area->cr_reads.cr0 = 1;
ctrl_area->cr_writes.cr0 = 1;
- ctrl_area->cr_reads.cr4 = 1;
+ //ctrl_area->cr_reads.cr4 = 1;
ctrl_area->cr_writes.cr4 = 1;
- if (vm_info->io_map.num_ports > 0) {
- struct vmm_io_hook * iter;
+ if ( !RB_EMPTY_ROOT(&(vm_info->io_map)) ) {
+ struct v3_io_hook * iter;
+ struct rb_node * io_node = v3_rb_first(&(vm_info->io_map));
addr_t io_port_bitmap;
-
+ int i = 0;
+
io_port_bitmap = (addr_t)V3_VAddr(V3_AllocPages(3));
memset((uchar_t*)io_port_bitmap, 0, PAGE_SIZE * 3);
//PrintDebug("Setting up IO Map at 0x%x\n", io_port_bitmap);
- FOREACH_IO_HOOK(vm_info->io_map, iter) {
+ do {
+ iter = rb_entry(io_node, struct v3_io_hook, tree_node);
+
ushort_t port = iter->port;
uchar_t * bitmap = (uchar_t *)io_port_bitmap;
+ //PrintDebug("%d: Hooking Port %d\n", i, port);
bitmap += (port / 8);
// PrintDebug("Setting Bit for port 0x%x\n", port);
*bitmap |= 1 << (port % 8);
- }
+
+ i++;
+ } while ((io_node = v3_rb_next(io_node)));
//PrintDebugMemDump((uchar_t*)io_port_bitmap, PAGE_SIZE *2);
/* Testing 64 bit page tables for long paged real mode guests */
// vm_info->direct_map_pt = (addr_t)V3_PAddr(create_passthrough_pts_64(vm_info));
- vm_info->direct_map_pt = (addr_t)V3_PAddr(create_passthrough_pts_32(vm_info));
+ vm_info->direct_map_pt = (addr_t)V3_PAddr(v3_create_direct_passthrough_pts(vm_info));
/* End Test */
vm_info->shdw_pg_state.guest_cr0 = 0x0000000000000010LL;
}
+ /* Safety locations for fs/gs */
+ vm_info->fs = 0;
+ vm_info->gs = 0;
+
+
}
while (1) {
ullong_t tmp_tsc;
- uint_t vm_cr_low = 0, vm_cr_high = 0;
+
+#define MSR_STAR 0xc0000081
+#define MSR_LSTAR 0xc0000082
+#define MSR_CSTAR 0xc0000083
+#define MSR_SF_MASK 0xc0000084
+#define MSR_GS_BASE 0xc0000101
+#define MSR_KERNGS_BASE 0xc0000102
+
+
+ struct v3_msr host_cstar;
+ struct v3_msr host_star;
+ struct v3_msr host_lstar;
+ struct v3_msr host_syscall_mask;
+ struct v3_msr host_gs_base;
+ struct v3_msr host_kerngs_base;
+
v3_enable_ints();
v3_clgi();
- //PrintDebug("SVM Entry to rip=%p...\n", (void *)info->rip);
+ /*
+ PrintDebug("SVM Entry to CS=%p rip=%p...\n",
+ (void *)(addr_t)info->segments.cs.base,
+ (void *)(addr_t)info->rip);
+ */
- v3_get_msr(0xc0000101, &vm_cr_high, &vm_cr_low);
- rdtscll(info->time_state.cached_host_tsc);
+ v3_get_msr(MSR_STAR, &(host_star.hi), &(host_star.lo));
+ v3_get_msr(MSR_LSTAR, &(host_lstar.hi), &(host_lstar.lo));
+ v3_get_msr(MSR_CSTAR, &(host_cstar.hi), &(host_cstar.lo));
+ v3_get_msr(MSR_SF_MASK, &(host_syscall_mask.hi), &(host_syscall_mask.lo));
+ v3_get_msr(MSR_GS_BASE, &(host_gs_base.hi), &(host_gs_base.lo));
+ v3_get_msr(MSR_KERNGS_BASE, &(host_kerngs_base.hi), &(host_kerngs_base.lo));
+
- guest_ctrl->TSC_OFFSET = info->time_state.guest_tsc - info->time_state.cached_host_tsc;
+ rdtscll(info->time_state.cached_host_tsc);
+ // guest_ctrl->TSC_OFFSET = info->time_state.guest_tsc - info->time_state.cached_host_tsc;
+ //v3_svm_launch((vmcb_t*)V3_PAddr(info->vmm_data), &(info->vm_regs), &(info->fs), &(info->gs));
v3_svm_launch((vmcb_t*)V3_PAddr(info->vmm_data), &(info->vm_regs));
+
rdtscll(tmp_tsc);
- v3_set_msr(0xc0000101, vm_cr_high, vm_cr_low);
+ v3_set_msr(MSR_STAR, host_star.hi, host_star.lo);
+ v3_set_msr(MSR_LSTAR, host_lstar.hi, host_lstar.lo);
+ v3_set_msr(MSR_CSTAR, host_cstar.hi, host_cstar.lo);
+ v3_set_msr(MSR_SF_MASK, host_syscall_mask.hi, host_syscall_mask.lo);
+ v3_set_msr(MSR_GS_BASE, host_gs_base.hi, host_gs_base.lo);
+ v3_set_msr(MSR_KERNGS_BASE, host_kerngs_base.hi, host_kerngs_base.lo);
+
//PrintDebug("SVM Returned\n");
v3_stgi();
- if (num_exits % 25 == 0) {
+ if ((num_exits % 5000) == 0) {
PrintDebug("SVM Exit number %d\n", num_exits);
+ if (info->enable_profiler)
+ v3_print_profile(info);
}
+
if (v3_handle_svm_exit(info) != 0) {
vmcb_ctrl_t * guest_ctrl = GET_VMCB_CTRL_AREA((vmcb_t*)(info->vmm_data));
PrintDebug("RIP Linear: %p\n", (void *)linear_addr);
v3_print_segments(info);
v3_print_ctrl_regs(info);
+ if (info->shdw_pg_mode == SHADOW_PAGING) {
+ PrintDebug("Shadow Paging Guest Registers:\n");
+ PrintDebug("\tGuest CR0=%p\n", (void *)(addr_t)(info->shdw_pg_state.guest_cr0));
+ PrintDebug("\tGuest CR3=%p\n", (void *)(addr_t)(info->shdw_pg_state.guest_cr3));
+ // efer
+ // CR4
+ }
v3_print_GPRs(info);
-
-
PrintDebug("SVM Exit Code: %p\n", (void *)(addr_t)guest_ctrl->exit_code);
PrintDebug("exit_info1 low = 0x%.8x\n", *(uint_t*)&(guest_ctrl->exit_info1));
break;
}
+
}
return 0;
}
v3_cpuid(CPUID_FEATURE_IDS, &eax, &ebx, &ecx, &edx);
- PrintDebug("CPUID_FEATURE_IDS_ecx=0x%p\n", (void *)ecx);
+ PrintDebug("CPUID_FEATURE_IDS_ecx=%p\n", (void *)ecx);
if ((ecx & CPUID_FEATURE_IDS_ecx_svm_avail) == 0) {
PrintDebug("SVM Not Available\n");
v3_cpuid(CPUID_SVM_REV_AND_FEATURE_IDS, &eax, &ebx, &ecx, &edx);
- PrintDebug("CPUID_FEATURE_IDS_edx=0x%p\n", (void *)edx);
+ PrintDebug("CPUID_FEATURE_IDS_edx=%p\n", (void *)edx);
if ((edx & CPUID_SVM_REV_AND_FEATURE_IDS_edx_svml) == 0) {
PrintDebug("SVM BIOS Disabled, not unlockable\n");
PrintDebug("SVM is available and enabled.\n");
v3_cpuid(CPUID_SVM_REV_AND_FEATURE_IDS, &eax, &ebx, &ecx, &edx);
-
- PrintDebug("CPUID_FEATURE_IDS_edx=0x%p\n", (void *)edx);
+ PrintDebug("CPUID_FEATURE_IDS_eax=%p\n", (void *)eax);
+ PrintDebug("CPUID_FEATURE_IDS_ebx=%p\n", (void *)ebx);
+ PrintDebug("CPUID_FEATURE_IDS_ecx=%p\n", (void *)ecx);
+ PrintDebug("CPUID_FEATURE_IDS_edx=%p\n", (void *)edx);
+
if ((edx & CPUID_SVM_REV_AND_FEATURE_IDS_edx_np) == 0) {
PrintDebug("SVM Nested Paging not supported\n");