X-Git-Url: http://v3vee.org/palacios/gitweb/gitweb.cgi?a=blobdiff_plain;f=palacios%2Fsrc%2Fpalacios%2Fvmm.c;h=0a281ab684c698a6988e87f37d0b7a925f6de44b;hb=6b9abb54ebafd8266f1711b803ccb027675a465f;hp=490f106dd9b7e06800f5c1ba3623c67db0caca9d;hpb=604e1d91e2cb21ca9b0afcb9a27cabd7119d67a4;p=palacios.git diff --git a/palacios/src/palacios/vmm.c b/palacios/src/palacios/vmm.c index 490f106..0a281ab 100644 --- a/palacios/src/palacios/vmm.c +++ b/palacios/src/palacios/vmm.c @@ -154,6 +154,14 @@ void Init_V3(struct v3_os_hooks * hooks, char * cpu_mask, int num_cpus, char *op // Parse host-os defined options into an easily-accessed format. v3_parse_options(options); +#ifdef V3_CONFIG_MULTIBOOT + v3_init_multiboot(); +#endif + +#ifdef V3_CONFIG_HVM + v3_init_hvm(); +#endif + // Memory manager initialization v3_init_mem(); @@ -255,6 +263,14 @@ void Shutdown_V3() { v3_deinit_mem(); +#ifdef V3_CONFIG_HVM + v3_deinit_hvm(); +#endif + +#ifdef V3_CONFIG_MULTIBOOT + v3_deinit_multiboot(); +#endif + v3_deinit_options(); @@ -265,9 +281,42 @@ v3_cpu_arch_t v3_get_cpu_type(int cpu_id) { return v3_cpu_types[cpu_id]; } +static int start_core(void * p) +{ + struct guest_info * core = (struct guest_info *)p; + + if (v3_scheduler_register_core(core) == -1){ + PrintError(core->vm_info, core,"Error initializing scheduling in core %d\n", core->vcpu_id); + } + + PrintDebug(core->vm_info,core,"virtual core %u (on logical core %u): in start_core (RIP=%p)\n", + core->vcpu_id, core->pcpu_id, (void *)(addr_t)core->rip); -struct v3_vm_info * v3_create_vm(void * cfg, void * priv_data, char * name) { + switch (v3_mach_type) { +#ifdef V3_CONFIG_SVM + case V3_SVM_CPU: + case V3_SVM_REV3_CPU: + return v3_start_svm_guest(core); + break; +#endif +#if V3_CONFIG_VMX + case V3_VMX_CPU: + case V3_VMX_EPT_CPU: + case V3_VMX_EPT_UG_CPU: + return v3_start_vmx_guest(core); + break; +#endif + default: + PrintError(core->vm_info, core, "Attempting to enter a guest on an invalid CPU\n"); + return -1; + } + // should not happen + return 0; +} + +struct v3_vm_info * v3_create_vm(void * cfg, void * priv_data, char * name, unsigned int cpu_mask) { struct v3_vm_info * vm = v3_config_guest(cfg, priv_data); + int vcore_id = 0; if (vm == NULL) { PrintError(VM_NONE, VCORE_NONE, "Could not configure guest\n"); @@ -299,43 +348,32 @@ struct v3_vm_info * v3_create_vm(void * cfg, void * priv_data, char * name) { PrintError(vm, VCORE_NONE,"Error registering VM with scheduler\n"); } - return vm; -} + if (v3_cpu_mapper_admit_vm(vm,cpu_mask) != 0){ + PrintError(vm, VCORE_NONE,"Error admitting VM %s for mapping", vm->name); + } + for (vcore_id = 0; vcore_id < vm->num_cores; vcore_id++) { + struct guest_info * core = &(vm->cores[vcore_id]); + PrintDebug(vm, VCORE_NONE, "Creating virtual core %u on logical core %u\n", + vcore_id, core->pcpu_id); -static int start_core(void * p) -{ - struct guest_info * core = (struct guest_info *)p; + sprintf(core->exec_name, "%s-%u", vm->name, vcore_id); - if (v3_scheduler_register_core(core) == -1){ - PrintError(core->vm_info, core,"Error initializing scheduling in core %d\n", core->vcpu_id); - } + PrintDebug(vm, VCORE_NONE, "run: core=%u, func=0x%p, arg=0x%p, name=%s\n", + core->pcpu_id, start_core, core, core->exec_name); - PrintDebug(core->vm_info,core,"virtual core %u (on logical core %u): in start_core (RIP=%p)\n", - core->vcpu_id, core->pcpu_id, (void *)(addr_t)core->rip); + core->core_thread = V3_CREATE_THREAD_ON_CPU(core->pcpu_id, start_core, core, core->exec_name); + + if (core->core_thread == NULL) { + PrintError(vm, VCORE_NONE, "Thread creation failed\n"); + v3_stop_vm(vm); + return NULL; + } - switch (v3_mach_type) { -#ifdef V3_CONFIG_SVM - case V3_SVM_CPU: - case V3_SVM_REV3_CPU: - return v3_start_svm_guest(core); - break; -#endif -#if V3_CONFIG_VMX - case V3_VMX_CPU: - case V3_VMX_EPT_CPU: - case V3_VMX_EPT_UG_CPU: - return v3_start_vmx_guest(core); - break; -#endif - default: - PrintError(core->vm_info, core, "Attempting to enter a guest on an invalid CPU\n"); - return -1; } - // should not happen - return 0; + return vm; } int v3_start_vm(struct v3_vm_info * vm, unsigned int cpu_mask) { @@ -355,6 +393,18 @@ int v3_start_vm(struct v3_vm_info * vm, unsigned int cpu_mask) { return -1; } +#if V3_CONFIG_MULTIBOOT + if (v3_setup_multiboot_vm_for_boot(vm)) { + PrintError(vm, VCORE_NONE, "Multiboot setup for boot failed\n"); + return -1; + } +#endif +#if V3_CONFIG_HVM + if (v3_setup_hvm_vm_for_boot(vm)) { + PrintError(vm, VCORE_NONE, "HVM setup for boot failed\n"); + return -1; + } +#endif /// CHECK IF WE ARE MULTICORE ENABLED.... @@ -377,10 +427,6 @@ int v3_start_vm(struct v3_vm_info * vm, unsigned int cpu_mask) { } vm->avail_cores = avail_cores; - - if (v3_cpu_mapper_admit_vm(vm,cpu_mask) != 0){ - PrintError(vm, VCORE_NONE,"Error admitting VM %s for mapping", vm->name); - } if (v3_scheduler_admit_vm(vm) != 0){ PrintError(vm, VCORE_NONE,"Error admitting VM %s for scheduling", vm->name); @@ -388,34 +434,23 @@ int v3_start_vm(struct v3_vm_info * vm, unsigned int cpu_mask) { vm->run_state = VM_RUNNING; - for (vcore_id = 0; vcore_id < vm->num_cores; vcore_id++) { struct guest_info * core = &(vm->cores[vcore_id]); PrintDebug(vm, VCORE_NONE, "Starting virtual core %u on logical core %u\n", vcore_id, core->pcpu_id); - - sprintf(core->exec_name, "%s-%u", vm->name, vcore_id); - PrintDebug(vm, VCORE_NONE, "run: core=%u, func=0x%p, arg=0x%p, name=%s\n", - core->pcpu_id, start_core, core, core->exec_name); - - if (core->core_run_state==CORE_INVALID) { + if (core->core_run_state==CORE_INVALID) { // launch of a fresh VM - core->core_run_state = CORE_STOPPED; + core->core_run_state = CORE_STOPPED; // core zero will turn itself on } else { // this is a resume - use whatever its current run_state is } - core->core_thread = V3_CREATE_THREAD_ON_CPU(core->pcpu_id, start_core, core, core->exec_name); + V3_START_THREAD(core->core_thread); - if (core->core_thread == NULL) { - PrintError(vm, VCORE_NONE, "Thread launch failed\n"); - v3_stop_vm(vm); - return -1; - } } return 0; @@ -449,6 +484,99 @@ int v3_reset_vm_core(struct guest_info * core, addr_t rip) { } +// resets the whole VM (non-HVM) or the ROS (HVM) +int v3_reset_vm(struct v3_vm_info *vm) +{ +#ifdef V3_CONFIG_HVM + if (vm->hvm_state.is_hvm) { + return v3_reset_vm_extended(vm,V3_VM_RESET_ROS,0); + } else { + return v3_reset_vm_extended(vm,V3_VM_RESET_ALL,0); + } +#else + return v3_reset_vm_extended(vm,V3_VM_RESET_ALL,0); +#endif +} + +int v3_reset_vm_extended(struct v3_vm_info *vm, v3_vm_reset_type t, void *data) +{ + uint32_t start, end, i; + uint32_t newcount; + + if (vm->run_state != VM_RUNNING) { + PrintError(vm,VCORE_NONE,"Attempt to reset VM in state %d (must be in running state)\n",vm->run_state); + return -1; + } + + + switch (t) { + case V3_VM_RESET_ALL: +#ifdef V3_CONFIG_HVM + if (vm->hvm_state.is_hvm) { + PrintError(vm,VCORE_NONE,"Attempt to do ALL reset of HVM (not allowed)\n"); + return -1; + } +#endif + start=0; end=vm->num_cores-1; + break; +#ifdef V3_CONFIG_HVM + case V3_VM_RESET_HRT: + case V3_VM_RESET_ROS: + if (vm->hvm_state.is_hvm) { + if (t==V3_VM_RESET_HRT) { + start = vm->hvm_state.first_hrt_core; + end = vm->num_cores-1; + } else { + start = 0; + end = vm->hvm_state.first_hrt_core-1; + } + } else { + PrintError(vm,VCORE_NONE,"This is not an HVM and so HVM-specific resets do not apply\n"); + return -1; + } +#endif + break; + case V3_VM_RESET_CORE_RANGE: + start = ((uint32_t*)data)[0]; + end = ((uint32_t*)data)[1]; + break; + default: + PrintError(vm,VCORE_NONE,"Unsupported reset type %d for this VM\n",t); + return -1; + break; + } + + PrintDebug(vm,VCORE_NONE,"Resetting cores %d through %d\n",start,end); + + newcount = end-start+1; + + for (i=start;i<=end;i++) { + if (!(vm->cores[i].core_run_state == CORE_RUNNING || vm->cores[i].core_run_state == CORE_STOPPED)) { + PrintError(vm,VCORE_NONE,"Cannot reset VM as core %u is in state %d (must be running or stopped)\n",i,vm->cores[i].core_run_state); + return -1; + } + } + + + // This had better be the only thread using the barrier at this point... + v3_init_counting_barrier(&vm->reset_barrier,newcount); + + // OK, I am the reseter, tell the relevant cores what to do + // each will atomically decrement the reset countdown and then + // spin waiting for it to hit zero. + + for (i=start;i<=end;i++) { + vm->cores[i].core_run_state = CORE_RESETTING; + // force exit of core + v3_interrupt_cpu(vm, vm->cores[i].pcpu_id, 0); + } + + // we don't wait for reset to finish + // because reset could have been initiated by a core + + return 0; +} + /* move a virtual core to different physical core */ int v3_move_vm_core(struct v3_vm_info * vm, int vcore_id, int target_cpu) { @@ -595,7 +723,7 @@ int v3_move_vm_mem(struct v3_vm_info * vm, void *gpa, int target_cpu) { new_hpa = V3_AllocPagesExtended(num_pages, PAGE_SIZE_4KB, new_node, - 0); // no constraints given new shadow pager impl + 0, 0); // no constraints given new shadow pager impl if (!new_hpa) { PrintError(vm, VCORE_NONE, "Cannot allocate memory for new base region...\n"); @@ -849,9 +977,10 @@ int v3_get_state_vm(struct v3_vm_info *vm, uint32_t numcores; uint32_t numregions; extern uint64_t v3_mem_block_size; + void *cur_gpa; if (!vm || !base || !core || !mem) { - PrintError(VM_NONE, VCORE_NONE, "Invalid rquest to v3_get_state_vm\n"); + PrintError(VM_NONE, VCORE_NONE, "Invalid request to v3_get_state_vm\n"); return -1; } @@ -865,14 +994,24 @@ int v3_get_state_vm(struct v3_vm_info *vm, case VM_PAUSED: base->state = V3_VM_PAUSED; break; case VM_ERROR: base->state = V3_VM_ERROR; break; case VM_SIMULATING: base->state = V3_VM_SIMULATING; break; + case VM_RESETTING: base->state = V3_VM_RESETTING; break; default: base->state = V3_VM_UNKNOWN; break; } + base->vm_type = V3_VM_GENERAL; + +#ifdef V3_CONFIG_HVM + if (vm->hvm_state.is_hvm) { + base->vm_type = V3_VM_HVM; + } +#endif + for (i=0;icores[i].core_run_state) { case CORE_INVALID: core->vcore[i].state = V3_VCORE_INVALID; break; case CORE_RUNNING: core->vcore[i].state = V3_VCORE_RUNNING; break; case CORE_STOPPED: core->vcore[i].state = V3_VCORE_STOPPED; break; + case CORE_RESETTING: core->vcore[i].state = V3_VCORE_RESETTING; break; default: core->vcore[i].state = V3_VCORE_UNKNOWN; break; } switch (vm->cores[i].cpu_mode) { @@ -894,6 +1033,18 @@ int v3_get_state_vm(struct v3_vm_info *vm, case VIRTUAL_MEM: core->vcore[i].mem_mode=V3_VCORE_MEM_MODE_VIRTUAL; break; default: core->vcore[i].mem_mode=V3_VCORE_MEM_MODE_UNKNOWN; break; } + + core->vcore[i].vcore_type = V3_VCORE_GENERAL; + +#ifdef V3_CONFIG_HVM + if (vm->hvm_state.is_hvm) { + if (v3_is_hvm_ros_core(&vm->cores[i])) { + core->vcore[i].vcore_type = V3_VCORE_ROS; + } else { + core->vcore[i].vcore_type = V3_VCORE_HRT; + } + } +#endif core->vcore[i].pcore=vm->cores[i].pcpu_id; core->vcore[i].last_rip=(void*)(vm->cores[i].rip); @@ -902,7 +1053,10 @@ int v3_get_state_vm(struct v3_vm_info *vm, core->num_vcores=numcores; - for (i=0;imem_map.num_base_regions;i++) { + cur_gpa=0; + + for (i=0;iregion[i].guest_paddr = cur_gpa; mem->region[i].host_paddr = (void*)(vm->mem_map.base_regions[i].host_addr); mem->region[i].size = v3_mem_block_size; #ifdef V3_CONFIG_SWAPPING @@ -913,9 +1067,35 @@ int v3_get_state_vm(struct v3_vm_info *vm, mem->region[i].pinned = 0; #endif + cur_gpa += mem->region[i].size; } mem->num_regions=numregions; + + + mem->mem_size=vm->mem_size; + mem->ros_mem_size=vm->mem_size; + +#ifdef V3_CONFIG_HVM + if (vm->hvm_state.is_hvm) { + mem->ros_mem_size=v3_get_hvm_ros_memsize(vm); + } +#endif + + return 0; +} + +int v3_get_state_sizes_vm(struct v3_vm_info *vm, + unsigned long long *num_vcores, + unsigned long long *num_regions) +{ + if (!vm || !num_vcores || !num_regions) { + PrintError(VM_NONE, VCORE_NONE, "Invalid request to v3_get_state_sizes\n"); + return -1; + } + + *num_vcores = vm->num_cores; + *num_regions = vm->mem_map.num_base_regions; return 0; }