#include <palacios/vmx.h>
#endif
+#ifdef CONFIG_VNET
+#include <palacios/vmm_vnet.h>
+#endif
+
v3_cpu_arch_t v3_cpu_types[CONFIG_MAX_CPUS];
struct v3_os_hooks * os_hooks = NULL;
} else
#endif
{
- PrintError("CPU has no virtualization Extensions\n");
+ PrintError("CPU has no virtualizationExtensions\n");
}
}
v3_init_instrumentation();
#endif
+
+#ifdef CONFIG_VNET
+ V3_init_vnet();
+#endif
+
if ((hooks) && (hooks->call_on_cpu)) {
for (i = 0; i < num_cpus; i++) {
}
-struct v3_vm_info * v3_create_vm(void * cfg) {
+struct v3_vm_info * v3_create_vm(void * cfg, void * priv_data) {
struct v3_vm_info * vm = v3_config_guest(cfg);
if (vm == NULL) {
return NULL;
}
+ vm->host_priv_data = priv_data;
+
return vm;
}
-int v3_start_vm(struct v3_vm_info * vm, unsigned int cpu_mask) {
- int i = 0;
- V3_Print("V3 -- Starting VM\n");
+static int start_core(void *p)
+{
+ struct guest_info * info = (struct guest_info *)p;
- for (i = 0; i < vm->num_cores; i++) {
- struct guest_info * info = &(vm->cores[i]);
+ PrintDebug("core %u: in start_core\n",info->cpu_id);
+
+ // we assume here that the APs are in INIT mode
+ // and only the BSP is in REAL
+ // the per-architecture code will rely on this
+ // assumption
- /* GRUESOM HACK... */
- // vm->cpu_id = v3_get_cpu_id();
- switch (v3_cpu_types[info->cpu_id]) {
+ switch (v3_cpu_types[info->cpu_id]) {
#ifdef CONFIG_SVM
- case V3_SVM_CPU:
- case V3_SVM_REV3_CPU:
- return v3_start_svm_guest(info);
- break;
+ case V3_SVM_CPU:
+ case V3_SVM_REV3_CPU:
+ return v3_start_svm_guest(info);
+ break;
#endif
#if CONFIG_VMX
- case V3_VMX_CPU:
- case V3_VMX_EPT_CPU:
- return v3_start_vmx_guest(info);
- break;
+ case V3_VMX_CPU:
+ case V3_VMX_EPT_CPU:
+ return v3_start_vmx_guest(info);
+ break;
#endif
- default:
- PrintError("Attemping to enter a guest on an invalid CPU\n");
- return -1;
+ default:
+ PrintError("Attempting to enter a guest on an invalid CPU\n");
+ return -1;
+ }
+ // should not happen
+ return 0;
+}
+
+
+static uint32_t get_next_core(unsigned int cpu_mask, uint32_t last_proc)
+{
+ uint32_t proc_to_use;
+
+ PrintDebug("In get_next_core cpu_mask=0x%x last_proc=%u\n",cpu_mask,last_proc);
+
+ proc_to_use=(last_proc+1) % 32; // only 32 procs
+ // This will wrap around, and eventually we can use proc 0,
+ // since that's clearly available
+ while (!((cpu_mask >> proc_to_use)&0x1)) {
+ proc_to_use=(proc_to_use+1)%32;
+ }
+ return proc_to_use;
+}
+
+int v3_start_vm(struct v3_vm_info * vm, unsigned int cpu_mask) {
+ uint32_t i;
+ uint32_t last_proc;
+ uint32_t proc_to_use;
+ char tname[16];
+
+ V3_Print("V3 -- Starting VM (%u cores)\n",vm->num_cores);
+
+ // We assume that we are running on CPU 0 of the underlying system
+ last_proc=0;
+
+ // We will fork off cores 1..n first, then boot core zero
+
+ // for the AP, we need to create threads
+
+ for (i = 1; i < vm->num_cores; i++) {
+ if (!os_hooks->start_thread_on_cpu) {
+ PrintError("Host OS does not support start_thread_on_cpu - FAILING\n");
+ return -1;
+ }
+
+ proc_to_use=get_next_core(cpu_mask,last_proc);
+ last_proc=proc_to_use;
+
+ // vm->cores[i].cpu_id=i;
+ // vm->cores[i].physical_cpu_id=proc_to_use;
+
+ PrintDebug("Starting virtual core %u on logical core %u\n",i,proc_to_use);
+
+ sprintf(tname,"core%u",i);
+
+ PrintDebug("run: core=%u, func=0x%p, arg=0x%p, name=%s\n",
+ proc_to_use, start_core, &(vm->cores[i]), tname);
+
+ // TODO: actually manage these threads instead of just launching them
+ if (!(os_hooks->start_thread_on_cpu(proc_to_use,start_core,&(vm->cores[i]),tname))) {
+ PrintError("Thread launch failed\n");
+ return -1;
+ }
+ }
+
+ // vm->cores[0].cpu_id=0;
+ // vm->cores[0].physical_cpu_id=0;
+
+ // Finally launch the BSP on core 0
+ sprintf(tname,"core%u",0);
+
+#if CONFIG_LINUX
+ if (vm->num_cores==1) {
+ start_core(&(vm->cores[0]));
+ return -1;
+ } else {
+ if (!os_hooks->start_thread_on_cpu(0,start_core,&(vm->cores[0]),tname)) {
+ PrintError("Thread launch failed\n");
+ return -1;
}
}
+#else
+ if (!os_hooks->start_thread_on_cpu(0,start_core,&(vm->cores[0]),tname)) {
+ PrintError("Thread launch failed\n");
+ return -1;
+ }
+#endif
return 0;
+
}
void v3_yield_cond(struct guest_info * info) {
uint64_t cur_cycle;
- rdtscll(cur_cycle);
+ cur_cycle = v3_get_host_time(&info->time_state);
if (cur_cycle > (info->yield_start_cycle + info->vm_info->yield_cycle_period)) {
(void *)cur_cycle, (void *)info->yield_start_cycle, (void *)info->yield_cycle_period);
*/
V3_Yield();
- rdtscll(info->yield_start_cycle);
+ info->yield_start_cycle = v3_get_host_time(&info->time_state);
}
}
V3_Yield();
if (info) {
- rdtscll(info->yield_start_cycle);
+ info->yield_start_cycle = v3_get_host_time(&info->time_state);
}
}
-void v3_interrupt_cpu(struct v3_vm_info * vm, int logical_cpu) {
+void v3_interrupt_cpu(struct v3_vm_info * vm, int logical_cpu, int vector) {
extern struct v3_os_hooks * os_hooks;
if ((os_hooks) && (os_hooks)->interrupt_cpu) {
- (os_hooks)->interrupt_cpu(vm, logical_cpu);
+ (os_hooks)->interrupt_cpu(vm, logical_cpu, vector);
}
}