*/
#include <palacios/vmm.h>
-#include <palacios/svm.h>
-#include <palacios/vmx.h>
#include <palacios/vmm_intr.h>
#include <palacios/vmm_config.h>
#include <palacios/vm_guest.h>
#include <palacios/vmm_instrument.h>
+#include <palacios/vmm_ctrl_regs.h>
+#include <palacios/vmm_lowlevel.h>
+#include <palacios/vmm_sprintf.h>
+
+#ifdef CONFIG_SVM
+#include <palacios/svm.h>
+#endif
+#ifdef CONFIG_VMX
+#include <palacios/vmx.h>
+#endif
+
+#ifdef CONFIG_VNET
+#include <palacios/vmm_vnet.h>
+#endif
-v3_cpu_arch_t v3_cpu_type;
+v3_cpu_arch_t v3_cpu_types[CONFIG_MAX_CPUS];
struct v3_os_hooks * os_hooks = NULL;
+int v3_dbg_enable = 0;
+
+
+
+static void init_cpu(void * arg) {
+ uint32_t cpu_id = (uint32_t)(addr_t)arg;
+
+#ifdef CONFIG_SVM
+ if (v3_is_svm_capable()) {
+ PrintDebug("Machine is SVM Capable\n");
+ v3_init_svm_cpu(cpu_id);
+
+ } else
+#endif
+#ifdef CONFIG_VMX
+ if (v3_is_vmx_capable()) {
+ PrintDebug("Machine is VMX Capable\n");
+ v3_init_vmx_cpu(cpu_id);
+
+ } else
+#endif
+ {
+ PrintError("CPU has no virtualizationExtensions\n");
+ }
+}
+
+
+
+void Init_V3(struct v3_os_hooks * hooks, int num_cpus) {
+ int i;
+
+ V3_Print("V3 Print statement to fix a Kitten page fault bug\n");
+
+ // Set global variables.
+ os_hooks = hooks;
+
+ for (i = 0; i < CONFIG_MAX_CPUS; i++) {
+ v3_cpu_types[i] = V3_INVALID_CPU;
+ }
+
+ // Register all the possible device types
+ v3_init_devices();
+
+ // Register all shadow paging handlers
+ V3_init_shdw_paging();
+
+
+#ifdef CONFIG_SYMMOD
+ V3_init_symmod();
+#endif
+
+#ifdef CONFIG_INSTRUMENT_VMM
+ v3_init_instrumentation();
+#endif
+
+
+#ifdef CONFIG_VNET
+ v3_init_vnet();
+#endif
+
+ if ((hooks) && (hooks->call_on_cpu)) {
+
+ for (i = 0; i < num_cpus; i++) {
+
+ V3_Print("Initializing VMM extensions on cpu %d\n", i);
+ hooks->call_on_cpu(i, &init_cpu, (void *)(addr_t)i);
+ }
+ }
+}
+
+
+v3_cpu_arch_t v3_get_cpu_type(int cpu_id) {
+ return v3_cpu_types[cpu_id];
+}
+
+
+struct v3_vm_info * v3_create_vm(void * cfg) {
+ struct v3_vm_info * vm = v3_config_guest(cfg);
+ if (vm == NULL) {
+ PrintError("Could not configure guest\n");
+ return NULL;
+ }
-static struct guest_info * allocate_guest() {
- void * info = V3_Malloc(sizeof(struct guest_info));
- memset(info, 0, sizeof(struct guest_info));
- return info;
+ return vm;
}
+static int start_core(void *p)
+{
+ struct guest_info * info = (struct guest_info*)p;
-void Init_V3(struct v3_os_hooks * hooks, struct v3_ctrl_ops * vmm_ops) {
- os_hooks = hooks;
- v3_cpu_type = V3_INVALID_CPU;
+ PrintDebug("core %u: in start_core\n",info->cpu_id);
+
+ // we assume here that the APs are in INIT mode
+ // and only the BSP is in REAL
+ // the per-architecture code will rely on this
+ // assumption
-#ifdef INSTRUMENT_VMM
- v3_init_instrumentation();
+
+ switch (v3_cpu_types[info->cpu_id]) {
+#ifdef CONFIG_SVM
+ case V3_SVM_CPU:
+ case V3_SVM_REV3_CPU:
+ return v3_start_svm_guest(info);
+ break;
+#endif
+#if CONFIG_VMX
+ case V3_VMX_CPU:
+ case V3_VMX_EPT_CPU:
+ return v3_start_vmx_guest(info);
+ break;
#endif
+ default:
+ PrintError("Attempting to enter a guest on an invalid CPU\n");
+ return -1;
+ }
+ // should not happen
+ return 0;
+}
+
- if (v3_is_svm_capable()) {
-
- PrintDebug("Machine is SVM Capable\n");
- vmm_ops->allocate_guest = &allocate_guest;
- vmm_ops->config_guest = &v3_config_guest;
- v3_init_SVM(vmm_ops);
-
- /*
- } else if (is_vmx_capable()) {
- vmm_cpu_type = VMM_VMX_CPU;
- PrintDebug("Machine is VMX Capable\n");
- //Init_VMX();*/
- } else {
- PrintDebug("CPU has no virtualization Extensions\n");
- }
+static uint32_t get_next_core(unsigned int cpu_mask, uint32_t last_proc)
+{
+ uint32_t proc_to_use;
+
+ PrintDebug("In get_next_core cpu_mask=0x%x last_proc=%u\n",cpu_mask,last_proc);
+
+ proc_to_use=(last_proc+1) % 32; // only 32 procs
+ // This will wrap around, and eventually we can use proc 0,
+ // since that's clearly available
+ while (!((cpu_mask >> proc_to_use)&0x1)) {
+ proc_to_use=(proc_to_use+1)%32;
+ }
+ return proc_to_use;
}
+int v3_start_vm(struct v3_vm_info * vm, unsigned int cpu_mask) {
+ uint32_t i;
+ uint32_t last_proc;
+ uint32_t proc_to_use;
+ char tname[16];
+
+ V3_Print("V3 -- Starting VM (%u cores)\n",vm->num_cores);
+
+ // We assume that we are running on CPU 0 of the underlying system
+ last_proc=0;
+
+ // We will fork off cores 1..n first, then boot core zero
+
+ // for the AP, we need to create threads
+
+ for (i = 1; i < vm->num_cores; i++) {
+ if (!os_hooks->start_thread_on_cpu) {
+ PrintError("Host OS does not support start_thread_on_cpu - FAILING\n");
+ return -1;
+ }
+
+ proc_to_use=get_next_core(cpu_mask,last_proc);
+ last_proc=proc_to_use;
+
+ // vm->cores[i].cpu_id=i;
+ // vm->cores[i].physical_cpu_id=proc_to_use;
+
+ PrintDebug("Starting virtual core %u on logical core %u\n",i,proc_to_use);
+
+ sprintf(tname,"core%u",i);
+
+ PrintDebug("run: core=%u, func=0x%p, arg=0x%p, name=%s\n",
+ proc_to_use, start_core, &(vm->cores[i]), tname);
+
+ // TODO: actually manage these threads instead of just launching them
+ if (!(os_hooks->start_thread_on_cpu(proc_to_use,start_core,&(vm->cores[i]),tname))) {
+ PrintError("Thread launch failed\n");
+ return -1;
+ }
+ }
-// Get CPU Type..
+ // vm->cores[0].cpu_id=0;
+ // vm->cores[0].physical_cpu_id=0;
+ // Finally launch the BSP on core 0
+ sprintf(tname,"core%u",0);
+
+#if CONFIG_LINUX
+ if (vm->num_cores==1) {
+ start_core(&(vm->cores[0]));
+ return -1;
+ } else {
+ if (!os_hooks->start_thread_on_cpu(0,start_core,&(vm->cores[0]),tname)) {
+ PrintError("Thread launch failed\n");
+ return -1;
+ }
+ }
+#else
+ if (!os_hooks->start_thread_on_cpu(0,start_core,&(vm->cores[0]),tname)) {
+ PrintError("Thread launch failed\n");
+ return -1;
+ }
+#endif
+
+ return 0;
+
+}
+
+
+#ifdef __V3_32BIT__
+
+v3_cpu_mode_t v3_get_host_cpu_mode() {
+ uint32_t cr4_val;
+ struct cr4_32 * cr4;
+
+ __asm__ (
+ "movl %%cr4, %0; "
+ : "=r"(cr4_val)
+ );
+
+
+ cr4 = (struct cr4_32 *)&(cr4_val);
+
+ if (cr4->pae == 1) {
+ return PROTECTED_PAE;
+ } else {
+ return PROTECTED;
+ }
+}
+
+#elif __V3_64BIT__
+
+v3_cpu_mode_t v3_get_host_cpu_mode() {
+ return LONG;
+}
+
+#endif
+
+
+#define V3_Yield(addr) \
+ do { \
+ extern struct v3_os_hooks * os_hooks; \
+ if ((os_hooks) && (os_hooks)->yield_cpu) { \
+ (os_hooks)->yield_cpu(); \
+ } \
+ } while (0) \
+
+
+
+void v3_yield_cond(struct guest_info * info) {
+ uint64_t cur_cycle;
+ rdtscll(cur_cycle);
+
+ if (cur_cycle > (info->yield_start_cycle + info->vm_info->yield_cycle_period)) {
+
+ /*
+ PrintDebug("Conditional Yield (cur_cyle=%p, start_cycle=%p, period=%p)\n",
+ (void *)cur_cycle, (void *)info->yield_start_cycle, (void *)info->yield_cycle_period);
+ */
+ V3_Yield();
+ rdtscll(info->yield_start_cycle);
+ }
+}
+
+
+/*
+ * unconditional cpu yield
+ * if the yielding thread is a guest context, the guest quantum is reset on resumption
+ * Non guest context threads should call this function with a NULL argument
+ */
+void v3_yield(struct guest_info * info) {
+ V3_Yield();
+
+ if (info) {
+ rdtscll(info->yield_start_cycle);
+ }
+}
+
+
+
+
+void v3_print_cond(const char * fmt, ...) {
+ if (v3_dbg_enable == 1) {
+ char buf[2048];
+ va_list ap;
+
+ va_start(ap, fmt);
+ vsnprintf(buf, 2048, fmt, ap);
+ va_end(ap);
+
+ V3_Print("%s", buf);
+ }
+}
+
+
+
+
+void v3_interrupt_cpu(struct v3_vm_info * vm, int logical_cpu, int vector) {
+ extern struct v3_os_hooks * os_hooks;
+
+ if ((os_hooks) && (os_hooks)->interrupt_cpu) {
+ (os_hooks)->interrupt_cpu(vm, logical_cpu, vector);
+ }
+}
+
+
+
+unsigned int v3_get_cpu_id() {
+ extern struct v3_os_hooks * os_hooks;
+ unsigned int ret = (unsigned int)-1;
+
+ if ((os_hooks) && (os_hooks)->get_cpu) {
+ ret = os_hooks->get_cpu();
+ }
+
+ return ret;
+}
+
+
+
+int v3_vm_enter(struct guest_info * info) {
+ switch (v3_cpu_types[info->cpu_id]) {
+#ifdef CONFIG_SVM
+ case V3_SVM_CPU:
+ case V3_SVM_REV3_CPU:
+ return v3_svm_enter(info);
+ break;
+#endif
+#if CONFIG_VMX
+ case V3_VMX_CPU:
+ case V3_VMX_EPT_CPU:
+ return v3_vmx_enter(info);
+ break;
+#endif
+ default:
+ PrintError("Attemping to enter a guest on an invalid CPU\n");
+ return -1;
+ }
+}