*/
#include <palacios/vmm.h>
-#include <palacios/svm.h>
-#include <palacios/vmx.h>
#include <palacios/vmm_intr.h>
#include <palacios/vmm_config.h>
#include <palacios/vm_guest.h>
#include <palacios/vmm_instrument.h>
+#include <palacios/vmm_ctrl_regs.h>
+#include <palacios/vmm_lowlevel.h>
+#include <palacios/vmm_sprintf.h>
+
+#ifdef CONFIG_SVM
+#include <palacios/svm.h>
+#endif
+#ifdef CONFIG_VMX
+#include <palacios/vmx.h>
+#endif
-/* These should be the only global variables in Palacios */
-/* They are architecture specific variables */
-v3_cpu_arch_t v3_cpu_type;
+v3_cpu_arch_t v3_cpu_types[CONFIG_MAX_CPUS];
struct v3_os_hooks * os_hooks = NULL;
+int v3_dbg_enable = 0;
static struct guest_info * allocate_guest() {
}
+static void init_cpu(void * arg) {
+ uint32_t cpu_id = (uint32_t)(addr_t)arg;
+
+#ifdef CONFIG_SVM
+ if (v3_is_svm_capable()) {
+ PrintDebug("Machine is SVM Capable\n");
+ v3_init_svm_cpu(cpu_id);
+
+ } else
+#endif
+#ifdef CONFIG_VMX
+ if (v3_is_vmx_capable()) {
+ PrintDebug("Machine is VMX Capable\n");
+ v3_init_vmx_cpu(cpu_id);
+
+ } else
+#endif
+ {
+ PrintError("CPU has no virtualization Extensions\n");
+ }
+}
+
+
+
+void Init_V3(struct v3_os_hooks * hooks, int num_cpus) {
+ int i;
+
+ V3_Print("V3 Print statement to fix a Kitten page fault bug\n");
-void Init_V3(struct v3_os_hooks * hooks, struct v3_ctrl_ops * vmm_ops) {
-
// Set global variables.
os_hooks = hooks;
- v3_cpu_type = V3_INVALID_CPU;
+ for (i = 0; i < CONFIG_MAX_CPUS; i++) {
+ v3_cpu_types[i] = V3_INVALID_CPU;
+ }
+
+ // Register all the possible device types
+ v3_init_devices();
#ifdef INSTRUMENT_VMM
v3_init_instrumentation();
#endif
- if (v3_is_svm_capable()) {
+ if ((hooks) && (hooks->call_on_cpu)) {
- PrintDebug("Machine is SVM Capable\n");
- vmm_ops->allocate_guest = &allocate_guest;
- v3_init_SVM(vmm_ops);
+ for (i = 0; i < num_cpus; i++) {
- /*
- } else if (is_vmx_capable()) {
- vmm_cpu_type = VMM_VMX_CPU;
- PrintDebug("Machine is VMX Capable\n");
- //Init_VMX();*/
+ V3_Print("Initializing VMM extensions on cpu %d\n", i);
+ hooks->call_on_cpu(i, &init_cpu, (void *)(addr_t)i);
+ }
+ }
+}
+
+
+v3_cpu_arch_t v3_get_cpu_type(int cpu_id) {
+ return v3_cpu_types[cpu_id];
+}
+
+
+struct guest_info * v3_create_vm(void * cfg) {
+ struct guest_info * info = allocate_guest();
+
+ if (!info) {
+ PrintError("Could not allocate Guest\n");
+ return NULL;
+ }
+
+ if (v3_config_guest(info, cfg) == -1) {
+ PrintError("Could not configure guest\n");
+ return NULL;
+ }
+
+ return info;
+}
+
+
+int v3_start_vm(struct guest_info * info, unsigned int cpu_mask) {
+
+ info->cpu_id = v3_get_cpu_id();
+
+ V3_Print("V3 -- Starting VM\n");
+
+ switch (v3_cpu_types[info->cpu_id]) {
+#ifdef CONFIG_SVM
+ case V3_SVM_CPU:
+ case V3_SVM_REV3_CPU:
+ return v3_start_svm_guest(info);
+ break;
+#endif
+#if CONFIG_VMX
+ case V3_VMX_CPU:
+ case V3_VMX_EPT_CPU:
+ return v3_start_vmx_guest(info);
+ break;
+#endif
+ default:
+ PrintError("Attemping to enter a guest on an invalid CPU\n");
+ return -1;
+ }
+
+ return 0;
+}
+
+
+#ifdef __V3_32BIT__
+
+v3_cpu_mode_t v3_get_host_cpu_mode() {
+ uint32_t cr4_val;
+ struct cr4_32 * cr4;
+
+ __asm__ (
+ "movl %%cr4, %0; "
+ : "=r"(cr4_val)
+ );
+
+
+ cr4 = (struct cr4_32 *)&(cr4_val);
+
+ if (cr4->pae == 1) {
+ return PROTECTED_PAE;
} else {
- PrintDebug("CPU has no virtualization Extensions\n");
+ return PROTECTED;
+ }
+}
+
+#elif __V3_64BIT__
+
+v3_cpu_mode_t v3_get_host_cpu_mode() {
+ return LONG;
+}
+
+#endif
+
+
+#define V3_Yield(addr) \
+ do { \
+ extern struct v3_os_hooks * os_hooks; \
+ if ((os_hooks) && (os_hooks)->yield_cpu) { \
+ (os_hooks)->yield_cpu(); \
+ } \
+ } while (0) \
+
+
+
+void v3_yield_cond(struct guest_info * info) {
+ uint64_t cur_cycle;
+ rdtscll(cur_cycle);
+
+ if (cur_cycle > (info->yield_start_cycle + info->yield_cycle_period)) {
+
+ /*
+ PrintDebug("Conditional Yield (cur_cyle=%p, start_cycle=%p, period=%p)\n",
+ (void *)cur_cycle, (void *)info->yield_start_cycle, (void *)info->yield_cycle_period);
+ */
+ V3_Yield();
+ rdtscll(info->yield_start_cycle);
+ }
+}
+
+
+/*
+ * unconditional cpu yield
+ * if the yielding thread is a guest context, the guest quantum is reset on resumption
+ * Non guest context threads should call this function with a NULL argument
+ */
+void v3_yield(struct guest_info * info) {
+ V3_Yield();
+
+ if (info) {
+ rdtscll(info->yield_start_cycle);
+ }
+}
+
+
+
+
+void v3_print_cond(const char * fmt, ...) {
+ if (v3_dbg_enable == 1) {
+ char buf[2048];
+ va_list ap;
+
+ va_start(ap, fmt);
+ vsnprintf(buf, 2048, fmt, ap);
+ va_end(ap);
+
+ V3_Print("%s", buf);
+ }
+}
+
+
+
+
+void v3_interrupt_cpu(struct guest_info * info, int logical_cpu) {
+ extern struct v3_os_hooks * os_hooks;
+
+ if ((os_hooks) && (os_hooks)->interrupt_cpu) {
+ (os_hooks)->interrupt_cpu(info, logical_cpu);
+ }
+}
+
+
+
+unsigned int v3_get_cpu_id() {
+ extern struct v3_os_hooks * os_hooks;
+ unsigned int ret = (unsigned int)-1;
+
+ if ((os_hooks) && (os_hooks)->get_cpu) {
+ ret = os_hooks->get_cpu();
+ }
+
+ return ret;
+}
+
+
+
+int v3_vm_enter(struct guest_info * info) {
+ switch (v3_cpu_types[info->cpu_id]) {
+#ifdef CONFIG_SVM
+ case V3_SVM_CPU:
+ case V3_SVM_REV3_CPU:
+ return v3_svm_enter(info);
+ break;
+#endif
+#if CONFIG_VMX
+ case V3_VMX_CPU:
+ case V3_VMX_EPT_CPU:
+ return v3_vmx_enter(info);
+ break;
+#endif
+ default:
+ PrintError("Attemping to enter a guest on an invalid CPU\n");
+ return -1;
}
}