* All rights reserved.
*
* Author: Jack Lange <jarusl@cs.northwestern.edu>
+ * Patrick G. Bridges <bridges@cs.unm.edu>
*
* This is free software. You are permitted to use,
* redistribute, and modify it as specified in the file "V3VEE_LICENSE".
*/
-#include <palacios/vmm_time.h>
#include <palacios/vmm.h>
+#include <palacios/vmm_time.h>
#include <palacios/vm_guest.h>
-void v3_init_time(struct guest_info * info) {
- struct vm_time * time_state = &(info->time_state);
+#ifndef V3_CONFIG_DEBUG_TIME
+#undef PrintDebug
+#define PrintDebug(fmt, args...)
+#endif
- time_state->cpu_freq = V3_CPU_KHZ();
-
- time_state->guest_tsc = 0;
- time_state->cached_host_tsc = 0;
- // time_state->pending_cycles = 0;
-
- INIT_LIST_HEAD(&(time_state->timers));
- time_state->num_timers = 0;
+/* Overview
+ *
+ * Time handling in VMMs is challenging, and Palacios uses the highest
+ * resolution, lowest overhead timer on modern CPUs that it can - the
+ * processor timestamp counter (TSC). Note that on somewhat old processors
+ * this can be problematic; in particular, older AMD processors did not
+ * have a constant rate timestamp counter in the face of power management
+ * events. However, the latest Intel and AMD CPUs all do (should...) have a
+ * constant rate TSC, and Palacios relies on this fact.
+ *
+ * Basically, Palacios keeps track of three quantities as it runs to manage
+ * the passage of time:
+ * (1) The host timestamp counter - read directly from HW and never written
+ * (2) A monotonic guest timestamp counter used to measure the progression of
+ * time in the guest. This is stored as an absolute number of cycles elapsed
+ * and is updated on guest entry and exit; it can also be updated explicitly
+ * in the monitor at times
+ * (3) The actual guest timestamp counter (which can be written by
+ * writing to the guest TSC MSR - MSR 0x10) from the monotonic guest TSC.
+ * This is also computed as an offset from (2) above when the TSC and
+ * this offset is updated when the TSC MSR is written.
+ *
+ * Because all other devices are slaved off of the passage of time in the guest,
+ * it is (2) above that drives the firing of other timers in the guest,
+ * including timer devices such as the Programmable Interrupt Timer (PIT).
+ *
+ * Future additions:
+ * (1) Add support for temporarily skewing guest time off of where it should
+ * be to support slack simulation of guests. The idea is that simulators
+ * set this skew to be the difference between how much time passed for a
+ * simulated feature and a real implementation of that feature, making time
+ * pass at a different rate from real time on this core. The VMM will then
+ * attempt to move this skew back towards 0 subject to resolution/accuracy
+ * constraints from various system timers.
+ *
+ * The main effort in doing this will be to get accuracy/resolution
+ * information from each local timer and to use this to bound how much skew
+ * is removed on each exit.
+ *
+ * (2) Look more into sychronizing the offsets *across* virtual and physical
+ * cores so that multicore guests stay mostly in sync.
+ *
+ * (3) Look into using the AMD TSC multiplier feature and adding explicit time
+ * dilation support to time handling.
+ */
+
+
+static int handle_cpufreq_hcall(struct guest_info * info, uint_t hcall_id, void * priv_data) {
+ struct vm_core_time * time_state = &(info->time_state);
+
+ info->vm_regs.rbx = time_state->guest_cpu_freq;
+
+ PrintDebug("Guest request cpu frequency: return %ld\n", (long)info->vm_regs.rbx);
+
+ return 0;
}
-int v3_add_timer(struct guest_info * info, struct vm_timer_ops * ops, void * private_data) {
- struct vm_timer * timer = NULL;
- timer = (struct vm_timer *)V3_Malloc(sizeof(struct vm_timer));
+
+int v3_start_time(struct guest_info * info) {
+ /* We start running with guest_time == host_time */
+ uint64_t t = v3_get_host_time(&info->time_state);
+
+ info->time_state.vm_enter_host_time = 0;
+ info->time_state.vm_pause_host_time = t;
+ info->time_state.initial_host_time = t;
+ info->yield_start_cycle = t;
+
+ info->time_state.last_update = 0;
+ info->time_state.guest_cycles = 0;
+ PrintDebug("Starting time for core %d at host time %llu/guest time %llu.\n",
+ info->vcpu_id, t, info->time_state.guest_cycles);
+ v3_yield(info);
+ return 0;
+}
+
+int v3_offset_time( struct guest_info * info, sint64_t offset )
+{
+ struct vm_core_time * time_state = &(info->time_state);
+ if (info->vm_info->time_state.follow_host_time) {
+ PrintError("Cannot offset guest time passage while slaved to host clock.\n");
+ return 1;
+ } else {
+ time_state->guest_cycles += offset;
+ }
+ return 0;
+}
+
+int v3_skip_time(struct guest_info * info) {
+ if (info->vm_info->time_state.follow_host_time) {
+ PrintError("Cannot skip guest time passage while slaved to host clock.\n");
+ return 1;
+ } else {
+ info->time_state.vm_pause_host_time = v3_get_host_time(&info->time_state);
+ }
+ return 0;
+}
+
+static sint64_t host_to_guest_cycles(struct guest_info * info, sint64_t host_cycles) {
+ return (host_cycles * info->time_state.clock_ratio_num) / info->time_state.clock_ratio_denom;
+}
+
+int v3_time_advance_cycles(struct guest_info * info, uint64_t *host_cycles)
+{
+ uint64_t t = v3_get_host_time(&info->time_state);
+
+ info->time_state.vm_pause_host_time = t;
+
+ if (info->vm_info->time_state.follow_host_time) {
+ /* How many guest cycles should have elapsed? */
+ sint64_t host_elapsed = t - info->time_state.initial_host_time;
+ sint64_t guest_elapsed = host_to_guest_cycles(info, host_elapsed);
+
+ info->time_state.guest_cycles = guest_elapsed;
+ } else {
+ uint64_t guest_cycles;
+ if (*host_cycles) {
+ guest_cycles = host_to_guest_cycles(info, *host_cycles);
+ } else {
+ guest_cycles = host_to_guest_cycles(info, (sint64_t)(t - info->time_state.vm_pause_host_time));
+ }
+ info->time_state.guest_cycles += guest_cycles;
+ }
+
+ return 0;
+}
+
+int v3_advance_time(struct guest_info * info) {
+ return v3_time_advance_cycles(info, NULL);
+}
+
+/* Called immediately upon entry in the the VMM */
+int
+v3_time_exit_vm( struct guest_info * info, uint64_t * host_cycles )
+{
+ return v3_time_advance_cycles(info, host_cycles);
+}
+
+/* Called immediately prior to entry to the VM */
+int
+v3_time_enter_vm( struct guest_info * info )
+{
+ struct vm_core_time * time_state = &(info->time_state);
+ uint64_t host_time = v3_get_host_time(&info->time_state);
+
+ time_state->vm_enter_host_time = host_time;
+ return 0;
+}
+
+
+
+struct v3_timer * v3_add_timer(struct guest_info * info,
+ struct v3_timer_ops * ops,
+ void * private_data) {
+ struct v3_timer * timer = NULL;
+ timer = (struct v3_timer *)V3_Malloc(sizeof(struct v3_timer));
V3_ASSERT(timer != NULL);
timer->ops = ops;
list_add(&(timer->timer_link), &(info->time_state.timers));
info->time_state.num_timers++;
- return 0;
+ return timer;
}
-
-int v3_remove_timer(struct guest_info * info, struct vm_timer * timer) {
+int v3_remove_timer(struct guest_info * info, struct v3_timer * timer) {
list_del(&(timer->timer_link));
info->time_state.num_timers--;
return 0;
}
+void v3_update_timers(struct guest_info * info) {
+ struct vm_core_time *time_state = &info->time_state;
+ struct v3_timer * tmp_timer;
+ sint64_t cycles;
+ uint64_t old_time = time_state->last_update;
+
+ time_state->last_update = v3_get_guest_time(time_state);
+ cycles = (sint64_t)(time_state->last_update - old_time);
+ if (cycles < 0) {
+ PrintError("Cycles appears to have rolled over - old time %lld, current time %lld.\n",
+ old_time, time_state->last_update);
+ return;
+ }
+
+ PrintDebug("Updating timers with %lld elapsed cycles.\n", cycles);
+ list_for_each_entry(tmp_timer, &(time_state->timers), timer_link) {
+ tmp_timer->ops->update_timer(info, cycles, time_state->guest_cpu_freq, tmp_timer->private_data);
+ }
+}
+
+/*
+ * Handle full virtualization of the time stamp counter. As noted
+ * above, we don't store the actual value of the TSC, only the guest's
+ * offset from monotonic guest's time. If the guest writes to the TSC, we
+ * handle this by changing that offset.
+ *
+ * Possible TODO: Proper hooking of TSC read/writes?
+ */
-void v3_update_time(struct guest_info * info, uint64_t cycles) {
- struct vm_timer * tmp_timer;
+int v3_rdtsc(struct guest_info * info) {
+ uint64_t tscval = v3_get_guest_tsc(&info->time_state);
+
+ info->vm_regs.rdx = tscval >> 32;
+ info->vm_regs.rax = tscval & 0xffffffffLL;
+
+ return 0;
+}
+
+int v3_handle_rdtsc(struct guest_info * info) {
+ v3_rdtsc(info);
+
+ info->vm_regs.rax &= 0x00000000ffffffffLL;
+ info->vm_regs.rdx &= 0x00000000ffffffffLL;
+
+ info->rip += 2;
- // cycles *= 8;
+ return 0;
+}
+
+int v3_rdtscp(struct guest_info * info) {
+ int ret;
+ /* First get the MSR value that we need. It's safe to futz with
+ * ra/c/dx here since they're modified by this instruction anyway. */
+ info->vm_regs.rcx = TSC_AUX_MSR;
+ ret = v3_handle_msr_read(info);
+
+ if (ret != 0) {
+ return ret;
+ }
-// cycles /= 150;
+ info->vm_regs.rcx = info->vm_regs.rax;
- info->time_state.guest_tsc += cycles;
+ /* Now do the TSC half of the instruction */
+ ret = v3_rdtsc(info);
- list_for_each_entry(tmp_timer, &(info->time_state.timers), timer_link) {
- tmp_timer->ops->update_time(info, cycles, info->time_state.cpu_freq, tmp_timer->private_data);
+ if (ret != 0) {
+ return ret;
}
-
+ return 0;
+}
+
+
+int v3_handle_rdtscp(struct guest_info * info) {
+ PrintDebug("Handling virtual RDTSCP call.\n");
+
+ v3_rdtscp(info);
+
+ info->vm_regs.rax &= 0x00000000ffffffffLL;
+ info->vm_regs.rcx &= 0x00000000ffffffffLL;
+ info->vm_regs.rdx &= 0x00000000ffffffffLL;
+
+ info->rip += 3;
+
+ return 0;
+}
+
+static int tsc_aux_msr_read_hook(struct guest_info *info, uint_t msr_num,
+ struct v3_msr *msr_val, void *priv) {
+ struct vm_core_time * time_state = &(info->time_state);
+
+ V3_ASSERT(msr_num == TSC_AUX_MSR);
+
+ msr_val->lo = time_state->tsc_aux.lo;
+ msr_val->hi = time_state->tsc_aux.hi;
+
+ return 0;
+}
+
+static int tsc_aux_msr_write_hook(struct guest_info *info, uint_t msr_num,
+ struct v3_msr msr_val, void *priv) {
+ struct vm_core_time * time_state = &(info->time_state);
+
+ V3_ASSERT(msr_num == TSC_AUX_MSR);
+
+ time_state->tsc_aux.lo = msr_val.lo;
+ time_state->tsc_aux.hi = msr_val.hi;
+
+ return 0;
+}
+
+static int tsc_msr_read_hook(struct guest_info *info, uint_t msr_num,
+ struct v3_msr *msr_val, void *priv) {
+ uint64_t time = v3_get_guest_tsc(&info->time_state);
+
+ V3_ASSERT(msr_num == TSC_MSR);
+
+ msr_val->hi = time >> 32;
+ msr_val->lo = time & 0xffffffffLL;
+
+ return 0;
+}
+
+static int tsc_msr_write_hook(struct guest_info *info, uint_t msr_num,
+ struct v3_msr msr_val, void *priv) {
+ struct vm_core_time * time_state = &(info->time_state);
+ uint64_t guest_time, new_tsc;
+
+ V3_ASSERT(msr_num == TSC_MSR);
+
+ new_tsc = (((uint64_t)msr_val.hi) << 32) | (uint64_t)msr_val.lo;
+ guest_time = v3_get_guest_time(time_state);
+ time_state->tsc_guest_offset = (sint64_t)(new_tsc - guest_time);
+
+ return 0;
+}
+
+
+int v3_init_time_vm(struct v3_vm_info * vm) {
+ int ret;
+
+ PrintDebug("Installing TSC MSR hook.\n");
+ ret = v3_hook_msr(vm, TSC_MSR,
+ tsc_msr_read_hook, tsc_msr_write_hook, NULL);
+
+ if (ret != 0) {
+ return ret;
+ }
+
+ PrintDebug("Installing TSC_AUX MSR hook.\n");
+ ret = v3_hook_msr(vm, TSC_AUX_MSR, tsc_aux_msr_read_hook,
+ tsc_aux_msr_write_hook, NULL);
+
+ if (ret != 0) {
+ return ret;
+ }
+
+ PrintDebug("Registering TIME_CPUFREQ hypercall.\n");
+ ret = v3_register_hypercall(vm, TIME_CPUFREQ_HCALL,
+ handle_cpufreq_hcall, NULL);
+
+ vm->time_state.td_num = 1;
+ vm->time_state.td_denom = 1;
+ PrintDebug("Setting base time dilation factor to %d/%d.\n",
+ vm->time_state.td_num, vm->time_state.td_denom);
+
+ vm->time_state.follow_host_time = 1;
+ PrintDebug("Locking guest time to host time.\n");
+ return ret;
+}
+
+void v3_deinit_time_vm(struct v3_vm_info * vm) {
+ v3_unhook_msr(vm, TSC_MSR);
+ v3_unhook_msr(vm, TSC_AUX_MSR);
+
+ v3_remove_hypercall(vm, TIME_CPUFREQ_HCALL);
+}
+
+void v3_init_time_core(struct guest_info * info) {
+ struct vm_core_time * time_state = &(info->time_state);
+ v3_cfg_tree_t * cfg_tree = info->core_cfg_data;
+ char * khz = NULL;
+
+ time_state->host_cpu_freq = V3_CPU_KHZ();
+ khz = v3_cfg_val(cfg_tree, "khz");
+
+ if (khz) {
+ time_state->guest_cpu_freq = atoi(khz);
+ PrintDebug("Logical Core %d (vcpu=%d) CPU frequency requested at %d khz.\n",
+ info->pcpu_id, info->vcpu_id, time_state->guest_cpu_freq);
+ }
+
+ if ( (khz == NULL) ||
+ (time_state->guest_cpu_freq <= 0) ||
+ (time_state->guest_cpu_freq > time_state->host_cpu_freq) ) {
+
+ time_state->guest_cpu_freq = time_state->host_cpu_freq;
+ }
+
+ /* Compute these using the GCD() of the guest and host CPU freq.
+ * If the GCD is too small, make it "big enough" */
+ time_state->clock_ratio_num = 1;
+ time_state->clock_ratio_denom = 1;
+
+ PrintDebug("Logical Core %d (vcpu=%d) CPU frequency set to %d KHz (host CPU frequency = %d KHz).\n",
+ info->pcpu_id, info->vcpu_id,
+ time_state->guest_cpu_freq,
+ time_state->host_cpu_freq);
+
+ time_state->guest_cycles = 0;
+ time_state->tsc_guest_offset = 0;
+ time_state->last_update = 0;
+
+ time_state->initial_host_time = 0;
+ time_state->vm_enter_host_time = 0;
+ time_state->vm_pause_host_time = 0;
+
+ INIT_LIST_HEAD(&(time_state->timers));
+ time_state->num_timers = 0;
+
+ time_state->tsc_aux.lo = 0;
+ time_state->tsc_aux.hi = 0;
+}
+
+
+void v3_deinit_time_core(struct guest_info * core) {
+ struct vm_core_time * time_state = &(core->time_state);
+ struct v3_timer * tmr = NULL;
+ struct v3_timer * tmp = NULL;
+
+ list_for_each_entry_safe(tmr, tmp, &(time_state->timers), timer_link) {
+ v3_remove_timer(core, tmr);
+ }
- //info->time_state.pending_cycles = 0;
}