X-Git-Url: http://v3vee.org/palacios/gitweb/gitweb.cgi?a=blobdiff_plain;f=palacios%2Finclude%2Fpalacios%2Fvmm_direct_paging.h;h=fc9d924942ef755c76cacf60929205962a2c1fd7;hb=9d346ee2cf8a7f2a79dc4929b8e0416de7d8f8c8;hp=22843fd767e8649acec1dbdc6a8f42d3ccaf9545;hpb=1fe82881720f7f9f64f789871f763aca93b47a7e;p=palacios.git diff --git a/palacios/include/palacios/vmm_direct_paging.h b/palacios/include/palacios/vmm_direct_paging.h index 22843fd..fc9d924 100644 --- a/palacios/include/palacios/vmm_direct_paging.h +++ b/palacios/include/palacios/vmm_direct_paging.h @@ -25,11 +25,144 @@ #include #include +#include +#include + -addr_t v3_create_direct_passthrough_pts(struct guest_info * guest_info); +/********************************** + PASSTHROUGH PAGING - CORE FUNC + **********************************/ + + +struct v3_passthrough_impl_state { + // currently there is only a single implementation + // that internally includes SVM and VMX support + // The externally visible state is just the callbacks + v3_rw_lock_t event_callback_lock; + struct list_head event_callback_list; + int inited; +}; + + +int v3_init_passthrough_paging(struct v3_vm_info *vm); +int v3_init_passthrough_paging_core(struct guest_info *core); +int v3_deinit_passthrough_paging(struct v3_vm_info *vm); +int v3_deinit_passthrough_paging_core(struct guest_info *core); + +int v3_init_passthrough_pts(struct guest_info * guest_info); +int v3_free_passthrough_pts(struct guest_info * core); + +int v3_reset_passthrough_pts(struct guest_info * guest_info); + +// actual_start/end may be null if you don't want this info +// If non-null, these return the actual affected GPA range +int v3_handle_passthrough_pagefault(struct guest_info * info, addr_t fault_addr, pf_error_t error_code, + addr_t *actual_start, addr_t *actual_end); + +int v3_activate_passthrough_pt(struct guest_info * info); + +int v3_invalidate_passthrough_addr(struct guest_info * info, addr_t inv_addr, + addr_t *actual_start, addr_t *actual_end); + +// The range invalidated is minimally [start, end] +int v3_invalidate_passthrough_addr_range(struct guest_info * info, + addr_t inv_addr_start, addr_t inv_addr_end, + addr_t *actual_start, addr_t *actual_end); + +/********************************** + PASSTHROUGH PAGING - EVENTS + **********************************/ + +struct v3_passthrough_pg_event { + enum {PASSTHROUGH_PAGEFAULT,PASSTHROUGH_INVALIDATE_RANGE,PASSTHROUGH_ACTIVATE} event_type; + enum {PASSTHROUGH_PREIMPL, PASSTHROUGH_POSTIMPL} event_order; + addr_t gpa; // for pf + pf_error_t error_code; // for pf + addr_t gpa_start; // for invalidation of range or page fault + addr_t gpa_end; // for invalidation of range or page fault (range is [start,end] ) + // PREIMPL: start/end is the requested range + // POSTIMPL: start/end is the actual range invalidated +}; + + + +int v3_register_passthrough_paging_event_callback(struct v3_vm_info *vm, + int (*callback)(struct guest_info *core, + struct v3_passthrough_pg_event *, + void *priv_data), + void *priv_data); + +int v3_unregister_passthrough_paging_event_callback(struct v3_vm_info *vm, + int (*callback)(struct guest_info *core, + struct v3_passthrough_pg_event *, + void *priv_data), + void *priv_data); + + + +/***************************** + NESTED PAGING - CORE FUNC + *****************************/ + + +struct v3_nested_impl_state { + // currently there is only a single implementation + // that internally includes SVM and VMX support + // The externally visible state is just the callbacks + v3_rw_lock_t event_callback_lock; + struct list_head event_callback_list; +}; + +int v3_init_nested_paging(struct v3_vm_info *vm); +int v3_init_nested_paging_core(struct guest_info *core, void *hwinfo); +int v3_deinit_nested_paging(struct v3_vm_info *vm); +int v3_deinit_nested_paging_core(struct guest_info *core); + + +// actual_start/end may be null if you don't want this info +// If non-null, these return the actual affected GPA range +int v3_handle_nested_pagefault(struct guest_info * info, addr_t fault_addr, void *pfinfo, + addr_t *actual_start, addr_t *actual_end); + +int v3_invalidate_nested_addr(struct guest_info * info, addr_t inv_addr, + addr_t *actual_start, addr_t *actual_end); + +// The range invalidated is minimally [start, end] +int v3_invalidate_nested_addr_range(struct guest_info * info, + addr_t inv_addr_start, addr_t inv_addr_end, + addr_t *actual_start, addr_t *actual_end); + + + +/***************************** + NESTED PAGING - EVENTS + *****************************/ + +struct v3_nested_pg_event { + enum {NESTED_PAGEFAULT,NESTED_INVALIDATE_RANGE} event_type; + enum {NESTED_PREIMPL, NESTED_POSTIMPL} event_order; + addr_t gpa; // for pf + pf_error_t error_code; // for pf + addr_t gpa_start; // for invalidation of range or page fault + addr_t gpa_end; // for invalidation of range or page fault (range is [start,end] ) + // PREIMPL: start/end is the requested range + // POSTIMPL: start/end is the actual range invalidated +}; + + + +int v3_register_nested_paging_event_callback(struct v3_vm_info *vm, + int (*callback)(struct guest_info *core, + struct v3_nested_pg_event *, + void *priv_data), + void *priv_data); + +int v3_unregister_nested_paging_event_callback(struct v3_vm_info *vm, + int (*callback)(struct guest_info *core, + struct v3_nested_pg_event *, + void *priv_data), + void *priv_data); -int v3_handle_passthrough_pagefault(struct guest_info * info, addr_t fault_addr, pf_error_t error_code); -int v3_handle_nested_pagefault(struct guest_info * info, addr_t fault_addr, pf_error_t error_code); #endif // ! __V3VEE__