Palacios Public Git Repository

To checkout Palacios execute

  git clone http://v3vee.org/palacios/palacios.web/palacios.git
This will give you the master branch. You probably want the devel branch or one of the release branches. To switch to the devel branch, simply execute
  cd palacios
  git checkout --track -b devel origin/devel
The other branches are similar.


various fixes. Hopefully this fixes the transient shutdown bug...
[palacios.git] / palacios / src / palacios / vmm_shadow_paging_32.h
index d60eb9f..bc83a7c 100644 (file)
@@ -27,7 +27,7 @@ static inline int activate_shadow_pt_32(struct guest_info * info) {
 
     shdw_page->cr3 = shdw_page->page_pa;
     
-    shadow_cr3->pdt_base_addr = PAGE_BASE_ADDR(shdw_page->page_pa);
+    shadow_cr3->pdt_base_addr = PAGE_BASE_ADDR_4KB(shdw_page->page_pa);
     PrintDebug( "Created new shadow page table %p\n", (void *)BASE_TO_PAGE_ADDR(shadow_cr3->pdt_base_addr));
   
     shadow_cr3->pwt = guest_cr3->pwt;
@@ -43,15 +43,11 @@ static inline int activate_shadow_pt_32(struct guest_info * info) {
  * *
  * *
  */
-static int handle_large_pagefault_32(struct guest_info * info, 
-                                    addr_t fault_addr, pf_error_t error_code, 
-                                    pte32_t * shadow_pt, pde32_4MB_t * large_guest_pde);
+static int handle_4MB_shadow_pagefault_32(struct guest_info * info,  addr_t fault_addr, pf_error_t error_code, 
+                                         pte32_t * shadow_pt, pde32_4MB_t * large_guest_pde);
 
-static int handle_shadow_pte32_fault(struct guest_info * info, 
-                                    addr_t fault_addr, 
-                                    pf_error_t error_code,
-                                    pte32_t * shadow_pt, 
-                                    pte32_t * guest_pt);
+static int handle_pte_shadow_pagefault_32(struct guest_info * info, addr_t fault_addr, pf_error_t error_code,
+                                         pte32_t * shadow_pt,  pte32_t * guest_pt);
 
 
 static inline int handle_shadow_pagefault_32(struct guest_info * info, addr_t fault_addr, pf_error_t error_code) {
@@ -64,6 +60,7 @@ static inline int handle_shadow_pagefault_32(struct guest_info * info, addr_t fa
     pde32_t * shadow_pde = (pde32_t *)&(shadow_pd[PDE32_INDEX(fault_addr)]);
 
     PrintDebug("Shadow page fault handler: %p\n", (void*) fault_addr );
+    PrintDebug("Handling PDE32 Fault\n");
 
     if (guest_pa_to_host_va(info, guest_cr3, (addr_t*)&guest_pd) == -1) {
        PrintError("Invalid Guest PDE Address: 0x%p\n",  (void *)guest_cr3);
@@ -83,235 +80,119 @@ static inline int handle_shadow_pagefault_32(struct guest_info * info, addr_t fa
     if (is_guest_pf(guest_pde_access, shadow_pde_access) == 1) {
        PrintDebug("Injecting PDE pf to guest: (guest access error=%d) (pf error code=%d)\n", 
                   *(uint_t *)&guest_pde_access, *(uint_t *)&error_code);
-       inject_guest_pf(info, fault_addr, error_code);
+       if (inject_guest_pf(info, fault_addr, error_code) == -1) {
+           PrintError("Could not inject guest page fault\n");
+           return -1;
+       }
        return 0;
     }
 
-  
-    if (shadow_pde_access == PT_ACCESS_NOT_PRESENT) 
-       {
-           struct shadow_page_data * shdw_page =  create_new_shadow_pt(info);
-           pte32_t * shadow_pt = (pte32_t *)V3_VAddr((void *)shdw_page->page_pa);
-
-           shadow_pde->present = 1;
-           shadow_pde->user_page = guest_pde->user_page;
-           //    shadow_pde->large_page = guest_pde->large_page;
-           shadow_pde->large_page = 0;
-      
-
-           // VMM Specific options
-           shadow_pde->write_through = guest_pde->write_through;
-           shadow_pde->cache_disable = guest_pde->cache_disable;
-           shadow_pde->global_page = guest_pde->global_page;
-           //
-      
-           guest_pde->accessed = 1;
-      
-           shadow_pde->pt_base_addr = PAGE_BASE_ADDR((addr_t)V3_PAddr(shadow_pt));
-      
-           if (guest_pde->large_page == 0) {
-               pte32_t * guest_pt = NULL;
-               shadow_pde->writable = guest_pde->writable;
-
-               if (guest_pa_to_host_va(info, BASE_TO_PAGE_ADDR(guest_pde->pt_base_addr), (addr_t*)&guest_pt) == -1) {
-                   // Machine check the guest
-                   PrintDebug("Invalid Guest PTE Address: 0x%p\n", (void *)BASE_TO_PAGE_ADDR(guest_pde->pt_base_addr));
-                   v3_raise_exception(info, MC_EXCEPTION);
-                   return 0;
-               }
-
-               if (handle_shadow_pte32_fault(info, fault_addr, error_code, shadow_pt, guest_pt)  == -1) {
-                   PrintError("Error handling Page fault caused by PTE\n");
-                   return -1;
-               }
-           } else {
-               // ??  What if guest pde is dirty a this point?
-               ((pde32_4MB_t *)guest_pde)->dirty = 0;
-               shadow_pde->writable = 0;
-
-               if (handle_large_pagefault_32(info, fault_addr, error_code, shadow_pt, (pde32_4MB_t *)guest_pde) == -1) {
-                   PrintError("Error handling large pagefault\n");
-                   return -1;
-               }       
 
-           }
-       }
-    else if (shadow_pde_access == PT_ACCESS_OK) 
-       {
-           //
-           // PTE fault
-           //
-           pte32_t * shadow_pt = (pte32_t *)V3_VAddr( (void*)(addr_t) BASE_TO_PAGE_ADDR(shadow_pde->pt_base_addr) );
-
-           if (guest_pde->large_page == 0) {
-               pte32_t * guest_pt = NULL;
 
-               if (guest_pa_to_host_va(info, BASE_TO_PAGE_ADDR(guest_pde->pt_base_addr), (addr_t*)&guest_pt) == -1) {
-                   // Machine check the guest
-                   PrintDebug("Invalid Guest PTE Address: 0x%p\n", (void *)BASE_TO_PAGE_ADDR(guest_pde->pt_base_addr));
-                   v3_raise_exception(info, MC_EXCEPTION);
-                   return 0;
-               }
+    if (shadow_pde_access == PT_ACCESS_USER_ERROR) {
+       // 
+       // PDE Entry marked non user
+       //
+       PrintDebug("Shadow Paging User access error (shadow_pde_access=0x%x, guest_pde_access=0x%x)\n", 
+                  shadow_pde_access, guest_pde_access);
        
-               if (handle_shadow_pte32_fault(info, fault_addr, error_code, shadow_pt, guest_pt)  == -1) {
-                   PrintError("Error handling Page fault caused by PTE\n");
-                   return -1;
-               }
-           } else if (guest_pde->large_page == 1) {
-               if (handle_large_pagefault_32(info, fault_addr, error_code, shadow_pt, (pde32_4MB_t *)guest_pde) == -1) {
-                   PrintError("Error handling large pagefault\n");
-                   return -1;
-               }
-           }
+       if (inject_guest_pf(info, fault_addr, error_code) == -1) {
+           PrintError("Could not inject guest page fault\n");
+           return -1;
        }
-    else if ((shadow_pde_access == PT_ACCESS_WRITE_ERROR) && 
-            (guest_pde->large_page == 1)) 
-       {
-           //
-           // Page Directory Entry marked read-only
-           // Its a large page and we need to update the dirty bit in the guest
-           //
-
-           PrintDebug("Large page write error... Setting dirty bit and returning\n");
-           ((pde32_4MB_t *)guest_pde)->dirty = 1;
-           shadow_pde->writable = guest_pde->writable;
-           return 0;
-      
-       } 
-    else if (shadow_pde_access == PT_ACCESS_USER_ERROR) 
-       {
-           //
-           // Page Directory Entry marked non-user
-           //      
-           PrintDebug("Shadow Paging User access error (shadow_pde_access=0x%x, guest_pde_access=0x%x)\n", 
-                      shadow_pde_access, guest_pde_access);
-           inject_guest_pf(info, fault_addr, error_code);
-           return 0;
-       }
-    else 
-       {
-           // inject page fault in guest
-           inject_guest_pf(info, fault_addr, error_code);
-           PrintDebug("Unknown Error occurred (shadow_pde_access=%d)\n", shadow_pde_access);
-           PrintDebug("Manual Says to inject page fault into guest\n");
-#ifdef DEBUG_SHADOW_PAGING
-           PrintDebug("Guest PDE: (access=%d)\n\t", guest_pde_access);
-           PrintPTEntry(info, PAGE_PD32, fault_addr, guest_pde);
-           PrintDebug("Shadow PDE: (access=%d)\n\t", shadow_pde_access);
-           PrintPTEntry(info, PAGE_PD32, fault_addr, shadow_pde);
-#endif
-
-           return 0; 
+       return 0;
+    } else if ((shadow_pde_access == PT_ACCESS_WRITE_ERROR) && 
+              (guest_pde->large_page == 1)) {
+       
+       ((pde32_4MB_t *)guest_pde)->dirty = 1;
+       shadow_pde->writable = guest_pde->writable;
+       return 0;
+    } else if ((shadow_pde_access != PT_ACCESS_NOT_PRESENT) &&
+              (shadow_pde_access != PT_ACCESS_OK)) {
+       // inject page fault in guest
+       if (inject_guest_pf(info, fault_addr, error_code) == -1) {
+           PrintError("Could not inject guest page fault\n");
+           return -1;
        }
-
-    PrintDebug("Returning end of PDE function (rip=%p)\n", (void *)(addr_t)(info->rip));
-    return 0;
-}
-
-
-
-/* The guest status checks have already been done,
- * only special case shadow checks remain
- */
-static int handle_large_pagefault_32(struct guest_info * info, 
-                                    addr_t fault_addr, pf_error_t error_code, 
-                                    pte32_t * shadow_pt, pde32_4MB_t * large_guest_pde) 
-{
-    pt_access_status_t shadow_pte_access = v3_can_access_pte32(shadow_pt, fault_addr, error_code);
-    pte32_t * shadow_pte = (pte32_t *)&(shadow_pt[PTE32_INDEX(fault_addr)]);
-    addr_t guest_fault_pa = BASE_TO_PAGE_ADDR_4MB(large_guest_pde->page_base_addr) + PAGE_OFFSET_4MB(fault_addr);  
-
-    struct v3_shadow_region * shdw_reg = v3_get_shadow_region(info, guest_fault_pa);
-
-    if (shdw_reg == NULL) {
-       // Inject a machine check in the guest
-       PrintDebug("Invalid Guest Address in page table (0x%p)\n", (void *)guest_fault_pa);
-       v3_raise_exception(info, MC_EXCEPTION);
-       return -1;
-    }
-
-    if (shadow_pte_access == PT_ACCESS_OK) {
-       // Inconsistent state...
-       // Guest Re-Entry will flush tables and everything should now workd
-       PrintDebug("Inconsistent state... Guest re-entry should flush tlb\n");
+       PrintDebug("Unknown Error occurred (shadow_pde_access=%d)\n", shadow_pde_access);
+       PrintDebug("Manual Says to inject page fault into guest\n");
        return 0;
     }
 
   
-    if (shadow_pte_access == PT_ACCESS_NOT_PRESENT) {
-       // Get the guest physical address of the fault
+    pte32_t * shadow_pt = NULL;
+    pte32_t * guest_pt = NULL;
 
-       if ((shdw_reg->host_type == SHDW_REGION_ALLOCATED) || 
-           (shdw_reg->host_type == SHDW_REGION_WRITE_HOOK)) {
-           addr_t shadow_pa = v3_get_shadow_addr(shdw_reg, guest_fault_pa);
+    // Get the next shadow page level, allocate if not present
 
-           shadow_pte->page_base_addr = PAGE_BASE_ADDR(shadow_pa);
-
-           shadow_pte->present = 1;
+    if (shadow_pde_access == PT_ACCESS_NOT_PRESENT) {
+       struct shadow_page_data * shdw_page =  create_new_shadow_pt(info);
+       shadow_pt = (pte32_t *)V3_VAddr((void *)shdw_page->page_pa);
 
-           /* We are assuming that the PDE entry has precedence
-            * so the Shadow PDE will mirror the guest PDE settings, 
-            * and we don't have to worry about them here
-            * Allow everything
-            */
-           shadow_pte->user_page = 1;
+       shadow_pde->present = 1;
+       shadow_pde->user_page = guest_pde->user_page;
 
-           if (shdw_reg->host_type == SHDW_REGION_WRITE_HOOK) {
-               shadow_pte->writable = 0;
-           } else {
-               shadow_pte->writable = 1;
-           }
 
-           //set according to VMM policy
-           shadow_pte->write_through = large_guest_pde->write_through;
-           shadow_pte->cache_disable = large_guest_pde->cache_disable;
-           shadow_pte->global_page = large_guest_pde->global_page;
-           //
-      
+       if (guest_pde->large_page == 0) {
+           shadow_pde->writable = guest_pde->writable;
        } else {
-           // Handle hooked pages as well as other special pages
-           //      if (handle_special_page_fault(info, fault_addr, guest_fault_pa, error_code) == -1) {
+           // This large page flag is temporary until we can get a working cache....
+           ((pde32_4MB_t *)guest_pde)->vmm_info = V3_LARGE_PG;
 
-           if (v3_handle_mem_full_hook(info, fault_addr, guest_fault_pa, shdw_reg, error_code) == -1) {
-               PrintError("Special Page Fault handler returned error for address: %p\n", (void *)fault_addr);
-               return -1;
+           if (error_code.write) {
+               shadow_pde->writable = guest_pde->writable;
+               ((pde32_4MB_t *)guest_pde)->dirty = 1;
+           } else {
+               shadow_pde->writable = 0;
+               ((pde32_4MB_t *)guest_pde)->dirty = 0;
            }
        }
-    } else if (shadow_pte_access == PT_ACCESS_WRITE_ERROR) {
+      
 
-       if (shdw_reg->host_type == SHDW_REGION_WRITE_HOOK) {
+       // VMM Specific options
+       shadow_pde->write_through = guest_pde->write_through;
+       shadow_pde->cache_disable = guest_pde->cache_disable;
+       shadow_pde->global_page = guest_pde->global_page;
+       //
+      
+       guest_pde->accessed = 1;
+      
 
-           if (v3_handle_mem_wr_hook(info, fault_addr, guest_fault_pa, shdw_reg, error_code) == -1) {
-               PrintError("Special Page Fault handler returned error for address: %p\n", (void *)fault_addr);
-               return -1;
-           }
-       }
 
 
+       shadow_pde->pt_base_addr = PAGE_BASE_ADDR(shdw_page->page_pa);
+    } else {
+       shadow_pt = (pte32_t *)V3_VAddr((void *)BASE_TO_PAGE_ADDR(shadow_pde->pt_base_addr));
+    }
+
 
+      
+    if (guest_pde->large_page == 0) {
+       if (guest_pa_to_host_va(info, BASE_TO_PAGE_ADDR(guest_pde->pt_base_addr), (addr_t*)&guest_pt) == -1) {
+           // Machine check the guest
+           PrintDebug("Invalid Guest PTE Address: 0x%p\n", (void *)BASE_TO_PAGE_ADDR(guest_pde->pt_base_addr));
+           v3_raise_exception(info, MC_EXCEPTION);
+           return 0;
+       }
 
+       if (handle_pte_shadow_pagefault_32(info, fault_addr, error_code, shadow_pt, guest_pt)  == -1) {
+           PrintError("Error handling Page fault caused by PTE\n");
+           return -1;
+       }
     } else {
-       PrintError("Error in large page fault handler...\n");
-       PrintError("This case should have been handled at the top level handler\n");
-       return -1;
+       if (handle_4MB_shadow_pagefault_32(info, fault_addr, error_code, shadow_pt, (pde32_4MB_t *)guest_pde) == -1) {
+           PrintError("Error handling large pagefault\n");
+           return -1;
+       }       
     }
 
-    PrintDebug("Returning from large page fault handler\n");
     return 0;
 }
 
 
 
-
-/* 
- * We assume the the guest pte pointer has already been translated to a host virtual address
- */
-static int handle_shadow_pte32_fault(struct guest_info * info, 
-                                    addr_t fault_addr, 
-                                    pf_error_t error_code,
-                                    pte32_t * shadow_pt, 
-                                    pte32_t * guest_pt) {
+static int handle_pte_shadow_pagefault_32(struct guest_info * info, addr_t fault_addr, pf_error_t error_code,
+                                         pte32_t * shadow_pt, pte32_t * guest_pt) {
 
     pt_access_status_t guest_pte_access;
     pt_access_status_t shadow_pte_access;
@@ -334,18 +215,18 @@ static int handle_shadow_pte32_fault(struct guest_info * info,
     // Check the shadow page permissions
     shadow_pte_access = v3_can_access_pte32(shadow_pt, fault_addr, error_code);
   
-#ifdef DEBUG_SHADOW_PAGING
-    PrintDebug("Guest PTE: (access=%d)\n\t", guest_pte_access);
-    PrintPTEntry(info, PAGE_PT32, fault_addr, guest_pte);
-    PrintDebug("Shadow PTE: (access=%d)\n\t", shadow_pte_access);
-    PrintPTEntry(info, PAGE_PT32, fault_addr, shadow_pte);
-#endif
   
     /* Was the page fault caused by the Guest's page tables? */
     if (is_guest_pf(guest_pte_access, shadow_pte_access) == 1) {
+
        PrintDebug("Access error injecting pf to guest (guest access error=%d) (pf error code=%d)\n", 
-                  guest_pte_access, *(uint_t*)&error_code);    
-       inject_guest_pf(info, fault_addr, error_code);
+                  guest_pte_access, *(uint_t*)&error_code);
+
+       if (inject_guest_pf(info, fault_addr, error_code) == -1) {
+           PrintError("Could not inject guest page fault\n");
+           return -1;
+       }
+
        return 0; 
     }
 
@@ -368,6 +249,8 @@ static int handle_shadow_pte32_fault(struct guest_info * info,
            addr_t shadow_pa = v3_get_shadow_addr(shdw_reg, guest_pa);
       
            shadow_pte->page_base_addr = PAGE_BASE_ADDR(shadow_pa);
+
+           PrintDebug("\tMapping shadow page (%p)\n", (void *)BASE_TO_PAGE_ADDR(shadow_pte->page_base_addr));
       
            shadow_pte->present = guest_pte->present;
            shadow_pte->user_page = guest_pte->user_page;
@@ -422,18 +305,120 @@ static int handle_shadow_pte32_fault(struct guest_info * info,
 
     } else {
        // Inject page fault into the guest     
-       inject_guest_pf(info, fault_addr, error_code);
+       if (inject_guest_pf(info, fault_addr, error_code) == -1) {
+           PrintError("Could not inject guest page fault\n");
+           return -1;
+       }
+
        PrintError("PTE Page fault fell through... Not sure if this should ever happen\n");
        PrintError("Manual Says to inject page fault into guest\n");
        return -1;
     }
 
-    PrintDebug("Returning end of function\n");
     return 0;
 }
 
 
 
+static int handle_4MB_shadow_pagefault_32(struct guest_info * info, 
+                                    addr_t fault_addr, pf_error_t error_code, 
+                                    pte32_t * shadow_pt, pde32_4MB_t * large_guest_pde) 
+{
+    pt_access_status_t shadow_pte_access = v3_can_access_pte32(shadow_pt, fault_addr, error_code);
+    pte32_t * shadow_pte = (pte32_t *)&(shadow_pt[PTE32_INDEX(fault_addr)]);
+    addr_t guest_fault_pa = BASE_TO_PAGE_ADDR_4MB(large_guest_pde->page_base_addr) + PAGE_OFFSET_4MB(fault_addr);  
+
+
+    PrintDebug("Handling 4MB fault (guest_fault_pa=%p) (error_code=%x)\n", (void *)guest_fault_pa, *(uint_t*)&error_code);
+    PrintDebug("ShadowPT=%p, LargeGuestPDE=%p\n", shadow_pt, large_guest_pde);
+
+    struct v3_shadow_region * shdw_reg = v3_get_shadow_region(info, guest_fault_pa);
+
+    if (shdw_reg == NULL) {
+       // Inject a machine check in the guest
+       PrintDebug("Invalid Guest Address in page table (0x%p)\n", (void *)guest_fault_pa);
+       v3_raise_exception(info, MC_EXCEPTION);
+       return -1;
+    }
+
+    if (shadow_pte_access == PT_ACCESS_OK) {
+       // Inconsistent state...
+       // Guest Re-Entry will flush tables and everything should now workd
+       PrintDebug("Inconsistent state... Guest re-entry should flush tlb\n");
+       return 0;
+    }
+
+  
+    if (shadow_pte_access == PT_ACCESS_NOT_PRESENT) {
+       // Get the guest physical address of the fault
+
+       if ((shdw_reg->host_type == SHDW_REGION_ALLOCATED) || 
+           (shdw_reg->host_type == SHDW_REGION_WRITE_HOOK)) {
+           addr_t shadow_pa = v3_get_shadow_addr(shdw_reg, guest_fault_pa);
+
+           shadow_pte->page_base_addr = PAGE_BASE_ADDR(shadow_pa);
+
+           PrintDebug("\tMapping shadow page (%p)\n", (void *)BASE_TO_PAGE_ADDR(shadow_pte->page_base_addr));
+
+           shadow_pte->present = 1;
+
+           /* We are assuming that the PDE entry has precedence
+            * so the Shadow PDE will mirror the guest PDE settings, 
+            * and we don't have to worry about them here
+            * Allow everything
+            */
+           shadow_pte->user_page = 1;
+
+           if (shdw_reg->host_type == SHDW_REGION_WRITE_HOOK) {
+               shadow_pte->writable = 0;
+           } else {
+               shadow_pte->writable = 1;
+           }
+
+           //set according to VMM policy
+           shadow_pte->write_through = large_guest_pde->write_through;
+           shadow_pte->cache_disable = large_guest_pde->cache_disable;
+           shadow_pte->global_page = large_guest_pde->global_page;
+           //
+      
+       } else {
+           if (v3_handle_mem_full_hook(info, fault_addr, guest_fault_pa, shdw_reg, error_code) == -1) {
+               PrintError("Special Page Fault handler returned error for address: %p\n", (void *)fault_addr);
+               return -1;
+           }
+       }
+    } else if (shadow_pte_access == PT_ACCESS_WRITE_ERROR) {
+
+       if (shdw_reg->host_type == SHDW_REGION_WRITE_HOOK) {
+
+           if (v3_handle_mem_wr_hook(info, fault_addr, guest_fault_pa, shdw_reg, error_code) == -1) {
+               PrintError("Special Page Fault handler returned error for address: %p\n", (void *)fault_addr);
+               return -1;
+           }
+       }
+
+    } else {
+       PrintError("Error in large page fault handler...\n");
+       PrintError("This case should have been handled at the top level handler\n");
+       return -1;
+    }
+
+    PrintDebug("Returning from large page fault handler\n");
+    return 0;
+}
+
+
+
+
+
+
+
+
+
+
+
+
 /* If we start to optimize we should look up the guest pages in the cache... */
 static inline int handle_shadow_invlpg_32(struct guest_info * info, addr_t vaddr) {
     pde32_t * shadow_pd = (pde32_t *)CR3_TO_PDE32_VA(info->ctrl_regs.cr3);