/* * Create the resume-time page table. This table maps the image(pig) area, * the kernel text area, and various utility pages for use during resume, * since we cannot overwrite the resuming kernel's page table during inflate * and expect things to work properly. */ void hibernate_populate_resume_pt(union hibernate_info *hib_info, paddr_t image_start, paddr_t image_end) { int phys_page_number, i; paddr_t pa, piglet_start, piglet_end; vaddr_t kern_start_4m_va, kern_end_4m_va, page; /* Identity map PD, PT, and stack pages */ pmap_kenter_pa(HIBERNATE_PT_PAGE, HIBERNATE_PT_PAGE, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_PD_PAGE, HIBERNATE_PD_PAGE, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_STACK_PAGE, HIBERNATE_STACK_PAGE, VM_PROT_ALL); pmap_activate(curproc); bzero((caddr_t)HIBERNATE_PT_PAGE, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PD_PAGE, PAGE_SIZE); bzero((caddr_t)HIBERNATE_STACK_PAGE, PAGE_SIZE); /* PDE for low pages */ hibernate_enter_resume_4k_pde(0); /* * Identity map first 640KB physical for tramps and special utility * pages using 4KB mappings */ for (i = 0; i < 160; i ++) { hibernate_enter_resume_mapping(i*PAGE_SIZE, i*PAGE_SIZE, 0); } /* * Map current kernel VA range using 4M pages */ kern_start_4m_va = (paddr_t)&start & ~(PAGE_MASK_4M); kern_end_4m_va = (paddr_t)&end & ~(PAGE_MASK_4M); /* i386 kernels load at 2MB phys (on the 0th 4mb page) */ phys_page_number = 0; for (page = kern_start_4m_va; page <= kern_end_4m_va; page += NBPD, phys_page_number++) { pa = (paddr_t)(phys_page_number * NBPD); hibernate_enter_resume_mapping(page, pa, 1); } /* * Identity map the image (pig) area */ phys_page_number = image_start / NBPD; image_start &= ~(PAGE_MASK_4M); image_end &= ~(PAGE_MASK_4M); for (page = image_start; page <= image_end ; page += NBPD, phys_page_number++) { pa = (paddr_t)(phys_page_number * NBPD); hibernate_enter_resume_mapping(page, pa, 1); } /* * Map the piglet */ phys_page_number = hib_info->piglet_pa / NBPD; piglet_start = hib_info->piglet_va; piglet_end = piglet_start + HIBERNATE_CHUNK_SIZE * 3; piglet_start &= ~(PAGE_MASK_4M); piglet_end &= ~(PAGE_MASK_4M); for (page = piglet_start; page <= piglet_end ; page += NBPD, phys_page_number++) { pa = (paddr_t)(phys_page_number * NBPD); hibernate_enter_resume_mapping(page, pa, 1); } }
/* * Create the resume-time page table. This table maps the image(pig) area, * the kernel text area, and various utility pages for use during resume, * since we cannot overwrite the resuming kernel's page table during inflate * and expect things to work properly. */ void hibernate_populate_resume_pt(union hibernate_info *hib_info, paddr_t image_start, paddr_t image_end) { int phys_page_number, i; paddr_t pa; vaddr_t kern_start_2m_va, kern_end_2m_va, page; vaddr_t piglet_start_va, piglet_end_va; pt_entry_t *pde, npde; /* Identity map MMU pages */ pmap_kenter_pa(HIBERNATE_PML4T, HIBERNATE_PML4T, PROT_MASK); pmap_kenter_pa(HIBERNATE_PDPT_LOW, HIBERNATE_PDPT_LOW, PROT_MASK); pmap_kenter_pa(HIBERNATE_PDPT_HI, HIBERNATE_PDPT_HI, PROT_MASK); pmap_kenter_pa(HIBERNATE_PD_LOW, HIBERNATE_PD_LOW, PROT_MASK); pmap_kenter_pa(HIBERNATE_PD_LOW2, HIBERNATE_PD_LOW2, PROT_MASK); pmap_kenter_pa(HIBERNATE_PD_HI, HIBERNATE_PD_HI, PROT_MASK); pmap_kenter_pa(HIBERNATE_PT_LOW, HIBERNATE_PT_LOW, PROT_MASK); pmap_kenter_pa(HIBERNATE_PT_LOW2, HIBERNATE_PT_LOW2, PROT_MASK); pmap_kenter_pa(HIBERNATE_PT_HI, HIBERNATE_PT_HI, PROT_MASK); /* Identity map 3 pages for stack */ pmap_kenter_pa(HIBERNATE_STACK_PAGE, HIBERNATE_STACK_PAGE, PROT_MASK); pmap_kenter_pa(HIBERNATE_STACK_PAGE - PAGE_SIZE, HIBERNATE_STACK_PAGE - PAGE_SIZE, PROT_MASK); pmap_kenter_pa(HIBERNATE_STACK_PAGE - 2*PAGE_SIZE, HIBERNATE_STACK_PAGE - 2*PAGE_SIZE, PROT_MASK); pmap_activate(curproc); bzero((caddr_t)HIBERNATE_PML4T, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PDPT_LOW, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PDPT_HI, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PD_LOW, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PD_LOW2, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PD_HI, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PT_LOW, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PT_LOW2, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PT_HI, PAGE_SIZE); bzero((caddr_t)(HIBERNATE_STACK_PAGE - 3*PAGE_SIZE) , 3*PAGE_SIZE); /* First 512GB PML4E */ pde = (pt_entry_t *)(HIBERNATE_PML4T + (pl4_pi(0) * sizeof(pt_entry_t))); npde = (HIBERNATE_PDPT_LOW) | PG_RW | PG_V; *pde = npde; /* First 1GB PDPTE */ pde = (pt_entry_t *)(HIBERNATE_PDPT_LOW + (pl3_pi(0) * sizeof(pt_entry_t))); npde = (HIBERNATE_PD_LOW) | PG_RW | PG_V; *pde = npde; /* PD for first 2MB */ pde = (pt_entry_t *)(HIBERNATE_PD_LOW + (pl2_pi(0) * sizeof(pt_entry_t))); npde = (HIBERNATE_PT_LOW) | PG_RW | PG_V; *pde = npde; /* * Identity map low physical pages. * See arch/amd64/include/hibernate_var.h for page ranges used here. */ for (i = ACPI_TRAMPOLINE; i <= HIBERNATE_HIBALLOC_PAGE; i += PAGE_SIZE) hibernate_enter_resume_mapping(i, i, 0); /* * Map current kernel VA range using 2MB pages */ kern_start_2m_va = (paddr_t)&start & ~(PAGE_MASK_L2); kern_end_2m_va = (paddr_t)&end & ~(PAGE_MASK_L2); /* amd64 kernels load at 16MB phys (on the 8th 2mb page) */ phys_page_number = 8; for (page = kern_start_2m_va; page <= kern_end_2m_va; page += NBPD_L2, phys_page_number++) { pa = (paddr_t)(phys_page_number * NBPD_L2); hibernate_enter_resume_mapping(page, pa, 1); } /* * Identity map the piglet using 2MB pages. */ phys_page_number = hib_info->piglet_pa / NBPD_L2; /* VA == PA */ piglet_start_va = hib_info->piglet_pa; piglet_end_va = piglet_start_va + HIBERNATE_CHUNK_SIZE * 4; for (page = piglet_start_va; page <= piglet_end_va; page += NBPD_L2, phys_page_number++) { pa = (paddr_t)(phys_page_number * NBPD_L2); hibernate_enter_resume_mapping(page, pa, 1); } /* Unmap MMU pages (stack remains mapped) */ pmap_kremove(HIBERNATE_PML4T, PAGE_SIZE); pmap_kremove(HIBERNATE_PDPT_LOW, PAGE_SIZE); pmap_kremove(HIBERNATE_PDPT_HI, PAGE_SIZE); pmap_kremove(HIBERNATE_PD_LOW, PAGE_SIZE); pmap_kremove(HIBERNATE_PD_LOW2, PAGE_SIZE); pmap_kremove(HIBERNATE_PD_HI, PAGE_SIZE); pmap_kremove(HIBERNATE_PT_LOW, PAGE_SIZE); pmap_kremove(HIBERNATE_PT_LOW2, PAGE_SIZE); pmap_kremove(HIBERNATE_PT_HI, PAGE_SIZE); pmap_activate(curproc); }
/* * Create the resume-time page table. This table maps the image(pig) area, * the kernel text area, and various utility pages for use during resume, * since we cannot overwrite the resuming kernel's page table during inflate * and expect things to work properly. */ void hibernate_populate_resume_pt(union hibernate_info *hib_info, paddr_t image_start, paddr_t image_end) { int phys_page_number, i; paddr_t pa, piglet_start, piglet_end; vaddr_t kern_start_2m_va, kern_end_2m_va, page; pt_entry_t *pde, npde; /* Identity map MMU pages */ pmap_kenter_pa(HIBERNATE_PML4T, HIBERNATE_PML4T, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_PDPT_LOW, HIBERNATE_PDPT_LOW, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_PDPT_HI, HIBERNATE_PDPT_HI, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_PD_LOW, HIBERNATE_PD_LOW, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_PD_LOW2, HIBERNATE_PD_LOW2, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_PD_HI, HIBERNATE_PD_HI, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_PT_LOW, HIBERNATE_PT_LOW, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_PT_LOW2, HIBERNATE_PT_LOW2, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_PT_HI, HIBERNATE_PT_HI, VM_PROT_ALL); /* Identity map 3 pages for stack */ pmap_kenter_pa(HIBERNATE_STACK_PAGE, HIBERNATE_STACK_PAGE, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_STACK_PAGE - PAGE_SIZE, HIBERNATE_STACK_PAGE - PAGE_SIZE, VM_PROT_ALL); pmap_kenter_pa(HIBERNATE_STACK_PAGE - 2*PAGE_SIZE, HIBERNATE_STACK_PAGE - 2*PAGE_SIZE, VM_PROT_ALL); pmap_activate(curproc); bzero((caddr_t)HIBERNATE_PML4T, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PDPT_LOW, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PDPT_HI, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PD_LOW, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PD_LOW2, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PD_HI, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PT_LOW, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PT_LOW2, PAGE_SIZE); bzero((caddr_t)HIBERNATE_PT_HI, PAGE_SIZE); bzero((caddr_t)(HIBERNATE_STACK_PAGE - 3*PAGE_SIZE) , 3*PAGE_SIZE); /* First 512GB PML4E */ pde = (pt_entry_t *)(HIBERNATE_PML4T + (pl4_pi(0) * sizeof(pt_entry_t))); npde = (HIBERNATE_PDPT_LOW) | PG_RW | PG_V; *pde = npde; /* First 1GB PDPTE */ pde = (pt_entry_t *)(HIBERNATE_PDPT_LOW + (pl3_pi(0) * sizeof(pt_entry_t))); npde = (HIBERNATE_PD_LOW) | PG_RW | PG_V; *pde = npde; /* PD for first 2MB */ pde = (pt_entry_t *)(HIBERNATE_PD_LOW + (pl2_pi(0) * sizeof(pt_entry_t))); npde = (HIBERNATE_PT_LOW) | PG_RW | PG_V; *pde = npde; /* * Identity map first 640KB physical for tramps and special utility * pages using 4KB mappings */ for (i = 0; i < 160; i ++) { hibernate_enter_resume_mapping(i*PAGE_SIZE, i*PAGE_SIZE, 0); } /* * Map current kernel VA range using 2MB pages */ kern_start_2m_va = (paddr_t)&start & ~(PAGE_MASK_2M); kern_end_2m_va = (paddr_t)&end & ~(PAGE_MASK_2M); /* amd64 kernels load at 16MB phys (on the 8th 2mb page) */ phys_page_number = 8; for (page = kern_start_2m_va; page <= kern_end_2m_va; page += NBPD_L2, phys_page_number++) { pa = (paddr_t)(phys_page_number * NBPD_L2); hibernate_enter_resume_mapping(page, pa, 1); } /* * Map the piglet */ phys_page_number = hib_info->piglet_pa / NBPD_L2; piglet_start = hib_info->piglet_va; piglet_end = piglet_start + HIBERNATE_CHUNK_SIZE * 3; piglet_start &= ~(PAGE_MASK_2M); piglet_end &= ~(PAGE_MASK_2M); for (page = piglet_start; page <= piglet_end ; page += NBPD_L2, phys_page_number++) { pa = (paddr_t)(phys_page_number * NBPD_L2); hibernate_enter_resume_mapping(page, pa, 1); } }