Index: sys/amd64/amd64/efirt_machdep.c =================================================================== --- sys/amd64/amd64/efirt_machdep.c +++ sys/amd64/amd64/efirt_machdep.c @@ -54,6 +54,7 @@ #include #include #include +#include static pml5_entry_t *efi_pml5; static pml4_entry_t *efi_pml4; @@ -64,11 +65,13 @@ void efi_destroy_1t1_map(void) { + struct pctrie_iter pages; vm_page_t m; if (obj_1t1_pt != NULL) { + vm_page_iter_init(&pages, obj_1t1_pt); VM_OBJECT_RLOCK(obj_1t1_pt); - TAILQ_FOREACH(m, &obj_1t1_pt->memq, listq) + VM_RADIX_FOREACH(m, &pages) m->ref_count = VPRC_OBJREF; vm_wire_sub(obj_1t1_pt->resident_page_count); VM_OBJECT_RUNLOCK(obj_1t1_pt); Index: sys/amd64/amd64/pmap.c =================================================================== --- sys/amd64/amd64/pmap.c +++ sys/amd64/amd64/pmap.c @@ -7697,30 +7697,32 @@ pmap_enter_object(pmap_t pmap, vm_offset_t start, vm_offset_t end, vm_page_t m_start, vm_prot_t prot) { + struct pctrie_iter pages; struct rwlock *lock; vm_offset_t va; vm_page_t m, mpte; - vm_pindex_t diff, psize; int rv; VM_OBJECT_ASSERT_LOCKED(m_start->object); - psize = atop(end - start); mpte = NULL; - m = m_start; + vm_page_iter_limit_init(&pages, m_start->object, + m_start->pindex + atop(end - start)); + m = vm_radix_iter_lookup(&pages, m_start->pindex); lock = NULL; PMAP_LOCK(pmap); - while (m != NULL && (diff = m->pindex - m_start->pindex) < psize) { - va = start + ptoa(diff); + while (m != NULL) { + va = start + ptoa(m->pindex - m_start->pindex); if ((va & PDRMASK) == 0 && va + NBPDR <= end && m->psind == 1 && pmap_ps_enabled(pmap) && ((rv = pmap_enter_2mpage(pmap, va, m, prot, &lock)) == KERN_SUCCESS || rv == KERN_NO_SPACE)) - m = &m[NBPDR / PAGE_SIZE - 1]; - else + m = vm_radix_iter_jump(&pages, NBPDR / PAGE_SIZE); + else { mpte = pmap_enter_quick_locked(pmap, va, m, prot, mpte, &lock); - m = TAILQ_NEXT(m, listq); + m = vm_radix_iter_step(&pages); + } } if (lock != NULL) rw_wunlock(lock); @@ -7894,6 +7896,7 @@ pmap_object_init_pt(pmap_t pmap, vm_offset_t addr, vm_object_t object, vm_pindex_t pindex, vm_size_t size) { + struct pctrie_iter pages; pd_entry_t *pde; pt_entry_t PG_A, PG_M, PG_RW, PG_V; vm_paddr_t pa, ptepa; @@ -7913,7 +7916,8 @@ return; if (!vm_object_populate(object, pindex, pindex + atop(size))) return; - p = vm_page_lookup(object, pindex); + vm_page_iter_init(&pages, object); + p = vm_radix_iter_lookup(&pages, pindex); KASSERT(vm_page_all_valid(p), ("pmap_object_init_pt: invalid page %p", p)); pat_mode = p->md.pat_mode; @@ -7931,15 +7935,15 @@ * the pages are not physically contiguous or have differing * memory attributes. */ - p = TAILQ_NEXT(p, listq); for (pa = ptepa + PAGE_SIZE; pa < ptepa + size; pa += PAGE_SIZE) { + p = vm_radix_iter_next(&pages); + if (p == NULL) + return; KASSERT(vm_page_all_valid(p), ("pmap_object_init_pt: invalid page %p", p)); - if (pa != VM_PAGE_TO_PHYS(p) || - pat_mode != p->md.pat_mode) + if (pat_mode != p->md.pat_mode) return; - p = TAILQ_NEXT(p, listq); } /*