Index: sys/vm/vm_page.h =================================================================== --- sys/vm/vm_page.h +++ sys/vm/vm_page.h @@ -566,7 +566,6 @@ void vm_page_deactivate_noreuse(vm_page_t); void vm_page_dequeue(vm_page_t m); void vm_page_dequeue_deferred(vm_page_t m); -void vm_page_drain_pqbatch(void); vm_page_t vm_page_find_least(vm_object_t, vm_pindex_t); bool vm_page_free_prep(vm_page_t m); vm_page_t vm_page_getfake(vm_paddr_t paddr, vm_memattr_t memattr); @@ -576,6 +575,8 @@ vm_page_t vm_page_lookup (vm_object_t, vm_pindex_t); vm_page_t vm_page_next(vm_page_t m); int vm_page_pa_tryrelock(pmap_t, vm_paddr_t, vm_paddr_t *); +void vm_page_pqbatch_drain(void); +void vm_page_pqbatch_submit(vm_page_t m, uint8_t queue); vm_page_t vm_page_prev(vm_page_t m); bool vm_page_ps_test(vm_page_t m, int flags, vm_page_t skip_m); void vm_page_putfake(vm_page_t m); Index: sys/vm/vm_page.c =================================================================== --- sys/vm/vm_page.c +++ sys/vm/vm_page.c @@ -3134,8 +3134,8 @@ vm_batchqueue_init(bq); } -static void -vm_pqbatch_submit_page(vm_page_t m, uint8_t queue) +void +vm_page_pqbatch_submit(vm_page_t m, uint8_t queue) { struct vm_batchqueue *bq; struct vm_pagequeue *pq; @@ -3183,14 +3183,14 @@ } /* - * vm_page_drain_pqbatch: [ internal use only ] + * vm_page_pqbatch_drain: [ internal use only ] * * Force all per-CPU page queue batch queues to be drained. This is * intended for use in severe memory shortages, to ensure that pages * do not remain stuck in the batch queues. */ void -vm_page_drain_pqbatch(void) +vm_page_pqbatch_drain(void) { struct thread *td; struct vm_domain *vmd; @@ -3263,7 +3263,7 @@ * will be created, but the second will have no effect. */ if (vm_page_pqstate_cmpset(m, queue, queue, PGA_DEQUEUE, PGA_DEQUEUE)) - vm_pqbatch_submit_page(m, queue); + vm_page_pqbatch_submit(m, queue); } /* @@ -3288,7 +3288,7 @@ if ((queue = m->queue) == PQ_NONE) return; vm_page_aflag_set(m, PGA_DEQUEUE); - vm_pqbatch_submit_page(m, queue); + vm_page_pqbatch_submit(m, queue); } /* @@ -3363,7 +3363,7 @@ m->queue = queue; if ((m->aflags & PGA_REQUEUE) == 0) vm_page_aflag_set(m, PGA_REQUEUE); - vm_pqbatch_submit_page(m, queue); + vm_page_pqbatch_submit(m, queue); } /* @@ -3383,7 +3383,7 @@ if ((m->aflags & PGA_REQUEUE) == 0) vm_page_aflag_set(m, PGA_REQUEUE); - vm_pqbatch_submit_page(m, atomic_load_8(&m->queue)); + vm_page_pqbatch_submit(m, atomic_load_8(&m->queue)); } /* @@ -3411,7 +3411,7 @@ */ if (vm_page_pqstate_cmpset(m, oldq, newq, PGA_DEQUEUE | PGA_REQUEUE, PGA_REQUEUE)) - vm_pqbatch_submit_page(m, newq); + vm_page_pqbatch_submit(m, newq); return; } @@ -3430,7 +3430,7 @@ vm_page_aflag_clear(m, PGA_ENQUEUED); } vm_pagequeue_unlock(pq); - vm_pqbatch_submit_page(m, newq); + vm_page_pqbatch_submit(m, newq); } /* @@ -3799,7 +3799,7 @@ } if ((m->aflags & PGA_REQUEUE_HEAD) == 0) vm_page_aflag_set(m, PGA_REQUEUE_HEAD); - vm_pqbatch_submit_page(m, PQ_INACTIVE); + vm_page_pqbatch_submit(m, PQ_INACTIVE); } void Index: sys/vm/vm_swapout.c =================================================================== --- sys/vm/vm_swapout.c +++ sys/vm/vm_swapout.c @@ -413,7 +413,7 @@ * avoidance measure. */ if ((swapout_flags & VM_SWAP_NORMAL) != 0) - vm_page_drain_pqbatch(); + vm_page_pqbatch_drain(); swapout_procs(swapout_flags); }