pmap_remove_l3() already invalidates any existing TLB entry for the address being unmapped. There is no need to repeat the invalidation in the caller.
I agree that the change is functionally correct, but perhaps the intent was to coalesce the invalidations ? I.e. either pmap_remove_l3() should stop doing the invalidation, or its variant is added which does not do the invalidation and used there.
I mean that invalidate page puts the barriers around tlbi, dsb+isb probably mean full CPU stop, so the existing code tried to avoid doing it for each page.
Then, for correctness, you have to introduce the delayed invalidation machinery from amd64. (Or perform a TLB range invalidation operation whenever you change PV list locks of the addresses whose PTEs were destroyed since the last PV list lock change.)
It will work well for 4KB mappings of physical pages that come from superpage reservations and those cases where the vm_phys buddy allocator just happens to give you contiguous physical pages.
In any case, the pmap_invalidate_range() calls would have to be removed from pmap_remove() and "pushed down" inside of pmap_remove_l3() where we perform the change lock.
I think the architecture allows us to issue the tlb invalidation, then later execute the dsb & isb instructions as long as we don't try to the old mapping, however will need to check with ARM. We should also just invalidate all entries in pmap_invalidate_range if the length is too long, however in this case it may be better to stop invalidating until after we have removed all the pagetable entries.
Here are two related observations:
- pmap_protect() could easily (and correctly) use pmap_invalidate_range(). Because pmap_protect() does not destroy PV entries, it does not face the above mentioned issues with range invalidations in pmap_remove().
- I can't imagine that pmap_remove_l2() needs to do a pmap_invalidate_range(). Given the break-before-make nonsense that arm requires for the creation of promoted mappings, there shouldn't be anything besides a single superpage mapping in the TLB when pmap_remove_l2() executes.
|4166 ↗||(On Diff #34136)|
I think that the removal of this particular redundant call to pmap_invalidate_page() in pmap_ts_referenced() is non-controversial, so I'm going to go ahead and eliminate it.
It actually gets executed with some frequency, because we are making up for the lack of a functioning access bit by removing mappings.
Introduce pmap_remove_l3_range() and use it in two places: (1) pmap_remove(), where it eliminates redundant TLB invalidations by pmap_remove() and pmap_remove_l3(), and (2) pmap_enter_l2(), where it may optimize the TLB invalidations by batching them.