Re: [PATCH v8 1/2] hugetlb: don't delete vma_lock in hugetlb MADV_DONTNEED processing

From: Nadav Amit
Date: Thu Nov 10 2022 - 17:22:45 EST


On Nov 10, 2022, at 1:48 PM, Mike Kravetz <mike.kravetz@xxxxxxxxxx> wrote:

>>> void unmap_hugepage_range(struct vm_area_struct *vma, unsigned long start,
>>> unsigned long end, struct page *ref_page,
>>> zap_flags_t zap_flags)
>>> {
>>> + struct mmu_notifier_range range;
>>> struct mmu_gather tlb;
>>>
>>> + mmu_notifier_range_init(&range, MMU_NOTIFY_UNMAP, 0, vma, vma->vm_mm,
>>> + start, end);
>>> + adjust_range_if_pmd_sharing_possible(vma, &range.start, &range.end);
>>> tlb_gather_mmu(&tlb, vma->vm_mm);
>>> +
>>> __unmap_hugepage_range(&tlb, vma, start, end, ref_page, zap_flags);
>>
>> Is there a reason for not using range.start and range.end?
>
> After calling adjust_range_if_pmd_sharing_possible, range.start - range.end
> could be much greater than the range we actually want to unmap. The range
> gets adjusted to account for pmd sharing if that is POSSIBLE. It does not
> know for sure if we will actually 'unshare a pmd'.
>
> I suppose adjust_range_if_pmd_sharing_possible could be modified to actually
> check if unmapping will result in unsharing, but it does not do that today.

Thanks for the explanation. It’s probably me, but I am still not sure that I
understand the the different between __unmap_hugepage_range() using (start,
end) and __zap_page_range_single() using (address, range.end). Perhaps it
worth a comment in the code?

But anyhow… shouldn’t unmap_hugepage_range() call
mmu_notifier_invalidate_range_start()?