mm/hmm: return -EFAULT when setting HMM_PFN_ERROR on requested valid pages

hmm_range_fault() should never return 0 if the caller requested a valid
page, but the pfns output for that page would be HMM_PFN_ERROR.

hmm_pte_need_fault() must always be called before setting HMM_PFN_ERROR to
detect if the page is in faulting mode or not.

Fix two cases in hmm_vma_walk_pmd() and reorganize some of the duplicated
code.

Fixes: d08faca018 ("mm/hmm: properly handle migration pmd")
Fixes: da4c3c735e ("mm/hmm/mirror: helper to snapshot CPU page table")
Reviewed-by: Ralph Campbell <rcampbell@nvidia.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jason Gunthorpe <jgg@mellanox.com>
This commit is contained in:
Jason Gunthorpe 2020-03-05 15:26:33 -04:00
parent 76612d6ce4
commit 2288a9a681
1 changed files with 21 additions and 17 deletions

View File

@ -371,8 +371,10 @@ static int hmm_vma_walk_pmd(pmd_t *pmdp,
{ {
struct hmm_vma_walk *hmm_vma_walk = walk->private; struct hmm_vma_walk *hmm_vma_walk = walk->private;
struct hmm_range *range = hmm_vma_walk->range; struct hmm_range *range = hmm_vma_walk->range;
uint64_t *pfns = range->pfns; uint64_t *pfns = &range->pfns[(start - range->start) >> PAGE_SHIFT];
unsigned long addr = start, i; unsigned long npages = (end - start) >> PAGE_SHIFT;
unsigned long addr = start;
bool fault, write_fault;
pte_t *ptep; pte_t *ptep;
pmd_t pmd; pmd_t pmd;
@ -382,14 +384,6 @@ static int hmm_vma_walk_pmd(pmd_t *pmdp,
return hmm_vma_walk_hole(start, end, -1, walk); return hmm_vma_walk_hole(start, end, -1, walk);
if (thp_migration_supported() && is_pmd_migration_entry(pmd)) { if (thp_migration_supported() && is_pmd_migration_entry(pmd)) {
bool fault, write_fault;
unsigned long npages;
uint64_t *pfns;
i = (addr - range->start) >> PAGE_SHIFT;
npages = (end - addr) >> PAGE_SHIFT;
pfns = &range->pfns[i];
hmm_range_need_fault(hmm_vma_walk, pfns, npages, hmm_range_need_fault(hmm_vma_walk, pfns, npages,
0, &fault, &write_fault); 0, &fault, &write_fault);
if (fault || write_fault) { if (fault || write_fault) {
@ -398,8 +392,15 @@ static int hmm_vma_walk_pmd(pmd_t *pmdp,
return -EBUSY; return -EBUSY;
} }
return hmm_pfns_fill(start, end, range, HMM_PFN_NONE); return hmm_pfns_fill(start, end, range, HMM_PFN_NONE);
} else if (!pmd_present(pmd)) }
if (!pmd_present(pmd)) {
hmm_range_need_fault(hmm_vma_walk, pfns, npages, 0, &fault,
&write_fault);
if (fault || write_fault)
return -EFAULT;
return hmm_pfns_fill(start, end, range, HMM_PFN_ERROR); return hmm_pfns_fill(start, end, range, HMM_PFN_ERROR);
}
if (pmd_devmap(pmd) || pmd_trans_huge(pmd)) { if (pmd_devmap(pmd) || pmd_trans_huge(pmd)) {
/* /*
@ -416,8 +417,7 @@ static int hmm_vma_walk_pmd(pmd_t *pmdp,
if (!pmd_devmap(pmd) && !pmd_trans_huge(pmd)) if (!pmd_devmap(pmd) && !pmd_trans_huge(pmd))
goto again; goto again;
i = (addr - range->start) >> PAGE_SHIFT; return hmm_vma_handle_pmd(walk, addr, end, pfns, pmd);
return hmm_vma_handle_pmd(walk, addr, end, &pfns[i], pmd);
} }
/* /*
@ -426,15 +426,19 @@ static int hmm_vma_walk_pmd(pmd_t *pmdp,
* entry pointing to pte directory or it is a bad pmd that will not * entry pointing to pte directory or it is a bad pmd that will not
* recover. * recover.
*/ */
if (pmd_bad(pmd)) if (pmd_bad(pmd)) {
hmm_range_need_fault(hmm_vma_walk, pfns, npages, 0, &fault,
&write_fault);
if (fault || write_fault)
return -EFAULT;
return hmm_pfns_fill(start, end, range, HMM_PFN_ERROR); return hmm_pfns_fill(start, end, range, HMM_PFN_ERROR);
}
ptep = pte_offset_map(pmdp, addr); ptep = pte_offset_map(pmdp, addr);
i = (addr - range->start) >> PAGE_SHIFT; for (; addr < end; addr += PAGE_SIZE, ptep++, pfns++) {
for (; addr < end; addr += PAGE_SIZE, ptep++, i++) {
int r; int r;
r = hmm_vma_handle_pte(walk, addr, end, pmdp, ptep, &pfns[i]); r = hmm_vma_handle_pte(walk, addr, end, pmdp, ptep, pfns);
if (r) { if (r) {
/* hmm_vma_handle_pte() did pte_unmap() */ /* hmm_vma_handle_pte() did pte_unmap() */
hmm_vma_walk->last = addr; hmm_vma_walk->last = addr;