@@ -1211,7 +1211,7 @@ struct page *follow_trans_huge_pmd(struct vm_area_struct *vma,
1211
1211
return ERR_PTR (- EFAULT );
1212
1212
1213
1213
/* Full NUMA hinting faults to serialise migration in fault paths */
1214
- if ((flags & FOLL_NUMA ) && pmd_numa (* pmd ))
1214
+ if ((flags & FOLL_NUMA ) && pmd_protnone (* pmd ))
1215
1215
goto out ;
1216
1216
1217
1217
page = pmd_page (* pmd );
@@ -1342,7 +1342,7 @@ int do_huge_pmd_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
1342
1342
1343
1343
/*
1344
1344
* Migrate the THP to the requested node, returns with page unlocked
1345
- * and pmd_numa cleared .
1345
+ * and access rights restored .
1346
1346
*/
1347
1347
spin_unlock (ptl );
1348
1348
migrated = migrate_misplaced_transhuge_page (mm , vma ,
@@ -1357,7 +1357,7 @@ int do_huge_pmd_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
1357
1357
BUG_ON (!PageLocked (page ));
1358
1358
pmd = pmd_mknonnuma (pmd );
1359
1359
set_pmd_at (mm , haddr , pmdp , pmd );
1360
- VM_BUG_ON (pmd_numa (* pmdp ));
1360
+ VM_BUG_ON (pmd_protnone (* pmdp ));
1361
1361
update_mmu_cache_pmd (vma , addr , pmdp );
1362
1362
unlock_page (page );
1363
1363
out_unlock :
@@ -1483,7 +1483,7 @@ int change_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd,
1483
1483
ret = 1 ;
1484
1484
if (!prot_numa ) {
1485
1485
entry = pmdp_get_and_clear_notify (mm , addr , pmd );
1486
- if (pmd_numa (entry ))
1486
+ if (pmd_protnone (entry ))
1487
1487
entry = pmd_mknonnuma (entry );
1488
1488
entry = pmd_modify (entry , newprot );
1489
1489
ret = HPAGE_PMD_NR ;
@@ -1499,7 +1499,7 @@ int change_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd,
1499
1499
* local vs remote hits on the zero page.
1500
1500
*/
1501
1501
if (!is_huge_zero_page (page ) &&
1502
- !pmd_numa (* pmd )) {
1502
+ !pmd_protnone (* pmd )) {
1503
1503
pmdp_set_numa (mm , addr , pmd );
1504
1504
ret = HPAGE_PMD_NR ;
1505
1505
}
@@ -1767,9 +1767,9 @@ static int __split_huge_page_map(struct page *page,
1767
1767
pte_t * pte , entry ;
1768
1768
BUG_ON (PageCompound (page + i ));
1769
1769
/*
1770
- * Note that pmd_numa is not transferred deliberately
1771
- * to avoid any possibility that pte_numa leaks to
1772
- * a PROT_NONE VMA by accident .
1770
+ * Note that NUMA hinting access restrictions are not
1771
+ * transferred to avoid any possibility of altering
1772
+ * permissions across VMAs .
1773
1773
*/
1774
1774
entry = mk_pte (page + i , vma -> vm_page_prot );
1775
1775
entry = maybe_mkwrite (pte_mkdirty (entry ), vma );
0 commit comments