mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-11-01 00:48:50 +00:00
mm: mask bits from pmd in pmd_lockptr/pmd_huge_pte
The pmd pointer passed to pmd_lockptr/pmd_huge_pte can point to any entry in a pmd table. With USE_SPLIT_PMD_PTLOCKS==1 the code uses virt_to_page to get a struct page for the pmd table. The virt_to_page function automatically masks the lower PAGE_SHIFT bits from the address. But if the size of a pmd table is larger than PAGE_SIZE the additional bits are not removed from the pmd address and the wrong page struct is used. Fix this by explicitely masking the offset in the pmd table from the pmd pointer. Signed-off-by: Martin Schwidefsky <schwidefsky@de.ibm.com>
This commit is contained in:
parent
db85eaeb52
commit
634391ace1
1 changed files with 8 additions and 2 deletions
|
@ -1477,9 +1477,15 @@ static inline void pgtable_page_dtor(struct page *page)
|
|||
|
||||
#if USE_SPLIT_PMD_PTLOCKS
|
||||
|
||||
static struct page *pmd_to_page(pmd_t *pmd)
|
||||
{
|
||||
unsigned long mask = ~(PTRS_PER_PMD * sizeof(pmd_t) - 1);
|
||||
return virt_to_page((void *)((unsigned long) pmd & mask));
|
||||
}
|
||||
|
||||
static inline spinlock_t *pmd_lockptr(struct mm_struct *mm, pmd_t *pmd)
|
||||
{
|
||||
return ptlock_ptr(virt_to_page(pmd));
|
||||
return ptlock_ptr(pmd_to_page(pmd));
|
||||
}
|
||||
|
||||
static inline bool pgtable_pmd_page_ctor(struct page *page)
|
||||
|
@ -1498,7 +1504,7 @@ static inline void pgtable_pmd_page_dtor(struct page *page)
|
|||
ptlock_free(page);
|
||||
}
|
||||
|
||||
#define pmd_huge_pte(mm, pmd) (virt_to_page(pmd)->pmd_huge_pte)
|
||||
#define pmd_huge_pte(mm, pmd) (pmd_to_page(pmd)->pmd_huge_pte)
|
||||
|
||||
#else
|
||||
|
||||
|
|
Loading…
Reference in a new issue