mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-09-19 09:04:57 +00:00
8495f7e673
When a user process exits, the kernel cleans up the mm_struct of the user process and during cleanup, check_mm() checks the page tables of the user process for corruption (E.g: unexpected page flags set/cleared). For corrupted page tables, the error message printed by check_mm() isn't very clear as it prints the loop index instead of page table type (E.g: Resident file mapping pages vs Resident shared memory pages). The loop index in check_mm() is used to index rss_stat[] which represents individual memory type stats. Hence, instead of printing index, print memory type, thereby improving error message. Without patch: -------------- [ 204.836425] mm/pgtable-generic.c:29: bad p4d 0000000089eb4e92(800000025f941467) [ 204.836544] BUG: Bad rss-counter state mm:00000000f75895ea idx:0 val:2 [ 204.836615] BUG: Bad rss-counter state mm:00000000f75895ea idx:1 val:5 [ 204.836685] BUG: non-zero pgtables_bytes on freeing mm: 20480 With patch: ----------- [ 69.815453] mm/pgtable-generic.c:29: bad p4d 0000000084653642(800000025ca37467) [ 69.815872] BUG: Bad rss-counter state mm:00000000014a6c03 type:MM_FILEPAGES val:2 [ 69.815962] BUG: Bad rss-counter state mm:00000000014a6c03 type:MM_ANONPAGES val:5 [ 69.816050] BUG: non-zero pgtables_bytes on freeing mm: 20480 Also, change print function (from printk(KERN_ALERT, ..) to pr_alert()) so that it matches the other print statement. Link: http://lkml.kernel.org/r/da75b5153f617f4c5739c08ee6ebeb3d19db0fbc.1565123758.git.sai.praneeth.prakhya@intel.com Signed-off-by: Sai Praneeth Prakhya <sai.praneeth.prakhya@intel.com> Reviewed-by: Anshuman Khandual <anshuman.khandual@arm.com> Suggested-by: Dave Hansen <dave.hansen@intel.com> Acked-by: Michal Hocko <mhocko@suse.com> Acked-by: Vlastimil Babka <vbabka@suse.cz> Acked-by: Dave Hansen <dave.hansen@intel.com> Cc: Ingo Molnar <mingo@kernel.org> Cc: Peter Zijlstra <peterz@infradead.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
99 lines
2.5 KiB
C
99 lines
2.5 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef _LINUX_MM_TYPES_TASK_H
|
|
#define _LINUX_MM_TYPES_TASK_H
|
|
|
|
/*
|
|
* Here are the definitions of the MM data types that are embedded in 'struct task_struct'.
|
|
*
|
|
* (These are defined separately to decouple sched.h from mm_types.h as much as possible.)
|
|
*/
|
|
|
|
#include <linux/types.h>
|
|
#include <linux/threads.h>
|
|
#include <linux/atomic.h>
|
|
#include <linux/cpumask.h>
|
|
|
|
#include <asm/page.h>
|
|
|
|
#ifdef CONFIG_ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH
|
|
#include <asm/tlbbatch.h>
|
|
#endif
|
|
|
|
#define USE_SPLIT_PTE_PTLOCKS (NR_CPUS >= CONFIG_SPLIT_PTLOCK_CPUS)
|
|
#define USE_SPLIT_PMD_PTLOCKS (USE_SPLIT_PTE_PTLOCKS && \
|
|
IS_ENABLED(CONFIG_ARCH_ENABLE_SPLIT_PMD_PTLOCK))
|
|
#define ALLOC_SPLIT_PTLOCKS (SPINLOCK_SIZE > BITS_PER_LONG/8)
|
|
|
|
/*
|
|
* The per task VMA cache array:
|
|
*/
|
|
#define VMACACHE_BITS 2
|
|
#define VMACACHE_SIZE (1U << VMACACHE_BITS)
|
|
#define VMACACHE_MASK (VMACACHE_SIZE - 1)
|
|
|
|
struct vmacache {
|
|
u64 seqnum;
|
|
struct vm_area_struct *vmas[VMACACHE_SIZE];
|
|
};
|
|
|
|
/*
|
|
* When updating this, please also update struct resident_page_types[] in
|
|
* kernel/fork.c
|
|
*/
|
|
enum {
|
|
MM_FILEPAGES, /* Resident file mapping pages */
|
|
MM_ANONPAGES, /* Resident anonymous pages */
|
|
MM_SWAPENTS, /* Anonymous swap entries */
|
|
MM_SHMEMPAGES, /* Resident shared memory pages */
|
|
NR_MM_COUNTERS
|
|
};
|
|
|
|
#if USE_SPLIT_PTE_PTLOCKS && defined(CONFIG_MMU)
|
|
#define SPLIT_RSS_COUNTING
|
|
/* per-thread cached information, */
|
|
struct task_rss_stat {
|
|
int events; /* for synchronization threshold */
|
|
int count[NR_MM_COUNTERS];
|
|
};
|
|
#endif /* USE_SPLIT_PTE_PTLOCKS */
|
|
|
|
struct mm_rss_stat {
|
|
atomic_long_t count[NR_MM_COUNTERS];
|
|
};
|
|
|
|
struct page_frag {
|
|
struct page *page;
|
|
#if (BITS_PER_LONG > 32) || (PAGE_SIZE >= 65536)
|
|
__u32 offset;
|
|
__u32 size;
|
|
#else
|
|
__u16 offset;
|
|
__u16 size;
|
|
#endif
|
|
};
|
|
|
|
/* Track pages that require TLB flushes */
|
|
struct tlbflush_unmap_batch {
|
|
#ifdef CONFIG_ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH
|
|
/*
|
|
* The arch code makes the following promise: generic code can modify a
|
|
* PTE, then call arch_tlbbatch_add_mm() (which internally provides all
|
|
* needed barriers), then call arch_tlbbatch_flush(), and the entries
|
|
* will be flushed on all CPUs by the time that arch_tlbbatch_flush()
|
|
* returns.
|
|
*/
|
|
struct arch_tlbflush_unmap_batch arch;
|
|
|
|
/* True if a flush is needed. */
|
|
bool flush_required;
|
|
|
|
/*
|
|
* If true then the PTE was dirty when unmapped. The entry must be
|
|
* flushed before IO is initiated or a stale TLB entry potentially
|
|
* allows an update without redirtying the page.
|
|
*/
|
|
bool writable;
|
|
#endif
|
|
};
|
|
|
|
#endif /* _LINUX_MM_TYPES_TASK_H */
|