mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-10-31 16:38:12 +00:00
Minor things, mainly - mailmap updates, MAINTAINERS updates, etc.
Fixes for post-5.18 changes: - fix for a damon boot hang, from SeongJae - fix for a kfence warning splat, from Jason Donenfeld - fix for zero-pfn pinning, from Alex Williamson - fix for fallocate hole punch clearing, from Mike Kravetz Fixes pre-5.18 material: - fix for a performance regression, from Marcelo - fix for a hwpoisining BUG from zhenwei pi -----BEGIN PGP SIGNATURE----- iHUEABYKAB0WIQTTMBEPP41GrTpTJgfdBJ7gKXxAjgUCYri4RgAKCRDdBJ7gKXxA jmhsAQDCvGqtIUhgkTwid8KBRNbowsg0LXd6k+gUjcxBhH403wEA0r0cxxkDAmgr QNXn/qZRzQP2ji+pdjH9NBOsd2g2XQA= =UGJ7 -----END PGP SIGNATURE----- Merge tag 'mm-hotfixes-stable-2022-06-26' of git://git.kernel.org/pub/scm/linux/kernel/git/akpm/mm Pull hotfixes from Andrew Morton: "Minor things, mainly - mailmap updates, MAINTAINERS updates, etc. Fixes for this merge window: - fix for a damon boot hang, from SeongJae - fix for a kfence warning splat, from Jason Donenfeld - fix for zero-pfn pinning, from Alex Williamson - fix for fallocate hole punch clearing, from Mike Kravetz Fixes for previous releases: - fix for a performance regression, from Marcelo - fix for a hwpoisining BUG from zhenwei pi" * tag 'mm-hotfixes-stable-2022-06-26' of git://git.kernel.org/pub/scm/linux/kernel/git/akpm/mm: mailmap: add entry for Christian Marangi mm/memory-failure: disable unpoison once hw error happens hugetlbfs: zero partial pages during fallocate hole punch mm: memcontrol: reference to tools/cgroup/memcg_slabinfo.py mm: re-allow pinning of zero pfns mm/kfence: select random number before taking raw lock MAINTAINERS: add maillist information for LoongArch MAINTAINERS: update MM tree references MAINTAINERS: update Abel Vesa's email MAINTAINERS: add MEMORY HOT(UN)PLUG section and add David as reviewer MAINTAINERS: add Miaohe Lin as a memory-failure reviewer mailmap: add alias for jarkko@profian.com mm/damon/reclaim: schedule 'damon_reclaim_timer' only after 'system_wq' is initialized kthread: make it clear that kthread_create_on_node() might be terminated by any fatal signal mm: lru_cache_disable: use synchronize_rcu_expedited mm/page_isolation.c: fix one kernel-doc comment
This commit is contained in:
commit
413c1f1491
15 changed files with 119 additions and 37 deletions
4
.mailmap
4
.mailmap
|
@ -10,6 +10,8 @@
|
||||||
# Please keep this list dictionary sorted.
|
# Please keep this list dictionary sorted.
|
||||||
#
|
#
|
||||||
Aaron Durbin <adurbin@google.com>
|
Aaron Durbin <adurbin@google.com>
|
||||||
|
Abel Vesa <abelvesa@kernel.org> <abel.vesa@nxp.com>
|
||||||
|
Abel Vesa <abelvesa@kernel.org> <abelvesa@gmail.com>
|
||||||
Abhinav Kumar <quic_abhinavk@quicinc.com> <abhinavk@codeaurora.org>
|
Abhinav Kumar <quic_abhinavk@quicinc.com> <abhinavk@codeaurora.org>
|
||||||
Adam Oldham <oldhamca@gmail.com>
|
Adam Oldham <oldhamca@gmail.com>
|
||||||
Adam Radford <aradford@gmail.com>
|
Adam Radford <aradford@gmail.com>
|
||||||
|
@ -85,6 +87,7 @@ Christian Borntraeger <borntraeger@linux.ibm.com> <borntrae@de.ibm.com>
|
||||||
Christian Brauner <brauner@kernel.org> <christian@brauner.io>
|
Christian Brauner <brauner@kernel.org> <christian@brauner.io>
|
||||||
Christian Brauner <brauner@kernel.org> <christian.brauner@canonical.com>
|
Christian Brauner <brauner@kernel.org> <christian.brauner@canonical.com>
|
||||||
Christian Brauner <brauner@kernel.org> <christian.brauner@ubuntu.com>
|
Christian Brauner <brauner@kernel.org> <christian.brauner@ubuntu.com>
|
||||||
|
Christian Marangi <ansuelsmth@gmail.com>
|
||||||
Christophe Ricard <christophe.ricard@gmail.com>
|
Christophe Ricard <christophe.ricard@gmail.com>
|
||||||
Christoph Hellwig <hch@lst.de>
|
Christoph Hellwig <hch@lst.de>
|
||||||
Colin Ian King <colin.king@intel.com> <colin.king@canonical.com>
|
Colin Ian King <colin.king@intel.com> <colin.king@canonical.com>
|
||||||
|
@ -165,6 +168,7 @@ Jan Glauber <jan.glauber@gmail.com> <jang@de.ibm.com>
|
||||||
Jan Glauber <jan.glauber@gmail.com> <jang@linux.vnet.ibm.com>
|
Jan Glauber <jan.glauber@gmail.com> <jang@linux.vnet.ibm.com>
|
||||||
Jan Glauber <jan.glauber@gmail.com> <jglauber@cavium.com>
|
Jan Glauber <jan.glauber@gmail.com> <jglauber@cavium.com>
|
||||||
Jarkko Sakkinen <jarkko@kernel.org> <jarkko.sakkinen@linux.intel.com>
|
Jarkko Sakkinen <jarkko@kernel.org> <jarkko.sakkinen@linux.intel.com>
|
||||||
|
Jarkko Sakkinen <jarkko@kernel.org> <jarkko@profian.com>
|
||||||
Jason Gunthorpe <jgg@ziepe.ca> <jgg@mellanox.com>
|
Jason Gunthorpe <jgg@ziepe.ca> <jgg@mellanox.com>
|
||||||
Jason Gunthorpe <jgg@ziepe.ca> <jgg@nvidia.com>
|
Jason Gunthorpe <jgg@ziepe.ca> <jgg@nvidia.com>
|
||||||
Jason Gunthorpe <jgg@ziepe.ca> <jgunthorpe@obsidianresearch.com>
|
Jason Gunthorpe <jgg@ziepe.ca> <jgunthorpe@obsidianresearch.com>
|
||||||
|
|
|
@ -120,7 +120,8 @@ Testing
|
||||||
unpoison-pfn
|
unpoison-pfn
|
||||||
Software-unpoison page at PFN echoed into this file. This way
|
Software-unpoison page at PFN echoed into this file. This way
|
||||||
a page can be reused again. This only works for Linux
|
a page can be reused again. This only works for Linux
|
||||||
injected failures, not for real memory failures.
|
injected failures, not for real memory failures. Once any hardware
|
||||||
|
memory failure happens, this feature is disabled.
|
||||||
|
|
||||||
Note these injection interfaces are not stable and might change between
|
Note these injection interfaces are not stable and might change between
|
||||||
kernel versions
|
kernel versions
|
||||||
|
|
21
MAINTAINERS
21
MAINTAINERS
|
@ -9152,6 +9152,7 @@ F: drivers/media/platform/st/sti/hva
|
||||||
|
|
||||||
HWPOISON MEMORY FAILURE HANDLING
|
HWPOISON MEMORY FAILURE HANDLING
|
||||||
M: Naoya Horiguchi <naoya.horiguchi@nec.com>
|
M: Naoya Horiguchi <naoya.horiguchi@nec.com>
|
||||||
|
R: Miaohe Lin <linmiaohe@huawei.com>
|
||||||
L: linux-mm@kvack.org
|
L: linux-mm@kvack.org
|
||||||
S: Maintained
|
S: Maintained
|
||||||
F: mm/hwpoison-inject.c
|
F: mm/hwpoison-inject.c
|
||||||
|
@ -11623,6 +11624,7 @@ F: drivers/gpu/drm/bridge/lontium-lt8912b.c
|
||||||
LOONGARCH
|
LOONGARCH
|
||||||
M: Huacai Chen <chenhuacai@kernel.org>
|
M: Huacai Chen <chenhuacai@kernel.org>
|
||||||
R: WANG Xuerui <kernel@xen0n.name>
|
R: WANG Xuerui <kernel@xen0n.name>
|
||||||
|
L: loongarch@lists.linux.dev
|
||||||
S: Maintained
|
S: Maintained
|
||||||
T: git git://git.kernel.org/pub/scm/linux/kernel/git/chenhuacai/linux-loongson.git
|
T: git git://git.kernel.org/pub/scm/linux/kernel/git/chenhuacai/linux-loongson.git
|
||||||
F: arch/loongarch/
|
F: arch/loongarch/
|
||||||
|
@ -12879,9 +12881,8 @@ M: Andrew Morton <akpm@linux-foundation.org>
|
||||||
L: linux-mm@kvack.org
|
L: linux-mm@kvack.org
|
||||||
S: Maintained
|
S: Maintained
|
||||||
W: http://www.linux-mm.org
|
W: http://www.linux-mm.org
|
||||||
T: quilt https://ozlabs.org/~akpm/mmotm/
|
T: git://git.kernel.org/pub/scm/linux/kernel/git/akpm/mm
|
||||||
T: quilt https://ozlabs.org/~akpm/mmots/
|
T: quilt git://git.kernel.org/pub/scm/linux/kernel/git/akpm/25-new
|
||||||
T: git git://github.com/hnaz/linux-mm.git
|
|
||||||
F: include/linux/gfp.h
|
F: include/linux/gfp.h
|
||||||
F: include/linux/memory_hotplug.h
|
F: include/linux/memory_hotplug.h
|
||||||
F: include/linux/mm.h
|
F: include/linux/mm.h
|
||||||
|
@ -12891,6 +12892,18 @@ F: include/linux/vmalloc.h
|
||||||
F: mm/
|
F: mm/
|
||||||
F: tools/testing/selftests/vm/
|
F: tools/testing/selftests/vm/
|
||||||
|
|
||||||
|
MEMORY HOT(UN)PLUG
|
||||||
|
M: David Hildenbrand <david@redhat.com>
|
||||||
|
M: Oscar Salvador <osalvador@suse.de>
|
||||||
|
L: linux-mm@kvack.org
|
||||||
|
S: Maintained
|
||||||
|
F: Documentation/admin-guide/mm/memory-hotplug.rst
|
||||||
|
F: Documentation/core-api/memory-hotplug.rst
|
||||||
|
F: drivers/base/memory.c
|
||||||
|
F: include/linux/memory_hotplug.h
|
||||||
|
F: mm/memory_hotplug.c
|
||||||
|
F: tools/testing/selftests/memory-hotplug/
|
||||||
|
|
||||||
MEMORY TECHNOLOGY DEVICES (MTD)
|
MEMORY TECHNOLOGY DEVICES (MTD)
|
||||||
M: Miquel Raynal <miquel.raynal@bootlin.com>
|
M: Miquel Raynal <miquel.raynal@bootlin.com>
|
||||||
M: Richard Weinberger <richard@nod.at>
|
M: Richard Weinberger <richard@nod.at>
|
||||||
|
@ -14295,7 +14308,7 @@ F: drivers/iio/gyro/fxas21002c_i2c.c
|
||||||
F: drivers/iio/gyro/fxas21002c_spi.c
|
F: drivers/iio/gyro/fxas21002c_spi.c
|
||||||
|
|
||||||
NXP i.MX CLOCK DRIVERS
|
NXP i.MX CLOCK DRIVERS
|
||||||
M: Abel Vesa <abel.vesa@nxp.com>
|
M: Abel Vesa <abelvesa@kernel.org>
|
||||||
L: linux-clk@vger.kernel.org
|
L: linux-clk@vger.kernel.org
|
||||||
L: linux-imx@nxp.com
|
L: linux-imx@nxp.com
|
||||||
S: Maintained
|
S: Maintained
|
||||||
|
|
|
@ -558,7 +558,7 @@ static ssize_t hard_offline_page_store(struct device *dev,
|
||||||
if (kstrtoull(buf, 0, &pfn) < 0)
|
if (kstrtoull(buf, 0, &pfn) < 0)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
pfn >>= PAGE_SHIFT;
|
pfn >>= PAGE_SHIFT;
|
||||||
ret = memory_failure(pfn, 0);
|
ret = memory_failure(pfn, MF_SW_SIMULATED);
|
||||||
if (ret == -EOPNOTSUPP)
|
if (ret == -EOPNOTSUPP)
|
||||||
ret = 0;
|
ret = 0;
|
||||||
return ret ? ret : count;
|
return ret ? ret : count;
|
||||||
|
|
|
@ -600,23 +600,43 @@ static void hugetlb_vmtruncate(struct inode *inode, loff_t offset)
|
||||||
remove_inode_hugepages(inode, offset, LLONG_MAX);
|
remove_inode_hugepages(inode, offset, LLONG_MAX);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void hugetlbfs_zero_partial_page(struct hstate *h,
|
||||||
|
struct address_space *mapping,
|
||||||
|
loff_t start,
|
||||||
|
loff_t end)
|
||||||
|
{
|
||||||
|
pgoff_t idx = start >> huge_page_shift(h);
|
||||||
|
struct folio *folio;
|
||||||
|
|
||||||
|
folio = filemap_lock_folio(mapping, idx);
|
||||||
|
if (!folio)
|
||||||
|
return;
|
||||||
|
|
||||||
|
start = start & ~huge_page_mask(h);
|
||||||
|
end = end & ~huge_page_mask(h);
|
||||||
|
if (!end)
|
||||||
|
end = huge_page_size(h);
|
||||||
|
|
||||||
|
folio_zero_segment(folio, (size_t)start, (size_t)end);
|
||||||
|
|
||||||
|
folio_unlock(folio);
|
||||||
|
folio_put(folio);
|
||||||
|
}
|
||||||
|
|
||||||
static long hugetlbfs_punch_hole(struct inode *inode, loff_t offset, loff_t len)
|
static long hugetlbfs_punch_hole(struct inode *inode, loff_t offset, loff_t len)
|
||||||
{
|
{
|
||||||
|
struct hugetlbfs_inode_info *info = HUGETLBFS_I(inode);
|
||||||
|
struct address_space *mapping = inode->i_mapping;
|
||||||
struct hstate *h = hstate_inode(inode);
|
struct hstate *h = hstate_inode(inode);
|
||||||
loff_t hpage_size = huge_page_size(h);
|
loff_t hpage_size = huge_page_size(h);
|
||||||
loff_t hole_start, hole_end;
|
loff_t hole_start, hole_end;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* For hole punch round up the beginning offset of the hole and
|
* hole_start and hole_end indicate the full pages within the hole.
|
||||||
* round down the end.
|
|
||||||
*/
|
*/
|
||||||
hole_start = round_up(offset, hpage_size);
|
hole_start = round_up(offset, hpage_size);
|
||||||
hole_end = round_down(offset + len, hpage_size);
|
hole_end = round_down(offset + len, hpage_size);
|
||||||
|
|
||||||
if (hole_end > hole_start) {
|
|
||||||
struct address_space *mapping = inode->i_mapping;
|
|
||||||
struct hugetlbfs_inode_info *info = HUGETLBFS_I(inode);
|
|
||||||
|
|
||||||
inode_lock(inode);
|
inode_lock(inode);
|
||||||
|
|
||||||
/* protected by i_rwsem */
|
/* protected by i_rwsem */
|
||||||
|
@ -626,15 +646,33 @@ static long hugetlbfs_punch_hole(struct inode *inode, loff_t offset, loff_t len)
|
||||||
}
|
}
|
||||||
|
|
||||||
i_mmap_lock_write(mapping);
|
i_mmap_lock_write(mapping);
|
||||||
|
|
||||||
|
/* If range starts before first full page, zero partial page. */
|
||||||
|
if (offset < hole_start)
|
||||||
|
hugetlbfs_zero_partial_page(h, mapping,
|
||||||
|
offset, min(offset + len, hole_start));
|
||||||
|
|
||||||
|
/* Unmap users of full pages in the hole. */
|
||||||
|
if (hole_end > hole_start) {
|
||||||
if (!RB_EMPTY_ROOT(&mapping->i_mmap.rb_root))
|
if (!RB_EMPTY_ROOT(&mapping->i_mmap.rb_root))
|
||||||
hugetlb_vmdelete_list(&mapping->i_mmap,
|
hugetlb_vmdelete_list(&mapping->i_mmap,
|
||||||
hole_start >> PAGE_SHIFT,
|
hole_start >> PAGE_SHIFT,
|
||||||
hole_end >> PAGE_SHIFT, 0);
|
hole_end >> PAGE_SHIFT, 0);
|
||||||
i_mmap_unlock_write(mapping);
|
|
||||||
remove_inode_hugepages(inode, hole_start, hole_end);
|
|
||||||
inode_unlock(inode);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* If range extends beyond last full page, zero partial page. */
|
||||||
|
if ((offset + len) > hole_end && (offset + len) > hole_start)
|
||||||
|
hugetlbfs_zero_partial_page(h, mapping,
|
||||||
|
hole_end, offset + len);
|
||||||
|
|
||||||
|
i_mmap_unlock_write(mapping);
|
||||||
|
|
||||||
|
/* Remove full pages from the file. */
|
||||||
|
if (hole_end > hole_start)
|
||||||
|
remove_inode_hugepages(inode, hole_start, hole_end);
|
||||||
|
|
||||||
|
inode_unlock(inode);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1600,7 +1600,7 @@ static inline bool is_pinnable_page(struct page *page)
|
||||||
if (mt == MIGRATE_CMA || mt == MIGRATE_ISOLATE)
|
if (mt == MIGRATE_CMA || mt == MIGRATE_ISOLATE)
|
||||||
return false;
|
return false;
|
||||||
#endif
|
#endif
|
||||||
return !(is_zone_movable_page(page) || is_zero_pfn(page_to_pfn(page)));
|
return !is_zone_movable_page(page) || is_zero_pfn(page_to_pfn(page));
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
static inline bool is_pinnable_page(struct page *page)
|
static inline bool is_pinnable_page(struct page *page)
|
||||||
|
@ -3232,6 +3232,7 @@ enum mf_flags {
|
||||||
MF_MUST_KILL = 1 << 2,
|
MF_MUST_KILL = 1 << 2,
|
||||||
MF_SOFT_OFFLINE = 1 << 3,
|
MF_SOFT_OFFLINE = 1 << 3,
|
||||||
MF_UNPOISON = 1 << 4,
|
MF_UNPOISON = 1 << 4,
|
||||||
|
MF_SW_SIMULATED = 1 << 5,
|
||||||
};
|
};
|
||||||
extern int memory_failure(unsigned long pfn, int flags);
|
extern int memory_failure(unsigned long pfn, int flags);
|
||||||
extern void memory_failure_queue(unsigned long pfn, int flags);
|
extern void memory_failure_queue(unsigned long pfn, int flags);
|
||||||
|
|
|
@ -340,7 +340,7 @@ static int kthread(void *_create)
|
||||||
|
|
||||||
self = to_kthread(current);
|
self = to_kthread(current);
|
||||||
|
|
||||||
/* If user was SIGKILLed, I release the structure. */
|
/* Release the structure when caller killed by a fatal signal. */
|
||||||
done = xchg(&create->done, NULL);
|
done = xchg(&create->done, NULL);
|
||||||
if (!done) {
|
if (!done) {
|
||||||
kfree(create);
|
kfree(create);
|
||||||
|
@ -398,7 +398,7 @@ static void create_kthread(struct kthread_create_info *create)
|
||||||
/* We want our own signal handler (we take no signals by default). */
|
/* We want our own signal handler (we take no signals by default). */
|
||||||
pid = kernel_thread(kthread, create, CLONE_FS | CLONE_FILES | SIGCHLD);
|
pid = kernel_thread(kthread, create, CLONE_FS | CLONE_FILES | SIGCHLD);
|
||||||
if (pid < 0) {
|
if (pid < 0) {
|
||||||
/* If user was SIGKILLed, I release the structure. */
|
/* Release the structure when caller killed by a fatal signal. */
|
||||||
struct completion *done = xchg(&create->done, NULL);
|
struct completion *done = xchg(&create->done, NULL);
|
||||||
|
|
||||||
if (!done) {
|
if (!done) {
|
||||||
|
@ -440,9 +440,9 @@ struct task_struct *__kthread_create_on_node(int (*threadfn)(void *data),
|
||||||
*/
|
*/
|
||||||
if (unlikely(wait_for_completion_killable(&done))) {
|
if (unlikely(wait_for_completion_killable(&done))) {
|
||||||
/*
|
/*
|
||||||
* If I was SIGKILLed before kthreadd (or new kernel thread)
|
* If I was killed by a fatal signal before kthreadd (or new
|
||||||
* calls complete(), leave the cleanup of this structure to
|
* kernel thread) calls complete(), leave the cleanup of this
|
||||||
* that thread.
|
* structure to that thread.
|
||||||
*/
|
*/
|
||||||
if (xchg(&create->done, NULL))
|
if (xchg(&create->done, NULL))
|
||||||
return ERR_PTR(-EINTR);
|
return ERR_PTR(-EINTR);
|
||||||
|
@ -876,7 +876,7 @@ __kthread_create_worker(int cpu, unsigned int flags,
|
||||||
*
|
*
|
||||||
* Returns a pointer to the allocated worker on success, ERR_PTR(-ENOMEM)
|
* Returns a pointer to the allocated worker on success, ERR_PTR(-ENOMEM)
|
||||||
* when the needed structures could not get allocated, and ERR_PTR(-EINTR)
|
* when the needed structures could not get allocated, and ERR_PTR(-EINTR)
|
||||||
* when the worker was SIGKILLed.
|
* when the caller was killed by a fatal signal.
|
||||||
*/
|
*/
|
||||||
struct kthread_worker *
|
struct kthread_worker *
|
||||||
kthread_create_worker(unsigned int flags, const char namefmt[], ...)
|
kthread_create_worker(unsigned int flags, const char namefmt[], ...)
|
||||||
|
@ -925,7 +925,7 @@ EXPORT_SYMBOL(kthread_create_worker);
|
||||||
* Return:
|
* Return:
|
||||||
* The pointer to the allocated worker on success, ERR_PTR(-ENOMEM)
|
* The pointer to the allocated worker on success, ERR_PTR(-ENOMEM)
|
||||||
* when the needed structures could not get allocated, and ERR_PTR(-EINTR)
|
* when the needed structures could not get allocated, and ERR_PTR(-EINTR)
|
||||||
* when the worker was SIGKILLed.
|
* when the caller was killed by a fatal signal.
|
||||||
*/
|
*/
|
||||||
struct kthread_worker *
|
struct kthread_worker *
|
||||||
kthread_create_worker_on_cpu(int cpu, unsigned int flags,
|
kthread_create_worker_on_cpu(int cpu, unsigned int flags,
|
||||||
|
|
|
@ -374,6 +374,8 @@ static void damon_reclaim_timer_fn(struct work_struct *work)
|
||||||
}
|
}
|
||||||
static DECLARE_DELAYED_WORK(damon_reclaim_timer, damon_reclaim_timer_fn);
|
static DECLARE_DELAYED_WORK(damon_reclaim_timer, damon_reclaim_timer_fn);
|
||||||
|
|
||||||
|
static bool damon_reclaim_initialized;
|
||||||
|
|
||||||
static int enabled_store(const char *val,
|
static int enabled_store(const char *val,
|
||||||
const struct kernel_param *kp)
|
const struct kernel_param *kp)
|
||||||
{
|
{
|
||||||
|
@ -382,6 +384,10 @@ static int enabled_store(const char *val,
|
||||||
if (rc < 0)
|
if (rc < 0)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
|
/* system_wq might not initialized yet */
|
||||||
|
if (!damon_reclaim_initialized)
|
||||||
|
return rc;
|
||||||
|
|
||||||
if (enabled)
|
if (enabled)
|
||||||
schedule_delayed_work(&damon_reclaim_timer, 0);
|
schedule_delayed_work(&damon_reclaim_timer, 0);
|
||||||
|
|
||||||
|
@ -449,6 +455,8 @@ static int __init damon_reclaim_init(void)
|
||||||
damon_add_target(ctx, target);
|
damon_add_target(ctx, target);
|
||||||
|
|
||||||
schedule_delayed_work(&damon_reclaim_timer, 0);
|
schedule_delayed_work(&damon_reclaim_timer, 0);
|
||||||
|
|
||||||
|
damon_reclaim_initialized = true;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -48,7 +48,7 @@ static int hwpoison_inject(void *data, u64 val)
|
||||||
|
|
||||||
inject:
|
inject:
|
||||||
pr_info("Injecting memory failure at pfn %#lx\n", pfn);
|
pr_info("Injecting memory failure at pfn %#lx\n", pfn);
|
||||||
err = memory_failure(pfn, 0);
|
err = memory_failure(pfn, MF_SW_SIMULATED);
|
||||||
return (err == -EOPNOTSUPP) ? 0 : err;
|
return (err == -EOPNOTSUPP) ? 0 : err;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -360,6 +360,9 @@ static void *kfence_guarded_alloc(struct kmem_cache *cache, size_t size, gfp_t g
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
struct slab *slab;
|
struct slab *slab;
|
||||||
void *addr;
|
void *addr;
|
||||||
|
const bool random_right_allocate = prandom_u32_max(2);
|
||||||
|
const bool random_fault = CONFIG_KFENCE_STRESS_TEST_FAULTS &&
|
||||||
|
!prandom_u32_max(CONFIG_KFENCE_STRESS_TEST_FAULTS);
|
||||||
|
|
||||||
/* Try to obtain a free object. */
|
/* Try to obtain a free object. */
|
||||||
raw_spin_lock_irqsave(&kfence_freelist_lock, flags);
|
raw_spin_lock_irqsave(&kfence_freelist_lock, flags);
|
||||||
|
@ -404,7 +407,7 @@ static void *kfence_guarded_alloc(struct kmem_cache *cache, size_t size, gfp_t g
|
||||||
* is that the out-of-bounds accesses detected are deterministic for
|
* is that the out-of-bounds accesses detected are deterministic for
|
||||||
* such allocations.
|
* such allocations.
|
||||||
*/
|
*/
|
||||||
if (prandom_u32_max(2)) {
|
if (random_right_allocate) {
|
||||||
/* Allocate on the "right" side, re-calculate address. */
|
/* Allocate on the "right" side, re-calculate address. */
|
||||||
meta->addr += PAGE_SIZE - size;
|
meta->addr += PAGE_SIZE - size;
|
||||||
meta->addr = ALIGN_DOWN(meta->addr, cache->align);
|
meta->addr = ALIGN_DOWN(meta->addr, cache->align);
|
||||||
|
@ -444,7 +447,7 @@ static void *kfence_guarded_alloc(struct kmem_cache *cache, size_t size, gfp_t g
|
||||||
if (cache->ctor)
|
if (cache->ctor)
|
||||||
cache->ctor(addr);
|
cache->ctor(addr);
|
||||||
|
|
||||||
if (CONFIG_KFENCE_STRESS_TEST_FAULTS && !prandom_u32_max(CONFIG_KFENCE_STRESS_TEST_FAULTS))
|
if (random_fault)
|
||||||
kfence_protect(meta->addr); /* Random "faults" by protecting the object. */
|
kfence_protect(meta->addr); /* Random "faults" by protecting the object. */
|
||||||
|
|
||||||
atomic_long_inc(&counters[KFENCE_COUNTER_ALLOCATED]);
|
atomic_long_inc(&counters[KFENCE_COUNTER_ALLOCATED]);
|
||||||
|
|
|
@ -1112,7 +1112,7 @@ static int madvise_inject_error(int behavior,
|
||||||
} else {
|
} else {
|
||||||
pr_info("Injecting memory failure for pfn %#lx at process virtual address %#lx\n",
|
pr_info("Injecting memory failure for pfn %#lx at process virtual address %#lx\n",
|
||||||
pfn, start);
|
pfn, start);
|
||||||
ret = memory_failure(pfn, MF_COUNT_INCREASED);
|
ret = memory_failure(pfn, MF_COUNT_INCREASED | MF_SW_SIMULATED);
|
||||||
if (ret == -EOPNOTSUPP)
|
if (ret == -EOPNOTSUPP)
|
||||||
ret = 0;
|
ret = 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -4859,7 +4859,7 @@ static int mem_cgroup_slab_show(struct seq_file *m, void *p)
|
||||||
{
|
{
|
||||||
/*
|
/*
|
||||||
* Deprecated.
|
* Deprecated.
|
||||||
* Please, take a look at tools/cgroup/slabinfo.py .
|
* Please, take a look at tools/cgroup/memcg_slabinfo.py .
|
||||||
*/
|
*/
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -69,6 +69,8 @@ int sysctl_memory_failure_recovery __read_mostly = 1;
|
||||||
|
|
||||||
atomic_long_t num_poisoned_pages __read_mostly = ATOMIC_LONG_INIT(0);
|
atomic_long_t num_poisoned_pages __read_mostly = ATOMIC_LONG_INIT(0);
|
||||||
|
|
||||||
|
static bool hw_memory_failure __read_mostly = false;
|
||||||
|
|
||||||
static bool __page_handle_poison(struct page *page)
|
static bool __page_handle_poison(struct page *page)
|
||||||
{
|
{
|
||||||
int ret;
|
int ret;
|
||||||
|
@ -1768,6 +1770,9 @@ int memory_failure(unsigned long pfn, int flags)
|
||||||
|
|
||||||
mutex_lock(&mf_mutex);
|
mutex_lock(&mf_mutex);
|
||||||
|
|
||||||
|
if (!(flags & MF_SW_SIMULATED))
|
||||||
|
hw_memory_failure = true;
|
||||||
|
|
||||||
p = pfn_to_online_page(pfn);
|
p = pfn_to_online_page(pfn);
|
||||||
if (!p) {
|
if (!p) {
|
||||||
res = arch_memory_failure(pfn, flags);
|
res = arch_memory_failure(pfn, flags);
|
||||||
|
@ -2103,6 +2108,13 @@ int unpoison_memory(unsigned long pfn)
|
||||||
|
|
||||||
mutex_lock(&mf_mutex);
|
mutex_lock(&mf_mutex);
|
||||||
|
|
||||||
|
if (hw_memory_failure) {
|
||||||
|
unpoison_pr_info("Unpoison: Disabled after HW memory failure %#lx\n",
|
||||||
|
pfn, &unpoison_rs);
|
||||||
|
ret = -EOPNOTSUPP;
|
||||||
|
goto unlock_mutex;
|
||||||
|
}
|
||||||
|
|
||||||
if (!PageHWPoison(p)) {
|
if (!PageHWPoison(p)) {
|
||||||
unpoison_pr_info("Unpoison: Page was already unpoisoned %#lx\n",
|
unpoison_pr_info("Unpoison: Page was already unpoisoned %#lx\n",
|
||||||
pfn, &unpoison_rs);
|
pfn, &unpoison_rs);
|
||||||
|
|
|
@ -286,6 +286,8 @@ __first_valid_page(unsigned long pfn, unsigned long nr_pages)
|
||||||
* @flags: isolation flags
|
* @flags: isolation flags
|
||||||
* @gfp_flags: GFP flags used for migrating pages
|
* @gfp_flags: GFP flags used for migrating pages
|
||||||
* @isolate_before: isolate the pageblock before the boundary_pfn
|
* @isolate_before: isolate the pageblock before the boundary_pfn
|
||||||
|
* @skip_isolation: the flag to skip the pageblock isolation in second
|
||||||
|
* isolate_single_pageblock()
|
||||||
*
|
*
|
||||||
* Free and in-use pages can be as big as MAX_ORDER-1 and contain more than one
|
* Free and in-use pages can be as big as MAX_ORDER-1 and contain more than one
|
||||||
* pageblock. When not all pageblocks within a page are isolated at the same
|
* pageblock. When not all pageblocks within a page are isolated at the same
|
||||||
|
|
|
@ -881,7 +881,7 @@ void lru_cache_disable(void)
|
||||||
* lru_disable_count = 0 will have exited the critical
|
* lru_disable_count = 0 will have exited the critical
|
||||||
* section when synchronize_rcu() returns.
|
* section when synchronize_rcu() returns.
|
||||||
*/
|
*/
|
||||||
synchronize_rcu();
|
synchronize_rcu_expedited();
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
__lru_add_drain_all(true);
|
__lru_add_drain_all(true);
|
||||||
#else
|
#else
|
||||||
|
|
Loading…
Reference in a new issue