mm/hwpoison: fix race between hugetlb free/demotion and memory_failure_hugetlb()

There is a race condition between memory_failure_hugetlb() and hugetlb
free/demotion, which causes setting PageHWPoison flag on the wrong page.
The one simple result is that wrong processes can be killed, but another
(more serious) one is that the actual error is left unhandled, so no one
prevents later access to it, and that might lead to more serious results
like consuming corrupted data.

Think about the below race window:

  CPU 1                                   CPU 2
  memory_failure_hugetlb
  struct page *head = compound_head(p);
                                          hugetlb page might be freed to
                                          buddy, or even changed to another
                                          compound page.

  get_hwpoison_page -- page is not what we want now...

The current code first does prechecks roughly and then reconfirms after
taking refcount, but it's found that it makes code overly complicated,
so move the prechecks in a single hugetlb_lock range.

A newly introduced function, try_memory_failure_hugetlb(), always takes
hugetlb_lock (even for non-hugetlb pages).  That can be improved, but
memory_failure() is rare in principle, so should not be a big problem.

Link: https://lkml.kernel.org/r/20220408135323.1559401-2-naoya.horiguchi@linux.dev
Fixes: 761ad8d7c7 ("mm: hwpoison: introduce memory_failure_hugetlb()")
Signed-off-by: Naoya Horiguchi <naoya.horiguchi@nec.com>
Reported-by: Mike Kravetz <mike.kravetz@oracle.com>
Reviewed-by: Miaohe Lin <linmiaohe@huawei.com>
Reviewed-by: Mike Kravetz <mike.kravetz@oracle.com>
Cc: Yang Shi <shy828301@gmail.com>
Cc: Dan Carpenter <dan.carpenter@oracle.com>
Cc: <stable@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
Naoya Horiguchi 2022-04-21 16:35:33 -07:00 committed by Linus Torvalds
parent b253435746
commit 405ce05123
4 changed files with 128 additions and 43 deletions

View File

@ -169,6 +169,7 @@ long hugetlb_unreserve_pages(struct inode *inode, long start, long end,
long freed); long freed);
bool isolate_huge_page(struct page *page, struct list_head *list); bool isolate_huge_page(struct page *page, struct list_head *list);
int get_hwpoison_huge_page(struct page *page, bool *hugetlb); int get_hwpoison_huge_page(struct page *page, bool *hugetlb);
int get_huge_page_for_hwpoison(unsigned long pfn, int flags);
void putback_active_hugepage(struct page *page); void putback_active_hugepage(struct page *page);
void move_hugetlb_state(struct page *oldpage, struct page *newpage, int reason); void move_hugetlb_state(struct page *oldpage, struct page *newpage, int reason);
void free_huge_page(struct page *page); void free_huge_page(struct page *page);
@ -378,6 +379,11 @@ static inline int get_hwpoison_huge_page(struct page *page, bool *hugetlb)
return 0; return 0;
} }
static inline int get_huge_page_for_hwpoison(unsigned long pfn, int flags)
{
return 0;
}
static inline void putback_active_hugepage(struct page *page) static inline void putback_active_hugepage(struct page *page)
{ {
} }

View File

@ -3197,6 +3197,14 @@ extern int sysctl_memory_failure_recovery;
extern void shake_page(struct page *p); extern void shake_page(struct page *p);
extern atomic_long_t num_poisoned_pages __read_mostly; extern atomic_long_t num_poisoned_pages __read_mostly;
extern int soft_offline_page(unsigned long pfn, int flags); extern int soft_offline_page(unsigned long pfn, int flags);
#ifdef CONFIG_MEMORY_FAILURE
extern int __get_huge_page_for_hwpoison(unsigned long pfn, int flags);
#else
static inline int __get_huge_page_for_hwpoison(unsigned long pfn, int flags)
{
return 0;
}
#endif
#ifndef arch_memory_failure #ifndef arch_memory_failure
static inline int arch_memory_failure(unsigned long pfn, int flags) static inline int arch_memory_failure(unsigned long pfn, int flags)

View File

@ -6785,6 +6785,16 @@ int get_hwpoison_huge_page(struct page *page, bool *hugetlb)
return ret; return ret;
} }
int get_huge_page_for_hwpoison(unsigned long pfn, int flags)
{
int ret;
spin_lock_irq(&hugetlb_lock);
ret = __get_huge_page_for_hwpoison(pfn, flags);
spin_unlock_irq(&hugetlb_lock);
return ret;
}
void putback_active_hugepage(struct page *page) void putback_active_hugepage(struct page *page)
{ {
spin_lock_irq(&hugetlb_lock); spin_lock_irq(&hugetlb_lock);

View File

@ -1498,50 +1498,113 @@ static int try_to_split_thp_page(struct page *page, const char *msg)
return 0; return 0;
} }
static int memory_failure_hugetlb(unsigned long pfn, int flags) /*
* Called from hugetlb code with hugetlb_lock held.
*
* Return values:
* 0 - free hugepage
* 1 - in-use hugepage
* 2 - not a hugepage
* -EBUSY - the hugepage is busy (try to retry)
* -EHWPOISON - the hugepage is already hwpoisoned
*/
int __get_huge_page_for_hwpoison(unsigned long pfn, int flags)
{ {
struct page *p = pfn_to_page(pfn); struct page *page = pfn_to_page(pfn);
struct page *head = compound_head(p); struct page *head = compound_head(page);
int res; int ret = 2; /* fallback to normal page handling */
unsigned long page_flags; bool count_increased = false;
if (!PageHeadHuge(head))
goto out;
if (flags & MF_COUNT_INCREASED) {
ret = 1;
count_increased = true;
} else if (HPageFreed(head) || HPageMigratable(head)) {
ret = get_page_unless_zero(head);
if (ret)
count_increased = true;
} else {
ret = -EBUSY;
goto out;
}
if (TestSetPageHWPoison(head)) { if (TestSetPageHWPoison(head)) {
pr_err("Memory failure: %#lx: already hardware poisoned\n", ret = -EHWPOISON;
pfn); goto out;
res = -EHWPOISON; }
if (flags & MF_ACTION_REQUIRED)
return ret;
out:
if (count_increased)
put_page(head);
return ret;
}
#ifdef CONFIG_HUGETLB_PAGE
/*
* Taking refcount of hugetlb pages needs extra care about race conditions
* with basic operations like hugepage allocation/free/demotion.
* So some of prechecks for hwpoison (pinning, and testing/setting
* PageHWPoison) should be done in single hugetlb_lock range.
*/
static int try_memory_failure_hugetlb(unsigned long pfn, int flags, int *hugetlb)
{
int res;
struct page *p = pfn_to_page(pfn);
struct page *head;
unsigned long page_flags;
bool retry = true;
*hugetlb = 1;
retry:
res = get_huge_page_for_hwpoison(pfn, flags);
if (res == 2) { /* fallback to normal page handling */
*hugetlb = 0;
return 0;
} else if (res == -EHWPOISON) {
pr_err("Memory failure: %#lx: already hardware poisoned\n", pfn);
if (flags & MF_ACTION_REQUIRED) {
head = compound_head(p);
res = kill_accessing_process(current, page_to_pfn(head), flags); res = kill_accessing_process(current, page_to_pfn(head), flags);
}
return res; return res;
} else if (res == -EBUSY) {
if (retry) {
retry = false;
goto retry;
}
action_result(pfn, MF_MSG_UNKNOWN, MF_IGNORED);
return res;
}
head = compound_head(p);
lock_page(head);
if (hwpoison_filter(p)) {
ClearPageHWPoison(head);
res = -EOPNOTSUPP;
goto out;
} }
num_poisoned_pages_inc(); num_poisoned_pages_inc();
if (!(flags & MF_COUNT_INCREASED)) { /*
res = get_hwpoison_page(p, flags); * Handling free hugepage. The possible race with hugepage allocation
if (!res) { * or demotion can be prevented by PageHWPoison flag.
lock_page(head); */
if (hwpoison_filter(p)) { if (res == 0) {
if (TestClearPageHWPoison(head)) unlock_page(head);
num_poisoned_pages_dec(); res = MF_FAILED;
unlock_page(head); if (__page_handle_poison(p)) {
return -EOPNOTSUPP; page_ref_inc(p);
} res = MF_RECOVERED;
unlock_page(head);
res = MF_FAILED;
if (__page_handle_poison(p)) {
page_ref_inc(p);
res = MF_RECOVERED;
}
action_result(pfn, MF_MSG_FREE_HUGE, res);
return res == MF_RECOVERED ? 0 : -EBUSY;
} else if (res < 0) {
action_result(pfn, MF_MSG_UNKNOWN, MF_IGNORED);
return -EBUSY;
} }
action_result(pfn, MF_MSG_FREE_HUGE, res);
return res == MF_RECOVERED ? 0 : -EBUSY;
} }
lock_page(head);
/* /*
* The page could have changed compound pages due to race window. * The page could have changed compound pages due to race window.
* If this happens just bail out. * If this happens just bail out.
@ -1554,14 +1617,6 @@ static int memory_failure_hugetlb(unsigned long pfn, int flags)
page_flags = head->flags; page_flags = head->flags;
if (hwpoison_filter(p)) {
if (TestClearPageHWPoison(head))
num_poisoned_pages_dec();
put_page(p);
res = -EOPNOTSUPP;
goto out;
}
/* /*
* TODO: hwpoison for pud-sized hugetlb doesn't work right now, so * TODO: hwpoison for pud-sized hugetlb doesn't work right now, so
* simply disable it. In order to make it work properly, we need * simply disable it. In order to make it work properly, we need
@ -1588,6 +1643,12 @@ static int memory_failure_hugetlb(unsigned long pfn, int flags)
unlock_page(head); unlock_page(head);
return res; return res;
} }
#else
static inline int try_memory_failure_hugetlb(unsigned long pfn, int flags, int *hugetlb)
{
return 0;
}
#endif
static int memory_failure_dev_pagemap(unsigned long pfn, int flags, static int memory_failure_dev_pagemap(unsigned long pfn, int flags,
struct dev_pagemap *pgmap) struct dev_pagemap *pgmap)
@ -1712,6 +1773,7 @@ int memory_failure(unsigned long pfn, int flags)
int res = 0; int res = 0;
unsigned long page_flags; unsigned long page_flags;
bool retry = true; bool retry = true;
int hugetlb = 0;
if (!sysctl_memory_failure_recovery) if (!sysctl_memory_failure_recovery)
panic("Memory failure on page %lx", pfn); panic("Memory failure on page %lx", pfn);
@ -1739,10 +1801,9 @@ int memory_failure(unsigned long pfn, int flags)
} }
try_again: try_again:
if (PageHuge(p)) { res = try_memory_failure_hugetlb(pfn, flags, &hugetlb);
res = memory_failure_hugetlb(pfn, flags); if (hugetlb)
goto unlock_mutex; goto unlock_mutex;
}
if (TestSetPageHWPoison(p)) { if (TestSetPageHWPoison(p)) {
pr_err("Memory failure: %#lx: already hardware poisoned\n", pr_err("Memory failure: %#lx: already hardware poisoned\n",