mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-02 08:34:20 +08:00
mm/memory-failure: convert free_raw_hwp_pages() to folios
Change free_raw_hwp_pages() to folio_free_raw_hwp(), converts two users of hugetlb specific page macro users to their folio equivalents. Link: https://lkml.kernel.org/r/20230112204608.80136-5-sidhartha.kumar@oracle.com Signed-off-by: Sidhartha Kumar <sidhartha.kumar@oracle.com> Acked-by: Naoya Horiguchi <naoya.horiguchi@nec.com> Cc: Matthew Wilcox <willy@infradead.org> Cc: Miaohe Lin <linmiaohe@huawei.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
2ff6cecee6
commit
9637d7dfb1
@ -1766,23 +1766,23 @@ static int hugetlb_set_page_hwpoison(struct page *hpage, struct page *page)
|
||||
return ret;
|
||||
}
|
||||
|
||||
static unsigned long free_raw_hwp_pages(struct page *hpage, bool move_flag)
|
||||
static unsigned long folio_free_raw_hwp(struct folio *folio, bool move_flag)
|
||||
{
|
||||
/*
|
||||
* HPageVmemmapOptimized hugepages can't be freed because struct
|
||||
* hugetlb_vmemmap_optimized hugepages can't be freed because struct
|
||||
* pages for tail pages are required but they don't exist.
|
||||
*/
|
||||
if (move_flag && HPageVmemmapOptimized(hpage))
|
||||
if (move_flag && folio_test_hugetlb_vmemmap_optimized(folio))
|
||||
return 0;
|
||||
|
||||
/*
|
||||
* HPageRawHwpUnreliable hugepages shouldn't be unpoisoned by
|
||||
* hugetlb_raw_hwp_unreliable hugepages shouldn't be unpoisoned by
|
||||
* definition.
|
||||
*/
|
||||
if (HPageRawHwpUnreliable(hpage))
|
||||
if (folio_test_hugetlb_raw_hwp_unreliable(folio))
|
||||
return 0;
|
||||
|
||||
return __free_raw_hwp_pages(hpage, move_flag);
|
||||
return __free_raw_hwp_pages(&folio->page, move_flag);
|
||||
}
|
||||
|
||||
void folio_clear_hugetlb_hwpoison(struct folio *folio)
|
||||
@ -1790,7 +1790,7 @@ void folio_clear_hugetlb_hwpoison(struct folio *folio)
|
||||
if (folio_test_hugetlb_raw_hwp_unreliable(folio))
|
||||
return;
|
||||
folio_clear_hwpoison(folio);
|
||||
free_raw_hwp_pages(&folio->page, true);
|
||||
folio_free_raw_hwp(folio, true);
|
||||
}
|
||||
|
||||
/*
|
||||
@ -1929,7 +1929,7 @@ static inline int try_memory_failure_hugetlb(unsigned long pfn, int flags, int *
|
||||
return 0;
|
||||
}
|
||||
|
||||
static inline unsigned long free_raw_hwp_pages(struct page *hpage, bool flag)
|
||||
static inline unsigned long folio_free_raw_hwp(struct folio *folio, bool flag)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
@ -2336,6 +2336,7 @@ core_initcall(memory_failure_init);
|
||||
int unpoison_memory(unsigned long pfn)
|
||||
{
|
||||
struct page *page;
|
||||
struct folio *folio;
|
||||
struct page *p;
|
||||
int ret = -EBUSY;
|
||||
unsigned long count = 1;
|
||||
@ -2348,6 +2349,7 @@ int unpoison_memory(unsigned long pfn)
|
||||
|
||||
p = pfn_to_page(pfn);
|
||||
page = compound_head(p);
|
||||
folio = page_folio(p);
|
||||
|
||||
mutex_lock(&mf_mutex);
|
||||
|
||||
@ -2389,7 +2391,7 @@ int unpoison_memory(unsigned long pfn)
|
||||
if (!ret) {
|
||||
if (PageHuge(p)) {
|
||||
huge = true;
|
||||
count = free_raw_hwp_pages(page, false);
|
||||
count = folio_free_raw_hwp(folio, false);
|
||||
if (count == 0) {
|
||||
ret = -EBUSY;
|
||||
goto unlock_mutex;
|
||||
@ -2405,7 +2407,7 @@ int unpoison_memory(unsigned long pfn)
|
||||
} else {
|
||||
if (PageHuge(p)) {
|
||||
huge = true;
|
||||
count = free_raw_hwp_pages(page, false);
|
||||
count = folio_free_raw_hwp(folio, false);
|
||||
if (count == 0) {
|
||||
ret = -EBUSY;
|
||||
put_page(page);
|
||||
|
Loading…
Reference in New Issue
Block a user