mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-11 21:38:32 +08:00
mm/hugetlb: convert restore_reserve_on_error() to folios
Use the hugetlb folio flag macros inside restore_reserve_on_error() and update the comments to reflect the use of folios. Link: https://lkml.kernel.org/r/20230113223057.173292-8-sidhartha.kumar@oracle.com Signed-off-by: Sidhartha Kumar <sidhartha.kumar@oracle.com> Reviewed-by: Mike Kravetz <mike.kravetz@oracle.com> Cc: John Hubbard <jhubbard@nvidia.com> Cc: Matthew Wilcox <willy@infradead.org> Cc: Muchun Song <songmuchun@bytedance.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
e37d3e838d
commit
0ffdc38eb5
27
mm/hugetlb.c
27
mm/hugetlb.c
@ -2819,22 +2819,23 @@ static long vma_del_reservation(struct hstate *h,
|
|||||||
void restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma,
|
void restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma,
|
||||||
unsigned long address, struct page *page)
|
unsigned long address, struct page *page)
|
||||||
{
|
{
|
||||||
|
struct folio *folio = page_folio(page);
|
||||||
long rc = vma_needs_reservation(h, vma, address);
|
long rc = vma_needs_reservation(h, vma, address);
|
||||||
|
|
||||||
if (HPageRestoreReserve(page)) {
|
if (folio_test_hugetlb_restore_reserve(folio)) {
|
||||||
if (unlikely(rc < 0))
|
if (unlikely(rc < 0))
|
||||||
/*
|
/*
|
||||||
* Rare out of memory condition in reserve map
|
* Rare out of memory condition in reserve map
|
||||||
* manipulation. Clear HPageRestoreReserve so that
|
* manipulation. Clear hugetlb_restore_reserve so
|
||||||
* global reserve count will not be incremented
|
* that global reserve count will not be incremented
|
||||||
* by free_huge_page. This will make it appear
|
* by free_huge_page. This will make it appear
|
||||||
* as though the reservation for this page was
|
* as though the reservation for this folio was
|
||||||
* consumed. This may prevent the task from
|
* consumed. This may prevent the task from
|
||||||
* faulting in the page at a later time. This
|
* faulting in the folio at a later time. This
|
||||||
* is better than inconsistent global huge page
|
* is better than inconsistent global huge page
|
||||||
* accounting of reserve counts.
|
* accounting of reserve counts.
|
||||||
*/
|
*/
|
||||||
ClearHPageRestoreReserve(page);
|
folio_clear_hugetlb_restore_reserve(folio);
|
||||||
else if (rc)
|
else if (rc)
|
||||||
(void)vma_add_reservation(h, vma, address);
|
(void)vma_add_reservation(h, vma, address);
|
||||||
else
|
else
|
||||||
@ -2845,7 +2846,7 @@ void restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma,
|
|||||||
* This indicates there is an entry in the reserve map
|
* This indicates there is an entry in the reserve map
|
||||||
* not added by alloc_huge_page. We know it was added
|
* not added by alloc_huge_page. We know it was added
|
||||||
* before the alloc_huge_page call, otherwise
|
* before the alloc_huge_page call, otherwise
|
||||||
* HPageRestoreReserve would be set on the page.
|
* hugetlb_restore_reserve would be set on the folio.
|
||||||
* Remove the entry so that a subsequent allocation
|
* Remove the entry so that a subsequent allocation
|
||||||
* does not consume a reservation.
|
* does not consume a reservation.
|
||||||
*/
|
*/
|
||||||
@ -2854,12 +2855,12 @@ void restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma,
|
|||||||
/*
|
/*
|
||||||
* VERY rare out of memory condition. Since
|
* VERY rare out of memory condition. Since
|
||||||
* we can not delete the entry, set
|
* we can not delete the entry, set
|
||||||
* HPageRestoreReserve so that the reserve
|
* hugetlb_restore_reserve so that the reserve
|
||||||
* count will be incremented when the page
|
* count will be incremented when the folio
|
||||||
* is freed. This reserve will be consumed
|
* is freed. This reserve will be consumed
|
||||||
* on a subsequent allocation.
|
* on a subsequent allocation.
|
||||||
*/
|
*/
|
||||||
SetHPageRestoreReserve(page);
|
folio_set_hugetlb_restore_reserve(folio);
|
||||||
} else if (rc < 0) {
|
} else if (rc < 0) {
|
||||||
/*
|
/*
|
||||||
* Rare out of memory condition from
|
* Rare out of memory condition from
|
||||||
@ -2875,12 +2876,12 @@ void restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma,
|
|||||||
/*
|
/*
|
||||||
* For private mappings, no entry indicates
|
* For private mappings, no entry indicates
|
||||||
* a reservation is present. Since we can
|
* a reservation is present. Since we can
|
||||||
* not add an entry, set SetHPageRestoreReserve
|
* not add an entry, set hugetlb_restore_reserve
|
||||||
* on the page so reserve count will be
|
* on the folio so reserve count will be
|
||||||
* incremented when freed. This reserve will
|
* incremented when freed. This reserve will
|
||||||
* be consumed on a subsequent allocation.
|
* be consumed on a subsequent allocation.
|
||||||
*/
|
*/
|
||||||
SetHPageRestoreReserve(page);
|
folio_set_hugetlb_restore_reserve(folio);
|
||||||
} else
|
} else
|
||||||
/*
|
/*
|
||||||
* No reservation present, do nothing
|
* No reservation present, do nothing
|
||||||
|
Loading…
Reference in New Issue
Block a user