mirror of
https://github.com/torvalds/linux.git
synced 2026-03-08 00:44:31 +01:00
While compaction depends on migration, the other direction is not the case. So let's make it clearer that this is all about migration of balloon pages. Adjust all comments/docs in the core to talk about "migration" instead of "compaction". While at it add some "/* CONFIG_BALLOON_MIGRATION */". Link: https://lkml.kernel.org/r/20260119230133.3551867-23-david@kernel.org Signed-off-by: David Hildenbrand (Red Hat) <david@kernel.org> Reviewed-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com> Acked-by: Michael S. Tsirkin <mst@redhat.com> Cc: Arnd Bergmann <arnd@arndb.de> Cc: Christophe Leroy <christophe.leroy@csgroup.eu> Cc: Eugenio Pérez <eperezma@redhat.com> Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org> Cc: Jason Wang <jasowang@redhat.com> Cc: Jerrin Shaji George <jerrin.shaji-george@broadcom.com> Cc: Jonathan Corbet <corbet@lwn.net> Cc: Liam Howlett <liam.howlett@oracle.com> Cc: Madhavan Srinivasan <maddy@linux.ibm.com> Cc: Michael Ellerman <mpe@ellerman.id.au> Cc: Michal Hocko <mhocko@suse.com> Cc: Mike Rapoport <rppt@kernel.org> Cc: Nicholas Piggin <npiggin@gmail.com> Cc: Oscar Salvador <osalvador@suse.de> Cc: SeongJae Park <sj@kernel.org> Cc: Suren Baghdasaryan <surenb@google.com> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: Xuan Zhuo <xuanzhuo@linux.alibaba.com> Cc: Zi Yan <ziy@nvidia.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
344 lines
10 KiB
C
344 lines
10 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Common interface for implementing a memory balloon, including support
|
|
* for migration of pages inflated in a memory balloon.
|
|
*
|
|
* Copyright (C) 2012, Red Hat, Inc. Rafael Aquini <aquini@redhat.com>
|
|
*/
|
|
#include <linux/mm.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/export.h>
|
|
#include <linux/balloon.h>
|
|
|
|
/*
|
|
* Lock protecting the balloon_dev_info of all devices. We don't really
|
|
* expect more than one device.
|
|
*/
|
|
static DEFINE_SPINLOCK(balloon_pages_lock);
|
|
|
|
/**
|
|
* balloon_page_insert - insert a page into the balloon's page list and make
|
|
* the page->private assignment accordingly.
|
|
* @balloon : pointer to balloon device
|
|
* @page : page to be assigned as a 'balloon page'
|
|
*
|
|
* Caller must ensure the balloon_pages_lock is held.
|
|
*/
|
|
static void balloon_page_insert(struct balloon_dev_info *balloon,
|
|
struct page *page)
|
|
{
|
|
lockdep_assert_held(&balloon_pages_lock);
|
|
__SetPageOffline(page);
|
|
if (IS_ENABLED(CONFIG_BALLOON_MIGRATION)) {
|
|
SetPageMovableOps(page);
|
|
set_page_private(page, (unsigned long)balloon);
|
|
}
|
|
list_add(&page->lru, &balloon->pages);
|
|
}
|
|
|
|
/**
|
|
* balloon_page_finalize - prepare a balloon page that was removed from the
|
|
* balloon list for release to the page allocator
|
|
* @page: page to be released to the page allocator
|
|
*
|
|
* Caller must ensure the balloon_pages_lock is held.
|
|
*/
|
|
static void balloon_page_finalize(struct page *page)
|
|
{
|
|
lockdep_assert_held(&balloon_pages_lock);
|
|
if (IS_ENABLED(CONFIG_BALLOON_MIGRATION))
|
|
set_page_private(page, 0);
|
|
/* PageOffline is sticky until the page is freed to the buddy. */
|
|
}
|
|
|
|
static void balloon_page_enqueue_one(struct balloon_dev_info *b_dev_info,
|
|
struct page *page)
|
|
{
|
|
balloon_page_insert(b_dev_info, page);
|
|
if (b_dev_info->adjust_managed_page_count)
|
|
adjust_managed_page_count(page, -1);
|
|
__count_vm_event(BALLOON_INFLATE);
|
|
inc_node_page_state(page, NR_BALLOON_PAGES);
|
|
}
|
|
|
|
/**
|
|
* balloon_page_list_enqueue() - inserts a list of pages into the balloon page
|
|
* list.
|
|
* @b_dev_info: balloon device descriptor where we will insert a new page to
|
|
* @pages: pages to enqueue - allocated using balloon_page_alloc.
|
|
*
|
|
* Driver must call this function to properly enqueue balloon pages before
|
|
* definitively removing them from the guest system.
|
|
*
|
|
* Return: number of pages that were enqueued.
|
|
*/
|
|
size_t balloon_page_list_enqueue(struct balloon_dev_info *b_dev_info,
|
|
struct list_head *pages)
|
|
{
|
|
struct page *page, *tmp;
|
|
unsigned long flags;
|
|
size_t n_pages = 0;
|
|
|
|
spin_lock_irqsave(&balloon_pages_lock, flags);
|
|
list_for_each_entry_safe(page, tmp, pages, lru) {
|
|
list_del(&page->lru);
|
|
balloon_page_enqueue_one(b_dev_info, page);
|
|
n_pages++;
|
|
}
|
|
spin_unlock_irqrestore(&balloon_pages_lock, flags);
|
|
return n_pages;
|
|
}
|
|
EXPORT_SYMBOL_GPL(balloon_page_list_enqueue);
|
|
|
|
/**
|
|
* balloon_page_list_dequeue() - removes pages from balloon's page list and
|
|
* returns a list of the pages.
|
|
* @b_dev_info: balloon device descriptor where we will grab a page from.
|
|
* @pages: pointer to the list of pages that would be returned to the caller.
|
|
* @n_req_pages: number of requested pages.
|
|
*
|
|
* Driver must call this function to properly de-allocate a previous enlisted
|
|
* balloon pages before definitively releasing it back to the guest system.
|
|
* This function tries to remove @n_req_pages from the ballooned pages and
|
|
* return them to the caller in the @pages list.
|
|
*
|
|
* Note that this function may fail to dequeue some pages even if the balloon
|
|
* isn't empty - since the page list can be temporarily empty due to compaction
|
|
* of isolated pages.
|
|
*
|
|
* Return: number of pages that were added to the @pages list.
|
|
*/
|
|
size_t balloon_page_list_dequeue(struct balloon_dev_info *b_dev_info,
|
|
struct list_head *pages, size_t n_req_pages)
|
|
{
|
|
struct page *page, *tmp;
|
|
unsigned long flags;
|
|
size_t n_pages = 0;
|
|
|
|
spin_lock_irqsave(&balloon_pages_lock, flags);
|
|
list_for_each_entry_safe(page, tmp, &b_dev_info->pages, lru) {
|
|
if (n_pages == n_req_pages)
|
|
break;
|
|
list_del(&page->lru);
|
|
if (b_dev_info->adjust_managed_page_count)
|
|
adjust_managed_page_count(page, 1);
|
|
balloon_page_finalize(page);
|
|
__count_vm_event(BALLOON_DEFLATE);
|
|
list_add(&page->lru, pages);
|
|
dec_node_page_state(page, NR_BALLOON_PAGES);
|
|
n_pages++;
|
|
}
|
|
spin_unlock_irqrestore(&balloon_pages_lock, flags);
|
|
|
|
return n_pages;
|
|
}
|
|
EXPORT_SYMBOL_GPL(balloon_page_list_dequeue);
|
|
|
|
/**
|
|
* balloon_page_alloc - allocates a new page for insertion into the balloon
|
|
* page list.
|
|
*
|
|
* Driver must call this function to properly allocate a new balloon page.
|
|
* Driver must call balloon_page_enqueue before definitively removing the page
|
|
* from the guest system.
|
|
*
|
|
* Return: struct page for the allocated page or NULL on allocation failure.
|
|
*/
|
|
struct page *balloon_page_alloc(void)
|
|
{
|
|
gfp_t gfp_flags = __GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN;
|
|
|
|
if (IS_ENABLED(CONFIG_BALLOON_MIGRATION))
|
|
gfp_flags |= GFP_HIGHUSER_MOVABLE;
|
|
else
|
|
gfp_flags |= GFP_HIGHUSER;
|
|
|
|
return alloc_page(gfp_flags);
|
|
}
|
|
EXPORT_SYMBOL_GPL(balloon_page_alloc);
|
|
|
|
/**
|
|
* balloon_page_enqueue - inserts a new page into the balloon page list.
|
|
*
|
|
* @b_dev_info: balloon device descriptor where we will insert a new page
|
|
* @page: new page to enqueue - allocated using balloon_page_alloc.
|
|
*
|
|
* Drivers must call this function to properly enqueue a new allocated balloon
|
|
* page before definitively removing the page from the guest system.
|
|
*
|
|
* Drivers must not enqueue pages while page->lru is still in
|
|
* use, and must not use page->lru until a page was unqueued again.
|
|
*/
|
|
void balloon_page_enqueue(struct balloon_dev_info *b_dev_info,
|
|
struct page *page)
|
|
{
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&balloon_pages_lock, flags);
|
|
balloon_page_enqueue_one(b_dev_info, page);
|
|
spin_unlock_irqrestore(&balloon_pages_lock, flags);
|
|
}
|
|
EXPORT_SYMBOL_GPL(balloon_page_enqueue);
|
|
|
|
/**
|
|
* balloon_page_dequeue - removes a page from balloon's page list and returns
|
|
* its address to allow the driver to release the page.
|
|
* @b_dev_info: balloon device descriptor where we will grab a page from.
|
|
*
|
|
* Driver must call this function to properly dequeue a previously enqueued page
|
|
* before definitively releasing it back to the guest system.
|
|
*
|
|
* Caller must perform its own accounting to ensure that this
|
|
* function is called only if some pages are actually enqueued.
|
|
*
|
|
* Note that this function may fail to dequeue some pages even if there are
|
|
* some enqueued pages - since the page list can be temporarily empty due to
|
|
* the compaction of isolated pages.
|
|
*
|
|
* TODO: remove the caller accounting requirements, and allow caller to wait
|
|
* until all pages can be dequeued.
|
|
*
|
|
* Return: struct page for the dequeued page, or NULL if no page was dequeued.
|
|
*/
|
|
struct page *balloon_page_dequeue(struct balloon_dev_info *b_dev_info)
|
|
{
|
|
unsigned long flags;
|
|
LIST_HEAD(pages);
|
|
int n_pages;
|
|
|
|
n_pages = balloon_page_list_dequeue(b_dev_info, &pages, 1);
|
|
|
|
if (n_pages != 1) {
|
|
/*
|
|
* If we are unable to dequeue a balloon page because the page
|
|
* list is empty and there are no isolated pages, then something
|
|
* went out of track and some balloon pages are lost.
|
|
* BUG() here, otherwise the balloon driver may get stuck in
|
|
* an infinite loop while attempting to release all its pages.
|
|
*/
|
|
spin_lock_irqsave(&balloon_pages_lock, flags);
|
|
if (unlikely(list_empty(&b_dev_info->pages) &&
|
|
!b_dev_info->isolated_pages))
|
|
BUG();
|
|
spin_unlock_irqrestore(&balloon_pages_lock, flags);
|
|
return NULL;
|
|
}
|
|
return list_first_entry(&pages, struct page, lru);
|
|
}
|
|
EXPORT_SYMBOL_GPL(balloon_page_dequeue);
|
|
|
|
#ifdef CONFIG_BALLOON_MIGRATION
|
|
static struct balloon_dev_info *balloon_page_device(struct page *page)
|
|
{
|
|
return (struct balloon_dev_info *)page_private(page);
|
|
}
|
|
|
|
static bool balloon_page_isolate(struct page *page, isolate_mode_t mode)
|
|
|
|
{
|
|
struct balloon_dev_info *b_dev_info;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&balloon_pages_lock, flags);
|
|
b_dev_info = balloon_page_device(page);
|
|
if (!b_dev_info) {
|
|
/*
|
|
* The page already got deflated and removed from the
|
|
* balloon list.
|
|
*/
|
|
spin_unlock_irqrestore(&balloon_pages_lock, flags);
|
|
return false;
|
|
}
|
|
list_del(&page->lru);
|
|
b_dev_info->isolated_pages++;
|
|
spin_unlock_irqrestore(&balloon_pages_lock, flags);
|
|
|
|
return true;
|
|
}
|
|
|
|
static void balloon_page_putback(struct page *page)
|
|
{
|
|
struct balloon_dev_info *b_dev_info = balloon_page_device(page);
|
|
unsigned long flags;
|
|
|
|
/*
|
|
* When we isolated the page, the page was still inflated in a balloon
|
|
* device. As isolated balloon pages cannot get deflated, we still have
|
|
* a balloon device here.
|
|
*/
|
|
if (WARN_ON_ONCE(!b_dev_info))
|
|
return;
|
|
|
|
spin_lock_irqsave(&balloon_pages_lock, flags);
|
|
list_add(&page->lru, &b_dev_info->pages);
|
|
b_dev_info->isolated_pages--;
|
|
spin_unlock_irqrestore(&balloon_pages_lock, flags);
|
|
}
|
|
|
|
static int balloon_page_migrate(struct page *newpage, struct page *page,
|
|
enum migrate_mode mode)
|
|
{
|
|
struct balloon_dev_info *b_dev_info = balloon_page_device(page);
|
|
unsigned long flags;
|
|
int rc;
|
|
|
|
/*
|
|
* When we isolated the page, the page was still inflated in a balloon
|
|
* device. As isolated balloon pages cannot get deflated, we still have
|
|
* a balloon device here.
|
|
*/
|
|
if (WARN_ON_ONCE(!b_dev_info))
|
|
return -EAGAIN;
|
|
|
|
rc = b_dev_info->migratepage(b_dev_info, newpage, page, mode);
|
|
if (rc < 0 && rc != -ENOENT)
|
|
return rc;
|
|
|
|
spin_lock_irqsave(&balloon_pages_lock, flags);
|
|
if (!rc) {
|
|
/* Insert the new page into the balloon list. */
|
|
get_page(newpage);
|
|
balloon_page_insert(b_dev_info, newpage);
|
|
__count_vm_event(BALLOON_MIGRATE);
|
|
|
|
if (b_dev_info->adjust_managed_page_count &&
|
|
page_zone(page) != page_zone(newpage)) {
|
|
/*
|
|
* When we migrate a page to a different zone we
|
|
* have to fixup the count of both involved zones.
|
|
*/
|
|
adjust_managed_page_count(page, 1);
|
|
adjust_managed_page_count(newpage, -1);
|
|
}
|
|
} else {
|
|
/* Old page was deflated but new page not inflated. */
|
|
__count_vm_event(BALLOON_DEFLATE);
|
|
|
|
if (b_dev_info->adjust_managed_page_count)
|
|
adjust_managed_page_count(page, 1);
|
|
}
|
|
|
|
b_dev_info->isolated_pages--;
|
|
|
|
/* Free the now-deflated page we isolated in balloon_page_isolate(). */
|
|
balloon_page_finalize(page);
|
|
spin_unlock_irqrestore(&balloon_pages_lock, flags);
|
|
|
|
put_page(page);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct movable_operations balloon_mops = {
|
|
.migrate_page = balloon_page_migrate,
|
|
.isolate_page = balloon_page_isolate,
|
|
.putback_page = balloon_page_putback,
|
|
};
|
|
|
|
static int __init balloon_init(void)
|
|
{
|
|
return set_movable_ops(&balloon_mops, PGTY_offline);
|
|
}
|
|
core_initcall(balloon_init);
|
|
|
|
#endif /* CONFIG_BALLOON_MIGRATION */
|