119 lines
3.0 KiB
C
119 lines
3.0 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
|
|
/******************************************************************************
|
|
* Xen memory reservation utilities.
|
|
*
|
|
* Copyright (c) 2003, B Dragovic
|
|
* Copyright (c) 2003-2004, M Williamson, K Fraser
|
|
* Copyright (c) 2005 Dan M. Smith, IBM Corporation
|
|
* Copyright (c) 2010 Daniel Kiper
|
|
* Copyright (c) 2018 Oleksandr Andrushchenko, EPAM Systems Inc.
|
|
*/
|
|
|
|
#include <asm/xen/hypercall.h>
|
|
|
|
#include <xen/interface/memory.h>
|
|
#include <xen/mem-reservation.h>
|
|
|
|
/*
|
|
* Use one extent per PAGE_SIZE to avoid to break down the page into
|
|
* multiple frame.
|
|
*/
|
|
#define EXTENT_ORDER (fls(XEN_PFN_PER_PAGE) - 1)
|
|
|
|
#ifdef CONFIG_XEN_HAVE_PVMMU
|
|
void __xenmem_reservation_va_mapping_update(unsigned long count,
|
|
struct page **pages,
|
|
xen_pfn_t *frames)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < count; i++) {
|
|
struct page *page = pages[i];
|
|
unsigned long pfn = page_to_pfn(page);
|
|
|
|
BUG_ON(!page);
|
|
|
|
/*
|
|
* We don't support PV MMU when Linux and Xen is using
|
|
* different page granularity.
|
|
*/
|
|
BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE);
|
|
|
|
set_phys_to_machine(pfn, frames[i]);
|
|
|
|
/* Link back into the page tables if not highmem. */
|
|
if (!PageHighMem(page)) {
|
|
int ret;
|
|
|
|
ret = HYPERVISOR_update_va_mapping(
|
|
(unsigned long)__va(pfn << PAGE_SHIFT),
|
|
mfn_pte(frames[i], PAGE_KERNEL),
|
|
0);
|
|
BUG_ON(ret);
|
|
}
|
|
}
|
|
}
|
|
EXPORT_SYMBOL_GPL(__xenmem_reservation_va_mapping_update);
|
|
|
|
void __xenmem_reservation_va_mapping_reset(unsigned long count,
|
|
struct page **pages)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < count; i++) {
|
|
struct page *page = pages[i];
|
|
unsigned long pfn = page_to_pfn(page);
|
|
|
|
/*
|
|
* We don't support PV MMU when Linux and Xen are using
|
|
* different page granularity.
|
|
*/
|
|
BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE);
|
|
|
|
if (!PageHighMem(page)) {
|
|
int ret;
|
|
|
|
ret = HYPERVISOR_update_va_mapping(
|
|
(unsigned long)__va(pfn << PAGE_SHIFT),
|
|
__pte_ma(0), 0);
|
|
BUG_ON(ret);
|
|
}
|
|
__set_phys_to_machine(pfn, INVALID_P2M_ENTRY);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL_GPL(__xenmem_reservation_va_mapping_reset);
|
|
#endif /* CONFIG_XEN_HAVE_PVMMU */
|
|
|
|
/* @frames is an array of PFNs */
|
|
int xenmem_reservation_increase(int count, xen_pfn_t *frames)
|
|
{
|
|
struct xen_memory_reservation reservation = {
|
|
.address_bits = 0,
|
|
.extent_order = EXTENT_ORDER,
|
|
.domid = DOMID_SELF
|
|
};
|
|
|
|
/* XENMEM_populate_physmap requires a PFN based on Xen granularity. */
|
|
set_xen_guest_handle(reservation.extent_start, frames);
|
|
reservation.nr_extents = count;
|
|
return HYPERVISOR_memory_op(XENMEM_populate_physmap, &reservation);
|
|
}
|
|
EXPORT_SYMBOL_GPL(xenmem_reservation_increase);
|
|
|
|
/* @frames is an array of GFNs */
|
|
int xenmem_reservation_decrease(int count, xen_pfn_t *frames)
|
|
{
|
|
struct xen_memory_reservation reservation = {
|
|
.address_bits = 0,
|
|
.extent_order = EXTENT_ORDER,
|
|
.domid = DOMID_SELF
|
|
};
|
|
|
|
/* XENMEM_decrease_reservation requires a GFN */
|
|
set_xen_guest_handle(reservation.extent_start, frames);
|
|
reservation.nr_extents = count;
|
|
return HYPERVISOR_memory_op(XENMEM_decrease_reservation, &reservation);
|
|
}
|
|
EXPORT_SYMBOL_GPL(xenmem_reservation_decrease);
|