Merge branch 'stable/vmalloc-3.2' of git://git.kernel.org/pub/scm/linux/kernel/git/konrad/xen
* 'stable/vmalloc-3.2' of git://git.kernel.org/pub/scm/linux/kernel/git/konrad/xen: net: xen-netback: use API provided by xenbus module to map rings block: xen-blkback: use API provided by xenbus module to map rings xen: use generic functions instead of xen_{alloc, free}_vm_area()
This commit is contained in:
commit
06d381484f
|
@ -1,29 +0,0 @@
|
||||||
/******************************************************************************
|
|
||||||
* arch/ia64/include/asm/xen/grant_table.h
|
|
||||||
*
|
|
||||||
* Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
|
|
||||||
* VA Linux Systems Japan K.K.
|
|
||||||
*
|
|
||||||
* This program is free software; you can redistribute it and/or modify
|
|
||||||
* it under the terms of the GNU General Public License as published by
|
|
||||||
* the Free Software Foundation; either version 2 of the License, or
|
|
||||||
* (at your option) any later version.
|
|
||||||
*
|
|
||||||
* This program is distributed in the hope that it will be useful,
|
|
||||||
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
||||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
||||||
* GNU General Public License for more details.
|
|
||||||
*
|
|
||||||
* You should have received a copy of the GNU General Public License
|
|
||||||
* along with this program; if not, write to the Free Software
|
|
||||||
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
||||||
*
|
|
||||||
*/
|
|
||||||
|
|
||||||
#ifndef _ASM_IA64_XEN_GRANT_TABLE_H
|
|
||||||
#define _ASM_IA64_XEN_GRANT_TABLE_H
|
|
||||||
|
|
||||||
struct vm_struct *xen_alloc_vm_area(unsigned long size);
|
|
||||||
void xen_free_vm_area(struct vm_struct *area);
|
|
||||||
|
|
||||||
#endif /* _ASM_IA64_XEN_GRANT_TABLE_H */
|
|
|
@ -31,68 +31,6 @@
|
||||||
|
|
||||||
#include <asm/xen/hypervisor.h>
|
#include <asm/xen/hypervisor.h>
|
||||||
|
|
||||||
struct vm_struct *xen_alloc_vm_area(unsigned long size)
|
|
||||||
{
|
|
||||||
int order;
|
|
||||||
unsigned long virt;
|
|
||||||
unsigned long nr_pages;
|
|
||||||
struct vm_struct *area;
|
|
||||||
|
|
||||||
order = get_order(size);
|
|
||||||
virt = __get_free_pages(GFP_KERNEL, order);
|
|
||||||
if (virt == 0)
|
|
||||||
goto err0;
|
|
||||||
nr_pages = 1 << order;
|
|
||||||
scrub_pages(virt, nr_pages);
|
|
||||||
|
|
||||||
area = kmalloc(sizeof(*area), GFP_KERNEL);
|
|
||||||
if (area == NULL)
|
|
||||||
goto err1;
|
|
||||||
|
|
||||||
area->flags = VM_IOREMAP;
|
|
||||||
area->addr = (void *)virt;
|
|
||||||
area->size = size;
|
|
||||||
area->pages = NULL;
|
|
||||||
area->nr_pages = nr_pages;
|
|
||||||
area->phys_addr = 0; /* xenbus_map_ring_valloc uses this field! */
|
|
||||||
|
|
||||||
return area;
|
|
||||||
|
|
||||||
err1:
|
|
||||||
free_pages(virt, order);
|
|
||||||
err0:
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL_GPL(xen_alloc_vm_area);
|
|
||||||
|
|
||||||
void xen_free_vm_area(struct vm_struct *area)
|
|
||||||
{
|
|
||||||
unsigned int order = get_order(area->size);
|
|
||||||
unsigned long i;
|
|
||||||
unsigned long phys_addr = __pa(area->addr);
|
|
||||||
|
|
||||||
/* This area is used for foreign page mappping.
|
|
||||||
* So underlying machine page may not be assigned. */
|
|
||||||
for (i = 0; i < (1 << order); i++) {
|
|
||||||
unsigned long ret;
|
|
||||||
unsigned long gpfn = (phys_addr >> PAGE_SHIFT) + i;
|
|
||||||
struct xen_memory_reservation reservation = {
|
|
||||||
.nr_extents = 1,
|
|
||||||
.address_bits = 0,
|
|
||||||
.extent_order = 0,
|
|
||||||
.domid = DOMID_SELF
|
|
||||||
};
|
|
||||||
set_xen_guest_handle(reservation.extent_start, &gpfn);
|
|
||||||
ret = HYPERVISOR_memory_op(XENMEM_populate_physmap,
|
|
||||||
&reservation);
|
|
||||||
BUG_ON(ret != 1);
|
|
||||||
}
|
|
||||||
free_pages((unsigned long)area->addr, order);
|
|
||||||
kfree(area);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL_GPL(xen_free_vm_area);
|
|
||||||
|
|
||||||
|
|
||||||
/****************************************************************************
|
/****************************************************************************
|
||||||
* grant table hack
|
* grant table hack
|
||||||
* cmd: GNTTABOP_xxx
|
* cmd: GNTTABOP_xxx
|
||||||
|
|
|
@ -1,7 +0,0 @@
|
||||||
#ifndef _ASM_X86_XEN_GRANT_TABLE_H
|
|
||||||
#define _ASM_X86_XEN_GRANT_TABLE_H
|
|
||||||
|
|
||||||
#define xen_alloc_vm_area(size) alloc_vm_area(size)
|
|
||||||
#define xen_free_vm_area(area) free_vm_area(area)
|
|
||||||
|
|
||||||
#endif /* _ASM_X86_XEN_GRANT_TABLE_H */
|
|
|
@ -71,7 +71,7 @@ int arch_gnttab_map_shared(unsigned long *frames, unsigned long nr_gframes,
|
||||||
|
|
||||||
if (shared == NULL) {
|
if (shared == NULL) {
|
||||||
struct vm_struct *area =
|
struct vm_struct *area =
|
||||||
xen_alloc_vm_area(PAGE_SIZE * max_nr_gframes);
|
alloc_vm_area(PAGE_SIZE * max_nr_gframes);
|
||||||
BUG_ON(area == NULL);
|
BUG_ON(area == NULL);
|
||||||
shared = area->addr;
|
shared = area->addr;
|
||||||
*__shared = shared;
|
*__shared = shared;
|
||||||
|
|
|
@ -170,7 +170,7 @@ struct xen_blkif {
|
||||||
enum blkif_protocol blk_protocol;
|
enum blkif_protocol blk_protocol;
|
||||||
enum blkif_backend_type blk_backend_type;
|
enum blkif_backend_type blk_backend_type;
|
||||||
union blkif_back_rings blk_rings;
|
union blkif_back_rings blk_rings;
|
||||||
struct vm_struct *blk_ring_area;
|
void *blk_ring;
|
||||||
/* The VBD attached to this interface. */
|
/* The VBD attached to this interface. */
|
||||||
struct xen_vbd vbd;
|
struct xen_vbd vbd;
|
||||||
/* Back pointer to the backend_info. */
|
/* Back pointer to the backend_info. */
|
||||||
|
@ -198,9 +198,6 @@ struct xen_blkif {
|
||||||
int st_wr_sect;
|
int st_wr_sect;
|
||||||
|
|
||||||
wait_queue_head_t waiting_to_free;
|
wait_queue_head_t waiting_to_free;
|
||||||
|
|
||||||
grant_handle_t shmem_handle;
|
|
||||||
grant_ref_t shmem_ref;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -122,38 +122,6 @@ static struct xen_blkif *xen_blkif_alloc(domid_t domid)
|
||||||
return blkif;
|
return blkif;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int map_frontend_page(struct xen_blkif *blkif, unsigned long shared_page)
|
|
||||||
{
|
|
||||||
struct gnttab_map_grant_ref op;
|
|
||||||
|
|
||||||
gnttab_set_map_op(&op, (unsigned long)blkif->blk_ring_area->addr,
|
|
||||||
GNTMAP_host_map, shared_page, blkif->domid);
|
|
||||||
|
|
||||||
if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
|
|
||||||
BUG();
|
|
||||||
|
|
||||||
if (op.status) {
|
|
||||||
DPRINTK("Grant table operation failure !\n");
|
|
||||||
return op.status;
|
|
||||||
}
|
|
||||||
|
|
||||||
blkif->shmem_ref = shared_page;
|
|
||||||
blkif->shmem_handle = op.handle;
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void unmap_frontend_page(struct xen_blkif *blkif)
|
|
||||||
{
|
|
||||||
struct gnttab_unmap_grant_ref op;
|
|
||||||
|
|
||||||
gnttab_set_unmap_op(&op, (unsigned long)blkif->blk_ring_area->addr,
|
|
||||||
GNTMAP_host_map, blkif->shmem_handle);
|
|
||||||
|
|
||||||
if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
|
|
||||||
BUG();
|
|
||||||
}
|
|
||||||
|
|
||||||
static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page,
|
static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page,
|
||||||
unsigned int evtchn)
|
unsigned int evtchn)
|
||||||
{
|
{
|
||||||
|
@ -163,35 +131,29 @@ static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page,
|
||||||
if (blkif->irq)
|
if (blkif->irq)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
blkif->blk_ring_area = alloc_vm_area(PAGE_SIZE);
|
err = xenbus_map_ring_valloc(blkif->be->dev, shared_page, &blkif->blk_ring);
|
||||||
if (!blkif->blk_ring_area)
|
if (err < 0)
|
||||||
return -ENOMEM;
|
|
||||||
|
|
||||||
err = map_frontend_page(blkif, shared_page);
|
|
||||||
if (err) {
|
|
||||||
free_vm_area(blkif->blk_ring_area);
|
|
||||||
return err;
|
return err;
|
||||||
}
|
|
||||||
|
|
||||||
switch (blkif->blk_protocol) {
|
switch (blkif->blk_protocol) {
|
||||||
case BLKIF_PROTOCOL_NATIVE:
|
case BLKIF_PROTOCOL_NATIVE:
|
||||||
{
|
{
|
||||||
struct blkif_sring *sring;
|
struct blkif_sring *sring;
|
||||||
sring = (struct blkif_sring *)blkif->blk_ring_area->addr;
|
sring = (struct blkif_sring *)blkif->blk_ring;
|
||||||
BACK_RING_INIT(&blkif->blk_rings.native, sring, PAGE_SIZE);
|
BACK_RING_INIT(&blkif->blk_rings.native, sring, PAGE_SIZE);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case BLKIF_PROTOCOL_X86_32:
|
case BLKIF_PROTOCOL_X86_32:
|
||||||
{
|
{
|
||||||
struct blkif_x86_32_sring *sring_x86_32;
|
struct blkif_x86_32_sring *sring_x86_32;
|
||||||
sring_x86_32 = (struct blkif_x86_32_sring *)blkif->blk_ring_area->addr;
|
sring_x86_32 = (struct blkif_x86_32_sring *)blkif->blk_ring;
|
||||||
BACK_RING_INIT(&blkif->blk_rings.x86_32, sring_x86_32, PAGE_SIZE);
|
BACK_RING_INIT(&blkif->blk_rings.x86_32, sring_x86_32, PAGE_SIZE);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case BLKIF_PROTOCOL_X86_64:
|
case BLKIF_PROTOCOL_X86_64:
|
||||||
{
|
{
|
||||||
struct blkif_x86_64_sring *sring_x86_64;
|
struct blkif_x86_64_sring *sring_x86_64;
|
||||||
sring_x86_64 = (struct blkif_x86_64_sring *)blkif->blk_ring_area->addr;
|
sring_x86_64 = (struct blkif_x86_64_sring *)blkif->blk_ring;
|
||||||
BACK_RING_INIT(&blkif->blk_rings.x86_64, sring_x86_64, PAGE_SIZE);
|
BACK_RING_INIT(&blkif->blk_rings.x86_64, sring_x86_64, PAGE_SIZE);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
@ -203,8 +165,7 @@ static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page,
|
||||||
xen_blkif_be_int, 0,
|
xen_blkif_be_int, 0,
|
||||||
"blkif-backend", blkif);
|
"blkif-backend", blkif);
|
||||||
if (err < 0) {
|
if (err < 0) {
|
||||||
unmap_frontend_page(blkif);
|
xenbus_unmap_ring_vfree(blkif->be->dev, blkif->blk_ring);
|
||||||
free_vm_area(blkif->blk_ring_area);
|
|
||||||
blkif->blk_rings.common.sring = NULL;
|
blkif->blk_rings.common.sring = NULL;
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
@ -230,8 +191,7 @@ static void xen_blkif_disconnect(struct xen_blkif *blkif)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (blkif->blk_rings.common.sring) {
|
if (blkif->blk_rings.common.sring) {
|
||||||
unmap_frontend_page(blkif);
|
xenbus_unmap_ring_vfree(blkif->be->dev, blkif->blk_ring);
|
||||||
free_vm_area(blkif->blk_ring_area);
|
|
||||||
blkif->blk_rings.common.sring = NULL;
|
blkif->blk_rings.common.sring = NULL;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -58,10 +58,6 @@ struct xenvif {
|
||||||
u8 fe_dev_addr[6];
|
u8 fe_dev_addr[6];
|
||||||
|
|
||||||
/* Physical parameters of the comms window. */
|
/* Physical parameters of the comms window. */
|
||||||
grant_handle_t tx_shmem_handle;
|
|
||||||
grant_ref_t tx_shmem_ref;
|
|
||||||
grant_handle_t rx_shmem_handle;
|
|
||||||
grant_ref_t rx_shmem_ref;
|
|
||||||
unsigned int irq;
|
unsigned int irq;
|
||||||
|
|
||||||
/* List of frontends to notify after a batch of frames sent. */
|
/* List of frontends to notify after a batch of frames sent. */
|
||||||
|
@ -70,8 +66,6 @@ struct xenvif {
|
||||||
/* The shared rings and indexes. */
|
/* The shared rings and indexes. */
|
||||||
struct xen_netif_tx_back_ring tx;
|
struct xen_netif_tx_back_ring tx;
|
||||||
struct xen_netif_rx_back_ring rx;
|
struct xen_netif_rx_back_ring rx;
|
||||||
struct vm_struct *tx_comms_area;
|
|
||||||
struct vm_struct *rx_comms_area;
|
|
||||||
|
|
||||||
/* Frontend feature information. */
|
/* Frontend feature information. */
|
||||||
u8 can_sg:1;
|
u8 can_sg:1;
|
||||||
|
@ -106,6 +100,11 @@ struct xenvif {
|
||||||
wait_queue_head_t waiting_to_free;
|
wait_queue_head_t waiting_to_free;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
static inline struct xenbus_device *xenvif_to_xenbus_device(struct xenvif *vif)
|
||||||
|
{
|
||||||
|
return to_xenbus_device(vif->dev->dev.parent);
|
||||||
|
}
|
||||||
|
|
||||||
#define XEN_NETIF_TX_RING_SIZE __CONST_RING_SIZE(xen_netif_tx, PAGE_SIZE)
|
#define XEN_NETIF_TX_RING_SIZE __CONST_RING_SIZE(xen_netif_tx, PAGE_SIZE)
|
||||||
#define XEN_NETIF_RX_RING_SIZE __CONST_RING_SIZE(xen_netif_rx, PAGE_SIZE)
|
#define XEN_NETIF_RX_RING_SIZE __CONST_RING_SIZE(xen_netif_rx, PAGE_SIZE)
|
||||||
|
|
||||||
|
|
|
@ -1589,88 +1589,42 @@ static int xen_netbk_kthread(void *data)
|
||||||
|
|
||||||
void xen_netbk_unmap_frontend_rings(struct xenvif *vif)
|
void xen_netbk_unmap_frontend_rings(struct xenvif *vif)
|
||||||
{
|
{
|
||||||
struct gnttab_unmap_grant_ref op;
|
if (vif->tx.sring)
|
||||||
|
xenbus_unmap_ring_vfree(xenvif_to_xenbus_device(vif),
|
||||||
if (vif->tx.sring) {
|
vif->tx.sring);
|
||||||
gnttab_set_unmap_op(&op, (unsigned long)vif->tx_comms_area->addr,
|
if (vif->rx.sring)
|
||||||
GNTMAP_host_map, vif->tx_shmem_handle);
|
xenbus_unmap_ring_vfree(xenvif_to_xenbus_device(vif),
|
||||||
|
vif->rx.sring);
|
||||||
if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
|
|
||||||
BUG();
|
|
||||||
}
|
|
||||||
|
|
||||||
if (vif->rx.sring) {
|
|
||||||
gnttab_set_unmap_op(&op, (unsigned long)vif->rx_comms_area->addr,
|
|
||||||
GNTMAP_host_map, vif->rx_shmem_handle);
|
|
||||||
|
|
||||||
if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
|
|
||||||
BUG();
|
|
||||||
}
|
|
||||||
if (vif->rx_comms_area)
|
|
||||||
free_vm_area(vif->rx_comms_area);
|
|
||||||
if (vif->tx_comms_area)
|
|
||||||
free_vm_area(vif->tx_comms_area);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
int xen_netbk_map_frontend_rings(struct xenvif *vif,
|
int xen_netbk_map_frontend_rings(struct xenvif *vif,
|
||||||
grant_ref_t tx_ring_ref,
|
grant_ref_t tx_ring_ref,
|
||||||
grant_ref_t rx_ring_ref)
|
grant_ref_t rx_ring_ref)
|
||||||
{
|
{
|
||||||
struct gnttab_map_grant_ref op;
|
void *addr;
|
||||||
struct xen_netif_tx_sring *txs;
|
struct xen_netif_tx_sring *txs;
|
||||||
struct xen_netif_rx_sring *rxs;
|
struct xen_netif_rx_sring *rxs;
|
||||||
|
|
||||||
int err = -ENOMEM;
|
int err = -ENOMEM;
|
||||||
|
|
||||||
vif->tx_comms_area = alloc_vm_area(PAGE_SIZE);
|
err = xenbus_map_ring_valloc(xenvif_to_xenbus_device(vif),
|
||||||
if (vif->tx_comms_area == NULL)
|
tx_ring_ref, &addr);
|
||||||
|
if (err)
|
||||||
goto err;
|
goto err;
|
||||||
|
|
||||||
vif->rx_comms_area = alloc_vm_area(PAGE_SIZE);
|
txs = (struct xen_netif_tx_sring *)addr;
|
||||||
if (vif->rx_comms_area == NULL)
|
|
||||||
goto err;
|
|
||||||
|
|
||||||
gnttab_set_map_op(&op, (unsigned long)vif->tx_comms_area->addr,
|
|
||||||
GNTMAP_host_map, tx_ring_ref, vif->domid);
|
|
||||||
|
|
||||||
if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
|
|
||||||
BUG();
|
|
||||||
|
|
||||||
if (op.status) {
|
|
||||||
netdev_warn(vif->dev,
|
|
||||||
"failed to map tx ring. err=%d status=%d\n",
|
|
||||||
err, op.status);
|
|
||||||
err = op.status;
|
|
||||||
goto err;
|
|
||||||
}
|
|
||||||
|
|
||||||
vif->tx_shmem_ref = tx_ring_ref;
|
|
||||||
vif->tx_shmem_handle = op.handle;
|
|
||||||
|
|
||||||
txs = (struct xen_netif_tx_sring *)vif->tx_comms_area->addr;
|
|
||||||
BACK_RING_INIT(&vif->tx, txs, PAGE_SIZE);
|
BACK_RING_INIT(&vif->tx, txs, PAGE_SIZE);
|
||||||
|
|
||||||
gnttab_set_map_op(&op, (unsigned long)vif->rx_comms_area->addr,
|
err = xenbus_map_ring_valloc(xenvif_to_xenbus_device(vif),
|
||||||
GNTMAP_host_map, rx_ring_ref, vif->domid);
|
rx_ring_ref, &addr);
|
||||||
|
if (err)
|
||||||
if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
|
|
||||||
BUG();
|
|
||||||
|
|
||||||
if (op.status) {
|
|
||||||
netdev_warn(vif->dev,
|
|
||||||
"failed to map rx ring. err=%d status=%d\n",
|
|
||||||
err, op.status);
|
|
||||||
err = op.status;
|
|
||||||
goto err;
|
goto err;
|
||||||
}
|
|
||||||
|
|
||||||
vif->rx_shmem_ref = rx_ring_ref;
|
rxs = (struct xen_netif_rx_sring *)addr;
|
||||||
vif->rx_shmem_handle = op.handle;
|
|
||||||
vif->rx_req_cons_peek = 0;
|
|
||||||
|
|
||||||
rxs = (struct xen_netif_rx_sring *)vif->rx_comms_area->addr;
|
|
||||||
BACK_RING_INIT(&vif->rx, rxs, PAGE_SIZE);
|
BACK_RING_INIT(&vif->rx, rxs, PAGE_SIZE);
|
||||||
|
|
||||||
|
vif->rx_req_cons_peek = 0;
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
err:
|
err:
|
||||||
|
|
|
@ -443,7 +443,7 @@ int xenbus_map_ring_valloc(struct xenbus_device *dev, int gnt_ref, void **vaddr)
|
||||||
|
|
||||||
*vaddr = NULL;
|
*vaddr = NULL;
|
||||||
|
|
||||||
area = xen_alloc_vm_area(PAGE_SIZE);
|
area = alloc_vm_area(PAGE_SIZE);
|
||||||
if (!area)
|
if (!area)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
@ -453,7 +453,7 @@ int xenbus_map_ring_valloc(struct xenbus_device *dev, int gnt_ref, void **vaddr)
|
||||||
BUG();
|
BUG();
|
||||||
|
|
||||||
if (op.status != GNTST_okay) {
|
if (op.status != GNTST_okay) {
|
||||||
xen_free_vm_area(area);
|
free_vm_area(area);
|
||||||
xenbus_dev_fatal(dev, op.status,
|
xenbus_dev_fatal(dev, op.status,
|
||||||
"mapping in shared page %d from domain %d",
|
"mapping in shared page %d from domain %d",
|
||||||
gnt_ref, dev->otherend_id);
|
gnt_ref, dev->otherend_id);
|
||||||
|
@ -552,7 +552,7 @@ int xenbus_unmap_ring_vfree(struct xenbus_device *dev, void *vaddr)
|
||||||
BUG();
|
BUG();
|
||||||
|
|
||||||
if (op.status == GNTST_okay)
|
if (op.status == GNTST_okay)
|
||||||
xen_free_vm_area(area);
|
free_vm_area(area);
|
||||||
else
|
else
|
||||||
xenbus_dev_error(dev, op.status,
|
xenbus_dev_error(dev, op.status,
|
||||||
"unmapping page at handle %d error %d",
|
"unmapping page at handle %d error %d",
|
||||||
|
|
|
@ -43,7 +43,6 @@
|
||||||
#include <xen/interface/grant_table.h>
|
#include <xen/interface/grant_table.h>
|
||||||
|
|
||||||
#include <asm/xen/hypervisor.h>
|
#include <asm/xen/hypervisor.h>
|
||||||
#include <asm/xen/grant_table.h>
|
|
||||||
|
|
||||||
#include <xen/features.h>
|
#include <xen/features.h>
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue