2020-10-01 09:22:22 +08:00
|
|
|
#ifndef KVM_DIRTY_RING_H
|
|
|
|
#define KVM_DIRTY_RING_H
|
|
|
|
|
|
|
|
#include <linux/kvm.h>
|
|
|
|
|
|
|
|
/**
|
|
|
|
* kvm_dirty_ring: KVM internal dirty ring structure
|
|
|
|
*
|
|
|
|
* @dirty_index: free running counter that points to the next slot in
|
|
|
|
* dirty_ring->dirty_gfns, where a new dirty page should go
|
|
|
|
* @reset_index: free running counter that points to the next dirty page
|
|
|
|
* in dirty_ring->dirty_gfns for which dirty trap needs to
|
|
|
|
* be reenabled
|
|
|
|
* @size: size of the compact list, dirty_ring->dirty_gfns
|
|
|
|
* @soft_limit: when the number of dirty pages in the list reaches this
|
|
|
|
* limit, vcpu that owns this ring should exit to userspace
|
|
|
|
* to allow userspace to harvest all the dirty pages
|
|
|
|
* @dirty_gfns: the array to keep the dirty gfns
|
|
|
|
* @index: index of this dirty ring
|
|
|
|
*/
|
|
|
|
struct kvm_dirty_ring {
|
|
|
|
u32 dirty_index;
|
|
|
|
u32 reset_index;
|
|
|
|
u32 size;
|
|
|
|
u32 soft_limit;
|
|
|
|
struct kvm_dirty_gfn *dirty_gfns;
|
|
|
|
int index;
|
|
|
|
};
|
|
|
|
|
2021-11-21 20:54:40 +08:00
|
|
|
#ifndef CONFIG_HAVE_KVM_DIRTY_RING
|
2020-10-01 09:22:22 +08:00
|
|
|
/*
|
2021-11-21 20:54:40 +08:00
|
|
|
* If CONFIG_HAVE_HVM_DIRTY_RING not defined, kvm_dirty_ring.o should
|
2020-10-01 09:22:22 +08:00
|
|
|
* not be included as well, so define these nop functions for the arch.
|
|
|
|
*/
|
|
|
|
static inline u32 kvm_dirty_ring_get_rsvd_entries(void)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int kvm_dirty_ring_alloc(struct kvm_dirty_ring *ring,
|
|
|
|
int index, u32 size)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int kvm_dirty_ring_reset(struct kvm *kvm,
|
|
|
|
struct kvm_dirty_ring *ring)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void kvm_dirty_ring_push(struct kvm_dirty_ring *ring,
|
|
|
|
u32 slot, u64 offset)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct page *kvm_dirty_ring_get_page(struct kvm_dirty_ring *ring,
|
|
|
|
u32 offset)
|
|
|
|
{
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void kvm_dirty_ring_free(struct kvm_dirty_ring *ring)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool kvm_dirty_ring_soft_full(struct kvm_dirty_ring *ring)
|
|
|
|
{
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2021-11-21 20:54:40 +08:00
|
|
|
#else /* CONFIG_HAVE_KVM_DIRTY_RING */
|
2020-10-01 09:22:22 +08:00
|
|
|
|
|
|
|
u32 kvm_dirty_ring_get_rsvd_entries(void);
|
|
|
|
int kvm_dirty_ring_alloc(struct kvm_dirty_ring *ring, int index, u32 size);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* called with kvm->slots_lock held, returns the number of
|
|
|
|
* processed pages.
|
|
|
|
*/
|
|
|
|
int kvm_dirty_ring_reset(struct kvm *kvm, struct kvm_dirty_ring *ring);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* returns =0: successfully pushed
|
|
|
|
* <0: unable to push, need to wait
|
|
|
|
*/
|
|
|
|
void kvm_dirty_ring_push(struct kvm_dirty_ring *ring, u32 slot, u64 offset);
|
|
|
|
|
|
|
|
/* for use in vm_operations_struct */
|
|
|
|
struct page *kvm_dirty_ring_get_page(struct kvm_dirty_ring *ring, u32 offset);
|
|
|
|
|
|
|
|
void kvm_dirty_ring_free(struct kvm_dirty_ring *ring);
|
|
|
|
bool kvm_dirty_ring_soft_full(struct kvm_dirty_ring *ring);
|
|
|
|
|
2021-11-21 20:54:40 +08:00
|
|
|
#endif /* CONFIG_HAVE_KVM_DIRTY_RING */
|
2020-10-01 09:22:22 +08:00
|
|
|
|
|
|
|
#endif /* KVM_DIRTY_RING_H */
|