mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2024-12-28 16:52:18 +00:00
86bdf3ebcf
ARM64 needs to dirty memory outside of a VCPU context when VGIC/ITS is enabled. It's conflicting with that ring-based dirty page tracking always requires a running VCPU context. Introduce a new flavor of dirty ring that requires the use of both VCPU dirty rings and a dirty bitmap. The expectation is that for non-VCPU sources of dirty memory (such as the VGIC/ITS on arm64), KVM writes to the dirty bitmap. Userspace should scan the dirty bitmap before migrating the VM to the target. Use an additional capability to advertise this behavior. The newly added capability (KVM_CAP_DIRTY_LOG_RING_WITH_BITMAP) can't be enabled before KVM_CAP_DIRTY_LOG_RING_ACQ_REL on ARM64. In this way, the newly added capability is treated as an extension of KVM_CAP_DIRTY_LOG_RING_ACQ_REL. Suggested-by: Marc Zyngier <maz@kernel.org> Suggested-by: Peter Xu <peterx@redhat.com> Co-developed-by: Oliver Upton <oliver.upton@linux.dev> Signed-off-by: Oliver Upton <oliver.upton@linux.dev> Signed-off-by: Gavin Shan <gshan@redhat.com> Acked-by: Peter Xu <peterx@redhat.com> Signed-off-by: Marc Zyngier <maz@kernel.org> Link: https://lore.kernel.org/r/20221110104914.31280-4-gshan@redhat.com
102 lines
2.7 KiB
C
102 lines
2.7 KiB
C
#ifndef KVM_DIRTY_RING_H
|
|
#define KVM_DIRTY_RING_H
|
|
|
|
#include <linux/kvm.h>
|
|
|
|
/**
|
|
* kvm_dirty_ring: KVM internal dirty ring structure
|
|
*
|
|
* @dirty_index: free running counter that points to the next slot in
|
|
* dirty_ring->dirty_gfns, where a new dirty page should go
|
|
* @reset_index: free running counter that points to the next dirty page
|
|
* in dirty_ring->dirty_gfns for which dirty trap needs to
|
|
* be reenabled
|
|
* @size: size of the compact list, dirty_ring->dirty_gfns
|
|
* @soft_limit: when the number of dirty pages in the list reaches this
|
|
* limit, vcpu that owns this ring should exit to userspace
|
|
* to allow userspace to harvest all the dirty pages
|
|
* @dirty_gfns: the array to keep the dirty gfns
|
|
* @index: index of this dirty ring
|
|
*/
|
|
struct kvm_dirty_ring {
|
|
u32 dirty_index;
|
|
u32 reset_index;
|
|
u32 size;
|
|
u32 soft_limit;
|
|
struct kvm_dirty_gfn *dirty_gfns;
|
|
int index;
|
|
};
|
|
|
|
#ifndef CONFIG_HAVE_KVM_DIRTY_RING
|
|
/*
|
|
* If CONFIG_HAVE_HVM_DIRTY_RING not defined, kvm_dirty_ring.o should
|
|
* not be included as well, so define these nop functions for the arch.
|
|
*/
|
|
static inline u32 kvm_dirty_ring_get_rsvd_entries(void)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline bool kvm_use_dirty_bitmap(struct kvm *kvm)
|
|
{
|
|
return true;
|
|
}
|
|
|
|
static inline int kvm_dirty_ring_alloc(struct kvm_dirty_ring *ring,
|
|
int index, u32 size)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int kvm_dirty_ring_reset(struct kvm *kvm,
|
|
struct kvm_dirty_ring *ring)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline void kvm_dirty_ring_push(struct kvm_vcpu *vcpu,
|
|
u32 slot, u64 offset)
|
|
{
|
|
}
|
|
|
|
static inline struct page *kvm_dirty_ring_get_page(struct kvm_dirty_ring *ring,
|
|
u32 offset)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
static inline void kvm_dirty_ring_free(struct kvm_dirty_ring *ring)
|
|
{
|
|
}
|
|
|
|
#else /* CONFIG_HAVE_KVM_DIRTY_RING */
|
|
|
|
int kvm_cpu_dirty_log_size(void);
|
|
bool kvm_use_dirty_bitmap(struct kvm *kvm);
|
|
bool kvm_arch_allow_write_without_running_vcpu(struct kvm *kvm);
|
|
u32 kvm_dirty_ring_get_rsvd_entries(void);
|
|
int kvm_dirty_ring_alloc(struct kvm_dirty_ring *ring, int index, u32 size);
|
|
|
|
/*
|
|
* called with kvm->slots_lock held, returns the number of
|
|
* processed pages.
|
|
*/
|
|
int kvm_dirty_ring_reset(struct kvm *kvm, struct kvm_dirty_ring *ring);
|
|
|
|
/*
|
|
* returns =0: successfully pushed
|
|
* <0: unable to push, need to wait
|
|
*/
|
|
void kvm_dirty_ring_push(struct kvm_vcpu *vcpu, u32 slot, u64 offset);
|
|
|
|
bool kvm_dirty_ring_check_request(struct kvm_vcpu *vcpu);
|
|
|
|
/* for use in vm_operations_struct */
|
|
struct page *kvm_dirty_ring_get_page(struct kvm_dirty_ring *ring, u32 offset);
|
|
|
|
void kvm_dirty_ring_free(struct kvm_dirty_ring *ring);
|
|
|
|
#endif /* CONFIG_HAVE_KVM_DIRTY_RING */
|
|
|
|
#endif /* KVM_DIRTY_RING_H */
|