mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2024-12-28 16:53:49 +00:00
2bcb52a360
Pin, as in FOLL_PIN, pages when mapping them for direct access by KVM. As per Documentation/core-api/pin_user_pages.rst, writing to a page that was gotten via FOLL_GET is explicitly disallowed. Correct (uses FOLL_PIN calls): pin_user_pages() write to the data within the pages unpin_user_pages() INCORRECT (uses FOLL_GET calls): get_user_pages() write to the data within the pages put_page() Unfortunately, FOLL_PIN is a "private" flag, and so kvm_follow_pfn must use a one-off bool instead of being able to piggyback the "flags" field. Link: https://lwn.net/Articles/930667 Link: https://lore.kernel.org/all/cover.1683044162.git.lstoakes@gmail.com Tested-by: Alex Bennée <alex.bennee@linaro.org> Signed-off-by: Sean Christopherson <seanjc@google.com> Tested-by: Dmitry Osipenko <dmitry.osipenko@collabora.com> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> Message-ID: <20241010182427.1434605-32-seanjc@google.com>
97 lines
2.6 KiB
C
97 lines
2.6 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
|
|
#ifndef __KVM_MM_H__
|
|
#define __KVM_MM_H__ 1
|
|
|
|
/*
|
|
* Architectures can choose whether to use an rwlock or spinlock
|
|
* for the mmu_lock. These macros, for use in common code
|
|
* only, avoids using #ifdefs in places that must deal with
|
|
* multiple architectures.
|
|
*/
|
|
|
|
#ifdef KVM_HAVE_MMU_RWLOCK
|
|
#define KVM_MMU_LOCK_INIT(kvm) rwlock_init(&(kvm)->mmu_lock)
|
|
#define KVM_MMU_LOCK(kvm) write_lock(&(kvm)->mmu_lock)
|
|
#define KVM_MMU_UNLOCK(kvm) write_unlock(&(kvm)->mmu_lock)
|
|
#else
|
|
#define KVM_MMU_LOCK_INIT(kvm) spin_lock_init(&(kvm)->mmu_lock)
|
|
#define KVM_MMU_LOCK(kvm) spin_lock(&(kvm)->mmu_lock)
|
|
#define KVM_MMU_UNLOCK(kvm) spin_unlock(&(kvm)->mmu_lock)
|
|
#endif /* KVM_HAVE_MMU_RWLOCK */
|
|
|
|
|
|
struct kvm_follow_pfn {
|
|
const struct kvm_memory_slot *slot;
|
|
const gfn_t gfn;
|
|
|
|
unsigned long hva;
|
|
|
|
/* FOLL_* flags modifying lookup behavior, e.g. FOLL_WRITE. */
|
|
unsigned int flags;
|
|
|
|
/*
|
|
* Pin the page (effectively FOLL_PIN, which is an mm/ internal flag).
|
|
* The page *must* be pinned if KVM will write to the page via a kernel
|
|
* mapping, e.g. via kmap(), mremap(), etc.
|
|
*/
|
|
bool pin;
|
|
|
|
/*
|
|
* If non-NULL, try to get a writable mapping even for a read fault.
|
|
* Set to true if a writable mapping was obtained.
|
|
*/
|
|
bool *map_writable;
|
|
|
|
/*
|
|
* Optional output. Set to a valid "struct page" if the returned pfn
|
|
* is for a refcounted or pinned struct page, NULL if the returned pfn
|
|
* has no struct page or if the struct page is not being refcounted
|
|
* (e.g. tail pages of non-compound higher order allocations from
|
|
* IO/PFNMAP mappings).
|
|
*/
|
|
struct page **refcounted_page;
|
|
};
|
|
|
|
kvm_pfn_t hva_to_pfn(struct kvm_follow_pfn *kfp);
|
|
|
|
#ifdef CONFIG_HAVE_KVM_PFNCACHE
|
|
void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm,
|
|
unsigned long start,
|
|
unsigned long end);
|
|
#else
|
|
static inline void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm,
|
|
unsigned long start,
|
|
unsigned long end)
|
|
{
|
|
}
|
|
#endif /* HAVE_KVM_PFNCACHE */
|
|
|
|
#ifdef CONFIG_KVM_PRIVATE_MEM
|
|
void kvm_gmem_init(struct module *module);
|
|
int kvm_gmem_create(struct kvm *kvm, struct kvm_create_guest_memfd *args);
|
|
int kvm_gmem_bind(struct kvm *kvm, struct kvm_memory_slot *slot,
|
|
unsigned int fd, loff_t offset);
|
|
void kvm_gmem_unbind(struct kvm_memory_slot *slot);
|
|
#else
|
|
static inline void kvm_gmem_init(struct module *module)
|
|
{
|
|
|
|
}
|
|
|
|
static inline int kvm_gmem_bind(struct kvm *kvm,
|
|
struct kvm_memory_slot *slot,
|
|
unsigned int fd, loff_t offset)
|
|
{
|
|
WARN_ON_ONCE(1);
|
|
return -EIO;
|
|
}
|
|
|
|
static inline void kvm_gmem_unbind(struct kvm_memory_slot *slot)
|
|
{
|
|
WARN_ON_ONCE(1);
|
|
}
|
|
#endif /* CONFIG_KVM_PRIVATE_MEM */
|
|
|
|
#endif /* __KVM_MM_H__ */
|