mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2024-12-29 09:13:38 +00:00
mm: make vma_prepare() and friends static and internal to vma.c
Now we have abstracted merge behaviour for new VMA ranges, we are able to render vma_prepare(), init_vma_prep(), vma_complete(), can_vma_merge_before() and can_vma_merge_after() static and internal to vma.c. These are internal implementation details of kernel VMA manipulation and merging mechanisms and thus should not be exposed. This also renders the functions userland testable. Link: https://lkml.kernel.org/r/7f7f1c34ce10405a6aab2714c505af3cf41b7851.1725040657.git.lorenzo.stoakes@oracle.com Signed-off-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com> Cc: Liam R. Howlett <Liam.Howlett@oracle.com> Cc: Mark Brown <broonie@kernel.org> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: Bert Karwatzki <spasswolf@web.de> Cc: Jeff Xu <jeffxu@chromium.org> Cc: Jiri Olsa <olsajiri@gmail.com> Cc: Kees Cook <kees@kernel.org> Cc: Lorenzo Stoakes <lstoakes@gmail.com> Cc: Matthew Wilcox <willy@infradead.org> Cc: "Paul E. McKenney" <paulmck@kernel.org> Cc: Paul Moore <paul@paul-moore.com> Cc: Sidhartha Kumar <sidhartha.kumar@oracle.com> Cc: Suren Baghdasaryan <surenb@google.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
cacded5e42
commit
25d3925fa5
318
mm/vma.c
318
mm/vma.c
@ -104,8 +104,7 @@ static void init_multi_vma_prep(struct vma_prepare *vp,
|
||||
*
|
||||
* We assume the vma may be removed as part of the merge.
|
||||
*/
|
||||
bool
|
||||
can_vma_merge_before(struct vma_merge_struct *vmg)
|
||||
static bool can_vma_merge_before(struct vma_merge_struct *vmg)
|
||||
{
|
||||
pgoff_t pglen = PHYS_PFN(vmg->end - vmg->start);
|
||||
|
||||
@ -127,7 +126,7 @@ can_vma_merge_before(struct vma_merge_struct *vmg)
|
||||
*
|
||||
* We assume that vma is not removed as part of the merge.
|
||||
*/
|
||||
bool can_vma_merge_after(struct vma_merge_struct *vmg)
|
||||
static bool can_vma_merge_after(struct vma_merge_struct *vmg)
|
||||
{
|
||||
if (is_mergeable_vma(vmg, /* merge_next = */ false) &&
|
||||
is_mergeable_anon_vma(vmg->anon_vma, vmg->prev->anon_vma, vmg->prev)) {
|
||||
@ -137,6 +136,162 @@ bool can_vma_merge_after(struct vma_merge_struct *vmg)
|
||||
return false;
|
||||
}
|
||||
|
||||
static void __vma_link_file(struct vm_area_struct *vma,
|
||||
struct address_space *mapping)
|
||||
{
|
||||
if (vma_is_shared_maywrite(vma))
|
||||
mapping_allow_writable(mapping);
|
||||
|
||||
flush_dcache_mmap_lock(mapping);
|
||||
vma_interval_tree_insert(vma, &mapping->i_mmap);
|
||||
flush_dcache_mmap_unlock(mapping);
|
||||
}
|
||||
|
||||
/*
|
||||
* Requires inode->i_mapping->i_mmap_rwsem
|
||||
*/
|
||||
static void __remove_shared_vm_struct(struct vm_area_struct *vma,
|
||||
struct address_space *mapping)
|
||||
{
|
||||
if (vma_is_shared_maywrite(vma))
|
||||
mapping_unmap_writable(mapping);
|
||||
|
||||
flush_dcache_mmap_lock(mapping);
|
||||
vma_interval_tree_remove(vma, &mapping->i_mmap);
|
||||
flush_dcache_mmap_unlock(mapping);
|
||||
}
|
||||
|
||||
/*
|
||||
* vma_prepare() - Helper function for handling locking VMAs prior to altering
|
||||
* @vp: The initialized vma_prepare struct
|
||||
*/
|
||||
static void vma_prepare(struct vma_prepare *vp)
|
||||
{
|
||||
if (vp->file) {
|
||||
uprobe_munmap(vp->vma, vp->vma->vm_start, vp->vma->vm_end);
|
||||
|
||||
if (vp->adj_next)
|
||||
uprobe_munmap(vp->adj_next, vp->adj_next->vm_start,
|
||||
vp->adj_next->vm_end);
|
||||
|
||||
i_mmap_lock_write(vp->mapping);
|
||||
if (vp->insert && vp->insert->vm_file) {
|
||||
/*
|
||||
* Put into interval tree now, so instantiated pages
|
||||
* are visible to arm/parisc __flush_dcache_page
|
||||
* throughout; but we cannot insert into address
|
||||
* space until vma start or end is updated.
|
||||
*/
|
||||
__vma_link_file(vp->insert,
|
||||
vp->insert->vm_file->f_mapping);
|
||||
}
|
||||
}
|
||||
|
||||
if (vp->anon_vma) {
|
||||
anon_vma_lock_write(vp->anon_vma);
|
||||
anon_vma_interval_tree_pre_update_vma(vp->vma);
|
||||
if (vp->adj_next)
|
||||
anon_vma_interval_tree_pre_update_vma(vp->adj_next);
|
||||
}
|
||||
|
||||
if (vp->file) {
|
||||
flush_dcache_mmap_lock(vp->mapping);
|
||||
vma_interval_tree_remove(vp->vma, &vp->mapping->i_mmap);
|
||||
if (vp->adj_next)
|
||||
vma_interval_tree_remove(vp->adj_next,
|
||||
&vp->mapping->i_mmap);
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
/*
|
||||
* vma_complete- Helper function for handling the unlocking after altering VMAs,
|
||||
* or for inserting a VMA.
|
||||
*
|
||||
* @vp: The vma_prepare struct
|
||||
* @vmi: The vma iterator
|
||||
* @mm: The mm_struct
|
||||
*/
|
||||
static void vma_complete(struct vma_prepare *vp, struct vma_iterator *vmi,
|
||||
struct mm_struct *mm)
|
||||
{
|
||||
if (vp->file) {
|
||||
if (vp->adj_next)
|
||||
vma_interval_tree_insert(vp->adj_next,
|
||||
&vp->mapping->i_mmap);
|
||||
vma_interval_tree_insert(vp->vma, &vp->mapping->i_mmap);
|
||||
flush_dcache_mmap_unlock(vp->mapping);
|
||||
}
|
||||
|
||||
if (vp->remove && vp->file) {
|
||||
__remove_shared_vm_struct(vp->remove, vp->mapping);
|
||||
if (vp->remove2)
|
||||
__remove_shared_vm_struct(vp->remove2, vp->mapping);
|
||||
} else if (vp->insert) {
|
||||
/*
|
||||
* split_vma has split insert from vma, and needs
|
||||
* us to insert it before dropping the locks
|
||||
* (it may either follow vma or precede it).
|
||||
*/
|
||||
vma_iter_store(vmi, vp->insert);
|
||||
mm->map_count++;
|
||||
}
|
||||
|
||||
if (vp->anon_vma) {
|
||||
anon_vma_interval_tree_post_update_vma(vp->vma);
|
||||
if (vp->adj_next)
|
||||
anon_vma_interval_tree_post_update_vma(vp->adj_next);
|
||||
anon_vma_unlock_write(vp->anon_vma);
|
||||
}
|
||||
|
||||
if (vp->file) {
|
||||
i_mmap_unlock_write(vp->mapping);
|
||||
uprobe_mmap(vp->vma);
|
||||
|
||||
if (vp->adj_next)
|
||||
uprobe_mmap(vp->adj_next);
|
||||
}
|
||||
|
||||
if (vp->remove) {
|
||||
again:
|
||||
vma_mark_detached(vp->remove, true);
|
||||
if (vp->file) {
|
||||
uprobe_munmap(vp->remove, vp->remove->vm_start,
|
||||
vp->remove->vm_end);
|
||||
fput(vp->file);
|
||||
}
|
||||
if (vp->remove->anon_vma)
|
||||
anon_vma_merge(vp->vma, vp->remove);
|
||||
mm->map_count--;
|
||||
mpol_put(vma_policy(vp->remove));
|
||||
if (!vp->remove2)
|
||||
WARN_ON_ONCE(vp->vma->vm_end < vp->remove->vm_end);
|
||||
vm_area_free(vp->remove);
|
||||
|
||||
/*
|
||||
* In mprotect's case 6 (see comments on vma_merge),
|
||||
* we are removing both mid and next vmas
|
||||
*/
|
||||
if (vp->remove2) {
|
||||
vp->remove = vp->remove2;
|
||||
vp->remove2 = NULL;
|
||||
goto again;
|
||||
}
|
||||
}
|
||||
if (vp->insert && vp->file)
|
||||
uprobe_mmap(vp->insert);
|
||||
}
|
||||
|
||||
/*
|
||||
* init_vma_prep() - Initializer wrapper for vma_prepare struct
|
||||
* @vp: The vma_prepare struct
|
||||
* @vma: The vma that will be altered once locked
|
||||
*/
|
||||
static void init_vma_prep(struct vma_prepare *vp, struct vm_area_struct *vma)
|
||||
{
|
||||
init_multi_vma_prep(vp, vma, NULL, NULL, NULL);
|
||||
}
|
||||
|
||||
/*
|
||||
* Can the proposed VMA be merged with the left (previous) VMA taking into
|
||||
* account the start position of the proposed range.
|
||||
@ -315,31 +470,6 @@ static int split_vma(struct vma_iterator *vmi, struct vm_area_struct *vma,
|
||||
return __split_vma(vmi, vma, addr, new_below);
|
||||
}
|
||||
|
||||
/*
|
||||
* init_vma_prep() - Initializer wrapper for vma_prepare struct
|
||||
* @vp: The vma_prepare struct
|
||||
* @vma: The vma that will be altered once locked
|
||||
*/
|
||||
void init_vma_prep(struct vma_prepare *vp,
|
||||
struct vm_area_struct *vma)
|
||||
{
|
||||
init_multi_vma_prep(vp, vma, NULL, NULL, NULL);
|
||||
}
|
||||
|
||||
/*
|
||||
* Requires inode->i_mapping->i_mmap_rwsem
|
||||
*/
|
||||
static void __remove_shared_vm_struct(struct vm_area_struct *vma,
|
||||
struct address_space *mapping)
|
||||
{
|
||||
if (vma_is_shared_maywrite(vma))
|
||||
mapping_unmap_writable(mapping);
|
||||
|
||||
flush_dcache_mmap_lock(mapping);
|
||||
vma_interval_tree_remove(vma, &mapping->i_mmap);
|
||||
flush_dcache_mmap_unlock(mapping);
|
||||
}
|
||||
|
||||
/*
|
||||
* vma has some anon_vma assigned, and is already inserted on that
|
||||
* anon_vma's interval trees.
|
||||
@ -372,60 +502,6 @@ anon_vma_interval_tree_post_update_vma(struct vm_area_struct *vma)
|
||||
anon_vma_interval_tree_insert(avc, &avc->anon_vma->rb_root);
|
||||
}
|
||||
|
||||
static void __vma_link_file(struct vm_area_struct *vma,
|
||||
struct address_space *mapping)
|
||||
{
|
||||
if (vma_is_shared_maywrite(vma))
|
||||
mapping_allow_writable(mapping);
|
||||
|
||||
flush_dcache_mmap_lock(mapping);
|
||||
vma_interval_tree_insert(vma, &mapping->i_mmap);
|
||||
flush_dcache_mmap_unlock(mapping);
|
||||
}
|
||||
|
||||
/*
|
||||
* vma_prepare() - Helper function for handling locking VMAs prior to altering
|
||||
* @vp: The initialized vma_prepare struct
|
||||
*/
|
||||
void vma_prepare(struct vma_prepare *vp)
|
||||
{
|
||||
if (vp->file) {
|
||||
uprobe_munmap(vp->vma, vp->vma->vm_start, vp->vma->vm_end);
|
||||
|
||||
if (vp->adj_next)
|
||||
uprobe_munmap(vp->adj_next, vp->adj_next->vm_start,
|
||||
vp->adj_next->vm_end);
|
||||
|
||||
i_mmap_lock_write(vp->mapping);
|
||||
if (vp->insert && vp->insert->vm_file) {
|
||||
/*
|
||||
* Put into interval tree now, so instantiated pages
|
||||
* are visible to arm/parisc __flush_dcache_page
|
||||
* throughout; but we cannot insert into address
|
||||
* space until vma start or end is updated.
|
||||
*/
|
||||
__vma_link_file(vp->insert,
|
||||
vp->insert->vm_file->f_mapping);
|
||||
}
|
||||
}
|
||||
|
||||
if (vp->anon_vma) {
|
||||
anon_vma_lock_write(vp->anon_vma);
|
||||
anon_vma_interval_tree_pre_update_vma(vp->vma);
|
||||
if (vp->adj_next)
|
||||
anon_vma_interval_tree_pre_update_vma(vp->adj_next);
|
||||
}
|
||||
|
||||
if (vp->file) {
|
||||
flush_dcache_mmap_lock(vp->mapping);
|
||||
vma_interval_tree_remove(vp->vma, &vp->mapping->i_mmap);
|
||||
if (vp->adj_next)
|
||||
vma_interval_tree_remove(vp->adj_next,
|
||||
&vp->mapping->i_mmap);
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
/*
|
||||
* dup_anon_vma() - Helper function to duplicate anon_vma
|
||||
* @dst: The destination VMA
|
||||
@ -715,84 +791,6 @@ int vma_shrink(struct vma_iterator *vmi, struct vm_area_struct *vma,
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* vma_complete- Helper function for handling the unlocking after altering VMAs,
|
||||
* or for inserting a VMA.
|
||||
*
|
||||
* @vp: The vma_prepare struct
|
||||
* @vmi: The vma iterator
|
||||
* @mm: The mm_struct
|
||||
*/
|
||||
void vma_complete(struct vma_prepare *vp,
|
||||
struct vma_iterator *vmi, struct mm_struct *mm)
|
||||
{
|
||||
if (vp->file) {
|
||||
if (vp->adj_next)
|
||||
vma_interval_tree_insert(vp->adj_next,
|
||||
&vp->mapping->i_mmap);
|
||||
vma_interval_tree_insert(vp->vma, &vp->mapping->i_mmap);
|
||||
flush_dcache_mmap_unlock(vp->mapping);
|
||||
}
|
||||
|
||||
if (vp->remove && vp->file) {
|
||||
__remove_shared_vm_struct(vp->remove, vp->mapping);
|
||||
if (vp->remove2)
|
||||
__remove_shared_vm_struct(vp->remove2, vp->mapping);
|
||||
} else if (vp->insert) {
|
||||
/*
|
||||
* split_vma has split insert from vma, and needs
|
||||
* us to insert it before dropping the locks
|
||||
* (it may either follow vma or precede it).
|
||||
*/
|
||||
vma_iter_store(vmi, vp->insert);
|
||||
mm->map_count++;
|
||||
}
|
||||
|
||||
if (vp->anon_vma) {
|
||||
anon_vma_interval_tree_post_update_vma(vp->vma);
|
||||
if (vp->adj_next)
|
||||
anon_vma_interval_tree_post_update_vma(vp->adj_next);
|
||||
anon_vma_unlock_write(vp->anon_vma);
|
||||
}
|
||||
|
||||
if (vp->file) {
|
||||
i_mmap_unlock_write(vp->mapping);
|
||||
uprobe_mmap(vp->vma);
|
||||
|
||||
if (vp->adj_next)
|
||||
uprobe_mmap(vp->adj_next);
|
||||
}
|
||||
|
||||
if (vp->remove) {
|
||||
again:
|
||||
vma_mark_detached(vp->remove, true);
|
||||
if (vp->file) {
|
||||
uprobe_munmap(vp->remove, vp->remove->vm_start,
|
||||
vp->remove->vm_end);
|
||||
fput(vp->file);
|
||||
}
|
||||
if (vp->remove->anon_vma)
|
||||
anon_vma_merge(vp->vma, vp->remove);
|
||||
mm->map_count--;
|
||||
mpol_put(vma_policy(vp->remove));
|
||||
if (!vp->remove2)
|
||||
WARN_ON_ONCE(vp->vma->vm_end < vp->remove->vm_end);
|
||||
vm_area_free(vp->remove);
|
||||
|
||||
/*
|
||||
* In mprotect's case 6 (see comments on vma_merge),
|
||||
* we are removing both mid and next vmas
|
||||
*/
|
||||
if (vp->remove2) {
|
||||
vp->remove = vp->remove2;
|
||||
vp->remove2 = NULL;
|
||||
goto again;
|
||||
}
|
||||
}
|
||||
if (vp->insert && vp->file)
|
||||
uprobe_mmap(vp->insert);
|
||||
}
|
||||
|
||||
static inline void vms_clear_ptes(struct vma_munmap_struct *vms,
|
||||
struct ma_state *mas_detach, bool mm_wr_locked)
|
||||
{
|
||||
|
25
mm/vma.h
25
mm/vma.h
@ -132,17 +132,6 @@ void anon_vma_interval_tree_pre_update_vma(struct vm_area_struct *vma);
|
||||
/* Required for expand_downwards(). */
|
||||
void anon_vma_interval_tree_post_update_vma(struct vm_area_struct *vma);
|
||||
|
||||
/* Required for do_brk_flags(). */
|
||||
void vma_prepare(struct vma_prepare *vp);
|
||||
|
||||
/* Required for do_brk_flags(). */
|
||||
void init_vma_prep(struct vma_prepare *vp,
|
||||
struct vm_area_struct *vma);
|
||||
|
||||
/* Required for do_brk_flags(). */
|
||||
void vma_complete(struct vma_prepare *vp,
|
||||
struct vma_iterator *vmi, struct mm_struct *mm);
|
||||
|
||||
int vma_expand(struct vma_merge_struct *vmg);
|
||||
int vma_shrink(struct vma_iterator *vmi, struct vm_area_struct *vma,
|
||||
unsigned long start, unsigned long end, pgoff_t pgoff);
|
||||
@ -277,20 +266,6 @@ void remove_vma(struct vm_area_struct *vma, bool unreachable, bool closed);
|
||||
void unmap_region(struct ma_state *mas, struct vm_area_struct *vma,
|
||||
struct vm_area_struct *prev, struct vm_area_struct *next);
|
||||
|
||||
/*
|
||||
* Can we merge the VMA described by vmg into the following VMA vmg->next?
|
||||
*
|
||||
* Required by mmap_region().
|
||||
*/
|
||||
bool can_vma_merge_before(struct vma_merge_struct *vmg);
|
||||
|
||||
/*
|
||||
* Can we merge the VMA described by vmg into the preceding VMA vmg->prev?
|
||||
*
|
||||
* Required by mmap_region() and do_brk_flags().
|
||||
*/
|
||||
bool can_vma_merge_after(struct vma_merge_struct *vmg);
|
||||
|
||||
/* We are about to modify the VMA's flags. */
|
||||
struct vm_area_struct *vma_modify_flags(struct vma_iterator *vmi,
|
||||
struct vm_area_struct *prev, struct vm_area_struct *vma,
|
||||
|
Loading…
Reference in New Issue
Block a user