mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2024-12-29 17:22:07 +00:00
maple_tree: Add mtree_alloc_cyclic()
I need a cyclic allocator for the simple_offset implementation in fs/libfs.c. Signed-off-by: Chuck Lever <chuck.lever@oracle.com> Link: https://lore.kernel.org/r/170820144179.6328.12838600511394432325.stgit@91.116.238.104.host.secureserver.net Signed-off-by: Christian Brauner <brauner@kernel.org>
This commit is contained in:
parent
ecba88a3b3
commit
9b6713cc75
@ -171,6 +171,7 @@ enum maple_type {
|
|||||||
#define MT_FLAGS_LOCK_IRQ 0x100
|
#define MT_FLAGS_LOCK_IRQ 0x100
|
||||||
#define MT_FLAGS_LOCK_BH 0x200
|
#define MT_FLAGS_LOCK_BH 0x200
|
||||||
#define MT_FLAGS_LOCK_EXTERN 0x300
|
#define MT_FLAGS_LOCK_EXTERN 0x300
|
||||||
|
#define MT_FLAGS_ALLOC_WRAPPED 0x0800
|
||||||
|
|
||||||
#define MAPLE_HEIGHT_MAX 31
|
#define MAPLE_HEIGHT_MAX 31
|
||||||
|
|
||||||
@ -319,6 +320,9 @@ int mtree_insert_range(struct maple_tree *mt, unsigned long first,
|
|||||||
int mtree_alloc_range(struct maple_tree *mt, unsigned long *startp,
|
int mtree_alloc_range(struct maple_tree *mt, unsigned long *startp,
|
||||||
void *entry, unsigned long size, unsigned long min,
|
void *entry, unsigned long size, unsigned long min,
|
||||||
unsigned long max, gfp_t gfp);
|
unsigned long max, gfp_t gfp);
|
||||||
|
int mtree_alloc_cyclic(struct maple_tree *mt, unsigned long *startp,
|
||||||
|
void *entry, unsigned long range_lo, unsigned long range_hi,
|
||||||
|
unsigned long *next, gfp_t gfp);
|
||||||
int mtree_alloc_rrange(struct maple_tree *mt, unsigned long *startp,
|
int mtree_alloc_rrange(struct maple_tree *mt, unsigned long *startp,
|
||||||
void *entry, unsigned long size, unsigned long min,
|
void *entry, unsigned long size, unsigned long min,
|
||||||
unsigned long max, gfp_t gfp);
|
unsigned long max, gfp_t gfp);
|
||||||
@ -499,6 +503,9 @@ void *mas_find_range(struct ma_state *mas, unsigned long max);
|
|||||||
void *mas_find_rev(struct ma_state *mas, unsigned long min);
|
void *mas_find_rev(struct ma_state *mas, unsigned long min);
|
||||||
void *mas_find_range_rev(struct ma_state *mas, unsigned long max);
|
void *mas_find_range_rev(struct ma_state *mas, unsigned long max);
|
||||||
int mas_preallocate(struct ma_state *mas, void *entry, gfp_t gfp);
|
int mas_preallocate(struct ma_state *mas, void *entry, gfp_t gfp);
|
||||||
|
int mas_alloc_cyclic(struct ma_state *mas, unsigned long *startp,
|
||||||
|
void *entry, unsigned long range_lo, unsigned long range_hi,
|
||||||
|
unsigned long *next, gfp_t gfp);
|
||||||
|
|
||||||
bool mas_nomem(struct ma_state *mas, gfp_t gfp);
|
bool mas_nomem(struct ma_state *mas, gfp_t gfp);
|
||||||
void mas_pause(struct ma_state *mas);
|
void mas_pause(struct ma_state *mas);
|
||||||
|
@ -4290,6 +4290,56 @@ static inline void *mas_insert(struct ma_state *mas, void *entry)
|
|||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* mas_alloc_cyclic() - Internal call to find somewhere to store an entry
|
||||||
|
* @mas: The maple state.
|
||||||
|
* @startp: Pointer to ID.
|
||||||
|
* @range_lo: Lower bound of range to search.
|
||||||
|
* @range_hi: Upper bound of range to search.
|
||||||
|
* @entry: The entry to store.
|
||||||
|
* @next: Pointer to next ID to allocate.
|
||||||
|
* @gfp: The GFP_FLAGS to use for allocations.
|
||||||
|
*
|
||||||
|
* Return: 0 if the allocation succeeded without wrapping, 1 if the
|
||||||
|
* allocation succeeded after wrapping, or -EBUSY if there are no
|
||||||
|
* free entries.
|
||||||
|
*/
|
||||||
|
int mas_alloc_cyclic(struct ma_state *mas, unsigned long *startp,
|
||||||
|
void *entry, unsigned long range_lo, unsigned long range_hi,
|
||||||
|
unsigned long *next, gfp_t gfp)
|
||||||
|
{
|
||||||
|
unsigned long min = range_lo;
|
||||||
|
int ret = 0;
|
||||||
|
|
||||||
|
range_lo = max(min, *next);
|
||||||
|
ret = mas_empty_area(mas, range_lo, range_hi, 1);
|
||||||
|
if ((mas->tree->ma_flags & MT_FLAGS_ALLOC_WRAPPED) && ret == 0) {
|
||||||
|
mas->tree->ma_flags &= ~MT_FLAGS_ALLOC_WRAPPED;
|
||||||
|
ret = 1;
|
||||||
|
}
|
||||||
|
if (ret < 0 && range_lo > min) {
|
||||||
|
ret = mas_empty_area(mas, min, range_hi, 1);
|
||||||
|
if (ret == 0)
|
||||||
|
ret = 1;
|
||||||
|
}
|
||||||
|
if (ret < 0)
|
||||||
|
return ret;
|
||||||
|
|
||||||
|
do {
|
||||||
|
mas_insert(mas, entry);
|
||||||
|
} while (mas_nomem(mas, gfp));
|
||||||
|
if (mas_is_err(mas))
|
||||||
|
return xa_err(mas->node);
|
||||||
|
|
||||||
|
*startp = mas->index;
|
||||||
|
*next = *startp + 1;
|
||||||
|
if (*next == 0)
|
||||||
|
mas->tree->ma_flags |= MT_FLAGS_ALLOC_WRAPPED;
|
||||||
|
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(mas_alloc_cyclic);
|
||||||
|
|
||||||
static __always_inline void mas_rewalk(struct ma_state *mas, unsigned long index)
|
static __always_inline void mas_rewalk(struct ma_state *mas, unsigned long index)
|
||||||
{
|
{
|
||||||
retry:
|
retry:
|
||||||
@ -6443,6 +6493,49 @@ int mtree_alloc_range(struct maple_tree *mt, unsigned long *startp,
|
|||||||
}
|
}
|
||||||
EXPORT_SYMBOL(mtree_alloc_range);
|
EXPORT_SYMBOL(mtree_alloc_range);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* mtree_alloc_cyclic() - Find somewhere to store this entry in the tree.
|
||||||
|
* @mt: The maple tree.
|
||||||
|
* @startp: Pointer to ID.
|
||||||
|
* @range_lo: Lower bound of range to search.
|
||||||
|
* @range_hi: Upper bound of range to search.
|
||||||
|
* @entry: The entry to store.
|
||||||
|
* @next: Pointer to next ID to allocate.
|
||||||
|
* @gfp: The GFP_FLAGS to use for allocations.
|
||||||
|
*
|
||||||
|
* Finds an empty entry in @mt after @next, stores the new index into
|
||||||
|
* the @id pointer, stores the entry at that index, then updates @next.
|
||||||
|
*
|
||||||
|
* @mt must be initialized with the MT_FLAGS_ALLOC_RANGE flag.
|
||||||
|
*
|
||||||
|
* Context: Any context. Takes and releases the mt.lock. May sleep if
|
||||||
|
* the @gfp flags permit.
|
||||||
|
*
|
||||||
|
* Return: 0 if the allocation succeeded without wrapping, 1 if the
|
||||||
|
* allocation succeeded after wrapping, -ENOMEM if memory could not be
|
||||||
|
* allocated, -EINVAL if @mt cannot be used, or -EBUSY if there are no
|
||||||
|
* free entries.
|
||||||
|
*/
|
||||||
|
int mtree_alloc_cyclic(struct maple_tree *mt, unsigned long *startp,
|
||||||
|
void *entry, unsigned long range_lo, unsigned long range_hi,
|
||||||
|
unsigned long *next, gfp_t gfp)
|
||||||
|
{
|
||||||
|
int ret;
|
||||||
|
|
||||||
|
MA_STATE(mas, mt, 0, 0);
|
||||||
|
|
||||||
|
if (!mt_is_alloc(mt))
|
||||||
|
return -EINVAL;
|
||||||
|
if (WARN_ON_ONCE(mt_is_reserved(entry)))
|
||||||
|
return -EINVAL;
|
||||||
|
mtree_lock(mt);
|
||||||
|
ret = mas_alloc_cyclic(&mas, startp, entry, range_lo, range_hi,
|
||||||
|
next, gfp);
|
||||||
|
mtree_unlock(mt);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(mtree_alloc_cyclic);
|
||||||
|
|
||||||
int mtree_alloc_rrange(struct maple_tree *mt, unsigned long *startp,
|
int mtree_alloc_rrange(struct maple_tree *mt, unsigned long *startp,
|
||||||
void *entry, unsigned long size, unsigned long min,
|
void *entry, unsigned long size, unsigned long min,
|
||||||
unsigned long max, gfp_t gfp)
|
unsigned long max, gfp_t gfp)
|
||||||
|
Loading…
Reference in New Issue
Block a user