mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-01-17 02:15:57 +00:00
lib: percpu_count_sum()
Provide an accurate version of percpu_counter_read. Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
52d9f3b409
commit
bf1d89c813
@ -34,13 +34,24 @@ void percpu_counter_init(struct percpu_counter *fbc, s64 amount);
|
|||||||
void percpu_counter_destroy(struct percpu_counter *fbc);
|
void percpu_counter_destroy(struct percpu_counter *fbc);
|
||||||
void percpu_counter_set(struct percpu_counter *fbc, s64 amount);
|
void percpu_counter_set(struct percpu_counter *fbc, s64 amount);
|
||||||
void __percpu_counter_add(struct percpu_counter *fbc, s64 amount, s32 batch);
|
void __percpu_counter_add(struct percpu_counter *fbc, s64 amount, s32 batch);
|
||||||
s64 percpu_counter_sum_positive(struct percpu_counter *fbc);
|
s64 __percpu_counter_sum(struct percpu_counter *fbc);
|
||||||
|
|
||||||
static inline void percpu_counter_add(struct percpu_counter *fbc, s64 amount)
|
static inline void percpu_counter_add(struct percpu_counter *fbc, s64 amount)
|
||||||
{
|
{
|
||||||
__percpu_counter_add(fbc, amount, FBC_BATCH);
|
__percpu_counter_add(fbc, amount, FBC_BATCH);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline s64 percpu_counter_sum_positive(struct percpu_counter *fbc)
|
||||||
|
{
|
||||||
|
s64 ret = __percpu_counter_sum(fbc);
|
||||||
|
return ret < 0 ? 0 : ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline s64 percpu_counter_sum(struct percpu_counter *fbc)
|
||||||
|
{
|
||||||
|
return __percpu_counter_sum(fbc);
|
||||||
|
}
|
||||||
|
|
||||||
static inline s64 percpu_counter_read(struct percpu_counter *fbc)
|
static inline s64 percpu_counter_read(struct percpu_counter *fbc)
|
||||||
{
|
{
|
||||||
return fbc->count;
|
return fbc->count;
|
||||||
@ -107,6 +118,11 @@ static inline s64 percpu_counter_sum_positive(struct percpu_counter *fbc)
|
|||||||
return percpu_counter_read_positive(fbc);
|
return percpu_counter_read_positive(fbc);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline s64 percpu_counter_sum(struct percpu_counter *fbc)
|
||||||
|
{
|
||||||
|
return percpu_counter_read(fbc);
|
||||||
|
}
|
||||||
|
|
||||||
#endif /* CONFIG_SMP */
|
#endif /* CONFIG_SMP */
|
||||||
|
|
||||||
static inline void percpu_counter_inc(struct percpu_counter *fbc)
|
static inline void percpu_counter_inc(struct percpu_counter *fbc)
|
||||||
|
@ -52,7 +52,7 @@ EXPORT_SYMBOL(__percpu_counter_add);
|
|||||||
* Add up all the per-cpu counts, return the result. This is a more accurate
|
* Add up all the per-cpu counts, return the result. This is a more accurate
|
||||||
* but much slower version of percpu_counter_read_positive()
|
* but much slower version of percpu_counter_read_positive()
|
||||||
*/
|
*/
|
||||||
s64 percpu_counter_sum_positive(struct percpu_counter *fbc)
|
s64 __percpu_counter_sum(struct percpu_counter *fbc)
|
||||||
{
|
{
|
||||||
s64 ret;
|
s64 ret;
|
||||||
int cpu;
|
int cpu;
|
||||||
@ -64,9 +64,9 @@ s64 percpu_counter_sum_positive(struct percpu_counter *fbc)
|
|||||||
ret += *pcount;
|
ret += *pcount;
|
||||||
}
|
}
|
||||||
spin_unlock(&fbc->lock);
|
spin_unlock(&fbc->lock);
|
||||||
return ret < 0 ? 0 : ret;
|
return ret;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(percpu_counter_sum_positive);
|
EXPORT_SYMBOL(__percpu_counter_sum);
|
||||||
|
|
||||||
void percpu_counter_init(struct percpu_counter *fbc, s64 amount)
|
void percpu_counter_init(struct percpu_counter *fbc, s64 amount)
|
||||||
{
|
{
|
||||||
|
Loading…
x
Reference in New Issue
Block a user