mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2024-12-28 16:52:18 +00:00
9519a9fab0
A rolling buffer is a series of folios held in a list of folio_queues. New folios and folio_queue structs may be inserted at the head simultaneously with spent ones being removed from the tail without the need for locking. The rolling buffer includes an iov_iter and it has to be careful managing this as the list of folio_queues is extended such that an oops doesn't incurred because the iterator was pointing to the end of a folio_queue segment that got appended to and then removed. We need to use the mechanism twice, once for read and once for write, and, in future patches, we will use a second rolling buffer to handle bounce buffering for content encryption. Signed-off-by: David Howells <dhowells@redhat.com> Link: https://lore.kernel.org/r/20241108173236.1382366-8-dhowells@redhat.com cc: Jeff Layton <jlayton@kernel.org> cc: netfs@lists.linux.dev cc: linux-fsdevel@vger.kernel.org Signed-off-by: Christian Brauner <brauner@kernel.org>
226 lines
5.6 KiB
C
226 lines
5.6 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/* Rolling buffer helpers
|
|
*
|
|
* Copyright (C) 2024 Red Hat, Inc. All Rights Reserved.
|
|
* Written by David Howells (dhowells@redhat.com)
|
|
*/
|
|
|
|
#include <linux/bitops.h>
|
|
#include <linux/pagemap.h>
|
|
#include <linux/rolling_buffer.h>
|
|
#include <linux/slab.h>
|
|
#include "internal.h"
|
|
|
|
static atomic_t debug_ids;
|
|
|
|
/**
|
|
* netfs_folioq_alloc - Allocate a folio_queue struct
|
|
* @rreq_id: Associated debugging ID for tracing purposes
|
|
* @gfp: Allocation constraints
|
|
* @trace: Trace tag to indicate the purpose of the allocation
|
|
*
|
|
* Allocate, initialise and account the folio_queue struct and log a trace line
|
|
* to mark the allocation.
|
|
*/
|
|
struct folio_queue *netfs_folioq_alloc(unsigned int rreq_id, gfp_t gfp,
|
|
unsigned int /*enum netfs_folioq_trace*/ trace)
|
|
{
|
|
struct folio_queue *fq;
|
|
|
|
fq = kmalloc(sizeof(*fq), gfp);
|
|
if (fq) {
|
|
netfs_stat(&netfs_n_folioq);
|
|
folioq_init(fq, rreq_id);
|
|
fq->debug_id = atomic_inc_return(&debug_ids);
|
|
trace_netfs_folioq(fq, trace);
|
|
}
|
|
return fq;
|
|
}
|
|
EXPORT_SYMBOL(netfs_folioq_alloc);
|
|
|
|
/**
|
|
* netfs_folioq_free - Free a folio_queue struct
|
|
* @folioq: The object to free
|
|
* @trace: Trace tag to indicate which free
|
|
*
|
|
* Free and unaccount the folio_queue struct.
|
|
*/
|
|
void netfs_folioq_free(struct folio_queue *folioq,
|
|
unsigned int /*enum netfs_trace_folioq*/ trace)
|
|
{
|
|
trace_netfs_folioq(folioq, trace);
|
|
netfs_stat_d(&netfs_n_folioq);
|
|
kfree(folioq);
|
|
}
|
|
EXPORT_SYMBOL(netfs_folioq_free);
|
|
|
|
/*
|
|
* Initialise a rolling buffer. We allocate an empty folio queue struct to so
|
|
* that the pointers can be independently driven by the producer and the
|
|
* consumer.
|
|
*/
|
|
int rolling_buffer_init(struct rolling_buffer *roll, unsigned int rreq_id,
|
|
unsigned int direction)
|
|
{
|
|
struct folio_queue *fq;
|
|
|
|
fq = netfs_folioq_alloc(rreq_id, GFP_NOFS, netfs_trace_folioq_rollbuf_init);
|
|
if (!fq)
|
|
return -ENOMEM;
|
|
|
|
roll->head = fq;
|
|
roll->tail = fq;
|
|
iov_iter_folio_queue(&roll->iter, direction, fq, 0, 0, 0);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Add another folio_queue to a rolling buffer if there's no space left.
|
|
*/
|
|
int rolling_buffer_make_space(struct rolling_buffer *roll)
|
|
{
|
|
struct folio_queue *fq, *head = roll->head;
|
|
|
|
if (!folioq_full(head))
|
|
return 0;
|
|
|
|
fq = netfs_folioq_alloc(head->rreq_id, GFP_NOFS, netfs_trace_folioq_make_space);
|
|
if (!fq)
|
|
return -ENOMEM;
|
|
fq->prev = head;
|
|
|
|
roll->head = fq;
|
|
if (folioq_full(head)) {
|
|
/* Make sure we don't leave the master iterator pointing to a
|
|
* block that might get immediately consumed.
|
|
*/
|
|
if (roll->iter.folioq == head &&
|
|
roll->iter.folioq_slot == folioq_nr_slots(head)) {
|
|
roll->iter.folioq = fq;
|
|
roll->iter.folioq_slot = 0;
|
|
}
|
|
}
|
|
|
|
/* Make sure the initialisation is stored before the next pointer.
|
|
*
|
|
* [!] NOTE: After we set head->next, the consumer is at liberty to
|
|
* immediately delete the old head.
|
|
*/
|
|
smp_store_release(&head->next, fq);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Decant the list of folios to read into a rolling buffer.
|
|
*/
|
|
ssize_t rolling_buffer_load_from_ra(struct rolling_buffer *roll,
|
|
struct readahead_control *ractl,
|
|
struct folio_batch *put_batch)
|
|
{
|
|
struct folio_queue *fq;
|
|
struct page **vec;
|
|
int nr, ix, to;
|
|
ssize_t size = 0;
|
|
|
|
if (rolling_buffer_make_space(roll) < 0)
|
|
return -ENOMEM;
|
|
|
|
fq = roll->head;
|
|
vec = (struct page **)fq->vec.folios;
|
|
nr = __readahead_batch(ractl, vec + folio_batch_count(&fq->vec),
|
|
folio_batch_space(&fq->vec));
|
|
ix = fq->vec.nr;
|
|
to = ix + nr;
|
|
fq->vec.nr = to;
|
|
for (; ix < to; ix++) {
|
|
struct folio *folio = folioq_folio(fq, ix);
|
|
unsigned int order = folio_order(folio);
|
|
|
|
fq->orders[ix] = order;
|
|
size += PAGE_SIZE << order;
|
|
trace_netfs_folio(folio, netfs_folio_trace_read);
|
|
if (!folio_batch_add(put_batch, folio))
|
|
folio_batch_release(put_batch);
|
|
}
|
|
WRITE_ONCE(roll->iter.count, roll->iter.count + size);
|
|
|
|
/* Store the counter after setting the slot. */
|
|
smp_store_release(&roll->next_head_slot, to);
|
|
|
|
for (; ix < folioq_nr_slots(fq); ix++)
|
|
folioq_clear(fq, ix);
|
|
|
|
return size;
|
|
}
|
|
|
|
/*
|
|
* Append a folio to the rolling buffer.
|
|
*/
|
|
ssize_t rolling_buffer_append(struct rolling_buffer *roll, struct folio *folio,
|
|
unsigned int flags)
|
|
{
|
|
ssize_t size = folio_size(folio);
|
|
int slot;
|
|
|
|
if (rolling_buffer_make_space(roll) < 0)
|
|
return -ENOMEM;
|
|
|
|
slot = folioq_append(roll->head, folio);
|
|
if (flags & ROLLBUF_MARK_1)
|
|
folioq_mark(roll->head, slot);
|
|
if (flags & ROLLBUF_MARK_2)
|
|
folioq_mark2(roll->head, slot);
|
|
|
|
WRITE_ONCE(roll->iter.count, roll->iter.count + size);
|
|
|
|
/* Store the counter after setting the slot. */
|
|
smp_store_release(&roll->next_head_slot, slot);
|
|
return size;
|
|
}
|
|
|
|
/*
|
|
* Delete a spent buffer from a rolling queue and return the next in line. We
|
|
* don't return the last buffer to keep the pointers independent, but return
|
|
* NULL instead.
|
|
*/
|
|
struct folio_queue *rolling_buffer_delete_spent(struct rolling_buffer *roll)
|
|
{
|
|
struct folio_queue *spent = roll->tail, *next = READ_ONCE(spent->next);
|
|
|
|
if (!next)
|
|
return NULL;
|
|
next->prev = NULL;
|
|
netfs_folioq_free(spent, netfs_trace_folioq_delete);
|
|
roll->tail = next;
|
|
return next;
|
|
}
|
|
|
|
/*
|
|
* Clear out a rolling queue. Folios that have mark 1 set are put.
|
|
*/
|
|
void rolling_buffer_clear(struct rolling_buffer *roll)
|
|
{
|
|
struct folio_batch fbatch;
|
|
struct folio_queue *p;
|
|
|
|
folio_batch_init(&fbatch);
|
|
|
|
while ((p = roll->tail)) {
|
|
roll->tail = p->next;
|
|
for (int slot = 0; slot < folioq_count(p); slot++) {
|
|
struct folio *folio = folioq_folio(p, slot);
|
|
if (!folio)
|
|
continue;
|
|
if (folioq_is_marked(p, slot)) {
|
|
trace_netfs_folio(folio, netfs_folio_trace_put);
|
|
if (!folio_batch_add(&fbatch, folio))
|
|
folio_batch_release(&fbatch);
|
|
}
|
|
}
|
|
|
|
netfs_folioq_free(p, netfs_trace_folioq_clear);
|
|
}
|
|
|
|
folio_batch_release(&fbatch);
|
|
}
|