mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2024-12-29 09:12:07 +00:00
rxrpc: Use irq-disabling spinlocks between app and I/O thread
Where a spinlock is used by both the application thread and the I/O thread, use irq-disabling locking so that an interrupt taken on the app thread doesn't also slow down the I/O thread. Signed-off-by: David Howells <dhowells@redhat.com> cc: Marc Dionne <marc.dionne@auristor.com> cc: linux-afs@lists.infradead.org Signed-off-by: Jakub Kicinski <kuba@kernel.org>
This commit is contained in:
parent
08d55d7cf3
commit
a2ea9a9072
@ -408,9 +408,9 @@ void rxrpc_kernel_shutdown_call(struct socket *sock, struct rxrpc_call *call)
|
||||
|
||||
/* Make sure we're not going to call back into a kernel service */
|
||||
if (call->notify_rx) {
|
||||
spin_lock(&call->notify_lock);
|
||||
spin_lock_irq(&call->notify_lock);
|
||||
call->notify_rx = rxrpc_dummy_notify_rx;
|
||||
spin_unlock(&call->notify_lock);
|
||||
spin_unlock_irq(&call->notify_lock);
|
||||
}
|
||||
}
|
||||
mutex_unlock(&call->user_mutex);
|
||||
|
@ -700,7 +700,6 @@ struct rxrpc_call {
|
||||
struct rxrpc_txqueue *send_queue; /* Queue that sendmsg is writing into */
|
||||
|
||||
/* Transmitted data tracking. */
|
||||
spinlock_t tx_lock; /* Transmit queue lock */
|
||||
struct rxrpc_txqueue *tx_queue; /* Start of transmission buffers */
|
||||
struct rxrpc_txqueue *tx_qtail; /* End of transmission buffers */
|
||||
rxrpc_seq_t tx_qbase; /* First slot in tx_queue */
|
||||
|
@ -188,8 +188,8 @@ void rxrpc_discard_prealloc(struct rxrpc_sock *rx)
|
||||
/* Make sure that there aren't any incoming calls in progress before we
|
||||
* clear the preallocation buffers.
|
||||
*/
|
||||
spin_lock(&rx->incoming_lock);
|
||||
spin_unlock(&rx->incoming_lock);
|
||||
spin_lock_irq(&rx->incoming_lock);
|
||||
spin_unlock_irq(&rx->incoming_lock);
|
||||
|
||||
head = b->peer_backlog_head;
|
||||
tail = b->peer_backlog_tail;
|
||||
@ -343,7 +343,7 @@ bool rxrpc_new_incoming_call(struct rxrpc_local *local,
|
||||
if (sp->hdr.type != RXRPC_PACKET_TYPE_DATA)
|
||||
return rxrpc_protocol_error(skb, rxrpc_eproto_no_service_call);
|
||||
|
||||
read_lock(&local->services_lock);
|
||||
read_lock_irq(&local->services_lock);
|
||||
|
||||
/* Weed out packets to services we're not offering. Packets that would
|
||||
* begin a call are explicitly rejected and the rest are just
|
||||
@ -399,12 +399,12 @@ bool rxrpc_new_incoming_call(struct rxrpc_local *local,
|
||||
spin_unlock(&conn->state_lock);
|
||||
|
||||
spin_unlock(&rx->incoming_lock);
|
||||
read_unlock(&local->services_lock);
|
||||
read_unlock_irq(&local->services_lock);
|
||||
|
||||
if (hlist_unhashed(&call->error_link)) {
|
||||
spin_lock(&call->peer->lock);
|
||||
spin_lock_irq(&call->peer->lock);
|
||||
hlist_add_head(&call->error_link, &call->peer->error_targets);
|
||||
spin_unlock(&call->peer->lock);
|
||||
spin_unlock_irq(&call->peer->lock);
|
||||
}
|
||||
|
||||
_leave(" = %p{%d}", call, call->debug_id);
|
||||
@ -413,20 +413,20 @@ bool rxrpc_new_incoming_call(struct rxrpc_local *local,
|
||||
return true;
|
||||
|
||||
unsupported_service:
|
||||
read_unlock(&local->services_lock);
|
||||
read_unlock_irq(&local->services_lock);
|
||||
return rxrpc_direct_abort(skb, rxrpc_abort_service_not_offered,
|
||||
RX_INVALID_OPERATION, -EOPNOTSUPP);
|
||||
unsupported_security:
|
||||
read_unlock(&local->services_lock);
|
||||
read_unlock_irq(&local->services_lock);
|
||||
return rxrpc_direct_abort(skb, rxrpc_abort_service_not_offered,
|
||||
RX_INVALID_OPERATION, -EKEYREJECTED);
|
||||
no_call:
|
||||
spin_unlock(&rx->incoming_lock);
|
||||
read_unlock(&local->services_lock);
|
||||
read_unlock_irq(&local->services_lock);
|
||||
_leave(" = f [%u]", skb->mark);
|
||||
return false;
|
||||
discard:
|
||||
read_unlock(&local->services_lock);
|
||||
read_unlock_irq(&local->services_lock);
|
||||
return true;
|
||||
}
|
||||
|
||||
|
@ -49,7 +49,7 @@ void rxrpc_poke_call(struct rxrpc_call *call, enum rxrpc_call_poke_trace what)
|
||||
bool busy;
|
||||
|
||||
if (!test_bit(RXRPC_CALL_DISCONNECTED, &call->flags)) {
|
||||
spin_lock_bh(&local->lock);
|
||||
spin_lock_irq(&local->lock);
|
||||
busy = !list_empty(&call->attend_link);
|
||||
trace_rxrpc_poke_call(call, busy, what);
|
||||
if (!busy && !rxrpc_try_get_call(call, rxrpc_call_get_poke))
|
||||
@ -57,7 +57,7 @@ void rxrpc_poke_call(struct rxrpc_call *call, enum rxrpc_call_poke_trace what)
|
||||
if (!busy) {
|
||||
list_add_tail(&call->attend_link, &local->call_attend_q);
|
||||
}
|
||||
spin_unlock_bh(&local->lock);
|
||||
spin_unlock_irq(&local->lock);
|
||||
if (!busy)
|
||||
rxrpc_wake_up_io_thread(local);
|
||||
}
|
||||
@ -151,7 +151,6 @@ struct rxrpc_call *rxrpc_alloc_call(struct rxrpc_sock *rx, gfp_t gfp,
|
||||
skb_queue_head_init(&call->rx_oos_queue);
|
||||
init_waitqueue_head(&call->waitq);
|
||||
spin_lock_init(&call->notify_lock);
|
||||
spin_lock_init(&call->tx_lock);
|
||||
refcount_set(&call->ref, 1);
|
||||
call->debug_id = debug_id;
|
||||
call->tx_total_len = -1;
|
||||
@ -302,9 +301,9 @@ static int rxrpc_connect_call(struct rxrpc_call *call, gfp_t gfp)
|
||||
|
||||
trace_rxrpc_client(NULL, -1, rxrpc_client_queue_new_call);
|
||||
rxrpc_get_call(call, rxrpc_call_get_io_thread);
|
||||
spin_lock(&local->client_call_lock);
|
||||
spin_lock_irq(&local->client_call_lock);
|
||||
list_add_tail(&call->wait_link, &local->new_client_calls);
|
||||
spin_unlock(&local->client_call_lock);
|
||||
spin_unlock_irq(&local->client_call_lock);
|
||||
rxrpc_wake_up_io_thread(local);
|
||||
return 0;
|
||||
|
||||
@ -434,7 +433,7 @@ struct rxrpc_call *rxrpc_new_client_call(struct rxrpc_sock *rx,
|
||||
|
||||
/*
|
||||
* Set up an incoming call. call->conn points to the connection.
|
||||
* This is called in BH context and isn't allowed to fail.
|
||||
* This is called with interrupts disabled and isn't allowed to fail.
|
||||
*/
|
||||
void rxrpc_incoming_call(struct rxrpc_sock *rx,
|
||||
struct rxrpc_call *call,
|
||||
@ -576,7 +575,7 @@ void rxrpc_release_call(struct rxrpc_sock *rx, struct rxrpc_call *call)
|
||||
rxrpc_put_call_slot(call);
|
||||
|
||||
/* Make sure we don't get any more notifications */
|
||||
spin_lock(&rx->recvmsg_lock);
|
||||
spin_lock_irq(&rx->recvmsg_lock);
|
||||
|
||||
if (!list_empty(&call->recvmsg_link)) {
|
||||
_debug("unlinking once-pending call %p { e=%lx f=%lx }",
|
||||
@ -589,7 +588,7 @@ void rxrpc_release_call(struct rxrpc_sock *rx, struct rxrpc_call *call)
|
||||
call->recvmsg_link.next = NULL;
|
||||
call->recvmsg_link.prev = NULL;
|
||||
|
||||
spin_unlock(&rx->recvmsg_lock);
|
||||
spin_unlock_irq(&rx->recvmsg_lock);
|
||||
if (put)
|
||||
rxrpc_put_call(call, rxrpc_call_put_unnotify);
|
||||
|
||||
|
@ -510,9 +510,9 @@ void rxrpc_connect_client_calls(struct rxrpc_local *local)
|
||||
struct rxrpc_call *call;
|
||||
LIST_HEAD(new_client_calls);
|
||||
|
||||
spin_lock(&local->client_call_lock);
|
||||
spin_lock_irq(&local->client_call_lock);
|
||||
list_splice_tail_init(&local->new_client_calls, &new_client_calls);
|
||||
spin_unlock(&local->client_call_lock);
|
||||
spin_unlock_irq(&local->client_call_lock);
|
||||
|
||||
while ((call = list_first_entry_or_null(&new_client_calls,
|
||||
struct rxrpc_call, wait_link))) {
|
||||
@ -547,9 +547,9 @@ void rxrpc_expose_client_call(struct rxrpc_call *call)
|
||||
set_bit(RXRPC_CONN_DONT_REUSE, &conn->flags);
|
||||
trace_rxrpc_client(conn, channel, rxrpc_client_exposed);
|
||||
|
||||
spin_lock(&call->peer->lock);
|
||||
spin_lock_irq(&call->peer->lock);
|
||||
hlist_add_head(&call->error_link, &call->peer->error_targets);
|
||||
spin_unlock(&call->peer->lock);
|
||||
spin_unlock_irq(&call->peer->lock);
|
||||
}
|
||||
}
|
||||
|
||||
@ -590,9 +590,9 @@ void rxrpc_disconnect_client_call(struct rxrpc_bundle *bundle, struct rxrpc_call
|
||||
ASSERTCMP(call->call_id, ==, 0);
|
||||
ASSERT(!test_bit(RXRPC_CALL_EXPOSED, &call->flags));
|
||||
/* May still be on ->new_client_calls. */
|
||||
spin_lock(&local->client_call_lock);
|
||||
spin_lock_irq(&local->client_call_lock);
|
||||
list_del_init(&call->wait_link);
|
||||
spin_unlock(&local->client_call_lock);
|
||||
spin_unlock_irq(&local->client_call_lock);
|
||||
return;
|
||||
}
|
||||
|
||||
|
@ -26,7 +26,7 @@ static bool rxrpc_set_conn_aborted(struct rxrpc_connection *conn, struct sk_buff
|
||||
bool aborted = false;
|
||||
|
||||
if (conn->state != RXRPC_CONN_ABORTED) {
|
||||
spin_lock(&conn->state_lock);
|
||||
spin_lock_irq(&conn->state_lock);
|
||||
if (conn->state != RXRPC_CONN_ABORTED) {
|
||||
conn->abort_code = abort_code;
|
||||
conn->error = err;
|
||||
@ -37,7 +37,7 @@ static bool rxrpc_set_conn_aborted(struct rxrpc_connection *conn, struct sk_buff
|
||||
set_bit(RXRPC_CONN_EV_ABORT_CALLS, &conn->events);
|
||||
aborted = true;
|
||||
}
|
||||
spin_unlock(&conn->state_lock);
|
||||
spin_unlock_irq(&conn->state_lock);
|
||||
}
|
||||
|
||||
return aborted;
|
||||
@ -261,10 +261,10 @@ static int rxrpc_process_event(struct rxrpc_connection *conn,
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
|
||||
spin_lock(&conn->state_lock);
|
||||
spin_lock_irq(&conn->state_lock);
|
||||
if (conn->state == RXRPC_CONN_SERVICE_CHALLENGING)
|
||||
conn->state = RXRPC_CONN_SERVICE;
|
||||
spin_unlock(&conn->state_lock);
|
||||
spin_unlock_irq(&conn->state_lock);
|
||||
|
||||
if (conn->state == RXRPC_CONN_SERVICE) {
|
||||
/* Offload call state flipping to the I/O thread. As
|
||||
|
@ -31,13 +31,13 @@ void rxrpc_poke_conn(struct rxrpc_connection *conn, enum rxrpc_conn_trace why)
|
||||
if (WARN_ON_ONCE(!local))
|
||||
return;
|
||||
|
||||
spin_lock_bh(&local->lock);
|
||||
spin_lock_irq(&local->lock);
|
||||
busy = !list_empty(&conn->attend_link);
|
||||
if (!busy) {
|
||||
rxrpc_get_connection(conn, why);
|
||||
list_add_tail(&conn->attend_link, &local->conn_attend_q);
|
||||
}
|
||||
spin_unlock_bh(&local->lock);
|
||||
spin_unlock_irq(&local->lock);
|
||||
rxrpc_wake_up_io_thread(local);
|
||||
}
|
||||
|
||||
@ -196,9 +196,9 @@ void rxrpc_disconnect_call(struct rxrpc_call *call)
|
||||
call->peer->cong_ssthresh = call->cong_ssthresh;
|
||||
|
||||
if (!hlist_unhashed(&call->error_link)) {
|
||||
spin_lock(&call->peer->lock);
|
||||
spin_lock_irq(&call->peer->lock);
|
||||
hlist_del_init(&call->error_link);
|
||||
spin_unlock(&call->peer->lock);
|
||||
spin_unlock_irq(&call->peer->lock);
|
||||
}
|
||||
|
||||
if (rxrpc_is_client_call(call)) {
|
||||
|
@ -424,7 +424,7 @@ static void rxrpc_input_queue_data(struct rxrpc_call *call, struct sk_buff *skb,
|
||||
struct rxrpc_skb_priv *sp = rxrpc_skb(skb);
|
||||
bool last = sp->hdr.flags & RXRPC_LAST_PACKET;
|
||||
|
||||
__skb_queue_tail(&call->recvmsg_queue, skb);
|
||||
skb_queue_tail(&call->recvmsg_queue, skb);
|
||||
rxrpc_input_update_ack_window(call, window, wtop);
|
||||
trace_rxrpc_receive(call, last ? why + 1 : why, sp->hdr.serial, sp->hdr.seq);
|
||||
if (last)
|
||||
@ -501,7 +501,6 @@ static void rxrpc_input_data_one(struct rxrpc_call *call, struct sk_buff *skb,
|
||||
|
||||
rxrpc_get_skb(skb, rxrpc_skb_get_to_recvmsg);
|
||||
|
||||
spin_lock(&call->recvmsg_queue.lock);
|
||||
rxrpc_input_queue_data(call, skb, window, wtop, rxrpc_receive_queue);
|
||||
*_notify = true;
|
||||
|
||||
@ -523,8 +522,6 @@ static void rxrpc_input_data_one(struct rxrpc_call *call, struct sk_buff *skb,
|
||||
rxrpc_receive_queue_oos);
|
||||
}
|
||||
|
||||
spin_unlock(&call->recvmsg_queue.lock);
|
||||
|
||||
call->ackr_sack_base = sack;
|
||||
} else {
|
||||
unsigned int slot;
|
||||
|
@ -500,9 +500,9 @@ int rxrpc_io_thread(void *data)
|
||||
}
|
||||
|
||||
/* Deal with connections that want immediate attention. */
|
||||
spin_lock_bh(&local->lock);
|
||||
spin_lock_irq(&local->lock);
|
||||
list_splice_tail_init(&local->conn_attend_q, &conn_attend_q);
|
||||
spin_unlock_bh(&local->lock);
|
||||
spin_unlock_irq(&local->lock);
|
||||
|
||||
while ((conn = list_first_entry_or_null(&conn_attend_q,
|
||||
struct rxrpc_connection,
|
||||
@ -519,9 +519,9 @@ int rxrpc_io_thread(void *data)
|
||||
rxrpc_discard_expired_client_conns(local);
|
||||
|
||||
/* Deal with calls that want immediate attention. */
|
||||
spin_lock_bh(&local->lock);
|
||||
spin_lock_irq(&local->lock);
|
||||
list_splice_tail_init(&local->call_attend_q, &call_attend_q);
|
||||
spin_unlock_bh(&local->lock);
|
||||
spin_unlock_irq(&local->lock);
|
||||
|
||||
while ((call = list_first_entry_or_null(&call_attend_q,
|
||||
struct rxrpc_call,
|
||||
|
@ -213,23 +213,23 @@ static void rxrpc_distribute_error(struct rxrpc_peer *peer, struct sk_buff *skb,
|
||||
struct rxrpc_call *call;
|
||||
HLIST_HEAD(error_targets);
|
||||
|
||||
spin_lock(&peer->lock);
|
||||
spin_lock_irq(&peer->lock);
|
||||
hlist_move_list(&peer->error_targets, &error_targets);
|
||||
|
||||
while (!hlist_empty(&error_targets)) {
|
||||
call = hlist_entry(error_targets.first,
|
||||
struct rxrpc_call, error_link);
|
||||
hlist_del_init(&call->error_link);
|
||||
spin_unlock(&peer->lock);
|
||||
spin_unlock_irq(&peer->lock);
|
||||
|
||||
rxrpc_see_call(call, rxrpc_call_see_distribute_error);
|
||||
rxrpc_set_call_completion(call, compl, 0, -err);
|
||||
rxrpc_input_call_event(call);
|
||||
|
||||
spin_lock(&peer->lock);
|
||||
spin_lock_irq(&peer->lock);
|
||||
}
|
||||
|
||||
spin_unlock(&peer->lock);
|
||||
spin_unlock_irq(&peer->lock);
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -320,6 +320,7 @@ static void rxrpc_free_peer(struct rxrpc_peer *peer)
|
||||
* Set up a new incoming peer. There shouldn't be any other matching peers
|
||||
* since we've already done a search in the list from the non-reentrant context
|
||||
* (the data_ready handler) that is the only place we can add new peers.
|
||||
* Called with interrupts disabled.
|
||||
*/
|
||||
void rxrpc_new_incoming_peer(struct rxrpc_local *local, struct rxrpc_peer *peer)
|
||||
{
|
||||
|
@ -36,16 +36,16 @@ void rxrpc_notify_socket(struct rxrpc_call *call)
|
||||
sk = &rx->sk;
|
||||
if (rx && sk->sk_state < RXRPC_CLOSE) {
|
||||
if (call->notify_rx) {
|
||||
spin_lock(&call->notify_lock);
|
||||
spin_lock_irq(&call->notify_lock);
|
||||
call->notify_rx(sk, call, call->user_call_ID);
|
||||
spin_unlock(&call->notify_lock);
|
||||
spin_unlock_irq(&call->notify_lock);
|
||||
} else {
|
||||
spin_lock(&rx->recvmsg_lock);
|
||||
spin_lock_irq(&rx->recvmsg_lock);
|
||||
if (list_empty(&call->recvmsg_link)) {
|
||||
rxrpc_get_call(call, rxrpc_call_get_notify_socket);
|
||||
list_add_tail(&call->recvmsg_link, &rx->recvmsg_q);
|
||||
}
|
||||
spin_unlock(&rx->recvmsg_lock);
|
||||
spin_unlock_irq(&rx->recvmsg_lock);
|
||||
|
||||
if (!sock_flag(sk, SOCK_DEAD)) {
|
||||
_debug("call %ps", sk->sk_data_ready);
|
||||
@ -337,14 +337,14 @@ int rxrpc_recvmsg(struct socket *sock, struct msghdr *msg, size_t len,
|
||||
* We also want to weed out calls that got requeued whilst we were
|
||||
* shovelling data out.
|
||||
*/
|
||||
spin_lock(&rx->recvmsg_lock);
|
||||
spin_lock_irq(&rx->recvmsg_lock);
|
||||
l = rx->recvmsg_q.next;
|
||||
call = list_entry(l, struct rxrpc_call, recvmsg_link);
|
||||
|
||||
if (!rxrpc_call_is_complete(call) &&
|
||||
skb_queue_empty(&call->recvmsg_queue)) {
|
||||
list_del_init(&call->recvmsg_link);
|
||||
spin_unlock(&rx->recvmsg_lock);
|
||||
spin_unlock_irq(&rx->recvmsg_lock);
|
||||
release_sock(&rx->sk);
|
||||
trace_rxrpc_recvmsg(call->debug_id, rxrpc_recvmsg_unqueue, 0);
|
||||
rxrpc_put_call(call, rxrpc_call_put_recvmsg);
|
||||
@ -355,7 +355,7 @@ int rxrpc_recvmsg(struct socket *sock, struct msghdr *msg, size_t len,
|
||||
list_del_init(&call->recvmsg_link);
|
||||
else
|
||||
rxrpc_get_call(call, rxrpc_call_get_recvmsg);
|
||||
spin_unlock(&rx->recvmsg_lock);
|
||||
spin_unlock_irq(&rx->recvmsg_lock);
|
||||
|
||||
call_debug_id = call->debug_id;
|
||||
trace_rxrpc_recvmsg(call_debug_id, rxrpc_recvmsg_dequeue, 0);
|
||||
@ -445,9 +445,9 @@ int rxrpc_recvmsg(struct socket *sock, struct msghdr *msg, size_t len,
|
||||
|
||||
error_requeue_call:
|
||||
if (!(flags & MSG_PEEK)) {
|
||||
spin_lock(&rx->recvmsg_lock);
|
||||
spin_lock_irq(&rx->recvmsg_lock);
|
||||
list_add(&call->recvmsg_link, &rx->recvmsg_q);
|
||||
spin_unlock(&rx->recvmsg_lock);
|
||||
spin_unlock_irq(&rx->recvmsg_lock);
|
||||
trace_rxrpc_recvmsg(call_debug_id, rxrpc_recvmsg_requeue, 0);
|
||||
} else {
|
||||
rxrpc_put_call(call, rxrpc_call_put_recvmsg);
|
||||
|
@ -114,10 +114,10 @@ int rxrpc_init_client_conn_security(struct rxrpc_connection *conn)
|
||||
if (conn->state == RXRPC_CONN_CLIENT_UNSECURED) {
|
||||
ret = conn->security->init_connection_security(conn, token);
|
||||
if (ret == 0) {
|
||||
spin_lock(&conn->state_lock);
|
||||
spin_lock_irq(&conn->state_lock);
|
||||
if (conn->state == RXRPC_CONN_CLIENT_UNSECURED)
|
||||
conn->state = RXRPC_CONN_CLIENT;
|
||||
spin_unlock(&conn->state_lock);
|
||||
spin_unlock_irq(&conn->state_lock);
|
||||
}
|
||||
}
|
||||
mutex_unlock(&conn->security_lock);
|
||||
|
@ -261,7 +261,6 @@ static void rxrpc_queue_packet(struct rxrpc_sock *rx, struct rxrpc_call *call,
|
||||
trace_rxrpc_tq(call, sq, seq, rxrpc_tq_queue);
|
||||
|
||||
/* Add the packet to the call's output buffer */
|
||||
spin_lock(&call->tx_lock);
|
||||
poke = (READ_ONCE(call->tx_bottom) == call->send_top);
|
||||
sq->bufs[ix] = txb;
|
||||
/* Order send_top after the queue->next pointer and txb content. */
|
||||
@ -270,7 +269,6 @@ static void rxrpc_queue_packet(struct rxrpc_sock *rx, struct rxrpc_call *call,
|
||||
rxrpc_notify_end_tx(rx, call, notify_end_tx);
|
||||
call->send_queue = NULL;
|
||||
}
|
||||
spin_unlock(&call->tx_lock);
|
||||
|
||||
if (poke)
|
||||
rxrpc_poke_call(call, rxrpc_call_poke_start);
|
||||
|
Loading…
Reference in New Issue
Block a user