net: tls: factor out tls_*crypt_async_wait()
[ Upstream commit c57ca512f3b68ddcd62bda9cc24a8f5584ab01b1 ] Factor out waiting for async encrypt and decrypt to finish. There are already multiple copies and a subsequent fix will need more. No functional changes. Note that crypto_wait_req() returns wait->err Signed-off-by: Jakub Kicinski <kuba@kernel.org> Reviewed-by: Simon Horman <horms@kernel.org> Reviewed-by: Sabrina Dubroca <sd@queasysnail.net> Signed-off-by: David S. Miller <davem@davemloft.net> Stable-dep-of: aec7961916f3 ("tls: fix race between async notify and socket close") Signed-off-by: Sasha Levin <sashal@kernel.org>
This commit is contained in:
parent
5e01c54ebc
commit
de31df6f87
|
@ -230,6 +230,20 @@ static void tls_decrypt_done(void *data, int err)
|
|||
spin_unlock_bh(&ctx->decrypt_compl_lock);
|
||||
}
|
||||
|
||||
static int tls_decrypt_async_wait(struct tls_sw_context_rx *ctx)
|
||||
{
|
||||
int pending;
|
||||
|
||||
spin_lock_bh(&ctx->decrypt_compl_lock);
|
||||
reinit_completion(&ctx->async_wait.completion);
|
||||
pending = atomic_read(&ctx->decrypt_pending);
|
||||
spin_unlock_bh(&ctx->decrypt_compl_lock);
|
||||
if (pending)
|
||||
crypto_wait_req(-EINPROGRESS, &ctx->async_wait);
|
||||
|
||||
return ctx->async_wait.err;
|
||||
}
|
||||
|
||||
static int tls_do_decryption(struct sock *sk,
|
||||
struct scatterlist *sgin,
|
||||
struct scatterlist *sgout,
|
||||
|
@ -495,6 +509,28 @@ static void tls_encrypt_done(void *data, int err)
|
|||
schedule_delayed_work(&ctx->tx_work.work, 1);
|
||||
}
|
||||
|
||||
static int tls_encrypt_async_wait(struct tls_sw_context_tx *ctx)
|
||||
{
|
||||
int pending;
|
||||
|
||||
spin_lock_bh(&ctx->encrypt_compl_lock);
|
||||
ctx->async_notify = true;
|
||||
|
||||
pending = atomic_read(&ctx->encrypt_pending);
|
||||
spin_unlock_bh(&ctx->encrypt_compl_lock);
|
||||
if (pending)
|
||||
crypto_wait_req(-EINPROGRESS, &ctx->async_wait);
|
||||
else
|
||||
reinit_completion(&ctx->async_wait.completion);
|
||||
|
||||
/* There can be no concurrent accesses, since we have no
|
||||
* pending encrypt operations
|
||||
*/
|
||||
WRITE_ONCE(ctx->async_notify, false);
|
||||
|
||||
return ctx->async_wait.err;
|
||||
}
|
||||
|
||||
static int tls_do_encryption(struct sock *sk,
|
||||
struct tls_context *tls_ctx,
|
||||
struct tls_sw_context_tx *ctx,
|
||||
|
@ -984,7 +1020,6 @@ static int tls_sw_sendmsg_locked(struct sock *sk, struct msghdr *msg,
|
|||
int num_zc = 0;
|
||||
int orig_size;
|
||||
int ret = 0;
|
||||
int pending;
|
||||
|
||||
if (!eor && (msg->msg_flags & MSG_EOR))
|
||||
return -EINVAL;
|
||||
|
@ -1163,24 +1198,12 @@ trim_sgl:
|
|||
if (!num_async) {
|
||||
goto send_end;
|
||||
} else if (num_zc) {
|
||||
int err;
|
||||
|
||||
/* Wait for pending encryptions to get completed */
|
||||
spin_lock_bh(&ctx->encrypt_compl_lock);
|
||||
ctx->async_notify = true;
|
||||
|
||||
pending = atomic_read(&ctx->encrypt_pending);
|
||||
spin_unlock_bh(&ctx->encrypt_compl_lock);
|
||||
if (pending)
|
||||
crypto_wait_req(-EINPROGRESS, &ctx->async_wait);
|
||||
else
|
||||
reinit_completion(&ctx->async_wait.completion);
|
||||
|
||||
/* There can be no concurrent accesses, since we have no
|
||||
* pending encrypt operations
|
||||
*/
|
||||
WRITE_ONCE(ctx->async_notify, false);
|
||||
|
||||
if (ctx->async_wait.err) {
|
||||
ret = ctx->async_wait.err;
|
||||
err = tls_encrypt_async_wait(ctx);
|
||||
if (err) {
|
||||
ret = err;
|
||||
copied = 0;
|
||||
}
|
||||
}
|
||||
|
@ -1229,7 +1252,6 @@ void tls_sw_splice_eof(struct socket *sock)
|
|||
ssize_t copied = 0;
|
||||
bool retrying = false;
|
||||
int ret = 0;
|
||||
int pending;
|
||||
|
||||
if (!ctx->open_rec)
|
||||
return;
|
||||
|
@ -1264,22 +1286,7 @@ retry:
|
|||
}
|
||||
|
||||
/* Wait for pending encryptions to get completed */
|
||||
spin_lock_bh(&ctx->encrypt_compl_lock);
|
||||
ctx->async_notify = true;
|
||||
|
||||
pending = atomic_read(&ctx->encrypt_pending);
|
||||
spin_unlock_bh(&ctx->encrypt_compl_lock);
|
||||
if (pending)
|
||||
crypto_wait_req(-EINPROGRESS, &ctx->async_wait);
|
||||
else
|
||||
reinit_completion(&ctx->async_wait.completion);
|
||||
|
||||
/* There can be no concurrent accesses, since we have no pending
|
||||
* encrypt operations
|
||||
*/
|
||||
WRITE_ONCE(ctx->async_notify, false);
|
||||
|
||||
if (ctx->async_wait.err)
|
||||
if (tls_encrypt_async_wait(ctx))
|
||||
goto unlock;
|
||||
|
||||
/* Transmit if any encryptions have completed */
|
||||
|
@ -2109,16 +2116,10 @@ put_on_rx_list:
|
|||
|
||||
recv_end:
|
||||
if (async) {
|
||||
int ret, pending;
|
||||
int ret;
|
||||
|
||||
/* Wait for all previously submitted records to be decrypted */
|
||||
spin_lock_bh(&ctx->decrypt_compl_lock);
|
||||
reinit_completion(&ctx->async_wait.completion);
|
||||
pending = atomic_read(&ctx->decrypt_pending);
|
||||
spin_unlock_bh(&ctx->decrypt_compl_lock);
|
||||
ret = 0;
|
||||
if (pending)
|
||||
ret = crypto_wait_req(-EINPROGRESS, &ctx->async_wait);
|
||||
ret = tls_decrypt_async_wait(ctx);
|
||||
__skb_queue_purge(&ctx->async_hold);
|
||||
|
||||
if (ret) {
|
||||
|
@ -2435,16 +2436,9 @@ void tls_sw_release_resources_tx(struct sock *sk)
|
|||
struct tls_context *tls_ctx = tls_get_ctx(sk);
|
||||
struct tls_sw_context_tx *ctx = tls_sw_ctx_tx(tls_ctx);
|
||||
struct tls_rec *rec, *tmp;
|
||||
int pending;
|
||||
|
||||
/* Wait for any pending async encryptions to complete */
|
||||
spin_lock_bh(&ctx->encrypt_compl_lock);
|
||||
ctx->async_notify = true;
|
||||
pending = atomic_read(&ctx->encrypt_pending);
|
||||
spin_unlock_bh(&ctx->encrypt_compl_lock);
|
||||
|
||||
if (pending)
|
||||
crypto_wait_req(-EINPROGRESS, &ctx->async_wait);
|
||||
tls_encrypt_async_wait(ctx);
|
||||
|
||||
tls_tx_records(sk, -1);
|
||||
|
||||
|
|
Loading…
Reference in New Issue