2017-06-15 02:37:39 +08:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2016-2017, Mellanox Technologies. All rights reserved.
|
|
|
|
* Copyright (c) 2016-2017, Dave Watson <davejwatson@fb.com>. All rights reserved.
|
|
|
|
*
|
|
|
|
* This software is available to you under a choice of one of two
|
|
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
|
|
* General Public License (GPL) Version 2, available from the file
|
|
|
|
* COPYING in the main directory of this source tree, or the
|
|
|
|
* OpenIB.org BSD license below:
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or
|
|
|
|
* without modification, are permitted provided that the following
|
|
|
|
* conditions are met:
|
|
|
|
*
|
|
|
|
* - Redistributions of source code must retain the above
|
|
|
|
* copyright notice, this list of conditions and the following
|
|
|
|
* disclaimer.
|
|
|
|
*
|
|
|
|
* - Redistributions in binary form must reproduce the above
|
|
|
|
* copyright notice, this list of conditions and the following
|
|
|
|
* disclaimer in the documentation and/or other materials
|
|
|
|
* provided with the distribution.
|
|
|
|
*
|
|
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
|
|
* SOFTWARE.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef _TLS_OFFLOAD_H
|
|
|
|
#define _TLS_OFFLOAD_H
|
|
|
|
|
|
|
|
#include <linux/types.h>
|
2017-11-14 11:30:11 +08:00
|
|
|
#include <asm/byteorder.h>
|
2018-02-01 00:04:37 +08:00
|
|
|
#include <linux/crypto.h>
|
2017-11-14 11:30:11 +08:00
|
|
|
#include <linux/socket.h>
|
|
|
|
#include <linux/tcp.h>
|
2019-11-06 06:24:35 +08:00
|
|
|
#include <linux/mutex.h>
|
2019-06-06 05:11:39 +08:00
|
|
|
#include <linux/netdevice.h>
|
2019-08-30 18:25:47 +08:00
|
|
|
#include <linux/rcupdate.h>
|
2018-10-13 08:45:59 +08:00
|
|
|
|
2019-10-05 07:19:24 +08:00
|
|
|
#include <net/net_namespace.h>
|
2017-11-14 11:30:11 +08:00
|
|
|
#include <net/tcp.h>
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 01:10:35 +08:00
|
|
|
#include <net/strparser.h>
|
2018-09-21 12:16:13 +08:00
|
|
|
#include <crypto/aead.h>
|
2017-06-15 02:37:39 +08:00
|
|
|
#include <uapi/linux/tls.h>
|
|
|
|
|
2022-07-08 09:03:13 +08:00
|
|
|
struct tls_rec;
|
2017-06-15 02:37:39 +08:00
|
|
|
|
2022-09-20 21:01:47 +08:00
|
|
|
struct tls_cipher_size_desc {
|
|
|
|
unsigned int iv;
|
|
|
|
unsigned int key;
|
|
|
|
unsigned int salt;
|
|
|
|
unsigned int tag;
|
|
|
|
unsigned int rec_seq;
|
|
|
|
};
|
|
|
|
|
|
|
|
extern const struct tls_cipher_size_desc tls_cipher_size_desc[];
|
|
|
|
|
2017-06-15 02:37:39 +08:00
|
|
|
/* Maximum data size carried in a TLS record */
|
|
|
|
#define TLS_MAX_PAYLOAD_SIZE ((size_t)1 << 14)
|
|
|
|
|
|
|
|
#define TLS_HEADER_SIZE 5
|
|
|
|
#define TLS_NONCE_OFFSET TLS_HEADER_SIZE
|
|
|
|
|
|
|
|
#define TLS_CRYPTO_INFO_READY(info) ((info)->cipher_type)
|
|
|
|
|
2023-04-17 21:42:14 +08:00
|
|
|
#define TLS_RECORD_TYPE_ALERT 0x15
|
|
|
|
#define TLS_RECORD_TYPE_HANDSHAKE 0x16
|
2017-06-15 02:37:39 +08:00
|
|
|
#define TLS_RECORD_TYPE_DATA 0x17
|
|
|
|
|
|
|
|
#define TLS_AAD_SPACE_SIZE 13
|
2018-04-01 00:11:52 +08:00
|
|
|
|
2019-03-20 10:03:36 +08:00
|
|
|
#define MAX_IV_SIZE 16
|
2022-04-08 11:38:19 +08:00
|
|
|
#define TLS_TAG_SIZE 16
|
2019-06-11 12:40:00 +08:00
|
|
|
#define TLS_MAX_REC_SEQ_SIZE 8
|
2022-07-08 09:03:10 +08:00
|
|
|
#define TLS_MAX_AAD_SIZE TLS_AAD_SPACE_SIZE
|
2019-03-20 10:03:36 +08:00
|
|
|
|
2021-09-28 14:28:43 +08:00
|
|
|
/* For CCM mode, the full 16-bytes of IV is made of '4' fields of given sizes.
|
2019-03-20 10:03:36 +08:00
|
|
|
*
|
|
|
|
* IV[16] = b0[1] || implicit nonce[4] || explicit nonce[8] || length[3]
|
|
|
|
*
|
|
|
|
* The field 'length' is encoded in field 'b0' as '(length width - 1)'.
|
|
|
|
* Hence b0 contains (3 - 1) = 2.
|
|
|
|
*/
|
|
|
|
#define TLS_AES_CCM_IV_B0_BYTE 2
|
2021-09-28 14:28:43 +08:00
|
|
|
#define TLS_SM4_CCM_IV_B0_BYTE 2
|
2019-03-20 10:03:36 +08:00
|
|
|
|
2018-07-13 19:33:43 +08:00
|
|
|
enum {
|
|
|
|
TLS_BASE,
|
|
|
|
TLS_SW,
|
|
|
|
TLS_HW,
|
|
|
|
TLS_HW_RECORD,
|
|
|
|
TLS_NUM_CONFIG,
|
|
|
|
};
|
|
|
|
|
2018-09-21 12:16:13 +08:00
|
|
|
struct tx_work {
|
|
|
|
struct delayed_work work;
|
|
|
|
struct sock *sk;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct tls_sw_context_tx {
|
|
|
|
struct crypto_aead *aead_send;
|
|
|
|
struct crypto_wait async_wait;
|
|
|
|
struct tx_work tx_work;
|
|
|
|
struct tls_rec *open_rec;
|
net/tls: Fixed race condition in async encryption
On processors with multi-engine crypto accelerators, it is possible that
multiple records get encrypted in parallel and their encryption
completion is notified to different cpus in multicore processor. This
leads to the situation where tls_encrypt_done() starts executing in
parallel on different cores. In current implementation, encrypted
records are queued to tx_ready_list in tls_encrypt_done(). This requires
addition to linked list 'tx_ready_list' to be protected. As
tls_decrypt_done() could be executing in irq content, it is not possible
to protect linked list addition operation using a lock.
To fix the problem, we remove linked list addition operation from the
irq context. We do tx_ready_list addition/removal operation from
application context only and get rid of possible multiple access to
the linked list. Before starting encryption on the record, we add it to
the tail of tx_ready_list. To prevent tls_tx_records() from transmitting
it, we mark the record with a new flag 'tx_ready' in 'struct tls_rec'.
When record encryption gets completed, tls_encrypt_done() has to only
update the 'tx_ready' flag to true & linked list add operation is not
required.
The changed logic brings some other side benefits. Since the records
are always submitted in tls sequence number order for encryption, the
tx_ready_list always remains sorted and addition of new records to it
does not have to traverse the linked list.
Lastly, we renamed tx_ready_list in 'struct tls_sw_context_tx' to
'tx_list'. This is because now, the some of the records at the tail are
not ready to transmit.
Fixes: a42055e8d2c3 ("net/tls: Add support for async encryption")
Signed-off-by: Vakul Garg <vakul.garg@nxp.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-09-24 18:05:56 +08:00
|
|
|
struct list_head tx_list;
|
2018-09-21 12:16:13 +08:00
|
|
|
atomic_t encrypt_pending;
|
2020-05-23 04:10:31 +08:00
|
|
|
/* protect crypto_wait with encrypt_pending */
|
|
|
|
spinlock_t encrypt_compl_lock;
|
2018-09-21 12:16:13 +08:00
|
|
|
int async_notify;
|
2019-10-07 12:09:31 +08:00
|
|
|
u8 async_capable:1;
|
2018-09-21 12:16:13 +08:00
|
|
|
|
|
|
|
#define BIT_TX_SCHEDULED 0
|
2019-07-20 01:29:16 +08:00
|
|
|
#define BIT_TX_CLOSING 1
|
2018-09-21 12:16:13 +08:00
|
|
|
unsigned long tx_bitmask;
|
2017-06-15 02:37:39 +08:00
|
|
|
};
|
|
|
|
|
2022-07-23 07:50:33 +08:00
|
|
|
struct tls_strparser {
|
|
|
|
struct sock *sk;
|
|
|
|
|
|
|
|
u32 mark : 8;
|
|
|
|
u32 stopped : 1;
|
|
|
|
u32 copy_mode : 1;
|
|
|
|
u32 msg_ready : 1;
|
|
|
|
|
|
|
|
struct strp_msg stm;
|
|
|
|
|
|
|
|
struct sk_buff *anchor;
|
|
|
|
struct work_struct work;
|
|
|
|
};
|
|
|
|
|
2018-04-30 15:16:15 +08:00
|
|
|
struct tls_sw_context_rx {
|
|
|
|
struct crypto_aead *aead_recv;
|
|
|
|
struct crypto_wait async_wait;
|
2019-01-16 18:40:16 +08:00
|
|
|
struct sk_buff_head rx_list; /* list of decrypted 'data' records */
|
2018-04-30 15:16:15 +08:00
|
|
|
void (*saved_data_ready)(struct sock *sk);
|
2018-10-13 08:46:00 +08:00
|
|
|
|
2022-07-15 13:22:25 +08:00
|
|
|
u8 reader_present;
|
2019-10-07 12:09:31 +08:00
|
|
|
u8 async_capable:1;
|
2022-07-06 07:59:24 +08:00
|
|
|
u8 zc_capable:1;
|
2022-07-15 13:22:25 +08:00
|
|
|
u8 reader_contended:1;
|
2022-07-23 07:50:33 +08:00
|
|
|
|
|
|
|
struct tls_strparser strp;
|
|
|
|
|
2018-08-29 17:56:55 +08:00
|
|
|
atomic_t decrypt_pending;
|
2020-05-23 04:10:31 +08:00
|
|
|
/* protect crypto_wait with decrypt_pending*/
|
|
|
|
spinlock_t decrypt_compl_lock;
|
2022-07-15 13:22:33 +08:00
|
|
|
struct sk_buff_head async_hold;
|
2022-07-15 13:22:25 +08:00
|
|
|
struct wait_queue_head wq;
|
2018-08-29 17:56:55 +08:00
|
|
|
};
|
|
|
|
|
2018-04-30 15:16:16 +08:00
|
|
|
struct tls_record_info {
|
|
|
|
struct list_head list;
|
|
|
|
u32 end_seq;
|
|
|
|
int len;
|
|
|
|
int num_frags;
|
|
|
|
skb_frag_t frags[MAX_SKB_FRAGS];
|
|
|
|
};
|
|
|
|
|
2018-07-13 19:33:39 +08:00
|
|
|
struct tls_offload_context_tx {
|
2018-04-30 15:16:16 +08:00
|
|
|
struct crypto_aead *aead_send;
|
|
|
|
spinlock_t lock; /* protects records list */
|
|
|
|
struct list_head records_list;
|
|
|
|
struct tls_record_info *open_record;
|
|
|
|
struct tls_record_info *retransmit_hint;
|
|
|
|
u64 hint_record_sn;
|
|
|
|
u64 unacked_record_sn;
|
|
|
|
|
|
|
|
struct scatterlist sg_tx_data[MAX_SKB_FRAGS];
|
|
|
|
void (*sk_destruct)(struct sock *sk);
|
2022-07-27 17:43:42 +08:00
|
|
|
struct work_struct destruct_work;
|
|
|
|
struct tls_context *ctx;
|
2019-06-06 05:11:39 +08:00
|
|
|
u8 driver_state[] __aligned(8);
|
2018-04-30 15:16:16 +08:00
|
|
|
/* The TLS layer reserves room for driver specific state
|
|
|
|
* Currently the belief is that there is not enough
|
|
|
|
* driver specific state to justify another layer of indirection
|
|
|
|
*/
|
2019-06-06 05:11:38 +08:00
|
|
|
#define TLS_DRIVER_STATE_SIZE_TX 16
|
2018-04-30 15:16:16 +08:00
|
|
|
};
|
|
|
|
|
2018-07-13 19:33:39 +08:00
|
|
|
#define TLS_OFFLOAD_CONTEXT_SIZE_TX \
|
2019-06-06 05:11:39 +08:00
|
|
|
(sizeof(struct tls_offload_context_tx) + TLS_DRIVER_STATE_SIZE_TX)
|
2018-04-30 15:16:16 +08:00
|
|
|
|
2019-06-05 03:00:12 +08:00
|
|
|
enum tls_context_flags {
|
net/tls: Fix use-after-free after the TLS device goes down and up
When a netdev with active TLS offload goes down, tls_device_down is
called to stop the offload and tear down the TLS context. However, the
socket stays alive, and it still points to the TLS context, which is now
deallocated. If a netdev goes up, while the connection is still active,
and the data flow resumes after a number of TCP retransmissions, it will
lead to a use-after-free of the TLS context.
This commit addresses this bug by keeping the context alive until its
normal destruction, and implements the necessary fallbacks, so that the
connection can resume in software (non-offloaded) kTLS mode.
On the TX side tls_sw_fallback is used to encrypt all packets. The RX
side already has all the necessary fallbacks, because receiving
non-decrypted packets is supported. The thing needed on the RX side is
to block resync requests, which are normally produced after receiving
non-decrypted packets.
The necessary synchronization is implemented for a graceful teardown:
first the fallbacks are deployed, then the driver resources are released
(it used to be possible to have a tls_dev_resync after tls_dev_del).
A new flag called TLS_RX_DEV_DEGRADED is added to indicate the fallback
mode. It's used to skip the RX resync logic completely, as it becomes
useless, and some objects may be released (for example, resync_async,
which is allocated and freed by the driver).
Fixes: e8f69799810c ("net/tls: Add generic NIC offload infrastructure")
Signed-off-by: Maxim Mikityanskiy <maximmi@nvidia.com>
Reviewed-by: Tariq Toukan <tariqt@nvidia.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2021-06-01 20:08:00 +08:00
|
|
|
/* tls_device_down was called after the netdev went down, device state
|
|
|
|
* was released, and kTLS works in software, even though rx_conf is
|
|
|
|
* still TLS_HW (needed for transition).
|
|
|
|
*/
|
|
|
|
TLS_RX_DEV_DEGRADED = 0,
|
2019-06-11 12:40:09 +08:00
|
|
|
/* Unlike RX where resync is driven entirely by the core in TX only
|
|
|
|
* the driver knows when things went out of sync, so we need the flag
|
|
|
|
* to be atomic.
|
|
|
|
*/
|
|
|
|
TLS_TX_SYNC_SCHED = 1,
|
2020-11-26 06:18:10 +08:00
|
|
|
/* tls_dev_del was called for the RX side, device state was released,
|
|
|
|
* but tls_ctx->netdev might still be kept, because TX-side driver
|
|
|
|
* resources might not be released yet. Used to prevent the second
|
|
|
|
* tls_dev_del call in tls_device_down if it happens simultaneously.
|
|
|
|
*/
|
|
|
|
TLS_RX_DEV_CLOSED = 2,
|
2019-06-05 03:00:12 +08:00
|
|
|
};
|
|
|
|
|
2018-03-23 01:10:06 +08:00
|
|
|
struct cipher_context {
|
|
|
|
char *iv;
|
|
|
|
char *rec_seq;
|
|
|
|
};
|
|
|
|
|
2018-09-12 23:44:42 +08:00
|
|
|
union tls_crypto_context {
|
|
|
|
struct tls_crypto_info info;
|
2019-01-31 05:58:05 +08:00
|
|
|
union {
|
|
|
|
struct tls12_crypto_info_aes_gcm_128 aes_gcm_128;
|
|
|
|
struct tls12_crypto_info_aes_gcm_256 aes_gcm_256;
|
2020-11-24 23:24:47 +08:00
|
|
|
struct tls12_crypto_info_chacha20_poly1305 chacha20_poly1305;
|
2021-10-25 21:04:39 +08:00
|
|
|
struct tls12_crypto_info_sm4_gcm sm4_gcm;
|
|
|
|
struct tls12_crypto_info_sm4_ccm sm4_ccm;
|
2019-01-31 05:58:05 +08:00
|
|
|
};
|
2018-09-12 23:44:42 +08:00
|
|
|
};
|
|
|
|
|
2019-02-14 15:11:35 +08:00
|
|
|
struct tls_prot_info {
|
|
|
|
u16 version;
|
|
|
|
u16 cipher_type;
|
|
|
|
u16 prepend_size;
|
|
|
|
u16 tag_size;
|
|
|
|
u16 overhead_size;
|
|
|
|
u16 iv_size;
|
2019-03-20 10:03:36 +08:00
|
|
|
u16 salt_size;
|
2019-02-14 15:11:35 +08:00
|
|
|
u16 rec_seq_size;
|
|
|
|
u16 aad_size;
|
|
|
|
u16 tail_size;
|
|
|
|
};
|
|
|
|
|
2017-06-15 02:37:39 +08:00
|
|
|
struct tls_context {
|
2019-06-04 06:17:04 +08:00
|
|
|
/* read-only cache line */
|
2019-02-14 15:11:35 +08:00
|
|
|
struct tls_prot_info prot_info;
|
|
|
|
|
2019-06-04 06:17:04 +08:00
|
|
|
u8 tx_conf:3;
|
|
|
|
u8 rx_conf:3;
|
2022-05-18 17:27:31 +08:00
|
|
|
u8 zerocopy_sendfile:1;
|
2022-07-06 07:59:24 +08:00
|
|
|
u8 rx_no_pad:1;
|
2017-06-15 02:37:39 +08:00
|
|
|
|
2019-06-04 06:17:04 +08:00
|
|
|
int (*push_pending_record)(struct sock *sk, int flags);
|
|
|
|
void (*sk_write_space)(struct sock *sk);
|
2018-04-30 15:16:15 +08:00
|
|
|
|
|
|
|
void *priv_ctx_tx;
|
|
|
|
void *priv_ctx_rx;
|
2017-06-15 02:37:39 +08:00
|
|
|
|
net/tls: Use RCU API to access tls_ctx->netdev
Currently, tls_device_down synchronizes with tls_device_resync_rx using
RCU, however, the pointer to netdev is stored using WRITE_ONCE and
loaded using READ_ONCE.
Although such approach is technically correct (rcu_dereference is
essentially a READ_ONCE, and rcu_assign_pointer uses WRITE_ONCE to store
NULL), using special RCU helpers for pointers is more valid, as it
includes additional checks and might change the implementation
transparently to the callers.
Mark the netdev pointer as __rcu and use the correct RCU helpers to
access it. For non-concurrent access pass the right conditions that
guarantee safe access (locks taken, refcount value). Also use the
correct helper in mlx5e, where even READ_ONCE was missing.
The transition to RCU exposes existing issues, fixed by this commit:
1. bond_tls_device_xmit could read netdev twice, and it could become
NULL the second time, after the NULL check passed.
2. Drivers shouldn't stop processing the last packet if tls_device_down
just set netdev to NULL, before tls_dev_del was called. This prevents a
possible packet drop when transitioning to the fallback software mode.
Fixes: 89df6a810470 ("net/bonding: Implement TLS TX device offload")
Fixes: c55dcdd435aa ("net/tls: Fix use-after-free after the TLS device goes down and up")
Signed-off-by: Maxim Mikityanskiy <maximmi@nvidia.com>
Link: https://lore.kernel.org/r/20220810081602.1435800-1-maximmi@nvidia.com
Signed-off-by: Jakub Kicinski <kuba@kernel.org>
2022-08-10 16:16:02 +08:00
|
|
|
struct net_device __rcu *netdev;
|
2017-11-13 16:22:45 +08:00
|
|
|
|
2019-06-04 06:17:04 +08:00
|
|
|
/* rw cache line */
|
2018-03-23 01:10:06 +08:00
|
|
|
struct cipher_context tx;
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 01:10:35 +08:00
|
|
|
struct cipher_context rx;
|
2017-06-15 02:37:39 +08:00
|
|
|
|
|
|
|
struct scatterlist *partially_sent_record;
|
|
|
|
u16 partially_sent_offset;
|
2018-09-21 12:16:13 +08:00
|
|
|
|
2018-05-02 04:05:39 +08:00
|
|
|
bool in_tcp_sendpages;
|
2018-10-13 08:45:59 +08:00
|
|
|
bool pending_open_record_frags;
|
2019-11-06 06:24:35 +08:00
|
|
|
|
|
|
|
struct mutex tx_lock; /* protects partially_sent_* fields and
|
|
|
|
* per-type TX fields
|
|
|
|
*/
|
2019-06-04 06:17:04 +08:00
|
|
|
unsigned long flags;
|
2017-06-15 02:37:39 +08:00
|
|
|
|
2019-06-04 06:17:04 +08:00
|
|
|
/* cache cold stuff */
|
2019-07-20 01:29:18 +08:00
|
|
|
struct proto *sk_proto;
|
net/tls: Fix use-after-free after the TLS device goes down and up
When a netdev with active TLS offload goes down, tls_device_down is
called to stop the offload and tear down the TLS context. However, the
socket stays alive, and it still points to the TLS context, which is now
deallocated. If a netdev goes up, while the connection is still active,
and the data flow resumes after a number of TCP retransmissions, it will
lead to a use-after-free of the TLS context.
This commit addresses this bug by keeping the context alive until its
normal destruction, and implements the necessary fallbacks, so that the
connection can resume in software (non-offloaded) kTLS mode.
On the TX side tls_sw_fallback is used to encrypt all packets. The RX
side already has all the necessary fallbacks, because receiving
non-decrypted packets is supported. The thing needed on the RX side is
to block resync requests, which are normally produced after receiving
non-decrypted packets.
The necessary synchronization is implemented for a graceful teardown:
first the fallbacks are deployed, then the driver resources are released
(it used to be possible to have a tls_dev_resync after tls_dev_del).
A new flag called TLS_RX_DEV_DEGRADED is added to indicate the fallback
mode. It's used to skip the RX resync logic completely, as it becomes
useless, and some objects may be released (for example, resync_async,
which is allocated and freed by the driver).
Fixes: e8f69799810c ("net/tls: Add generic NIC offload infrastructure")
Signed-off-by: Maxim Mikityanskiy <maximmi@nvidia.com>
Reviewed-by: Tariq Toukan <tariqt@nvidia.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2021-06-01 20:08:00 +08:00
|
|
|
struct sock *sk;
|
2019-07-20 01:29:18 +08:00
|
|
|
|
2018-07-13 19:33:43 +08:00
|
|
|
void (*sk_destruct)(struct sock *sk);
|
2019-06-04 06:17:04 +08:00
|
|
|
|
|
|
|
union tls_crypto_context crypto_send;
|
|
|
|
union tls_crypto_context crypto_recv;
|
|
|
|
|
|
|
|
struct list_head list;
|
|
|
|
refcount_t refcount;
|
2019-08-30 18:25:47 +08:00
|
|
|
struct rcu_head rcu;
|
2017-06-15 02:37:39 +08:00
|
|
|
};
|
|
|
|
|
2019-04-26 03:32:03 +08:00
|
|
|
enum tls_offload_ctx_dir {
|
|
|
|
TLS_OFFLOAD_CTX_DIR_RX,
|
|
|
|
TLS_OFFLOAD_CTX_DIR_TX,
|
|
|
|
};
|
|
|
|
|
|
|
|
struct tlsdev_ops {
|
|
|
|
int (*tls_dev_add)(struct net_device *netdev, struct sock *sk,
|
|
|
|
enum tls_offload_ctx_dir direction,
|
|
|
|
struct tls_crypto_info *crypto_info,
|
|
|
|
u32 start_offload_tcp_sn);
|
|
|
|
void (*tls_dev_del)(struct net_device *netdev,
|
|
|
|
struct tls_context *ctx,
|
|
|
|
enum tls_offload_ctx_dir direction);
|
2019-07-09 10:53:13 +08:00
|
|
|
int (*tls_dev_resync)(struct net_device *netdev,
|
|
|
|
struct sock *sk, u32 seq, u8 *rcd_sn,
|
|
|
|
enum tls_offload_ctx_dir direction);
|
2019-04-26 03:32:03 +08:00
|
|
|
};
|
|
|
|
|
2019-06-11 12:40:02 +08:00
|
|
|
enum tls_offload_sync_type {
|
|
|
|
TLS_OFFLOAD_SYNC_TYPE_DRIVER_REQ = 0,
|
|
|
|
TLS_OFFLOAD_SYNC_TYPE_CORE_NEXT_HINT = 1,
|
net/tls: Add asynchronous resync
This patch adds support for asynchronous resynchronization in tls_device.
Async resync follows two distinct stages:
1. The NIC driver indicates that it would like to resync on some TLS
record within the received packet (P), but the driver does not
know (yet) which of the TLS records within the packet.
At this stage, the NIC driver will query the device to find the exact
TCP sequence for resync (tcpsn), however, the driver does not wait
for the device to provide the response.
2. Eventually, the device responds, and the driver provides the tcpsn
within the resync packet to KTLS. Now, KTLS can check the tcpsn against
any processed TLS records within packet P, and also against any record
that is processed in the future within packet P.
The asynchronous resync path simplifies the device driver, as it can
save bits on the packet completion (32-bit TCP sequence), and pass this
information on an asynchronous command instead.
Signed-off-by: Boris Pismenny <borisp@mellanox.com>
Signed-off-by: Tariq Toukan <tariqt@mellanox.com>
Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
2020-06-09 00:11:38 +08:00
|
|
|
TLS_OFFLOAD_SYNC_TYPE_DRIVER_REQ_ASYNC = 2,
|
2019-06-11 12:40:02 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
#define TLS_DEVICE_RESYNC_NH_START_IVAL 2
|
|
|
|
#define TLS_DEVICE_RESYNC_NH_MAX_IVAL 128
|
|
|
|
|
net/tls: Add asynchronous resync
This patch adds support for asynchronous resynchronization in tls_device.
Async resync follows two distinct stages:
1. The NIC driver indicates that it would like to resync on some TLS
record within the received packet (P), but the driver does not
know (yet) which of the TLS records within the packet.
At this stage, the NIC driver will query the device to find the exact
TCP sequence for resync (tcpsn), however, the driver does not wait
for the device to provide the response.
2. Eventually, the device responds, and the driver provides the tcpsn
within the resync packet to KTLS. Now, KTLS can check the tcpsn against
any processed TLS records within packet P, and also against any record
that is processed in the future within packet P.
The asynchronous resync path simplifies the device driver, as it can
save bits on the packet completion (32-bit TCP sequence), and pass this
information on an asynchronous command instead.
Signed-off-by: Boris Pismenny <borisp@mellanox.com>
Signed-off-by: Tariq Toukan <tariqt@mellanox.com>
Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
2020-06-09 00:11:38 +08:00
|
|
|
#define TLS_DEVICE_RESYNC_ASYNC_LOGMAX 13
|
|
|
|
struct tls_offload_resync_async {
|
|
|
|
atomic64_t req;
|
2020-11-15 21:14:48 +08:00
|
|
|
u16 loglen;
|
|
|
|
u16 rcd_delta;
|
net/tls: Add asynchronous resync
This patch adds support for asynchronous resynchronization in tls_device.
Async resync follows two distinct stages:
1. The NIC driver indicates that it would like to resync on some TLS
record within the received packet (P), but the driver does not
know (yet) which of the TLS records within the packet.
At this stage, the NIC driver will query the device to find the exact
TCP sequence for resync (tcpsn), however, the driver does not wait
for the device to provide the response.
2. Eventually, the device responds, and the driver provides the tcpsn
within the resync packet to KTLS. Now, KTLS can check the tcpsn against
any processed TLS records within packet P, and also against any record
that is processed in the future within packet P.
The asynchronous resync path simplifies the device driver, as it can
save bits on the packet completion (32-bit TCP sequence), and pass this
information on an asynchronous command instead.
Signed-off-by: Boris Pismenny <borisp@mellanox.com>
Signed-off-by: Tariq Toukan <tariqt@mellanox.com>
Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
2020-06-09 00:11:38 +08:00
|
|
|
u32 log[TLS_DEVICE_RESYNC_ASYNC_LOGMAX];
|
|
|
|
};
|
|
|
|
|
2018-07-13 19:33:43 +08:00
|
|
|
struct tls_offload_context_rx {
|
|
|
|
/* sw must be the first member of tls_offload_context_rx */
|
|
|
|
struct tls_sw_context_rx sw;
|
2019-06-11 12:40:02 +08:00
|
|
|
enum tls_offload_sync_type resync_type;
|
|
|
|
/* this member is set regardless of resync_type, to avoid branches */
|
|
|
|
u8 resync_nh_reset:1;
|
|
|
|
/* CORE_NEXT_HINT-only member, but use the hole here */
|
|
|
|
u8 resync_nh_do_now:1;
|
|
|
|
union {
|
|
|
|
/* TLS_OFFLOAD_SYNC_TYPE_DRIVER_REQ */
|
|
|
|
struct {
|
|
|
|
atomic64_t resync_req;
|
|
|
|
};
|
|
|
|
/* TLS_OFFLOAD_SYNC_TYPE_CORE_NEXT_HINT */
|
|
|
|
struct {
|
|
|
|
u32 decrypted_failed;
|
|
|
|
u32 decrypted_tgt;
|
|
|
|
} resync_nh;
|
net/tls: Add asynchronous resync
This patch adds support for asynchronous resynchronization in tls_device.
Async resync follows two distinct stages:
1. The NIC driver indicates that it would like to resync on some TLS
record within the received packet (P), but the driver does not
know (yet) which of the TLS records within the packet.
At this stage, the NIC driver will query the device to find the exact
TCP sequence for resync (tcpsn), however, the driver does not wait
for the device to provide the response.
2. Eventually, the device responds, and the driver provides the tcpsn
within the resync packet to KTLS. Now, KTLS can check the tcpsn against
any processed TLS records within packet P, and also against any record
that is processed in the future within packet P.
The asynchronous resync path simplifies the device driver, as it can
save bits on the packet completion (32-bit TCP sequence), and pass this
information on an asynchronous command instead.
Signed-off-by: Boris Pismenny <borisp@mellanox.com>
Signed-off-by: Tariq Toukan <tariqt@mellanox.com>
Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
2020-06-09 00:11:38 +08:00
|
|
|
/* TLS_OFFLOAD_SYNC_TYPE_DRIVER_REQ_ASYNC */
|
|
|
|
struct {
|
|
|
|
struct tls_offload_resync_async *resync_async;
|
|
|
|
};
|
2019-06-11 12:40:02 +08:00
|
|
|
};
|
2019-06-06 05:11:39 +08:00
|
|
|
u8 driver_state[] __aligned(8);
|
2018-07-13 19:33:43 +08:00
|
|
|
/* The TLS layer reserves room for driver specific state
|
|
|
|
* Currently the belief is that there is not enough
|
|
|
|
* driver specific state to justify another layer of indirection
|
|
|
|
*/
|
2019-06-06 05:11:38 +08:00
|
|
|
#define TLS_DRIVER_STATE_SIZE_RX 8
|
2018-07-13 19:33:43 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
#define TLS_OFFLOAD_CONTEXT_SIZE_RX \
|
2019-06-06 05:11:39 +08:00
|
|
|
(sizeof(struct tls_offload_context_rx) + TLS_DRIVER_STATE_SIZE_RX)
|
2018-07-13 19:33:43 +08:00
|
|
|
|
2018-07-13 19:33:39 +08:00
|
|
|
struct tls_record_info *tls_get_record(struct tls_offload_context_tx *context,
|
2018-04-30 15:16:16 +08:00
|
|
|
u32 seq, u64 *p_record_sn);
|
|
|
|
|
|
|
|
static inline bool tls_record_is_start_marker(struct tls_record_info *rec)
|
|
|
|
{
|
|
|
|
return rec->len == 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 tls_record_start_seq(struct tls_record_info *rec)
|
|
|
|
{
|
|
|
|
return rec->end_seq - rec->len;
|
|
|
|
}
|
2017-06-15 02:37:39 +08:00
|
|
|
|
2018-07-13 19:33:43 +08:00
|
|
|
struct sk_buff *
|
|
|
|
tls_validate_xmit_skb(struct sock *sk, struct net_device *dev,
|
|
|
|
struct sk_buff *skb);
|
net/tls: Fix use-after-free after the TLS device goes down and up
When a netdev with active TLS offload goes down, tls_device_down is
called to stop the offload and tear down the TLS context. However, the
socket stays alive, and it still points to the TLS context, which is now
deallocated. If a netdev goes up, while the connection is still active,
and the data flow resumes after a number of TCP retransmissions, it will
lead to a use-after-free of the TLS context.
This commit addresses this bug by keeping the context alive until its
normal destruction, and implements the necessary fallbacks, so that the
connection can resume in software (non-offloaded) kTLS mode.
On the TX side tls_sw_fallback is used to encrypt all packets. The RX
side already has all the necessary fallbacks, because receiving
non-decrypted packets is supported. The thing needed on the RX side is
to block resync requests, which are normally produced after receiving
non-decrypted packets.
The necessary synchronization is implemented for a graceful teardown:
first the fallbacks are deployed, then the driver resources are released
(it used to be possible to have a tls_dev_resync after tls_dev_del).
A new flag called TLS_RX_DEV_DEGRADED is added to indicate the fallback
mode. It's used to skip the RX resync logic completely, as it becomes
useless, and some objects may be released (for example, resync_async,
which is allocated and freed by the driver).
Fixes: e8f69799810c ("net/tls: Add generic NIC offload infrastructure")
Signed-off-by: Maxim Mikityanskiy <maximmi@nvidia.com>
Reviewed-by: Tariq Toukan <tariqt@nvidia.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2021-06-01 20:08:00 +08:00
|
|
|
struct sk_buff *
|
|
|
|
tls_validate_xmit_skb_sw(struct sock *sk, struct net_device *dev,
|
|
|
|
struct sk_buff *skb);
|
2018-07-13 19:33:43 +08:00
|
|
|
|
2018-04-30 15:16:16 +08:00
|
|
|
static inline bool tls_is_sk_tx_device_offloaded(struct sock *sk)
|
|
|
|
{
|
2018-07-13 19:33:43 +08:00
|
|
|
#ifdef CONFIG_SOCK_VALIDATE_XMIT
|
2019-04-09 08:59:50 +08:00
|
|
|
return sk_fullsock(sk) &&
|
2018-07-13 19:33:43 +08:00
|
|
|
(smp_load_acquire(&sk->sk_validate_xmit_skb) ==
|
|
|
|
&tls_validate_xmit_skb);
|
|
|
|
#else
|
|
|
|
return false;
|
|
|
|
#endif
|
2018-04-30 15:16:16 +08:00
|
|
|
}
|
|
|
|
|
2019-02-14 15:11:35 +08:00
|
|
|
static inline struct tls_context *tls_get_ctx(const struct sock *sk)
|
|
|
|
{
|
|
|
|
struct inet_connection_sock *icsk = inet_csk(sk);
|
|
|
|
|
2019-08-30 18:25:47 +08:00
|
|
|
/* Use RCU on icsk_ulp_data only for sock diag code,
|
|
|
|
* TLS data path doesn't need rcu_dereference().
|
|
|
|
*/
|
|
|
|
return (__force void *)icsk->icsk_ulp_data;
|
2019-02-14 15:11:35 +08:00
|
|
|
}
|
|
|
|
|
2018-04-30 15:16:15 +08:00
|
|
|
static inline struct tls_sw_context_rx *tls_sw_ctx_rx(
|
|
|
|
const struct tls_context *tls_ctx)
|
|
|
|
{
|
|
|
|
return (struct tls_sw_context_rx *)tls_ctx->priv_ctx_rx;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct tls_sw_context_tx *tls_sw_ctx_tx(
|
2017-06-15 02:37:39 +08:00
|
|
|
const struct tls_context *tls_ctx)
|
|
|
|
{
|
2018-04-30 15:16:15 +08:00
|
|
|
return (struct tls_sw_context_tx *)tls_ctx->priv_ctx_tx;
|
2017-06-15 02:37:39 +08:00
|
|
|
}
|
|
|
|
|
2018-07-13 19:33:39 +08:00
|
|
|
static inline struct tls_offload_context_tx *
|
|
|
|
tls_offload_ctx_tx(const struct tls_context *tls_ctx)
|
2017-06-15 02:37:39 +08:00
|
|
|
{
|
2018-07-13 19:33:39 +08:00
|
|
|
return (struct tls_offload_context_tx *)tls_ctx->priv_ctx_tx;
|
2017-06-15 02:37:39 +08:00
|
|
|
}
|
|
|
|
|
bpf: sk_msg, sock{map|hash} redirect through ULP
A sockmap program that redirects through a kTLS ULP enabled socket
will not work correctly because the ULP layer is skipped. This
fixes the behavior to call through the ULP layer on redirect to
ensure any operations required on the data stream at the ULP layer
continue to be applied.
To do this we add an internal flag MSG_SENDPAGE_NOPOLICY to avoid
calling the BPF layer on a redirected message. This is
required to avoid calling the BPF layer multiple times (possibly
recursively) which is not the current/expected behavior without
ULPs. In the future we may add a redirect flag if users _do_
want the policy applied again but this would need to work for both
ULP and non-ULP sockets and be opt-in to avoid breaking existing
programs.
Also to avoid polluting the flag space with an internal flag we
reuse the flag space overlapping MSG_SENDPAGE_NOPOLICY with
MSG_WAITFORONE. Here WAITFORONE is specific to recv path and
SENDPAGE_NOPOLICY is only used for sendpage hooks. The last thing
to verify is user space API is masked correctly to ensure the flag
can not be set by user. (Note this needs to be true regardless
because we have internal flags already in-use that user space
should not be able to set). But for completeness we have two UAPI
paths into sendpage, sendfile and splice.
In the sendfile case the function do_sendfile() zero's flags,
./fs/read_write.c:
static ssize_t do_sendfile(int out_fd, int in_fd, loff_t *ppos,
size_t count, loff_t max)
{
...
fl = 0;
#if 0
/*
* We need to debate whether we can enable this or not. The
* man page documents EAGAIN return for the output at least,
* and the application is arguably buggy if it doesn't expect
* EAGAIN on a non-blocking file descriptor.
*/
if (in.file->f_flags & O_NONBLOCK)
fl = SPLICE_F_NONBLOCK;
#endif
file_start_write(out.file);
retval = do_splice_direct(in.file, &pos, out.file, &out_pos, count, fl);
}
In the splice case the pipe_to_sendpage "actor" is used which
masks flags with SPLICE_F_MORE.
./fs/splice.c:
static int pipe_to_sendpage(struct pipe_inode_info *pipe,
struct pipe_buffer *buf, struct splice_desc *sd)
{
...
more = (sd->flags & SPLICE_F_MORE) ? MSG_MORE : 0;
...
}
Confirming what we expect that internal flags are in fact internal
to socket side.
Fixes: d3b18ad31f93 ("tls: add bpf support to sk_msg handling")
Signed-off-by: John Fastabend <john.fastabend@gmail.com>
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
2018-12-21 03:35:35 +08:00
|
|
|
static inline bool tls_sw_has_ctx_tx(const struct sock *sk)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
|
|
|
|
|
|
|
if (!ctx)
|
|
|
|
return false;
|
|
|
|
return !!tls_sw_ctx_tx(ctx);
|
|
|
|
}
|
|
|
|
|
2020-05-30 07:06:59 +08:00
|
|
|
static inline bool tls_sw_has_ctx_rx(const struct sock *sk)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
|
|
|
|
|
|
|
if (!ctx)
|
|
|
|
return false;
|
|
|
|
return !!tls_sw_ctx_rx(ctx);
|
|
|
|
}
|
|
|
|
|
2018-07-13 19:33:43 +08:00
|
|
|
static inline struct tls_offload_context_rx *
|
|
|
|
tls_offload_ctx_rx(const struct tls_context *tls_ctx)
|
|
|
|
{
|
|
|
|
return (struct tls_offload_context_rx *)tls_ctx->priv_ctx_rx;
|
|
|
|
}
|
|
|
|
|
2019-06-06 05:11:39 +08:00
|
|
|
static inline void *__tls_driver_ctx(struct tls_context *tls_ctx,
|
|
|
|
enum tls_offload_ctx_dir direction)
|
|
|
|
{
|
|
|
|
if (direction == TLS_OFFLOAD_CTX_DIR_TX)
|
|
|
|
return tls_offload_ctx_tx(tls_ctx)->driver_state;
|
|
|
|
else
|
|
|
|
return tls_offload_ctx_rx(tls_ctx)->driver_state;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void *
|
|
|
|
tls_driver_ctx(const struct sock *sk, enum tls_offload_ctx_dir direction)
|
|
|
|
{
|
|
|
|
return __tls_driver_ctx(tls_get_ctx(sk), direction);
|
|
|
|
}
|
|
|
|
|
net/tls: Add asynchronous resync
This patch adds support for asynchronous resynchronization in tls_device.
Async resync follows two distinct stages:
1. The NIC driver indicates that it would like to resync on some TLS
record within the received packet (P), but the driver does not
know (yet) which of the TLS records within the packet.
At this stage, the NIC driver will query the device to find the exact
TCP sequence for resync (tcpsn), however, the driver does not wait
for the device to provide the response.
2. Eventually, the device responds, and the driver provides the tcpsn
within the resync packet to KTLS. Now, KTLS can check the tcpsn against
any processed TLS records within packet P, and also against any record
that is processed in the future within packet P.
The asynchronous resync path simplifies the device driver, as it can
save bits on the packet completion (32-bit TCP sequence), and pass this
information on an asynchronous command instead.
Signed-off-by: Boris Pismenny <borisp@mellanox.com>
Signed-off-by: Tariq Toukan <tariqt@mellanox.com>
Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
2020-06-09 00:11:38 +08:00
|
|
|
#define RESYNC_REQ BIT(0)
|
|
|
|
#define RESYNC_REQ_ASYNC BIT(1)
|
2018-07-13 19:33:43 +08:00
|
|
|
/* The TLS context is valid until sk_destruct is called */
|
|
|
|
static inline void tls_offload_rx_resync_request(struct sock *sk, __be32 seq)
|
|
|
|
{
|
|
|
|
struct tls_context *tls_ctx = tls_get_ctx(sk);
|
|
|
|
struct tls_offload_context_rx *rx_ctx = tls_offload_ctx_rx(tls_ctx);
|
|
|
|
|
net/tls: Add asynchronous resync
This patch adds support for asynchronous resynchronization in tls_device.
Async resync follows two distinct stages:
1. The NIC driver indicates that it would like to resync on some TLS
record within the received packet (P), but the driver does not
know (yet) which of the TLS records within the packet.
At this stage, the NIC driver will query the device to find the exact
TCP sequence for resync (tcpsn), however, the driver does not wait
for the device to provide the response.
2. Eventually, the device responds, and the driver provides the tcpsn
within the resync packet to KTLS. Now, KTLS can check the tcpsn against
any processed TLS records within packet P, and also against any record
that is processed in the future within packet P.
The asynchronous resync path simplifies the device driver, as it can
save bits on the packet completion (32-bit TCP sequence), and pass this
information on an asynchronous command instead.
Signed-off-by: Boris Pismenny <borisp@mellanox.com>
Signed-off-by: Tariq Toukan <tariqt@mellanox.com>
Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
2020-06-09 00:11:38 +08:00
|
|
|
atomic64_set(&rx_ctx->resync_req, ((u64)ntohl(seq) << 32) | RESYNC_REQ);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Log all TLS record header TCP sequences in [seq, seq+len] */
|
|
|
|
static inline void
|
|
|
|
tls_offload_rx_resync_async_request_start(struct sock *sk, __be32 seq, u16 len)
|
|
|
|
{
|
|
|
|
struct tls_context *tls_ctx = tls_get_ctx(sk);
|
|
|
|
struct tls_offload_context_rx *rx_ctx = tls_offload_ctx_rx(tls_ctx);
|
|
|
|
|
|
|
|
atomic64_set(&rx_ctx->resync_async->req, ((u64)ntohl(seq) << 32) |
|
2020-06-30 22:27:46 +08:00
|
|
|
((u64)len << 16) | RESYNC_REQ | RESYNC_REQ_ASYNC);
|
net/tls: Add asynchronous resync
This patch adds support for asynchronous resynchronization in tls_device.
Async resync follows two distinct stages:
1. The NIC driver indicates that it would like to resync on some TLS
record within the received packet (P), but the driver does not
know (yet) which of the TLS records within the packet.
At this stage, the NIC driver will query the device to find the exact
TCP sequence for resync (tcpsn), however, the driver does not wait
for the device to provide the response.
2. Eventually, the device responds, and the driver provides the tcpsn
within the resync packet to KTLS. Now, KTLS can check the tcpsn against
any processed TLS records within packet P, and also against any record
that is processed in the future within packet P.
The asynchronous resync path simplifies the device driver, as it can
save bits on the packet completion (32-bit TCP sequence), and pass this
information on an asynchronous command instead.
Signed-off-by: Boris Pismenny <borisp@mellanox.com>
Signed-off-by: Tariq Toukan <tariqt@mellanox.com>
Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
2020-06-09 00:11:38 +08:00
|
|
|
rx_ctx->resync_async->loglen = 0;
|
2020-11-15 21:14:48 +08:00
|
|
|
rx_ctx->resync_async->rcd_delta = 0;
|
net/tls: Add asynchronous resync
This patch adds support for asynchronous resynchronization in tls_device.
Async resync follows two distinct stages:
1. The NIC driver indicates that it would like to resync on some TLS
record within the received packet (P), but the driver does not
know (yet) which of the TLS records within the packet.
At this stage, the NIC driver will query the device to find the exact
TCP sequence for resync (tcpsn), however, the driver does not wait
for the device to provide the response.
2. Eventually, the device responds, and the driver provides the tcpsn
within the resync packet to KTLS. Now, KTLS can check the tcpsn against
any processed TLS records within packet P, and also against any record
that is processed in the future within packet P.
The asynchronous resync path simplifies the device driver, as it can
save bits on the packet completion (32-bit TCP sequence), and pass this
information on an asynchronous command instead.
Signed-off-by: Boris Pismenny <borisp@mellanox.com>
Signed-off-by: Tariq Toukan <tariqt@mellanox.com>
Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
2020-06-09 00:11:38 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
tls_offload_rx_resync_async_request_end(struct sock *sk, __be32 seq)
|
|
|
|
{
|
|
|
|
struct tls_context *tls_ctx = tls_get_ctx(sk);
|
|
|
|
struct tls_offload_context_rx *rx_ctx = tls_offload_ctx_rx(tls_ctx);
|
|
|
|
|
|
|
|
atomic64_set(&rx_ctx->resync_async->req,
|
|
|
|
((u64)ntohl(seq) << 32) | RESYNC_REQ);
|
2018-07-13 19:33:43 +08:00
|
|
|
}
|
|
|
|
|
2019-06-11 12:40:02 +08:00
|
|
|
static inline void
|
|
|
|
tls_offload_rx_resync_set_type(struct sock *sk, enum tls_offload_sync_type type)
|
|
|
|
{
|
|
|
|
struct tls_context *tls_ctx = tls_get_ctx(sk);
|
|
|
|
|
|
|
|
tls_offload_ctx_rx(tls_ctx)->resync_type = type;
|
|
|
|
}
|
2018-07-13 19:33:43 +08:00
|
|
|
|
2019-06-11 12:40:09 +08:00
|
|
|
/* Driver's seq tracking has to be disabled until resync succeeded */
|
|
|
|
static inline bool tls_offload_tx_resync_pending(struct sock *sk)
|
|
|
|
{
|
|
|
|
struct tls_context *tls_ctx = tls_get_ctx(sk);
|
|
|
|
bool ret;
|
|
|
|
|
|
|
|
ret = test_bit(TLS_TX_SYNC_SCHED, &tls_ctx->flags);
|
|
|
|
smp_mb__after_atomic();
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2019-06-06 05:11:40 +08:00
|
|
|
struct sk_buff *tls_encrypt_skb(struct sk_buff *skb);
|
2017-06-15 02:37:39 +08:00
|
|
|
|
2019-09-03 12:31:05 +08:00
|
|
|
#ifdef CONFIG_TLS_DEVICE
|
2019-12-18 06:12:01 +08:00
|
|
|
void tls_device_sk_destruct(struct sock *sk);
|
2019-10-05 07:19:22 +08:00
|
|
|
void tls_offload_tx_resync_request(struct sock *sk, u32 got_seq, u32 exp_seq);
|
2019-12-18 06:12:01 +08:00
|
|
|
|
|
|
|
static inline bool tls_is_sk_rx_device_offloaded(struct sock *sk)
|
|
|
|
{
|
|
|
|
if (!sk_fullsock(sk) ||
|
|
|
|
smp_load_acquire(&sk->sk_destruct) != tls_device_sk_destruct)
|
|
|
|
return false;
|
|
|
|
return tls_get_ctx(sk)->rx_conf == TLS_HW;
|
|
|
|
}
|
2019-09-03 12:31:05 +08:00
|
|
|
#endif
|
2017-06-15 02:37:39 +08:00
|
|
|
#endif /* _TLS_OFFLOAD_H */
|