mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-17 01:04:19 +08:00
ebeeb1ad9b
An rds_connection can get added during netns deletion between lines 528
and 529 of
506 static void rds_tcp_kill_sock(struct net *net)
:
/* code to pull out all the rds_connections that should be destroyed */
:
528 spin_unlock_irq(&rds_tcp_conn_lock);
529 list_for_each_entry_safe(tc, _tc, &tmp_list, t_tcp_node)
530 rds_conn_destroy(tc->t_cpath->cp_conn);
Such an rds_connection would miss out the rds_conn_destroy()
loop (that cancels all pending work) and (if it was scheduled
after netns deletion) could trigger the use-after-free.
A similar race-window exists for the module unload path
in rds_tcp_exit -> rds_tcp_destroy_conns
Concurrency with netns deletion (rds_tcp_kill_sock()) must be handled
by checking check_net() before enqueuing new work or adding new
connections.
Concurrency with module-unload is handled by maintaining a module
specific flag that is set at the start of the module exit function,
and must be checked before enqueuing new work or adding new connections.
This commit refactors existing RDS_DESTROY_PENDING checks added by
commit 3db6e0d172
("rds: use RCU to synchronize work-enqueue with
connection teardown") and consolidates all the concurrency checks
listed above into the function rds_destroy_pending().
Signed-off-by: Sowmini Varadhan <sowmini.varadhan@oracle.com>
Acked-by: Santosh Shilimkar <santosh.shilimkar@oracle.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
190 lines
5.3 KiB
C
190 lines
5.3 KiB
C
/*
|
|
* Copyright (c) 2006 Oracle. All rights reserved.
|
|
*
|
|
* This software is available to you under a choice of one of two
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
* General Public License (GPL) Version 2, available from the file
|
|
* COPYING in the main directory of this source tree, or the
|
|
* OpenIB.org BSD license below:
|
|
*
|
|
* Redistribution and use in source and binary forms, with or
|
|
* without modification, are permitted provided that the following
|
|
* conditions are met:
|
|
*
|
|
* - Redistributions of source code must retain the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer.
|
|
*
|
|
* - Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*
|
|
*/
|
|
#include <linux/kernel.h>
|
|
#include <linux/in.h>
|
|
#include <net/tcp.h>
|
|
|
|
#include "rds.h"
|
|
#include "tcp.h"
|
|
|
|
void rds_tcp_state_change(struct sock *sk)
|
|
{
|
|
void (*state_change)(struct sock *sk);
|
|
struct rds_conn_path *cp;
|
|
struct rds_tcp_connection *tc;
|
|
|
|
read_lock_bh(&sk->sk_callback_lock);
|
|
cp = sk->sk_user_data;
|
|
if (!cp) {
|
|
state_change = sk->sk_state_change;
|
|
goto out;
|
|
}
|
|
tc = cp->cp_transport_data;
|
|
state_change = tc->t_orig_state_change;
|
|
|
|
rdsdebug("sock %p state_change to %d\n", tc->t_sock, sk->sk_state);
|
|
|
|
switch (sk->sk_state) {
|
|
/* ignore connecting sockets as they make progress */
|
|
case TCP_SYN_SENT:
|
|
case TCP_SYN_RECV:
|
|
break;
|
|
case TCP_ESTABLISHED:
|
|
/* Force the peer to reconnect so that we have the
|
|
* TCP ports going from <smaller-ip>.<transient> to
|
|
* <larger-ip>.<RDS_TCP_PORT>. We avoid marking the
|
|
* RDS connection as RDS_CONN_UP until the reconnect,
|
|
* to avoid RDS datagram loss.
|
|
*/
|
|
if (!IS_CANONICAL(cp->cp_conn->c_laddr, cp->cp_conn->c_faddr) &&
|
|
rds_conn_path_transition(cp, RDS_CONN_CONNECTING,
|
|
RDS_CONN_ERROR)) {
|
|
rds_conn_path_drop(cp, false);
|
|
} else {
|
|
rds_connect_path_complete(cp, RDS_CONN_CONNECTING);
|
|
}
|
|
break;
|
|
case TCP_CLOSE_WAIT:
|
|
case TCP_CLOSE:
|
|
rds_conn_path_drop(cp, false);
|
|
default:
|
|
break;
|
|
}
|
|
out:
|
|
read_unlock_bh(&sk->sk_callback_lock);
|
|
state_change(sk);
|
|
}
|
|
|
|
int rds_tcp_conn_path_connect(struct rds_conn_path *cp)
|
|
{
|
|
struct socket *sock = NULL;
|
|
struct sockaddr_in src, dest;
|
|
int ret;
|
|
struct rds_connection *conn = cp->cp_conn;
|
|
struct rds_tcp_connection *tc = cp->cp_transport_data;
|
|
|
|
/* for multipath rds,we only trigger the connection after
|
|
* the handshake probe has determined the number of paths.
|
|
*/
|
|
if (cp->cp_index > 0 && cp->cp_conn->c_npaths < 2)
|
|
return -EAGAIN;
|
|
|
|
mutex_lock(&tc->t_conn_path_lock);
|
|
|
|
if (rds_conn_path_up(cp)) {
|
|
mutex_unlock(&tc->t_conn_path_lock);
|
|
return 0;
|
|
}
|
|
ret = sock_create_kern(rds_conn_net(conn), PF_INET,
|
|
SOCK_STREAM, IPPROTO_TCP, &sock);
|
|
if (ret < 0)
|
|
goto out;
|
|
|
|
rds_tcp_tune(sock);
|
|
|
|
src.sin_family = AF_INET;
|
|
src.sin_addr.s_addr = (__force u32)conn->c_laddr;
|
|
src.sin_port = (__force u16)htons(0);
|
|
|
|
ret = sock->ops->bind(sock, (struct sockaddr *)&src, sizeof(src));
|
|
if (ret) {
|
|
rdsdebug("bind failed with %d at address %pI4\n",
|
|
ret, &conn->c_laddr);
|
|
goto out;
|
|
}
|
|
|
|
dest.sin_family = AF_INET;
|
|
dest.sin_addr.s_addr = (__force u32)conn->c_faddr;
|
|
dest.sin_port = (__force u16)htons(RDS_TCP_PORT);
|
|
|
|
/*
|
|
* once we call connect() we can start getting callbacks and they
|
|
* own the socket
|
|
*/
|
|
rds_tcp_set_callbacks(sock, cp);
|
|
ret = sock->ops->connect(sock, (struct sockaddr *)&dest, sizeof(dest),
|
|
O_NONBLOCK);
|
|
|
|
rdsdebug("connect to address %pI4 returned %d\n", &conn->c_faddr, ret);
|
|
if (ret == -EINPROGRESS)
|
|
ret = 0;
|
|
if (ret == 0) {
|
|
rds_tcp_keepalive(sock);
|
|
sock = NULL;
|
|
} else {
|
|
rds_tcp_restore_callbacks(sock, cp->cp_transport_data);
|
|
}
|
|
|
|
out:
|
|
mutex_unlock(&tc->t_conn_path_lock);
|
|
if (sock)
|
|
sock_release(sock);
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Before killing the tcp socket this needs to serialize with callbacks. The
|
|
* caller has already grabbed the sending sem so we're serialized with other
|
|
* senders.
|
|
*
|
|
* TCP calls the callbacks with the sock lock so we hold it while we reset the
|
|
* callbacks to those set by TCP. Our callbacks won't execute again once we
|
|
* hold the sock lock.
|
|
*/
|
|
void rds_tcp_conn_path_shutdown(struct rds_conn_path *cp)
|
|
{
|
|
struct rds_tcp_connection *tc = cp->cp_transport_data;
|
|
struct socket *sock = tc->t_sock;
|
|
|
|
rdsdebug("shutting down conn %p tc %p sock %p\n",
|
|
cp->cp_conn, tc, sock);
|
|
|
|
if (sock) {
|
|
if (rds_destroy_pending(cp->cp_conn))
|
|
rds_tcp_set_linger(sock);
|
|
sock->ops->shutdown(sock, RCV_SHUTDOWN | SEND_SHUTDOWN);
|
|
lock_sock(sock->sk);
|
|
rds_tcp_restore_callbacks(sock, tc); /* tc->tc_sock = NULL */
|
|
|
|
release_sock(sock->sk);
|
|
sock_release(sock);
|
|
}
|
|
|
|
if (tc->t_tinc) {
|
|
rds_inc_put(&tc->t_tinc->ti_inc);
|
|
tc->t_tinc = NULL;
|
|
}
|
|
tc->t_tinc_hdr_rem = sizeof(struct rds_header);
|
|
tc->t_tinc_data_rem = 0;
|
|
}
|