forked from luck/tmp_suning_uos_patched
net: shrink struct sock and request_sock by 8 bytes
One 32bit hole is following skc_refcnt, use it. skc_incoming_cpu can also be an union for request_sock rcv_wnd. Signed-off-by: Eric Dumazet <edumazet@google.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
8e5eb54d30
commit
ed53d0ab76
|
@ -51,15 +51,14 @@ struct request_sock {
|
|||
#define rsk_refcnt __req_common.skc_refcnt
|
||||
#define rsk_hash __req_common.skc_hash
|
||||
#define rsk_listener __req_common.skc_listener
|
||||
#define rsk_window_clamp __req_common.skc_window_clamp
|
||||
#define rsk_rcv_wnd __req_common.skc_rcv_wnd
|
||||
|
||||
struct request_sock *dl_next;
|
||||
u16 mss;
|
||||
u8 num_retrans; /* number of retransmits */
|
||||
u8 cookie_ts:1; /* syncookie: encode tcpopts in timestamp */
|
||||
u8 num_timeout:7; /* number of timeouts */
|
||||
/* The following two fields can be easily recomputed I think -AK */
|
||||
u32 window_clamp; /* window clamp at creation time */
|
||||
u32 rcv_wnd; /* rcv_wnd offered first time */
|
||||
u32 ts_recent;
|
||||
struct timer_list rsk_timer;
|
||||
const struct request_sock_ops *rsk_ops;
|
||||
|
|
|
@ -226,11 +226,18 @@ struct sock_common {
|
|||
struct hlist_nulls_node skc_nulls_node;
|
||||
};
|
||||
int skc_tx_queue_mapping;
|
||||
int skc_incoming_cpu;
|
||||
union {
|
||||
int skc_incoming_cpu;
|
||||
u32 skc_rcv_wnd;
|
||||
};
|
||||
|
||||
atomic_t skc_refcnt;
|
||||
/* private: */
|
||||
int skc_dontcopy_end[0];
|
||||
union {
|
||||
u32 skc_rxhash;
|
||||
u32 skc_window_clamp;
|
||||
};
|
||||
/* public: */
|
||||
};
|
||||
|
||||
|
@ -287,7 +294,6 @@ struct cg_proto;
|
|||
* @sk_rcvlowat: %SO_RCVLOWAT setting
|
||||
* @sk_rcvtimeo: %SO_RCVTIMEO setting
|
||||
* @sk_sndtimeo: %SO_SNDTIMEO setting
|
||||
* @sk_rxhash: flow hash received from netif layer
|
||||
* @sk_txhash: computed flow hash for use on transmit
|
||||
* @sk_filter: socket filtering instructions
|
||||
* @sk_timer: sock cleanup timer
|
||||
|
@ -346,6 +352,7 @@ struct sock {
|
|||
#define sk_cookie __sk_common.skc_cookie
|
||||
#define sk_incoming_cpu __sk_common.skc_incoming_cpu
|
||||
#define sk_flags __sk_common.skc_flags
|
||||
#define sk_rxhash __sk_common.skc_rxhash
|
||||
|
||||
socket_lock_t sk_lock;
|
||||
struct sk_buff_head sk_receive_queue;
|
||||
|
@ -365,9 +372,6 @@ struct sock {
|
|||
} sk_backlog;
|
||||
#define sk_rmem_alloc sk_backlog.rmem_alloc
|
||||
int sk_forward_alloc;
|
||||
#ifdef CONFIG_RPS
|
||||
__u32 sk_rxhash;
|
||||
#endif
|
||||
|
||||
__u32 sk_txhash;
|
||||
#ifdef CONFIG_NET_RX_BUSY_POLL
|
||||
|
|
|
@ -382,10 +382,10 @@ struct sock *cookie_v4_check(struct sock *sk, struct sk_buff *skb)
|
|||
}
|
||||
|
||||
/* Try to redo what tcp_v4_send_synack did. */
|
||||
req->window_clamp = tp->window_clamp ? :dst_metric(&rt->dst, RTAX_WINDOW);
|
||||
req->rsk_window_clamp = tp->window_clamp ? :dst_metric(&rt->dst, RTAX_WINDOW);
|
||||
|
||||
tcp_select_initial_window(tcp_full_space(sk), req->mss,
|
||||
&req->rcv_wnd, &req->window_clamp,
|
||||
&req->rsk_rcv_wnd, &req->rsk_window_clamp,
|
||||
ireq->wscale_ok, &rcv_wscale,
|
||||
dst_metric(&rt->dst, RTAX_INITRWND));
|
||||
|
||||
|
|
|
@ -6022,7 +6022,7 @@ static void tcp_openreq_init(struct request_sock *req,
|
|||
{
|
||||
struct inet_request_sock *ireq = inet_rsk(req);
|
||||
|
||||
req->rcv_wnd = 0; /* So that tcp_send_synack() knows! */
|
||||
req->rsk_rcv_wnd = 0; /* So that tcp_send_synack() knows! */
|
||||
req->cookie_ts = 0;
|
||||
tcp_rsk(req)->rcv_isn = TCP_SKB_CB(skb)->seq;
|
||||
tcp_rsk(req)->rcv_nxt = TCP_SKB_CB(skb)->seq + 1;
|
||||
|
|
|
@ -803,7 +803,7 @@ static void tcp_v4_reqsk_send_ack(const struct sock *sk, struct sk_buff *skb,
|
|||
*/
|
||||
tcp_v4_send_ack(skb, (sk->sk_state == TCP_LISTEN) ?
|
||||
tcp_rsk(req)->snt_isn + 1 : tcp_sk(sk)->snd_nxt,
|
||||
tcp_rsk(req)->rcv_nxt, req->rcv_wnd,
|
||||
tcp_rsk(req)->rcv_nxt, req->rsk_rcv_wnd,
|
||||
tcp_time_stamp,
|
||||
req->ts_recent,
|
||||
0,
|
||||
|
|
|
@ -381,18 +381,18 @@ void tcp_openreq_init_rwin(struct request_sock *req,
|
|||
|
||||
window_clamp = READ_ONCE(tp->window_clamp);
|
||||
/* Set this up on the first call only */
|
||||
req->window_clamp = window_clamp ? : dst_metric(dst, RTAX_WINDOW);
|
||||
req->rsk_window_clamp = window_clamp ? : dst_metric(dst, RTAX_WINDOW);
|
||||
|
||||
/* limit the window selection if the user enforce a smaller rx buffer */
|
||||
if (sk_listener->sk_userlocks & SOCK_RCVBUF_LOCK &&
|
||||
(req->window_clamp > full_space || req->window_clamp == 0))
|
||||
req->window_clamp = full_space;
|
||||
(req->rsk_window_clamp > full_space || req->rsk_window_clamp == 0))
|
||||
req->rsk_window_clamp = full_space;
|
||||
|
||||
/* tcp_full_space because it is guaranteed to be the first packet */
|
||||
tcp_select_initial_window(full_space,
|
||||
mss - (ireq->tstamp_ok ? TCPOLEN_TSTAMP_ALIGNED : 0),
|
||||
&req->rcv_wnd,
|
||||
&req->window_clamp,
|
||||
&req->rsk_rcv_wnd,
|
||||
&req->rsk_window_clamp,
|
||||
ireq->wscale_ok,
|
||||
&rcv_wscale,
|
||||
dst_metric(dst, RTAX_INITRWND));
|
||||
|
@ -512,9 +512,9 @@ struct sock *tcp_create_openreq_child(const struct sock *sk,
|
|||
if (sysctl_tcp_fack)
|
||||
tcp_enable_fack(newtp);
|
||||
}
|
||||
newtp->window_clamp = req->window_clamp;
|
||||
newtp->rcv_ssthresh = req->rcv_wnd;
|
||||
newtp->rcv_wnd = req->rcv_wnd;
|
||||
newtp->window_clamp = req->rsk_window_clamp;
|
||||
newtp->rcv_ssthresh = req->rsk_rcv_wnd;
|
||||
newtp->rcv_wnd = req->rsk_rcv_wnd;
|
||||
newtp->rx_opt.wscale_ok = ireq->wscale_ok;
|
||||
if (newtp->rx_opt.wscale_ok) {
|
||||
newtp->rx_opt.snd_wscale = ireq->snd_wscale;
|
||||
|
@ -707,7 +707,7 @@ struct sock *tcp_check_req(struct sock *sk, struct sk_buff *skb,
|
|||
/* RFC793: "first check sequence number". */
|
||||
|
||||
if (paws_reject || !tcp_in_window(TCP_SKB_CB(skb)->seq, TCP_SKB_CB(skb)->end_seq,
|
||||
tcp_rsk(req)->rcv_nxt, tcp_rsk(req)->rcv_nxt + req->rcv_wnd)) {
|
||||
tcp_rsk(req)->rcv_nxt, tcp_rsk(req)->rcv_nxt + req->rsk_rcv_wnd)) {
|
||||
/* Out of window: send ACK and drop. */
|
||||
if (!(flg & TCP_FLAG_RST))
|
||||
req->rsk_ops->send_ack(sk, skb, req);
|
||||
|
|
|
@ -3023,7 +3023,7 @@ struct sk_buff *tcp_make_synack(const struct sock *sk, struct dst_entry *dst,
|
|||
th->ack_seq = htonl(tcp_rsk(req)->rcv_nxt);
|
||||
|
||||
/* RFC1323: The window in SYN & SYN/ACK segments is never scaled. */
|
||||
th->window = htons(min(req->rcv_wnd, 65535U));
|
||||
th->window = htons(min(req->rsk_rcv_wnd, 65535U));
|
||||
tcp_options_write((__be32 *)(th + 1), NULL, &opts);
|
||||
th->doff = (tcp_header_size >> 2);
|
||||
TCP_INC_STATS_BH(sock_net(sk), TCP_MIB_OUTSEGS);
|
||||
|
|
|
@ -235,9 +235,9 @@ struct sock *cookie_v6_check(struct sock *sk, struct sk_buff *skb)
|
|||
goto out_free;
|
||||
}
|
||||
|
||||
req->window_clamp = tp->window_clamp ? :dst_metric(dst, RTAX_WINDOW);
|
||||
req->rsk_window_clamp = tp->window_clamp ? :dst_metric(dst, RTAX_WINDOW);
|
||||
tcp_select_initial_window(tcp_full_space(sk), req->mss,
|
||||
&req->rcv_wnd, &req->window_clamp,
|
||||
&req->rsk_rcv_wnd, &req->rsk_window_clamp,
|
||||
ireq->wscale_ok, &rcv_wscale,
|
||||
dst_metric(dst, RTAX_INITRWND));
|
||||
|
||||
|
|
|
@ -931,7 +931,7 @@ static void tcp_v6_reqsk_send_ack(const struct sock *sk, struct sk_buff *skb,
|
|||
*/
|
||||
tcp_v6_send_ack(sk, skb, (sk->sk_state == TCP_LISTEN) ?
|
||||
tcp_rsk(req)->snt_isn + 1 : tcp_sk(sk)->snd_nxt,
|
||||
tcp_rsk(req)->rcv_nxt, req->rcv_wnd,
|
||||
tcp_rsk(req)->rcv_nxt, req->rsk_rcv_wnd,
|
||||
tcp_time_stamp, req->ts_recent, sk->sk_bound_dev_if,
|
||||
tcp_v6_md5_do_lookup(sk, &ipv6_hdr(skb)->daddr),
|
||||
0, 0);
|
||||
|
|
Loading…
Reference in New Issue
Block a user