Lines Matching refs:req
337 struct request_sock *req;
379 req = tp->fastopen_rsk;
383 (req == NULL || seq != tcp_rsk(req)->snt_isn)) {
468 if (req && req->sk == NULL)
472 struct request_sock *req, **prev;
477 req = inet_csk_search_req(sk, &prev, th->dest,
479 if (!req)
485 WARN_ON(req->sk);
487 if (seq != tcp_rsk(req)->snt_isn) {
498 inet_csk_reqsk_queue_drop(sk, req, prev);
817 struct request_sock *req)
823 tcp_rsk(req)->snt_isn + 1 : tcp_sk(sk)->snd_nxt,
824 tcp_rsk(req)->rcv_nxt, req->rcv_wnd,
826 req->ts_recent,
830 inet_rsk(req)->no_srccheck ? IP_REPLY_ARG_NOSRCCHECK : 0,
840 struct request_sock *req,
844 const struct inet_request_sock *ireq = inet_rsk(req);
850 if (!dst && (dst = inet_csk_route_req(sk, &fl4, req)) == NULL)
853 skb = tcp_make_synack(sk, dst, req, NULL);
863 if (!tcp_rsk(req)->snt_synack && !err)
864 tcp_rsk(req)->snt_synack = tcp_time_stamp;
870 static int tcp_v4_rtx_synack(struct sock *sk, struct request_sock *req)
872 int res = tcp_v4_send_synack(sk, NULL, req, 0, false);
882 static void tcp_v4_reqsk_destructor(struct request_sock *req)
884 kfree(inet_rsk(req)->opt);
989 struct request_sock *req)
993 addr = (union tcp_md5_addr *)&inet_rsk(req)->rmt_addr;
1166 const struct sock *sk, const struct request_sock *req,
1177 } else if (req) {
1178 saddr = inet_rsk(req)->loc_addr;
1179 daddr = inet_rsk(req)->rmt_addr;
1289 struct request_sock *req,
1341 tcp_rsk(req)->rcv_nxt = TCP_SKB_CB(skb)->end_seq;
1354 tcp_rsk(req)->rcv_nxt = TCP_SKB_CB(skb)->end_seq;
1372 struct request_sock *req)
1376 const struct inet_request_sock *ireq = inet_rsk(req);
1380 req->num_retrans = 0;
1381 req->num_timeout = 0;
1382 req->sk = NULL;
1384 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL);
1395 tcp_rsk(req)->snt_synack = tcp_time_stamp;
1408 tp->fastopen_rsk = req;
1414 tcp_rsk(req)->listener = sk;
1429 inet_csk_reqsk_queue_add(sk, req, child);
1461 WARN_ON(req->sk == NULL);
1468 struct request_sock *req;
1506 req = inet_reqsk_alloc(&tcp_request_sock_ops);
1507 if (!req)
1511 tcp_rsk(req)->af_specific = &tcp_request_sock_ipv4_ops;
1523 tcp_openreq_init(req, &tmp_opt, skb);
1525 ireq = inet_rsk(req);
1532 if (security_inet_conn_request(sk, skb, req))
1536 TCP_ECN_create_request(req, skb, sock_net(sk));
1539 isn = cookie_v4_init_sequence(sk, skb, &req->mss);
1540 req->cookie_ts = tmp_opt.tstamp_ok;
1553 (dst = inet_csk_route_req(sk, &fl4, req)) != NULL &&
1555 if (!tcp_peer_is_proven(req, dst, true)) {
1564 !tcp_peer_is_proven(req, dst, false)) {
1579 tcp_rsk(req)->snt_isn = isn;
1582 dst = inet_csk_route_req(sk, &fl4, req);
1586 do_fastopen = tcp_fastopen_check(sk, skb, req, &foc, &valid_foc);
1599 skb_synack = tcp_make_synack(sk, dst, req,
1616 tcp_rsk(req)->snt_synack = tcp_time_stamp;
1617 tcp_rsk(req)->listener = NULL;
1619 inet_csk_reqsk_queue_hash_add(sk, req, TCP_TIMEOUT_INIT);
1623 } else if (tcp_v4_conn_req_fastopen(sk, skb, skb_synack, req))
1631 reqsk_free(req);
1644 struct request_sock *req,
1659 newsk = tcp_create_openreq_child(sk, req, skb);
1668 ireq = inet_rsk(req);
1684 dst = inet_csk_route_child_sock(sk, newsk, req);
1700 tcp_synack_rtt_meas(newsk, req);
1701 newtp->total_retrans = req->num_retrans;
1747 struct request_sock *req = inet_csk_search_req(sk, &prev, th->source,
1749 if (req)
1750 return tcp_check_req(sk, skb, req, prev, false);
2261 struct request_sock *req = cur;
2264 req = req->dl_next;
2266 while (req) {
2267 if (req->rsk_ops->family == st->family) {
2268 cur = req;
2271 req = req->dl_next;
2276 req = icsk->icsk_accept_queue.listen_opt->syn_table[st->sbucket];
2626 static void get_openreq4(const struct sock *sk, const struct request_sock *req,
2629 const struct inet_request_sock *ireq = inet_rsk(req);
2630 long delta = req->expires - jiffies;
2643 req->num_timeout,
2648 req,