Searched defs:start_seq (Results 1 - 1 of 1) sorted by relevance
/net/ipv4/ |
H A D | tcp_input.c | 891 * Implements also blockage to start_seq wrap-around. Problem lies in the 892 * fact that though start_seq (s) is before end_seq (i.e., not reversed), 894 * happens when start_seq resides between end_seq wrap (e_w) and snd_nxt 905 * crazy SACK blocks. Doing this check for start_seq alone closes somewhat 924 u32 start_seq, u32 end_seq) 927 if (after(end_seq, tp->snd_nxt) || !before(start_seq, end_seq)) 930 /* Nasty start_seq wrap-around check (see comments above) */ 931 if (!before(start_seq, tp->snd_nxt)) 935 * start_seq == snd_una is non-sensical (see comments above) 937 if (after(start_seq, t 923 tcp_is_sackblock_valid(struct tcp_sock *tp, bool is_dsack, u32 start_seq, u32 end_seq) argument 1073 tcp_match_skb_to_sack(struct sock *sk, struct sk_buff *skb, u32 start_seq, u32 end_seq) argument 1120 tcp_sacktag_one(struct sock *sk, struct tcp_sacktag_state *state, u8 sacked, u32 start_seq, u32 end_seq, bool dup_sack, int pcount) argument 1208 u32 start_seq = TCP_SKB_CB(skb)->seq; /* start of newly-SACKed */ local 1297 tcp_shift_skb_data(struct sock *sk, struct sk_buff *skb, struct tcp_sacktag_state *state, u32 start_seq, u32 end_seq, bool dup_sack) argument 1435 tcp_sacktag_walk(struct sk_buff *skb, struct sock *sk, struct tcp_sack_block *next_dup, struct tcp_sacktag_state *state, u32 start_seq, u32 end_seq, bool dup_sack_in) argument 1661 u32 start_seq = sp[i].start_seq; local [all...] |
Completed in 24 milliseconds