/net/atm/ |
H A D | atm_misc.c | 75 int atm_pcr_goal(const struct atm_trafprm *tp) argument 77 if (tp->pcr && tp->pcr != ATM_MAX_PCR) 78 return -tp->pcr; 79 if (tp->min_pcr && !tp->pcr) 80 return tp->min_pcr; 81 if (tp->max_pcr != ATM_MAX_PCR) 82 return -tp->max_pcr;
|
/net/ipv4/ |
H A D | tcp_scalable.c | 20 struct tcp_sock *tp = tcp_sk(sk); local 25 if (tp->snd_cwnd <= tp->snd_ssthresh) 26 tcp_slow_start(tp, acked); 28 tcp_cong_avoid_ai(tp, min(tp->snd_cwnd, TCP_SCALABLE_AI_CNT)); 33 const struct tcp_sock *tp = tcp_sk(sk); local 35 return max(tp->snd_cwnd - (tp->snd_cwnd>>TCP_SCALABLE_MD_SCALE), 2U);
|
H A D | tcp_diag.c | 22 const struct tcp_sock *tp = tcp_sk(sk); local 29 r->idiag_rqueue = max_t(int, tp->rcv_nxt - tp->copied_seq, 0); 30 r->idiag_wqueue = tp->write_seq - tp->snd_una;
|
H A D | tcp_highspeed.c | 101 struct tcp_sock *tp = tcp_sk(sk); local 108 tp->snd_cwnd_clamp = min_t(u32, tp->snd_cwnd_clamp, 0xffffffff/128); 113 struct tcp_sock *tp = tcp_sk(sk); local 119 if (tp->snd_cwnd <= tp->snd_ssthresh) 120 tcp_slow_start(tp, acked); 129 if (tp->snd_cwnd > hstcp_aimd_vals[ca->ai].cwnd) { 130 while (tp->snd_cwnd > hstcp_aimd_vals[ca->ai].cwnd && 133 } else if (ca->ai && tp 152 const struct tcp_sock *tp = tcp_sk(sk); local [all...] |
H A D | tcp_bic.c | 143 struct tcp_sock *tp = tcp_sk(sk); local 149 if (tp->snd_cwnd <= tp->snd_ssthresh) 150 tcp_slow_start(tp, acked); 152 bictcp_update(ca, tp->snd_cwnd); 153 tcp_cong_avoid_ai(tp, ca->cnt); 163 const struct tcp_sock *tp = tcp_sk(sk); local 169 if (tp->snd_cwnd < ca->last_max_cwnd && fast_convergence) 170 ca->last_max_cwnd = (tp->snd_cwnd * (BICTCP_BETA_SCALE + beta)) 173 ca->last_max_cwnd = tp 185 const struct tcp_sock *tp = tcp_sk(sk); local [all...] |
H A D | tcp_cong.c | 294 void tcp_slow_start(struct tcp_sock *tp, u32 acked) argument 296 u32 cwnd = tp->snd_cwnd + acked; 298 if (cwnd > tp->snd_ssthresh) 299 cwnd = tp->snd_ssthresh + 1; 300 tp->snd_cwnd = min(cwnd, tp->snd_cwnd_clamp); 304 /* In theory this is tp->snd_cwnd += 1 / tp->snd_cwnd (or alternative w) */ 305 void tcp_cong_avoid_ai(struct tcp_sock *tp, u32 w) argument 307 if (tp 326 struct tcp_sock *tp = tcp_sk(sk); local 343 const struct tcp_sock *tp = tcp_sk(sk); local [all...] |
H A D | tcp_dctcp.c | 76 static void dctcp_reset(const struct tcp_sock *tp, struct dctcp *ca) argument 78 ca->next_seq = tp->snd_nxt; 86 const struct tcp_sock *tp = tcp_sk(sk); local 88 if ((tp->ecn_flags & TCP_ECN_OK) || 93 ca->prior_snd_una = tp->snd_una; 94 ca->prior_rcv_nxt = tp->rcv_nxt; 101 dctcp_reset(tp, ca); 115 struct tcp_sock *tp = tcp_sk(sk); local 117 return max(tp->snd_cwnd - ((tp 129 struct tcp_sock *tp = tcp_sk(sk); local 159 struct tcp_sock *tp = tcp_sk(sk); local 188 const struct tcp_sock *tp = tcp_sk(sk); local [all...] |
H A D | tcp_fastopen.c | 134 struct tcp_sock *tp; local 154 tp = tcp_sk(child); 156 tp->fastopen_rsk = req; 167 tp->snd_wnd = ntohs(tcp_hdr(skb)->window); 199 tp->syn_data_acked = 1; 201 tcp_rsk(req)->rcv_nxt = tp->rcv_nxt = TCP_SKB_CB(skb)->end_seq;
|
H A D | tcp_hybla.c | 47 struct tcp_sock *tp = tcp_sk(sk); local 56 tp->snd_cwnd = 2; 57 tp->snd_cwnd_clamp = 65535; 63 ca->minrtt_us = tp->srtt_us; 64 tp->snd_cwnd = ca->rho; 91 struct tcp_sock *tp = tcp_sk(sk); local 97 if (tp->srtt_us < ca->minrtt_us) { 99 ca->minrtt_us = tp->srtt_us; 115 if (tp->snd_cwnd < tp [all...] |
H A D | tcp_lp.c | 135 struct tcp_sock *tp = tcp_sk(sk); local 146 if (tp->rx_opt.rcv_tsval == lp->remote_ref_time || 147 tp->rx_opt.rcv_tsecr == lp->local_ref_time) 150 m = HZ * (tp->rx_opt.rcv_tsval - 151 lp->remote_ref_time) / (tp->rx_opt.rcv_tsecr - 170 lp->remote_ref_time = tp->rx_opt.rcv_tsval; 171 lp->local_ref_time = tp->rx_opt.rcv_tsecr; 188 struct tcp_sock *tp = tcp_sk(sk); local 196 tp->rx_opt.rcv_tsval * (LP_RESOL / lp->remote_hz) - 197 tp 265 struct tcp_sock *tp = tcp_sk(sk); local [all...] |
H A D | tcp_vegas.c | 72 const struct tcp_sock *tp = tcp_sk(sk); local 79 vegas->beg_snd_nxt = tp->snd_nxt; 159 static inline u32 tcp_vegas_ssthresh(struct tcp_sock *tp) argument 161 return min(tp->snd_ssthresh, tp->snd_cwnd-1); 166 struct tcp_sock *tp = tcp_sk(sk); local 180 vegas->beg_snd_nxt = tp->snd_nxt; 219 target_cwnd = (u64)tp->snd_cwnd * vegas->baseRTT; 226 diff = tp->snd_cwnd * (rtt-vegas->baseRTT) / vegas->baseRTT; 228 if (diff > gamma && tp [all...] |
H A D | tcp_veno.c | 119 struct tcp_sock *tp = tcp_sk(sk); local 147 target_cwnd = (u64)tp->snd_cwnd * veno->basertt; 151 veno->diff = (tp->snd_cwnd << V_PARAM_SHIFT) - target_cwnd; 153 if (tp->snd_cwnd <= tp->snd_ssthresh) { 155 tcp_slow_start(tp, acked); 162 tcp_cong_avoid_ai(tp, tp->snd_cwnd); 167 if (tp->snd_cwnd_cnt >= tp 192 const struct tcp_sock *tp = tcp_sk(sk); local [all...] |
H A D | tcp_yeah.c | 44 struct tcp_sock *tp = tcp_sk(sk); local 56 tp->snd_cwnd_clamp = min_t(u32, tp->snd_cwnd_clamp, 0xffffffff/128); 72 struct tcp_sock *tp = tcp_sk(sk); local 78 if (tp->snd_cwnd <= tp->snd_ssthresh) 79 tcp_slow_start(tp, acked); 84 tp->snd_cwnd_cnt += yeah->pkts_acked; 85 if (tp->snd_cwnd_cnt > min(tp 206 const struct tcp_sock *tp = tcp_sk(sk); local [all...] |
H A D | syncookies.c | 264 struct tcp_sock *tp = tcp_sk(sk); local 350 req->window_clamp = tp->window_clamp ? :dst_metric(&rt->dst, RTAX_WINDOW);
|
H A D | tcp_htcp.c | 69 const struct tcp_sock *tp = tcp_sk(sk); local 79 return max(tp->snd_cwnd, (tp->snd_ssthresh << 7) / ca->beta); 105 const struct tcp_sock *tp = tcp_sk(sk); local 127 if (ca->packetcount >= tp->snd_cwnd - (ca->alpha >> 7 ? : 1) && 224 const struct tcp_sock *tp = tcp_sk(sk); local 228 return max((tp->snd_cwnd * ca->beta) >> 7, 2U); 233 struct tcp_sock *tp = tcp_sk(sk); local 239 if (tp->snd_cwnd <= tp [all...] |
H A D | tcp_illinois.c | 58 struct tcp_sock *tp = tcp_sk(sk); local 61 ca->end_seq = tp->snd_nxt; 222 struct tcp_sock *tp = tcp_sk(sk); local 225 if (tp->snd_cwnd < win_thresh) { 260 struct tcp_sock *tp = tcp_sk(sk); local 271 if (tp->snd_cwnd <= tp->snd_ssthresh) 272 tcp_slow_start(tp, acked); 278 tp->snd_cwnd_cnt += ca->acked; 282 * tp 295 struct tcp_sock *tp = tcp_sk(sk); local [all...] |
H A D | tcp_minisocks.c | 270 const struct tcp_sock *tp = tcp_sk(sk); local 273 if (tcp_death_row.sysctl_tw_recycle && tp->rx_opt.ts_recent_stamp) 285 tw->tw_rcv_wscale = tp->rx_opt.rcv_wscale; 286 tcptw->tw_rcv_nxt = tp->rcv_nxt; 287 tcptw->tw_snd_nxt = tp->snd_nxt; 288 tcptw->tw_rcv_wnd = tcp_receive_window(tp); 289 tcptw->tw_ts_recent = tp->rx_opt.ts_recent; 290 tcptw->tw_ts_recent_stamp = tp->rx_opt.ts_recent_stamp; 291 tcptw->tw_ts_offset = tp->tsoffset; 315 key = tp 369 struct tcp_sock *tp = tcp_sk(sk); local 396 tcp_ecn_openreq_child(struct tcp_sock *tp, const struct request_sock *req) argument [all...] |
H A D | tcp_probe.c | 110 const struct tcp_sock *tp = tcp_sk(sk); local 118 (full || tp->snd_cwnd != tcp_probe.lastcwnd)) { 149 p->snd_nxt = tp->snd_nxt; 150 p->snd_una = tp->snd_una; 151 p->snd_cwnd = tp->snd_cwnd; 152 p->snd_wnd = tp->snd_wnd; 153 p->rcv_wnd = tp->rcv_wnd; 155 p->srtt = tp->srtt_us >> 3; 159 tcp_probe.lastcwnd = tp->snd_cwnd;
|
H A D | tcp_timer.c | 57 struct tcp_sock *tp = tcp_sk(sk); local 62 if ((s32)(tcp_time_stamp - tp->lsndtime) > 2*TCP_RTO_MAX || !do_reset) 72 if ((s32)(tcp_time_stamp - tp->lsndtime) <= TCP_TIMEWAIT_LEN || 74 (!tp->snd_wnd && !tp->packets_out)) 110 struct tcp_sock *tp = tcp_sk(sk); local 115 mss = max(mss, 68 - tp->tcp_header_len); 158 struct tcp_sock *tp = tcp_sk(sk); local 165 if (tp->syn_fastopen || tp 205 struct tcp_sock *tp = tcp_sk(sk); local 271 struct tcp_sock *tp = tcp_sk(sk); local 349 struct tcp_sock *tp = tcp_sk(sk); local 568 struct tcp_sock *tp = tcp_sk(sk); local [all...] |
H A D | tcp_westwood.c | 121 * to fix mismatch between tp->snd_una and w->snd_una for the first 163 const struct tcp_sock *tp = tcp_sk(sk); local 168 w->bk += tp->snd_una - w->snd_una; 169 w->snd_una = tp->snd_una; 180 const struct tcp_sock *tp = tcp_sk(sk); local 183 w->cumul_ack = tp->snd_una - w->snd_una; 186 * tp->snd_una. 189 w->accounted += tp->mss_cache; 190 w->cumul_ack = tp->mss_cache; 193 if (w->cumul_ack > tp 217 const struct tcp_sock *tp = tcp_sk(sk); local 240 struct tcp_sock *tp = tcp_sk(sk); local [all...] |
/net/ipv6/netfilter/ |
H A D | ip6t_hbh.c | 58 const u_int8_t *tp = NULL; local 108 tp = skb_header_pointer(skb, ptr, sizeof(_opttype), 110 if (tp == NULL) 114 if (*tp != (optinfo->opts[temp] & 0xFF00) >> 8) { 115 pr_debug("Tbad %02X %02X\n", *tp, 122 if (*tp) {
|
/net/mac80211/ |
H A D | rc80211_minstrel_debugfs.c | 62 unsigned int i, tp, prob, eprob; local 85 tp = MINSTREL_TRUNC(mrs->cur_tp / 10); 91 tp / 10, tp % 10,
|
H A D | rc80211_minstrel_ht_debugfs.c | 23 unsigned int j, tp, prob, eprob; local 62 tp = mr->cur_tp / 10; 68 tp / 10, tp % 10,
|
/net/sched/ |
H A D | cls_cgroup.c | 25 struct tcf_proto *tp; member in struct:cls_cgroup_head 29 static int cls_cgroup_classify(struct sk_buff *skb, const struct tcf_proto *tp, argument 32 struct cls_cgroup_head *head = rcu_dereference_bh(tp->root); 65 static unsigned long cls_cgroup_get(struct tcf_proto *tp, u32 handle) argument 70 static void cls_cgroup_put(struct tcf_proto *tp, unsigned long f) argument 74 static int cls_cgroup_init(struct tcf_proto *tp) argument 95 struct tcf_proto *tp, unsigned long base, 100 struct cls_cgroup_head *head = rtnl_dereference(tp->root); 125 new->tp = tp; 94 cls_cgroup_change(struct net *net, struct sk_buff *in_skb, struct tcf_proto *tp, unsigned long base, u32 handle, struct nlattr **tca, unsigned long *arg, bool ovr) argument 154 cls_cgroup_destroy(struct tcf_proto *tp) argument 164 cls_cgroup_delete(struct tcf_proto *tp, unsigned long arg) argument 169 cls_cgroup_walk(struct tcf_proto *tp, struct tcf_walker *arg) argument 184 cls_cgroup_dump(struct net *net, struct tcf_proto *tp, unsigned long fh, struct sk_buff *skb, struct tcmsg *t) argument [all...] |
/net/ipv6/ |
H A D | syncookies.c | 161 struct tcp_sock *tp = tcp_sk(sk); local 258 req->window_clamp = tp->window_clamp ? :dst_metric(dst, RTAX_WINDOW);
|