/net/ipv4/netfilter/ |
H A D | nf_conntrack_l3proto_ipv4_compat.c | 35 struct ct_iter_state *st = seq->private; local 38 for (st->bucket = 0; 39 st->bucket < net->ct.htable_size; 40 st->bucket++) { 42 hlist_nulls_first_rcu(&net->ct.hash[st->bucket])); 53 struct ct_iter_state *st = seq->private; local 57 if (likely(get_nulls_value(head) == st->bucket)) { 58 if (++st->bucket >= net->ct.htable_size) 62 hlist_nulls_first_rcu(&net->ct.hash[st->bucket])); 221 struct ct_expect_iter_state *st local 237 struct ct_expect_iter_state *st = seq->private; local 365 const struct ip_conntrack_stat *st = v; local [all...] |
/net/netfilter/ |
H A D | nf_conntrack_standalone.c | 57 struct ct_iter_state *st = seq->private; local 60 for (st->bucket = 0; 61 st->bucket < net->ct.htable_size; 62 st->bucket++) { 63 n = rcu_dereference(hlist_nulls_first_rcu(&net->ct.hash[st->bucket])); 74 struct ct_iter_state *st = seq->private; local 78 if (likely(get_nulls_value(head) == st->bucket)) { 79 if (++st->bucket >= net->ct.htable_size) 84 &net->ct.hash[st->bucket])); 102 struct ct_iter_state *st local 147 struct ct_iter_state *st = s->private; local 314 const struct ip_conntrack_stat *st = v; local [all...] |
H A D | xt_qtaguid_print.h | 23 char *pp_sock_tag(struct sock_tag *st); 69 static inline char *pp_sock_tag(struct sock_tag *st) argument
|
H A D | xt_qtaguid_print.c | 227 char *pp_sock_tag(struct sock_tag *st) argument 232 if (!st) { 237 tag_str = pp_tag_t(&st->tag); 242 st, st->sk, st->socket, atomic_long_read( 243 &st->socket->file->f_count), 244 st->pid, tag_str);
|
H A D | xt_recent.c | 461 struct recent_iter_state *st = seq->private; local 462 const struct recent_table *t = st->table; 468 for (st->bucket = 0; st->bucket < ip_list_hash_size; st->bucket++) 469 list_for_each_entry(e, &t->iphash[st->bucket], list) 477 struct recent_iter_state *st = seq->private; local 478 const struct recent_table *t = st->table; 482 while (head == &t->iphash[st->bucket]) { 483 if (++st 524 struct recent_iter_state *st; local [all...] |
H A D | nf_conntrack_expect.c | 468 struct ct_expect_iter_state *st = seq->private; local 471 for (st->bucket = 0; st->bucket < nf_ct_expect_hsize; st->bucket++) { 472 n = rcu_dereference(hlist_first_rcu(&net->ct.expect_hash[st->bucket])); 483 struct ct_expect_iter_state *st = seq->private; local 487 if (++st->bucket >= nf_ct_expect_hsize) 489 head = rcu_dereference(hlist_first_rcu(&net->ct.expect_hash[st->bucket]));
|
H A D | xt_cluster.c | 65 __be32 st = addr->s6_addr32[0]; local 66 return ((st & htonl(0xFF000000)) == htonl(0xFF000000));
|
H A D | nfnetlink_log.c | 933 static struct hlist_node *get_first(struct net *net, struct iter_state *st) argument 936 if (!st) 941 for (st->bucket = 0; st->bucket < INSTANCE_BUCKETS; st->bucket++) { 942 struct hlist_head *head = &log->instance_table[st->bucket]; 950 static struct hlist_node *get_next(struct net *net, struct iter_state *st, argument 958 if (++st->bucket >= INSTANCE_BUCKETS) 962 head = &log->instance_table[st->bucket]; 968 static struct hlist_node *get_idx(struct net *net, struct iter_state *st, argument [all...] |
H A D | nfnetlink_queue_core.c | 1177 struct iter_state *st = seq->private; local 1181 if (!st) 1186 for (st->bucket = 0; st->bucket < INSTANCE_BUCKETS; st->bucket++) { 1187 if (!hlist_empty(&q->instance_table[st->bucket])) 1188 return q->instance_table[st->bucket].first; 1195 struct iter_state *st = seq->private; local 1202 if (++st->bucket >= INSTANCE_BUCKETS) 1206 h = q->instance_table[st [all...] |
H A D | nf_conntrack_netlink.c | 1855 __u16 cpu, const struct ip_conntrack_stat *st) 1871 if (nla_put_be32(skb, CTA_STATS_SEARCHED, htonl(st->searched)) || 1872 nla_put_be32(skb, CTA_STATS_FOUND, htonl(st->found)) || 1873 nla_put_be32(skb, CTA_STATS_NEW, htonl(st->new)) || 1874 nla_put_be32(skb, CTA_STATS_INVALID, htonl(st->invalid)) || 1875 nla_put_be32(skb, CTA_STATS_IGNORE, htonl(st->ignore)) || 1876 nla_put_be32(skb, CTA_STATS_DELETE, htonl(st->delete)) || 1877 nla_put_be32(skb, CTA_STATS_DELETE_LIST, htonl(st->delete_list)) || 1878 nla_put_be32(skb, CTA_STATS_INSERT, htonl(st->insert)) || 1880 htonl(st 1854 ctnetlink_ct_stat_cpu_fill_info(struct sk_buff *skb, u32 portid, u32 seq, __u16 cpu, const struct ip_conntrack_stat *st) argument 1907 const struct ip_conntrack_stat *st; local 3020 ctnetlink_exp_stat_fill_info(struct sk_buff *skb, u32 portid, u32 seq, int cpu, const struct ip_conntrack_stat *st) argument 3061 const struct ip_conntrack_stat *st; local [all...] |
/net/ipv6/ |
H A D | addrconf_core.c | 38 __be32 st; local 40 st = addr->s6_addr32[0]; 45 if ((st & htonl(0xE0000000)) != htonl(0x00000000) && 46 (st & htonl(0xE0000000)) != htonl(0xE0000000)) 50 if ((st & htonl(0xFF000000)) == htonl(0xFF000000)) { 57 if ((st & htonl(0xFFC00000)) == htonl(0xFE800000)) 60 if ((st & htonl(0xFFC00000)) == htonl(0xFEC00000)) 63 if ((st & htonl(0xFE000000)) == htonl(0xFC000000))
|
/net/ipv4/ |
H A D | tcp_ipv4.c | 1858 * starting from bucket given in st->bucket; when st->bucket is zero the 1867 struct tcp_iter_state *st = seq->private; local 1871 ilb = &tcp_hashinfo.listening_hash[st->bucket]; 1874 st->offset = 0; 1877 ilb = &tcp_hashinfo.listening_hash[st->bucket]; 1878 ++st->num; 1879 ++st->offset; 1881 if (st->state == TCP_SEQ_STATE_OPENREQ) { 1884 icsk = inet_csk(st 1945 struct tcp_iter_state *st = seq->private; local 1959 empty_bucket(const struct tcp_iter_state *st) argument 1970 struct tcp_iter_state *st = seq->private; local 2003 struct tcp_iter_state *st = seq->private; local 2023 struct tcp_iter_state *st = seq->private; local 2039 struct tcp_iter_state *st = seq->private; local 2054 struct tcp_iter_state *st = seq->private; local 2088 struct tcp_iter_state *st = seq->private; local 2110 struct tcp_iter_state *st = seq->private; local 2141 struct tcp_iter_state *st = seq->private; local 2308 struct tcp_iter_state *st; local [all...] |
H A D | ip_input.c | 346 struct ip_rt_acct *st = this_cpu_ptr(ip_rt_acct); local 348 st[idx&0xFF].o_packets++; 349 st[idx&0xFF].o_bytes += skb->len; 350 st[(idx>>16)&0xFF].i_packets++; 351 st[(idx>>16)&0xFF].i_bytes += skb->len;
|
/net/sched/ |
H A D | sch_codel.c | 215 struct tc_codel_xstats st = { local 229 st.drop_next = codel_time_to_us(delta); 231 st.drop_next = -codel_time_to_us(-delta); 234 return gnet_stats_copy_app(d, &st, sizeof(st));
|
H A D | sch_fq_codel.c | 460 struct tc_fq_codel_xstats st = { local 465 st.qdisc_stats.maxpacket = q->cstats.maxpacket; 466 st.qdisc_stats.drop_overlimit = q->drop_overlimit; 467 st.qdisc_stats.ecn_mark = q->cstats.ecn_mark; 468 st.qdisc_stats.new_flow_count = q->new_flow_count; 471 st.qdisc_stats.new_flows_len++; 474 st.qdisc_stats.old_flows_len++; 476 return gnet_stats_copy_app(d, &st, sizeof(st));
|
H A D | sch_sfb.c | 589 struct tc_sfb_xstats st = { local 598 st.maxqlen = sfb_compute_qlen(&st.maxprob, &st.avgprob, q); 600 return gnet_stats_copy_app(d, &st, sizeof(st));
|
H A D | sch_red.c | 288 struct tc_red_xstats st = { local 295 return gnet_stats_copy_app(d, &st, sizeof(st));
|
H A D | sch_choke.c | 530 struct tc_choke_xstats st = { local 538 return gnet_stats_copy_app(d, &st, sizeof(st));
|
H A D | sch_hhf.c | 699 struct tc_hhf_xstats st = { local 706 return gnet_stats_copy_app(d, &st, sizeof(st));
|
H A D | sch_pie.c | 492 struct tc_pie_xstats st = { local 506 return gnet_stats_copy_app(d, &st, sizeof(st));
|
/net/bridge/netfilter/ |
H A D | ebt_stp.c | 144 const struct stp_config_pdu *st; local 147 st = skb_header_pointer(skb, sizeof(_stph), 149 if (st == NULL) 151 return ebt_filter_config(info, st);
|
/net/ipx/ |
H A D | ipx_route.c | 261 struct sockaddr_ipx *sg, *st; local 268 st = (struct sockaddr_ipx *)&rt.rt_dst; 273 st->sipx_family != AF_IPX) 278 rc = ipxrtr_delete(st->sipx_network); 282 f.ipx_network = st->sipx_network;
|
/net/core/ |
H A D | skbuff.c | 2630 * @st: state variable 2636 unsigned int to, struct skb_seq_state *st) 2638 st->lower_offset = from; 2639 st->upper_offset = to; 2640 st->root_skb = st->cur_skb = skb; 2641 st->frag_idx = st->stepped_offset = 0; 2642 st->frag_data = NULL; 2650 * @st 2635 skb_prepare_seq_read(struct sk_buff *skb, unsigned int from, unsigned int to, struct skb_seq_state *st) argument 2671 skb_seq_read(unsigned int consumed, const u8 **data, struct skb_seq_state *st) argument 2745 skb_abort_seq_read(struct skb_seq_state *st) argument [all...] |
H A D | gen_stats.c | 294 * @st: application specific statistics data 305 gnet_stats_copy_app(struct gnet_dump *d, void *st, int len) argument 308 d->xstats = st; 313 return gnet_stats_copy(d, TCA_STATS_APP, st, len);
|
H A D | neighbour.c | 1898 struct neigh_statistics *st; local 1900 st = per_cpu_ptr(tbl->stats, cpu); 1901 ndst.ndts_allocs += st->allocs; 1902 ndst.ndts_destroys += st->destroys; 1903 ndst.ndts_hash_grows += st->hash_grows; 1904 ndst.ndts_res_failed += st->res_failed; 1905 ndst.ndts_lookups += st->lookups; 1906 ndst.ndts_hits += st->hits; 1907 ndst.ndts_rcv_probes_mcast += st->rcv_probes_mcast; 1908 ndst.ndts_rcv_probes_ucast += st 2721 struct neigh_statistics *st = v; local [all...] |