/net/netfilter/ipset/ |
H A D | ip_set_bitmap_gen.h | 272 .head = mtype_head,
|
H A D | ip_set_bitmap_ipmac.c | 231 if (skb_mac_header(skb) < skb->head ||
|
/net/sctp/ |
H A D | outqueue.c | 361 static void sctp_insert_list(struct list_head *head, struct list_head *new) argument 371 list_for_each(pos, head) { 381 list_add_tail(new, head); 1024 SCTP_DEBUG_PRINTK("TX TSN 0x%x skb->head " 1027 chunk->skb ?chunk->skb->head : NULL,
|
H A D | associola.c | 1311 struct list_head *head = &asoc->peer.transport_addr_list; local 1323 /* Skip the head. */ 1324 if (pos->next == head) 1325 pos = head->next;
|
/net/batman-adv/ |
H A D | routing.c | 41 struct hlist_head *head; local 49 head = &hash->table[i]; 52 hlist_for_each_entry_rcu(orig_node, head, hash_entry) {
|
/net/bluetooth/rfcomm/ |
H A D | sock.c | 111 sk_for_each(sk, &rfcomm_sk_list.head) { 129 sk_for_each(sk, &rfcomm_sk_list.head) { 975 sk_for_each(sk, &rfcomm_sk_list.head) {
|
/net/openvswitch/ |
H A D | datapath.c | 184 struct hlist_head *head; local 186 head = vport_hash_bucket(dp, port_no); 187 hlist_for_each_entry_rcu(vport, head, dp_hash_node) { 202 struct hlist_head *head = vport_hash_bucket(dp, vport->port_no); local 204 hlist_add_head_rcu(&vport->dp_hash_node, head);
|
/net/ipv4/ |
H A D | tcp.c | 1141 /* We have some space in skb head. Superb! */ 3017 struct sk_buff **tcp_gro_receive(struct sk_buff **head, struct sk_buff *skb) argument 3057 for (; (p = *head); head = &p->next) { 3088 if (flush || skb_gro_receive(head, skb)) { 3093 p = *head; 3104 pp = head; 3117 skb->csum_start = skb_transport_header(skb) - skb->head;
|
H A D | tcp_input.c | 852 * A'. Reno "three dupacks" marks head of queue lost. 853 * A''. Its FACK modification, head until snd.fack is lost. 886 * in case of head skb. Due to potentional receiver driven attacks, we 888 * reneging and defer head skb's loss recovery to standard loss recovery 1288 /* Shifting pages past head area doesn't work */ 1355 /* TODO: head merge to next could be attempted here 1654 /* Skip empty blocks in at head of the cache */ 1730 /* Clear the head of the cache sack blocks so we can skip it next time */ 2091 * retransmit can be triggered by timeout of queue head. 2170 /* Detect loss in event "A" above by marking head o 4440 tcp_collapse(struct sock *sk, struct sk_buff_head *list, struct sk_buff *head, struct sk_buff *tail, u32 start, u32 end) argument 4549 struct sk_buff *head; local [all...] |
H A D | ip_vti.c | 631 static void vti_destroy_tunnels(struct vti_net *ipn, struct list_head *head) argument 642 unregister_netdevice_queue(t->dev, head);
|
H A D | af_inet.c | 1367 static struct sk_buff **inet_gro_receive(struct sk_buff **head, argument 1406 for (p = *head; p; p = p->next) { 1434 pp = ops->callbacks.gro_receive(head, skb);
|
/net/bluetooth/ |
H A D | l2cap_core.c | 265 static struct sk_buff *l2cap_ertm_seq_in_queue(struct sk_buff_head *head, argument 270 skb_queue_walk(head, skb) { 285 * and removed from the head in constant time, without further memory 304 seq_list->head = L2CAP_SEQ_LIST_CLEAR; 328 if (seq_list->head == L2CAP_SEQ_LIST_CLEAR) { 329 /* In case someone tries to pop the head of an empty list */ 331 } else if (seq_list->head == seq) { 333 seq_list->head = seq_list->list[seq & mask]; 336 if (seq_list->head == L2CAP_SEQ_LIST_TAIL) { 337 seq_list->head [all...] |
/net/caif/ |
H A D | cfpkt_skbuff.c | 26 struct sk_buff_head head; member in struct:cfpktq
|
/net/ipv6/ |
H A D | sit.c | 395 static void prl_list_destroy_rcu(struct rcu_head *head) argument 399 p = container_of(head, struct ip_tunnel_prl_entry, rcu_head); 1474 static void __net_exit sit_destroy_tunnels(struct sit_net *sitn, struct list_head *head) argument 1485 unregister_netdevice_queue(t->dev, head);
|
H A D | udp.c | 213 udp_portaddr_for_each_entry_rcu(sk, node, &hslot2->head) { 295 sk_nulls_for_each_rcu(sk, node, &hslot->head) { 754 sk = sk_nulls_head(&hslot->head); 923 skb->csum_start = skb_transport_header(skb) - skb->head;
|
H A D | addrconf.c | 619 struct hlist_head *head; local 626 head = &net->dev_index_head[h]; 630 hlist_for_each_entry_rcu(dev, head, index_hlist) { 4063 struct hlist_head *head; local 4073 head = &net->dev_index_head[h]; 4074 hlist_for_each_entry_rcu(dev, head, index_hlist) { 4502 struct hlist_head *head; local 4510 head = &net->dev_index_head[h]; 4511 hlist_for_each_entry_rcu(dev, head, index_hlist) {
|
/net/mac80211/ |
H A D | cfg.c | 853 /* Need to have a beacon head if we don't have one yet */ 854 if (!params->head && !old) 857 /* new or old head? */ 858 if (params->head) 880 * memory is | beacon_data | head | tail | 882 new->head = ((u8 *) new) + sizeof(*new); 883 new->tail = new->head + new_head_len; 887 /* copy in head */ 888 if (params->head) 889 memcpy(new->head, param [all...] |
/net/packet/ |
H A D | af_packet.c | 419 return packet_lookup_frame(po, rb, rb->head, status); 1005 po->rx_ring.head, status); 1077 unsigned int previous = rb->head ? rb->head - 1 : rb->frame_max; 1083 buff->head = buff->head != buff->frame_max ? buff->head+1 : 0; 1102 po->rx_ring.head, 1620 * structure, so that corresponding packet head is 1626 /* Special case: outgoing packets have ll header at head */ [all...] |
/net/sched/ |
H A D | sch_hfsc.c | 1152 struct hfsc_class *head, *cl; local 1163 head = &q->root; 1180 if (cl->level >= head->level) 1189 head = cl;
|
/net/sunrpc/ |
H A D | clnt.c | 915 xbufp->len = xbufp->head[0].iov_len + xbufp->page_len + 1493 buf->head[0].iov_base = start; 1494 buf->head[0].iov_len = len; 2059 struct kvec *iov = &task->tk_rqstp->rq_rcv_buf.head[0];
|
/net/core/ |
H A D | fib_rules.c | 162 static void fib_rules_put_rcu(struct rcu_head *head) argument 164 struct fib_rules_ops *ops = container_of(head, struct fib_rules_ops, rcu);
|
/net/netfilter/ |
H A D | xt_hashlimit.c | 202 static void dsthash_free_rcu(struct rcu_head *head) argument 204 struct dsthash_ent *ent = container_of(head, struct dsthash_ent, rcu);
|
/net/atm/ |
H A D | lec.c | 222 pr_debug("skbuff head:%lx data:%lx tail:%lx end:%lx\n", 223 (long)skb->head, (long)skb->data, (long)skb_tail_pointer(skb), 1545 struct hlist_head *head; local 1550 head = &priv->lec_arp_tables[HASH(mac_addr[ETH_ALEN - 1])]; 1551 hlist_for_each_entry(entry, head, next) {
|
/net/bridge/ |
H A D | br_private.h | 469 extern void br_multicast_free_pg(struct rcu_head *head);
|
/net/bridge/netfilter/ |
H A D | ebtables.c | 322 find_inlist_lock_noload(struct list_head *head, const char *name, int *error, argument 334 list_for_each_entry(e, head, list) { 344 find_inlist_lock(struct list_head *head, const char *name, const char *prefix, argument 348 find_inlist_lock_noload(head, name, error, mutex),
|