/net/netlabel/ |
H A D | netlabel_addrlist.h | 96 #define netlbl_af4list_foreach(iter, head) \ 97 for (iter = __af4list_valid((head)->next, head); \ 98 &iter->list != (head); \ 99 iter = __af4list_valid(iter->list.next, head)) 101 #define netlbl_af4list_foreach_rcu(iter, head) \ 102 for (iter = __af4list_valid_rcu((head)->next, head); \ 103 &iter->list != (head); \ 104 iter = __af4list_valid_rcu(iter->list.next, head)) [all...] |
H A D | netlabel_addrlist.c | 52 * @head: the list head 55 * Searches the IPv4 address list given by @head. If a matching address entry 61 struct list_head *head) 65 list_for_each_entry_rcu(iter, head, list) 76 * @head: the list head 79 * Searches the IPv4 address list given by @head. If an exact match if found 86 struct list_head *head) 90 list_for_each_entry_rcu(iter, head, lis 60 netlbl_af4list_search(__be32 addr, struct list_head *head) argument 84 netlbl_af4list_search_exact(__be32 addr, __be32 mask, struct list_head *head) argument 110 netlbl_af6list_search(const struct in6_addr *addr, struct list_head *head) argument 135 netlbl_af6list_search_exact(const struct in6_addr *addr, const struct in6_addr *mask, struct list_head *head) argument 162 netlbl_af4list_add(struct netlbl_af4list *entry, struct list_head *head) argument 199 netlbl_af6list_add(struct netlbl_af6list *entry, struct list_head *head) argument 253 netlbl_af4list_remove(__be32 addr, __be32 mask, struct list_head *head) argument 293 netlbl_af6list_remove(const struct in6_addr *addr, const struct in6_addr *mask, struct list_head *head) argument [all...] |
/net/sched/ |
H A D | cls_cgroup.c | 32 struct cls_cgroup_head *head = rcu_dereference_bh(tp->root); local 57 if (!tcf_em_tree_match(skb, &head->ematches, NULL)) 62 return tcf_exts_exec(skb, &head->exts, res); 85 struct cls_cgroup_head *head = container_of(root, local 89 tcf_exts_destroy(&head->exts); 90 tcf_em_tree_destroy(&head->ematches); 91 kfree(head); 100 struct cls_cgroup_head *head = rtnl_dereference(tp->root); local 109 if (!head && !handle) 112 if (head 156 struct cls_cgroup_head *head = rtnl_dereference(tp->root); local 171 struct cls_cgroup_head *head = rtnl_dereference(tp->root); local 187 struct cls_cgroup_head *head = rtnl_dereference(tp->root); local [all...] |
H A D | cls_basic.c | 44 struct basic_head *head = rcu_dereference_bh(tp->root); local 47 list_for_each_entry_rcu(f, &head->flist, link) { 62 struct basic_head *head = rtnl_dereference(tp->root); local 65 if (head == NULL) 68 list_for_each_entry(f, &head->flist, link) 81 struct basic_head *head; local 83 head = kzalloc(sizeof(*head), GFP_KERNEL); 84 if (head == NULL) 86 INIT_LIST_HEAD(&head 91 basic_delete_filter(struct rcu_head *head) argument 102 struct basic_head *head = rtnl_dereference(tp->root); local 116 struct basic_head *head = rtnl_dereference(tp->root); local 173 struct basic_head *head = rtnl_dereference(tp->root); local 239 struct basic_head *head = rtnl_dereference(tp->root); local [all...] |
H A D | cls_fw.c | 62 struct fw_head *head = rcu_dereference_bh(tp->root); local 67 if (head != NULL) { 68 id &= head->mask; 70 for (f = rcu_dereference_bh(head->ht[fw_hash(id)]); f; 100 struct fw_head *head = rtnl_dereference(tp->root); local 103 if (head == NULL) 106 f = rtnl_dereference(head->ht[fw_hash(handle)]); 123 static void fw_delete_filter(struct rcu_head *head) argument 125 struct fw_filter *f = container_of(head, struct fw_filter, rcu); 133 struct fw_head *head local 154 struct fw_head *head = rtnl_dereference(tp->root); local 187 struct fw_head *head = rtnl_dereference(tp->root); local 236 struct fw_head *head = rtnl_dereference(tp->root); local 331 struct fw_head *head = rtnl_dereference(tp->root); local 361 struct fw_head *head = rtnl_dereference(tp->root); local [all...] |
H A D | cls_bpf.c | 55 struct cls_bpf_head *head = rcu_dereference_bh(tp->root); local 59 list_for_each_entry_rcu(prog, &head->plist, link) { 81 struct cls_bpf_head *head; local 83 head = kzalloc(sizeof(*head), GFP_KERNEL); 84 if (head == NULL) 87 INIT_LIST_HEAD_RCU(&head->plist); 88 rcu_assign_pointer(tp->root, head); 112 struct cls_bpf_head *head = rtnl_dereference(tp->root); local 115 list_for_each_entry(prog, &head 129 struct cls_bpf_head *head = rtnl_dereference(tp->root); local 144 struct cls_bpf_head *head = rtnl_dereference(tp->root); local 225 cls_bpf_grab_new_handle(struct tcf_proto *tp, struct cls_bpf_head *head) argument 245 struct cls_bpf_head *head = rtnl_dereference(tp->root); local 344 struct cls_bpf_head *head = rtnl_dereference(tp->root); local [all...] |
H A D | cls_route.c | 72 route4_reset_fastmap(struct route4_head *head) argument 75 memset(head->fastmap, 0, sizeof(head->fastmap)); 80 route4_set_fastmap(struct route4_head *head, u32 id, int iif, argument 87 head->fastmap[h].id = id; 88 head->fastmap[h].iif = iif; 89 head->fastmap[h].filter = f; 124 route4_set_fastmap(head, id, iif, f); \ 131 struct route4_head *head = rcu_dereference_bh(tp->root); local 143 if (head 232 struct route4_head *head = rtnl_dereference(tp->root); local 269 route4_delete_filter(struct rcu_head *head) argument 279 struct route4_head *head = rtnl_dereference(tp->root); local 312 struct route4_head *head = rtnl_dereference(tp->root); local 369 route4_set_parms(struct net *net, struct tcf_proto *tp, unsigned long base, struct route4_filter *f, u32 handle, struct route4_head *head, struct nlattr **tb, struct nlattr *est, int new, bool ovr) argument 469 struct route4_head *head = rtnl_dereference(tp->root); local 562 struct route4_head *head = rtnl_dereference(tp->root); local [all...] |
H A D | sch_hhf.c | 121 struct sk_buff *head; member in struct:wdrr_bucket 197 struct list_head *head, 203 if (list_empty(head)) 206 list_for_each_entry_safe(flow, next, head, flowchain) { 213 if (list_is_last(&flow->flowchain, head)) 228 static struct hh_flow_state *alloc_new_hh(struct list_head *head, argument 234 if (!list_empty(head)) { 236 list_for_each_entry(flow, head, flowchain) { 255 list_add_tail(&flow->flowchain, head); 344 /* Removes one skb from head o 196 seek_list(const u32 hash, struct list_head *head, struct hhf_sched_data *q) argument 437 struct list_head *head; local 512 struct list_head *head = &q->hh_flows[i]; local [all...] |
H A D | sch_choke.c | 71 unsigned int head; member in struct:choke_sched_data 82 return (q->tail - q->head) & q->tab_mask; 97 /* Move head pointer forward to skip over holes */ 101 q->head = (q->head + 1) & q->tab_mask; 102 if (q->head == q->tail) 104 } while (q->tab[q->head] == NULL); 112 if (q->head == q->tail) 125 if (idx == q->head) 231 * times to find a random skb but then just give up and return the head [all...] |
/net/ipv6/netfilter/ |
H A D | nf_conntrack_reasm.c | 381 struct sk_buff *fp, *op, *head = fq->q.fragments; local 387 WARN_ON(head == NULL); 388 WARN_ON(NFCT_FRAG6_CB(head)->offset != 0); 395 payload_len = ((head->data - skb_network_header(head)) - 404 if (skb_unclone(head, GFP_ATOMIC)) { 405 pr_debug("skb is cloned but can't expand head"); 412 if (skb_has_frag_list(head)) { 420 clone->next = head->next; 421 head [all...] |
/net/batman-adv/ |
H A D | hash.h | 60 struct hlist_head *head; local 66 head = &hash->table[i]; 70 hlist_for_each_safe(node, node_tmp, head) { 123 struct hlist_head *head; local 131 head = &hash->table[index]; 136 hlist_for_each(node, head) { 145 hlist_add_head_rcu(data_node, head); 167 struct hlist_head *head; local 171 head = &hash->table[index]; 174 hlist_for_each(node, head) { [all...] |
H A D | fragmentation.h | 41 if (!hlist_empty(&frags_entry->head) &&
|
H A D | fragmentation.c | 29 * @head: head of chain with entries. 33 static void batadv_frag_clear_chain(struct hlist_head *head) argument 38 hlist_for_each_entry_safe(entry, node, head, list) { 61 batadv_frag_clear_chain(&orig_node->fragments[i].head); 103 if (!hlist_empty(&chain->head)) 104 batadv_frag_clear_chain(&chain->head); 116 * @chain_out: list head to attach complete chains of fragments to 122 * fragments needed to merge the packet, the chain is moved to the passed head 162 hlist_add_head(&frag_entry_new->list, &chain->head); 297 struct hlist_head head = HLIST_HEAD_INIT; local [all...] |
/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_recvfrom.c | 72 /* Set up the XDR head */ 73 rqstp->rq_arg.head[0].iov_base = page_address(page); 74 rqstp->rq_arg.head[0].iov_len = 79 /* Compute bytes past head in the SGL */ 80 bc = byte_count - rqstp->rq_arg.head[0].iov_len; 102 BUG_ON((rqstp->rq_arg.head[0].iov_len + rqstp->rq_arg.page_len) 130 struct svc_rdma_op_ctxt *head, 141 struct svc_rdma_op_ctxt *head, 157 ctxt->read_hdr = head; 165 head 139 rdma_read_chunk_lcl(struct svcxprt_rdma *xprt, struct svc_rqst *rqstp, struct svc_rdma_op_ctxt *head, int *page_no, u32 *page_offset, u32 rs_handle, u32 rs_length, u64 rs_offset, int last) argument 232 rdma_read_chunk_frmr(struct svcxprt_rdma *xprt, struct svc_rqst *rqstp, struct svc_rdma_op_ctxt *head, int *page_no, u32 *page_offset, u32 rs_handle, u32 rs_length, u64 rs_offset, int last) argument 368 rdma_read_chunks(struct svcxprt_rdma *xprt, struct rpcrdma_msg *rmsgp, struct svc_rqst *rqstp, struct svc_rdma_op_ctxt *head) argument 465 rdma_read_complete(struct svc_rqst *rqstp, struct svc_rdma_op_ctxt *head) argument [all...] |
/net/atm/ |
H A D | addr.c | 51 struct list_head *head; local 55 head = &dev->lecs; 57 head = &dev->local; 58 list_for_each_entry_safe(this, p, head, entry) { 63 if (head == &dev->local) 72 struct list_head *head; local 80 head = &dev->lecs; 82 head = &dev->local; 83 list_for_each_entry(this, head, entry) { 95 list_add(&this->entry, head); 107 struct list_head *head; local 137 struct list_head *head; local [all...] |
/net/dccp/ccids/lib/ |
H A D | packet_history.h | 56 tfrc_tx_hist_find_entry(struct tfrc_tx_hist_entry *head, u64 seqno) argument 58 while (head != NULL && head->seqno != seqno) 59 head = head->next; 60 return head;
|
/net/ipv6/ |
H A D | reassembly.c | 379 struct sk_buff *fp, *head = fq->q.fragments; local 391 /* Make the one we just received the head. */ 393 head = prev->next; 394 fp = skb_clone(head, GFP_ATOMIC); 399 fp->next = head->next; 404 skb_morph(head, fq->q.fragments); 405 head->next = fq->q.fragments->next; 408 fq->q.fragments = head; 411 WARN_ON(head == NULL); 412 WARN_ON(FRAG6_CB(head) [all...] |
/net/ceph/ |
H A D | pagelist.c | 11 struct page *page = list_entry(pl->head.prev, struct page, lru); 22 while (!list_empty(&pl->head)) { 23 struct page *page = list_first_entry(&pl->head, struct page, 48 list_add_tail(&page->lru, &pl->head); 119 c->page_lru = pl->head.prev; 137 while (pl->head.prev != c->page_lru) { 138 page = list_entry(pl->head.prev, struct page, lru); 144 if (!list_empty(&pl->head)) { 145 page = list_entry(pl->head.prev, struct page, lru);
|
/net/ipv4/ |
H A D | inet_hashtables.c | 58 struct inet_bind_hashbucket *head, 70 hlist_add_head(&tb->node, &head->chain); 108 struct inet_bind_hashbucket *head = &hashinfo->bhash[bhash]; local 113 spin_lock(&head->lock); 120 spin_unlock(&head->lock); 137 struct inet_bind_hashbucket *head = &table->bhash[bhash]; local 140 spin_lock(&head->lock); 148 inet_bind_bucket_for_each(tb, &head->chain) { 155 sock_net(sk), head, port); 157 spin_unlock(&head 56 inet_bind_bucket_create(struct kmem_cache *cachep, struct net *net, struct inet_bind_hashbucket *head, const unsigned short snum) argument 286 struct inet_ehash_bucket *head = &hashinfo->ehash[slot]; local 335 struct inet_ehash_bucket *head = inet_ehash_bucket(hinfo, hash); local 404 struct inet_ehash_bucket *head; local 485 struct inet_bind_hashbucket *head; local [all...] |
H A D | ip_fragment.c | 196 struct sk_buff *head = qp->q.fragments; local 206 head->dev = dev_get_by_index_rcu(net, qp->iif); 207 if (!head->dev) 211 iph = ip_hdr(head); 212 err = ip_route_input_noref(head, iph->daddr, iph->saddr, 213 iph->tos, head->dev); 223 (skb_rtable(head)->rt_type != RTN_LOCAL))) 227 icmp_send(head, ICMP_TIME_EXCEEDED, ICMP_EXC_FRAGTIME, 0); 425 /* Eat head of the next overlapped fragment 507 struct sk_buff *fp, *head local [all...] |
/net/ipv4/netfilter/ |
H A D | nf_conntrack_l3proto_ipv4_compat.c | 50 struct hlist_nulls_node *head) 55 head = rcu_dereference(hlist_nulls_next_rcu(head)); 56 while (is_a_nulls(head)) { 57 if (likely(get_nulls_value(head) == st->bucket)) { 61 head = rcu_dereference( 64 return head; 69 struct hlist_nulls_node *head = ct_get_first(seq); local 71 if (head) 72 while (pos && (head 49 ct_get_next(struct seq_file *seq, struct hlist_nulls_node *head) argument 233 ct_expect_get_next(struct seq_file *seq, struct hlist_node *head) argument 251 struct hlist_node *head = ct_expect_get_first(seq); local [all...] |
/net/ieee802154/ |
H A D | reassembly.c | 240 struct sk_buff *fp, *head = fq->q.fragments; local 245 /* Make the one we just received the head. */ 247 head = prev->next; 248 fp = skb_clone(head, GFP_ATOMIC); 253 fp->next = head->next; 258 skb_morph(head, fq->q.fragments); 259 head->next = fq->q.fragments->next; 262 fq->q.fragments = head; 266 if (skb_unclone(head, GFP_ATOMIC)) 273 if (skb_has_frag_list(head)) { [all...] |
/net/sunrpc/ |
H A D | xdr.c | 135 struct kvec *head = xdr->head; local 137 char *buf = (char *)head->iov_base; 138 unsigned int buflen = head->iov_len; 140 head->iov_len = offset; 309 * @len: bytes to remove from buf->head[0] 311 * Shrinks XDR buffer's header kvec buf->head[0] by 318 struct kvec *head, *tail; local 323 head = buf->head; 606 struct kvec *head = buf->head; local [all...] |
/net/tipc/ |
H A D | msg.c | 94 * out: head buf after sucessful complete reassembly, otherwise NULL 99 struct sk_buff *head = *headbuf; local 116 if (unlikely(head)) 120 head = *headbuf = frag; 121 skb_frag_list_init(head); 122 TIPC_SKB_CB(head)->tail = NULL; 127 if (!head) 130 if (skb_try_coalesce(head, frag, &headstolen, &delta)) { 133 tail = TIPC_SKB_CB(head)->tail; 134 if (!skb_has_frag_list(head)) 436 struct sk_buff *head = NULL; local [all...] |
/net/netfilter/ |
H A D | nf_internals.h | 16 unsigned int nf_iterate(struct list_head *head, struct sk_buff *skb,
|