/net/netlabel/ |
H A D | netlabel_addrlist.h | 97 #define netlbl_af4list_foreach(iter, head) \ 98 for (iter = __af4list_valid((head)->next, head); \ 99 &iter->list != (head); \ 100 iter = __af4list_valid(iter->list.next, head)) 102 #define netlbl_af4list_foreach_rcu(iter, head) \ 103 for (iter = __af4list_valid_rcu((head)->next, head); \ 104 &iter->list != (head); \ 105 iter = __af4list_valid_rcu(iter->list.next, head)) [all...] |
H A D | netlabel_addrlist.c | 53 * @head: the list head 56 * Searches the IPv4 address list given by @head. If a matching address entry 62 struct list_head *head) 66 list_for_each_entry_rcu(iter, head, list) 77 * @head: the list head 80 * Searches the IPv4 address list given by @head. If an exact match if found 87 struct list_head *head) 91 list_for_each_entry_rcu(iter, head, lis 61 netlbl_af4list_search(__be32 addr, struct list_head *head) argument 85 netlbl_af4list_search_exact(__be32 addr, __be32 mask, struct list_head *head) argument 111 netlbl_af6list_search(const struct in6_addr *addr, struct list_head *head) argument 136 netlbl_af6list_search_exact(const struct in6_addr *addr, const struct in6_addr *mask, struct list_head *head) argument 163 netlbl_af4list_add(struct netlbl_af4list *entry, struct list_head *head) argument 200 netlbl_af6list_add(struct netlbl_af6list *entry, struct list_head *head) argument 254 netlbl_af4list_remove(__be32 addr, __be32 mask, struct list_head *head) argument 294 netlbl_af6list_remove(const struct in6_addr *addr, const struct in6_addr *mask, struct list_head *head) argument [all...] |
/net/sched/ |
H A D | cls_cgroup.c | 123 struct cls_cgroup_head *head = tp->root; local 150 if (!tcf_em_tree_match(skb, &head->ematches, NULL)) 155 return tcf_exts_exec(skb, &head->exts, res); 187 struct cls_cgroup_head *head = tp->root; local 195 if (head == NULL) { 199 head = kzalloc(sizeof(*head), GFP_KERNEL); 200 if (head == NULL) 203 head->handle = handle; 206 tp->root = head; 235 struct cls_cgroup_head *head = tp->root; local 251 struct cls_cgroup_head *head = tp->root; local 267 struct cls_cgroup_head *head = tp->root; local [all...] |
H A D | cls_basic.c | 46 struct basic_head *head = (struct basic_head *) tp->root; local 49 list_for_each_entry(f, &head->flist, link) { 64 struct basic_head *head = (struct basic_head *) tp->root; local 67 if (head == NULL) 70 list_for_each_entry(f, &head->flist, link) 83 struct basic_head *head; local 85 head = kzalloc(sizeof(*head), GFP_KERNEL); 86 if (head == NULL) 88 INIT_LIST_HEAD(&head 103 struct basic_head *head = tp->root; local 115 struct basic_head *head = (struct basic_head *) tp->root; local 171 struct basic_head *head = (struct basic_head *) tp->root; local 231 struct basic_head *head = (struct basic_head *) tp->root; local [all...] |
H A D | cls_fw.c | 83 struct fw_head *head = (struct fw_head *)tp->root; local 88 if (head != NULL) { 89 id &= head->mask; 90 for (f = head->ht[fw_hash(id)]; f; f = f->next) { 119 struct fw_head *head = (struct fw_head *)tp->root; local 122 if (head == NULL) 125 for (f = head->ht[fw_hash(handle)]; f; f = f->next) { 150 struct fw_head *head = tp->root; local 154 if (head == NULL) 158 while ((f = head 168 struct fw_head *head = (struct fw_head *)tp->root; local 198 struct fw_head *head = (struct fw_head *)tp->root; local 242 struct fw_head *head = (struct fw_head *)tp->root; local 304 struct fw_head *head = (struct fw_head *)tp->root; local 333 struct fw_head *head = (struct fw_head *)tp->root; local [all...] |
H A D | cls_route.c | 73 route4_reset_fastmap(struct Qdisc *q, struct route4_head *head, u32 id) argument 78 memset(head->fastmap, 0, sizeof(head->fastmap)); 83 route4_set_fastmap(struct route4_head *head, u32 id, int iif, argument 88 head->fastmap[h].id = id; 89 head->fastmap[h].iif = iif; 90 head->fastmap[h].filter = f; 124 route4_set_fastmap(head, id, iif, f); \ 131 struct route4_head *head = (struct route4_head *)tp->root; local 143 if (head 221 struct route4_head *head = (struct route4_head *)tp->root; local 265 struct route4_head *head = tp->root; local 292 struct route4_head *head = (struct route4_head *)tp->root; local 338 route4_set_parms(struct net *net, struct tcf_proto *tp, unsigned long base, struct route4_filter *f, u32 handle, struct route4_head *head, struct nlattr **tb, struct nlattr *est, int new) argument 437 struct route4_head *head = tp->root; local 525 struct route4_head *head = tp->root; local [all...] |
H A D | sch_choke.c | 72 unsigned int head; member in struct:choke_sched_data 89 return (q->tail - q->head) & q->tab_mask; 104 /* Move head pointer forward to skip over holes */ 108 q->head = (q->head + 1) & q->tab_mask; 109 if (q->head == q->tail) 111 } while (q->tab[q->head] == NULL); 119 if (q->head == q->tail) 132 if (idx == q->head) 226 * times to find a random skb but then just give up and return the head [all...] |
/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_recvfrom.c | 70 /* Set up the XDR head */ 71 rqstp->rq_arg.head[0].iov_base = page_address(page); 72 rqstp->rq_arg.head[0].iov_len = min(byte_count, ctxt->sge[0].length); 76 /* Compute bytes past head in the SGL */ 77 bc = byte_count - rqstp->rq_arg.head[0].iov_len; 98 BUG_ON((rqstp->rq_arg.head[0].iov_len + rqstp->rq_arg.page_len) 130 struct svc_rdma_op_ctxt *head, 151 head->arg.head[0] = rqstp->rq_arg.head[ 128 map_read_chunks(struct svcxprt_rdma *xprt, struct svc_rqst *rqstp, struct svc_rdma_op_ctxt *head, struct rpcrdma_msg *rmsgp, struct svc_rdma_req_map *rpl_map, struct svc_rdma_req_map *chl_map, int ch_count, int byte_count) argument 228 fast_reg_read_chunks(struct svcxprt_rdma *xprt, struct svc_rqst *rqstp, struct svc_rdma_op_ctxt *head, struct rpcrdma_msg *rmsgp, struct svc_rdma_req_map *rpl_map, struct svc_rdma_req_map *chl_map, int ch_count, int byte_count) argument 533 rdma_read_complete(struct svc_rqst *rqstp, struct svc_rdma_op_ctxt *head) argument [all...] |
/net/ipv6/netfilter/ |
H A D | nf_conntrack_reasm.c | 364 struct sk_buff *fp, *op, *head = fq->q.fragments; local 370 WARN_ON(head == NULL); 371 WARN_ON(NFCT_FRAG6_CB(head)->offset != 0); 378 payload_len = ((head->data - skb_network_header(head)) - 387 if (skb_unclone(head, GFP_ATOMIC)) { 388 pr_debug("skb is cloned but can't expand head"); 395 if (skb_has_frag_list(head)) { 403 clone->next = head->next; 404 head [all...] |
/net/batman-adv/ |
H A D | hash.h | 62 struct hlist_head *head; local 68 head = &hash->table[i]; 72 hlist_for_each_safe(node, node_tmp, head) { 125 struct hlist_head *head; local 133 head = &hash->table[index]; 138 hlist_for_each(node, head) { 147 hlist_add_head_rcu(data_node, head); 169 struct hlist_head *head; local 173 head = &hash->table[index]; 176 hlist_for_each(node, head) { [all...] |
H A D | originator.h | 71 struct hlist_head *head; local 79 head = &hash->table[index]; 82 hlist_for_each_entry_rcu(orig_node, head, hash_entry) {
|
H A D | unicast.c | 33 batadv_frag_merge_packet(struct list_head *head, argument 63 list_move_tail(&tfp->list, head); 81 static void batadv_frag_create_entry(struct list_head *head, argument 90 tfp = list_entry((head)->prev, typeof(*tfp), list); 95 list_move(&tfp->list, head); 99 static int batadv_frag_create_buffer(struct list_head *head) argument 107 batadv_frag_list_free(head); 113 list_add(&tfp->list, head); 120 batadv_frag_search_packet(struct list_head *head, argument 135 list_for_each_entry(tfp, head, lis 159 batadv_frag_list_free(struct list_head *head) argument [all...] |
/net/atm/ |
H A D | addr.c | 51 struct list_head *head; local 55 head = &dev->lecs; 57 head = &dev->local; 58 list_for_each_entry_safe(this, p, head, entry) { 63 if (head == &dev->local) 72 struct list_head *head; local 80 head = &dev->lecs; 82 head = &dev->local; 83 list_for_each_entry(this, head, entry) { 95 list_add(&this->entry, head); 107 struct list_head *head; local 137 struct list_head *head; local [all...] |
/net/dccp/ccids/lib/ |
H A D | packet_history.h | 56 tfrc_tx_hist_find_entry(struct tfrc_tx_hist_entry *head, u64 seqno) argument 58 while (head != NULL && head->seqno != seqno) 59 head = head->next; 60 return head;
|
/net/ipv6/ |
H A D | reassembly.c | 381 struct sk_buff *fp, *head = fq->q.fragments; local 393 /* Make the one we just received the head. */ 395 head = prev->next; 396 fp = skb_clone(head, GFP_ATOMIC); 401 fp->next = head->next; 406 skb_morph(head, fq->q.fragments); 407 head->next = fq->q.fragments->next; 410 fq->q.fragments = head; 413 WARN_ON(head == NULL); 414 WARN_ON(FRAG6_CB(head) [all...] |
/net/ceph/ |
H A D | pagelist.c | 10 struct page *page = list_entry(pl->head.prev, struct page, lru); 19 while (!list_empty(&pl->head)) { 20 struct page *page = list_first_entry(&pl->head, struct page, 45 list_add_tail(&page->lru, &pl->head); 116 c->page_lru = pl->head.prev; 134 while (pl->head.prev != c->page_lru) { 135 page = list_entry(pl->head.prev, struct page, lru); 141 if (!list_empty(&pl->head)) { 142 page = list_entry(pl->head.prev, struct page, lru);
|
/net/ipv4/ |
H A D | inet_hashtables.c | 33 struct inet_bind_hashbucket *head, 45 hlist_add_head(&tb->node, &head->chain); 83 struct inet_bind_hashbucket *head = &hashinfo->bhash[bhash]; local 88 spin_lock(&head->lock); 95 spin_unlock(&head->lock); 112 struct inet_bind_hashbucket *head = &table->bhash[bhash]; local 115 spin_lock(&head->lock); 123 inet_bind_bucket_for_each(tb, &head->chain) { 130 sock_net(sk), head, port); 132 spin_unlock(&head 31 inet_bind_bucket_create(struct kmem_cache *cachep, struct net *net, struct inet_bind_hashbucket *head, const unsigned short snum) argument 248 struct inet_ehash_bucket *head = &hashinfo->ehash[slot]; local 325 struct inet_ehash_bucket *head = inet_ehash_bucket(hinfo, hash); local 404 struct inet_ehash_bucket *head; local 485 struct inet_bind_hashbucket *head; local [all...] |
H A D | ip_fragment.c | 213 struct sk_buff *head = qp->q.fragments; local 218 head->dev = dev_get_by_index_rcu(net, qp->iif); 219 if (!head->dev) 223 iph = ip_hdr(head); 224 err = ip_route_input_noref(head, iph->daddr, iph->saddr, 225 iph->tos, head->dev); 235 skb_rtable(head)->rt_type != RTN_LOCAL)) 240 icmp_send(head, ICMP_TIME_EXCEEDED, ICMP_EXC_FRAGTIME, 0); 439 /* Eat head of the next overlapped fragment 522 struct sk_buff *fp, *head local [all...] |
/net/ipv4/netfilter/ |
H A D | nf_conntrack_l3proto_ipv4_compat.c | 50 struct hlist_nulls_node *head) 55 head = rcu_dereference(hlist_nulls_next_rcu(head)); 56 while (is_a_nulls(head)) { 57 if (likely(get_nulls_value(head) == st->bucket)) { 61 head = rcu_dereference( 64 return head; 69 struct hlist_nulls_node *head = ct_get_first(seq); local 71 if (head) 72 while (pos && (head 49 ct_get_next(struct seq_file *seq, struct hlist_nulls_node *head) argument 233 ct_expect_get_next(struct seq_file *seq, struct hlist_node *head) argument 251 struct hlist_node *head = ct_expect_get_first(seq); local [all...] |
/net/sunrpc/ |
H A D | xdr.c | 135 struct kvec *head = xdr->head; local 137 char *buf = (char *)head->iov_base; 138 unsigned int buflen = head->iov_len; 140 head->iov_len = offset; 306 * @len: bytes to remove from buf->head[0] 308 * Shrinks XDR buffer's header kvec buf->head[0] by 315 struct kvec *head, *tail; local 320 head = buf->head; [all...] |
/net/netfilter/ |
H A D | nf_internals.h | 16 extern unsigned int nf_iterate(struct list_head *head,
|
H A D | nfnetlink_acct.c | 35 struct list_head head; member in struct:nf_acct 55 list_for_each_entry(nfacct, &nfnl_acct_list, head) { 91 list_add_tail_rcu(&nfacct->head, &nfnl_acct_list); 151 list_for_each_entry_rcu(cur, &nfnl_acct_list, head) { 191 list_for_each_entry(cur, &nfnl_acct_list, head) { 230 list_del_rcu(&cur->head); 249 list_for_each_entry(cur, &nfnl_acct_list, head) 256 list_for_each_entry(cur, &nfnl_acct_list, head) { 304 list_for_each_entry_rcu(cur, &nfnl_acct_list, head) { 361 list_for_each_entry_safe(cur, tmp, &nfnl_acct_list, head) { [all...] |
/net/sctp/ |
H A D | proc.c | 212 struct sctp_hashbucket *head; local 221 head = &sctp_ep_hashtable[hash]; 223 read_lock(&head->lock); 224 sctp_for_each_hentry(epb, &head->chain) { 238 read_unlock(&head->lock); 320 struct sctp_hashbucket *head; local 329 head = &sctp_assoc_hashtable[hash]; 331 read_lock(&head->lock); 332 sctp_for_each_hentry(epb, &head->chain) { 365 read_unlock(&head 440 struct sctp_hashbucket *head; local [all...] |
/net/sunrpc/auth_gss/ |
H A D | gss_krb5_wrap.c | 58 iov = &buf->head[0]; 72 if (len <= buf->head[0].iov_len) { 73 pad = *(u8 *)(buf->head[0].iov_base + len - 1); 74 if (pad > buf->head[0].iov_len) 76 buf->head[0].iov_len -= pad; 79 len -= buf->head[0].iov_len; 97 * However adjusting the head length, as we do above, is harmless. 99 * also uses length and head length together to determine the original 101 * easier on the server if we adjust head and tail length in tandem. 149 /* Assumptions: the head an 397 char head[LOCAL_BUF_LEN]; local [all...] |
/net/rds/ |
H A D | connection.c | 67 static struct rds_connection *rds_conn_lookup(struct hlist_head *head, argument 73 hlist_for_each_entry_rcu(conn, head, c_hash_node) { 119 struct hlist_head *head = rds_conn_bucket(laddr, faddr); local 125 conn = rds_conn_lookup(head, laddr, faddr, trans); 226 found = rds_conn_lookup(head, laddr, faddr, trans); 232 hlist_add_head_rcu(&conn->c_hash_node, head); 377 struct hlist_head *head; local 389 for (i = 0, head = rds_conn_hash; i < ARRAY_SIZE(rds_conn_hash); 390 i++, head++) { 391 hlist_for_each_entry_rcu(conn, head, c_hash_nod 439 struct hlist_head *head; local [all...] |