/net/bridge/ |
H A D | br_forward.c | 24 static int deliver_clone(const struct net_bridge_port *prev, 132 static int deliver_clone(const struct net_bridge_port *prev, argument 145 __packet_hook(prev, skb); 150 struct net_bridge_port *prev, struct net_bridge_port *p, 158 return prev; 160 if (!prev) 163 err = deliver_clone(prev, skb, __packet_hook); 179 struct net_bridge_port *prev; local 181 prev = NULL; 187 prev 149 maybe_deliver( struct net_bridge_port *prev, struct net_bridge_port *p, struct sk_buff *skb, void (*__packet_hook)(const struct net_bridge_port *p, struct sk_buff *skb)) argument 230 struct net_bridge_port *prev = NULL; local [all...] |
/net/netfilter/ |
H A D | xt_limit.c | 21 unsigned long prev; member in struct:xt_limit_priv 43 `prev' tracks the last packet hit: you gain one credit per jiffy. 76 priv->credit += (now - xchg(&priv->prev, now)) * CREDITS_PER_JIFFY; 123 priv->prev = jiffies; 144 compat_ulong_t prev; member in struct:compat_xt_rateinfo 151 /* To keep the full "prev" timestamp, the upper 32 bits are stored in the 159 .prev = cm->prev | (unsigned long)cm->master << 32, 173 .prev = m->prev, [all...] |
/net/ceph/ |
H A D | pagelist.c | 11 struct page *page = list_entry(pl->head.prev, struct page, lru); 119 c->page_lru = pl->head.prev; 137 while (pl->head.prev != c->page_lru) { 138 page = list_entry(pl->head.prev, struct page, lru); 145 page = list_entry(pl->head.prev, struct page, lru);
|
/net/bridge/netfilter/ |
H A D | ebt_limit.c | 41 info->credit += (now - xchg(&info->prev, now)) * CREDITS_PER_JIFFY; 81 info->prev = jiffies; 96 compat_ulong_t prev; member in struct:ebt_compat_limit_info
|
/net/core/ |
H A D | dst.c | 300 unsigned long prev, new; local 305 prev = cmpxchg(&dst->_metrics, old, new); 307 if (prev != old) { 309 p = __DST_METRICS_PTR(prev); 310 if (prev & DST_METRICS_READ_ONLY) 321 unsigned long prev, new; local 324 prev = cmpxchg(&dst->_metrics, old, new); 325 if (prev == old)
|
/net/ieee802154/ |
H A D | reassembly.c | 49 struct sk_buff *prev, struct net_device *dev); 143 struct sk_buff *prev, *next; local 176 prev = fq->q.fragments_tail; 177 if (!prev || lowpan_cb(prev)->d_offset < lowpan_cb(skb)->d_offset) { 181 prev = NULL; 185 prev = next; 193 if (prev) 194 prev->next = skb; 218 res = lowpan_frag_reasm(fq, prev, de 237 lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *prev, struct net_device *dev) argument [all...] |
H A D | dgram.c | 384 struct sock *sk, *prev = NULL; local 400 if (prev) { 405 dgram_rcv_skb(prev, clone); 408 prev = sk; 412 if (prev) { 413 dgram_rcv_skb(prev, skb);
|
/net/ipv4/ |
H A D | ip_fragment.c | 95 static int ip_frag_reasm(struct ipq *qp, struct sk_buff *prev, 320 struct sk_buff *prev, *next; local 386 prev = qp->q.fragments_tail; 387 if (!prev || FRAG_CB(prev)->offset < offset) { 391 prev = NULL; 395 prev = next; 403 if (prev) { 404 int i = (FRAG_CB(prev)->offset + prev 502 ip_frag_reasm(struct ipq *qp, struct sk_buff *prev, struct net_device *dev) argument [all...] |
/net/atm/ |
H A D | mpoa_caches.c | 112 entry->prev = NULL; 114 client->in_cache->prev = entry; 200 if (entry->prev != NULL) 201 entry->prev->next = entry->next; 205 entry->next->prev = entry->prev; 435 if (entry->prev != NULL) 436 entry->prev->next = entry->next; 440 entry->next->prev = entry->prev; [all...] |
H A D | mpoa_caches.h | 16 struct in_cache_entry *prev; member in struct:in_cache_entry 53 struct eg_cache_entry *prev; member in struct:eg_cache_entry
|
/net/ipv6/netfilter/ |
H A D | nf_conntrack_reasm.c | 221 struct sk_buff *prev, *next; local 301 prev = fq->q.fragments_tail; 302 if (!prev || NFCT_FRAG6_CB(prev)->offset < offset) { 306 prev = NULL; 310 prev = next; 323 if (prev && 324 (NFCT_FRAG6_CB(prev)->offset + prev->len) > offset) 337 if (prev) [all...] |
/net/ipv6/ |
H A D | reassembly.c | 79 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *prev, 208 struct sk_buff *prev, *next; local 285 prev = fq->q.fragments_tail; 286 if (!prev || FRAG6_CB(prev)->offset < offset) { 290 prev = NULL; 294 prev = next; 306 if (prev && 307 (FRAG6_CB(prev)->offset + prev 375 ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *prev, struct net_device *dev) argument [all...] |
H A D | inet6_connection_sock.c | 125 struct request_sock *req, **prev; local 127 for (prev = &lopt->syn_table[inet6_synq_hash(raddr, rport, 130 (req = *prev) != NULL; 131 prev = &req->dl_next) { 140 *prevp = prev;
|
/net/rds/ |
H A D | ib_send.c | 532 struct rds_ib_send_work *prev; local 660 prev = NULL; 721 if (prev) 722 prev->s_wr.next = &send->s_wr; 723 prev = send; 739 prev->s_op = ic->i_data_op; 740 prev->s_wr.send_flags |= IB_SEND_SOLICITED; 766 if (prev->s_op) { 767 ic->i_data_op = prev->s_op; 768 prev 880 struct rds_ib_send_work *prev; local [all...] |
H A D | iw_send.c | 503 struct rds_iw_send_work *prev; local 624 prev = NULL; 719 if (prev) 720 prev->s_wr.next = &send->s_wr; 721 prev = send; 733 prev->s_rm = ic->i_rm; 734 prev->s_wr.send_flags |= IB_SEND_SIGNALED | IB_SEND_SOLICITED; 755 if (prev->s_rm) { 756 ic->i_rm = prev->s_rm; 757 prev 793 struct rds_iw_send_work *prev; local [all...] |
/net/netlabel/ |
H A D | netlabel_kapi.c | 432 struct netlbl_lsm_catmap *prev = NULL; local 439 prev = iter; 459 if (prev == NULL) { 463 iter->next = prev->next; 464 prev->next = iter; 537 struct netlbl_lsm_catmap *prev = NULL; local 563 if (prev && idx == 0 && bit == 0) 564 return prev->startbit + NETLBL_CATMAP_SIZE - 1; 571 prev = iter;
|
H A D | netlabel_addrlist.c | 179 iter->list.prev, 217 iter->list.prev,
|
/net/sched/ |
H A D | sch_teql.c | 144 struct Qdisc *q, *prev; local 148 prev = master->slaves; 149 if (prev) { 151 q = NEXT_SLAVE(prev); 153 NEXT_SLAVE(prev) = NEXT_SLAVE(q); 173 } while ((prev = q) != master->slaves);
|
H A D | sch_sfq.c | 101 sfq_index prev; member in struct:sfq_head 235 slot->dep.prev = p; 238 sfq_dep_head(q, n)->prev = x; 244 p = q->slots[x].dep.prev; \ 246 sfq_dep_head(q, n)->prev = p; \ 283 slot->skblist_prev = skb->prev; 284 skb->prev->next = (struct sk_buff *)slot; 285 skb->next = skb->prev = NULL; 295 skb->next->prev = (struct sk_buff *)slot; 296 skb->next = skb->prev [all...] |
H A D | sch_hhf.c | 207 u32 prev = flow->hit_timestamp + q->hhf_evict_timeout; local 209 if (hhf_time_before(prev, now)) { 237 u32 prev = flow->hit_timestamp + q->hhf_evict_timeout; local 239 if (hhf_time_before(prev, now)) 271 u32 prev; local 275 prev = q->hhf_arrays_reset_timestamp + q->hhf_reset_timeout; 276 if (hhf_time_before(prev, now)) {
|
H A D | sch_fq.c | 321 struct sk_buff *prev, *head = flow->head; local 338 prev = NULL; 340 prev = head; 345 if (!prev) { /* no rtx packet in queue, become the new head */ 349 if (prev == flow->tail) 352 skb->next = prev->next; 353 prev->next = skb;
|
/net/netfilter/ipvs/ |
H A D | ip_vs_rr.c | 44 /* dest is already unlinked, so p->prev is not valid but 48 svc->sched_data = p->next->prev;
|
/net/packet/ |
H A D | internal.h | 35 char *prev; member in struct:tpacket_kbdq_core
|
/net/dccp/ |
H A D | minisocks.c | 156 struct request_sock **prev) 203 inet_csk_reqsk_queue_unlink(sk, req, prev); 215 inet_csk_reqsk_queue_drop(sk, req, prev); 154 dccp_check_req(struct sock *sk, struct sk_buff *skb, struct request_sock *req, struct request_sock **prev) argument
|
/net/rfkill/ |
H A D | core.c | 240 bool prev, any; local 245 prev = !!(rfkill->state & RFKILL_BLOCK_HW); 250 *change = prev != blocked; 271 bool prev, curr; local 286 prev = rfkill->state & RFKILL_BLOCK_SW; 322 if (prev != curr) 506 bool prev, hwblock; local 511 prev = !!(rfkill->state & RFKILL_BLOCK_SW); 520 if (prev != blocked && !hwblock) 553 * No need to care about prev/setbloc [all...] |