/net/core/ |
H A D | scm.c | 91 fpp = &fpl->fp[fpl->count]; 115 struct scm_fp_list *fpl = scm->fp; 119 scm->fp = NULL; 121 fput(fpl->fp[i]); 155 err=scm_fp_copy(cmsg, &p->fp); 197 if (p->fp && !p->fp->count) 199 kfree(p->fp); 200 p->fp = NULL; 255 int fdnum = scm->fp 256 struct file **fp = scm->fp->fp; local [all...] |
H A D | filter.c | 644 struct sk_filter *fp = container_of(rcu, struct sk_filter, rcu); local 646 bpf_jit_free(fp); 647 kfree(fp); 651 static int __sk_prepare_filter(struct sk_filter *fp) argument 655 fp->bpf_func = sk_run_filter; 657 err = sk_chk_filter(fp->insns, fp->len); 661 bpf_jit_compile(fp); 678 struct sk_filter *fp; local 686 fp 706 sk_unattached_filter_destroy(struct sk_filter *fp) argument 724 struct sk_filter *fp, *old_fp; local [all...] |
/net/ipv6/ |
H A D | exthdrs_core.c | 88 __be16 _frag_off, *fp; local 89 fp = skb_header_pointer(skb, 94 if (fp == NULL) 97 *frag_offp = *fp; 239 __be16 *fp; local 243 fp = skb_header_pointer(skb, 248 if (fp == NULL) 251 _frag_off = ntohs(*fp) & ~0x7;
|
H A D | reassembly.c | 381 struct sk_buff *fp, *head = fq->q.fragments; local 396 fp = skb_clone(head, GFP_ATOMIC); 398 if (!fp) 401 fp->next = head->next; 402 if (!fp->next) 403 fq->q.fragments_tail = fp; 404 prev->next = fp; 463 for (fp = head->next; fp;) { 466 struct sk_buff *next = fp [all...] |
/net/irda/ |
H A D | irlmp_frame.c | 96 __u8 *fp; local 104 fp = skb->data; 110 slsap_sel = fp[0] & LSAP_MASK; 111 dlsap_sel = fp[1]; 117 if ((fp[0] & CONTROL_BIT) && (fp[2] == CONNECT_CMD)) { 140 if (fp[0] & CONTROL_BIT) { 142 __func__, fp[2]); 152 if (fp[0] & CONTROL_BIT) { 153 switch (fp[ 206 __u8 *fp; local [all...] |
H A D | iriap.c | 449 __u8 *fp; local 457 fp = skb->data; 461 len = get_unaligned_be16(fp + n); 467 obj_id = get_unaligned_be16(fp + n); 470 type = fp[n++]; 475 memcpy(&tmp_cpu32, fp+n, 4); n += 4; 483 charset = fp[n++]; 510 value_len = fp[n++]; 515 fp[n + value_len] = 0x00; 516 IRDA_DEBUG(4, "Got string %s\n", fp 562 __u8 *fp; local 655 __u8 *fp; local 935 __u8 *fp; local [all...] |
/net/bridge/netfilter/ |
H A D | ebt_vlan.c | 53 const struct vlan_hdr *fp; local 56 fp = skb_header_pointer(skb, 0, sizeof(_frame), &_frame); 57 if (fp == NULL) 60 TCI = ntohs(fp->h_vlan_TCI); 61 encap = fp->h_vlan_encapsulated_proto;
|
/net/unix/ |
H A D | garbage.c | 125 void unix_inflight(struct file *fp) argument 127 struct sock *s = unix_get_socket(fp); 142 void unix_notinflight(struct file *fp) argument 144 struct sock *s = unix_get_socket(fp); 167 if (UNIXCB(skb).fp) { 172 int nfd = UNIXCB(skb).fp->count; 173 struct file **fp = UNIXCB(skb).fp->fp; local 179 struct sock *sk = unix_get_socket(*fp [all...] |
H A D | af_unix.c | 1332 scm->fp = UNIXCB(skb).fp; 1333 UNIXCB(skb).fp = NULL; 1335 for (i = scm->fp->count-1; i >= 0; i--) 1336 unix_notinflight(scm->fp->fp[i]); 1344 if (UNIXCB(skb).fp) 1361 for (i = scm->fp->count - 1; i >= 0; i--) { 1362 struct sock *sk = unix_get_socket(scm->fp->fp[ [all...] |
/net/decnet/ |
H A D | dn_table.c | 71 #define DN_FIB_SCAN(f, fp) \ 72 for( ; ((f) = *(fp)) != NULL; (fp) = &(f)->fn_next) 74 #define DN_FIB_SCAN_KEY(f, fp, key) \ 75 for( ; ((f) = *(fp)) != NULL && dn_key_eq((f)->fn_key, (key)); (fp) = &(f)->fn_next) 125 struct dn_fib_node *f, **fp, *next; local 131 for(fp = dn_chain_p(f->fn_key, dz); 132 *fp && dn_key_leq((*fp) 527 struct dn_fib_node *new_f, *f, **fp, **del_fp; local 664 struct dn_fib_node **fp, **del_fp, *f; local 742 dn_flush_list(struct dn_fib_node **fp, int z, struct dn_hash *table) argument [all...] |
/net/sched/ |
H A D | cls_route.c | 293 struct route4_filter **fp, *f = (struct route4_filter *)arg; local 304 for (fp = &b->ht[from_hash(h >> 16)]; *fp; fp = &(*fp)->next) { 305 if (*fp == f) { 307 *fp = f->next; 345 struct route4_filter *fp; local 400 for (fp = b->ht[h2]; fp; f 438 struct route4_filter *f, *f1, **fp; local [all...] |
H A D | cls_rsvp.h | 301 struct rsvp_filter **fp, *f = (struct rsvp_filter *)arg; local 307 for (fp = &s->ht[(h >> 8) & 0xFF]; *fp; fp = &(*fp)->next) { 308 if (*fp == f) { 310 *fp = f->next; 426 struct rsvp_filter *f, **fp; local 519 fp = &s->ht[h2]; 527 for (fp [all...] |
H A D | cls_fw.c | 170 struct fw_filter **fp; local 175 for (fp = &head->ht[fw_hash(f->id)]; *fp; fp = &(*fp)->next) { 176 if (*fp == f) { 178 *fp = f->next;
|
H A D | cls_tcindex.c | 312 struct tcindex_filter **fp; local 317 for (fp = p->h+(handle % p->hash); *fp; fp = &(*fp)->next) 319 *fp = f;
|
/net/ipv6/netfilter/ |
H A D | nf_conntrack_reasm.c | 364 struct sk_buff *fp, *op, *head = fq->q.fragments; local 431 for (fp=head->next; fp; fp = fp->next) { 432 head->data_len += fp->len; 433 head->len += fp->len; 434 if (head->ip_summed != fp->ip_summed) 437 head->csum = csum_add(head->csum, fp->csum); 438 head->truesize += fp [all...] |
/net/ipv4/ |
H A D | ip_fragment.c | 302 struct sk_buff *fp; local 310 fp = qp->q.fragments; 312 struct sk_buff *xp = fp->next; 314 sum_truesize += fp->truesize; 315 kfree_skb(fp); 316 fp = xp; 317 } while (fp); 522 struct sk_buff *fp, *head = qp->q.fragments; local 539 fp = skb_clone(head, GFP_ATOMIC); 540 if (!fp) [all...] |
H A D | inet_fragment.c | 174 struct sk_buff *fp; local 182 fp = q->fragments; 184 while (fp) { 185 struct sk_buff *xp = fp->next; 187 sum_truesize += fp->truesize; 188 frag_kfree_skb(nf, f, fp); 189 fp = xp;
|
H A D | inetpeer.c | 104 atomic_t *fp = inetpeer_seq_ptr(family); local 106 if (unlikely(base->flush_seq != atomic_read(fp))) { 108 base->flush_seq = atomic_read(fp); 114 atomic_t *fp = inetpeer_seq_ptr(family); local 116 atomic_inc(fp);
|
/net/802/ |
H A D | hippi.c | 60 * have to set the fp field this way. 62 hip->fp.fixed = htonl(0x04800018); 63 hip->fp.d2_size = htonl(len + 8);
|
/net/ |
H A D | compat.c | 280 int fdnum = scm->fp->count; 281 struct file **fp = scm->fp->fp; local 290 err = security_file_receive(fp[i]); 304 fd_install(new_fd, get_file(fp[i]));
|