/net/core/ |
H A D | tso.c | 47 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; local 50 tso->size = frag->size; 51 tso->data = page_address(frag->page.p) + frag->page_offset; 70 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; local 73 tso->size = frag->size; 74 tso->data = page_address(frag->page.p) + frag->page_offset;
|
H A D | datagram.c | 343 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; local 347 end = start + skb_frag_size(frag); 351 struct page *page = skb_frag_page(frag); 356 err = memcpy_toiovec(to, vaddr + frag->page_offset + 429 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; local 433 end = start + skb_frag_size(frag); 437 struct page *page = skb_frag_page(frag); 442 err = memcpy_toiovecend(to, vaddr + frag->page_offset + 519 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; local 523 end = start + skb_frag_size(frag); 675 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; local [all...] |
H A D | skbuff.c | 332 struct page_frag frag; member in struct:netdev_alloc_cache 349 if (unlikely(!nc->frag.page)) { 356 nc->frag.page = alloc_pages(gfp, order); 357 if (likely(nc->frag.page)) 362 nc->frag.size = PAGE_SIZE << order; 367 &nc->frag.page->_count); 369 nc->frag.offset = 0; 372 if (nc->frag.offset + fragsz > nc->frag.size) { 373 if (atomic_read(&nc->frag 467 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; local 1355 struct sk_buff *frag; local 1901 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; local 1975 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; local 2069 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; local 2675 skb_frag_t *frag; local 2889 skb_frag_t *frag = skb_shinfo(head_skb)->frags; local 3116 skb_frag_t *frag; local 3149 skb_frag_t *frag = pinfo->frags + nr_frags; local 3302 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; local [all...] |
/net/tipc/ |
H A D | msg.c | 91 * @*headbuf: in: NULL for first frag, otherwise value returned from prev call 100 struct sk_buff *frag = *buf; local 107 if (!frag) 110 msg = buf_msg(frag); 112 frag->next = NULL; 113 skb_pull(frag, msg_hdr_sz(msg)); 118 if (unlikely(skb_unclone(frag, GFP_ATOMIC))) 120 head = *headbuf = frag; 130 if (skb_try_coalesce(head, frag, &headstolen, &delta)) { 131 kfree_skb_partial(frag, headstole 435 struct sk_buff *frag = buf; local [all...] |
/net/sctp/ |
H A D | chunk.c | 175 __u8 frag; local 258 frag = SCTP_DATA_MIDDLE_FRAG; 261 frag |= SCTP_DATA_FIRST_FRAG; 264 frag |= SCTP_DATA_LAST_FRAG; 272 frag |= SCTP_DATA_SACK_IMM; 275 chunk = sctp_make_datafrag_empty(asoc, sinfo, len, frag, 0); 305 frag = SCTP_DATA_NOT_FRAG; 307 frag = SCTP_DATA_LAST_FRAG; 311 frag |= SCTP_DATA_SACK_IMM; 313 chunk = sctp_make_datafrag_empty(asoc, sinfo, over, frag, [all...] |
H A D | ulpevent.c | 958 struct sk_buff *skb, *frag; local 974 skb_walk_frags(skb, frag) 975 sctp_ulpevent_receive_data(sctp_skb2event(frag), asoc); 983 struct sk_buff *skb, *frag; local 1000 skb_walk_frags(skb, frag) { 1005 sctp_ulpevent_release_frag_data(sctp_skb2event(frag)); 1015 struct sk_buff *skb, *frag; local 1023 skb_walk_frags(skb, frag) { 1028 sctp_ulpevent_release_frag_data(sctp_skb2event(frag));
|
/net/rds/ |
H A D | ib_recv.c | 158 struct rds_page_frag *frag; local 176 list_for_each_entry_safe(frag, frag_tmp, &list, f_cache_entry) { 177 list_del(&frag->f_cache_entry); 178 WARN_ON(!list_empty(&frag->f_item)); 179 kmem_cache_free(rds_ib_frag_slab, frag); 189 /* Recycle frag and attached recv buffer f_sg */ 191 struct rds_page_frag *frag) 193 rdsdebug("frag %p page %p\n", frag, sg_page(&frag 190 rds_ib_frag_free(struct rds_ib_connection *ic, struct rds_page_frag *frag) argument 202 struct rds_page_frag *frag; local 273 struct rds_page_frag *frag; local 479 struct rds_page_frag *frag; local 743 struct rds_page_frag *frag; local [all...] |
H A D | iw_recv.c | 46 static void rds_iw_frag_drop_page(struct rds_page_frag *frag) argument 48 rdsdebug("frag %p page %p\n", frag, frag->f_page); 49 __free_page(frag->f_page); 50 frag->f_page = NULL; 53 static void rds_iw_frag_free(struct rds_page_frag *frag) argument 55 rdsdebug("frag %p page %p\n", frag, frag 72 struct rds_page_frag *frag = recv->r_frag; local 279 struct rds_page_frag *frag; local 310 struct rds_page_frag *frag; local 574 struct rds_page_frag *frag; local [all...] |
/net/ieee802154/ |
H A D | 6lowpan_rtnl.c | 218 struct sk_buff *frag; local 221 frag = alloc_skb(real_dev->hard_header_len + 225 if (likely(frag)) { 226 frag->dev = real_dev; 227 frag->priority = skb->priority; 228 skb_reserve(frag, real_dev->hard_header_len); 229 skb_reset_network_header(frag); 230 *mac_cb(frag) = *mac_cb(skb); 232 rc = dev_hard_header(frag, real_dev, 0, &master_hdr->dest, 235 kfree_skb(frag); 250 struct sk_buff *frag; local [all...] |
/net/ipv6/ |
H A D | ip6_output.c | 23 * Imran Patel : frag id should be in NBO 556 struct sk_buff *frag; local 605 skb_walk_frags(skb, frag) { 607 if (frag->len > mtu || 608 ((frag->len & 7) && frag->next) || 609 skb_headroom(frag) < hlen) 613 if (skb_shared(frag)) 616 BUG_ON(frag->sk); 618 frag [all...] |
/net/xfrm/ |
H A D | xfrm_ipcomp.c | 72 skb_frag_t *frag; local 79 frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags; 86 __skb_frag_set_page(frag, page); 92 frag->page_offset = 0; 93 skb_frag_size_set(frag, len); 94 memcpy(skb_frag_address(frag), scratch, len);
|
/net/ipv4/ |
H A D | ip_output.c | 533 struct sk_buff *frag, *frag2; local 542 skb_walk_frags(skb, frag) { 544 if (frag->len > mtu || 545 ((frag->len & 7) && frag->next) || 546 skb_headroom(frag) < hlen) 550 if (skb_shared(frag)) 553 BUG_ON(frag->sk); 555 frag->sk = skb->sk; 556 frag [all...] |
H A D | icmp.c | 797 info = ntohs(icmph->un.frag.mtu);
|
/net/bluetooth/ |
H A D | af_bluetooth.c | 377 struct sk_buff *frag; local 382 skb_walk_frags(skb, frag) { 383 if (chunk <= frag->len) { 387 __skb_pull(frag, chunk); 389 } else if (frag->len) { 390 /* Pulling all frag data */ 391 chunk -= frag->len; 392 skb->len -= frag->len; 393 skb->data_len -= frag->len; 394 __skb_pull(frag, fra [all...] |
H A D | amp.c | 296 memcpy(cp->frag, ctrl->assoc, frag_len);
|
H A D | l2cap_core.c | 2092 struct sk_buff **frag; local 2103 frag = &skb_shinfo(skb)->frag_list; 2114 *frag = tmp; 2116 if (chan->ops->memcpy_fromiovec(chan, skb_put(*frag, count), 2123 skb->len += (*frag)->len; 2124 skb->data_len += (*frag)->len; 2126 frag = &(*frag)->next; 2849 struct sk_buff *skb, **frag; local 2889 frag [all...] |
/net/irda/ |
H A D | irttp.c | 237 struct sk_buff *skb, *frag; local 260 while ((frag = skb_dequeue(&self->rx_fragments)) != NULL) { 261 skb_copy_to_linear_data_offset(skb, n, frag->data, frag->len); 262 n += frag->len; 264 dev_kfree_skb(frag); 295 struct sk_buff *frag; local 311 frag = alloc_skb(self->max_seg_size+self->max_header_size, 313 if (!frag) 316 skb_reserve(frag, sel [all...] |
/net/openvswitch/ |
H A D | flow.c | 274 key->ip.frag = OVS_FRAG_TYPE_LATER; 276 key->ip.frag = OVS_FRAG_TYPE_FIRST; 278 key->ip.frag = OVS_FRAG_TYPE_NONE; 511 key->ip.frag = OVS_FRAG_TYPE_LATER; 516 key->ip.frag = OVS_FRAG_TYPE_FIRST; 518 key->ip.frag = OVS_FRAG_TYPE_NONE; 603 if (key->ip.frag == OVS_FRAG_TYPE_LATER) 606 key->ip.frag = OVS_FRAG_TYPE_FIRST;
|
H A D | flow.h | 109 u8 frag; /* One of OVS_FRAG_TYPE_*. */ member in struct:sw_flow_key::__anon117
|
H A D | flow_netlink.c | 157 if (match->key->ip.frag != OVS_FRAG_TYPE_LATER) { 192 if (match->key->ip.frag != OVS_FRAG_TYPE_LATER) { 674 SW_FLOW_KEY_PUT(match, ip.frag, 707 SW_FLOW_KEY_PUT(match, ip.frag, 1119 ipv4_key->ipv4_frag = output->ip.frag; 1135 ipv6_key->ipv6_frag = output->ip.frag; 1154 swkey->ip.frag != OVS_FRAG_TYPE_LATER) { 1535 if (ipv4_key->ipv4_frag != flow_key->ip.frag) 1551 if (ipv6_key->ipv6_frag != flow_key->ip.frag)
|
/net/wireless/ |
H A D | wext-compat.c | 324 struct iw_param *frag, char *extra) 331 if (frag->disabled || !frag->fixed) 333 else if (frag->value < 256) 337 wdev->wiphy->frag_threshold = frag->value & ~0x1; 350 struct iw_param *frag, char *extra) 354 frag->value = wdev->wiphy->frag_threshold; 355 frag->disabled = frag->value == (u32) -1; 356 frag 322 cfg80211_wext_siwfrag(struct net_device *dev, struct iw_request_info *info, struct iw_param *frag, char *extra) argument 348 cfg80211_wext_giwfrag(struct net_device *dev, struct iw_request_info *info, struct iw_param *frag, char *extra) argument [all...] |
/net/ipv4/netfilter/ |
H A D | nf_log_ipv4.c | 192 ntohs(ich->un.frag.mtu));
|
/net/mac80211/ |
H A D | rx.c | 1594 unsigned int frag, unsigned int seq, int rx_queue, 1611 entry->last_frag = frag; 1620 unsigned int frag, unsigned int seq, 1637 entry->last_frag + 1 != frag) 1667 unsigned int frag, seq; local 1679 frag = sc & IEEE80211_SCTL_FRAG; 1681 if (likely(!ieee80211_has_morefrags(fc) && frag == 0)) 1702 if (frag == 0) { 1704 entry = ieee80211_reassemble_add(rx->sdata, frag, seq, 1722 entry = ieee80211_reassemble_find(rx->sdata, frag, se 1593 ieee80211_reassemble_add(struct ieee80211_sub_if_data *sdata, unsigned int frag, unsigned int seq, int rx_queue, struct sk_buff **skb) argument 1619 ieee80211_reassemble_find(struct ieee80211_sub_if_data *sdata, unsigned int frag, unsigned int seq, int rx_queue, struct ieee80211_hdr *hdr) argument [all...] |
/net/appletalk/ |
H A D | ddp.c | 952 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; local 955 end = start + skb_frag_size(frag); 961 vaddr = kmap_atomic(skb_frag_page(frag)); 962 sum = atalk_sum_partial(vaddr + frag->page_offset +
|
/net/ipv6/netfilter/ |
H A D | Kconfig | 131 tristate '"frag" Fragmentation header match support' 134 frag matching allows you to match packets based on the fragmentation
|