/net/8021q/ |
H A D | Kconfig | 13 See the VLAN web page for more information:
|
/net/sunrpc/auth_gss/ |
H A D | gss_rpc_xdr.h | 137 struct page **pages; /* Array of contiguous pages */ 138 unsigned int page_base; /* Start of page data */ 139 unsigned int page_len; /* Length of page data */ 150 struct page **pages;
|
H A D | gss_krb5_crypto.c | 393 struct page **pages; 405 struct page *in_page; 410 /* Worst case is 4 fragments: head, end of page 1, start 411 * of page 2, tail. Anything more is a bug. */ 463 int offset, struct page **pages) 502 /* Worst case is 4 fragments: head, end of page 1, start 503 * of page 2, tail. Anything more is a bug. */ 563 * separate page before calling gss_wrap. 598 u32 offset, u8 *iv, struct page **pages, int encrypt) 604 struct page **save_page [all...] |
H A D | gss_krb5_wrap.c | 94 /* XXX: NOTE: we do not adjust the page lengths--they represent 95 * a range of data in the real filesystem page cache, and we need 98 * In the case of a request that fits into a single page, the server 102 * It's not really a problem that we don't fool with the page and 150 * The pages, however, may be real pages in the page cache and we replace 159 struct xdr_buf *buf, struct page **pages) 168 struct page **tmp_pages; 441 struct xdr_buf *buf, struct page **pages) 535 * doesn't want it checked; see page 6 of rfc 2203. 592 struct xdr_buf *buf, struct page **page [all...] |
H A D | gss_mech_switch.c | 445 struct page **inpages)
|
H A D | gss_rpc_upcall.c | 225 arg->pages = kzalloc(arg->npages * sizeof(struct page *), GFP_KERNEL);
|
/net/sunrpc/ |
H A D | xdr.c | 133 struct page **pages, unsigned int base, unsigned int len) 160 * @pgto_base: page vector address of destination 161 * @pgfrom_base: page vector address of source 166 * if a memory area starts at byte 'base' in page 'pages[i]', 172 _shift_data_right_pages(struct page **pages, size_t pgto_base, 175 struct page **pgfrom, **pgto; 191 /* Are any pointers crossing a page boundary? */ 225 * @pgbase: page vector address of destination 233 _copy_to_pages(struct page **pages, size_t pgbase, const char *p, size_t len) 235 struct page **pgt 501 void *page; local [all...] |
H A D | socklib.c | 74 struct page **ppage = xdr->pages;
|
H A D | cache.c | 872 struct page *page; local 879 page = find_or_create_page(mapping, 0, GFP_KERNEL); 880 if (!page) 883 kaddr = kmap(page); 885 kunmap(page); 886 unlock_page(page); 887 page_cache_release(page); 1159 * Each request is at most one page long.
|
H A D | backchannel_rqst.c | 83 * requests. Each deviceID is 16 bytes, so allocate one page 90 struct page *page_rcv = NULL, *page_snd = NULL;
|
H A D | xprtsock.c | 404 ssize_t (*do_sendpage)(struct socket *sock, struct page *page, 406 struct page **ppage; 2447 struct page *page; local 2454 page = alloc_page(GFP_KERNEL); 2455 if (!page) 2458 buf = page_address(page);
|
/net/mac802154/ |
H A D | wpan.c | 318 u8 chan, page; local 325 page = priv->page; 329 page >= WPAN_NUM_PAGES || 346 return mac802154_tx(priv->hw, skb, page, chan); 395 priv->page = 0;
|
/net/rds/ |
H A D | rds.h | 690 /* page.c */ 693 int rds_page_copy_user(struct page *page, unsigned long offset, 696 #define rds_page_copy_to_user(page, offset, ptr, bytes) \ 697 rds_page_copy_user(page, offset, ptr, bytes, 1) 698 #define rds_page_copy_from_user(page, offset, ptr, bytes) \ 699 rds_page_copy_user(page, offset, ptr, bytes, 0)
|
H A D | ib_rdma.c | 484 struct page *page = sg_page(&ibmr->sg[i]); local 489 set_page_dirty(page); 490 put_page(page);
|
H A D | iw.h | 28 /* page offset of the final full frag that fits in the page */ 32 struct page *f_page;
|
/net/openvswitch/ |
H A D | datapath.h | 22 #include <asm/page.h>
|
/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_transport.c | 490 struct page *svc_rdma_get_page(void) 492 struct page *page; local 494 while ((page = alloc_page(GFP_KERNEL)) == NULL) { 499 return page; 506 struct page *page; local 517 page = svc_rdma_get_page(); 518 ctxt->pages[sge_no] = page; 520 page, [all...] |
H A D | xprt_rdma.h | 152 * An external memory region is any buffer or page that is registered 219 struct page *mr_page; /* owning page, if any */ 220 char *mr_offset; /* kva if no page, else offset */
|
/net/ieee802154/ |
H A D | nl-mac.c | 211 u32 unscanned, u8 page, 229 nla_put_u8(msg, IEEE802154_ATTR_PAGE, page) || 362 u8 page; local 391 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]); 393 page = 0; 397 page, 482 u8 page; local 517 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]); 519 page = 0; 528 ret = ieee802154_mlme_ops(dev)->start_req(dev, &addr, channel, page, 209 ieee802154_nl_scan_confirm(struct net_device *dev, u8 status, u8 scan_type, u32 unscanned, u8 page, u8 *edl ) argument 543 u8 page; local [all...] |
/net/ipv4/ |
H A D | ip_output.c | 773 csum_page(struct page *page, int offset, int copy) argument 777 kaddr = kmap(page); 779 kunmap(page); 1045 if (!skb_can_coalesce(skb, i, pfrag->page, 1051 __skb_fill_page_desc(skb, i, pfrag->page, 1054 get_page(pfrag->page); 1058 page_address(pfrag->page) + pfrag->offset, 1126 * until ip_push_pending_frames() is called. Each piece can be a page 1127 * or non-page dat 1160 ip_append_page(struct sock *sk, struct flowi4 *fl4, struct page *page, int offset, size_t size, int flags) argument [all...] |
H A D | tcp.c | 895 static ssize_t do_tcp_sendpages(struct sock *sk, struct page *page, int offset, argument 945 can_coalesce = skb_can_coalesce(skb, i, page, offset); 956 get_page(page); 957 skb_fill_page_desc(skb, i, page, offset, copy); 1015 int tcp_sendpage(struct sock *sk, struct page *page, int offset, argument 1022 return sock_no_sendpage(sk->sk_socket, page, offset, size, 1026 res = do_tcp_sendpages(sk, page, offset, size, flags); 1039 /* Small frames wont use a full page 2988 struct page *page = skb_frag_page(f) + (offset >> PAGE_SHIFT); local [all...] |
/net/9p/ |
H A D | trans_virtio.c | 105 /* How many bytes left in this page. */ 225 struct page **pdata, int nr_pages, char *data, int count) 233 * if the first page doesn't start at 234 * page boundary find the offset 317 struct page **pages, char *data, 373 struct page **in_pages = NULL, **out_pages = NULL; 381 out_pages = kmalloc(sizeof(struct page *) * out_nr_pages, 398 in_pages = kmalloc(sizeof(struct page *) * in_nr_pages, 701 * that are not at page boundary, that can result in an extra 702 * page i [all...] |
/net/ceph/ |
H A D | crypto.c | 138 struct page *page; local 142 page = vmalloc_to_page(buf); 144 page = virt_to_page(buf); 146 sg_set_page(sg, page, len, off);
|
/net/core/ |
H A D | sock.c | 1826 if (pfrag->page) { 1827 if (atomic_read(&pfrag->page->_count) == 1) { 1833 put_page(pfrag->page); 1838 pfrag->page = alloc_pages(gfp | __GFP_COMP | 1841 if (likely(pfrag->page)) { 1846 pfrag->page = alloc_page(gfp); 1847 if (likely(pfrag->page)) { 2150 ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset, size_t size, int flags) argument 2155 char *kaddr = kmap(page); [all...] |
H A D | pktgen.c | 261 struct page *page; member in struct:pktgen_dev 1151 if (pkt_dev->page) { 1152 put_page(pkt_dev->page); 1153 pkt_dev->page = NULL; 2678 if (unlikely(!pkt_dev->page)) { 2683 pkt_dev->page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0); 2684 if (!pkt_dev->page) 2687 get_page(pkt_dev->page); 2688 skb_frag_set_page(skb, i, pkt_dev->page); [all...] |