Searched refs:page (Results 26 - 50 of 67) sorted by relevance

123

/net/8021q/
H A DKconfig13 See the VLAN web page for more information:
/net/sunrpc/auth_gss/
H A Dgss_rpc_xdr.h137 struct page **pages; /* Array of contiguous pages */
138 unsigned int page_base; /* Start of page data */
139 unsigned int page_len; /* Length of page data */
150 struct page **pages;
H A Dgss_krb5_crypto.c393 struct page **pages;
405 struct page *in_page;
410 /* Worst case is 4 fragments: head, end of page 1, start
411 * of page 2, tail. Anything more is a bug. */
463 int offset, struct page **pages)
502 /* Worst case is 4 fragments: head, end of page 1, start
503 * of page 2, tail. Anything more is a bug. */
563 * separate page before calling gss_wrap.
598 u32 offset, u8 *iv, struct page **pages, int encrypt)
604 struct page **save_page
[all...]
H A Dgss_krb5_wrap.c94 /* XXX: NOTE: we do not adjust the page lengths--they represent
95 * a range of data in the real filesystem page cache, and we need
98 * In the case of a request that fits into a single page, the server
102 * It's not really a problem that we don't fool with the page and
150 * The pages, however, may be real pages in the page cache and we replace
159 struct xdr_buf *buf, struct page **pages)
168 struct page **tmp_pages;
441 struct xdr_buf *buf, struct page **pages)
535 * doesn't want it checked; see page 6 of rfc 2203.
592 struct xdr_buf *buf, struct page **page
[all...]
H A Dgss_mech_switch.c445 struct page **inpages)
H A Dgss_rpc_upcall.c225 arg->pages = kzalloc(arg->npages * sizeof(struct page *), GFP_KERNEL);
/net/sunrpc/
H A Dxdr.c133 struct page **pages, unsigned int base, unsigned int len)
160 * @pgto_base: page vector address of destination
161 * @pgfrom_base: page vector address of source
166 * if a memory area starts at byte 'base' in page 'pages[i]',
172 _shift_data_right_pages(struct page **pages, size_t pgto_base,
175 struct page **pgfrom, **pgto;
191 /* Are any pointers crossing a page boundary? */
225 * @pgbase: page vector address of destination
233 _copy_to_pages(struct page **pages, size_t pgbase, const char *p, size_t len)
235 struct page **pgt
501 void *page; local
[all...]
H A Dsocklib.c74 struct page **ppage = xdr->pages;
H A Dcache.c872 struct page *page; local
879 page = find_or_create_page(mapping, 0, GFP_KERNEL);
880 if (!page)
883 kaddr = kmap(page);
885 kunmap(page);
886 unlock_page(page);
887 page_cache_release(page);
1159 * Each request is at most one page long.
H A Dbackchannel_rqst.c83 * requests. Each deviceID is 16 bytes, so allocate one page
90 struct page *page_rcv = NULL, *page_snd = NULL;
H A Dxprtsock.c404 ssize_t (*do_sendpage)(struct socket *sock, struct page *page,
406 struct page **ppage;
2447 struct page *page; local
2454 page = alloc_page(GFP_KERNEL);
2455 if (!page)
2458 buf = page_address(page);
/net/mac802154/
H A Dwpan.c318 u8 chan, page; local
325 page = priv->page;
329 page >= WPAN_NUM_PAGES ||
346 return mac802154_tx(priv->hw, skb, page, chan);
395 priv->page = 0;
/net/rds/
H A Drds.h690 /* page.c */
693 int rds_page_copy_user(struct page *page, unsigned long offset,
696 #define rds_page_copy_to_user(page, offset, ptr, bytes) \
697 rds_page_copy_user(page, offset, ptr, bytes, 1)
698 #define rds_page_copy_from_user(page, offset, ptr, bytes) \
699 rds_page_copy_user(page, offset, ptr, bytes, 0)
H A Dib_rdma.c484 struct page *page = sg_page(&ibmr->sg[i]); local
489 set_page_dirty(page);
490 put_page(page);
H A Diw.h28 /* page offset of the final full frag that fits in the page */
32 struct page *f_page;
/net/openvswitch/
H A Ddatapath.h22 #include <asm/page.h>
/net/sunrpc/xprtrdma/
H A Dsvc_rdma_transport.c490 struct page *svc_rdma_get_page(void)
492 struct page *page; local
494 while ((page = alloc_page(GFP_KERNEL)) == NULL) {
499 return page;
506 struct page *page; local
517 page = svc_rdma_get_page();
518 ctxt->pages[sge_no] = page;
520 page,
[all...]
H A Dxprt_rdma.h152 * An external memory region is any buffer or page that is registered
219 struct page *mr_page; /* owning page, if any */
220 char *mr_offset; /* kva if no page, else offset */
/net/ieee802154/
H A Dnl-mac.c211 u32 unscanned, u8 page,
229 nla_put_u8(msg, IEEE802154_ATTR_PAGE, page) ||
362 u8 page; local
391 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]);
393 page = 0;
397 page,
482 u8 page; local
517 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]);
519 page = 0;
528 ret = ieee802154_mlme_ops(dev)->start_req(dev, &addr, channel, page,
209 ieee802154_nl_scan_confirm(struct net_device *dev, u8 status, u8 scan_type, u32 unscanned, u8 page, u8 *edl ) argument
543 u8 page; local
[all...]
/net/ipv4/
H A Dip_output.c773 csum_page(struct page *page, int offset, int copy) argument
777 kaddr = kmap(page);
779 kunmap(page);
1045 if (!skb_can_coalesce(skb, i, pfrag->page,
1051 __skb_fill_page_desc(skb, i, pfrag->page,
1054 get_page(pfrag->page);
1058 page_address(pfrag->page) + pfrag->offset,
1126 * until ip_push_pending_frames() is called. Each piece can be a page
1127 * or non-page dat
1160 ip_append_page(struct sock *sk, struct flowi4 *fl4, struct page *page, int offset, size_t size, int flags) argument
[all...]
H A Dtcp.c895 static ssize_t do_tcp_sendpages(struct sock *sk, struct page *page, int offset, argument
945 can_coalesce = skb_can_coalesce(skb, i, page, offset);
956 get_page(page);
957 skb_fill_page_desc(skb, i, page, offset, copy);
1015 int tcp_sendpage(struct sock *sk, struct page *page, int offset, argument
1022 return sock_no_sendpage(sk->sk_socket, page, offset, size,
1026 res = do_tcp_sendpages(sk, page, offset, size, flags);
1039 /* Small frames wont use a full page
2988 struct page *page = skb_frag_page(f) + (offset >> PAGE_SHIFT); local
[all...]
/net/9p/
H A Dtrans_virtio.c105 /* How many bytes left in this page. */
225 struct page **pdata, int nr_pages, char *data, int count)
233 * if the first page doesn't start at
234 * page boundary find the offset
317 struct page **pages, char *data,
373 struct page **in_pages = NULL, **out_pages = NULL;
381 out_pages = kmalloc(sizeof(struct page *) * out_nr_pages,
398 in_pages = kmalloc(sizeof(struct page *) * in_nr_pages,
701 * that are not at page boundary, that can result in an extra
702 * page i
[all...]
/net/ceph/
H A Dcrypto.c138 struct page *page; local
142 page = vmalloc_to_page(buf);
144 page = virt_to_page(buf);
146 sg_set_page(sg, page, len, off);
/net/core/
H A Dsock.c1826 if (pfrag->page) {
1827 if (atomic_read(&pfrag->page->_count) == 1) {
1833 put_page(pfrag->page);
1838 pfrag->page = alloc_pages(gfp | __GFP_COMP |
1841 if (likely(pfrag->page)) {
1846 pfrag->page = alloc_page(gfp);
1847 if (likely(pfrag->page)) {
2150 ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset, size_t size, int flags) argument
2155 char *kaddr = kmap(page);
[all...]
H A Dpktgen.c261 struct page *page; member in struct:pktgen_dev
1151 if (pkt_dev->page) {
1152 put_page(pkt_dev->page);
1153 pkt_dev->page = NULL;
2678 if (unlikely(!pkt_dev->page)) {
2683 pkt_dev->page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0);
2684 if (!pkt_dev->page)
2687 get_page(pkt_dev->page);
2688 skb_frag_set_page(skb, i, pkt_dev->page);
[all...]

Completed in 798 milliseconds

123