/net/9p/ |
H A D | trans_common.h | 15 void p9_release_pages(struct page **, int); 16 int p9_payload_gup(char *, int *, struct page **, int);
|
H A D | trans_common.c | 25 void p9_release_pages(struct page **pages, int nr_pages) 51 * @pdata_off: data offset into the first page after translation (gup). 57 int p9_payload_gup(char *data, int *nr_pages, struct page **pages, int write)
|
/net/ceph/ |
H A D | pagelist.c | 10 struct page *page = list_entry(pl->head.prev, struct page, lru); local 11 kunmap(page); 20 struct page *page = list_first_entry(&pl->head, struct page, local 22 list_del(&page->lru); 23 __free_page(page); 32 struct page *pag local 86 struct page *page = __page_cache_alloc(GFP_NOFS); local 100 struct page *page = list_first_entry(&pl->free_list, local 129 struct page *page; local [all...] |
H A D | pagevec.c | 15 struct page **ceph_get_direct_page_vector(const void __user *data, 18 struct page **pages; 47 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) 60 void ceph_release_page_vector(struct page **pages, int num_pages) 73 struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags) 75 struct page **pages; 93 * copy user data into a page vector 95 int ceph_copy_user_to_page_vector(struct page **pages, 121 void ceph_copy_to_page_vector(struct page **pages, 144 void ceph_copy_from_page_vector(struct page **page [all...] |
/net/rds/ |
H A D | page.c | 41 struct page *r_page; 56 int rds_page_copy_user(struct page *page, unsigned long offset, argument 63 addr = kmap(page); 71 kunmap(page); 87 * If @bytes is at least a full page then this just returns a page from 90 * If @bytes is a partial page then this stores the unused region of the 91 * page in a per-cpu structure. Future partial-page allocation 102 struct page *page; local [all...] |
H A D | rdma.c | 49 * get the number of pages by looking at the page indices that the start and 158 struct page **pages, int write) 178 struct page **pages = NULL; 207 pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL); 232 * Pin the pages that make up the user buffer and transfer the page 234 * the whole region after transferring the partial page references 235 * to the sg array so that we can have one page ref cleanup path. 239 * the zero page. 281 * map page aligned regions. So we keep the offset, and build 448 struct page *pag local 467 struct page *page = sg_page(ao->op_sg); local 769 struct page *page = NULL; local [all...] |
H A D | info.c | 65 struct page **pages; 126 rdsdebug("page %p addr %p offset %lu this %lu data %p " 167 struct page **pages = NULL; 191 pages = kmalloc(nr_pages * sizeof(struct page *), GFP_KERNEL);
|
H A D | Makefile | 4 loop.o page.o rdma.o
|
/net/sunrpc/ |
H A D | sunrpc.h | 47 struct page *headpage, unsigned long headoffset, 48 struct page *tailpage, unsigned long tailoffset);
|
/net/mac802154/ |
H A D | tx.c | 42 u8 page; member in struct:xmit_work 53 xw->priv->phy->current_page != xw->page) { 55 xw->page, 63 xw->priv->phy->current_page = xw->page; 85 u8 page, u8 chan) 90 if (!(priv->phy->channels_supported[page] & (1 << chan))) { 125 work->page = page; 84 mac802154_tx(struct mac802154_priv *priv, struct sk_buff *skb, u8 page, u8 chan) argument
|
H A D | mac802154.h | 81 u8 page; member in struct:mac802154_sub_if_data 106 u8 page, u8 chan); 114 void mac802154_dev_set_page_channel(struct net_device *dev, u8 page, u8 chan);
|
H A D | monitor.c | 41 u8 chan, page; local 47 page = priv->hw->phy->current_page; 52 if (WARN_ON(page >= WPAN_NUM_PAGES) || 60 return mac802154_tx(priv->hw, skb, page, chan); 115 priv->page = 0;
|
H A D | mac_cmd.c | 38 u8 channel, u8 page, 48 mac802154_dev_set_page_channel(dev, page, channel); 36 mac802154_mlme_start_req(struct net_device *dev, struct ieee802154_addr *addr, u8 channel, u8 page, u8 bcn_ord, u8 sf_ord, u8 pan_coord, u8 blx, u8 coord_realign) argument
|
H A D | mib.c | 180 res = hw->ops->set_channel(&hw->hw, priv->page, priv->chan); 185 priv->hw->phy->current_page = priv->page; 192 void mac802154_dev_set_page_channel(struct net_device *dev, u8 page, u8 chan) argument 200 priv->page = page; 206 priv->hw->phy->current_page != priv->page) {
|
/net/core/ |
H A D | user_dma.c | 82 struct page *page = skb_frag_page(frag); local 87 cookie = dma_memcpy_pg_to_iovec(chan, to, pinned_list, page,
|
H A D | datagram.c | 345 struct page *page = skb_frag_page(frag); local 349 vaddr = kmap(page); 352 kunmap(page); 431 struct page *page = skb_frag_page(frag); local 435 vaddr = kmap(page); 438 kunmap(page); 521 struct page *page local 607 struct page *page = skb_frag_page(frag); local [all...] |
H A D | skbuff.c | 80 put_page(buf->page); 86 get_page(buf->page); 365 * containing page->_count every time we allocate a fragment. 380 if (unlikely(!nc->frag.page)) { 387 nc->frag.page = alloc_pages(gfp, order); 388 if (likely(nc->frag.page)) 395 atomic_set(&nc->frag.page->_count, NETDEV_PAGECNT_MAX_BIAS); 402 if ((atomic_read(&nc->frag.page->_count) == nc->pagecnt_bias) || 403 atomic_sub_and_test(nc->pagecnt_bias, &nc->frag.page->_count)) 408 data = page_address(nc->frag.page) 474 skb_add_rx_frag(struct sk_buff *skb, int i, struct page *page, int off, int size, unsigned int truesize) argument 818 struct page *page, *head = NULL; local 1668 linear_to_page(struct page *page, unsigned int *len, unsigned int *offset, struct sock *sk) argument 1687 spd_can_coalesce(const struct splice_pipe_desc *spd, struct page *page, unsigned int offset) argument 1700 spd_fill_page(struct splice_pipe_desc *spd, struct pipe_inode_info *pipe, struct page *page, unsigned int *len, unsigned int offset, bool linear, struct sock *sk) argument 1727 __splice_segment(struct page *page, unsigned int poff, unsigned int plen, unsigned int *off, unsigned int *len, struct splice_pipe_desc *spd, bool linear, struct sock *sk, struct pipe_inode_info *pipe) argument 2995 struct page *page = virt_to_head_page(skb->head); local 3456 struct page *page; local [all...] |
/net/atm/ |
H A D | mpoa_proc.c | 209 char *page, *p; local 218 page = (char *)__get_free_page(GFP_KERNEL); 219 if (!page) 222 for (p = page, len = 0; len < nbytes; p++, len++) { 224 free_page((unsigned long)page); 233 if (!parse_qos(page)) 234 printk("mpoa: proc_mpc_write: could not parse '%s'\n", page); 236 free_page((unsigned long)page);
|
H A D | proc.c | 380 unsigned long page; local 385 page = get_zeroed_page(GFP_KERNEL); 386 if (!page) 392 length = dev->ops->proc_read(dev, pos, (char *)page); 397 if (copy_to_user(buf, (char *)page, length)) 401 free_page(page);
|
/net/ipv4/ |
H A D | udp_impl.h | 26 extern int udp_sendpage(struct sock *sk, struct page *page, int offset,
|
/net/ieee802154/ |
H A D | nl-mac.c | 195 u8 status, u8 scan_type, u32 unscanned, u8 page, 213 nla_put_u8(msg, IEEE802154_ATTR_PAGE, page) || 317 u8 page; local 346 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]); 348 page = 0; 352 page, 442 u8 page; local 476 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]); 478 page = 0; 487 ret = ieee802154_mlme_ops(dev)->start_req(dev, &addr, channel, page, 194 ieee802154_nl_scan_confirm(struct net_device *dev, u8 status, u8 scan_type, u32 unscanned, u8 page, u8 *edl ) argument 502 u8 page; local [all...] |
/net/xfrm/ |
H A D | xfrm_ipcomp.c | 73 struct page *page; local 80 page = alloc_page(GFP_ATOMIC); 83 if (!page) 86 __skb_frag_set_page(frag, page);
|
/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_sendto.c | 68 * reserves a page for both the request and the reply header, and this 70 * on extra page for the RPCRMDA header. 117 /* Map the XDR page list */ 128 struct page *page; local 130 page = xdr->pages[page_no++]; 136 page, page_off, 160 * If head and tail use the same page, we don't need 167 /* Map another page for the tail */ 258 struct page *pag local 538 send_reply(struct svcxprt_rdma *rdma, struct svc_rqst *rqstp, struct page *page, struct rpcrdma_msg *rdma_resp, struct svc_rdma_op_ctxt *ctxt, struct svc_rdma_req_map *vec, int byte_count) argument [all...] |
H A D | svc_rdma_recvfrom.c | 61 struct page *page; local 65 /* Swap the page in the SGE with the page in argpages */ 66 page = ctxt->pages[0]; 68 rqstp->rq_pages[0] = page; 71 rqstp->rq_arg.head[0].iov_base = page_address(page); 85 page = ctxt->pages[sge_no]; 87 rqstp->rq_pages[sge_no] = page; 105 page [all...] |
/net/bridge/netfilter/ |
H A D | Kconfig | 24 brouter. See the man page for ebtables(8) and examples on the ebtables 33 local input, forwarding and local output. See the man page for 43 See the man page for ebtables(8).
|