/net/ceph/ |
H A D | pagelist.c | 10 struct page *page = list_entry(pl->head.prev, struct page, lru); local 11 kunmap(page); 20 struct page *page = list_first_entry(&pl->head, struct page, local 22 list_del(&page->lru); 23 __free_page(page); 32 struct page *pag local 86 struct page *page = __page_cache_alloc(GFP_NOFS); local 100 struct page *page = list_first_entry(&pl->free_list, local 129 struct page *page; local [all...] |
H A D | messenger.c | 191 static struct page *zero_page; /* used in certain error cases */ 522 static int ceph_tcp_recvpage(struct socket *sock, struct page *page, argument 530 kaddr = kmap(page); 533 kunmap(page); 559 static int ceph_tcp_sendpage(struct socket *sock, struct page *page, argument 565 ret = kernel_sendpage(sock, page, offset, size, flags); 789 static struct page *ceph_msg_data_bio_next(struct ceph_msg_data_cursor *cursor, 868 * For a page arra 945 struct page *page; local 1078 struct page *page; local 1452 ceph_crc32c_page(u32 crc, struct page *page, unsigned int page_offset, unsigned int length) argument 1494 struct page *page; local 2175 struct page *page; local [all...] |
/net/mac802154/ |
H A D | mac_cmd.c | 38 u8 channel, u8 page, 48 mac802154_dev_set_page_channel(dev, page, channel); 36 mac802154_mlme_start_req(struct net_device *dev, struct ieee802154_addr *addr, u8 channel, u8 page, u8 bcn_ord, u8 sf_ord, u8 pan_coord, u8 blx, u8 coord_realign) argument
|
H A D | monitor.c | 41 u8 chan, page; local 47 page = priv->hw->phy->current_page; 52 if (WARN_ON(page >= WPAN_NUM_PAGES) || 60 return mac802154_tx(priv->hw, skb, page, chan); 115 priv->page = 0;
|
H A D | tx.c | 42 u8 page; member in struct:xmit_work 53 xw->priv->phy->current_page != xw->page) { 55 xw->page, 63 xw->priv->phy->current_page = xw->page; 85 u8 page, u8 chan) 90 if (!(priv->phy->channels_supported[page] & (1 << chan))) { 125 work->page = page; 84 mac802154_tx(struct mac802154_priv *priv, struct sk_buff *skb, u8 page, u8 chan) argument
|
H A D | mac802154.h | 81 u8 page; member in struct:mac802154_sub_if_data 106 u8 page, u8 chan); 114 void mac802154_dev_set_page_channel(struct net_device *dev, u8 page, u8 chan);
|
H A D | mib.c | 180 res = hw->ops->set_channel(&hw->hw, priv->page, priv->chan); 185 priv->hw->phy->current_page = priv->page; 192 void mac802154_dev_set_page_channel(struct net_device *dev, u8 page, u8 chan) argument 200 priv->page = page; 206 priv->hw->phy->current_page != priv->page) {
|
H A D | wpan.c | 316 u8 chan, page; local 322 page = priv->page; 326 page >= WPAN_NUM_PAGES || 336 return mac802154_tx(priv->hw, skb, page, chan); 376 priv->page = 0;
|
/net/core/ |
H A D | user_dma.c | 82 struct page *page = skb_frag_page(frag); local 87 cookie = dma_memcpy_pg_to_iovec(chan, to, pinned_list, page,
|
H A D | datagram.c | 345 struct page *page = skb_frag_page(frag); local 349 vaddr = kmap(page); 352 kunmap(page); 431 struct page *page = skb_frag_page(frag); local 435 vaddr = kmap(page); 438 kunmap(page); 521 struct page *page local 607 struct page *page = skb_frag_page(frag); local [all...] |
H A D | skbuff.c | 80 put_page(buf->page); 86 get_page(buf->page); 365 * containing page->_count every time we allocate a fragment. 380 if (unlikely(!nc->frag.page)) { 387 nc->frag.page = alloc_pages(gfp, order); 388 if (likely(nc->frag.page)) 395 atomic_set(&nc->frag.page->_count, NETDEV_PAGECNT_MAX_BIAS); 402 if ((atomic_read(&nc->frag.page->_count) == nc->pagecnt_bias) || 403 atomic_sub_and_test(nc->pagecnt_bias, &nc->frag.page->_count)) 408 data = page_address(nc->frag.page) 474 skb_add_rx_frag(struct sk_buff *skb, int i, struct page *page, int off, int size, unsigned int truesize) argument 818 struct page *page, *head = NULL; local 1668 linear_to_page(struct page *page, unsigned int *len, unsigned int *offset, struct sock *sk) argument 1687 spd_can_coalesce(const struct splice_pipe_desc *spd, struct page *page, unsigned int offset) argument 1700 spd_fill_page(struct splice_pipe_desc *spd, struct pipe_inode_info *pipe, struct page *page, unsigned int *len, unsigned int offset, bool linear, struct sock *sk) argument 1727 __splice_segment(struct page *page, unsigned int poff, unsigned int plen, unsigned int *off, unsigned int *len, struct splice_pipe_desc *spd, bool linear, struct sock *sk, struct pipe_inode_info *pipe) argument 2995 struct page *page = virt_to_head_page(skb->head); local 3456 struct page *page; local [all...] |
H A D | sock.c | 1741 struct page *page; local 1743 page = alloc_pages(sk->sk_allocation, 0); 1744 if (!page) { 1752 page, 0, 1800 if (pfrag->page) { 1801 if (atomic_read(&pfrag->page->_count) == 1) { 1807 put_page(pfrag->page); 1818 pfrag->page = alloc_pages(gfp, order); 1819 if (likely(pfrag->page)) { 2116 sock_no_sendpage(struct socket *sock, struct page *page, int offset, size_t size, int flags) argument [all...] |
/net/rds/ |
H A D | page.c | 41 struct page *r_page; 56 int rds_page_copy_user(struct page *page, unsigned long offset, argument 63 addr = kmap(page); 71 kunmap(page); 87 * If @bytes is at least a full page then this just returns a page from 90 * If @bytes is a partial page then this stores the unused region of the 91 * page in a per-cpu structure. Future partial-page allocation 102 struct page *page; local [all...] |
H A D | rdma.c | 49 * get the number of pages by looking at the page indices that the start and 158 struct page **pages, int write) 178 struct page **pages = NULL; 207 pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL); 232 * Pin the pages that make up the user buffer and transfer the page 234 * the whole region after transferring the partial page references 235 * to the sg array so that we can have one page ref cleanup path. 239 * the zero page. 281 * map page aligned regions. So we keep the offset, and build 448 struct page *pag local 467 struct page *page = sg_page(ao->op_sg); local 769 struct page *page = NULL; local [all...] |
H A D | ib_rdma.c | 484 struct page *page = sg_page(&ibmr->sg[i]); local 489 set_page_dirty(page); 490 put_page(page);
|
/net/xfrm/ |
H A D | xfrm_ipcomp.c | 73 struct page *page; local 80 page = alloc_page(GFP_ATOMIC); 83 if (!page) 86 __skb_frag_set_page(frag, page);
|
/net/atm/ |
H A D | mpoa_proc.c | 209 char *page, *p; local 218 page = (char *)__get_free_page(GFP_KERNEL); 219 if (!page) 222 for (p = page, len = 0; len < nbytes; p++, len++) { 224 free_page((unsigned long)page); 233 if (!parse_qos(page)) 234 printk("mpoa: proc_mpc_write: could not parse '%s'\n", page); 236 free_page((unsigned long)page);
|
H A D | proc.c | 380 unsigned long page; local 385 page = get_zeroed_page(GFP_KERNEL); 386 if (!page) 392 length = dev->ops->proc_read(dev, pos, (char *)page); 397 if (copy_to_user(buf, (char *)page, length)) 401 free_page(page);
|
/net/ieee802154/ |
H A D | nl-mac.c | 195 u8 status, u8 scan_type, u32 unscanned, u8 page, 213 nla_put_u8(msg, IEEE802154_ATTR_PAGE, page) || 317 u8 page; local 346 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]); 348 page = 0; 352 page, 442 u8 page; local 476 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]); 478 page = 0; 487 ret = ieee802154_mlme_ops(dev)->start_req(dev, &addr, channel, page, 194 ieee802154_nl_scan_confirm(struct net_device *dev, u8 status, u8 scan_type, u32 unscanned, u8 page, u8 *edl ) argument 502 u8 page; local [all...] |
/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_recvfrom.c | 61 struct page *page; local 65 /* Swap the page in the SGE with the page in argpages */ 66 page = ctxt->pages[0]; 68 rqstp->rq_pages[0] = page; 71 rqstp->rq_arg.head[0].iov_base = page_address(page); 85 page = ctxt->pages[sge_no]; 87 rqstp->rq_pages[sge_no] = page; 105 page [all...] |
H A D | svc_rdma_sendto.c | 68 * reserves a page for both the request and the reply header, and this 70 * on extra page for the RPCRMDA header. 117 /* Map the XDR page list */ 128 struct page *page; local 130 page = xdr->pages[page_no++]; 136 page, page_off, 160 * If head and tail use the same page, we don't need 167 /* Map another page for the tail */ 258 struct page *pag local 538 send_reply(struct svcxprt_rdma *rdma, struct svc_rqst *rqstp, struct page *page, struct rpcrdma_msg *rdma_resp, struct svc_rdma_op_ctxt *ctxt, struct svc_rdma_req_map *vec, int byte_count) argument [all...] |
H A D | svc_rdma_transport.c | 474 struct page *svc_rdma_get_page(void) 476 struct page *page; local 478 while ((page = alloc_page(GFP_KERNEL)) == NULL) { 484 return page; 491 struct page *page; local 502 page = svc_rdma_get_page(); 503 ctxt->pages[sge_no] = page; 505 page, [all...] |
/net/ipv4/ |
H A D | ip_output.c | 739 csum_page(struct page *page, int offset, int copy) argument 743 kaddr = kmap(page); 745 kunmap(page); 1000 if (!skb_can_coalesce(skb, i, pfrag->page, 1006 __skb_fill_page_desc(skb, i, pfrag->page, 1009 get_page(pfrag->page); 1013 page_address(pfrag->page) + pfrag->offset, 1079 * until ip_push_pending_frames() is called. Each piece can be a page 1080 * or non-page dat 1113 ip_append_page(struct sock *sk, struct flowi4 *fl4, struct page *page, int offset, size_t size, int flags) argument [all...] |
H A D | af_inet.c | 791 ssize_t inet_sendpage(struct socket *sock, struct page *page, int offset, argument 804 return sk->sk_prot->sendpage(sk, page, offset, size, flags); 805 return sock_no_sendpage(sock, page, offset, size, flags);
|
/net/netfilter/ |
H A D | nfnetlink_queue_core.c | 235 struct page *page; local 250 page = virt_to_head_page(from->head); 251 offset = from->data - (unsigned char *)page_address(page); 252 __skb_fill_page_desc(to, 0, page, offset, plen); 253 get_page(page);
|