Searched defs:page (Results 1 - 25 of 33) sorted by relevance

12

/net/ceph/
H A Dpagelist.c10 struct page *page = list_entry(pl->head.prev, struct page, lru); local
11 kunmap(page);
20 struct page *page = list_first_entry(&pl->head, struct page, local
22 list_del(&page->lru);
23 __free_page(page);
32 struct page *pag local
86 struct page *page = __page_cache_alloc(GFP_NOFS); local
100 struct page *page = list_first_entry(&pl->free_list, local
129 struct page *page; local
[all...]
H A Dmessenger.c191 static struct page *zero_page; /* used in certain error cases */
522 static int ceph_tcp_recvpage(struct socket *sock, struct page *page, argument
530 kaddr = kmap(page);
533 kunmap(page);
559 static int ceph_tcp_sendpage(struct socket *sock, struct page *page, argument
565 ret = kernel_sendpage(sock, page, offset, size, flags);
789 static struct page *ceph_msg_data_bio_next(struct ceph_msg_data_cursor *cursor,
868 * For a page arra
945 struct page *page; local
1078 struct page *page; local
1452 ceph_crc32c_page(u32 crc, struct page *page, unsigned int page_offset, unsigned int length) argument
1494 struct page *page; local
2175 struct page *page; local
[all...]
/net/mac802154/
H A Dmac_cmd.c38 u8 channel, u8 page,
48 mac802154_dev_set_page_channel(dev, page, channel);
36 mac802154_mlme_start_req(struct net_device *dev, struct ieee802154_addr *addr, u8 channel, u8 page, u8 bcn_ord, u8 sf_ord, u8 pan_coord, u8 blx, u8 coord_realign) argument
H A Dmonitor.c41 u8 chan, page; local
47 page = priv->hw->phy->current_page;
52 if (WARN_ON(page >= WPAN_NUM_PAGES) ||
60 return mac802154_tx(priv->hw, skb, page, chan);
115 priv->page = 0;
H A Dtx.c42 u8 page; member in struct:xmit_work
53 xw->priv->phy->current_page != xw->page) {
55 xw->page,
63 xw->priv->phy->current_page = xw->page;
85 u8 page, u8 chan)
90 if (!(priv->phy->channels_supported[page] & (1 << chan))) {
125 work->page = page;
84 mac802154_tx(struct mac802154_priv *priv, struct sk_buff *skb, u8 page, u8 chan) argument
H A Dmac802154.h81 u8 page; member in struct:mac802154_sub_if_data
106 u8 page, u8 chan);
114 void mac802154_dev_set_page_channel(struct net_device *dev, u8 page, u8 chan);
H A Dmib.c180 res = hw->ops->set_channel(&hw->hw, priv->page, priv->chan);
185 priv->hw->phy->current_page = priv->page;
192 void mac802154_dev_set_page_channel(struct net_device *dev, u8 page, u8 chan) argument
200 priv->page = page;
206 priv->hw->phy->current_page != priv->page) {
H A Dwpan.c316 u8 chan, page; local
322 page = priv->page;
326 page >= WPAN_NUM_PAGES ||
336 return mac802154_tx(priv->hw, skb, page, chan);
376 priv->page = 0;
/net/core/
H A Duser_dma.c82 struct page *page = skb_frag_page(frag); local
87 cookie = dma_memcpy_pg_to_iovec(chan, to, pinned_list, page,
H A Ddatagram.c345 struct page *page = skb_frag_page(frag); local
349 vaddr = kmap(page);
352 kunmap(page);
431 struct page *page = skb_frag_page(frag); local
435 vaddr = kmap(page);
438 kunmap(page);
521 struct page *page local
607 struct page *page = skb_frag_page(frag); local
[all...]
H A Dskbuff.c80 put_page(buf->page);
86 get_page(buf->page);
365 * containing page->_count every time we allocate a fragment.
380 if (unlikely(!nc->frag.page)) {
387 nc->frag.page = alloc_pages(gfp, order);
388 if (likely(nc->frag.page))
395 atomic_set(&nc->frag.page->_count, NETDEV_PAGECNT_MAX_BIAS);
402 if ((atomic_read(&nc->frag.page->_count) == nc->pagecnt_bias) ||
403 atomic_sub_and_test(nc->pagecnt_bias, &nc->frag.page->_count))
408 data = page_address(nc->frag.page)
474 skb_add_rx_frag(struct sk_buff *skb, int i, struct page *page, int off, int size, unsigned int truesize) argument
818 struct page *page, *head = NULL; local
1668 linear_to_page(struct page *page, unsigned int *len, unsigned int *offset, struct sock *sk) argument
1687 spd_can_coalesce(const struct splice_pipe_desc *spd, struct page *page, unsigned int offset) argument
1700 spd_fill_page(struct splice_pipe_desc *spd, struct pipe_inode_info *pipe, struct page *page, unsigned int *len, unsigned int offset, bool linear, struct sock *sk) argument
1727 __splice_segment(struct page *page, unsigned int poff, unsigned int plen, unsigned int *off, unsigned int *len, struct splice_pipe_desc *spd, bool linear, struct sock *sk, struct pipe_inode_info *pipe) argument
2995 struct page *page = virt_to_head_page(skb->head); local
3456 struct page *page; local
[all...]
H A Dsock.c1741 struct page *page; local
1743 page = alloc_pages(sk->sk_allocation, 0);
1744 if (!page) {
1752 page, 0,
1800 if (pfrag->page) {
1801 if (atomic_read(&pfrag->page->_count) == 1) {
1807 put_page(pfrag->page);
1818 pfrag->page = alloc_pages(gfp, order);
1819 if (likely(pfrag->page)) {
2116 sock_no_sendpage(struct socket *sock, struct page *page, int offset, size_t size, int flags) argument
[all...]
/net/rds/
H A Dpage.c41 struct page *r_page;
56 int rds_page_copy_user(struct page *page, unsigned long offset, argument
63 addr = kmap(page);
71 kunmap(page);
87 * If @bytes is at least a full page then this just returns a page from
90 * If @bytes is a partial page then this stores the unused region of the
91 * page in a per-cpu structure. Future partial-page allocation
102 struct page *page; local
[all...]
H A Drdma.c49 * get the number of pages by looking at the page indices that the start and
158 struct page **pages, int write)
178 struct page **pages = NULL;
207 pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL);
232 * Pin the pages that make up the user buffer and transfer the page
234 * the whole region after transferring the partial page references
235 * to the sg array so that we can have one page ref cleanup path.
239 * the zero page.
281 * map page aligned regions. So we keep the offset, and build
448 struct page *pag local
467 struct page *page = sg_page(ao->op_sg); local
769 struct page *page = NULL; local
[all...]
H A Dib_rdma.c484 struct page *page = sg_page(&ibmr->sg[i]); local
489 set_page_dirty(page);
490 put_page(page);
/net/xfrm/
H A Dxfrm_ipcomp.c73 struct page *page; local
80 page = alloc_page(GFP_ATOMIC);
83 if (!page)
86 __skb_frag_set_page(frag, page);
/net/atm/
H A Dmpoa_proc.c209 char *page, *p; local
218 page = (char *)__get_free_page(GFP_KERNEL);
219 if (!page)
222 for (p = page, len = 0; len < nbytes; p++, len++) {
224 free_page((unsigned long)page);
233 if (!parse_qos(page))
234 printk("mpoa: proc_mpc_write: could not parse '%s'\n", page);
236 free_page((unsigned long)page);
H A Dproc.c380 unsigned long page; local
385 page = get_zeroed_page(GFP_KERNEL);
386 if (!page)
392 length = dev->ops->proc_read(dev, pos, (char *)page);
397 if (copy_to_user(buf, (char *)page, length))
401 free_page(page);
/net/ieee802154/
H A Dnl-mac.c195 u8 status, u8 scan_type, u32 unscanned, u8 page,
213 nla_put_u8(msg, IEEE802154_ATTR_PAGE, page) ||
317 u8 page; local
346 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]);
348 page = 0;
352 page,
442 u8 page; local
476 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]);
478 page = 0;
487 ret = ieee802154_mlme_ops(dev)->start_req(dev, &addr, channel, page,
194 ieee802154_nl_scan_confirm(struct net_device *dev, u8 status, u8 scan_type, u32 unscanned, u8 page, u8 *edl ) argument
502 u8 page; local
[all...]
/net/sunrpc/xprtrdma/
H A Dsvc_rdma_recvfrom.c61 struct page *page; local
65 /* Swap the page in the SGE with the page in argpages */
66 page = ctxt->pages[0];
68 rqstp->rq_pages[0] = page;
71 rqstp->rq_arg.head[0].iov_base = page_address(page);
85 page = ctxt->pages[sge_no];
87 rqstp->rq_pages[sge_no] = page;
105 page
[all...]
H A Dsvc_rdma_sendto.c68 * reserves a page for both the request and the reply header, and this
70 * on extra page for the RPCRMDA header.
117 /* Map the XDR page list */
128 struct page *page; local
130 page = xdr->pages[page_no++];
136 page, page_off,
160 * If head and tail use the same page, we don't need
167 /* Map another page for the tail */
258 struct page *pag local
538 send_reply(struct svcxprt_rdma *rdma, struct svc_rqst *rqstp, struct page *page, struct rpcrdma_msg *rdma_resp, struct svc_rdma_op_ctxt *ctxt, struct svc_rdma_req_map *vec, int byte_count) argument
[all...]
H A Dsvc_rdma_transport.c474 struct page *svc_rdma_get_page(void)
476 struct page *page; local
478 while ((page = alloc_page(GFP_KERNEL)) == NULL) {
484 return page;
491 struct page *page; local
502 page = svc_rdma_get_page();
503 ctxt->pages[sge_no] = page;
505 page,
[all...]
/net/ipv4/
H A Dip_output.c739 csum_page(struct page *page, int offset, int copy) argument
743 kaddr = kmap(page);
745 kunmap(page);
1000 if (!skb_can_coalesce(skb, i, pfrag->page,
1006 __skb_fill_page_desc(skb, i, pfrag->page,
1009 get_page(pfrag->page);
1013 page_address(pfrag->page) + pfrag->offset,
1079 * until ip_push_pending_frames() is called. Each piece can be a page
1080 * or non-page dat
1113 ip_append_page(struct sock *sk, struct flowi4 *fl4, struct page *page, int offset, size_t size, int flags) argument
[all...]
H A Daf_inet.c791 ssize_t inet_sendpage(struct socket *sock, struct page *page, int offset, argument
804 return sk->sk_prot->sendpage(sk, page, offset, size, flags);
805 return sock_no_sendpage(sock, page, offset, size, flags);
/net/netfilter/
H A Dnfnetlink_queue_core.c235 struct page *page; local
250 page = virt_to_head_page(from->head);
251 offset = from->data - (unsigned char *)page_address(page);
252 __skb_fill_page_desc(to, 0, page, offset, plen);
253 get_page(page);

Completed in 405 milliseconds

12