Searched refs:page (Results 176 - 200 of 649) sorted by relevance

1234567891011>>

/drivers/md/
H A Draid10.h66 struct page *tmppage;
H A Dmd.h29 /* Bad block numbers are stored sorted in a single page.
54 struct page *sb_page, *bb_page;
120 u64 *page; /* badblock list */ member in struct:md_rdev::badblocks
576 static inline void safe_put_page(struct page *p)
598 sector_t sector, int size, struct page *page);
601 struct page *page, int rw, bool metadata_op);
/drivers/staging/zsmalloc/
H A Dzsmalloc_int.h23 * span more than 1 page which avoids complex case of mapping 2 pages simply
40 * page <PFN> it is stored in, so for each sub-page belonging
68 * On systems with 4K page size, this gives 254 size classes! There is a
70 * - Large number of size classes is potentially wasteful as free page are
97 * We assign a page to ZS_ALMOST_EMPTY fullness group when:
134 struct page *fullness_list[_ZS_NR_FULLNESS_GROUPS];
/drivers/xen/xenbus/
H A Dxenbus_dev_backend.c10 #include <xen/page.h>
/drivers/xen/xenfs/
H A Dxenstored.c6 #include <xen/page.h>
/drivers/mtd/
H A Dmtdswap.c168 * page from the current block and is then pre-empted by the GC
611 static int mtdswap_map_free_block(struct mtdswap_dev *d, unsigned int page, argument
646 d->revmap[*block] = page;
664 unsigned int page, unsigned int *bp, int gc_context)
678 ret = mtdswap_map_free_block(d, page, bp);
731 unsigned int page, retries; local
734 page = d->revmap[oldblock];
761 ret = mtdswap_write_block(d, d->page_buf, page, newblock, 1);
763 d->page_data[page] = BLOCK_ERROR;
769 d->page_data[page]
663 mtdswap_write_block(struct mtdswap_dev *d, char *buf, unsigned int page, unsigned int *bp, int gc_context) argument
1069 mtdswap_writesect(struct mtd_blktrans_dev *dev, unsigned long page, char *buf) argument
1127 mtdswap_readsect(struct mtd_blktrans_dev *dev, unsigned long page, char *buf) argument
1195 unsigned long page; local
[all...]
/drivers/scsi/fcoe/
H A Dfcoe_transport.c340 * fcoe_get_paged_crc_eof() - Allocate a page to be used for the trailer CRC
345 * This routine allocates a page for frame trailers. The page is re-used if
347 * enough buffer left a new page is allocated for the trailer. Reference to
348 * the page from this function as well as the skbs using the page fragments
349 * ensure that the page is freed at the appropriate time.
356 struct page *page; local
358 page
[all...]
/drivers/target/tcm_fc/
H A Dtfc_conf.c143 char *page)
148 return ft_wwn_show(&acl->node_auth.port_name, page);
153 const char *page,
159 return ft_wwn_store(&acl->node_auth.port_name, page, count);
166 char *page)
171 return ft_wwn_show(&acl->node_auth.node_name, page);
176 const char *page,
182 return ft_wwn_store(&acl->node_auth.node_name, page, count);
448 char *page)
450 return sprintf(page, "TC
141 ft_nacl_show_port_name( struct se_node_acl *se_nacl, char *page) argument
151 ft_nacl_store_port_name( struct se_node_acl *se_nacl, const char *page, size_t count) argument
164 ft_nacl_show_node_name( struct se_node_acl *se_nacl, char *page) argument
174 ft_nacl_store_node_name( struct se_node_acl *se_nacl, const char *page, size_t count) argument
446 ft_wwn_show_attr_version( struct target_fabric_configfs *tf, char *page) argument
[all...]
/drivers/gpu/drm/
H A Ddrm_agpsupport.c318 int page; local
326 page = (request->offset + PAGE_SIZE - 1) / PAGE_SIZE;
327 if ((retcode = drm_bind_agp(entry->memory, page)))
329 entry->bound = dev->agp->base + (page << PAGE_SHIFT);
435 struct page **pages,
/drivers/media/dvb/ttpci/
H A Dav7110_ca.c28 * the project's page is at http://www.linuxtv.org/
155 u8 *page = (u8 *)__get_free_page(GFP_USER); local
158 if (!page)
166 if (copy_from_user(page, buf, count))
183 res = dvb_ringbuffer_write(cibuf, page, count);
185 free_page((unsigned long)page);
/drivers/staging/keucr/
H A Dscsiglue.c424 struct page *page = sg_page(sg) + local
442 unsigned char *ptr = kmap(page);
448 kunmap(page);
450 /* Start at the beginning of the next page */
452 ++page;
/drivers/net/ethernet/dec/
H A Dewrk3.c694 u_char csr, page; local
721 for (page = 1; page < lp->mPage; page++) { /* Write the free page numbers */
722 outb(page, EWRK3_FMQ); /* to the Free Memory Queue */
766 ** Writes a socket buffer to the free page queue
774 u_char page; local
790 ** Get a free page from the FMQ
792 if ((page
940 u_char page; local
[all...]
/drivers/gpu/drm/ttm/
H A Dttm_page_alloc_dma.c32 * - Tracks whether the page is UC, WB or cached (and reverts to WB
56 #define NUM_PAGES_TO_ALLOC (PAGE_SIZE/sizeof(struct page *))
120 * The accounting page keeping track of the allocated page along with
123 * @vaddr: The virtual address of the page
124 * @dma: The bus address of the page. If the page is not allocated
130 struct page *p;
271 static int set_pages_array_wb(struct page **pages, int addrinarray)
282 static int set_pages_array_wc(struct page **page
[all...]
H A Dttm_bo_vm.c83 struct page *page; local
180 /* Allocate all page at once, most common usage */
189 * first page.
195 page = ttm->pages[page_offset];
196 if (unlikely(!page && i == 0)) {
199 } else if (unlikely(!page)) {
202 pfn = page_to_pfn(page);
/drivers/s390/block/
H A Ddasd_eckd.c2548 /* Track based I/O needs IDAWs for each page, and not just for
2559 /* on 31-bit we need space for two 32 bit addresses per page
2906 * A tidaw can address 4k of memory, but must not cross page boundaries
2908 * blk_queue_segment_boundary to page boundaries and
2909 * blk_max_segment_size to page size when setting up the request queue.
3108 * Raw track based I/O needs IDAWs for each page,
3813 * Dump the range of CCWs into 'page' buffer
3817 dasd_eckd_dump_ccw_range(struct ccw1 *from, struct ccw1 *to, char *page) argument
3824 len += sprintf(page + len, KERN_ERR PRINTK_HEADER
3836 if (count % 8 == 0) len += sprintf(page
3874 char *page; local
3977 char *page; local
[all...]
/drivers/net/wireless/hostap/
H A Dhostap_download.c89 u16 page, offset; local
93 page = addr >> 7;
96 HFA384X_OUTW(page, HFA384X_AUXPAGE_OFF);
120 u16 page, offset; local
124 page = addr >> 7;
127 HFA384X_OUTW(page, HFA384X_AUXPAGE_OFF);
547 __le16 page; member in struct:__anon3483
570 le16_to_cpu(dlbuffer.page),
573 bufaddr = (le16_to_cpu(dlbuffer.page) << 7) + le16_to_cpu(dlbuffer.offset);
/drivers/net/
H A Dxen-netfront.c49 #include <xen/page.h>
60 struct page *page; member in struct:netfront_cb
239 struct page *page; local
266 page = alloc_page(GFP_ATOMIC | __GFP_NOWARN);
267 if (!page) {
279 __skb_fill_page_desc(skb, 0, page, 0, 0);
426 /* While the header overlaps a page boundary (including being
427 larger than a page), spli
870 struct page *page = NETFRONT_SKB_CB(skb)->page; local
1148 const struct page *page = local
1412 xennet_end_access(int ref, void *page) argument
1666 const struct page *page; local
[all...]
/drivers/net/wireless/iwlwifi/
H A Diwl-trans-pcie-rx.c213 BUG_ON(rxb && rxb->page);
259 struct page *page; local
278 page = alloc_pages(gfp_mask,
280 if (!page) {
303 __free_pages(page, hw_params(trans).rx_page_order);
312 BUG_ON(rxb->page);
313 rxb->page = page;
315 rxb->page_dma = dma_map_page(trans->dev, page,
[all...]
/drivers/staging/omapdrm/
H A Domap_gem.c28 struct page ** _drm_gem_get_pages(struct drm_gem_object *obj, gfp_t gfpmask);
29 void _drm_gem_put_pages(struct drm_gem_object *obj, struct page **pages,
87 struct page **pages;
107 * objects can be packed in a page, and not waste GPU virtual address
121 static int get_pages(struct drm_gem_object *obj, struct page ***pages);
126 * they are not necessarily page aligned, we reserve one or more small
128 * can create a second page-aligned mapping of parts of the buffer
140 pgoff_t obj_pgoff; /* page offset of obj currently
216 struct page **pages;
352 struct page *page
[all...]
/drivers/misc/ibmasm/
H A Dibmasmfs.c511 unsigned char *page; local
523 page = (unsigned char *)__get_free_page(GFP_KERNEL);
524 if (!page)
528 len = sprintf(page, "%d\n", value);
530 if (copy_to_user(buf, page, len)) {
538 free_page((unsigned long)page);
/drivers/mmc/host/
H A Dsdricoh_cs.c322 struct page *page; local
324 page = sg_page(data->sg);
326 buf = kmap(page) + data->sg->offset + (len * i);
330 kunmap(page);
331 flush_dcache_page(page);
/drivers/mtd/onenand/
H A Donenand_sim.c373 int block = -1, page = -1, bufferram = -1; local
398 page = (int) readw(this->base + ONENAND_REG_START_ADDRESS8);
399 page = (page >> ONENAND_FPA_SHIFT);
410 if (page != -1)
411 offset += page << this->page_shift;
478 buffer_size = 0x0400; /* 1KiB page */
480 buffer_size = 0x0800; /* 2KiB page */
/drivers/net/wan/
H A Dc101.c76 u8 page; member in struct:card_s
109 return card->page;
112 static inline void openwin(card_t *card, u8 page) argument
114 card->page = page;
115 writeb(page, card->win0base + C101_PAGE);
/drivers/net/ethernet/sun/
H A Dcassini.c30 * page-based RX descriptor engine with separate completion rings
34 * RX is handled by page sized buffers that are attached as fragments to
43 * -- on page reclamation, the driver swaps the page with a spare page.
44 * if that page is still in use, it frees its reference to that page,
45 * and allocates a new page for use. otherwise, it just recycles the
46 * the page.
169 * max mtu == 2 * page siz
470 cas_page_free(struct cas *cp, cas_page_t *page) argument
492 cas_page_t *page; local
575 cas_page_t *page = list_entry(elem, cas_page_t, list); local
1375 cas_page_t *page = cp->rx_pages[1][index]; local
1426 cas_page_t *page = cas_page_swap(cp, 0, i); local
1961 struct cas_page *page; local
2224 cas_page_t **page = cp->rx_pages[ring]; local
3418 struct page *page = alloc_pages(GFP_ATOMIC, local
3947 cas_page_t **page = cp->rx_pages[ring]; local
3987 cas_page_t **page = cp->rx_pages[ring]; local
[all...]
/drivers/atm/
H A Dadummy.c123 adummy_proc_read(struct atm_dev *dev, loff_t *pos, char *page) argument
128 return sprintf(page, "version %s\n", DRV_VERSION);

Completed in 717 milliseconds

1234567891011>>