Searched defs:page (Results 1 - 25 of 412) sorted by relevance

1234567891011>>

/drivers/gpu/drm/
H A Ddrm_cache.c36 drm_clflush_page(struct page *page) argument
41 if (unlikely(page == NULL))
44 page_virtual = kmap_atomic(page, KM_USER0);
50 static void drm_cache_flush_clflush(struct page *pages[],
69 drm_clflush_pages(struct page *pages[], unsigned long num_pages)
84 struct page *page = pages[i];
87 if (unlikely(page == NULL))
90 page_virtual = kmap_atomic(page, KM_USER
[all...]
H A Ddrm_buffer.c29 * Multipart buffer for coping data which is larger than the page size.
71 DRM_ERROR("Failed to allocate %dth page for drm"
121 " (%p) %dth page.\n",
166 int page = drm_buffer_page(buf); local
170 obj = &buf->data[page][idx];
174 memcpy(stack_obj, &buf->data[page][idx], beginsz);
176 memcpy(stack_obj + beginsz, &buf->data[page + 1][0],
H A Ddrm_scatter.c51 struct page *page; local
55 page = entry->pagelist[i];
56 if (page)
57 ClearPageReserved(page);
115 /* This also forces the mapping of COW pages, so our page list
138 /* Verify that each page points to its virtual address, and vice
/drivers/gpu/drm/i915/
H A Di915_gem_debug.c142 int page; local
157 for (page = 0; page < obj->size / PAGE_SIZE; page++) {
160 backing_map = kmap_atomic(obj->pages[page]);
163 DRM_ERROR("failed to map backing page\n");
170 page * 1024 + i);
176 page * PAGE_SIZE + i * 4),
/drivers/hwmon/pmbus/
H A Dpmbus.c31 * Find sensor groups and status registers on each page.
36 int page; local
38 /* Sensors detected on page 0 only */
75 for (page = 0; page < info->pages; page++) {
76 if (pmbus_check_word_register(client, page, PMBUS_READ_VOUT)) {
77 info->func[page] |= PMBUS_HAVE_VOUT;
78 if (pmbus_check_byte_register(client, page,
80 info->func[page] |
109 int page; local
[all...]
H A Dmax16064.c31 static int max16064_read_word_data(struct i2c_client *client, int page, int reg) argument
37 ret = pmbus_read_word_data(client, page,
41 ret = pmbus_read_word_data(client, page,
55 static int max16064_write_word_data(struct i2c_client *client, int page, argument
62 ret = pmbus_write_word_data(client, page,
66 ret = pmbus_write_word_data(client, page,
H A Dmax34440.c39 static int max34440_read_word_data(struct i2c_client *client, int page, int reg) argument
45 ret = pmbus_read_word_data(client, page,
49 ret = pmbus_read_word_data(client, page,
53 ret = pmbus_read_word_data(client, page,
68 static int max34440_write_word_data(struct i2c_client *client, int page, argument
75 ret = pmbus_write_word_data(client, page,
79 ret = pmbus_write_word_data(client, page,
83 ret = pmbus_write_word_data(client, page,
94 static int max34440_read_byte_data(struct i2c_client *client, int page, int reg) argument
99 if (page >
[all...]
/drivers/infiniband/hw/ehca/
H A Dehca_pd.c89 struct ipz_small_queue_page *page, *tmp; local
93 list_for_each_entry_safe(page, tmp, &my_pd->free[i], list) {
95 free_page(page->page);
96 kmem_cache_free(small_qp_cache, page);
/drivers/infiniband/hw/mlx4/
H A Ddoorbell.c47 struct mlx4_ib_user_db_page *page; local
53 list_for_each_entry(page, &context->db_page_list, list)
54 if (page->user_virt == (virt & PAGE_MASK))
57 page = kmalloc(sizeof *page, GFP_KERNEL);
58 if (!page) {
63 page->user_virt = (virt & PAGE_MASK);
64 page->refcnt = 0;
65 page->umem = ib_umem_get(&context->ibucontext, virt & PAGE_MASK,
67 if (IS_ERR(page
[all...]
/drivers/staging/rtl8192u/ieee80211/
H A Dinternal.h33 static inline void *crypto_kmap(struct page *page, int out) argument
35 return kmap_atomic(page, crypto_kmap_type(out));
H A Dscatterwalk.h22 struct page *page; member in struct:scatter_walk
40 return walk_in->page == walk_out->page &&
/drivers/infiniband/hw/qib/
H A Dqib_user_pages.c39 static void __qib_release_user_pages(struct page **p, size_t num_pages,
55 struct page **p, struct vm_area_struct **vma)
96 * have to bother with retries or mapping a dummy page to insure we
101 dma_addr_t qib_map_page(struct pci_dev *hwdev, struct page *page, argument
106 phys = pci_map_page(hwdev, page, offset, size, direction);
110 phys = pci_map_page(hwdev, page, offset, size, direction);
112 * FIXME: If we get 0 again, we should keep this page,
113 * map another, then free the 0 page.
122 * @start_page: the start page
[all...]
/drivers/md/
H A Ddm-sysfs.c22 char *page)
36 ret = dm_attr->show(md, page);
21 dm_attr_show(struct kobject *kobj, struct attribute *attr, char *page) argument
/drivers/mtd/nand/
H A Dbcm_umi_bch.c25 struct nand_chip *chip, uint8_t *buf, int page);
48 ** We treat the OOB for a 2K page as if it were 4 512 byte oobs,
70 /* We treat the OOB for a 4K page as if it were 8 512 byte oobs,
102 * bcm_umi_bch_read_page_hwecc - hardware ecc based page read function
110 int page)
123 /* Seek to page location within sector */
187 * bcm_umi_bch_write_page_hwecc - hardware ecc based page write function
108 bcm_umi_bch_read_page_hwecc(struct mtd_info *mtd, struct nand_chip *chip, uint8_t * buf, int page) argument
/drivers/usb/storage/
H A Dprotocol.c143 * each page has to be kmap()'ed separately. If the page is already
145 * If the page is not directly accessible -- such as a user buffer
154 * include multiple pages. Find the initial page structure
155 * and the starting offset within the page, and update
160 struct page *page = sg_page(sg) + local
178 * s-g entry. For each page: call kmap(), do the
183 unsigned char *ptr = kmap(page);
189 kunmap(page);
[all...]
/drivers/auxdisplay/
H A Dks0108.c101 void ks0108_page(unsigned char page) argument
103 ks0108_writedata(min(page,(unsigned char)7) | bit(3) | bit(4) | bit(5) | bit(7));
/drivers/base/
H A Ddma-coherent.c171 int page = (vaddr - mem->virt_base) >> PAGE_SHIFT; local
173 bitmap_release_region(mem->bitmap, page, order);
/drivers/dma/
H A Diovlock.c53 struct page **pages;
71 /* single kmalloc for pinned list, page_list[], and the page arrays */
74 + (iovec_pages_used * sizeof (struct page*)), GFP_KERNEL);
78 /* list of pages starts right after the page list array */
79 pages = (struct page **) &local_list->page_list[nr_iovecs];
177 /* break up copies to not cross page boundary */
213 struct dma_pinned_list *pinned_list, struct page *page,
226 u8 *vaddr = kmap(page);
228 kunmap(page);
212 dma_memcpy_pg_to_iovec(struct dma_chan *chan, struct iovec *iov, struct dma_pinned_list *pinned_list, struct page *page, unsigned int offset, size_t len) argument
[all...]
/drivers/gpu/drm/ttm/
H A Dttm_agp_backend.c63 struct page *page = ttm->pages[i]; local
65 if (!page)
66 page = ttm->dummy_read_page;
68 mem->pages[mem->page_count++] = page;
114 struct page *dummy_read_page)
/drivers/infiniband/hw/ipath/
H A Dipath_user_pages.c41 static void __ipath_release_user_pages(struct page **p, size_t num_pages,
57 struct page **p, struct vm_area_struct **vma)
101 * have to bother with retries or mapping a dummy page to insure we
106 dma_addr_t ipath_map_page(struct pci_dev *hwdev, struct page *page, argument
111 phys = pci_map_page(hwdev, page, offset, size, direction);
115 phys = pci_map_page(hwdev, page, offset, size, direction);
117 * FIXME: If we get 0 again, we should keep this page,
118 * map another, then free the 0 page.
141 * FIXME: If we get 0 again, we should keep this page,
[all...]
/drivers/mca/
H A Dmca-proc.c35 static int get_mca_info_helper(struct mca_device *mca_dev, char *page, int len) argument
40 len += sprintf(page+len, "%02x ",
42 len += sprintf(page+len, " %s\n", mca_dev ? mca_dev->name : "");
46 static int get_mca_info(char *page, char **start, off_t off, argument
58 len += sprintf(page+len, "Slot %d: ", i+1);
59 len = get_mca_info_helper(mca_dev, page, len);
65 len += sprintf(page+len, "Video : ");
66 len = get_mca_info_helper(mca_dev, page, len);
71 len += sprintf(page+len, "SCSI : ");
72 len = get_mca_info_helper(mca_dev, page, le
136 get_mca_machine_info(char* page, char **start, off_t off, int count, int *eof, void *data) argument
153 mca_read_proc(char *page, char **start, off_t off, int count, int *eof, void *data) argument
[all...]
/drivers/misc/
H A Duid_stat.c55 static int tcp_snd_read_proc(char *page, char **start, off_t off, argument
60 char *p = page;
67 len = (p - page) - off;
69 *start = page + off;
73 static int tcp_rcv_read_proc(char *page, char **start, off_t off, argument
78 char *p = page;
85 len = (p - page) - off;
87 *start = page + off;
/drivers/mtd/maps/
H A Dvmax301.c9 to the lower 8k of the device the second is paged. Writing a 16 bit page
10 value to anywhere in the first 8k will cause the second 8k to page around.
13 of flash that is smart enough to copy itself down, page in the rest of
41 static void __vmax301_page(struct map_info *map, unsigned long page) argument
43 writew(page, map->map_priv_2 - WINDOW_LENGTH);
44 map->map_priv_1 = page;
50 unsigned long page = (ofs >> WINDOW_SHIFT); local
51 if (map->map_priv_1 != page)
52 __vmax301_page(map, page);
/drivers/target/tcm_fc/
H A Dtfc_io.c77 struct page *page = NULL; local
98 page = sg_page(sg);
109 page = sg_page(sg);
138 BUG_ON(!page);
139 get_page(page);
142 page, off_in_page, tlen);
146 PAGE_SIZE << compound_order(page);
148 BUG_ON(!page);
149 from = kmap_atomic(page
201 struct page *page = NULL; local
[all...]
/drivers/atm/
H A Dadummy.c123 adummy_proc_read(struct atm_dev *dev, loff_t *pos, char *page) argument
128 return sprintf(page, "version %s\n", DRV_VERSION);

Completed in 687 milliseconds

1234567891011>>