/drivers/gpu/drm/ |
H A D | drm_cache.c | 36 drm_clflush_page(struct page *page) argument 41 if (unlikely(page == NULL)) 44 page_virtual = kmap_atomic(page, KM_USER0); 50 static void drm_cache_flush_clflush(struct page *pages[], 69 drm_clflush_pages(struct page *pages[], unsigned long num_pages) 84 struct page *page = pages[i]; 87 if (unlikely(page == NULL)) 90 page_virtual = kmap_atomic(page, KM_USER [all...] |
H A D | drm_buffer.c | 29 * Multipart buffer for coping data which is larger than the page size. 71 DRM_ERROR("Failed to allocate %dth page for drm" 121 " (%p) %dth page.\n", 166 int page = drm_buffer_page(buf); local 170 obj = &buf->data[page][idx]; 174 memcpy(stack_obj, &buf->data[page][idx], beginsz); 176 memcpy(stack_obj + beginsz, &buf->data[page + 1][0],
|
H A D | drm_scatter.c | 51 struct page *page; local 55 page = entry->pagelist[i]; 56 if (page) 57 ClearPageReserved(page); 115 /* This also forces the mapping of COW pages, so our page list 138 /* Verify that each page points to its virtual address, and vice
|
/drivers/gpu/drm/i915/ |
H A D | i915_gem_debug.c | 142 int page; local 157 for (page = 0; page < obj->size / PAGE_SIZE; page++) { 160 backing_map = kmap_atomic(obj->pages[page]); 163 DRM_ERROR("failed to map backing page\n"); 170 page * 1024 + i); 176 page * PAGE_SIZE + i * 4),
|
/drivers/hwmon/pmbus/ |
H A D | pmbus.c | 31 * Find sensor groups and status registers on each page. 36 int page; local 38 /* Sensors detected on page 0 only */ 75 for (page = 0; page < info->pages; page++) { 76 if (pmbus_check_word_register(client, page, PMBUS_READ_VOUT)) { 77 info->func[page] |= PMBUS_HAVE_VOUT; 78 if (pmbus_check_byte_register(client, page, 80 info->func[page] | 109 int page; local [all...] |
H A D | max16064.c | 31 static int max16064_read_word_data(struct i2c_client *client, int page, int reg) argument 37 ret = pmbus_read_word_data(client, page, 41 ret = pmbus_read_word_data(client, page, 55 static int max16064_write_word_data(struct i2c_client *client, int page, argument 62 ret = pmbus_write_word_data(client, page, 66 ret = pmbus_write_word_data(client, page,
|
H A D | max34440.c | 39 static int max34440_read_word_data(struct i2c_client *client, int page, int reg) argument 45 ret = pmbus_read_word_data(client, page, 49 ret = pmbus_read_word_data(client, page, 53 ret = pmbus_read_word_data(client, page, 68 static int max34440_write_word_data(struct i2c_client *client, int page, argument 75 ret = pmbus_write_word_data(client, page, 79 ret = pmbus_write_word_data(client, page, 83 ret = pmbus_write_word_data(client, page, 94 static int max34440_read_byte_data(struct i2c_client *client, int page, int reg) argument 99 if (page > [all...] |
/drivers/infiniband/hw/ehca/ |
H A D | ehca_pd.c | 89 struct ipz_small_queue_page *page, *tmp; local 93 list_for_each_entry_safe(page, tmp, &my_pd->free[i], list) { 95 free_page(page->page); 96 kmem_cache_free(small_qp_cache, page);
|
/drivers/infiniband/hw/mlx4/ |
H A D | doorbell.c | 47 struct mlx4_ib_user_db_page *page; local 53 list_for_each_entry(page, &context->db_page_list, list) 54 if (page->user_virt == (virt & PAGE_MASK)) 57 page = kmalloc(sizeof *page, GFP_KERNEL); 58 if (!page) { 63 page->user_virt = (virt & PAGE_MASK); 64 page->refcnt = 0; 65 page->umem = ib_umem_get(&context->ibucontext, virt & PAGE_MASK, 67 if (IS_ERR(page [all...] |
/drivers/staging/rtl8192u/ieee80211/ |
H A D | internal.h | 33 static inline void *crypto_kmap(struct page *page, int out) argument 35 return kmap_atomic(page, crypto_kmap_type(out));
|
H A D | scatterwalk.h | 22 struct page *page; member in struct:scatter_walk 40 return walk_in->page == walk_out->page &&
|
/drivers/infiniband/hw/qib/ |
H A D | qib_user_pages.c | 39 static void __qib_release_user_pages(struct page **p, size_t num_pages, 55 struct page **p, struct vm_area_struct **vma) 96 * have to bother with retries or mapping a dummy page to insure we 101 dma_addr_t qib_map_page(struct pci_dev *hwdev, struct page *page, argument 106 phys = pci_map_page(hwdev, page, offset, size, direction); 110 phys = pci_map_page(hwdev, page, offset, size, direction); 112 * FIXME: If we get 0 again, we should keep this page, 113 * map another, then free the 0 page. 122 * @start_page: the start page [all...] |
/drivers/md/ |
H A D | dm-sysfs.c | 22 char *page) 36 ret = dm_attr->show(md, page); 21 dm_attr_show(struct kobject *kobj, struct attribute *attr, char *page) argument
|
/drivers/mtd/nand/ |
H A D | bcm_umi_bch.c | 25 struct nand_chip *chip, uint8_t *buf, int page); 48 ** We treat the OOB for a 2K page as if it were 4 512 byte oobs, 70 /* We treat the OOB for a 4K page as if it were 8 512 byte oobs, 102 * bcm_umi_bch_read_page_hwecc - hardware ecc based page read function 110 int page) 123 /* Seek to page location within sector */ 187 * bcm_umi_bch_write_page_hwecc - hardware ecc based page write function 108 bcm_umi_bch_read_page_hwecc(struct mtd_info *mtd, struct nand_chip *chip, uint8_t * buf, int page) argument
|
/drivers/usb/storage/ |
H A D | protocol.c | 143 * each page has to be kmap()'ed separately. If the page is already 145 * If the page is not directly accessible -- such as a user buffer 154 * include multiple pages. Find the initial page structure 155 * and the starting offset within the page, and update 160 struct page *page = sg_page(sg) + local 178 * s-g entry. For each page: call kmap(), do the 183 unsigned char *ptr = kmap(page); 189 kunmap(page); [all...] |
/drivers/auxdisplay/ |
H A D | ks0108.c | 101 void ks0108_page(unsigned char page) argument 103 ks0108_writedata(min(page,(unsigned char)7) | bit(3) | bit(4) | bit(5) | bit(7));
|
/drivers/base/ |
H A D | dma-coherent.c | 171 int page = (vaddr - mem->virt_base) >> PAGE_SHIFT; local 173 bitmap_release_region(mem->bitmap, page, order);
|
/drivers/dma/ |
H A D | iovlock.c | 53 struct page **pages; 71 /* single kmalloc for pinned list, page_list[], and the page arrays */ 74 + (iovec_pages_used * sizeof (struct page*)), GFP_KERNEL); 78 /* list of pages starts right after the page list array */ 79 pages = (struct page **) &local_list->page_list[nr_iovecs]; 177 /* break up copies to not cross page boundary */ 213 struct dma_pinned_list *pinned_list, struct page *page, 226 u8 *vaddr = kmap(page); 228 kunmap(page); 212 dma_memcpy_pg_to_iovec(struct dma_chan *chan, struct iovec *iov, struct dma_pinned_list *pinned_list, struct page *page, unsigned int offset, size_t len) argument [all...] |
/drivers/gpu/drm/ttm/ |
H A D | ttm_agp_backend.c | 63 struct page *page = ttm->pages[i]; local 65 if (!page) 66 page = ttm->dummy_read_page; 68 mem->pages[mem->page_count++] = page; 114 struct page *dummy_read_page)
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_user_pages.c | 41 static void __ipath_release_user_pages(struct page **p, size_t num_pages, 57 struct page **p, struct vm_area_struct **vma) 101 * have to bother with retries or mapping a dummy page to insure we 106 dma_addr_t ipath_map_page(struct pci_dev *hwdev, struct page *page, argument 111 phys = pci_map_page(hwdev, page, offset, size, direction); 115 phys = pci_map_page(hwdev, page, offset, size, direction); 117 * FIXME: If we get 0 again, we should keep this page, 118 * map another, then free the 0 page. 141 * FIXME: If we get 0 again, we should keep this page, [all...] |
/drivers/mca/ |
H A D | mca-proc.c | 35 static int get_mca_info_helper(struct mca_device *mca_dev, char *page, int len) argument 40 len += sprintf(page+len, "%02x ", 42 len += sprintf(page+len, " %s\n", mca_dev ? mca_dev->name : ""); 46 static int get_mca_info(char *page, char **start, off_t off, argument 58 len += sprintf(page+len, "Slot %d: ", i+1); 59 len = get_mca_info_helper(mca_dev, page, len); 65 len += sprintf(page+len, "Video : "); 66 len = get_mca_info_helper(mca_dev, page, len); 71 len += sprintf(page+len, "SCSI : "); 72 len = get_mca_info_helper(mca_dev, page, le 136 get_mca_machine_info(char* page, char **start, off_t off, int count, int *eof, void *data) argument 153 mca_read_proc(char *page, char **start, off_t off, int count, int *eof, void *data) argument [all...] |
/drivers/misc/ |
H A D | uid_stat.c | 55 static int tcp_snd_read_proc(char *page, char **start, off_t off, argument 60 char *p = page; 67 len = (p - page) - off; 69 *start = page + off; 73 static int tcp_rcv_read_proc(char *page, char **start, off_t off, argument 78 char *p = page; 85 len = (p - page) - off; 87 *start = page + off;
|
/drivers/mtd/maps/ |
H A D | vmax301.c | 9 to the lower 8k of the device the second is paged. Writing a 16 bit page 10 value to anywhere in the first 8k will cause the second 8k to page around. 13 of flash that is smart enough to copy itself down, page in the rest of 41 static void __vmax301_page(struct map_info *map, unsigned long page) argument 43 writew(page, map->map_priv_2 - WINDOW_LENGTH); 44 map->map_priv_1 = page; 50 unsigned long page = (ofs >> WINDOW_SHIFT); local 51 if (map->map_priv_1 != page) 52 __vmax301_page(map, page);
|
/drivers/target/tcm_fc/ |
H A D | tfc_io.c | 77 struct page *page = NULL; local 98 page = sg_page(sg); 109 page = sg_page(sg); 138 BUG_ON(!page); 139 get_page(page); 142 page, off_in_page, tlen); 146 PAGE_SIZE << compound_order(page); 148 BUG_ON(!page); 149 from = kmap_atomic(page 201 struct page *page = NULL; local [all...] |
/drivers/atm/ |
H A D | adummy.c | 123 adummy_proc_read(struct atm_dev *dev, loff_t *pos, char *page) argument 128 return sprintf(page, "version %s\n", DRV_VERSION);
|