/drivers/media/pci/ivtv/ |
H A D | ivtv-udma.c | 33 dma_page->page_count = dma_page->last - dma_page->first + 1; 34 if (dma_page->page_count == 1) dma_page->tail -= dma_page->offset; 48 for (i = 0; i < dma_page->page_count; i++) { 49 unsigned int len = (i == dma_page->page_count - 1) ? 112 if (dma->SG_length || dma->page_count) { 113 IVTV_DEBUG_WARN("ivtv_udma_setup: SG_length %d page_count %d still full?\n", 114 dma->SG_length, dma->page_count); 120 if (user_dma.page_count <= 0) { 121 IVTV_DEBUG_WARN("ivtv_udma_setup: Error %d page_count from %d bytes %d offset\n", 122 user_dma.page_count, size_in_byte [all...] |
H A D | ivtv-yuv.c | 67 if (dma->SG_length || dma->page_count) { 69 ("prep_user_dma: SG_length %d page_count %d still full?\n", 70 dma->SG_length, dma->page_count); 79 y_pages = get_user_pages(current, current->mm, y_dma.uaddr, y_dma.page_count, 0, 1, &dma->map[0], NULL); 81 if (y_pages == y_dma.page_count) { 83 uv_dma.uaddr, uv_dma.page_count, 0, 1, 88 if (y_pages != y_dma.page_count || uv_pages != uv_dma.page_count) { 91 if (y_pages == y_dma.page_count) { 94 "expecting %d\n", uv_pages, uv_dma.page_count); [all...] |
/drivers/firewire/ |
H A D | core-iso.c | 42 int fw_iso_buffer_alloc(struct fw_iso_buffer *buffer, int page_count) argument 46 buffer->page_count = 0; 48 buffer->pages = kmalloc(page_count * sizeof(buffer->pages[0]), 53 for (i = 0; i < page_count; i++) { 58 buffer->page_count = i; 59 if (i < page_count) { 75 for (i = 0; i < buffer->page_count; i++) { 84 if (i < buffer->page_count) 91 int page_count, enum dma_data_direction direction) 95 ret = fw_iso_buffer_alloc(buffer, page_count); 90 fw_iso_buffer_init(struct fw_iso_buffer *buffer, struct fw_card *card, int page_count, enum dma_data_direction direction) argument [all...] |
/drivers/char/agp/ |
H A D | generic.c | 197 if (curr->page_count != 0) { 202 for (i = 0; i < curr->page_count; i++) { 207 for (i = 0; i < curr->page_count; i++) { 225 * @page_count: size_t argument of the number of pages 234 size_t page_count, u32 type) 245 if ((cur_memory + page_count > bridge->max_memory_agp) || 246 (cur_memory + page_count < page_count)) 250 new = agp_generic_alloc_user(page_count, type); 257 new = bridge->driver->alloc_by_type(page_count, typ 233 agp_allocate_memory(struct agp_bridge_data *bridge, size_t page_count, u32 type) argument 1164 agp_generic_alloc_by_type(size_t page_count, int type) argument 1178 agp_generic_alloc_user(size_t page_count, int type) argument [all...] |
H A D | sgi-agp.c | 173 if ((pg_start + mem->page_count) > num_entries) 178 while (j < (pg_start + mem->page_count)) { 189 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { 217 for (i = pg_start; i < (mem->page_count + pg_start); i++) {
|
H A D | nvidia-agp.c | 210 if (mem->page_count == 0) 213 if ((pg_start + mem->page_count) > 217 for (j = pg_start; j < (pg_start + mem->page_count); j++) { 226 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { 250 if (mem->page_count == 0) 253 for (i = pg_start; i < (mem->page_count + pg_start); i++)
|
H A D | i460-agp.c | 311 if ((io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count) > num_entries) { 317 while (j < (io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count)) { 327 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { 346 for (i = pg_start; i < (pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count); i++) 415 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; 417 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; 473 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; 475 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE;
|
H A D | efficeon-agp.c | 240 int i, count = mem->page_count, num_entries; 248 if ((pg_start + mem->page_count) > num_entries) 289 int i, count = mem->page_count, num_entries; 295 if ((pg_start + mem->page_count) > num_entries)
|
H A D | agp.h | 199 struct agp_memory *agp_generic_alloc_by_type(size_t page_count, int type); 203 struct agp_memory *memory, size_t page_count); 220 struct agp_memory *agp_generic_alloc_user(size_t page_count, int type);
|
H A D | intel-gtt.c | 126 DBG("try unmapping %lu pages\n", (unsigned long)mem->page_count); 218 if ((pg_start + mem->page_count) 225 for (i = pg_start; i < (pg_start + mem->page_count); i++) { 270 new->page_count = pg_count; 281 if (curr->page_count == 4) 900 if (mem->page_count == 0) 903 if (pg_start + mem->page_count > intel_private.gtt_total_entries) 918 ret = intel_gtt_map_memory(mem->pages, mem->page_count, &st); 926 intel_gtt_insert_pages(pg_start, mem->page_count, mem->pages, 953 if (mem->page_count [all...] |
H A D | ati-agp.c | 280 if (mem->page_count == 0) 283 if ((pg_start + mem->page_count) > num_entries) 287 while (j < (pg_start + mem->page_count)) { 301 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { 326 if (mem->page_count == 0) 329 for (i = pg_start; i < (mem->page_count + pg_start); i++) {
|
H A D | uninorth-agp.c | 165 if (mem->page_count == 0) 171 if ((pg_start + mem->page_count) > num_entries) 175 for (i = 0; i < mem->page_count; ++i) { 184 for (i = 0; i < mem->page_count; i++) { 214 if (mem->page_count == 0) 218 for (i = 0; i < mem->page_count; ++i) {
|
H A D | ali-agp.c | 128 int i, page_count; local 133 page_count = 1 << A_SIZE_32(agp_bridge->current_size)->page_order; 134 for (i = 0; i < PAGE_SIZE * page_count; i += PAGE_SIZE) {
|
H A D | parisc-agp.c | 138 io_pg_count = info->io_pages_per_kpage * mem->page_count; 155 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { 185 io_pg_count = info->io_pages_per_kpage * mem->page_count;
|
/drivers/staging/lustre/lustre/llite/ |
H A D | lloop.c | 195 u32 page_count = 0; local 225 pages[page_count] = bvec.bv_page; 226 offsets[page_count] = offset; 227 page_count++; 230 LASSERT(page_count <= LLOOP_MAX_SEGMENTS); 235 page_count); 237 pvec->ldp_size = page_count << PAGE_CACHE_SHIFT; 238 pvec->ldp_nr = page_count; 253 * 2. Reserve the # of (page_count * depth) cl_pages from the reserved 296 unsigned int page_count local [all...] |
H A D | rw26.c | 139 if (page_count(vmpage) > 3) 239 int page_count = pv->ldp_nr; local 247 for (i = 0; i < page_count; i++) { 337 struct page **pages, int page_count) 340 .ldp_nr = page_count, 333 ll_direct_IO_26_seg(const struct lu_env *env, struct cl_io *io, int rw, struct inode *inode, struct address_space *mapping, size_t size, loff_t file_offset, struct page **pages, int page_count) argument
|
/drivers/gpu/drm/ttm/ |
H A D | ttm_agp_backend.c | 63 mem->page_count = 0; 70 mem->pages[mem->page_count++] = page;
|
/drivers/target/ |
H A D | target_core_rd.c | 86 u32 i, j, page_count = 0, sg_per_table; local 96 page_count++; 103 return page_count; 108 u32 page_count; local 113 page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_table_array, 118 rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count, 119 rd_dev->sg_table_count, (unsigned long)page_count * PAGE_SIZE); 228 u32 page_count; local 233 page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_prot_array, 238 rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count, [all...] |
/drivers/staging/lustre/lustre/osc/ |
H A D | osc_cache.c | 178 int page_count; local 242 page_count = 0; 245 ++page_count; 249 if (page_count != ext->oe_nr_pages) 1032 int page_count = 0; local 1044 ++page_count; 1067 LASSERT(page_count == ext->oe_nr_pages); 1822 int page_count = 0; local 1830 if (!try_to_add_extent_for_io(cli, ext, rpclist, &page_count, 1832 return page_count; 1888 u32 page_count = 0; local 1959 int page_count = 0; local 2508 int page_count = 0; local [all...] |
H A D | osc_request.c | 1094 static void handle_short_read(int nob_read, u32 page_count, argument 1102 LASSERT (page_count > 0); 1110 page_count--; 1116 page_count--; 1121 while (page_count-- > 0) { 1131 u32 page_count, struct brw_page **pga) 1222 (long)pga[i]->pg->flags, page_count(pga[i]->pg), 1247 struct lov_stripe_md *lsm, u32 page_count, 1280 for (niocount = i = 1; i < page_count; i++) { 1303 desc = ptlrpc_prep_bulk_imp(req, page_count, 1129 check_write_rcs(struct ptlrpc_request *req, int requested_nob, int niocount, u32 page_count, struct brw_page **pga) argument 1245 osc_brw_prep_request(int cmd, struct client_obd *cli, struct obdo *oa, struct lov_stripe_md *lsm, u32 page_count, struct brw_page **pga, struct ptlrpc_request **reqp, struct obd_capa *ocapa, int reserve, int resend) argument 1450 check_write_checksum(struct obdo *oa, const lnet_process_id_t *peer, __u32 client_cksum, __u32 server_cksum, int nob, u32 page_count, struct brw_page **pga, cksum_type_t client_cksum_type) argument 1885 int page_count = 0; local [all...] |
/drivers/gpu/drm/nouveau/ |
H A D | nouveau_bo.c | 624 u32 page_count = new_mem->num_pages; local 627 page_count = new_mem->num_pages; 628 while (page_count) { 629 int line_count = (page_count > 8191) ? 8191 : page_count; 647 page_count -= line_count; 662 u32 page_count = new_mem->num_pages; local 665 page_count = new_mem->num_pages; 666 while (page_count) { 667 int line_count = (page_count > 204 701 u32 page_count = new_mem->num_pages; local 886 u32 page_count = new_mem->num_pages; local [all...] |
/drivers/iommu/ |
H A D | tegra-gart.c | 58 u32 page_count; /* total remappable size */ member in struct:gart_device 80 iova < gart->iovmm_base + GART_PAGE_SIZE * gart->page_count; \ 147 gart_end = gart_start + gart->page_count * GART_PAGE_SIZE - 1; 170 gart->page_count * GART_PAGE_SIZE - 1; 386 gart->page_count = (resource_size(res_remap) >> GART_PAGE_SHIFT); 388 gart->savedata = vmalloc(sizeof(u32) * gart->page_count);
|
/drivers/block/ |
H A D | rbd.c | 264 u32 page_count; member in struct:rbd_obj_request::__anon355::__anon356 2052 obj_request->page_count); 2276 obj_request->page_count = 0; 2465 unsigned int page_count; local 2468 page_count = (u32)calc_pages_for(offset, length); 2469 obj_request->page_count = page_count; 2471 page_count--; /* more on last page */ 2472 pages += page_count; 2508 u32 page_count; local 2550 u32 page_count; local 2659 u32 page_count; local 2805 u32 page_count; local 3238 u32 page_count; local 3529 u32 page_count; local [all...] |
/drivers/gpu/drm/ |
H A D | drm_bufs.c | 742 dma->page_count += byte_count >> PAGE_SHIFT; 777 int page_count; local 845 temp_pagelist = kmalloc((dma->page_count + (count << page_order)) * 855 dma->pagelist, dma->page_count * sizeof(*dma->pagelist)); 857 dma->page_count + (count << page_order)); 862 page_count = 0; 881 dma->page_count + page_count, 883 temp_pagelist[dma->page_count + page_count [all...] |
/drivers/gpu/drm/udl/ |
H A D | udl_gem.c | 140 int page_count = obj->base.size / PAGE_SIZE; local 154 obj->vmapping = vmap(obj->pages, page_count, 0, PAGE_KERNEL);
|