/drivers/base/ |
H A D | dma-coherent.c | 22 struct dma_coherent_mem **mem) 52 *mem = dma_mem; 66 static void dma_release_coherent_memory(struct dma_coherent_mem *mem) argument 68 if (!mem) 70 iounmap(mem->virt_base); 71 kfree(mem->bitmap); 72 kfree(mem); 76 struct dma_coherent_mem *mem) 81 dev->dma_mem = mem; 90 struct dma_coherent_mem *mem; local 20 dma_init_coherent_memory(phys_addr_t phys_addr, dma_addr_t device_addr, size_t size, int flags, struct dma_coherent_mem **mem) argument 75 dma_assign_coherent_memory(struct device *dev, struct dma_coherent_mem *mem) argument 108 struct dma_coherent_mem *mem = dev->dma_mem; local 120 struct dma_coherent_mem *mem = dev->dma_mem; local 158 struct dma_coherent_mem *mem; local 215 struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL; local 249 struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL; local 281 struct dma_coherent_mem *mem = rmem->priv; local [all...] |
H A D | memory.c | 81 struct memory_block *mem = to_memory_block(dev); local 83 kfree(mem); 114 struct memory_block *mem = to_memory_block(dev); local 117 phys_index = mem->start_section_nr / sections_per_block; 129 struct memory_block *mem = to_memory_block(dev); local 132 if (!present_section_nr(mem->start_section_nr + i)) 134 pfn = section_nr_to_pfn(mem->start_section_nr + i); 147 struct memory_block *mem = to_memory_block(dev); local 154 switch (mem->state) { 166 mem 253 memory_block_change_state(struct memory_block *mem, unsigned long to_state, unsigned long from_state_req) argument 275 struct memory_block *mem = to_memory_block(dev); local 299 struct memory_block *mem = to_memory_block(dev); local 311 struct memory_block *mem = to_memory_block(dev); local 372 struct memory_block *mem = to_memory_block(dev); local 380 struct memory_block *mem = to_memory_block(dev); local 598 struct memory_block *mem; local 624 struct memory_block *mem; local 654 struct memory_block *mem; local 689 struct memory_block *mem; local 715 is_memblock_offlined(struct memory_block *mem) argument [all...] |
/drivers/media/v4l2-core/ |
H A D | videobuf-dma-contig.c | 41 struct videobuf_dma_contig_memory *mem, 44 mem->size = size; 45 mem->vaddr = dma_alloc_coherent(dev, mem->size, 46 &mem->dma_handle, flags); 48 if (!mem->vaddr) { 49 dev_err(dev, "memory alloc size %ld failed\n", mem->size); 53 dev_dbg(dev, "dma mapped data is at %p (%ld)\n", mem->vaddr, mem->size); 59 struct videobuf_dma_contig_memory *mem) 40 __videobuf_dc_alloc(struct device *dev, struct videobuf_dma_contig_memory *mem, unsigned long size, gfp_t flags) argument 58 __videobuf_dc_free(struct device *dev, struct videobuf_dma_contig_memory *mem) argument 87 struct videobuf_dma_contig_memory *mem; local 144 videobuf_dma_contig_user_put(struct videobuf_dma_contig_memory *mem) argument 160 videobuf_dma_contig_user_get(struct videobuf_dma_contig_memory *mem, struct videobuf_buffer *vb) argument 213 struct videobuf_dma_contig_memory *mem; local 228 struct videobuf_dma_contig_memory *mem = buf->priv; local 240 struct videobuf_dma_contig_memory *mem = vb->priv; local 280 struct videobuf_dma_contig_memory *mem; local 368 struct videobuf_dma_contig_memory *mem = buf->priv; local 380 struct videobuf_dma_contig_memory *mem = buf->priv; local [all...] |
H A D | videobuf-vmalloc.c | 75 struct videobuf_vmalloc_memory *mem; local 91 mem = q->bufs[i]->priv; 92 if (mem) { 99 MAGIC_CHECK(mem->magic, MAGIC_VMAL_MEM); 105 __func__, i, mem->vaddr); 107 vfree(mem->vaddr); 108 mem->vaddr = NULL; 140 struct videobuf_vmalloc_memory *mem; local 143 vb = kzalloc(size + sizeof(*mem), GFP_KERNEL); 147 mem 161 struct videobuf_vmalloc_memory *mem = vb->priv; local 236 struct videobuf_vmalloc_memory *mem; local 317 struct videobuf_vmalloc_memory *mem = buf->priv; local 327 struct videobuf_vmalloc_memory *mem = buf->priv; local [all...] |
H A D | videobuf-dma-sg.c | 139 struct videobuf_dma_sg_memory *mem = buf->priv; local 140 BUG_ON(!mem); 142 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); 144 return &mem->dma; 403 struct videobuf_dma_sg_memory *mem; local 416 mem = q->bufs[i]->priv; 417 if (!mem) 420 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); 475 struct videobuf_dma_sg_memory *mem; local 478 vb = kzalloc(size + sizeof(*mem), GFP_KERNE 496 struct videobuf_dma_sg_memory *mem = buf->priv; local 510 struct videobuf_dma_sg_memory *mem = vb->priv; local 580 struct videobuf_dma_sg_memory *mem = buf->priv; local 596 struct videobuf_dma_sg_memory *mem = buf->priv; local [all...] |
/drivers/gpu/drm/nouveau/core/subdev/fb/ |
H A D | ramgk20a.c | 38 struct gk20a_mem *mem = to_gk20a_mem(*pmem); local 41 if (unlikely(mem == NULL)) 44 if (likely(mem->cpuaddr)) 45 dma_free_coherent(dev, mem->base.size << PAGE_SHIFT, 46 mem->cpuaddr, mem->handle); 48 kfree(mem->base.pages); 49 kfree(mem); 57 struct gk20a_mem *mem; local 82 mem [all...] |
H A D | ramnv1a.c | 35 u32 mem, mib; local 50 pci_read_config_dword(bridge, 0x7c, &mem); 51 mib = ((mem >> 6) & 31) + 1; 53 pci_read_config_dword(bridge, 0x84, &mem); 54 mib = ((mem >> 4) & 127) + 1;
|
/drivers/gpu/drm/ttm/ |
H A D | ttm_agp_backend.c | 47 struct agp_memory *mem; member in struct:ttm_agp_backend 55 struct agp_memory *mem; local 59 mem = agp_allocate_memory(agp_be->bridge, ttm->num_pages, AGP_USER_MEMORY); 60 if (unlikely(mem == NULL)) 63 mem->page_count = 0; 70 mem->pages[mem->page_count++] = page; 72 agp_be->mem = mem; 74 mem [all...] |
H A D | ttm_bo_util.c | 44 ttm_bo_mem_put(bo, &bo->mem); 52 struct ttm_mem_reg *old_mem = &bo->mem; 120 struct ttm_mem_reg *mem) 122 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; 128 return bdev->driver->io_mem_reserve(bdev, mem); 131 mem->bus.io_reserved_count++ == 0) { 133 ret = bdev->driver->io_mem_reserve(bdev, mem); 145 struct ttm_mem_reg *mem) 147 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; 153 --mem 119 ttm_mem_io_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) argument 144 ttm_mem_io_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) argument 162 struct ttm_mem_reg *mem = &bo->mem; local 182 struct ttm_mem_reg *mem = &bo->mem; local 191 ttm_mem_reg_ioremap(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem, void **virtual) argument 223 ttm_mem_reg_iounmap(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem, void *virtual) argument 511 struct ttm_mem_reg *mem = &bo->mem; local 533 struct ttm_mem_reg *mem = &bo->mem; pgprot_t prot; local [all...] |
H A D | ttm_bo_manager.c | 53 struct ttm_mem_reg *mem) 74 ret = drm_mm_insert_node_in_range_generic(mm, node, mem->num_pages, 75 mem->page_alignment, 0, 84 mem->mm_node = node; 85 mem->start = node->start; 92 struct ttm_mem_reg *mem) 96 if (mem->mm_node) { 98 drm_mm_remove_node(mem->mm_node); 101 kfree(mem->mm_node); 102 mem 50 ttm_bo_man_get_node(struct ttm_mem_type_manager *man, struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) argument 91 ttm_bo_man_put_node(struct ttm_mem_type_manager *man, struct ttm_mem_reg *mem) argument [all...] |
H A D | ttm_bo.c | 91 bo, bo->mem.num_pages, bo->mem.size >> 10, 92 bo->mem.size >> 20); 146 BUG_ON(bo->mem.mm_node != NULL); 171 if (!(bo->mem.placement & TTM_PL_FLAG_NO_EVICT)) { 175 man = &bdev->man[bo->mem.mem_type]; 277 struct ttm_mem_reg *mem, 282 bool old_is_pci = ttm_mem_reg_is_pci(bdev, &bo->mem); 283 bool new_is_pci = ttm_mem_reg_is_pci(bdev, mem); 284 struct ttm_mem_type_manager *old_man = &bdev->man[bo->mem 276 ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem, bool evict, bool interruptible, bool no_wait_gpu) argument 769 ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) argument 782 ttm_bo_mem_force_space(struct ttm_buffer_object *bo, uint32_t mem_type, const struct ttm_place *place, struct ttm_mem_reg *mem, bool interruptible, bool no_wait_gpu) argument 862 ttm_bo_mem_space(struct ttm_buffer_object *bo, struct ttm_placement *placement, struct ttm_mem_reg *mem, bool interruptible, bool no_wait_gpu) argument 971 struct ttm_mem_reg mem; local 1003 ttm_bo_mem_compat(struct ttm_placement *placement, struct ttm_mem_reg *mem, uint32_t *new_flags) argument 1496 ttm_mem_reg_is_pci(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) argument [all...] |
/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_alloc.h | 47 struct i40e_dma_mem *mem, 51 struct i40e_dma_mem *mem); 53 struct i40e_virt_mem *mem, 56 struct i40e_virt_mem *mem);
|
/drivers/net/ethernet/intel/i40evf/ |
H A D | i40e_alloc.h | 47 struct i40e_dma_mem *mem, 51 struct i40e_dma_mem *mem); 53 struct i40e_virt_mem *mem, 56 struct i40e_virt_mem *mem);
|
/drivers/char/hw_random/ |
H A D | geode-rng.c | 57 void __iomem *mem = (void __iomem *)rng->priv; local 59 *data = readl(mem + GEODE_RNG_DATA_REG); 66 void __iomem *mem = (void __iomem *)rng->priv; local 70 data = !!(readl(mem + GEODE_RNG_STATUS_REG)); 91 void __iomem *mem; local 107 mem = ioremap(rng_base, 0x58); 108 if (!mem) 110 geode_rng.priv = (unsigned long)mem; 123 iounmap(mem); 129 void __iomem *mem local [all...] |
H A D | intel-rng.c | 155 static inline u8 hwstatus_get(void __iomem *mem) argument 157 return readb(mem + INTEL_RNG_HW_STATUS); 160 static inline u8 hwstatus_set(void __iomem *mem, argument 163 writeb(hw_status, mem + INTEL_RNG_HW_STATUS); 164 return hwstatus_get(mem); 169 void __iomem *mem = (void __iomem *)rng->priv; local 173 data = !!(readb(mem + INTEL_RNG_STATUS) & 184 void __iomem *mem = (void __iomem *)rng->priv; local 186 *data = readb(mem + INTEL_RNG_DATA); 193 void __iomem *mem local 212 void __iomem *mem = (void __iomem *)rng->priv; local 233 void __iomem *mem; member in struct:intel_rng_hw 333 void __iomem *mem = mem; local 408 void __iomem *mem = (void __iomem *)intel_rng.priv; local [all...] |
/drivers/firmware/efi/ |
H A D | cper.c | 210 static int cper_mem_err_location(struct cper_mem_err_compact *mem, char *msg) argument 219 if (mem->validation_bits & CPER_MEM_VALID_NODE) 220 n += scnprintf(msg + n, len - n, "node: %d ", mem->node); 221 if (mem->validation_bits & CPER_MEM_VALID_CARD) 222 n += scnprintf(msg + n, len - n, "card: %d ", mem->card); 223 if (mem->validation_bits & CPER_MEM_VALID_MODULE) 224 n += scnprintf(msg + n, len - n, "module: %d ", mem->module); 225 if (mem->validation_bits & CPER_MEM_VALID_RANK_NUMBER) 226 n += scnprintf(msg + n, len - n, "rank: %d ", mem->rank); 227 if (mem 252 cper_dimm_err_location(struct cper_mem_err_compact *mem, char *msg) argument 274 cper_mem_err_pack(const struct cper_sec_mem_err *mem, struct cper_mem_err_compact *cmem) argument 308 cper_print_mem(const char *pfx, const struct cper_sec_mem_err *mem) argument [all...] |
/drivers/uio/ |
H A D | uio_mf624.c | 48 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; 76 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; 103 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; 150 info->mem[0].name = "PCI chipset, interrupts, status " 152 info->mem[0].addr = pci_resource_start(dev, 0); 153 if (!info->mem[0].addr) 155 info->mem[0].size = pci_resource_len(dev, 0); 156 info->mem[0].memtype = UIO_MEM_PHYS; 157 info->mem[0].internal_addr = pci_ioremap_bar(dev, 0); 158 if (!info->mem[ [all...] |
H A D | uio_cif.c | 31 void __iomem *plx_intscr = dev_info->mem[0].internal_addr 58 info->mem[0].addr = pci_resource_start(dev, 0); 59 if (!info->mem[0].addr) 61 info->mem[0].internal_addr = pci_ioremap_bar(dev, 0); 62 if (!info->mem[0].internal_addr) 65 info->mem[0].size = pci_resource_len(dev, 0); 66 info->mem[0].memtype = UIO_MEM_PHYS; 67 info->mem[1].addr = pci_resource_start(dev, 2); 68 info->mem[1].size = pci_resource_len(dev, 2); 69 info->mem[ [all...] |
H A D | uio_netx.c | 35 void __iomem *int_enable_reg = dev_info->mem[0].internal_addr 37 void __iomem *int_status_reg = dev_info->mem[0].internal_addr 82 info->mem[0].addr = pci_resource_start(dev, bar); 83 if (!info->mem[0].addr) 85 info->mem[0].internal_addr = ioremap(pci_resource_start(dev, bar), 88 if (!info->mem[0].internal_addr) 91 info->mem[0].size = pci_resource_len(dev, bar); 92 info->mem[0].memtype = UIO_MEM_PHYS; 99 iowrite32(0, info->mem[0].internal_addr + DPM_HOST_INT_EN0); 111 iounmap(info->mem[ [all...] |
/drivers/isdn/pcbit/ |
H A D | module.c | 25 static int mem[MAX_PCBIT_CARDS]; variable 28 module_param_array(mem, int, NULL, 0); 44 if (mem[0] || irq[0]) 46 for (board = 0; board < MAX_PCBIT_CARDS && mem[board] && irq[board]; board++) 48 if (!mem[board]) 49 mem[board] = 0xD0000; 53 if (pcbit_init_dev(board, mem[board], irq[board]) == 0) 108 mem[i] = ints[j];
|
/drivers/ata/ |
H A D | pata_palmld.c | 55 void __iomem *mem; local 66 mem = devm_ioremap(&pdev->dev, PALMLD_IDE_PHYS, 0x1000); 67 if (!mem) { 91 ap->ioaddr.cmd_addr = mem + 0x10; 92 ap->ioaddr.altstatus_addr = mem + 0xe; 93 ap->ioaddr.ctl_addr = mem + 0xe;
|
/drivers/char/agp/ |
H A D | sgi-agp.c | 59 * Flush GART tlb's. Cannot selectively flush based on memory so the mem 63 static void sgi_tioca_tlbflush(struct agp_memory *mem) argument 65 tioca_tlbflush(mem->bridge->dev_private_data); 127 static int sgi_tioca_insert_memory(struct agp_memory *mem, off_t pg_start, argument 137 bridge = mem->bridge; 169 if (type != 0 || mem->type != 0) { 173 if ((pg_start + mem->page_count) > num_entries) 178 while (j < (pg_start + mem->page_count)) { 184 if (!mem->is_flushed) { 186 mem 200 sgi_tioca_remove_memory(struct agp_memory *mem, off_t pg_start, int type) argument [all...] |
/drivers/gpu/drm/nouveau/core/subdev/bar/ |
H A D | nvc0.c | 34 struct nouveau_gpuobj *mem; member in struct:nvc0_bar_priv_vm 46 nvc0_bar_kmap(struct nouveau_bar *bar, struct nouveau_mem *mem, argument 52 ret = nouveau_vm_get(priv->bar[0].vm, mem->size << 12, 12, flags, vma); 56 nouveau_vm_map(vma, mem); 61 nvc0_bar_umap(struct nouveau_bar *bar, struct nouveau_mem *mem, argument 67 ret = nouveau_vm_get(priv->bar[1].vm, mem->size << 12, 68 mem->page_shift, flags, vma); 72 nouveau_vm_map(vma, mem); 93 &bar_vm->mem); 128 nv_wo32(bar_vm->mem, [all...] |
/drivers/iio/adc/ |
H A D | men_z188_adc.c | 28 struct resource *mem; member in struct:z188_adc 109 struct resource *mem; local 123 mem = mcb_request_mem(dev, "z188-adc"); 124 if (IS_ERR(mem)) 125 return PTR_ERR(mem); 127 adc->base = ioremap(mem->start, resource_size(mem)); 133 adc->mem = mem; 139 mcb_release_mem(mem); [all...] |
/drivers/net/wireless/ti/wlcore/ |
H A D | io.c | 83 if ((addr >= part->mem.start) && 84 (addr < part->mem.start + part->mem.size)) 85 return addr - part->mem.start; 88 return addr - part->reg.start + part->mem.size; 91 return addr - part->mem2.start + part->mem.size + 95 return addr - part->mem3.start + part->mem.size + 117 * ...+----+--> mem.start 121 * 00000000 <--+----+... ...+----+--> mem.start + mem [all...] |