/drivers/media/platform/s5p-mfc/ |
H A D | s5p_mfc_opr.c | 46 b->virt = dma_alloc_coherent(dev, b->size, &b->dma, GFP_KERNEL); 48 if (!b->virt) { 53 mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); 60 if (b->virt) { 61 dma_free_coherent(dev, b->size, b->virt, b->dma); 62 b->virt = NULL;
|
/drivers/mtd/maps/ |
H A D | ts5500_flash.c | 72 ts5500_map.virt = ioremap_nocache(ts5500_map.phys, ts5500_map.size); 74 if (!ts5500_map.virt) { 97 iounmap(ts5500_map.virt); 109 if (ts5500_map.virt) { 110 iounmap(ts5500_map.virt); 111 ts5500_map.virt = NULL;
|
H A D | cfi_flagadm.c | 97 flagadm_map.virt = ioremap(FLASH_PHYS_ADDR, 100 if (!flagadm_map.virt) { 115 iounmap((void __iomem *)flagadm_map.virt); 125 if (flagadm_map.virt) { 126 iounmap((void __iomem *)flagadm_map.virt); 127 flagadm_map.virt = NULL;
|
H A D | netsc520.c | 98 netsc520_map.virt = ioremap_nocache(netsc520_map.phys, netsc520_map.size); 100 if (!netsc520_map.virt) { 114 iounmap(netsc520_map.virt); 129 if (netsc520_map.virt) { 130 iounmap(netsc520_map.virt); 131 netsc520_map.virt = NULL;
|
H A D | uclinux.c | 59 size_t *retlen, void **virt, resource_size_t *phys) 62 *virt = map->virt + from; 96 mapp->virt = phys_to_virt(mapp->phys); 98 if (mapp->virt == 0) { 130 if (uclinux_ram_map.virt) 131 uclinux_ram_map.virt = 0; 58 uclinux_point(struct mtd_info *mtd, loff_t from, size_t len, size_t *retlen, void **virt, resource_size_t *phys) argument
|
H A D | l440gx.c | 80 l440gx_map.virt = ioremap_nocache(WINDOW_ADDR, WINDOW_SIZE); 82 if (!l440gx_map.virt) { 88 printk(KERN_NOTICE "window_addr = 0x%08lx\n", (unsigned long)l440gx_map.virt); 115 iounmap(l440gx_map.virt); 149 iounmap(l440gx_map.virt); 158 iounmap(l440gx_map.virt);
|
H A D | nettel.c | 219 nettel_amd_map.virt = ioremap_nocache(amdaddr, maxsize); 220 if (!nettel_amd_map.virt) { 279 iounmap(nettel_amd_map.virt); 280 nettel_amd_map.virt = NULL; 305 nettel_intel_map.virt = ioremap_nocache(intel0addr, maxsize); 306 if (!nettel_intel_map.virt) { 336 iounmap(nettel_intel_map.virt); 339 nettel_intel_map.virt = ioremap_nocache(intel0addr, maxsize); 340 if (!nettel_intel_map.virt) { 403 iounmap(nettel_intel_map.virt); [all...] |
H A D | solutionengine.c | 40 soleng_flash_map.virt = (void __iomem *)P2SEGADDR(0); 42 soleng_eprom_map.virt = (void __iomem *)P1SEGADDR(0x01000000); 52 soleng_flash_map.virt = P2SEGADDR(0x01000000); 54 soleng_eprom_map.virt = P1SEGADDR(0);
|
H A D | impa7.c | 75 impa7_map[i].virt = ioremap(pt[i].addr, pt[i].size); 76 if (!impa7_map[i].virt) { 95 iounmap((void __iomem *)impa7_map[i].virt); 108 iounmap((void __iomem *)impa7_map[i].virt); 109 impa7_map[i].virt = NULL;
|
H A D | dc21285.c | 57 val.x[0] = *(uint8_t*)(map->virt + ofs); 64 val.x[0] = *(uint16_t*)(map->virt + ofs); 71 val.x[0] = *(uint32_t*)(map->virt + ofs); 77 memcpy(to, (void*)(map->virt + from), len); 86 *(uint8_t*)(map->virt + adr) = d.x[0]; 95 *(uint16_t*)(map->virt + adr) = d.x[0]; 102 *(uint32_t*)(map->virt + adr) = d.x[0]; 179 dc21285_map.virt = ioremap(DC21285_FLASH, 16*1024*1024); 180 if (!dc21285_map.virt) { 192 iounmap(dc21285_map.virt); [all...] |
H A D | sc520cdp.c | 242 sc520cdp_map[i].virt = ioremap_nocache(sc520cdp_map[i].phys, sc520cdp_map[i].size); 244 if (!sc520cdp_map[i].virt) { 262 iounmap(sc520cdp_map[i].virt); 290 if (sc520cdp_map[i].virt) { 291 iounmap(sc520cdp_map[i].virt); 292 sc520cdp_map[i].virt = NULL;
|
H A D | sun_uflash.c | 81 up->map.virt = of_ioremap(&op->resource[0], 0, up->map.size, 83 if (!up->map.virt) { 95 of_iounmap(&op->resource[0], up->map.virt, up->map.size); 131 if (up->map.virt) { 132 of_iounmap(&op->resource[0], up->map.virt, up->map.size); 133 up->map.virt = NULL;
|
H A D | scx200_docflash.c | 178 scx200_docflash_map.virt = ioremap(docmem.start, scx200_docflash_map.size); 179 if (!scx200_docflash_map.virt) { 188 iounmap(scx200_docflash_map.virt); 211 if (scx200_docflash_map.virt) { 212 iounmap(scx200_docflash_map.virt);
|
/drivers/infiniband/hw/mlx4/ |
H A D | doorbell.c | 44 int mlx4_ib_db_map_user(struct mlx4_ib_ucontext *context, unsigned long virt, argument 53 if (page->user_virt == (virt & PAGE_MASK)) 62 page->user_virt = (virt & PAGE_MASK); 64 page->umem = ib_umem_get(&context->ibucontext, virt & PAGE_MASK, 75 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK);
|
/drivers/infiniband/hw/mlx5/ |
H A D | doorbell.c | 46 int mlx5_ib_db_map_user(struct mlx5_ib_ucontext *context, unsigned long virt, argument 55 if (page->user_virt == (virt & PAGE_MASK)) 64 page->user_virt = (virt & PAGE_MASK); 66 page->umem = ib_umem_get(&context->ibucontext, virt & PAGE_MASK, 77 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK);
|
/drivers/gpio/ |
H A D | gpio-stp-xway.c | 83 void __iomem *virt; member in struct:xway_stp 110 xway_stp_w32(chip->virt, chip->shadow, XWAY_STP_CPU0); 111 xway_stp_w32_mask(chip->virt, 0, XWAY_STP_CON_SWU, XWAY_STP_CON0); 151 * @virt: pointer to the remapped register range 156 xway_stp_w32(chip->virt, 0, XWAY_STP_AR); 157 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU0); 158 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU1); 159 xway_stp_w32(chip->virt, XWAY_STP_CON_SWU, XWAY_STP_CON0); 160 xway_stp_w32(chip->virt, 0, XWAY_STP_CON1); 163 xway_stp_w32_mask(chip->virt, XWAY_STP_EDGE_MAS [all...] |
/drivers/firmware/ |
H A D | iscsi_ibft_find.c | 62 void *virt; local 70 virt = isa_bus_to_virt(pos); 73 if (memcmp(virt, ibft_signs[i].sign, IBFT_SIGN_LEN) == 81 ibft_addr = (struct acpi_table_ibft *)virt;
|
/drivers/char/tpm/ |
H A D | tpm_acpi.c | 52 void __iomem *virt; local 98 virt = acpi_os_map_iomem(start, len); 99 if (!virt) { 105 memcpy_fromio(log->bios_event_log, virt, len); 107 acpi_os_unmap_iomem(virt, len);
|
/drivers/scsi/lpfc/ |
H A D | lpfc_mem.c | 123 pool->elements[i].virt = pci_pool_alloc(phba->lpfc_mbuf_pool, 125 if (!pool->elements[i].virt) 193 pci_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, 253 pci_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, 302 lpfc_mbuf_free(phba, mp->virt, mp->phys); 312 lpfc_mbuf_free(phba, mp->virt, mp->phys); 326 lpfc_mbuf_free(phba, mp->virt, mp->phys); 373 ret = pool->elements[pool->current_count].virt; 383 * @virt: mbuf to free 395 __lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_ argument 423 lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_t dma) argument [all...] |
/drivers/scsi/ |
H A D | sd_dif.c | 102 * avoid virt->phys remapping during write. However, at read time we 103 * don't know whether the virt sector is the same as when we wrote it 115 u32 phys, virt; local 134 virt = bip_get_seed(bip) & 0xffffffff; 141 if (be32_to_cpu(pi->ref_tag) == virt) 144 virt++; 166 u32 phys, virt; local 181 virt = bip_get_seed(bip) & 0xffffffff; 194 pi->ref_tag = cpu_to_be32(virt); 196 virt [all...] |
/drivers/scsi/isci/ |
H A D | unsolicited_frame_control.c | 65 void *virt = ihost->ufi_buf; local 85 uf_control->headers.array = virt + SCI_UFI_BUF_SIZE; 95 uf_control->address_table.array = virt + SCI_UFI_BUF_SIZE + SCI_UFI_HDR_SIZE; 114 uf->buffer = virt; 123 virt += SCU_UNSOLICITED_FRAME_BUFFER_SIZE;
|
/drivers/isdn/hardware/eicon/ |
H A D | diva_dma.h | 31 int nr, void *virt, dword phys,
|
H A D | diva_dma.c | 59 int nr, void *virt, dword phys, 62 pmap[nr].local_ram_addr = virt; 58 diva_init_dma_map_entry(struct _diva_dma_map_entry *pmap, int nr, void *virt, dword phys, void *addr_handle) argument
|
/drivers/acpi/ |
H A D | osl.c | 93 void __iomem *virt; member in struct:acpi_ioremap 304 return map->virt + (phys - map->phys); 312 void __iomem *virt = NULL; local 317 virt = map->virt + (phys - map->phys); 321 return virt; 327 acpi_map_lookup_virt(void __iomem *virt, acpi_size size) argument 332 if (map->virt <= virt && 333 virt 374 void __iomem *virt; local 445 acpi_os_unmap_iomem(void __iomem *virt, acpi_size size) argument 468 acpi_os_unmap_memory(void *virt, acpi_size size) argument 474 early_acpi_os_unmap_memory(void __iomem *virt, acpi_size size) argument 483 void __iomem *virt; local 529 acpi_os_get_physical_address(void *virt, acpi_physical_address * phys) argument [all...] |
/drivers/message/i2o/ |
H A D | memory.c | 32 i2o_status_block *sb = c->status_block.virt; 196 addr->virt = dma_alloc_coherent(dev, len, &addr->phys, GFP_KERNEL); 203 if (!addr->virt) 206 memset(addr->virt, 0, len); 223 if (addr->virt) { 225 dma_free_coherent(dev, addr->len, addr->virt, 228 kfree(addr->virt); 229 addr->virt = NULL;
|