Searched refs:dma_addr (Results 26 - 50 of 319) sorted by relevance

1234567891011>>

/drivers/message/i2o/
H A Dmemory.c78 dma_addr_t dma_addr; local
91 dma_addr = dma_map_single(&c->pdev->dev, ptr, size, direction);
92 if (!dma_mapping_error(&c->pdev->dev, dma_addr)) {
101 *mptr++ = cpu_to_le32(i2o_dma_low(dma_addr));
104 *mptr++ = cpu_to_le32(i2o_dma_high(dma_addr));
108 return dma_addr;
/drivers/net/ethernet/altera/
H A Daltera_msgdma.c122 csrwr32(lower_32_bits(buffer->dma_addr), priv->tx_dma_desc,
124 csrwr32(upper_32_bits(buffer->dma_addr), priv->tx_dma_desc,
166 dma_addr_t dma_addr = rxbuffer->dma_addr; local
176 csrwr32(lower_32_bits(dma_addr), priv->rx_dma_desc,
178 csrwr32(upper_32_bits(dma_addr), priv->rx_dma_desc,
H A Daltera_tse_main.c219 rxbuffer->dma_addr = dma_map_single(priv->device, rxbuffer->skb->data,
223 if (dma_mapping_error(priv->device, rxbuffer->dma_addr)) {
228 rxbuffer->dma_addr &= (dma_addr_t)~3;
237 dma_addr_t dma_addr = rxbuffer->dma_addr; local
240 if (dma_addr)
241 dma_unmap_single(priv->device, dma_addr,
246 rxbuffer->dma_addr = 0;
255 if (buffer->dma_addr) {
257 dma_unmap_page(priv->device, buffer->dma_addr,
566 dma_addr_t dma_addr; local
[all...]
/drivers/media/usb/dvb-usb/
H A Dusb-urb.c101 stream->dma_addr[stream->buf_num]);
121 &stream->dma_addr[stream->buf_num]) ) == NULL) {
128 stream->buf_list[stream->buf_num], (long long)stream->dma_addr[stream->buf_num]);
161 stream->urb_list[i]->transfer_dma = stream->dma_addr[i];
199 urb->transfer_dma = stream->dma_addr[i];
/drivers/media/platform/ti-vpe/
H A Dvpdma.h22 dma_addr_t dma_addr; member in struct:vpdma_buf
191 const struct vpdma_data_format *fmt, dma_addr_t dma_addr,
195 const struct vpdma_data_format *fmt, dma_addr_t dma_addr,
H A Dvpdma.c353 buf->dma_addr = dma_map_single(dev, buf->addr, buf->size,
355 if (dma_mapping_error(dev, buf->dma_addr)) {
374 dma_unmap_single(dev, buf->dma_addr, buf->size, DMA_TO_DEVICE);
440 write_reg(vpdma, VPDMA_LIST_ADDR, (u32) list->buf.dma_addr);
486 WARN_ON(blk->dma_addr & VPDMA_DESC_ALIGN);
493 cfd->payload_addr = (u32) blk->dma_addr;
514 WARN_ON(adb->dma_addr & VPDMA_DESC_ALIGN);
521 cfd->payload_addr = (u32) adb->dma_addr;
622 * dma_addr: dma address as seen by VPDMA
628 const struct vpdma_data_format *fmt, dma_addr_t dma_addr,
626 vpdma_add_out_dtd(struct vpdma_desc_list *list, int width, const struct v4l2_rect *c_rect, const struct vpdma_data_format *fmt, dma_addr_t dma_addr, enum vpdma_channel chan, u32 flags) argument
697 vpdma_add_in_dtd(struct vpdma_desc_list *list, int width, const struct v4l2_rect *c_rect, const struct vpdma_data_format *fmt, dma_addr_t dma_addr, enum vpdma_channel chan, int field, u32 flags, int frame_width, int frame_height, int start_h, int start_v) argument
[all...]
/drivers/media/v4l2-core/
H A Dvideobuf2-dma-contig.c32 dma_addr_t dma_addr; member in struct:vb2_dc_buf
94 return &buf->dma_addr;
153 dma_free_coherent(buf->dev, buf->size, buf->vaddr, buf->dma_addr);
168 buf->vaddr = dma_alloc_coherent(dev, size, &buf->dma_addr,
206 buf->dma_addr, buf->size);
220 __func__, (unsigned long)buf->dma_addr, vma->vm_start,
388 ret = dma_get_sgtable(buf->dev, sgt, buf->vaddr, buf->dma_addr,
625 buf->dma_addr = vb2_dc_pfn_to_dma(buf->dev, pfn);
669 buf->dma_addr = sg_dma_address(sgt->sgl);
739 buf->dma_addr
[all...]
H A Dvideobuf-dma-sg.c224 dma->dma_addr = kcalloc(nr_pages, sizeof(*dma->dma_addr), GFP_KERNEL);
225 if (!dma->dma_addr) {
233 &(dma->dma_addr[i]), GFP_KERNEL);
260 dma_free_coherent(dma->dev, PAGE_SIZE, addr, dma->dma_addr[i]);
262 kfree(dma->dma_addr);
263 dma->dma_addr = NULL;
363 if (dma->dma_addr) {
369 dma->dma_addr[i]);
371 kfree(dma->dma_addr);
[all...]
/drivers/block/rsxx/
H A Ddma.c36 dma_addr_t dma_addr; member in struct:rsxx_dma
227 if (!pci_dma_mapping_error(ctrl->card->dev, dma->dma_addr)) {
228 pci_unmap_page(ctrl->card->dev, dma->dma_addr,
455 dma->dma_addr = pci_map_page(ctrl->card->dev, dma->page,
457 if (pci_dma_mapping_error(ctrl->card->dev, dma->dma_addr)) {
476 cpu_to_le64(dma->dma_addr);
628 dma->dma_addr = 0;
781 &ctrl->status.dma_addr);
783 &ctrl->cmd.dma_addr);
788 iowrite32(lower_32_bits(ctrl->status.dma_addr),
[all...]
/drivers/infiniband/hw/ipath/
H A Dipath_dma.c50 static int ipath_mapping_error(struct ib_device *dev, u64 dma_addr) argument
52 return dma_addr == BAD_DMA_ADDRESS;
H A Dipath_user_sdma.c151 void *kvaddr, dma_addr_t dma_addr)
159 pkt->addr[i].addr = dma_addr;
166 void *kvaddr, dma_addr_t dma_addr)
171 kvaddr, dma_addr);
185 dma_addr_t dma_addr; local
208 dma_addr = dma_map_page(&dd->pcidev->dev, page, 0, len,
210 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) {
216 dma_addr);
298 dma_addr_t dma_addr = local
303 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) {
147 ipath_user_sdma_init_frag(struct ipath_user_sdma_pkt *pkt, int i, size_t offset, size_t len, int put_page, int dma_mapped, struct page *page, void *kvaddr, dma_addr_t dma_addr) argument
162 ipath_user_sdma_init_header(struct ipath_user_sdma_pkt *pkt, u32 counter, size_t offset, size_t len, int dma_mapped, struct page *page, void *kvaddr, dma_addr_t dma_addr) argument
403 dma_addr_t dma_addr; local
[all...]
/drivers/infiniband/hw/qib/
H A Dqib_dma.c48 static int qib_mapping_error(struct ib_device *dev, u64 dma_addr) argument
50 return dma_addr == BAD_DMA_ADDRESS;
/drivers/usb/musb/
H A Dux500_dma.c84 dma_addr_t dma_addr, u32 len)
99 "packet_sz=%d, mode=%d, dma_addr=0x%llx, len=%d is_tx=%d\n",
100 packet_sz, mode, (unsigned long long) dma_addr,
106 sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_addr)), len,
107 offset_in_page(dma_addr));
108 sg_dma_address(&sg) = dma_addr;
204 dma_addr_t dma_addr, u32 len)
211 if (!ux500_dma_is_compatible(channel, packet_sz, (void *)dma_addr, len))
216 ret = ux500_configure_channel(channel, packet_sz, mode, dma_addr, len);
82 ux500_configure_channel(struct dma_channel *channel, u16 packet_sz, u8 mode, dma_addr_t dma_addr, u32 len) argument
202 ux500_dma_channel_program(struct dma_channel *channel, u16 packet_sz, u8 mode, dma_addr_t dma_addr, u32 len) argument
H A Dmusb_dma.h168 dma_addr_t dma_addr,
/drivers/net/ethernet/aeroflex/
H A Dgreth.c239 u32 dma_addr; local
256 dma_addr = dma_map_single(greth->dev,
261 if (dma_mapping_error(greth->dev, dma_addr)) {
267 greth_write_bd(&rx_bd[i].addr, dma_addr);
284 dma_addr = dma_map_single(greth->dev,
289 if (dma_mapping_error(greth->dev, dma_addr)) {
294 greth_write_bd(&rx_bd[i].addr, dma_addr);
307 dma_addr = dma_map_single(greth->dev,
312 if (dma_mapping_error(greth->dev, dma_addr)) {
317 greth_write_bd(&tx_bd[i].addr, dma_addr);
401 u32 status, dma_addr, ctrl; local
469 u32 status, dma_addr; local
748 u32 status, dma_addr; local
863 u32 status, dma_addr; local
[all...]
/drivers/infiniband/ulp/iser/
H A Diser_initiator.c169 tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE);
259 u64 dma_addr; local
284 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc,
286 if (ib_dma_mapping_error(device->ib_device, dma_addr))
289 rx_desc->dma_addr = dma_addr;
292 rx_sg->addr = rx_desc->dma_addr;
303 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr,
331 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr,
592 rx_dma = rx_desc->dma_addr;
[all...]
/drivers/misc/genwqe/
H A Dcard_dev.c97 * @dma_addr: DMA address to be updated
153 * @dma_addr: DMA address to be updated
159 dma_addr_t *dma_addr,
174 if (dma_addr)
175 *dma_addr = m->dma_addr +
216 "[%s] %d. cleanup mapping: u_vaddr=%p u_kaddr=%016lx dma_addr=%lx\n",
219 (unsigned long)dma_map->dma_addr);
225 dma_map->dma_addr);
417 dma_map->dma_addr);
156 __genwqe_search_mapping(struct genwqe_file *cfile, unsigned long u_addr, unsigned int size, dma_addr_t *dma_addr, void **virt_addr) argument
511 dma_addr_t dma_addr; local
647 dma_addr_t dma_addr; local
[all...]
/drivers/crypto/qat/qat_common/
H A Dadf_transport_internal.h66 dma_addr_t dma_addr; member in struct:adf_etr_ring_data
/drivers/net/ethernet/cisco/enic/
H A Dvnic_rq.h69 dma_addr_t dma_addr; member in struct:vnic_rq_buf
129 dma_addr_t dma_addr, unsigned int len,
136 buf->dma_addr = dma_addr;
127 vnic_rq_post(struct vnic_rq *rq, void *os_buf, unsigned int os_buf_index, dma_addr_t dma_addr, unsigned int len, uint64_t wrid) argument
/drivers/net/ethernet/sfc/
H A Dvfdi.h154 * @u.set_status_page.dma_addr: Base address for the &struct vfdi_status.
204 u64 dma_addr; member in struct:vfdi_req::__anon3880::__anon3885
H A Drx.c91 dma_sync_single_for_cpu(&efx->pci_dev->dev, rx_buf->dma_addr, len,
132 dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
159 dma_addr_t dma_addr; local
171 dma_addr =
176 dma_addr))) {
181 state->dma_addr = dma_addr;
184 dma_addr = state->dma_addr;
187 dma_addr
[all...]
/drivers/media/common/saa7146/
H A Dsaa7146_vbi.c10 dma_addr_t dma_addr; local
24 cpu = pci_alloc_consistent(dev->pci, 4096, &dma_addr);
29 saa7146_write(dev, BASE_EVEN3, dma_addr);
30 saa7146_write(dev, BASE_ODD3, dma_addr+vbi_pixel_to_capture);
31 saa7146_write(dev, PROT_ADDR3, dma_addr+4096);
125 pci_free_consistent(dev->pci, 4096, cpu, dma_addr);
130 pci_free_consistent(dev->pci, 4096, cpu, dma_addr);
/drivers/gpu/drm/exynos/
H A Dexynos_drm_fbdev.c56 buffer->dma_addr, buffer->size, &buffer->dma_attrs);
106 phys_addr_t dma_addr = buffer->dma_addr; local
107 if (dma_addr)
108 buffer->kvaddr = (void __iomem *)phys_to_virt(dma_addr);
/drivers/net/ethernet/samsung/sxgbe/
H A Dsxgbe_dma.c51 dma_addr_t dma_addr; local
83 dma_addr = dma_tx + ((t_rsize - 1) * SXGBE_DESC_SIZE_BYTES);
84 writel(lower_32_bits(dma_addr),
87 dma_addr = dma_rx + ((r_rsize - 1) * SXGBE_DESC_SIZE_BYTES);
88 writel(lower_32_bits(dma_addr),
/drivers/net/ethernet/ibm/
H A Dibmveth.c164 pool->dma_addr = kmalloc(sizeof(dma_addr_t) * pool->size, GFP_KERNEL);
165 if (!pool->dma_addr) {
174 kfree(pool->dma_addr);
175 pool->dma_addr = NULL;
182 memset(pool->dma_addr, 0, sizeof(dma_addr_t) * pool->size);
215 dma_addr_t dma_addr; local
240 dma_addr = dma_map_single(&adapter->vdev->dev, skb->data,
243 if (dma_mapping_error(&adapter->vdev->dev, dma_addr))
247 pool->dma_addr[index] = dma_addr;
951 dma_addr_t dma_addr; local
[all...]

Completed in 715 milliseconds

1234567891011>>