Searched refs:dmabuf (Results 1 - 25 of 30) sorted by relevance

12

/drivers/base/
H A Ddma-buf.c35 struct dma_buf *dmabuf; local
40 dmabuf = file->private_data;
42 dmabuf->ops->release(dmabuf);
43 kfree(dmabuf);
49 struct dma_buf *dmabuf; local
54 dmabuf = file->private_data;
58 dmabuf->size >> PAGE_SHIFT)
61 return dmabuf->ops->mmap(dmabuf, vm
95 struct dma_buf *dmabuf; local
135 dma_buf_fd(struct dma_buf *dmabuf, int flags) argument
185 dma_buf_put(struct dma_buf *dmabuf) argument
204 dma_buf_attach(struct dma_buf *dmabuf, struct device *dev) argument
246 dma_buf_detach(struct dma_buf *dmabuf, struct dma_buf_attachment *attach) argument
322 dma_buf_begin_cpu_access(struct dma_buf *dmabuf, size_t start, size_t len, enum dma_data_direction direction) argument
349 dma_buf_end_cpu_access(struct dma_buf *dmabuf, size_t start, size_t len, enum dma_data_direction direction) argument
368 dma_buf_kmap_atomic(struct dma_buf *dmabuf, unsigned long page_num) argument
384 dma_buf_kunmap_atomic(struct dma_buf *dmabuf, unsigned long page_num, void *vaddr) argument
403 dma_buf_kmap(struct dma_buf *dmabuf, unsigned long page_num) argument
419 dma_buf_kunmap(struct dma_buf *dmabuf, unsigned long page_num, void *vaddr) argument
444 dma_buf_mmap(struct dma_buf *dmabuf, struct vm_area_struct *vma, unsigned long pgoff) argument
[all...]
/drivers/gpu/ion/
H A Dion.c737 struct dma_buf *dmabuf = attachment->dmabuf; local
738 struct ion_buffer *buffer = dmabuf->priv;
856 static int ion_mmap(struct dma_buf *dmabuf, struct vm_area_struct *vma) argument
858 struct ion_buffer *buffer = dmabuf->priv;
889 static void ion_dma_buf_release(struct dma_buf *dmabuf) argument
891 struct ion_buffer *buffer = dmabuf->priv;
895 static void *ion_dma_buf_kmap(struct dma_buf *dmabuf, unsigned long offset) argument
897 struct ion_buffer *buffer = dmabuf->priv;
901 static void ion_dma_buf_kunmap(struct dma_buf *dmabuf, unsigne argument
907 ion_dma_buf_begin_cpu_access(struct dma_buf *dmabuf, size_t start, size_t len, enum dma_data_direction direction) argument
930 ion_dma_buf_end_cpu_access(struct dma_buf *dmabuf, size_t start, size_t len, enum dma_data_direction direction) argument
957 struct dma_buf *dmabuf; local
986 struct dma_buf *dmabuf; local
1023 struct dma_buf *dmabuf; local
[all...]
/drivers/gpu/drm/savage/
H A Dsavage_state.c280 const struct drm_buf * dmabuf)
290 if (!dmabuf) {
342 if (start + n > dmabuf->total / 32) {
344 start, start + n - 1, dmabuf->total / 32);
353 if (dmabuf->bus_address != dev_priv->state.common.vbaddr) {
356 BCI_WRITE(dmabuf->bus_address | dev_priv->dma_type);
357 dev_priv->state.common.vbaddr = dmabuf->bus_address;
539 const struct drm_buf * dmabuf)
548 if (!dmabuf) {
603 if (dmabuf
278 savage_dispatch_dma_prim(drm_savage_private_t * dev_priv, const drm_savage_cmd_header_t * cmd_header, const struct drm_buf * dmabuf) argument
536 savage_dispatch_dma_idx(drm_savage_private_t * dev_priv, const drm_savage_cmd_header_t * cmd_header, const uint16_t *idx, const struct drm_buf * dmabuf) argument
892 savage_dispatch_draw(drm_savage_private_t * dev_priv, const drm_savage_cmd_header_t *start, const drm_savage_cmd_header_t *end, const struct drm_buf * dmabuf, const unsigned int *vtxbuf, unsigned int vb_size, unsigned int vb_stride, unsigned int nbox, const struct drm_clip_rect *boxes) argument
960 struct drm_buf *dmabuf; local
[all...]
/drivers/scsi/lpfc/
H A Dlpfc_bsg.c554 * dmabuf, context3 holds the data dmabuf
778 struct lpfc_dmabuf *dmabuf = NULL; local
807 dmabuf = bdeBuf1;
811 dmabuf = lpfc_sli_ringpostbuf_get(phba, pring, dma_addr);
813 if (dmabuf == NULL)
815 ct_req = (struct lpfc_sli_ct_request *)dmabuf->virt;
820 lpfc_sli_ringpostbuf_put(phba, pring, dmabuf);
878 dmabuf = bdeBuf1;
884 dmabuf
2346 struct lpfc_dmabuf *dmabuf; local
2484 struct lpfc_dmabuf *dmabuf; local
2516 lpfc_bsg_dma_page_free(struct lpfc_hba *phba, struct lpfc_dmabuf *dmabuf) argument
2542 struct lpfc_dmabuf *dmabuf, *next_dmabuf; local
3332 struct lpfc_dmabuf *dmabuf; local
3583 lpfc_bsg_sli_cfg_read_cmd_ext(struct lpfc_hba *phba, struct fc_bsg_job *job, enum nemb_type nemb_tp, struct lpfc_dmabuf *dmabuf) argument
3770 lpfc_bsg_sli_cfg_write_cmd_ext(struct lpfc_hba *phba, struct fc_bsg_job *job, enum nemb_type nemb_tp, struct lpfc_dmabuf *dmabuf) argument
3933 lpfc_bsg_handle_sli_cfg_mbox(struct lpfc_hba *phba, struct fc_bsg_job *job, struct lpfc_dmabuf *dmabuf) argument
4080 struct lpfc_dmabuf *dmabuf; local
4145 lpfc_bsg_write_ebuf_set(struct lpfc_hba *phba, struct fc_bsg_job *job, struct lpfc_dmabuf *dmabuf) argument
4270 lpfc_bsg_handle_sli_cfg_ebuf(struct lpfc_hba *phba, struct fc_bsg_job *job, struct lpfc_dmabuf *dmabuf) argument
4315 lpfc_bsg_handle_sli_cfg_ext(struct lpfc_hba *phba, struct fc_bsg_job *job, struct lpfc_dmabuf *dmabuf) argument
4400 struct lpfc_dmabuf *dmabuf = NULL; local
[all...]
H A Dlpfc_sli.c1638 struct lpfc_dmabuf *dmabuf, *next_dmabuf; local
1648 list_for_each_entry_safe(dmabuf, next_dmabuf,
1650 hbq_buf = container_of(dmabuf, struct hbq_dmabuf, dbuf);
1657 list_for_each_entry_safe(dmabuf, next_dmabuf, &phba->rb_pend_list,
1659 hbq_buf = container_of(dmabuf, struct hbq_dmabuf, dbuf);
3365 struct hbq_dmabuf *dmabuf; local
3392 dmabuf = container_of(cq_event, struct hbq_dmabuf,
3394 lpfc_sli4_handle_received_buffer(phba, dmabuf);
4665 struct lpfc_dmabuf *dmabuf; local
4668 dmabuf
7533 struct lpfc_dmabuf *dmabuf; local
11773 struct lpfc_dmabuf *dmabuf; local
11804 struct lpfc_dmabuf *dmabuf; local
11888 struct lpfc_dmabuf *dmabuf; local
12006 struct lpfc_dmabuf *dmabuf; local
12124 struct lpfc_dmabuf *dmabuf; local
12191 struct lpfc_dmabuf *dmabuf; local
12343 struct lpfc_dmabuf *dmabuf; local
12487 struct lpfc_dmabuf *dmabuf; local
13872 struct hbq_dmabuf *dmabuf = NULL; local
13896 struct hbq_dmabuf *dmabuf = NULL; local
13928 struct hbq_dmabuf *dmabuf = NULL; local
13970 lpfc_fc_frame_add(struct lpfc_vport *vport, struct hbq_dmabuf *dmabuf) argument
14054 lpfc_sli4_abort_partial_seq(struct lpfc_vport *vport, struct hbq_dmabuf *dmabuf) argument
14262 lpfc_sli4_handle_unsol_abort(struct lpfc_vport *vport, struct hbq_dmabuf *dmabuf) argument
14309 lpfc_seq_complete(struct hbq_dmabuf *dmabuf) argument
14503 lpfc_sli4_handle_received_buffer(struct lpfc_hba *phba, struct hbq_dmabuf *dmabuf) argument
15788 struct lpfc_dmabuf *dmabuf; local
[all...]
H A Dlpfc_init.c5209 struct lpfc_dmabuf *dmabuf; local
5252 dmabuf = kzalloc(sizeof(struct lpfc_dmabuf), GFP_KERNEL);
5253 if (!dmabuf)
5256 dmabuf->virt = dma_alloc_coherent(&phba->pcidev->dev,
5258 &dmabuf->phys,
5260 if (!dmabuf->virt) {
5265 memset(dmabuf->virt, 0, LPFC_HDR_TEMPLATE_SIZE);
5266 if (!IS_ALIGNED(dmabuf->phys, LPFC_HDR_TEMPLATE_SIZE)) {
5276 rpi_hdr->dmabuf = dmabuf;
5973 struct lpfc_dmabuf *dmabuf; local
9140 struct lpfc_dmabuf *dmabuf, *next; local
[all...]
H A Dlpfc_sli4.h277 struct lpfc_dmabuf *dmabuf; member in struct:lpfc_bmbx
564 struct lpfc_dmabuf *dmabuf; member in struct:lpfc_rpi_hdr
/drivers/scsi/
H A Dsun3_scsi.c140 static unsigned char *dmabuf = NULL; /* dma memory buffer */ variable
248 if((dmabuf = dvma_malloc_align(SUN3_DVMA_BUFSIZE, 0x10000)) == NULL) {
427 memcpy(dmabuf, data, count);
438 // addr = sun3_dvma_page((unsigned long)data, (unsigned long)dmabuf);
475 udc_regs->addr_hi = ((dvma_vtob(dmabuf) & 0xff0000) >> 8);
476 udc_regs->addr_lo = (dvma_vtob(dmabuf) & 0xffff);
586 memcpy(sun3_dma_orig_addr, dmabuf, sun3_dma_orig_count);
/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_kms.h51 bool dmabuf; member in struct:vmw_framebuffer
67 struct vmw_dma_buffer *dmabuf,
H A Dvmwgfx_kms.c124 struct vmw_dma_buffer *dmabuf,
138 ret = ttm_bo_reserve(&dmabuf->base, true, false, false, 0);
144 ret = ttm_bo_kmap(&dmabuf->base, kmap_offset, kmap_num, &map);
154 ttm_bo_unreserve(&dmabuf->base);
180 struct vmw_dma_buffer *dmabuf = NULL; local
189 handle, &surface, &dmabuf);
191 DRM_ERROR("failed to find surface or dmabuf: %i\n", ret);
220 } else if (dmabuf) {
222 du->cursor_dmabuf = dmabuf;
224 ret = vmw_cursor_update_dmabuf(dev_priv, dmabuf, widt
123 vmw_cursor_update_dmabuf(struct vmw_private *dev_priv, struct vmw_dma_buffer *dmabuf, u32 width, u32 height, u32 hotspotX, u32 hotspotY) argument
1005 vmw_kms_new_framebuffer_dmabuf(struct vmw_private *dev_priv, struct vmw_dma_buffer *dmabuf, struct vmw_framebuffer **out, const struct drm_mode_fb_cmd *mode_cmd) argument
1330 struct vmw_dma_buffer *dmabuf = vfbd->buffer; local
[all...]
H A Dvmwgfx_ioctl.c263 if (!vfb->dmabuf) {
264 DRM_ERROR("Framebuffer not dmabuf backed.\n");
/drivers/gpu/drm/
H A Ddrm_prime.c51 * (avoiding loops): importing_gem -> dmabuf -> exporting_gem
79 get_dma_buf(obj->import_attach->dmabuf);
80 *prime_fd = dma_buf_fd(obj->import_attach->dmabuf, flags);
238 dma_buf = attach->dmabuf;
239 dma_buf_detach(attach->dmabuf, attach);
H A Ddrm_gem.c238 obj->import_attach->dmabuf);
537 obj->import_attach->dmabuf);
/drivers/staging/comedi/drivers/
H A Dpcl816.c194 unsigned long dmabuf[2]; /* pointers to begin of DMA buffers */ member in struct:pcl816_private
433 ptr = (short *)devpriv->dmabuf[this_dma_buf];
801 (short *)devpriv->dmabuf[devpriv->next_dma_buf],
1090 if (devpriv->dmabuf[0])
1091 free_pages(devpriv->dmabuf[0], devpriv->dmapages[0]);
1092 if (devpriv->dmabuf[1])
1093 free_pages(devpriv->dmabuf[1], devpriv->dmapages[1]);
1245 devpriv->dmabuf[0] = __get_dma_pages(GFP_KERNEL, pages);
1247 if (!devpriv->dmabuf[0]) {
1256 devpriv->hwdmaptr[0] = virt_to_bus((void *)devpriv->dmabuf[
[all...]
H A Dpcl818.c343 unsigned long dmabuf[2]; /* pointers to begin of DMA buffers */ member in struct:pcl818_private
626 ptr = (short *)devpriv->dmabuf[1 - devpriv->next_dma_buf];
683 short *dmabuf = (short *)devpriv->dmabuf[0]; local
713 if (dmabuf[i] != MAGIC_DMA_WORD) { /* DMA overflow! */
715 /* printk("I %d dmabuf[i] %d %d\n",i,dmabuf[i],devpriv->dmasamplsize); */
726 if ((dmabuf[bufptr] & 0xf) != devpriv->act_chanlist[devpriv->act_chanlist_pos]) { /* dropout! */
729 (dmabuf[bufptr] & 0xf),
739 comedi_buf_put(s->async, dmabuf[bufpt
[all...]
H A Dpcl812.c449 unsigned long dmabuf[2]; /* PTR to DMA buf */ member in struct:pcl812_private
1105 ptr = (short *)devpriv->dmabuf[devpriv->next_dma_buf];
1195 (void *)devpriv->dmabuf[1 -
1282 if (devpriv->dmabuf[0])
1283 free_pages(devpriv->dmabuf[0], devpriv->dmapages[0]);
1284 if (devpriv->dmabuf[1])
1285 free_pages(devpriv->dmabuf[1], devpriv->dmapages[1]);
1434 devpriv->dmabuf[0] = __get_dma_pages(GFP_KERNEL, pages);
1435 if (!devpriv->dmabuf[0]) {
1445 devpriv->hwdmaptr[0] = virt_to_bus((void *)devpriv->dmabuf[
[all...]
/drivers/net/tokenring/
H A Dtms380tr.c371 dma_addr_t dmabuf; local
437 dmabuf = dma_map_single(tp->pdev, tp->Rpl[i].Skb->data, tp->MaxPacketSize, DMA_FROM_DEVICE);
438 if(tp->dmalimit && (dmabuf + tp->MaxPacketSize > tp->dmalimit))
447 tp->Rpl[i].FragList[0].DataAddr = htonl(dmabuf);
449 tp->Rpl[i].DMABuff = dmabuf;
608 dma_addr_t dmabuf, newbuf; local
624 dmabuf = 0;
629 dmabuf = dma_map_single(tp->pdev, skb->data, length, DMA_TO_DEVICE);
630 if(tp->dmalimit && (dmabuf + length > tp->dmalimit)) {
632 dma_unmap_single(tp->pdev, dmabuf, lengt
2025 dma_addr_t dmabuf; local
[all...]
/drivers/mmc/host/
H A Dat91_mci.c216 unsigned *dmabuf = host->buffer; local
224 memset(dmabuf, 0, 12);
246 *dmabuf++ = swab32(sgbuffer[index]);
248 char *tmpv = (char *)dmabuf;
251 dmabuf = (unsigned *)tmpv;
275 unsigned *dmabuf = host->buffer; local
311 sgbuffer[index] = swab32(*dmabuf++);
313 char *tmpv = (char *)dmabuf;
316 dmabuf = (unsigned *)tmpv;
H A Dwbsd.c280 char *dmabuf = host->dma_buffer; local
288 memcpy(dmabuf, sgbuf, sg[i].length);
289 dmabuf += sg[i].length;
297 char *dmabuf = host->dma_buffer; local
305 memcpy(sgbuf, dmabuf, sg[i].length);
306 dmabuf += sg[i].length;
/drivers/mtd/nand/
H A Dcafe_nand.c71 unsigned char *dmabuf; member in struct:cafe_priv
123 memcpy(cafe->dmabuf + cafe->datalen, buf, len);
138 memcpy(buf, cafe->dmabuf + cafe->datalen, len);
663 cafe->dmabuf = dma_alloc_coherent(&cafe->pdev->dev, 2112 + sizeof(struct nand_buffers),
665 if (!cafe->dmabuf) {
669 cafe->nand.buffers = (void *)cafe->dmabuf + 2112;
752 cafe_readl(cafe, NAND_DMA_ADDR0), cafe->dmabuf);
812 dma_free_coherent(&cafe->pdev->dev, 2112, cafe->dmabuf, cafe->dmaaddr);
832 dma_free_coherent(&cafe->pdev->dev, 2112, cafe->dmabuf, cafe->dmaaddr);
/drivers/media/video/omap/
H A Domap_vout_vrfb.c218 dma_addr_t dmabuf; local
228 dmabuf = vout->buf_phy_addr[vb->i];
257 dmabuf, src_element_index, src_frame_index);
/drivers/net/ethernet/smsc/
H A Dsmc91x.h378 dma_addr_t dmabuf; local
394 dmabuf = dma_map_single(lp->device, buf, len, DMA_FROM_DEVICE);
396 DTADR(dma) = dmabuf;
404 dma_unmap_single(lp->device, dmabuf, len, DMA_FROM_DEVICE);
417 dma_addr_t dmabuf; local
433 dmabuf = dma_map_single(lp->device, buf, len, DMA_FROM_DEVICE);
435 DTADR(dma) = dmabuf;
443 dma_unmap_single(lp->device, dmabuf, len, DMA_FROM_DEVICE);
/drivers/isdn/hardware/avm/
H A Db1.c739 p->recvbuf.dmabuf = buf;
748 p->sendbuf.dmabuf = buf;
755 p->recvbuf.dmabuf, p->recvbuf.dmaaddr);
765 p->recvbuf.dmabuf, p->recvbuf.dmaaddr);
767 p->sendbuf.dmabuf, p->sendbuf.dmaaddr);
H A Db1dma.c385 p = dma->sendbuf.dmabuf;
396 txlen = (u8 *)p - (u8 *)dma->sendbuf.dmabuf;
410 skb_copy_from_linear_data_offset(skb, 2, dma->sendbuf.dmabuf,
453 void *p = dma->recvbuf.dmabuf + 4;
608 dma->recvlen = *((u32 *)dma->recvbuf.dmabuf);
H A Dc4.c438 p = dma->sendbuf.dmabuf;
449 txlen = (u8 *)p - (u8 *)dma->sendbuf.dmabuf;
463 skb_copy_from_linear_data_offset(skb, 2, dma->sendbuf.dmabuf,
509 void *p = dma->recvbuf.dmabuf;

Completed in 588 milliseconds

12