/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_dmabuf.c | 59 dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, 100 rd = buf->sgt->sgl; 101 wr = sgt->sgl; 109 nents = dma_map_sg(attach->dev, sgt->sgl, sgt->orig_nents, dir); 111 DRM_ERROR("failed to map sgl with iommu.\n"); 198 struct scatterlist *sgl; local 244 sgl = sgt->sgl; 247 buffer->dma_addr = sg_dma_address(sgl);
|
H A D | exynos_drm_gem.c | 86 struct scatterlist *sgl; local 98 sgl = buf->sgt->sgl; 99 for_each_sg(buf->sgt->sgl, sgl, buf->sgt->nents, i) { 100 if (page_offset < (sgl->length >> PAGE_SHIFT)) 102 page_offset -= (sgl->length >> PAGE_SHIFT); 105 pfn = __phys_to_pfn(sg_phys(sgl)) + page_offset; 486 nents = dma_map_sg(drm_dev->dev, sgt->sgl, sgt->nents, dir); 488 DRM_ERROR("failed to map sgl wit [all...] |
/drivers/infiniband/hw/ipath/ |
H A D | ipath_dma.c | 101 static int ipath_map_sg(struct ib_device *dev, struct scatterlist *sgl, argument 111 for_each_sg(sgl, sg, nents, i) {
|
/drivers/infiniband/hw/qib/ |
H A D | qib_dma.c | 94 static int qib_map_sg(struct ib_device *dev, struct scatterlist *sgl, argument 104 for_each_sg(sgl, sg, nents, i) {
|
/drivers/net/ethernet/intel/ixgbe/ |
H A D | ixgbe_fcoe.h | 63 struct scatterlist *sgl; member in struct:ixgbe_fcoe_ddp
|
H A D | ixgbe_fcoe.c | 53 ddp->sgl = NULL; 107 if (ddp->sgl) 108 dma_unmap_sg(&adapter->pdev->dev, ddp->sgl, ddp->sgc, 124 * @sgl: the scatter-gather list for this request 130 struct scatterlist *sgl, unsigned int sgc, 149 if (!netdev || !sgl) 165 if (ddp->sgl) { 166 e_err(drv, "xid 0x%x w/ non-null sgl=%p nents=%d\n", 167 xid, ddp->sgl, ddp->sgc); 184 /* setup dma from scsi command sgl */ 129 ixgbe_fcoe_ddp_setup(struct net_device *netdev, u16 xid, struct scatterlist *sgl, unsigned int sgc, int target_mode) argument 324 ixgbe_fcoe_ddp_get(struct net_device *netdev, u16 xid, struct scatterlist *sgl, unsigned int sgc) argument 345 ixgbe_fcoe_ddp_target(struct net_device *netdev, u16 xid, struct scatterlist *sgl, unsigned int sgc) argument [all...] |
/drivers/dma/ |
H A D | coh901318_lli.c | 233 struct scatterlist *sgl, unsigned int nents, 258 for_each_sg(sgl, sg, nents, i) { 231 coh901318_lli_fill_sg(struct coh901318_pool *pool, struct coh901318_lli *lli, struct scatterlist *sgl, unsigned int nents, dma_addr_t dev_addr, u32 ctrl_chained, u32 ctrl, u32 ctrl_last, enum dma_transfer_direction dir, u32 ctrl_irq_mask) argument
|
H A D | timb_dma.c | 509 struct scatterlist *sgl, unsigned int sg_len, 520 if (!sgl || !sg_len) { 540 for_each_sg(sgl, sg, sg_len, i) { 508 td_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) argument
|
/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_fcoe.h | 110 struct scatterlist *sgl; member in struct:i40e_fcoe_ddp
|
/drivers/scsi/esas2r/ |
H A D | esas2r_io.c | 224 struct esas2r_mem_desc *sgl; local 231 sgl = esas2r_alloc_sgl(a); 233 if (unlikely(sgl == NULL)) 244 memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen); 248 (struct atto_vda_sge *)((u8 *)sgl->virt_addr + 253 (struct atto_vda_sge *)((u8 *)sgl->virt_addr 260 cpu_to_le64(sgl->phys_addr); 302 list_add(&sgl->next_desc, &rq->sg_table_head); 376 struct esas2r_mem_desc *sgl; local 449 sgl [all...] |
H A D | esas2r_init.c | 840 struct esas2r_mem_desc *sgl; local 908 for (i = 0, sgl = a->sg_list_mds; i < num_sg_lists; i++, sgl++) { 909 sgl->size = sgl_page_size; 911 list_add_tail(&sgl->next_desc, &a->free_sg_list_head); 913 if (!esas2r_initmem_alloc(a, sgl, ESAS2R_SGL_ALIGN)) {
|
/drivers/scsi/isci/ |
H A D | request.h | 69 * @sgl - track pio transfer progress as we iterate through the sgl 79 } sgl; member in struct:isci_stp_request
|
/drivers/staging/android/ion/ |
H A D | ion_heap.c | 48 for_each_sg(table->sgl, sg, table->nents, i) { 81 for_each_sg(table->sgl, sg, table->nents, i) { 118 static int ion_heap_sglist_zero(struct scatterlist *sgl, unsigned int nents, argument 126 for_each_sg_page(sgl, &piter, nents, 0) { 151 return ion_heap_sglist_zero(table->sgl, table->nents, pgprot);
|
/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_qp.c | 77 wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); 78 wqe->send.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); 79 wqe->send.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr); 103 wqe->write.sgl[0].stag = wr->ex.imm_data; 104 wqe->write.sgl[0].len = cpu_to_be32(0); 114 wqe->write.sgl[i].stag = 116 wqe->write.sgl[i].len = 118 wqe->write.sgl[i].to = 264 wqe->recv.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); 265 wqe->recv.sgl[ 540 struct ib_sge sgl; local [all...] |
/drivers/mmc/host/ |
H A D | mxs-mmc.c | 230 struct scatterlist * sgl; local 237 sgl = data->sg; 241 sgl = (struct scatterlist *) ssp->ssp_pio_words; 246 sgl, sg_len, ssp->slave_dirn, flags); 356 struct scatterlist *sgl = data->sg, *sg; local 402 for_each_sg(sgl, sg, sg_len, i)
|
/drivers/rapidio/devices/ |
H A D | tsi721_dma.c | 757 struct scatterlist *sgl, unsigned int sg_len, 767 if (!sgl || !sg_len) { 806 desc->sg = sgl; 756 tsi721_prep_rio_sg(struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *tinfo) argument
|
/drivers/dma/sh/ |
H A D | shdma-base.c | 538 struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr, 548 for_each_sg(sgl, sg, sg_len, i) 565 for_each_sg(sgl, sg, sg_len, i) { 639 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, 662 return shdma_prep_sg(schan, sgl, sg_len, &slave_addr, 680 struct scatterlist *sgl; local 708 sgl = kcalloc(sg_len, sizeof(*sgl), GFP_KERNEL); 709 if (!sgl) 712 sg_init_table(sgl, sg_le 537 shdma_prep_sg(struct shdma_chan *schan, struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr, enum dma_transfer_direction direction, unsigned long flags, bool cyclic) argument 638 shdma_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction direction, unsigned long flags, void *context) argument [all...] |
/drivers/infiniband/hw/cxgb4/ |
H A D | mem.c | 58 struct ulptx_sgl *sgl; local 67 wr_len = roundup(sizeof(*req) + sizeof(*sgl), 16); 87 sgl = (struct ulptx_sgl *)(req + 1); 88 sgl->cmd_nsge = cpu_to_be32(ULPTX_CMD(ULP_TX_SC_DSGL) | 90 sgl->len0 = cpu_to_be32(len); 91 sgl->addr0 = cpu_to_be64(data); 730 for_each_sg(mhp->umem->sg_head.sgl, sg, mhp->umem->nmap, entry) {
|
/drivers/infiniband/ulp/iser/ |
H A D | iser_memory.c | 53 struct scatterlist *sgl = (struct scatterlist *)data->buf; local 59 for_each_sg(sgl, sg, data->size, i) 79 sgl = (struct scatterlist *)data->buf; 81 for_each_sg(sgl, sg, data->size, i) { 127 struct scatterlist *sgl, *sg; local 135 sgl = (struct scatterlist *)data->buf; 139 for_each_sg(sgl, sg, sg_size, i) { 179 struct scatterlist *sg, *sgl = (struct scatterlist *)data->buf; local 186 *offset = (u64) sgl[0].offset & ~MASK_4K; 190 for_each_sg(sgl, s 231 struct scatterlist *sgl, *sg, *next_sg = NULL; local 270 struct scatterlist *sgl = (struct scatterlist *)data->buf; local [all...] |
/drivers/mtd/nand/gpmi-nand/ |
H A D | gpmi-lib.c | 1123 struct scatterlist *sgl; local 1143 sgl = &this->cmd_sgl; 1145 sg_init_one(sgl, this->cmd_buffer, this->command_length); 1146 dma_map_sg(this->dev, sgl, 1, DMA_TO_DEVICE); 1148 sgl, 1, DMA_MEM_TO_DEV,
|
/drivers/target/ |
H A D | target_core_file.c | 326 static int fd_do_rw(struct se_cmd *cmd, struct scatterlist *sgl, argument 344 for_each_sg(sgl, sg, sgl_nents, i) { 359 for_each_sg(sgl, sg, sgl_nents, i) 616 fd_execute_rw(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, argument 637 ret = fd_do_rw(cmd, sgl, sgl_nents, 0); 671 ret = fd_do_rw(cmd, sgl, sgl_nents, 1);
|
H A D | target_core_iblock.c | 669 iblock_execute_rw(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, argument 744 for_each_sg(sgl, sg, sgl_nents, i) {
|
H A D | target_core_pscsi.c | 869 pscsi_map_sg(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, argument 877 int nr_pages = (cmd->data_length + sgl[0].offset + 886 for_each_sg(sgl, sg, sgl_nents, i) { 1035 struct scatterlist *sgl = cmd->t_data_sg; local 1057 if (!sgl) { 1071 ret = pscsi_map_sg(cmd, sgl, sgl_nents, data_direction, &hbio);
|
H A D | target_core_rd.c | 385 rd_execute_rw(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, argument 444 sg_miter_start(&m, sgl, sgl_nents, 453 pr_debug("RD[%u]: invalid sgl %p len %zu\n",
|
/drivers/xen/ |
H A D | swiotlb-xen.c | 546 xen_swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, argument 555 for_each_sg(sgl, sg, nelems, i) { 571 xen_swiotlb_unmap_sg_attrs(hwdev, sgl, i, dir, 573 sg_dma_len(sgl) = 0; 604 xen_swiotlb_unmap_sg_attrs(struct device *hwdev, struct scatterlist *sgl, argument 613 for_each_sg(sgl, sg, nelems, i) 627 xen_swiotlb_sync_sg(struct device *hwdev, struct scatterlist *sgl, argument 634 for_each_sg(sgl, sg, nelems, i)
|