/drivers/crypto/caam/ |
H A D | sg_sw_sec4.h | 89 unsigned int nents, enum dma_data_direction dir, 94 for (i = 0; i < nents; i++) { 99 dma_map_sg(dev, sg, nents, dir); 101 return nents; 105 unsigned int nents, enum dma_data_direction dir, 110 for (i = 0; i < nents; i++) { 115 dma_unmap_sg(dev, sg, nents, dir); 117 return nents; 88 dma_map_sg_chained(struct device *dev, struct scatterlist *sg, unsigned int nents, enum dma_data_direction dir, bool chained) argument 104 dma_unmap_sg_chained(struct device *dev, struct scatterlist *sg, unsigned int nents, enum dma_data_direction dir, bool chained) argument
|
/drivers/infiniband/hw/cxgb3/ |
H A D | cxio_dbg.c | 139 void cxio_dump_rqt(struct cxio_rdev *rdev, u32 hwtid, int nents) argument 142 int size = nents * 64;
|
/drivers/parisc/ |
H A D | iommu-helpers.h | 7 * @nents: The number of entries in the scatter/gather list. 14 iommu_fill_pdir(struct ioc *ioc, struct scatterlist *startsg, int nents, argument 29 while (nents-- > 0) { 33 DBG_RUN_SG(" %d : %08lx/%05x %08lx/%05x\n", nents, 101 struct scatterlist *startsg, int nents, 109 while (nents > 0) { 126 while(--nents > 0) { 100 iommu_coalesce_chunks(struct ioc *ioc, struct device *dev, struct scatterlist *startsg, int nents, int (*iommu_alloc_range)(struct ioc *, struct device *, size_t)) argument
|
H A D | ccio-rm-dma.c | 115 static int ccio_map_sg(struct pci_dev *dev, struct scatterlist *sglist, int nents, int direction) argument 117 int tmp = nents; 120 while (nents) { 123 nents--; 131 static void ccio_unmap_sg(struct pci_dev *dev, struct scatterlist *sglist, int nents, int direction) argument 134 while (nents) { 136 nents--;
|
/drivers/usb/storage/ |
H A D | protocol.c | 141 unsigned int nents = scsi_sg_count(srb); local 144 nents = sg_nents(sg); 148 sg_miter_start(&miter, sg, nents, dir == FROM_XFER_BUF ?
|
/drivers/crypto/qce/ |
H A D | dma.c | 57 int qce_mapsg(struct device *dev, struct scatterlist *sg, int nents, argument 70 err = dma_map_sg(dev, sg, nents, dir); 75 return nents; 78 void qce_unmapsg(struct device *dev, struct scatterlist *sg, int nents, argument 87 dma_unmap_sg(dev, sg, nents, dir); 93 int nents = 0; local 99 nents++; 106 return nents; 135 int nents, unsigned long flags, 142 if (!sg || !nents) 134 qce_dma_prep_sg(struct dma_chan *chan, struct scatterlist *sg, int nents, unsigned long flags, enum dma_transfer_direction dir, dma_async_tx_callback cb, void *cb_param) argument [all...] |
/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_dmabuf.c | 59 dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, 78 int nents, ret; local 109 nents = dma_map_sg(attach->dev, sgt->sgl, sgt->orig_nents, dir); 110 if (!nents) { 249 if (sgt->nents == 1) { 250 /* always physically continuous memory if sgt->nents is 1. */
|
H A D | exynos_drm_gem.c | 99 for_each_sg(buf->sgt->sgl, sgl, buf->sgt->nents, i) { 482 int nents; local 486 nents = dma_map_sg(drm_dev->dev, sgt->sgl, sgt->nents, dir); 487 if (!nents) { 490 return nents; 501 dma_unmap_sg(drm_dev->dev, sgt->sgl, sgt->nents, dir);
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_dma.c | 102 int nents, enum dma_data_direction direction) 107 int ret = nents; 111 for_each_sg(sgl, sg, nents, i) { 127 struct scatterlist *sg, int nents, 101 ipath_map_sg(struct ib_device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction direction) argument 126 ipath_unmap_sg(struct ib_device *dev, struct scatterlist *sg, int nents, enum dma_data_direction direction) argument
|
/drivers/infiniband/hw/qib/ |
H A D | qib_dma.c | 95 int nents, enum dma_data_direction direction) 100 int ret = nents; 104 for_each_sg(sgl, sg, nents, i) { 120 struct scatterlist *sg, int nents, 94 qib_map_sg(struct ib_device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction direction) argument 119 qib_unmap_sg(struct ib_device *dev, struct scatterlist *sg, int nents, enum dma_data_direction direction) argument
|
/drivers/infiniband/hw/usnic/ |
H A D | usnic_uiom.h | 63 int nents; member in struct:usnic_uiom_chunk
|
/drivers/mmc/core/ |
H A D | sdio_ops.c | 129 unsigned int nents, left_size, i; local 160 nents = (left_size - 1) / seg_size + 1; 161 if (nents > 1) { 162 if (sg_alloc_table(&sgtable, nents, GFP_KERNEL)) 166 data.sg_len = nents; 185 if (nents > 1)
|
/drivers/video/adf/ |
H A D | adf_memblock.c | 31 int nents, ret; local 43 nents = dma_map_sg(attach->dev, table->sgl, 1, direction); 44 if (!nents) {
|
/drivers/dma/ |
H A D | coh901318_lli.c | 233 struct scatterlist *sgl, unsigned int nents, 258 for_each_sg(sgl, sg, nents, i) { 265 } else if (i == nents - 1) 231 coh901318_lli_fill_sg(struct coh901318_pool *pool, struct coh901318_lli *lli, struct scatterlist *sgl, unsigned int nents, dma_addr_t dev_addr, u32 ctrl_chained, u32 ctrl, u32 ctrl_last, enum dma_transfer_direction dir, u32 ctrl_irq_mask) argument
|
/drivers/misc/ |
H A D | tifm_core.c | 298 int tifm_map_sg(struct tifm_dev *sock, struct scatterlist *sg, int nents, argument 301 return pci_map_sg(to_pci_dev(sock->dev.parent), sg, nents, direction); 305 void tifm_unmap_sg(struct tifm_dev *sock, struct scatterlist *sg, int nents, argument 308 pci_unmap_sg(to_pci_dev(sock->dev.parent), sg, nents, direction);
|
/drivers/scsi/libfc/ |
H A D | fc_libfc.c | 106 * @nents: pointer to the remaining number of entries in the SG list. 113 u32 *nents, size_t *offset, 128 if (!(*nents)) 130 --(*nents); 111 fc_copy_buffer_to_sglist(void *buf, size_t len, struct scatterlist *sg, u32 *nents, size_t *offset, u32 *crc) argument
|
/drivers/spi/ |
H A D | spi-pxa2xx-dma.c | 26 int i, nents, len = drv_data->len; local 44 nents = DIV_ROUND_UP(len, SZ_2K); 45 if (nents != sgt->nents) { 49 ret = sg_alloc_table(sgt, nents, GFP_ATOMIC); 55 for_each_sg(sgt->sgl, sg, sgt->nents, i) { 67 nents = dma_map_sg(dmadev, sgt->sgl, sgt->nents, dir); 68 if (!nents) 71 return nents; 165 int nents, ret; local [all...] |
/drivers/staging/android/ion/ |
H A D | ion_heap.c | 48 for_each_sg(table->sgl, sg, table->nents, i) { 81 for_each_sg(table->sgl, sg, table->nents, i) { 118 static int ion_heap_sglist_zero(struct scatterlist *sgl, unsigned int nents, argument 126 for_each_sg_page(sgl, &piter, nents, 0) { 151 return ion_heap_sglist_zero(table->sgl, table->nents, pgprot);
|
/drivers/hsi/ |
H A D | hsi.c | 532 * @nents: Number of memory entries 535 * nents can be 0. This mainly makes sense for read transfer. 541 struct hsi_msg *hsi_alloc_msg(unsigned int nents, gfp_t flags) argument 550 if (!nents) 553 err = sg_alloc_table(&msg->sgt, nents, flags); 569 * fields set beforehand. If nents > 0 then the client has to initialize
|
/drivers/crypto/ |
H A D | bfin_crc.c | 122 static struct scatterlist *sg_get(struct scatterlist *sg_list, unsigned int nents, argument 128 for_each_sg(sg_list, sg, nents, i)
|
H A D | mxs-dcp.c | 259 const int nents = sg_nents(req->src); local 289 for_each_sg(req->src, src, nents, i) { 568 const int nents = sg_nents(req->src); local 583 for_each_sg(req->src, src, nents, i) {
|
/drivers/message/i2o/ |
H A D | i2o_block.c | 320 int nents; local 324 nents = blk_rq_map_sg(ireq->req->q, ireq->req, ireq->sg_table); 331 ireq->sg_nents = nents; 333 return i2o_dma_map_sg(c, ireq->sg_table, nents, direction, mptr);
|
/drivers/crypto/ccp/ |
H A D | ccp-ops.c | 56 unsigned int nents; member in struct:ccp_sg_workarea 485 dma_unmap_sg(wa->dma_dev, wa->dma_sg, wa->nents, wa->dma_dir); 500 wa->nents = sg_nents(sg); 514 wa->dma_count = dma_map_sg(dev, sg, wa->nents, dma_dir);
|
/drivers/crypto/ux500/cryp/ |
H A D | cryp_core.c | 813 int nents = 0; local 818 nents++; 821 return nents; 847 /* We have the device now, so store the nents in the dma struct. */
|
/drivers/crypto/ux500/hash/ |
H A D | hash_alg.h | 307 * @nents: Number of sg entries. 316 int nents; member in struct:hash_dma
|