/drivers/target/ |
H A D | target_core_file.c | 266 struct scatterlist *sg; local 293 for_each_sg(fd_prot->prot_sg, sg, fd_prot->prot_sg_nents, i) { 296 sg_set_buf(sg, buf, len); 332 struct scatterlist *sg; local 344 for_each_sg(sgl, sg, sgl_nents, i) { 345 iov[i].iov_len = sg->length; 346 iov[i].iov_base = kmap(sg_page(sg)) + sg->offset; 359 for_each_sg(sgl, sg, sgl_nents, i) 360 kunmap(sg_page(sg)); 439 fd_setup_write_same_buf(struct se_cmd *cmd, struct scatterlist *sg, unsigned int len) argument 483 struct scatterlist *sg; local [all...] |
H A D | target_core_rd.c | 85 struct scatterlist *sg; local 89 sg = sg_table[i].sg_table; 93 pg = sg_page(&sg[j]); 99 kfree(sg); 137 struct scatterlist *sg; local 144 sg = kzalloc(sg_per_table * sizeof(struct scatterlist), 146 if (!sg) { 152 sg_init_table(sg, sg_per_table); 154 sg_table[i].sg_table = sg; 167 sg_assign_page(&sg[ [all...] |
/drivers/usb/host/whci/ |
H A D | qset.c | 266 struct scatterlist *sg; local 273 sg = std->bounce_sg; 279 len = min(sg->length - offset, remaining); 280 memcpy(sg_virt(sg) + offset, bounce, len); 286 if (offset >= sg->length) { 287 sg = sg_next(sg); 435 struct scatterlist *sg; local 446 for_each_sg(urb->sg, sg, ur 559 struct scatterlist *sg; local [all...] |
/drivers/char/hw_random/ |
H A D | virtio-rng.c | 57 struct scatterlist sg; local 59 sg_init_one(&sg, buf, size); 62 virtqueue_add_inbuf(vi->vq, &sg, 1, buf, GFP_KERNEL);
|
/drivers/staging/rtl8192e/rtl8192e/ |
H A D | rtl_crypto.h | 165 struct scatterlist *sg, unsigned int nsg); 167 void (*dit_digest)(struct crypto_tfm *tfm, struct scatterlist *sg, 277 struct scatterlist *sg, 281 tfm->crt_digest.dit_update(tfm, sg, nsg); 291 struct scatterlist *sg, 295 tfm->crt_digest.dit_digest(tfm, sg, nsg, out); 276 crypto_digest_update(struct crypto_tfm *tfm, struct scatterlist *sg, unsigned int nsg) argument 290 crypto_digest_digest(struct crypto_tfm *tfm, struct scatterlist *sg, unsigned int nsg, u8 *out) argument
|
/drivers/crypto/ |
H A D | omap-sham.c | 152 struct scatterlist *sg; member in struct:omap_sham_reqctx 154 unsigned int offset; /* offset in current sg */ 585 ctx->sgl.page_link = ctx->sg->page_link; 586 ctx->sgl.offset = ctx->sg->offset; 588 sg_dma_address(&ctx->sgl) = sg_dma_address(ctx->sg); 641 while (ctx->sg) { 642 vaddr = kmap_atomic(sg_page(ctx->sg)); 646 ctx->sg->length - ctx->offset); 654 if (ctx->offset == ctx->sg->length) { 655 ctx->sg 720 struct scatterlist *sg; local [all...] |
/drivers/crypto/qce/ |
H A D | ablkcipher.c | 70 struct scatterlist *sg; local 104 sg = qce_sgtable_add(&rctx->dst_tbl, req->dst); 105 if (IS_ERR(sg)) { 106 ret = PTR_ERR(sg); 110 sg = qce_sgtable_add(&rctx->dst_tbl, &rctx->result_sg); 111 if (IS_ERR(sg)) { 112 ret = PTR_ERR(sg); 116 sg_mark_end(sg);
|
/drivers/gpu/drm/radeon/ |
H A D | radeon_prime.c | 62 struct sg_table *sg) 71 RADEON_GEM_DOMAIN_GTT, 0, sg, resv, &bo); 60 radeon_gem_prime_import_sg_table(struct drm_device *dev, struct dma_buf_attachment *attach, struct sg_table *sg) argument
|
/drivers/net/irda/ |
H A D | sa1100_ir.c | 50 struct scatterlist sg; member in struct:sa1100_buf 115 return sg_dma_len(&buf->sg) - state.residue; 150 desc = dmaengine_prep_slave_sg(chan, &buf->sg, 1, dir, 180 sg_set_buf(&si->dma_rx.sg, si->dma_rx.skb->data, HPSIR_MAX_RXLEN); 181 if (dma_map_sg(si->dma_rx.dev, &si->dma_rx.sg, 1, DMA_FROM_DEVICE) == 0) { 230 dma_unmap_sg(si->dma_tx.dev, &si->dma_tx.sg, 1, DMA_TO_DEVICE); 235 dev->stats.tx_bytes += sg_dma_len(&si->dma_tx.sg); 263 sg_set_buf(&si->dma_tx.sg, si->tx_buff.data, si->tx_buff.len); 264 if (dma_map_sg(si->dma_tx.dev, &si->dma_tx.sg, 1, DMA_TO_DEVICE) == 0) { 378 dma_unmap_sg(si->dma_tx.dev, &si->dma_tx.sg, [all...] |
/drivers/staging/android/ion/ |
H A D | ion_system_heap.c | 131 struct scatterlist *sg; local 162 sg = table->sgl; 164 sg_set_page(sg, page, PAGE_SIZE << compound_order(page), 0); 165 sg = sg_next(sg); 187 struct scatterlist *sg; local 195 for_each_sg(table->sgl, sg, table->nents, i) 196 free_buffer_page(sys_heap, buffer, sg_page(sg));
|
/drivers/usb/wusbcore/ |
H A D | crypto.c | 196 * what sg[4] is for. Maybe there is a smarter way to do this. 209 struct scatterlist sg[4], sg_dst; local 254 sg_init_table(sg, ARRAY_SIZE(sg)); 255 sg_set_buf(&sg[0], &b0, sizeof(b0)); 256 sg_set_buf(&sg[1], &b1, sizeof(b1)); 257 sg_set_buf(&sg[2], b, blen); 259 sg_set_buf(&sg[3], bzero, zero_padding); 264 result = crypto_blkcipher_encrypt(&desc, &sg_dst, sg, dst_size);
|
/drivers/crypto/amcc/ |
H A D | crypto4xx_core.c | 60 /* setup pe dma, include reset sg, pdr and pe, then release reset */ 71 /* un reset pe,sg and pdr */ 112 /* un reset pe,sg and pdr */ 551 struct scatterlist *sg; local 558 sg = &dst[i]; 559 sg_len = sg->length; 560 addr = dma_map_page(dev->core_dev->device, sg_page(sg), 561 sg->offset, sg->length, DMA_TO_DEVICE); 564 len = (nbytes <= sg 749 struct scatterlist *sg = sg_list; local 789 struct scatterlist *sg; local [all...] |
/drivers/crypto/ccp/ |
H A D | ccp-ops.c | 55 struct scatterlist *sg; member in struct:ccp_sg_workarea 491 struct scatterlist *sg, u64 len, 496 wa->sg = sg; 497 if (!sg) 500 wa->nents = sg_nents(sg); 501 wa->length = sg->length; 511 wa->dma_sg = sg; 514 wa->dma_count = dma_map_sg(dev, sg, wa->nents, dma_dir); 526 if (!wa->sg) 490 ccp_init_sg_workarea(struct ccp_sg_workarea *wa, struct device *dev, struct scatterlist *sg, u64 len, enum dma_data_direction dma_dir) argument 595 ccp_set_dm_area(struct ccp_dm_workarea *wa, unsigned int wa_offset, struct scatterlist *sg, unsigned int sg_offset, unsigned int len) argument 605 ccp_get_dm_area(struct ccp_dm_workarea *wa, unsigned int wa_offset, struct scatterlist *sg, unsigned int sg_offset, unsigned int len) argument 615 ccp_reverse_set_dm_area(struct ccp_dm_workarea *wa, struct scatterlist *sg, unsigned int len, unsigned int se_len, bool sign_extend) argument 648 ccp_reverse_get_dm_area(struct ccp_dm_workarea *wa, struct scatterlist *sg, unsigned int len) argument 677 ccp_init_data(struct ccp_data *data, struct ccp_cmd_queue *cmd_q, struct scatterlist *sg, u64 sg_len, unsigned int dm_len, enum dma_data_direction dir) argument 1499 struct scatterlist sg; local [all...] |
/drivers/block/ |
H A D | xen-blkfront.c | 79 struct scatterlist *sg; member in struct:blk_shadow 406 struct scatterlist *sg; local 455 nseg = blk_rq_map_sg(req->q, req, info->shadow[id].sg); 486 for_each_sg(info->shadow[id].sg, sg, nseg, i) { 487 fsect = sg->offset >> 9; 488 lsect = fsect + (sg->length >> 9) - 1; 514 gnt_list_entry = get_grant(&gref_head, page_to_pfn(sg_page(sg)), info); 523 BUG_ON(sg->offset + sg 1033 struct scatterlist *sg; local [all...] |
/drivers/crypto/nx/ |
H A D | nx.c | 97 struct nx_sg *sg; local 121 for (sg = sg_head; sg_len < len; sg++) { 124 sg->addr = sg_addr; 128 next_page = (sg->addr & PAGE_MASK) + PAGE_SIZE; 129 sg->len = min_t(u64, sg_addr, next_page) - sg->addr; 130 sg_len += sg->len; 139 if ((sg - sg_head) == sgmax) { 147 return sg; [all...] |
/drivers/media/pci/cx88/ |
H A D | cx88-core.c | 81 struct scatterlist *sg; local 89 sg = sglist; 91 while (offset && offset >= sg_dma_len(sg)) { 92 offset -= sg_dma_len(sg); 93 sg++; 99 if (bpl <= sg_dma_len(sg)-offset) { 102 *(rp++)=cpu_to_le32(sg_dma_address(sg)+offset); 108 (sg_dma_len(sg)-offset)); 109 *(rp++)=cpu_to_le32(sg_dma_address(sg)+offset); 110 todo -= (sg_dma_len(sg) [all...] |
/drivers/dma/ |
H A D | omap-dma.c | 75 struct omap_sg sg[0]; member in struct:omap_desc 362 struct omap_sg *sg = d->sg + idx; local 375 omap_dma_chan_write(c, cxsa, sg->addr); 378 omap_dma_chan_write(c, CEN, sg->en); 379 omap_dma_chan_write(c, CFN, sg->fn); 594 static size_t omap_dma_sg_size(struct omap_sg *sg) argument 596 return sg->en * sg->fn; 605 size += omap_dma_sg_size(&d->sg[ [all...] |
H A D | imx-dma.c | 142 /* For slave sg and cyclic */ 143 struct scatterlist *sg; member in struct:imxdma_desc 291 struct scatterlist *sg = d->sg; local 294 now = min(d->len, sg_dma_len(sg)); 299 imx_dmav1_writel(imxdma, sg->dma_address, 302 imx_dmav1_writel(imxdma, sg->dma_address, 334 d->sg && imxdma_hw_chain(imxdmac)) { 335 d->sg = sg_next(d->sg); 821 struct scatterlist *sg; local [all...] |
H A D | coh901318.h | 123 * @sg: scatter gather list 124 * @nents: number of entries in sg 136 struct scatterlist *sg, unsigned int nents,
|
/drivers/mmc/host/ |
H A D | sh_mmcif.c | 289 struct scatterlist *sg = data->sg; local 295 ret = dma_map_sg(chan->device->dev, sg, data->sg_len, 299 desc = dmaengine_prep_slave_sg(chan, sg, ret, 331 dev_dbg(&host->pd->dev, "%s(): desc %p, cookie %d, sg[%d]\n", __func__, 338 struct scatterlist *sg = data->sg; local 344 ret = dma_map_sg(chan->device->dev, sg, data->sg_len, 348 desc = dmaengine_prep_slave_sg(chan, sg, ret, 571 /* data->sg [all...] |
H A D | tifm_sd.c | 170 struct scatterlist *sg = r_data->sg; local 178 cnt = sg[host->sg_pos].length - host->block_pos; 191 cnt = sg[host->sg_pos].length; 193 off = sg[host->sg_pos].offset + host->block_pos; 195 pg = nth_page(sg_page(&sg[host->sg_pos]), off >> PAGE_SHIFT); 226 struct scatterlist *sg = r_data->sg; local 234 cnt = sg[host->sg_pos].length - host->block_pos; 240 cnt = sg[hos 268 struct scatterlist *sg = NULL; local [all...] |
/drivers/crypto/qat/qat_common/ |
H A D | qat_algs.c | 592 struct scatterlist *sg; local 608 for_each_sg(assoc, sg, assoc_n, i) { 609 if (!sg->length) 612 sg_virt(sg), 613 sg->length, 615 bufl->bufers[bufs].len = sg->length; 627 for_each_sg(sgl, sg, n, i) { 630 bufl->bufers[y].addr = dma_map_single(dev, sg_virt(sg), 631 sg->length, 633 bufl->bufers[y].len = sg [all...] |
/drivers/hsi/controllers/ |
H A D | omap_ssi.h | 119 * @sg: Pointer to the current sg entry being served 123 struct scatterlist *sg; member in struct:gdd_trn
|
/drivers/ide/ |
H A D | ide-dma-sff.c | 120 struct scatterlist *sg; local 123 for_each_sg(hwif->sg_table, sg, cmd->sg_nents, i) { 126 cur_addr = sg_dma_address(sg); 127 cur_len = sg_dma_len(sg);
|
/drivers/mmc/card/ |
H A D | queue.h | 37 struct scatterlist *sg; member in struct:mmc_queue_req
|