Lines Matching refs:sg
152 struct scatterlist *sg;
154 unsigned int offset; /* offset in current sg */
585 ctx->sgl.page_link = ctx->sg->page_link;
586 ctx->sgl.offset = ctx->sg->offset;
588 sg_dma_address(&ctx->sgl) = sg_dma_address(ctx->sg);
641 while (ctx->sg) {
642 vaddr = kmap_atomic(sg_page(ctx->sg));
646 ctx->sg->length - ctx->offset);
654 if (ctx->offset == ctx->sg->length) {
655 ctx->sg = sg_next(ctx->sg);
656 if (ctx->sg)
712 #define SG_AA(sg) (IS_ALIGNED(sg->offset, sizeof(u32)))
714 #define SG_SA(sg, bs) (IS_ALIGNED(sg->length, bs))
720 struct scatterlist *sg;
730 * Don't use the sg interface when the transfer size is less
741 sg = ctx->sg;
744 if (!SG_AA(sg))
747 if (!sg_is_last(sg) && !SG_SA(sg, bs))
751 length = min(ctx->total, sg->length);
753 if (sg_is_last(sg)) {
755 /* not last sg must be BLOCK_SIZE aligned */
764 if (!dma_map_sg(dd->dev, ctx->sg, 1, DMA_TO_DEVICE)) {
776 ret = omap_sham_xmit_dma(dd, sg_dma_address(ctx->sg), length, final, 1);
778 dma_unmap_sg(dd->dev, ctx->sg, 1, DMA_TO_DEVICE);
814 dma_unmap_sg(dd->dev, ctx->sg, 1, DMA_TO_DEVICE);
815 if (ctx->sg->length == ctx->offset) {
816 ctx->sg = sg_next(ctx->sg);
817 if (ctx->sg)
1087 ctx->sg = req->src;