/drivers/crypto/nx/ |
H A D | nx-sha256.c | 57 struct nx_sg *in_sg; local 76 in_sg = nx_ctx->in_sg; 93 in_sg = nx_build_sg_list(nx_ctx->in_sg, 97 in_sg = nx_build_sg_list(in_sg, (u8 *) data, 100 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * 134 in_sg 151 struct nx_sg *in_sg, *out_sg; local [all...] |
H A D | nx-sha512.c | 57 struct nx_sg *in_sg; local 76 in_sg = nx_ctx->in_sg; 93 in_sg = nx_build_sg_list(nx_ctx->in_sg, 97 in_sg = nx_build_sg_list(in_sg, (u8 *) data, 100 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * 136 in_sg 153 struct nx_sg *in_sg, *out_sg; local [all...] |
H A D | nx-aes-gcm.c | 131 struct nx_sg *nx_sg = nx_ctx->in_sg; 164 nx_sg = nx_walk_and_build(nx_ctx->in_sg, nx_ctx->ap->sglen, 166 nx_ctx->op_aead.inlen = (nx_ctx->in_sg - nx_sg) 227 nx_sg = nx_walk_and_build(nx_ctx->in_sg, nx_ctx->ap->sglen, 229 nx_ctx->op.inlen = (nx_ctx->in_sg - nx_sg) 266 struct nx_sg *in_sg, *out_sg; local 283 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) desc->info, 287 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeo [all...] |
H A D | nx-aes-xcbc.c | 74 struct nx_sg *in_sg, *out_sg; local 90 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) keys, sizeof(keys), 94 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); 108 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) keys[1], sizeof(keys[1]), 112 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); 161 struct nx_sg *in_sg; local 263 struct nx_sg *in_sg, *out_sg; local [all...] |
H A D | nx.h | 128 struct nx_sg *in_sg; /* aligned pointer into kmem to an sg list */ member in struct:nx_crypto_ctx
|
/drivers/crypto/ |
H A D | atmel-tdes.c | 115 struct scatterlist *in_sg; member in struct:atmel_tdes_dev 325 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 512 in = IS_ALIGNED((u32)dd->in_sg->offset, sizeof(u32)) && 513 IS_ALIGNED(dd->in_sg->length, dd->ctx->block_size); 518 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) 524 count = min(dd->total, sg_dma_len(dd->in_sg)); 527 err = dma_map_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 537 dma_unmap_sg(dd->dev, dd->in_sg, 1, 542 addr_in = sg_dma_address(dd->in_sg); 549 count = atmel_tdes_sg_copy(&dd->in_sg, [all...] |
H A D | atmel-aes.c | 122 struct scatterlist *in_sg; member in struct:atmel_aes_dev 395 dd->nb_in_sg = atmel_aes_sg_length(dd->req, dd->in_sg); 403 dd->bufcnt = sg_copy_to_buffer(dd->in_sg, dd->nb_in_sg, 426 in = IS_ALIGNED((u32)dd->in_sg->offset, sizeof(u32)) && 427 IS_ALIGNED(dd->in_sg->length, dd->ctx->block_size); 432 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) 438 count = min(dd->total, sg_dma_len(dd->in_sg)); 441 err = dma_map_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 451 dma_unmap_sg(dd->dev, dd->in_sg, 1, 456 addr_in = sg_dma_address(dd->in_sg); [all...] |
H A D | omap-aes.c | 169 struct scatterlist *in_sg; member in struct:omap_aes_dev 417 struct scatterlist *in_sg, struct scatterlist *out_sg, 427 scatterwalk_start(&dd->in_walk, dd->in_sg); 436 dma_sync_sg_for_device(dd->dev, dd->in_sg, in_sg_len, DMA_TO_DEVICE); 455 tx_in = dmaengine_prep_slave_sg(dd->dma_lch_in, in_sg, in_sg_len, 506 err = dma_map_sg(dd->dev, dd->in_sg, dd->in_sg_len, 521 err = omap_aes_crypt_dma(tfm, dd->in_sg, dd->out_sg, dd->in_sg_len, 524 dma_unmap_sg(dd->dev, dd->in_sg, dd->in_sg_len, DMA_TO_DEVICE); 586 sg_copy_buf(buf_in, dd->in_sg, 0, dd->total, 0); 590 dd->in_sg 416 omap_aes_crypt_dma(struct crypto_tfm *tfm, struct scatterlist *in_sg, struct scatterlist *out_sg, int in_sg_len, int out_sg_len) argument [all...] |
H A D | omap-des.c | 150 struct scatterlist *in_sg; member in struct:omap_des_dev 401 struct scatterlist *in_sg, struct scatterlist *out_sg, 411 scatterwalk_start(&dd->in_walk, dd->in_sg); 420 dma_sync_sg_for_device(dd->dev, dd->in_sg, in_sg_len, DMA_TO_DEVICE); 439 tx_in = dmaengine_prep_slave_sg(dd->dma_lch_in, in_sg, in_sg_len, 490 err = dma_map_sg(dd->dev, dd->in_sg, dd->in_sg_len, 505 err = omap_des_crypt_dma(tfm, dd->in_sg, dd->out_sg, dd->in_sg_len, 508 dma_unmap_sg(dd->dev, dd->in_sg, dd->in_sg_len, DMA_TO_DEVICE); 539 dma_unmap_sg(dd->dev, dd->in_sg, dd->in_sg_len, DMA_TO_DEVICE); 579 sg_copy_buf(buf_in, dd->in_sg, 400 omap_des_crypt_dma(struct crypto_tfm *tfm, struct scatterlist *in_sg, struct scatterlist *out_sg, int in_sg_len, int out_sg_len) argument [all...] |
H A D | sahara.c | 159 struct scatterlist *in_sg; member in struct:sahara_dev 410 dma_unmap_sg(dev->device, dev->in_sg, dev->nb_in_sg, 457 dev->nb_in_sg = sahara_sg_length(dev->in_sg, dev->total); 465 ret = dma_map_sg(dev->device, dev->in_sg, dev->nb_in_sg, 480 sg = dev->in_sg; 526 dma_unmap_sg(dev->device, dev->in_sg, dev->nb_in_sg, 564 dev->in_sg = req->src;
|
/drivers/usb/wusbcore/ |
H A D | wa-xfer.c | 1016 * subset of the in_sg that matches the buffer subset 1019 static struct scatterlist *wa_xfer_create_subset_sg(struct scatterlist *in_sg, argument 1026 struct scatterlist *current_xfer_sg = in_sg;
|