Searched refs:async_tx (Results 1 - 12 of 12) sorted by relevance
/drivers/dma/ |
H A D | mv_xor.c | 40 container_of(tx, struct mv_xor_desc_slot, async_tx) 281 mv_chan_set_next_descriptor(mv_chan, sw_desc->async_tx.phys); 291 BUG_ON(desc->async_tx.cookie < 0); 293 if (desc->async_tx.cookie > 0) { 294 cookie = desc->async_tx.cookie; 299 if (desc->async_tx.callback) 300 desc->async_tx.callback( 301 desc->async_tx.callback_param); 311 enum dma_ctrl_flags flags = desc->async_tx.flags; 343 dma_run_dependencies(&desc->async_tx); [all...] |
H A D | fsldma.c | 392 set_desc_next(chan, &tail->hw, desc->async_tx.phys); 422 child->async_tx.cookie = cookie; 454 dma_async_tx_descriptor_init(&desc->async_tx, &chan->common); 455 desc->async_tx.tx_submit = fsl_dma_tx_submit; 456 desc->async_tx.phys = pdesc; 514 dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys); 528 dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys); 568 new->async_tx.cookie = -EBUSY; 569 new->async_tx.flags = flags; 577 return &new->async_tx; [all...] |
H A D | iop-adma.c | 43 container_of(tx, struct iop_adma_desc_slot, async_tx) 65 struct dma_async_tx_descriptor *tx = &desc->async_tx; 101 struct dma_async_tx_descriptor *tx = &desc->async_tx; 140 struct dma_async_tx_descriptor *tx = &desc->async_tx; 177 if (!async_tx_test_ack(&desc->async_tx)) 212 iter->async_tx.cookie, iter->idx, busy, 213 iter->async_tx.phys, iop_desc_get_next_desc(iter), 214 async_tx_test_ack(&iter->async_tx)); 216 prefetch(&_iter->async_tx); 229 if (iter->async_tx [all...] |
H A D | shdma.c | 314 chunk->async_tx.cookie > 0 || 315 chunk->async_tx.cookie == -EBUSY || 320 chunk->async_tx.callback = NULL; 326 last->async_tx.callback = callback; 327 last->async_tx.callback_param = tx->callback_param; 330 tx->cookie, &last->async_tx, sh_chan->id, 435 dma_async_tx_descriptor_init(&desc->async_tx, 437 desc->async_tx.tx_submit = sh_dmae_tx_submit; 538 new->async_tx.cookie = -EBUSY; 542 new->async_tx [all...] |
H A D | shdma.h | 63 #define tx_to_sh_desc(tx) container_of(tx, struct sh_desc, async_tx)
|
H A D | fsldma.h | 101 struct dma_async_tx_descriptor async_tx; member in struct:fsl_desc_sw 162 #define tx_to_fsl_desc(tx) container_of(tx, struct fsl_desc_sw, async_tx)
|
H A D | mv_xor.h | 130 * @async_tx: support for the async_tx api 148 struct dma_async_tx_descriptor async_tx; member in struct:mv_xor_desc_slot
|
H A D | Kconfig | 273 bool "Async_tx: Offload support for the async_tx api" 276 This allows the async_tx api to take advantage of offload engines for
|
/drivers/dma/ppc4xx/ |
H A D | adma.h | 23 container_of(tx, struct ppc440spe_adma_desc_slot, async_tx) 128 * @async_tx: support for the async_tx api 152 struct dma_async_tx_descriptor async_tx; member in struct:ppc440spe_adma_desc_slot
|
H A D | adma.c | 36 #include <linux/async_tx.h> 1722 if (!(desc->async_tx.flags & DMA_COMPL_SKIP_DEST_UNMAP)) { 1733 if (!(desc->async_tx.flags & DMA_COMPL_SKIP_SRC_UNMAP)) { 1755 BUG_ON(desc->async_tx.cookie < 0); 1756 if (desc->async_tx.cookie > 0) { 1757 cookie = desc->async_tx.cookie; 1758 desc->async_tx.cookie = 0; 1763 if (desc->async_tx.callback) 1764 desc->async_tx.callback( 1765 desc->async_tx [all...] |
/drivers/tty/serial/ |
H A D | sh-sci.c | 1397 async_tx);
|
/drivers/md/ |
H A D | raid5.c | 49 #include <linux/async_tx.h>
|
Completed in 193 milliseconds