/drivers/net/irda/ |
H A D | sa1100_ir.c | 64 struct sa1100_buf dma_tx; member in struct:sa1100_irda 230 dma_unmap_sg(si->dma_tx.dev, &si->dma_tx.sg, 1, DMA_TO_DEVICE); 231 dev_kfree_skb(si->dma_tx.skb); 232 si->dma_tx.skb = NULL; 235 dev->stats.tx_bytes += sg_dma_len(&si->dma_tx.sg); 262 si->dma_tx.skb = skb; 263 sg_set_buf(&si->dma_tx.sg, si->tx_buff.data, si->tx_buff.len); 264 if (dma_map_sg(si->dma_tx.dev, &si->dma_tx [all...] |
/drivers/spi/ |
H A D | spi-rockchip.c | 200 struct rockchip_spi_dma_data dma_tx; member in struct:rockchip_spi 326 dmaengine_terminate_all(rs->dma_tx.ch); 457 txconf.direction = rs->dma_tx.direction; 458 txconf.dst_addr = rs->dma_tx.addr; 461 dmaengine_slave_config(rs->dma_tx.ch, &txconf); 464 rs->dma_tx.ch, 466 rs->dma_tx.direction, DMA_PREP_INTERRUPT); 486 dma_async_issue_pending(rs->dma_tx.ch); 690 rs->dma_tx.ch = dma_request_slave_channel(rs->dev, "tx"); 691 if (!rs->dma_tx [all...] |
H A D | spi-omap2-mcspi.c | 106 struct dma_chan *dma_tx; member in struct:omap2_mcspi_dma 407 if (mcspi_dma->dma_tx) { 411 dmaengine_slave_config(mcspi_dma->dma_tx, &cfg); 417 tx = dmaengine_prep_slave_sg(mcspi_dma->dma_tx, &sg, 1, 427 dma_async_issue_pending(mcspi_dma->dma_tx); 971 mcspi_dma->dma_tx = 976 if (!mcspi_dma->dma_tx) { 1013 if (!mcspi_dma->dma_rx || !mcspi_dma->dma_tx) { 1053 if (mcspi_dma->dma_tx) { 1054 dma_release_channel(mcspi_dma->dma_tx); [all...] |
H A D | spi-sh-msiof.c | 657 dma_sync_single_for_device(p->master->dma_tx->device->dev, 659 desc_tx = dmaengine_prep_slave_single(p->master->dma_tx, 695 dma_async_issue_pending(p->master->dma_tx); 732 dmaengine_terminate_all(p->master->dma_tx); 804 while (master->dma_tx && len > 15) { 1035 master->dma_tx = sh_msiof_request_dma_chan(dev, DMA_MEM_TO_DEV, 1038 if (!master->dma_tx) 1055 tx_dev = master->dma_tx->device->dev; 1079 dma_release_channel(master->dma_tx); 1080 master->dma_tx [all...] |
H A D | spi-imx.c | 811 if (master->dma_tx) { 812 dma_release_channel(master->dma_tx); 813 master->dma_tx = NULL; 827 master->dma_tx = dma_request_slave_channel(dev, "tx"); 828 if (!master->dma_tx) { 838 ret = dmaengine_slave_config(master->dma_tx, &slave_config); 901 desc_tx = dmaengine_prep_slave_sg(master->dma_tx, 939 dma_async_issue_pending(master->dma_tx); 948 dmaengine_terminate_all(master->dma_tx);
|
H A D | spi-rspi.c | 497 desc_tx = dmaengine_prep_slave_sg(rspi->master->dma_tx, 537 dma_async_issue_pending(rspi->master->dma_tx); 547 dmaengine_terminate_all(rspi->master->dma_tx); 959 master->dma_tx = rspi_request_dma_chan(dev, DMA_MEM_TO_DEV, dma_tx_id, 961 if (!master->dma_tx) 967 dma_release_channel(master->dma_tx); 968 master->dma_tx = NULL; 979 if (master->dma_tx) 980 dma_release_channel(master->dma_tx);
|
H A D | spi-ep93xx.c | 83 * @dma_tx: TX DMA channel 102 struct dma_chan *dma_tx; member in struct:ep93xx_spi 461 chan = espi->dma_tx; 543 chan = espi->dma_tx; 584 dma_async_issue_pending(espi->dma_tx); 825 espi->dma_tx = dma_request_channel(mask, ep93xx_spi_dma_filter, 827 if (!espi->dma_tx) { 849 if (espi->dma_tx) { 850 dma_release_channel(espi->dma_tx);
|
H A D | spi-davinci.c | 136 struct dma_chan *dma_tx; member in struct:davinci_spi 653 dmaengine_slave_config(dspi->dma_tx, &dma_tx_conf); 689 txdesc = dmaengine_prep_slave_sg(dspi->dma_tx, 707 dma_async_issue_pending(dspi->dma_tx); 822 dspi->dma_tx = dma_request_channel(mask, edma_filter_fn, 824 if (!dspi->dma_tx) { 1085 dma_release_channel(dspi->dma_tx);
|
/drivers/net/ethernet/stmicro/stmmac/ |
H A D | dwmac1000_dma.c | 34 int burst_len, u32 dma_tx, u32 dma_rx, int atds) 107 writel(dma_tx, ioaddr + DMA_TX_BASE_ADDR); 33 dwmac1000_dma_init(void __iomem *ioaddr, int pbl, int fb, int mb, int burst_len, u32 dma_tx, u32 dma_rx, int atds) argument
|
H A D | dwmac100_dma.c | 36 int burst_len, u32 dma_tx, u32 dma_rx, int atds) 63 writel(dma_tx, ioaddr + DMA_TX_BASE_ADDR); 35 dwmac100_dma_init(void __iomem *ioaddr, int pbl, int fb, int mb, int burst_len, u32 dma_tx, u32 dma_rx, int atds) argument
|
H A D | chain_mode.c | 36 struct dma_desc *desc = priv->dma_tx + entry; 58 desc = priv->dma_tx + entry;
|
H A D | ring_mode.c | 43 desc = priv->dma_tx + entry; 70 desc = priv->dma_tx + entry;
|
H A D | stmmac.h | 45 struct dma_desc *dma_tx; member in struct:stmmac_priv
|
H A D | stmmac_main.c | 366 desc = (priv->dma_tx + entry); 913 stmmac_display_ring((void *)priv->dma_tx, txsize, 0); 961 priv->hw->desc->init_tx_desc(&priv->dma_tx[i], 1072 priv->hw->mode->init(priv->dma_tx, priv->dma_tx_phy, 1083 p = priv->dma_tx + i; 1123 p = priv->dma_tx + i; 1202 priv->dma_tx = dma_alloc_coherent(priv->device, txsize * 1206 if (!priv->dma_tx) { 1237 priv->dma_tx, priv->dma_tx_phy); 1301 p = priv->dma_tx [all...] |
/drivers/net/ethernet/samsung/sxgbe/ |
H A D | sxgbe_dma.h | 27 int pbl, dma_addr_t dma_tx, dma_addr_t dma_rx,
|
H A D | sxgbe_dma.c | 47 int fix_burst, int pbl, dma_addr_t dma_tx, 69 writel(upper_32_bits(dma_tx), 71 writel(lower_32_bits(dma_tx), 83 dma_addr = dma_tx + ((t_rsize - 1) * SXGBE_DESC_SIZE_BYTES); 46 sxgbe_dma_channel_init(void __iomem *ioaddr, int cha_num, int fix_burst, int pbl, dma_addr_t dma_tx, dma_addr_t dma_rx, int t_rsize, int r_rsize) argument
|
H A D | sxgbe_main.c | 338 priv->hw->desc->init_tx_desc(&priv->txq[j]->dma_tx[i]); 385 tx_ring->dma_tx = dma_zalloc_coherent(dev, 388 if (!tx_ring->dma_tx) 417 tx_ring->dma_tx, tx_ring->dma_tx_phy); 526 tx_ring->dma_tx, tx_ring->dma_tx_phy); 595 struct sxgbe_tx_norm_desc *tdesc = txqueue->dma_tx + dma_desc; 727 p = tqueue->dma_tx + entry; 1300 tx_desc = tqueue->dma_tx + entry; 1325 first_desc = tqueue->dma_tx + entry; 1347 tx_desc = tqueue->dma_tx [all...] |
H A D | sxgbe_common.h | 382 struct sxgbe_tx_norm_desc *dma_tx; member in struct:sxgbe_tx_queue
|
/drivers/net/ethernet/micrel/ |
H A D | ks8842.c | 159 #define KS8842_USE_DMA(adapter) (((adapter)->dma_tx.channel != -1) && \ 171 struct ks8842_tx_dma_ctl dma_tx; member in struct:ks8842_adapter 434 struct ks8842_tx_dma_ctl *ctl = &adapter->dma_tx; 857 struct ks8842_tx_dma_ctl *ctl = &adapter->dma_tx; 873 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; 895 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; 927 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; 1002 adapter->dma_tx.channel = -1; 1056 if (adapter->dma_tx.adesc) 1183 adapter->dma_tx [all...] |
/drivers/net/wan/ |
H A D | z85230.c | 541 if(!chan->dma_tx) 572 if(chan->dma_tx) 807 c->dma_tx = 0; 897 c->dma_tx = 1; 1098 c->dma_tx = 1; 1444 if(c->dma_tx) 1465 if(c->dma_tx) 1747 if(c->dma_tx && ((unsigned long)(virt_to_bus(skb->data+skb->len))>=16*1024*1024 || spans_boundary(skb)))
|
H A D | z85230.h | 302 u8 dma_tx; /* TX is to use DMA */ member in struct:z8530_channel
|
/drivers/net/ethernet/calxeda/ |
H A D | xgmac.c | 370 struct xgmac_dma_desc *dma_tx; member in struct:xgmac_priv 760 priv->dma_tx = dma_alloc_coherent(priv->device, 765 if (!priv->dma_tx) 770 priv->dma_rx, priv->dma_tx, 781 desc_init_tx_desc(priv->dma_tx, DMA_TX_RING_SZ); 832 p = priv->dma_tx + i; 853 if (priv->dma_tx) { 856 priv->dma_tx, priv->dma_tx_phy); 857 priv->dma_tx = NULL; 881 struct xgmac_dma_desc *p = priv->dma_tx [all...] |
/drivers/mmc/host/ |
H A D | jz4740_mmc.c | 150 struct dma_chan *dma_tx; member in struct:jz4740_mmc_host 170 dma_release_channel(host->dma_tx); 181 host->dma_tx = dma_request_channel(mask, NULL, host); 182 if (!host->dma_tx) { 183 dev_err(mmc_dev(host->mmc), "Failed to get dma_tx channel\n"); 199 dma_release_channel(host->dma_tx); 211 return (data->flags & MMC_DATA_READ) ? host->dma_rx : host->dma_tx; 287 chan = host->dma_tx;
|
H A D | davinci_mmc.c | 206 struct dma_chan *dma_tx; member in struct:mmc_davinci_host 413 sync_dev = host->dma_tx; 433 chan = host->dma_tx; 434 dmaengine_slave_config(host->dma_tx, &dma_tx_conf); 436 desc = dmaengine_prep_slave_sg(host->dma_tx, 514 dma_release_channel(host->dma_tx); 526 host->dma_tx = 529 if (!host->dma_tx) { 530 dev_err(mmc_dev(host->mmc), "Can't get dma_tx channel\n"); 546 dma_release_channel(host->dma_tx); [all...] |
H A D | omap.c | 132 struct dma_chan *dma_tx; member in struct:mmc_omap_host 414 c = host->dma_tx; 1003 c = host->dma_tx; 1077 host->dma_tx : host->dma_rx; 1393 host->dma_tx = dma_request_slave_channel_compat(mask, 1395 if (!host->dma_tx) 1445 if (host->dma_tx) 1446 dma_release_channel(host->dma_tx); 1475 if (host->dma_tx) 1476 dma_release_channel(host->dma_tx); [all...] |