/drivers/spi/ |
H A D | spi-ep93xx.c | 78 * @tx: current byte in transfer to transmit 98 size_t tx; member in struct:ep93xx_spi 344 tx_val = ((u16 *)t->tx_buf)[espi->tx]; 346 espi->tx += sizeof(tx_val); 351 tx_val = ((u8 *)t->tx_buf)[espi->tx]; 353 espi->tx += sizeof(tx_val); 399 while (espi->fifo_level < SPI_FIFO_SIZE && espi->tx < t->len) { 620 espi->tx = 0; 823 espi->dma_tx_data.name = "ep93xx-spi-tx";
|
H A D | spi-fsl-dspi.c | 119 void *tx; member in struct:fsl_dspi 196 d16 = *(u16 *)dspi->tx; 197 dspi->tx += 2; 211 d8 = *(u8 *)dspi->tx; 212 dspi->tx++; 288 dspi->tx = (void *)t->tx_buf; 289 dspi->tx_end = dspi->tx + t->len; 297 if (!dspi->tx)
|
H A D | spi-fsl-espi.c | 115 const u32 *tx = mpc8xxx_spi->tx; local 117 if (!tx) 120 data = *tx++ << mpc8xxx_spi->tx_shift; 127 mpc8xxx_spi->tx = tx; 227 mpc8xxx_spi->tx = t->tx_buf; 573 /* Get interrupt events(tx/rx) */
|
H A D | spi-sirf.c | 118 /* 256 bytes rx/tx FIFO */ 128 * only if the rx/tx buffer and transfer size are 4-bytes aligned, we use dma 147 /* rx & tx bufs from the spi_transfer */ 148 const void *tx; member in struct:sirfsoc_spi 153 /* get word from tx buffer for sending */ 160 /* rx & tx DMA channels */ 169 * if tx size is not more than 4 and rx size is NULL, use 194 const u8 *tx = sspi->tx; local 196 if (tx) { 223 const u16 *tx = sspi->tx; local 253 const u32 *tx = sspi->tx; local [all...] |
/drivers/staging/comedi/drivers/ |
H A D | jr3_pci.h | 246 tx, enumerator in enum:link_types
|
H A D | vmk80xx.c | 172 unsigned char tx[1]; local 178 tx[0] = VMK8061_CMD_RD_PWR_STAT; 185 usb_bulk_msg(usb, tx_pipe, tx, 1, NULL, devpriv->ep_tx->bInterval); 197 unsigned char tx[1]; local 204 tx[0] = VMK8061_CMD_RD_VERSION; 210 usb_bulk_msg(usb, tx_pipe, tx, 1, NULL, devpriv->ep_tx->bInterval);
|
/drivers/staging/lustre/lnet/klnds/socklnd/ |
H A D | socklnd_cb.c | 32 ksock_tx_t *tx = NULL; local 37 /* searching for a noop tx in free list */ 41 tx = list_entry(ksocknal_data.ksnd_idle_noop_txs. \ 43 LASSERT(tx->tx_desc_size == size); 44 list_del(&tx->tx_list); 50 if (tx == NULL) 51 LIBCFS_ALLOC(tx, size); 53 if (tx == NULL) 56 atomic_set(&tx->tx_refcount, 1); 57 tx 70 ksock_tx_t *tx; local 94 ksocknal_free_tx(ksock_tx_t *tx) argument 111 ksocknal_send_iov(ksock_conn_t *conn, ksock_tx_t *tx) argument 148 ksocknal_send_kiov(ksock_conn_t *conn, ksock_tx_t *tx) argument 186 ksocknal_transmit(ksock_conn_t *conn, ksock_tx_t *tx) argument 389 ksocknal_tx_done(lnet_ni_t *ni, ksock_tx_t *tx) argument 410 ksock_tx_t *tx; local 433 ksocknal_check_zc_req(ksock_tx_t *tx) argument 477 ksocknal_uncheck_zc_req(ksock_tx_t *tx) argument 503 ksocknal_process_transmit(ksock_conn_t *conn, ksock_tx_t *tx) argument 621 ksocknal_find_conn_locked(ksock_peer_t *peer, ksock_tx_t *tx, int nonblk) argument 678 ksocknal_tx_prep(ksock_conn_t *conn, ksock_tx_t *tx) argument 688 ksocknal_queue_tx_locked(ksock_tx_t *tx, ksock_conn_t *conn) argument 836 ksocknal_launch_packet(lnet_ni_t *ni, ksock_tx_t *tx, lnet_process_id_t id) argument 937 ksock_tx_t *tx; local 1395 ksock_tx_t *tx; local 2333 ksock_tx_t *tx; local 2360 ksock_tx_t *tx; local 2422 ksock_tx_t *tx; local 2457 ksock_tx_t *tx = local [all...] |
/drivers/staging/media/lirc/ |
H A D | lirc_zilog.c | 115 struct IR_tx *tx; member in struct:IR 194 * ir->tx set to NULL and deallocated - happens before ir->tx->ir put() 287 struct IR_tx *tx; local 290 tx = ir->tx; 291 if (tx != NULL) 292 kref_get(&tx->ref); 294 return tx; 299 struct IR_tx *tx local 308 put_ir_tx(struct IR_tx *tx, bool ir_devices_lock_held) argument 333 struct IR_tx *tx; local 652 send_data_block(struct IR_tx *tx, unsigned char *data_block) argument 677 send_boot_data(struct IR_tx *tx) argument 754 fw_load(struct IR_tx *tx) argument 983 send_code(struct IR_tx *tx, unsigned int code, unsigned int key) argument 1104 struct IR_tx *tx; local 1407 struct IR_tx *tx = i2c_get_clientdata(client); local 1449 struct IR_tx *tx; local [all...] |
/drivers/staging/rtl8188eu/core/ |
H A D | rtw_ap.c | 1357 void update_beacon(struct adapter *padapter, u8 ie_id, u8 *oui, u8 tx) argument 1403 if (tx)
|
/drivers/staging/rtl8723au/core/ |
H A D | rtw_ap.c | 1244 void update_beacon23a(struct rtw_adapter *padapter, u8 ie_id, u8 *oui, u8 tx) argument 1303 if (tx)
|
/drivers/tty/serial/ |
H A D | 68328serial.c | 353 unsigned short tx; local 359 tx = uart->utx.w; 363 if (tx & UTX_TX_AVAIL)
|
H A D | max3100.c | 196 static int max3100_sr(struct max3100_port *s, u16 tx, u16 *rx) argument 207 etx = cpu_to_be16(tx); 217 dev_dbg(&s->spi->dev, "%04x - %04x\n", tx, *rx); 265 u16 tx, rx; local 293 tx = 0xffff; 295 tx = s->port.x_char; 296 s->port.icount.tx++; 300 tx = xmit->buf[xmit->tail]; 303 s->port.icount.tx++; 305 if (tx ! 578 u16 tx, rx; local 623 u16 tx, rx; local 748 u16 tx, rx; local 869 u16 tx, rx; local [all...] |
/drivers/ata/ |
H A D | pata_arasan_cf.c | 397 struct dma_async_tx_descriptor *tx; local 403 tx = chan->device->device_prep_dma_memcpy(chan, dest, src, len, flags); 404 if (!tx) { 409 tx->callback = dma_callback; 410 tx->callback_param = acdev; 411 cookie = tx->tx_submit(tx);
|
/drivers/atm/ |
H A D | ambassador.c | 450 static void tx_complete (amb_dev * dev, tx_out * tx) { argument 451 tx_simple * tx_descr = bus_to_virt (tx->handle); 454 PRINTD (DBG_FLOW|DBG_TX, "tx_complete %p %p", dev, tx); 457 atomic_inc(&ATM_SKB(skb)->vcc->stats->tx); 627 static int tx_give (amb_dev * dev, tx_in * tx) { argument 641 *txq->in.ptr = *tx; 1315 tx_in tx; local 1372 tx.vc = cpu_to_be16 (vcc->tx_frame_bits | vc); 1373 tx.tx_descr_length = cpu_to_be16 (sizeof(tx_frag)+sizeof(tx_frag_end)); 1374 tx [all...] |
H A D | eni.c | 188 if (eni_dev->tx[i].send) 190 eni_dev->tx[i].send,eni_dev->tx[i].words*4); 1033 struct eni_tx *tx; local 1047 tx = eni_vcc->tx; 1048 NULLCHECK(tx); 1087 if (!NEPMOK(tx->tx_pos,size+TX_GAP, 1088 eni_in(MID_TX_RDPTR(tx->index)),tx 1181 struct eni_tx *tx; local 1208 struct eni_tx *tx; local 1289 struct eni_tx *tx; local 1955 struct eni_tx *tx = ENI_VCC(vcc)->tx; local 2154 struct eni_tx *tx = eni_dev->tx+i; local [all...] |
/drivers/dma/ |
H A D | amba-pl08x.c | 300 static inline struct pl08x_txd *to_pl08x_txd(struct dma_async_tx_descriptor *tx) argument 302 return container_of(tx, struct pl08x_txd, vd.tx); 397 struct pl08x_txd *txd = to_pl08x_txd(&vd->tx); 1170 struct pl08x_txd *txd = to_pl08x_txd(&vd->tx); 1171 struct pl08x_dma_chan *plchan = to_pl08x_chan(vd->tx.chan); 1173 dma_descriptor_unmap(&vd->tx); 1245 struct pl08x_txd *txd = to_pl08x_txd(&vd->tx); 1815 struct pl08x_txd *tx; local 1825 tx [all...] |
H A D | cppi41.c | 339 static dma_cookie_t cppi41_tx_submit(struct dma_async_tx_descriptor *tx) argument 343 cookie = dma_cookie_assign(tx);
|
H A D | ep93xx_dma.c | 815 * @tx: descriptor to be executed 821 static dma_cookie_t ep93xx_dma_tx_submit(struct dma_async_tx_descriptor *tx) argument 823 struct ep93xx_dma_chan *edmac = to_ep93xx_dma_chan(tx->chan); 829 cookie = dma_cookie_assign(tx); 831 desc = container_of(tx, struct ep93xx_dma_desc, txd); 1094 * @flags: tx descriptor status flags
|
H A D | imx-dma.c | 744 static dma_cookie_t imxdma_tx_submit(struct dma_async_tx_descriptor *tx) argument 746 struct imxdma_channel *imxdmac = to_imxdma_chan(tx->chan); 753 cookie = dma_cookie_assign(tx);
|
H A D | mmp_pdma.c | 135 #define tx_to_mmp_pdma_desc(tx) \ 136 container_of(tx, struct mmp_pdma_desc_sw, async_tx) 337 static dma_cookie_t mmp_pdma_tx_submit(struct dma_async_tx_descriptor *tx) argument 339 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(tx->chan); 340 struct mmp_pdma_desc_sw *desc = tx_to_mmp_pdma_desc(tx);
|
H A D | nbpfaxi.c | 649 static dma_cookie_t nbpf_tx_submit(struct dma_async_tx_descriptor *tx) argument 651 struct nbpf_desc *desc = container_of(tx, struct nbpf_desc, async_tx); 657 cookie = dma_cookie_assign(tx);
|
H A D | s3c24xx-dma.c | 408 struct s3c24xx_txd *to_s3c24xx_txd(struct dma_async_tx_descriptor *tx) argument 410 return container_of(tx, struct s3c24xx_txd, vd.tx); 505 struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx); 635 struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx); 636 struct s3c24xx_dma_chan *s3cchan = to_s3c24xx_dma_chan(vd->tx.chan); 639 dma_descriptor_unmap(&vd->tx); 796 txd = to_s3c24xx_txd(&vd->tx);
|
/drivers/dma/ipu/ |
H A D | ipu_idmac.c | 822 static dma_cookie_t idmac_tx_submit(struct dma_async_tx_descriptor *tx) argument 824 struct idmac_tx_desc *desc = to_tx_desc(tx); 825 struct idmac_channel *ichan = to_idmac_chan(tx->chan); 826 struct idmac *idmac = to_idmac(tx->chan->device); 836 dev_err(dev, "Descriptor %p not prepared!\n", tx); 842 async_tx_clear_ack(tx); 869 cookie = dma_cookie_assign(tx); 900 tx->cookie = cookie;
|
/drivers/dma/xilinx/ |
H A D | xilinx_vdma.c | 255 #define to_vdma_tx_descriptor(tx) \ 256 container_of(tx, struct xilinx_vdma_tx_descriptor, async_tx) 883 * @tx: Async transaction descriptor 887 static dma_cookie_t xilinx_vdma_tx_submit(struct dma_async_tx_descriptor *tx) argument 889 struct xilinx_vdma_tx_descriptor *desc = to_vdma_tx_descriptor(tx); 890 struct xilinx_vdma_chan *chan = to_xilinx_chan(tx->chan); 907 cookie = dma_cookie_assign(tx);
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_verbs.c | 1004 struct ipath_verbs_txreq *tx = NULL; local 1012 tx = list_entry(l, struct ipath_verbs_txreq, txreq.list); 1015 return tx; 1019 struct ipath_verbs_txreq *tx) 1024 list_add(&tx->txreq.list, &dev->txreq_free); 1030 struct ipath_verbs_txreq *tx = cookie; local 1031 struct ipath_qp *qp = tx->qp; 1039 if (tx->wqe) 1040 ipath_send_complete(qp, tx->wqe, ibs); 1047 } else if (tx 1018 put_txreq(struct ipath_ibdev *dev, struct ipath_verbs_txreq *tx) argument 1104 struct ipath_verbs_txreq *tx; local 1993 struct ipath_verbs_txreq *tx; local [all...] |