Searched defs:tx (Results 151 - 175 of 288) sorted by relevance

1234567891011>>

/drivers/spi/
H A Dspi-ep93xx.c78 * @tx: current byte in transfer to transmit
98 size_t tx; member in struct:ep93xx_spi
344 tx_val = ((u16 *)t->tx_buf)[espi->tx];
346 espi->tx += sizeof(tx_val);
351 tx_val = ((u8 *)t->tx_buf)[espi->tx];
353 espi->tx += sizeof(tx_val);
399 while (espi->fifo_level < SPI_FIFO_SIZE && espi->tx < t->len) {
620 espi->tx = 0;
823 espi->dma_tx_data.name = "ep93xx-spi-tx";
H A Dspi-fsl-dspi.c119 void *tx; member in struct:fsl_dspi
196 d16 = *(u16 *)dspi->tx;
197 dspi->tx += 2;
211 d8 = *(u8 *)dspi->tx;
212 dspi->tx++;
288 dspi->tx = (void *)t->tx_buf;
289 dspi->tx_end = dspi->tx + t->len;
297 if (!dspi->tx)
H A Dspi-fsl-espi.c115 const u32 *tx = mpc8xxx_spi->tx; local
117 if (!tx)
120 data = *tx++ << mpc8xxx_spi->tx_shift;
127 mpc8xxx_spi->tx = tx;
227 mpc8xxx_spi->tx = t->tx_buf;
573 /* Get interrupt events(tx/rx) */
H A Dspi-sirf.c118 /* 256 bytes rx/tx FIFO */
128 * only if the rx/tx buffer and transfer size are 4-bytes aligned, we use dma
147 /* rx & tx bufs from the spi_transfer */
148 const void *tx; member in struct:sirfsoc_spi
153 /* get word from tx buffer for sending */
160 /* rx & tx DMA channels */
169 * if tx size is not more than 4 and rx size is NULL, use
194 const u8 *tx = sspi->tx; local
196 if (tx) {
223 const u16 *tx = sspi->tx; local
253 const u32 *tx = sspi->tx; local
[all...]
/drivers/staging/comedi/drivers/
H A Djr3_pci.h246 tx, enumerator in enum:link_types
H A Dvmk80xx.c172 unsigned char tx[1]; local
178 tx[0] = VMK8061_CMD_RD_PWR_STAT;
185 usb_bulk_msg(usb, tx_pipe, tx, 1, NULL, devpriv->ep_tx->bInterval);
197 unsigned char tx[1]; local
204 tx[0] = VMK8061_CMD_RD_VERSION;
210 usb_bulk_msg(usb, tx_pipe, tx, 1, NULL, devpriv->ep_tx->bInterval);
/drivers/staging/lustre/lnet/klnds/socklnd/
H A Dsocklnd_cb.c32 ksock_tx_t *tx = NULL; local
37 /* searching for a noop tx in free list */
41 tx = list_entry(ksocknal_data.ksnd_idle_noop_txs. \
43 LASSERT(tx->tx_desc_size == size);
44 list_del(&tx->tx_list);
50 if (tx == NULL)
51 LIBCFS_ALLOC(tx, size);
53 if (tx == NULL)
56 atomic_set(&tx->tx_refcount, 1);
57 tx
70 ksock_tx_t *tx; local
94 ksocknal_free_tx(ksock_tx_t *tx) argument
111 ksocknal_send_iov(ksock_conn_t *conn, ksock_tx_t *tx) argument
148 ksocknal_send_kiov(ksock_conn_t *conn, ksock_tx_t *tx) argument
186 ksocknal_transmit(ksock_conn_t *conn, ksock_tx_t *tx) argument
389 ksocknal_tx_done(lnet_ni_t *ni, ksock_tx_t *tx) argument
410 ksock_tx_t *tx; local
433 ksocknal_check_zc_req(ksock_tx_t *tx) argument
477 ksocknal_uncheck_zc_req(ksock_tx_t *tx) argument
503 ksocknal_process_transmit(ksock_conn_t *conn, ksock_tx_t *tx) argument
621 ksocknal_find_conn_locked(ksock_peer_t *peer, ksock_tx_t *tx, int nonblk) argument
678 ksocknal_tx_prep(ksock_conn_t *conn, ksock_tx_t *tx) argument
688 ksocknal_queue_tx_locked(ksock_tx_t *tx, ksock_conn_t *conn) argument
836 ksocknal_launch_packet(lnet_ni_t *ni, ksock_tx_t *tx, lnet_process_id_t id) argument
937 ksock_tx_t *tx; local
1395 ksock_tx_t *tx; local
2333 ksock_tx_t *tx; local
2360 ksock_tx_t *tx; local
2422 ksock_tx_t *tx; local
2457 ksock_tx_t *tx = local
[all...]
/drivers/staging/media/lirc/
H A Dlirc_zilog.c115 struct IR_tx *tx; member in struct:IR
194 * ir->tx set to NULL and deallocated - happens before ir->tx->ir put()
287 struct IR_tx *tx; local
290 tx = ir->tx;
291 if (tx != NULL)
292 kref_get(&tx->ref);
294 return tx;
299 struct IR_tx *tx local
308 put_ir_tx(struct IR_tx *tx, bool ir_devices_lock_held) argument
333 struct IR_tx *tx; local
652 send_data_block(struct IR_tx *tx, unsigned char *data_block) argument
677 send_boot_data(struct IR_tx *tx) argument
754 fw_load(struct IR_tx *tx) argument
983 send_code(struct IR_tx *tx, unsigned int code, unsigned int key) argument
1104 struct IR_tx *tx; local
1407 struct IR_tx *tx = i2c_get_clientdata(client); local
1449 struct IR_tx *tx; local
[all...]
/drivers/staging/rtl8188eu/core/
H A Drtw_ap.c1357 void update_beacon(struct adapter *padapter, u8 ie_id, u8 *oui, u8 tx) argument
1403 if (tx)
/drivers/staging/rtl8723au/core/
H A Drtw_ap.c1244 void update_beacon23a(struct rtw_adapter *padapter, u8 ie_id, u8 *oui, u8 tx) argument
1303 if (tx)
/drivers/tty/serial/
H A D68328serial.c353 unsigned short tx; local
359 tx = uart->utx.w;
363 if (tx & UTX_TX_AVAIL)
H A Dmax3100.c196 static int max3100_sr(struct max3100_port *s, u16 tx, u16 *rx) argument
207 etx = cpu_to_be16(tx);
217 dev_dbg(&s->spi->dev, "%04x - %04x\n", tx, *rx);
265 u16 tx, rx; local
293 tx = 0xffff;
295 tx = s->port.x_char;
296 s->port.icount.tx++;
300 tx = xmit->buf[xmit->tail];
303 s->port.icount.tx++;
305 if (tx !
578 u16 tx, rx; local
623 u16 tx, rx; local
748 u16 tx, rx; local
869 u16 tx, rx; local
[all...]
/drivers/ata/
H A Dpata_arasan_cf.c397 struct dma_async_tx_descriptor *tx; local
403 tx = chan->device->device_prep_dma_memcpy(chan, dest, src, len, flags);
404 if (!tx) {
409 tx->callback = dma_callback;
410 tx->callback_param = acdev;
411 cookie = tx->tx_submit(tx);
/drivers/atm/
H A Dambassador.c450 static void tx_complete (amb_dev * dev, tx_out * tx) { argument
451 tx_simple * tx_descr = bus_to_virt (tx->handle);
454 PRINTD (DBG_FLOW|DBG_TX, "tx_complete %p %p", dev, tx);
457 atomic_inc(&ATM_SKB(skb)->vcc->stats->tx);
627 static int tx_give (amb_dev * dev, tx_in * tx) { argument
641 *txq->in.ptr = *tx;
1315 tx_in tx; local
1372 tx.vc = cpu_to_be16 (vcc->tx_frame_bits | vc);
1373 tx.tx_descr_length = cpu_to_be16 (sizeof(tx_frag)+sizeof(tx_frag_end));
1374 tx
[all...]
H A Deni.c188 if (eni_dev->tx[i].send)
190 eni_dev->tx[i].send,eni_dev->tx[i].words*4);
1033 struct eni_tx *tx; local
1047 tx = eni_vcc->tx;
1048 NULLCHECK(tx);
1087 if (!NEPMOK(tx->tx_pos,size+TX_GAP,
1088 eni_in(MID_TX_RDPTR(tx->index)),tx
1181 struct eni_tx *tx; local
1208 struct eni_tx *tx; local
1289 struct eni_tx *tx; local
1955 struct eni_tx *tx = ENI_VCC(vcc)->tx; local
2154 struct eni_tx *tx = eni_dev->tx+i; local
[all...]
/drivers/dma/
H A Damba-pl08x.c300 static inline struct pl08x_txd *to_pl08x_txd(struct dma_async_tx_descriptor *tx) argument
302 return container_of(tx, struct pl08x_txd, vd.tx);
397 struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
1170 struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
1171 struct pl08x_dma_chan *plchan = to_pl08x_chan(vd->tx.chan);
1173 dma_descriptor_unmap(&vd->tx);
1245 struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
1815 struct pl08x_txd *tx; local
1825 tx
[all...]
H A Dcppi41.c339 static dma_cookie_t cppi41_tx_submit(struct dma_async_tx_descriptor *tx) argument
343 cookie = dma_cookie_assign(tx);
H A Dep93xx_dma.c815 * @tx: descriptor to be executed
821 static dma_cookie_t ep93xx_dma_tx_submit(struct dma_async_tx_descriptor *tx) argument
823 struct ep93xx_dma_chan *edmac = to_ep93xx_dma_chan(tx->chan);
829 cookie = dma_cookie_assign(tx);
831 desc = container_of(tx, struct ep93xx_dma_desc, txd);
1094 * @flags: tx descriptor status flags
H A Dimx-dma.c744 static dma_cookie_t imxdma_tx_submit(struct dma_async_tx_descriptor *tx) argument
746 struct imxdma_channel *imxdmac = to_imxdma_chan(tx->chan);
753 cookie = dma_cookie_assign(tx);
H A Dmmp_pdma.c135 #define tx_to_mmp_pdma_desc(tx) \
136 container_of(tx, struct mmp_pdma_desc_sw, async_tx)
337 static dma_cookie_t mmp_pdma_tx_submit(struct dma_async_tx_descriptor *tx) argument
339 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(tx->chan);
340 struct mmp_pdma_desc_sw *desc = tx_to_mmp_pdma_desc(tx);
H A Dnbpfaxi.c649 static dma_cookie_t nbpf_tx_submit(struct dma_async_tx_descriptor *tx) argument
651 struct nbpf_desc *desc = container_of(tx, struct nbpf_desc, async_tx);
657 cookie = dma_cookie_assign(tx);
H A Ds3c24xx-dma.c408 struct s3c24xx_txd *to_s3c24xx_txd(struct dma_async_tx_descriptor *tx) argument
410 return container_of(tx, struct s3c24xx_txd, vd.tx);
505 struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx);
635 struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx);
636 struct s3c24xx_dma_chan *s3cchan = to_s3c24xx_dma_chan(vd->tx.chan);
639 dma_descriptor_unmap(&vd->tx);
796 txd = to_s3c24xx_txd(&vd->tx);
/drivers/dma/ipu/
H A Dipu_idmac.c822 static dma_cookie_t idmac_tx_submit(struct dma_async_tx_descriptor *tx) argument
824 struct idmac_tx_desc *desc = to_tx_desc(tx);
825 struct idmac_channel *ichan = to_idmac_chan(tx->chan);
826 struct idmac *idmac = to_idmac(tx->chan->device);
836 dev_err(dev, "Descriptor %p not prepared!\n", tx);
842 async_tx_clear_ack(tx);
869 cookie = dma_cookie_assign(tx);
900 tx->cookie = cookie;
/drivers/dma/xilinx/
H A Dxilinx_vdma.c255 #define to_vdma_tx_descriptor(tx) \
256 container_of(tx, struct xilinx_vdma_tx_descriptor, async_tx)
883 * @tx: Async transaction descriptor
887 static dma_cookie_t xilinx_vdma_tx_submit(struct dma_async_tx_descriptor *tx) argument
889 struct xilinx_vdma_tx_descriptor *desc = to_vdma_tx_descriptor(tx);
890 struct xilinx_vdma_chan *chan = to_xilinx_chan(tx->chan);
907 cookie = dma_cookie_assign(tx);
/drivers/infiniband/hw/ipath/
H A Dipath_verbs.c1004 struct ipath_verbs_txreq *tx = NULL; local
1012 tx = list_entry(l, struct ipath_verbs_txreq, txreq.list);
1015 return tx;
1019 struct ipath_verbs_txreq *tx)
1024 list_add(&tx->txreq.list, &dev->txreq_free);
1030 struct ipath_verbs_txreq *tx = cookie; local
1031 struct ipath_qp *qp = tx->qp;
1039 if (tx->wqe)
1040 ipath_send_complete(qp, tx->wqe, ibs);
1047 } else if (tx
1018 put_txreq(struct ipath_ibdev *dev, struct ipath_verbs_txreq *tx) argument
1104 struct ipath_verbs_txreq *tx; local
1993 struct ipath_verbs_txreq *tx; local
[all...]

Completed in 6718 milliseconds

1234567891011>>