Searched refs:txq (Results 1 - 25 of 137) sorted by relevance

123456

/drivers/net/wireless/iwlwifi/pcie/
H A Dtx.c149 struct iwl_txq *txq = (void *)data; local
150 struct iwl_queue *q = &txq->q;
151 struct iwl_trans_pcie *trans_pcie = txq->trans_pcie;
154 SCD_TX_STTS_QUEUE_OFFSET(txq->q.id);
158 spin_lock(&txq->lock);
160 if (txq->q.read_ptr == txq->q.write_ptr) {
161 spin_unlock(&txq->lock);
164 spin_unlock(&txq->lock);
166 IWL_ERR(trans, "Queue %d stuck for %u ms.\n", txq
212 iwl_pcie_txq_update_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt) argument
257 iwl_pcie_txq_inval_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq) argument
286 iwl_pcie_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) argument
335 struct iwl_txq *txq = &trans_pcie->txq[i]; local
413 iwl_pcie_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) argument
447 iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, dma_addr_t addr, u16 len, bool reset) argument
479 iwl_pcie_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, u32 txq_id) argument
549 iwl_pcie_txq_init(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, u32 txq_id) argument
583 struct iwl_txq *txq = &trans_pcie->txq[txq_id]; local
611 struct iwl_txq *txq = &trans_pcie->txq[txq_id]; local
712 struct iwl_txq *txq = &trans_pcie->txq[txq_id]; local
909 iwl_pcie_txq_progress(struct iwl_trans_pcie *trans_pcie, struct iwl_txq *txq) argument
930 struct iwl_txq *txq = &trans_pcie->txq[txq_id]; local
1002 struct iwl_txq *txq = &trans_pcie->txq[txq_id]; local
1204 struct iwl_txq *txq = &trans_pcie->txq[trans_pcie->cmd_queue]; local
1485 struct iwl_txq *txq = &trans_pcie->txq[trans_pcie->cmd_queue]; local
1592 struct iwl_txq *txq = &trans_pcie->txq[trans_pcie->cmd_queue]; local
1680 struct iwl_txq *txq; local
[all...]
H A Dinternal.h238 iwl_pcie_get_scratchbuf_dma(struct iwl_txq *txq, int idx) argument
240 return txq->scratchbufs_dma +
291 struct iwl_txq *txq; member in struct:iwl_trans_pcie
433 struct iwl_txq *txq)
437 if (test_and_clear_bit(txq->q.id, trans_pcie->queue_stopped)) {
438 IWL_DEBUG_TX_QUEUES(trans, "Wake hwq %d\n", txq->q.id);
439 iwl_op_mode_queue_not_full(trans->op_mode, txq->q.id);
444 struct iwl_txq *txq)
448 if (!test_and_set_bit(txq->q.id, trans_pcie->queue_stopped)) {
449 iwl_op_mode_queue_full(trans->op_mode, txq
432 iwl_wake_queue(struct iwl_trans *trans, struct iwl_txq *txq) argument
443 iwl_stop_queue(struct iwl_trans *trans, struct iwl_txq *txq) argument
[all...]
/drivers/net/ethernet/marvell/
H A Dmvneta.c268 #define IS_TSO_HEADER(txq, addr) \
269 ((addr >= txq->tso_hdrs_phys) && \
270 (addr < txq->tso_hdrs_phys + txq->size * TSO_HEADER_SIZE))
487 /* Increment txq get counter */
488 static void mvneta_txq_inc_get(struct mvneta_tx_queue *txq) argument
490 txq->txq_get_index++;
491 if (txq->txq_get_index == txq->size)
492 txq
496 mvneta_txq_inc_put(struct mvneta_tx_queue *txq) argument
674 mvneta_txq_pend_desc_add(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int pend_desc) argument
689 mvneta_txq_next_desc_get(struct mvneta_tx_queue *txq) argument
700 mvneta_txq_desc_put(struct mvneta_tx_queue *txq) argument
744 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local
1125 mvneta_tx_done_pkts_coal_set(struct mvneta_port *pp, struct mvneta_tx_queue *txq, u32 value) argument
1149 mvneta_txq_sent_desc_dec(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int sent_desc) argument
1167 mvneta_txq_sent_desc_num_get(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
1183 mvneta_txq_sent_desc_proc(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
1287 mvneta_txq_bufs_free(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int num) argument
1310 mvneta_txq_done(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
1544 mvneta_tso_put_hdr(struct sk_buff *skb, struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
1561 mvneta_tso_put_data(struct net_device *dev, struct mvneta_tx_queue *txq, struct sk_buff *skb, char *data, int size, bool last_tcp, bool is_last) argument
1592 mvneta_tx_tso(struct sk_buff *skb, struct net_device *dev, struct mvneta_tx_queue *txq) argument
1664 mvneta_tx_frag_process(struct mvneta_port *pp, struct sk_buff *skb, struct mvneta_tx_queue *txq) argument
1722 struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; local
1803 mvneta_txq_done_force(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
1822 struct mvneta_tx_queue *txq; local
2220 mvneta_txq_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
2279 mvneta_txq_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
2713 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local
2826 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local
[all...]
H A Dmv643xx_eth.c192 #define IS_TSO_HEADER(txq, addr) \
193 ((addr >= txq->tso_hdrs_dma) && \
194 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE))
418 struct tx_queue txq[8]; member in struct:mv643xx_eth_private
450 /* rxq/txq helper functions *************************************************/
456 static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq) argument
458 return container_of(txq, struct mv643xx_eth_private, txq[txq
477 txq_reset_hw_ptr(struct tx_queue *txq) argument
487 txq_enable(struct tx_queue *txq) argument
493 txq_disable(struct tx_queue *txq) argument
503 txq_maybe_wake(struct tx_queue *txq) argument
741 txq_put_data_tso(struct net_device *dev, struct tx_queue *txq, struct sk_buff *skb, char *data, int length, bool last_tcp, bool is_last) argument
776 txq_put_hdr_tso(struct sk_buff *skb, struct tx_queue *txq, int length) argument
808 txq_submit_tso(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) argument
872 txq_submit_frag_skb(struct tx_queue *txq, struct sk_buff *skb) argument
910 txq_submit_skb(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) argument
977 struct tx_queue *txq; local
1012 txq_kick(struct tx_queue *txq) argument
1037 txq_reclaim(struct tx_queue *txq, int budget, int force) argument
1132 txq_set_rate(struct tx_queue *txq, int rate, int burst) argument
1150 txq_set_fixed_prio_mode(struct tx_queue *txq) argument
1229 struct tx_queue *txq = mp->txq + i; local
1996 struct tx_queue *txq = mp->txq + index; local
2065 txq_deinit(struct tx_queue *txq) argument
2153 struct tx_queue *txq = mp->txq + i; local
2294 struct tx_queue *txq = mp->txq + i; local
[all...]
H A Dmvpp2.c131 #define MVPP2_TXQ_SENT_REG(txq) (0x3c00 + 4 * (txq))
981 static inline int mvpp2_txq_phys(int port, int txq) argument
983 return (MVPP2_MAX_TCONT + port) * MVPP2_MAX_TXQ + txq;
4016 struct mvpp2_tx_queue *txq = port->txqs[queue]; local
4018 if (txq->descs != NULL)
4133 struct mvpp2_tx_queue *txq)
4137 mvpp2_write(port->priv, MVPP2_TXQ_NUM_REG, txq->id);
4145 mvpp2_txq_next_desc_get(struct mvpp2_tx_queue *txq) argument
4147 int tx_desc = txq
4132 mvpp2_txq_pend_desc_num_get(struct mvpp2_port *port, struct mvpp2_tx_queue *txq) argument
4182 mvpp2_txq_alloc_reserved_desc(struct mvpp2 *priv, struct mvpp2_tx_queue *txq, int num) argument
4198 mvpp2_txq_reserved_desc_num_proc(struct mvpp2 *priv, struct mvpp2_tx_queue *txq, struct mvpp2_txq_pcpu *txq_pcpu, int num) argument
4240 mvpp2_txq_desc_put(struct mvpp2_tx_queue *txq) argument
4285 mvpp2_txq_sent_desc_proc(struct mvpp2_port *port, struct mvpp2_tx_queue *txq) argument
4313 int txq, tx_port_num; local
4393 struct mvpp2_tx_queue *txq = port->txqs[queue]; local
4403 mvpp2_txq_bufs_free(struct mvpp2_port *port, struct mvpp2_tx_queue *txq, struct mvpp2_txq_pcpu *txq_pcpu, int num) argument
4442 mvpp2_txq_done(struct mvpp2_port *port, struct mvpp2_tx_queue *txq, struct mvpp2_txq_pcpu *txq_pcpu) argument
4585 mvpp2_txq_init(struct mvpp2_port *port, struct mvpp2_tx_queue *txq) argument
4669 mvpp2_txq_deinit(struct mvpp2_port *port, struct mvpp2_tx_queue *txq) argument
4700 mvpp2_txq_clean(struct mvpp2_port *port, struct mvpp2_tx_queue *txq) argument
4747 struct mvpp2_tx_queue *txq; local
4798 struct mvpp2_tx_queue *txq; local
5107 tx_desc_unmap_put(struct device *dev, struct mvpp2_tx_queue *txq, struct mvpp2_tx_desc *desc) argument
5116 mvpp2_tx_frag_process(struct mvpp2_port *port, struct sk_buff *skb, struct mvpp2_tx_queue *aggr_txq, struct mvpp2_tx_queue *txq) argument
5173 struct mvpp2_tx_queue *txq, *aggr_txq; local
5302 struct mvpp2_tx_queue *txq = mvpp2_get_tx_queue(port, cause_tx); local
5782 struct mvpp2_tx_queue *txq = port->txqs[queue]; local
5944 struct mvpp2_tx_queue *txq; local
[all...]
/drivers/net/wireless/ath/ath9k/
H A Dxmit.c50 static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq,
53 int tx_flags, struct ath_txq *txq);
55 struct ath_txq *txq, struct list_head *bf_q,
57 static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq,
65 struct ath_txq *txq,
80 void ath_txq_lock(struct ath_softc *sc, struct ath_txq *txq)
81 __acquires(&txq->axq_lock)
83 spin_lock_bh(&txq->axq_lock);
86 void ath_txq_unlock(struct ath_softc *sc, struct ath_txq *txq)
87 __releases(&txq
106 ath_tx_queue_tid(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) argument
156 ath_txq_skb_done(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb) argument
211 struct ath_txq *txq = tid->ac->txq; local
240 struct ath_txq *txq = tid->ac->txq; local
314 ath_tid_drain(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) argument
341 ath_tx_set_retry(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb, int count) argument
438 ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_buf *bf, struct list_head *bf_q, struct ath_tx_status *ts, int txok) argument
674 ath_tx_process_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_tx_status *ts, struct ath_buf *bf, struct list_head *bf_head) argument
871 ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff_head **q) argument
948 ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first, struct sk_buff_head *tid_q, int *aggr_len) argument
1230 ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, int len) argument
1324 ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first, struct sk_buff_head *tid_q) argument
1358 ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, bool *stop) argument
1409 struct ath_txq *txq; local
1449 struct ath_txq *txq = txtid->ac->txq; local
1463 struct ath_txq *txq; local
1500 struct ath_txq *txq; local
1526 struct ath_txq *txq; local
1552 struct ath_txq *txq = sc->tx.uapsdq; local
1665 struct ath_txq *txq = &sc->tx.txq[axq_qnum]; local
1727 ath_drain_txq_list(struct ath_softc *sc, struct ath_txq *txq, struct list_head *list) argument
1760 ath_draintxq(struct ath_softc *sc, struct ath_txq *txq) argument
1786 struct ath_txq *txq; local
1827 ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq) argument
1836 ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq) argument
1917 struct ath_txq *txq; local
1937 ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, struct list_head *head, bool internal) argument
2004 ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) argument
2093 ath_tx_setup_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) argument
2224 struct ath_txq *txq = txctl->txq; local
2397 ath_tx_complete(struct ath_softc *sc, struct sk_buff *skb, int tx_flags, struct ath_txq *txq) argument
2442 ath_tx_complete_buf(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, struct list_head *bf_q, struct ath_tx_status *ts, int txok) argument
2542 ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq) argument
2634 struct ath_txq *txq; local
2809 struct ath_txq *txq; local
[all...]
H A Ddebug_sta.c30 struct ath_txq *txq; local
56 txq = ac->txq;
57 ath_txq_lock(sc, txq);
61 ath_txq_unlock(sc, txq);
71 txq = tid->ac->txq;
72 ath_txq_lock(sc, txq);
85 ath_txq_unlock(sc, txq);
H A Dlink.c27 struct ath_txq *txq; local
39 txq = sc->tx.txq_map[i];
41 ath_txq_lock(sc, txq);
42 if (txq->axq_depth) {
43 if (txq->axq_tx_inprogress) {
45 ath_txq_unlock(sc, txq);
48 txq->axq_tx_inprogress = true;
51 ath_txq_unlock(sc, txq);
178 txctl.txq = sc->tx.txq_map[IEEE80211_AC_BE];
/drivers/bluetooth/
H A Dhci_h4.c55 struct sk_buff_head txq; member in struct:h4_struct
69 skb_queue_head_init(&h4->txq);
82 skb_queue_purge(&h4->txq);
96 skb_queue_purge(&h4->txq);
115 skb_queue_tail(&h4->txq, skb);
140 return skb_dequeue(&h4->txq);
H A Dhci_ath.c48 struct sk_buff_head txq; member in struct:ath_struct
119 skb_queue_head_init(&ath->txq);
136 skb_queue_purge(&ath->txq);
148 skb_queue_purge(&ath->txq);
186 skb_queue_tail(&ath->txq, skb);
198 return skb_dequeue(&ath->txq);
H A Dhci_ll.c83 struct sk_buff_head txq; member in struct:ll_struct
115 skb_queue_tail(&ll->txq, skb);
131 skb_queue_head_init(&ll->txq);
150 skb_queue_purge(&ll->txq);
163 skb_queue_purge(&ll->txq);
176 * 1. places all pending packets (waiting in tx_wait_q list) in txq list.
186 skb_queue_tail(&ll->txq, skb);
319 skb_queue_tail(&ll->txq, skb);
505 return skb_dequeue(&ll->txq);
H A Dbtsdio.c63 struct sk_buff_head txq; member in struct:btsdio_data
115 while ((skb = skb_dequeue(&data->txq))) {
119 skb_queue_head(&data->txq, skb);
253 skb_queue_purge(&data->txq);
284 skb_queue_tail(&data->txq, skb);
314 skb_queue_head_init(&data->txq);
/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dsge.c1048 static void txq_stop(struct sge_eth_txq *txq) argument
1050 netif_tx_stop_queue(txq->txq);
1051 txq->q.stops++;
1079 struct sge_eth_txq *txq; local
1106 txq = &adapter->sge.ethtxq[pi->first_qset + qidx];
1112 reclaim_completed_tx(adapter, &txq->q, true);
1121 credits = txq_avail(&txq->q) - ndesc;
1130 txq_stop(txq);
1144 txq
2006 struct sge_eth_txq *txq = &s->ethtxq[i]; local
2219 t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq, struct net_device *dev, struct netdev_queue *devq, unsigned int iqid) argument
2366 struct sge_eth_txq *txq = s->ethtxq; local
[all...]
/drivers/net/ethernet/freescale/
H A Dfec_main.c206 #define IS_TSO_HEADER(txq, addr) \
207 ((addr >= txq->tso_hdrs_dma) && \
208 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE))
219 struct fec_enet_priv_tx_q *txq = fep->tx_queue[queue_id]; local
225 if (bdp >= txq->tx_bd_base) {
226 base = txq->tx_bd_base;
227 ring_size = txq->tx_ring_size;
228 ex_base = (struct bufdesc_ex *)txq->tx_bd_base;
250 struct fec_enet_priv_tx_q *txq local
279 fec_enet_get_free_txdesc_num(struct fec_enet_private *fep, struct fec_enet_priv_tx_q *txq) argument
315 struct fec_enet_priv_tx_q *txq; local
359 fec_enet_txq_submit_frag_skb(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument
449 fec_enet_txq_submit_skb(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument
570 fec_enet_txq_put_data_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev, struct bufdesc *bdp, int index, char *data, int size, bool last_tcp, bool is_last) argument
633 fec_enet_txq_put_hdr_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev, struct bufdesc *bdp, int index) argument
689 fec_enet_txq_submit_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument
784 struct fec_enet_priv_tx_q *txq; local
811 struct fec_enet_priv_tx_q *txq; local
875 struct fec_enet_priv_tx_q *txq; local
903 struct fec_enet_priv_tx_q *txq; local
1190 struct fec_enet_priv_tx_q *txq; local
2508 struct fec_enet_priv_tx_q *txq; local
2546 struct fec_enet_priv_tx_q *txq; local
2571 struct fec_enet_priv_tx_q *txq; local
2664 struct fec_enet_priv_tx_q *txq; local
2956 struct fec_enet_priv_tx_q *txq; local
[all...]
/drivers/net/ethernet/atheros/alx/
H A Dmain.c57 struct alx_buffer *txb = &alx->txq.bufs[entry];
132 struct alx_tx_queue *txq = &alx->txq; local
134 if (txq->write_idx >= txq->read_idx)
135 return alx->tx_ringsz + txq->read_idx - txq->write_idx - 1;
136 return txq->read_idx - txq->write_idx - 1;
141 struct alx_tx_queue *txq local
386 struct alx_tx_queue *txq = &alx->txq; local
1004 struct alx_tx_queue *txq = &alx->txq; local
1070 struct alx_tx_queue *txq = &alx->txq; local
[all...]
/drivers/net/wireless/ath/ath5k/
H A Dbase.c719 struct ath5k_txq *txq, int padsize,
816 spin_lock_bh(&txq->lock);
817 list_add_tail(&bf->list, &txq->q);
818 txq->txq_len++;
819 if (txq->link == NULL) /* is this first packet? */
820 ath5k_hw_set_txdp(ah, txq->qnum, bf->daddr);
822 *txq->link = bf->daddr;
824 txq->link = &ds->ds_link;
825 ath5k_hw_start_tx_dma(ah, txq->qnum);
827 spin_unlock_bh(&txq
718 ath5k_txbuf_setup(struct ath5k_hw *ah, struct ath5k_buf *bf, struct ath5k_txq *txq, int padsize, struct ieee80211_tx_control *control) argument
964 struct ath5k_txq *txq; local
1099 struct ath5k_txq *txq; local
1128 struct ath5k_txq *txq = ah->txqs; local
1594 ath5k_tx_queue(struct ieee80211_hw *hw, struct sk_buff *skb, struct ath5k_txq *txq, struct ieee80211_tx_control *control) argument
1650 ath5k_tx_frame_completed(struct ath5k_hw *ah, struct sk_buff *skb, struct ath5k_txq *txq, struct ath5k_tx_status *ts, struct ath5k_buf *bf) argument
1716 ath5k_tx_processq(struct ath5k_hw *ah, struct ath5k_txq *txq) argument
2450 struct ath5k_txq *txq; local
2949 struct ath5k_txq *txq; local
[all...]
/drivers/atm/
H A Dambassador.c628 amb_txq * txq = &dev->txq; local
636 spin_lock_irqsave (&txq->lock, flags);
638 if (txq->pending < txq->maximum) {
639 PRINTD (DBG_TX, "TX in slot %p", txq->in.ptr);
641 *txq->in.ptr = *tx;
642 txq->pending++;
643 txq->in.ptr = NEXTQ (txq
660 amb_txq * txq = &dev->txq; local
1564 amb_txq * txq = &dev->txq; local
[all...]
/drivers/net/wireless/iwlegacy/
H A Dcommon.c382 il->txq[il->cmd_queue].meta[cmd_idx].flags &= ~CMD_WANT_SKB;
2724 il_txq_update_write_ptr(struct il_priv *il, struct il_tx_queue *txq) argument
2727 int txq_id = txq->q.id;
2729 if (txq->need_update == 0)
2747 il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8));
2755 _il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8));
2756 txq->need_update = 0;
2766 struct il_tx_queue *txq = &il->txq[txq_id]; local
2767 struct il_queue *q = &txq
2790 struct il_tx_queue *txq = &il->txq[txq_id]; local
2826 struct il_tx_queue *txq = &il->txq[il->cmd_queue]; local
2869 struct il_tx_queue *txq = &il->txq[il->cmd_queue]; local
2975 il_tx_queue_alloc(struct il_priv *il, struct il_tx_queue *txq, u32 id) argument
3019 struct il_tx_queue *txq = &il->txq[txq_id]; local
3093 struct il_tx_queue *txq = &il->txq[txq_id]; local
3128 struct il_tx_queue *txq = &il->txq[il->cmd_queue]; local
3254 struct il_tx_queue *txq = &il->txq[txq_id]; local
3296 struct il_tx_queue *txq = &il->txq[il->cmd_queue]; local
4803 struct il_tx_queue *txq = &il->txq[cnt]; local
[all...]
H A D3945.c289 struct il_tx_queue *txq = &il->txq[txq_id]; local
290 struct il_queue *q = &txq->q;
298 skb = txq->skbs[txq->q.read_ptr];
300 txq->skbs[txq->q.read_ptr] = NULL;
301 il->ops->txq_free_tfd(il, txq);
306 il_wake_queue(il, txq);
319 struct il_tx_queue *txq local
612 il3945_hw_txq_attach_buf_to_tfd(struct il_priv *il, struct il_tx_queue *txq, dma_addr_t addr, u16 len, u8 reset, u8 pad) argument
651 il3945_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq) argument
2213 il3945_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq) argument
[all...]
/drivers/net/
H A Difb.c67 struct netdev_queue *txq; local
70 txq = netdev_get_tx_queue(_dev, 0);
72 if (__netif_tx_trylock(txq)) {
74 __netif_tx_unlock(txq);
114 if (__netif_tx_trylock(txq)) {
120 __netif_tx_unlock(txq);
123 __netif_tx_unlock(txq);
/drivers/net/wireless/mwifiex/
H A Dtxrx.c161 struct netdev_queue *txq; local
197 txq = netdev_get_tx_queue(priv->netdev, index);
198 if (netif_tx_queue_stopped(txq)) {
199 netif_tx_wake_queue(txq);
/drivers/net/ethernet/chelsio/cxgb4/
H A Dsge.c1023 netif_tx_stop_queue(q->txq);
2203 struct sge_ofld_txq *txq = s->egr_map[id]; local
2206 tasklet_schedule(&txq->qresume_tsk);
2215 time_after_eq(jiffies, q->txq->trans_start + HZ / 100) &&
2216 __netif_tx_trylock(q->txq)) {
2227 __netif_tx_unlock(q->txq);
2406 int t4_sge_alloc_eth_txq(struct adapter *adap, struct sge_eth_txq *txq, argument
2416 nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc);
2418 txq->q.desc = alloc_ring(adap->pdev_dev, txq
2461 t4_sge_alloc_ctrl_txq(struct adapter *adap, struct sge_ctrl_txq *txq, struct net_device *dev, unsigned int iqid, unsigned int cmplqid) argument
2514 t4_sge_alloc_ofld_txq(struct adapter *adap, struct sge_ofld_txq *txq, struct net_device *dev, unsigned int iqid) argument
[all...]
/drivers/net/ethernet/brocade/bna/
H A Dbna_tx_rx.c3145 struct bna_txq *txq; local
3150 txq = (struct bna_txq *)qe;
3151 txq->tcb->priority = txq->priority;
3153 bna_ib_start(tx->bna, &txq->ib, is_regular);
3373 struct bna_txq *txq = NULL; local
3386 txq = (struct bna_txq *)qe;
3388 bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].q.q, &txq->qpt);
3389 cfg_req->q_cfg[i].q.priority = txq->priority;
3392 txq
3438 struct bna_txq *txq; local
3451 bna_txq_qpt_setup(struct bna_txq *txq, int page_count, int page_size, struct bna_mem_descr *qpt_mem, struct bna_mem_descr *swqpt_mem, struct bna_mem_descr *page_mem) argument
3510 struct bna_txq *txq; local
3586 struct bna_txq *txq = NULL; local
3690 struct bna_txq *txq; local
3835 struct bna_txq *txq; local
4010 struct bna_txq *txq; local
[all...]
/drivers/net/ethernet/chelsio/cxgb3/
H A Dsge.c179 return container_of(q, struct sge_qset, txq[qidx]);
654 memset(q->txq, 0, sizeof(struct sge_txq) * SGE_TXQ_PER_SET);
691 if (q->txq[i].desc) {
693 t3_sge_enable_ecntxt(adapter, q->txq[i].cntxt_id, 0);
695 if (q->txq[i].sdesc) {
696 free_tx_desc(adapter, &q->txq[i],
697 q->txq[i].in_use);
698 kfree(q->txq[i].sdesc);
701 q->txq[i].size *
703 q->txq[
1206 t3_stop_tx_queue(struct netdev_queue *txq, struct sge_qset *qs, struct sge_txq *q) argument
1227 struct netdev_queue *txq; local
2189 struct sge_txq *txq = &qs->txq[TXQ_ETH]; local
2200 struct sge_txq *txq = &qs->txq[TXQ_OFLD]; local
[all...]
/drivers/net/wireless/ath/ath6kl/
H A Dhtc_pipe.c106 if (list_empty(&ep->txq))
110 packet = list_first_entry(&ep->txq, struct htc_packet, list);
114 __func__, packet, get_queue_depth(&ep->txq));
159 packet = list_first_entry(&ep->txq, struct htc_packet, list);
183 if (list_empty(&ep->txq))
186 packet = list_first_entry(&ep->txq, struct htc_packet, list);
191 __func__, packet, get_queue_depth(&ep->txq));
303 struct list_head *txq)
313 __func__, txq,
314 (txq
301 htc_try_send(struct htc_target *target, struct htc_endpoint *ep, struct list_head *txq) argument
[all...]

Completed in 530 milliseconds

123456