/drivers/net/wireless/iwlwifi/pcie/ |
H A D | tx.c | 149 struct iwl_txq *txq = (void *)data; local 150 struct iwl_queue *q = &txq->q; 151 struct iwl_trans_pcie *trans_pcie = txq->trans_pcie; 154 SCD_TX_STTS_QUEUE_OFFSET(txq->q.id); 158 spin_lock(&txq->lock); 160 if (txq->q.read_ptr == txq->q.write_ptr) { 161 spin_unlock(&txq->lock); 164 spin_unlock(&txq->lock); 166 IWL_ERR(trans, "Queue %d stuck for %u ms.\n", txq 212 iwl_pcie_txq_update_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt) argument 257 iwl_pcie_txq_inval_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq) argument 286 iwl_pcie_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) argument 335 struct iwl_txq *txq = &trans_pcie->txq[i]; local 413 iwl_pcie_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) argument 447 iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, dma_addr_t addr, u16 len, bool reset) argument 479 iwl_pcie_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, u32 txq_id) argument 549 iwl_pcie_txq_init(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, u32 txq_id) argument 583 struct iwl_txq *txq = &trans_pcie->txq[txq_id]; local 611 struct iwl_txq *txq = &trans_pcie->txq[txq_id]; local 712 struct iwl_txq *txq = &trans_pcie->txq[txq_id]; local 909 iwl_pcie_txq_progress(struct iwl_trans_pcie *trans_pcie, struct iwl_txq *txq) argument 930 struct iwl_txq *txq = &trans_pcie->txq[txq_id]; local 1002 struct iwl_txq *txq = &trans_pcie->txq[txq_id]; local 1204 struct iwl_txq *txq = &trans_pcie->txq[trans_pcie->cmd_queue]; local 1485 struct iwl_txq *txq = &trans_pcie->txq[trans_pcie->cmd_queue]; local 1592 struct iwl_txq *txq = &trans_pcie->txq[trans_pcie->cmd_queue]; local 1680 struct iwl_txq *txq; local [all...] |
H A D | internal.h | 238 iwl_pcie_get_scratchbuf_dma(struct iwl_txq *txq, int idx) argument 240 return txq->scratchbufs_dma + 291 struct iwl_txq *txq; member in struct:iwl_trans_pcie 433 struct iwl_txq *txq) 437 if (test_and_clear_bit(txq->q.id, trans_pcie->queue_stopped)) { 438 IWL_DEBUG_TX_QUEUES(trans, "Wake hwq %d\n", txq->q.id); 439 iwl_op_mode_queue_not_full(trans->op_mode, txq->q.id); 444 struct iwl_txq *txq) 448 if (!test_and_set_bit(txq->q.id, trans_pcie->queue_stopped)) { 449 iwl_op_mode_queue_full(trans->op_mode, txq 432 iwl_wake_queue(struct iwl_trans *trans, struct iwl_txq *txq) argument 443 iwl_stop_queue(struct iwl_trans *trans, struct iwl_txq *txq) argument [all...] |
/drivers/net/ethernet/marvell/ |
H A D | mvneta.c | 268 #define IS_TSO_HEADER(txq, addr) \ 269 ((addr >= txq->tso_hdrs_phys) && \ 270 (addr < txq->tso_hdrs_phys + txq->size * TSO_HEADER_SIZE)) 487 /* Increment txq get counter */ 488 static void mvneta_txq_inc_get(struct mvneta_tx_queue *txq) argument 490 txq->txq_get_index++; 491 if (txq->txq_get_index == txq->size) 492 txq 496 mvneta_txq_inc_put(struct mvneta_tx_queue *txq) argument 674 mvneta_txq_pend_desc_add(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int pend_desc) argument 689 mvneta_txq_next_desc_get(struct mvneta_tx_queue *txq) argument 700 mvneta_txq_desc_put(struct mvneta_tx_queue *txq) argument 744 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local 1125 mvneta_tx_done_pkts_coal_set(struct mvneta_port *pp, struct mvneta_tx_queue *txq, u32 value) argument 1149 mvneta_txq_sent_desc_dec(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int sent_desc) argument 1167 mvneta_txq_sent_desc_num_get(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 1183 mvneta_txq_sent_desc_proc(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 1287 mvneta_txq_bufs_free(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int num) argument 1310 mvneta_txq_done(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 1544 mvneta_tso_put_hdr(struct sk_buff *skb, struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 1561 mvneta_tso_put_data(struct net_device *dev, struct mvneta_tx_queue *txq, struct sk_buff *skb, char *data, int size, bool last_tcp, bool is_last) argument 1592 mvneta_tx_tso(struct sk_buff *skb, struct net_device *dev, struct mvneta_tx_queue *txq) argument 1664 mvneta_tx_frag_process(struct mvneta_port *pp, struct sk_buff *skb, struct mvneta_tx_queue *txq) argument 1722 struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; local 1803 mvneta_txq_done_force(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 1822 struct mvneta_tx_queue *txq; local 2220 mvneta_txq_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 2279 mvneta_txq_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 2713 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local 2826 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local [all...] |
H A D | mv643xx_eth.c | 192 #define IS_TSO_HEADER(txq, addr) \ 193 ((addr >= txq->tso_hdrs_dma) && \ 194 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE)) 418 struct tx_queue txq[8]; member in struct:mv643xx_eth_private 450 /* rxq/txq helper functions *************************************************/ 456 static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq) argument 458 return container_of(txq, struct mv643xx_eth_private, txq[txq 477 txq_reset_hw_ptr(struct tx_queue *txq) argument 487 txq_enable(struct tx_queue *txq) argument 493 txq_disable(struct tx_queue *txq) argument 503 txq_maybe_wake(struct tx_queue *txq) argument 741 txq_put_data_tso(struct net_device *dev, struct tx_queue *txq, struct sk_buff *skb, char *data, int length, bool last_tcp, bool is_last) argument 776 txq_put_hdr_tso(struct sk_buff *skb, struct tx_queue *txq, int length) argument 808 txq_submit_tso(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) argument 872 txq_submit_frag_skb(struct tx_queue *txq, struct sk_buff *skb) argument 910 txq_submit_skb(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) argument 977 struct tx_queue *txq; local 1012 txq_kick(struct tx_queue *txq) argument 1037 txq_reclaim(struct tx_queue *txq, int budget, int force) argument 1132 txq_set_rate(struct tx_queue *txq, int rate, int burst) argument 1150 txq_set_fixed_prio_mode(struct tx_queue *txq) argument 1229 struct tx_queue *txq = mp->txq + i; local 1996 struct tx_queue *txq = mp->txq + index; local 2065 txq_deinit(struct tx_queue *txq) argument 2153 struct tx_queue *txq = mp->txq + i; local 2294 struct tx_queue *txq = mp->txq + i; local [all...] |
H A D | mvpp2.c | 131 #define MVPP2_TXQ_SENT_REG(txq) (0x3c00 + 4 * (txq)) 981 static inline int mvpp2_txq_phys(int port, int txq) argument 983 return (MVPP2_MAX_TCONT + port) * MVPP2_MAX_TXQ + txq; 4016 struct mvpp2_tx_queue *txq = port->txqs[queue]; local 4018 if (txq->descs != NULL) 4133 struct mvpp2_tx_queue *txq) 4137 mvpp2_write(port->priv, MVPP2_TXQ_NUM_REG, txq->id); 4145 mvpp2_txq_next_desc_get(struct mvpp2_tx_queue *txq) argument 4147 int tx_desc = txq 4132 mvpp2_txq_pend_desc_num_get(struct mvpp2_port *port, struct mvpp2_tx_queue *txq) argument 4182 mvpp2_txq_alloc_reserved_desc(struct mvpp2 *priv, struct mvpp2_tx_queue *txq, int num) argument 4198 mvpp2_txq_reserved_desc_num_proc(struct mvpp2 *priv, struct mvpp2_tx_queue *txq, struct mvpp2_txq_pcpu *txq_pcpu, int num) argument 4240 mvpp2_txq_desc_put(struct mvpp2_tx_queue *txq) argument 4285 mvpp2_txq_sent_desc_proc(struct mvpp2_port *port, struct mvpp2_tx_queue *txq) argument 4313 int txq, tx_port_num; local 4393 struct mvpp2_tx_queue *txq = port->txqs[queue]; local 4403 mvpp2_txq_bufs_free(struct mvpp2_port *port, struct mvpp2_tx_queue *txq, struct mvpp2_txq_pcpu *txq_pcpu, int num) argument 4442 mvpp2_txq_done(struct mvpp2_port *port, struct mvpp2_tx_queue *txq, struct mvpp2_txq_pcpu *txq_pcpu) argument 4585 mvpp2_txq_init(struct mvpp2_port *port, struct mvpp2_tx_queue *txq) argument 4669 mvpp2_txq_deinit(struct mvpp2_port *port, struct mvpp2_tx_queue *txq) argument 4700 mvpp2_txq_clean(struct mvpp2_port *port, struct mvpp2_tx_queue *txq) argument 4747 struct mvpp2_tx_queue *txq; local 4798 struct mvpp2_tx_queue *txq; local 5107 tx_desc_unmap_put(struct device *dev, struct mvpp2_tx_queue *txq, struct mvpp2_tx_desc *desc) argument 5116 mvpp2_tx_frag_process(struct mvpp2_port *port, struct sk_buff *skb, struct mvpp2_tx_queue *aggr_txq, struct mvpp2_tx_queue *txq) argument 5173 struct mvpp2_tx_queue *txq, *aggr_txq; local 5302 struct mvpp2_tx_queue *txq = mvpp2_get_tx_queue(port, cause_tx); local 5782 struct mvpp2_tx_queue *txq = port->txqs[queue]; local 5944 struct mvpp2_tx_queue *txq; local [all...] |
/drivers/net/wireless/ath/ath9k/ |
H A D | xmit.c | 50 static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, 53 int tx_flags, struct ath_txq *txq); 55 struct ath_txq *txq, struct list_head *bf_q, 57 static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, 65 struct ath_txq *txq, 80 void ath_txq_lock(struct ath_softc *sc, struct ath_txq *txq) 81 __acquires(&txq->axq_lock) 83 spin_lock_bh(&txq->axq_lock); 86 void ath_txq_unlock(struct ath_softc *sc, struct ath_txq *txq) 87 __releases(&txq 106 ath_tx_queue_tid(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) argument 156 ath_txq_skb_done(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb) argument 211 struct ath_txq *txq = tid->ac->txq; local 240 struct ath_txq *txq = tid->ac->txq; local 314 ath_tid_drain(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) argument 341 ath_tx_set_retry(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb, int count) argument 438 ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_buf *bf, struct list_head *bf_q, struct ath_tx_status *ts, int txok) argument 674 ath_tx_process_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_tx_status *ts, struct ath_buf *bf, struct list_head *bf_head) argument 871 ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff_head **q) argument 948 ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first, struct sk_buff_head *tid_q, int *aggr_len) argument 1230 ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, int len) argument 1324 ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first, struct sk_buff_head *tid_q) argument 1358 ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, bool *stop) argument 1409 struct ath_txq *txq; local 1449 struct ath_txq *txq = txtid->ac->txq; local 1463 struct ath_txq *txq; local 1500 struct ath_txq *txq; local 1526 struct ath_txq *txq; local 1552 struct ath_txq *txq = sc->tx.uapsdq; local 1665 struct ath_txq *txq = &sc->tx.txq[axq_qnum]; local 1727 ath_drain_txq_list(struct ath_softc *sc, struct ath_txq *txq, struct list_head *list) argument 1760 ath_draintxq(struct ath_softc *sc, struct ath_txq *txq) argument 1786 struct ath_txq *txq; local 1827 ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq) argument 1836 ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq) argument 1917 struct ath_txq *txq; local 1937 ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, struct list_head *head, bool internal) argument 2004 ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) argument 2093 ath_tx_setup_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) argument 2224 struct ath_txq *txq = txctl->txq; local 2397 ath_tx_complete(struct ath_softc *sc, struct sk_buff *skb, int tx_flags, struct ath_txq *txq) argument 2442 ath_tx_complete_buf(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, struct list_head *bf_q, struct ath_tx_status *ts, int txok) argument 2542 ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq) argument 2634 struct ath_txq *txq; local 2809 struct ath_txq *txq; local [all...] |
H A D | debug_sta.c | 30 struct ath_txq *txq; local 56 txq = ac->txq; 57 ath_txq_lock(sc, txq); 61 ath_txq_unlock(sc, txq); 71 txq = tid->ac->txq; 72 ath_txq_lock(sc, txq); 85 ath_txq_unlock(sc, txq);
|
H A D | link.c | 27 struct ath_txq *txq; local 39 txq = sc->tx.txq_map[i]; 41 ath_txq_lock(sc, txq); 42 if (txq->axq_depth) { 43 if (txq->axq_tx_inprogress) { 45 ath_txq_unlock(sc, txq); 48 txq->axq_tx_inprogress = true; 51 ath_txq_unlock(sc, txq); 178 txctl.txq = sc->tx.txq_map[IEEE80211_AC_BE];
|
/drivers/bluetooth/ |
H A D | hci_h4.c | 55 struct sk_buff_head txq; member in struct:h4_struct 69 skb_queue_head_init(&h4->txq); 82 skb_queue_purge(&h4->txq); 96 skb_queue_purge(&h4->txq); 115 skb_queue_tail(&h4->txq, skb); 140 return skb_dequeue(&h4->txq);
|
H A D | hci_ath.c | 48 struct sk_buff_head txq; member in struct:ath_struct 119 skb_queue_head_init(&ath->txq); 136 skb_queue_purge(&ath->txq); 148 skb_queue_purge(&ath->txq); 186 skb_queue_tail(&ath->txq, skb); 198 return skb_dequeue(&ath->txq);
|
H A D | hci_ll.c | 83 struct sk_buff_head txq; member in struct:ll_struct 115 skb_queue_tail(&ll->txq, skb); 131 skb_queue_head_init(&ll->txq); 150 skb_queue_purge(&ll->txq); 163 skb_queue_purge(&ll->txq); 176 * 1. places all pending packets (waiting in tx_wait_q list) in txq list. 186 skb_queue_tail(&ll->txq, skb); 319 skb_queue_tail(&ll->txq, skb); 505 return skb_dequeue(&ll->txq);
|
H A D | btsdio.c | 63 struct sk_buff_head txq; member in struct:btsdio_data 115 while ((skb = skb_dequeue(&data->txq))) { 119 skb_queue_head(&data->txq, skb); 253 skb_queue_purge(&data->txq); 284 skb_queue_tail(&data->txq, skb); 314 skb_queue_head_init(&data->txq);
|
/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | sge.c | 1048 static void txq_stop(struct sge_eth_txq *txq) argument 1050 netif_tx_stop_queue(txq->txq); 1051 txq->q.stops++; 1079 struct sge_eth_txq *txq; local 1106 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; 1112 reclaim_completed_tx(adapter, &txq->q, true); 1121 credits = txq_avail(&txq->q) - ndesc; 1130 txq_stop(txq); 1144 txq 2006 struct sge_eth_txq *txq = &s->ethtxq[i]; local 2219 t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq, struct net_device *dev, struct netdev_queue *devq, unsigned int iqid) argument 2366 struct sge_eth_txq *txq = s->ethtxq; local [all...] |
/drivers/net/ethernet/freescale/ |
H A D | fec_main.c | 206 #define IS_TSO_HEADER(txq, addr) \ 207 ((addr >= txq->tso_hdrs_dma) && \ 208 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE)) 219 struct fec_enet_priv_tx_q *txq = fep->tx_queue[queue_id]; local 225 if (bdp >= txq->tx_bd_base) { 226 base = txq->tx_bd_base; 227 ring_size = txq->tx_ring_size; 228 ex_base = (struct bufdesc_ex *)txq->tx_bd_base; 250 struct fec_enet_priv_tx_q *txq local 279 fec_enet_get_free_txdesc_num(struct fec_enet_private *fep, struct fec_enet_priv_tx_q *txq) argument 315 struct fec_enet_priv_tx_q *txq; local 359 fec_enet_txq_submit_frag_skb(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument 449 fec_enet_txq_submit_skb(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument 570 fec_enet_txq_put_data_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev, struct bufdesc *bdp, int index, char *data, int size, bool last_tcp, bool is_last) argument 633 fec_enet_txq_put_hdr_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev, struct bufdesc *bdp, int index) argument 689 fec_enet_txq_submit_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument 784 struct fec_enet_priv_tx_q *txq; local 811 struct fec_enet_priv_tx_q *txq; local 875 struct fec_enet_priv_tx_q *txq; local 903 struct fec_enet_priv_tx_q *txq; local 1190 struct fec_enet_priv_tx_q *txq; local 2508 struct fec_enet_priv_tx_q *txq; local 2546 struct fec_enet_priv_tx_q *txq; local 2571 struct fec_enet_priv_tx_q *txq; local 2664 struct fec_enet_priv_tx_q *txq; local 2956 struct fec_enet_priv_tx_q *txq; local [all...] |
/drivers/net/ethernet/atheros/alx/ |
H A D | main.c | 57 struct alx_buffer *txb = &alx->txq.bufs[entry]; 132 struct alx_tx_queue *txq = &alx->txq; local 134 if (txq->write_idx >= txq->read_idx) 135 return alx->tx_ringsz + txq->read_idx - txq->write_idx - 1; 136 return txq->read_idx - txq->write_idx - 1; 141 struct alx_tx_queue *txq local 386 struct alx_tx_queue *txq = &alx->txq; local 1004 struct alx_tx_queue *txq = &alx->txq; local 1070 struct alx_tx_queue *txq = &alx->txq; local [all...] |
/drivers/net/wireless/ath/ath5k/ |
H A D | base.c | 719 struct ath5k_txq *txq, int padsize, 816 spin_lock_bh(&txq->lock); 817 list_add_tail(&bf->list, &txq->q); 818 txq->txq_len++; 819 if (txq->link == NULL) /* is this first packet? */ 820 ath5k_hw_set_txdp(ah, txq->qnum, bf->daddr); 822 *txq->link = bf->daddr; 824 txq->link = &ds->ds_link; 825 ath5k_hw_start_tx_dma(ah, txq->qnum); 827 spin_unlock_bh(&txq 718 ath5k_txbuf_setup(struct ath5k_hw *ah, struct ath5k_buf *bf, struct ath5k_txq *txq, int padsize, struct ieee80211_tx_control *control) argument 964 struct ath5k_txq *txq; local 1099 struct ath5k_txq *txq; local 1128 struct ath5k_txq *txq = ah->txqs; local 1594 ath5k_tx_queue(struct ieee80211_hw *hw, struct sk_buff *skb, struct ath5k_txq *txq, struct ieee80211_tx_control *control) argument 1650 ath5k_tx_frame_completed(struct ath5k_hw *ah, struct sk_buff *skb, struct ath5k_txq *txq, struct ath5k_tx_status *ts, struct ath5k_buf *bf) argument 1716 ath5k_tx_processq(struct ath5k_hw *ah, struct ath5k_txq *txq) argument 2450 struct ath5k_txq *txq; local 2949 struct ath5k_txq *txq; local [all...] |
/drivers/atm/ |
H A D | ambassador.c | 628 amb_txq * txq = &dev->txq; local 636 spin_lock_irqsave (&txq->lock, flags); 638 if (txq->pending < txq->maximum) { 639 PRINTD (DBG_TX, "TX in slot %p", txq->in.ptr); 641 *txq->in.ptr = *tx; 642 txq->pending++; 643 txq->in.ptr = NEXTQ (txq 660 amb_txq * txq = &dev->txq; local 1564 amb_txq * txq = &dev->txq; local [all...] |
/drivers/net/wireless/iwlegacy/ |
H A D | common.c | 382 il->txq[il->cmd_queue].meta[cmd_idx].flags &= ~CMD_WANT_SKB; 2724 il_txq_update_write_ptr(struct il_priv *il, struct il_tx_queue *txq) argument 2727 int txq_id = txq->q.id; 2729 if (txq->need_update == 0) 2747 il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8)); 2755 _il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8)); 2756 txq->need_update = 0; 2766 struct il_tx_queue *txq = &il->txq[txq_id]; local 2767 struct il_queue *q = &txq 2790 struct il_tx_queue *txq = &il->txq[txq_id]; local 2826 struct il_tx_queue *txq = &il->txq[il->cmd_queue]; local 2869 struct il_tx_queue *txq = &il->txq[il->cmd_queue]; local 2975 il_tx_queue_alloc(struct il_priv *il, struct il_tx_queue *txq, u32 id) argument 3019 struct il_tx_queue *txq = &il->txq[txq_id]; local 3093 struct il_tx_queue *txq = &il->txq[txq_id]; local 3128 struct il_tx_queue *txq = &il->txq[il->cmd_queue]; local 3254 struct il_tx_queue *txq = &il->txq[txq_id]; local 3296 struct il_tx_queue *txq = &il->txq[il->cmd_queue]; local 4803 struct il_tx_queue *txq = &il->txq[cnt]; local [all...] |
H A D | 3945.c | 289 struct il_tx_queue *txq = &il->txq[txq_id]; local 290 struct il_queue *q = &txq->q; 298 skb = txq->skbs[txq->q.read_ptr]; 300 txq->skbs[txq->q.read_ptr] = NULL; 301 il->ops->txq_free_tfd(il, txq); 306 il_wake_queue(il, txq); 319 struct il_tx_queue *txq local 612 il3945_hw_txq_attach_buf_to_tfd(struct il_priv *il, struct il_tx_queue *txq, dma_addr_t addr, u16 len, u8 reset, u8 pad) argument 651 il3945_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq) argument 2213 il3945_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq) argument [all...] |
/drivers/net/ |
H A D | ifb.c | 67 struct netdev_queue *txq; local 70 txq = netdev_get_tx_queue(_dev, 0); 72 if (__netif_tx_trylock(txq)) { 74 __netif_tx_unlock(txq); 114 if (__netif_tx_trylock(txq)) { 120 __netif_tx_unlock(txq); 123 __netif_tx_unlock(txq);
|
/drivers/net/wireless/mwifiex/ |
H A D | txrx.c | 161 struct netdev_queue *txq; local 197 txq = netdev_get_tx_queue(priv->netdev, index); 198 if (netif_tx_queue_stopped(txq)) { 199 netif_tx_wake_queue(txq);
|
/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 1023 netif_tx_stop_queue(q->txq); 2203 struct sge_ofld_txq *txq = s->egr_map[id]; local 2206 tasklet_schedule(&txq->qresume_tsk); 2215 time_after_eq(jiffies, q->txq->trans_start + HZ / 100) && 2216 __netif_tx_trylock(q->txq)) { 2227 __netif_tx_unlock(q->txq); 2406 int t4_sge_alloc_eth_txq(struct adapter *adap, struct sge_eth_txq *txq, argument 2416 nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc); 2418 txq->q.desc = alloc_ring(adap->pdev_dev, txq 2461 t4_sge_alloc_ctrl_txq(struct adapter *adap, struct sge_ctrl_txq *txq, struct net_device *dev, unsigned int iqid, unsigned int cmplqid) argument 2514 t4_sge_alloc_ofld_txq(struct adapter *adap, struct sge_ofld_txq *txq, struct net_device *dev, unsigned int iqid) argument [all...] |
/drivers/net/ethernet/brocade/bna/ |
H A D | bna_tx_rx.c | 3145 struct bna_txq *txq; local 3150 txq = (struct bna_txq *)qe; 3151 txq->tcb->priority = txq->priority; 3153 bna_ib_start(tx->bna, &txq->ib, is_regular); 3373 struct bna_txq *txq = NULL; local 3386 txq = (struct bna_txq *)qe; 3388 bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].q.q, &txq->qpt); 3389 cfg_req->q_cfg[i].q.priority = txq->priority; 3392 txq 3438 struct bna_txq *txq; local 3451 bna_txq_qpt_setup(struct bna_txq *txq, int page_count, int page_size, struct bna_mem_descr *qpt_mem, struct bna_mem_descr *swqpt_mem, struct bna_mem_descr *page_mem) argument 3510 struct bna_txq *txq; local 3586 struct bna_txq *txq = NULL; local 3690 struct bna_txq *txq; local 3835 struct bna_txq *txq; local 4010 struct bna_txq *txq; local [all...] |
/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | sge.c | 179 return container_of(q, struct sge_qset, txq[qidx]); 654 memset(q->txq, 0, sizeof(struct sge_txq) * SGE_TXQ_PER_SET); 691 if (q->txq[i].desc) { 693 t3_sge_enable_ecntxt(adapter, q->txq[i].cntxt_id, 0); 695 if (q->txq[i].sdesc) { 696 free_tx_desc(adapter, &q->txq[i], 697 q->txq[i].in_use); 698 kfree(q->txq[i].sdesc); 701 q->txq[i].size * 703 q->txq[ 1206 t3_stop_tx_queue(struct netdev_queue *txq, struct sge_qset *qs, struct sge_txq *q) argument 1227 struct netdev_queue *txq; local 2189 struct sge_txq *txq = &qs->txq[TXQ_ETH]; local 2200 struct sge_txq *txq = &qs->txq[TXQ_OFLD]; local [all...] |
/drivers/net/wireless/ath/ath6kl/ |
H A D | htc_pipe.c | 106 if (list_empty(&ep->txq)) 110 packet = list_first_entry(&ep->txq, struct htc_packet, list); 114 __func__, packet, get_queue_depth(&ep->txq)); 159 packet = list_first_entry(&ep->txq, struct htc_packet, list); 183 if (list_empty(&ep->txq)) 186 packet = list_first_entry(&ep->txq, struct htc_packet, list); 191 __func__, packet, get_queue_depth(&ep->txq)); 303 struct list_head *txq) 313 __func__, txq, 314 (txq 301 htc_try_send(struct htc_target *target, struct htc_endpoint *ep, struct list_head *txq) argument [all...] |