Searched refs:rxq (Results 1 - 25 of 46) sorted by relevance

12

/drivers/net/wireless/iwlwifi/
H A Diwl-trans-pcie-rx.c80 * + A list of pre-allocated SKBs is stored in iwl->rxq->rx_free. When
81 * iwl->rxq->free_count drops to or below RX_LOW_WATERMARK, work is scheduled
82 * to replenish the iwl->rxq->rx_free.
84 * iwl->rxq is replenished and the READ INDEX is updated (updating the
87 * detached from the iwl->rxq. The driver 'processed' index is updated.
88 * + The Host/Firmware iwl->rxq is replenished at tasklet time from the rx_free
89 * list. If there are no allocated buffers in iwl->rxq->rx_free, the READ
204 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local
209 spin_lock_irqsave(&rxq
256 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local
363 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local
470 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local
[all...]
H A Diwl-trans-pcie.c89 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local
92 memset(&trans_pcie->rxq, 0, sizeof(trans_pcie->rxq));
94 spin_lock_init(&rxq->lock);
96 if (WARN_ON(rxq->bd || rxq->rb_stts))
100 rxq->bd = dma_zalloc_coherent(dev, sizeof(__le32) * RX_QUEUE_SIZE,
101 &rxq->bd_dma, GFP_KERNEL);
102 if (!rxq
126 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local
145 iwl_trans_rx_hw_init(struct iwl_trans *trans, struct iwl_rx_queue *rxq) argument
196 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local
239 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local
1982 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local
[all...]
/drivers/atm/
H A Dambassador.c688 amb_rxq * rxq = &dev->rxq[pool]; local
693 spin_lock_irqsave (&rxq->lock, flags);
695 if (rxq->pending < rxq->maximum) {
696 PRINTD (DBG_RX, "RX in slot %p", rxq->in.ptr);
698 *rxq->in.ptr = *rx;
699 rxq->pending++;
700 rxq->in.ptr = NEXTQ (rxq
713 amb_rxq * rxq = &dev->rxq[pool]; local
746 amb_rxq * rxq = &dev->rxq[pool]; local
784 amb_rxq * rxq; local
1590 amb_rxq * rxq = &dev->rxq[pool]; local
[all...]
H A Dfore200e.c1181 struct host_rxq* rxq = &fore200e->host_rxq; local
1188 entry = &rxq->host_entry[ rxq->head ];
1219 FORE200E_NEXT_ENTRY(rxq->head, QUEUE_SIZE_RX);
2215 struct host_rxq* rxq = &fore200e->host_rxq; local
2223 &rxq->status,
2232 &rxq->rpd,
2237 fore200e->bus->dma_chunk_free(fore200e, &rxq->status);
2247 rxq->host_entry[ i ].status =
2248 FORE200E_INDEX(rxq
[all...]
/drivers/net/ethernet/marvell/
H A Dmv643xx_eth.c424 struct rx_queue rxq[8]; member in struct:mv643xx_eth_private
459 /* rxq/txq helper functions *************************************************/
460 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) argument
462 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]);
470 static void rxq_enable(struct rx_queue *rxq) argument
472 struct mv643xx_eth_private *mp = rxq_to_mp(rxq);
473 wrlp(mp, RXQ_COMMAND, 1 << rxq->index);
476 static void rxq_disable(struct rx_queue *rxq) argument
553 rxq_process(struct rx_queue *rxq, int budget) argument
655 rxq_refill(struct rx_queue *rxq, int budget) argument
1209 struct rx_queue *rxq = mp->rxq + i; local
1849 struct rx_queue *rxq = mp->rxq + index; local
1929 rxq_deinit(struct rx_queue *rxq) argument
2282 struct rx_queue *rxq = mp->rxq + i; local
[all...]
/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dcxgb4vf_main.c286 int rxq, msi, err; local
300 for_each_ethrxq(s, rxq) {
304 &s->ethrxq[rxq].rspq);
312 while (--rxq >= 0)
313 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq);
324 int rxq, msi; local
328 for_each_ethrxq(s, rxq)
330 &s->ethrxq[rxq].rspq);
355 int rxq; local
358 for_each_ethrxq(s, rxq)
380 int rxq; local
521 struct sge_eth_rxq *rxq = &s->ethrxq[pi->first_qset]; local
552 struct sge_eth_rxq *rxq = &s->ethrxq[pi->first_qset]; local
594 struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; local
1156 struct sge_eth_rxq *rxq; local
1455 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; local
1618 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL]; local
1775 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL]; local
2329 struct sge_eth_rxq *rxq = &s->ethrxq[qs]; local
[all...]
H A Dsge.c1454 * @rxq: ingress RX Ethernet Queue
1461 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, argument
1467 skb = napi_get_frags(&rxq->rspq.napi);
1470 rxq->stats.rx_drops++;
1479 skb_record_rx_queue(skb, rxq->rspq.idx);
1483 ret = napi_gro_frags(&rxq->rspq.napi);
1486 rxq->stats.lro_pkts++;
1488 rxq->stats.lro_merged++;
1489 rxq->stats.pkts++;
1490 rxq
1507 struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); local
1639 struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); local
1956 struct sge_eth_rxq *rxq; local
2346 struct sge_eth_rxq *rxq = s->ethrxq; local
[all...]
/drivers/net/wireless/iwlegacy/
H A D3945-mac.c893 * + A list of pre-allocated SKBs is stored in iwl->rxq->rx_free. When
894 * iwl->rxq->free_count drops to or below RX_LOW_WATERMARK, work is scheduled
895 * to replenish the iwl->rxq->rx_free.
897 * iwl->rxq is replenished and the READ IDX is updated (updating the
900 * detached from the iwl->rxq. The driver 'processed' idx is updated.
901 * + The Host/Firmware iwl->rxq is replenished at tasklet time from the rx_free
902 * list. If there are no allocated buffers in iwl->rxq->rx_free, the READ
949 struct il_rx_queue *rxq = &il->rxq; local
955 spin_lock_irqsave(&rxq
998 struct il_rx_queue *rxq = &il->rxq; local
1065 il3945_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq) argument
1121 il3945_rx_queue_free(struct il_priv *il, struct il_rx_queue *rxq) argument
1192 struct il_rx_queue *rxq = &il->rxq; local
[all...]
H A D4965-mac.c101 il4965_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq) argument
105 spin_lock_irqsave(&rxq->lock, flags);
106 INIT_LIST_HEAD(&rxq->rx_free);
107 INIT_LIST_HEAD(&rxq->rx_used);
112 if (rxq->pool[i].page != NULL) {
113 pci_unmap_page(il->pci_dev, rxq->pool[i].page_dma,
116 __il_free_pages(il, rxq->pool[i].page);
117 rxq->pool[i].page = NULL;
119 list_add_tail(&rxq->pool[i].list, &rxq
134 il4965_rx_init(struct il_priv *il, struct il_rx_queue *rxq) argument
199 struct il_rx_queue *rxq = &il->rxq; local
268 struct il_rx_queue *rxq = &il->rxq; local
318 struct il_rx_queue *rxq = &il->rxq; local
420 il4965_rx_queue_free(struct il_priv *il, struct il_rx_queue *rxq) argument
4185 struct il_rx_queue *rxq = &il->rxq; local
[all...]
H A D4965.h58 void il4965_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq);
59 int il4965_rx_init(struct il_priv *il, struct il_rx_queue *rxq);
69 void il4965_rx_queue_free(struct il_priv *il, struct il_rx_queue *rxq);
H A D3945.c769 il3945_rx_init(struct il_priv *il, struct il_rx_queue *rxq) argument
771 il_wr(il, FH39_RCSR_RBD_BASE(0), rxq->bd_dma);
772 il_wr(il, FH39_RCSR_RPTR_ADDR(0), rxq->rb_stts_dma);
948 struct il_rx_queue *rxq = &il->rxq; local
958 if (!rxq->bd) {
965 il3945_rx_queue_reset(il, rxq);
969 il3945_rx_init(il, rxq);
972 rxq->need_update = 1;
973 il_rx_queue_update_write_ptr(il, rxq);
[all...]
H A Dcommon.c2463 * + A list of pre-allocated SKBs is stored in iwl->rxq->rx_free. When
2464 * iwl->rxq->free_count drops to or below RX_LOW_WATERMARK, work is scheduled
2465 * to replenish the iwl->rxq->rx_free.
2467 * iwl->rxq is replenished and the READ IDX is updated (updating the
2470 * detached from the iwl->rxq. The driver 'processed' idx is updated.
2471 * + The Host/Firmware iwl->rxq is replenished at tasklet time from the rx_free
2472 * list. If there are no allocated buffers in iwl->rxq->rx_free, the READ
2561 struct il_rx_queue *rxq = &il->rxq; local
2565 spin_lock_init(&rxq
[all...]
/drivers/net/ethernet/sfc/
H A Dvfdi.h151 * @u.mac_filter.rxq: Insert MAC filter at VF local address/VLAN targetting
198 u32 rxq; member in struct:vfdi_req::__anon2918::__anon2922
/drivers/net/ethernet/brocade/bna/
H A Dbnad_ethtool.c648 sprintf(string, "rxq%d_packets", q_num);
650 sprintf(string, "rxq%d_bytes", q_num);
652 sprintf(string, "rxq%d_packets_with_error",
655 sprintf(string, "rxq%d_allocbuf_failed", q_num);
657 sprintf(string, "rxq%d_producer_index", q_num);
659 sprintf(string, "rxq%d_consumer_index", q_num);
666 rcb[1]->rxq) {
667 sprintf(string, "rxq%d_packets", q_num);
669 sprintf(string, "rxq%d_bytes", q_num);
672 "rxq
[all...]
H A Dbna.h311 (q0) = rxp->rxq.single.only; \
315 (q0) = rxp->rxq.slr.large; \
316 (q1) = rxp->rxq.slr.small; \
319 (q0) = rxp->rxq.hds.data; \
320 (q1) = rxp->rxq.hds.hdr; \
H A Dbna_tx_rx.c1793 struct bna_rxq *rxq = NULL; local
1798 rxq = (struct bna_rxq *)qe;
1799 bfa_q_qe_init(&rxq->qe);
1801 return rxq;
1805 bna_rxq_put(struct bna_rx_mod *rx_mod, struct bna_rxq *rxq) argument
1807 bfa_q_qe_init(&rxq->qe);
1808 list_add_tail(&rxq->qe, &rx_mod->rxq_free_q);
1892 rxp->rxq.single.only = q0;
1893 rxp->rxq.single.reserved = NULL;
1896 rxp->rxq
1909 bna_rxq_qpt_setup(struct bna_rxq *rxq, struct bna_rxp *rxp, u32 page_count, u32 page_size, struct bna_mem_descr *qpt_mem, struct bna_mem_descr *swqpt_mem, struct bna_mem_descr *page_mem) argument
[all...]
H A Dbna_types.h606 struct bna_rxq *rxq; member in struct:bna_rcb
732 union bna_rxq_u rxq; member in struct:bna_rxp
879 struct bna_rxq *rxq; /* BFI_MAX_RXQ entries */ member in struct:bna_rx_mod
/drivers/net/ethernet/ibm/
H A Dibmveth.h51 #define h_register_logical_lan(ua, buflst, rxq, fltlst, mac) \
52 plpar_hcall_norets(H_REGISTER_LOGICAL_LAN, ua, buflst, rxq, fltlst, mac)
/drivers/net/
H A Dmacvtap.c163 * Select a queue based on the rxq of the device on which this packet
175 __u32 rxq; local
181 rxq = skb_get_rxhash(skb);
182 if (rxq) {
183 tap = rcu_dereference(vlan->taps[rxq % numvtaps]);
189 rxq = skb_get_rx_queue(skb);
191 while (unlikely(rxq >= numvtaps))
192 rxq -= numvtaps;
194 tap = rcu_dereference(vlan->taps[rxq]);
200 for (rxq
[all...]
/drivers/net/ethernet/chelsio/cxgb4/
H A Dsge.c1514 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, argument
1520 skb = napi_get_frags(&rxq->rspq.napi);
1523 rxq->stats.rx_drops++;
1532 skb_record_rx_queue(skb, rxq->rspq.idx);
1533 if (rxq->rspq.netdev->features & NETIF_F_RXHASH)
1538 rxq->stats.vlan_ex++;
1540 ret = napi_gro_frags(&rxq->rspq.napi);
1542 rxq->stats.lro_pkts++;
1544 rxq->stats.lro_merged++;
1545 rxq
1563 struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq); local
1693 struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq); local
1895 struct sge_eth_rxq *rxq; local
[all...]
/drivers/net/wireless/
H A Dmwl8k.c229 struct mwl8k_rx_queue rxq[MWL8K_RX_QUEUES]; member in struct:mwl8k_priv
1117 struct mwl8k_rx_queue *rxq = priv->rxq + index; local
1121 rxq->rxd_count = 0;
1122 rxq->head = 0;
1123 rxq->tail = 0;
1127 rxq->rxd = pci_alloc_consistent(priv->pdev, size, &rxq->rxd_dma);
1128 if (rxq->rxd == NULL) {
1132 memset(rxq
1164 struct mwl8k_rx_queue *rxq = priv->rxq + index; local
1201 struct mwl8k_rx_queue *rxq = priv->rxq + index; local
1278 struct mwl8k_rx_queue *rxq = priv->rxq + index; local
[all...]
/drivers/net/usb/
H A Dusbnet.c365 spin_lock_irqsave (&dev->rxq.lock, lockflags);
391 __usbnet_queue_skb(&dev->rxq, skb, rx_start);
397 spin_unlock_irqrestore (&dev->rxq.lock, lockflags);
508 state = defer_bh(dev, skb, &dev->rxq, state);
648 (void) unlink_urbs (dev, &dev->rxq);
668 unlink_urbs(dev, &dev->rxq);
671 while (!skb_queue_empty(&dev->rxq)
951 unlink_urbs (dev, &dev->rxq);
1234 if ((dev->txq.qlen + dev->rxq.qlen + dev->done.qlen) == 0) {
1243 int temp = dev->rxq
[all...]
/drivers/net/ethernet/xscale/
H A Dixp4xx_eth.c662 qmgr_disable_irq(port->plat->rxq);
670 unsigned int rxq = port->plat->rxq, rxfreeq = RXFREE_QUEUE(port->id); local
686 if ((n = queue_get_desc(rxq, port, 0)) < 0) {
692 qmgr_enable_irq(rxq);
693 if (!qmgr_stat_below_low_watermark(rxq) &&
700 qmgr_disable_irq(rxq);
1023 err = qmgr_request_queue(port->plat->rxq, RX_DESCS, 0, 0,
1052 qmgr_release_queue(port->plat->rxq);
1063 qmgr_release_queue(port->plat->rxq);
[all...]
/drivers/net/wireless/ipw2x00/
H A Dipw2200.c3442 struct ipw_rx_queue *rxq)
3447 spin_lock_irqsave(&rxq->lock, flags);
3449 INIT_LIST_HEAD(&rxq->rx_free);
3450 INIT_LIST_HEAD(&rxq->rx_used);
3456 if (rxq->pool[i].skb != NULL) {
3457 pci_unmap_single(priv->pci_dev, rxq->pool[i].dma_addr,
3459 dev_kfree_skb(rxq->pool[i].skb);
3460 rxq->pool[i].skb = NULL;
3462 list_add_tail(&rxq->pool[i].list, &rxq
3441 ipw_rx_queue_reset(struct ipw_priv *priv, struct ipw_rx_queue *rxq) argument
5171 struct ipw_rx_queue *rxq = priv->rxq; local
5211 struct ipw_rx_queue *rxq = priv->rxq; local
5257 ipw_rx_queue_free(struct ipw_priv *priv, struct ipw_rx_queue *rxq) argument
5277 struct ipw_rx_queue *rxq; local
[all...]
/drivers/net/ethernet/emulex/benet/
H A Dbe_main.c1110 struct be_queue_info *rxq = &rxo->q; local
1122 atomic_dec(&rxq->used);
1130 struct be_queue_info *rxq = &rxo->q; local
1138 index_inc(&rxcp->rxq_idx, rxq->len);
1149 struct be_queue_info *rxq = &rxo->q; local
1189 index_inc(&rxcp->rxq_idx, rxq->len);
1213 index_inc(&rxcp->rxq_idx, rxq->len);
1259 struct be_queue_info *rxq = &rxo->q; local
1289 index_inc(&rxcp->rxq_idx, rxq->len);
1428 struct be_queue_info *rxq local
1577 struct be_queue_info *rxq = &rxo->q; local
[all...]

Completed in 4332 milliseconds

12