/drivers/net/wireless/iwlwifi/ |
H A D | iwl-trans-pcie-rx.c | 80 * + A list of pre-allocated SKBs is stored in iwl->rxq->rx_free. When 81 * iwl->rxq->free_count drops to or below RX_LOW_WATERMARK, work is scheduled 82 * to replenish the iwl->rxq->rx_free. 84 * iwl->rxq is replenished and the READ INDEX is updated (updating the 87 * detached from the iwl->rxq. The driver 'processed' index is updated. 88 * + The Host/Firmware iwl->rxq is replenished at tasklet time from the rx_free 89 * list. If there are no allocated buffers in iwl->rxq->rx_free, the READ 204 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local 209 spin_lock_irqsave(&rxq 256 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local 363 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local 470 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local [all...] |
H A D | iwl-trans-pcie.c | 89 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local 92 memset(&trans_pcie->rxq, 0, sizeof(trans_pcie->rxq)); 94 spin_lock_init(&rxq->lock); 96 if (WARN_ON(rxq->bd || rxq->rb_stts)) 100 rxq->bd = dma_zalloc_coherent(dev, sizeof(__le32) * RX_QUEUE_SIZE, 101 &rxq->bd_dma, GFP_KERNEL); 102 if (!rxq 126 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local 145 iwl_trans_rx_hw_init(struct iwl_trans *trans, struct iwl_rx_queue *rxq) argument 196 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local 239 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local 1982 struct iwl_rx_queue *rxq = &trans_pcie->rxq; local [all...] |
/drivers/atm/ |
H A D | ambassador.c | 688 amb_rxq * rxq = &dev->rxq[pool]; local 693 spin_lock_irqsave (&rxq->lock, flags); 695 if (rxq->pending < rxq->maximum) { 696 PRINTD (DBG_RX, "RX in slot %p", rxq->in.ptr); 698 *rxq->in.ptr = *rx; 699 rxq->pending++; 700 rxq->in.ptr = NEXTQ (rxq 713 amb_rxq * rxq = &dev->rxq[pool]; local 746 amb_rxq * rxq = &dev->rxq[pool]; local 784 amb_rxq * rxq; local 1590 amb_rxq * rxq = &dev->rxq[pool]; local [all...] |
H A D | fore200e.c | 1181 struct host_rxq* rxq = &fore200e->host_rxq; local 1188 entry = &rxq->host_entry[ rxq->head ]; 1219 FORE200E_NEXT_ENTRY(rxq->head, QUEUE_SIZE_RX); 2215 struct host_rxq* rxq = &fore200e->host_rxq; local 2223 &rxq->status, 2232 &rxq->rpd, 2237 fore200e->bus->dma_chunk_free(fore200e, &rxq->status); 2247 rxq->host_entry[ i ].status = 2248 FORE200E_INDEX(rxq [all...] |
/drivers/net/ethernet/marvell/ |
H A D | mv643xx_eth.c | 424 struct rx_queue rxq[8]; member in struct:mv643xx_eth_private 459 /* rxq/txq helper functions *************************************************/ 460 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) argument 462 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); 470 static void rxq_enable(struct rx_queue *rxq) argument 472 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); 473 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); 476 static void rxq_disable(struct rx_queue *rxq) argument 553 rxq_process(struct rx_queue *rxq, int budget) argument 655 rxq_refill(struct rx_queue *rxq, int budget) argument 1209 struct rx_queue *rxq = mp->rxq + i; local 1849 struct rx_queue *rxq = mp->rxq + index; local 1929 rxq_deinit(struct rx_queue *rxq) argument 2282 struct rx_queue *rxq = mp->rxq + i; local [all...] |
/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | cxgb4vf_main.c | 286 int rxq, msi, err; local 300 for_each_ethrxq(s, rxq) { 304 &s->ethrxq[rxq].rspq); 312 while (--rxq >= 0) 313 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); 324 int rxq, msi; local 328 for_each_ethrxq(s, rxq) 330 &s->ethrxq[rxq].rspq); 355 int rxq; local 358 for_each_ethrxq(s, rxq) 380 int rxq; local 521 struct sge_eth_rxq *rxq = &s->ethrxq[pi->first_qset]; local 552 struct sge_eth_rxq *rxq = &s->ethrxq[pi->first_qset]; local 594 struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; local 1156 struct sge_eth_rxq *rxq; local 1455 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; local 1618 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL]; local 1775 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL]; local 2329 struct sge_eth_rxq *rxq = &s->ethrxq[qs]; local [all...] |
H A D | sge.c | 1454 * @rxq: ingress RX Ethernet Queue 1461 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, argument 1467 skb = napi_get_frags(&rxq->rspq.napi); 1470 rxq->stats.rx_drops++; 1479 skb_record_rx_queue(skb, rxq->rspq.idx); 1483 ret = napi_gro_frags(&rxq->rspq.napi); 1486 rxq->stats.lro_pkts++; 1488 rxq->stats.lro_merged++; 1489 rxq->stats.pkts++; 1490 rxq 1507 struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); local 1639 struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); local 1956 struct sge_eth_rxq *rxq; local 2346 struct sge_eth_rxq *rxq = s->ethrxq; local [all...] |
/drivers/net/wireless/iwlegacy/ |
H A D | 3945-mac.c | 893 * + A list of pre-allocated SKBs is stored in iwl->rxq->rx_free. When 894 * iwl->rxq->free_count drops to or below RX_LOW_WATERMARK, work is scheduled 895 * to replenish the iwl->rxq->rx_free. 897 * iwl->rxq is replenished and the READ IDX is updated (updating the 900 * detached from the iwl->rxq. The driver 'processed' idx is updated. 901 * + The Host/Firmware iwl->rxq is replenished at tasklet time from the rx_free 902 * list. If there are no allocated buffers in iwl->rxq->rx_free, the READ 949 struct il_rx_queue *rxq = &il->rxq; local 955 spin_lock_irqsave(&rxq 998 struct il_rx_queue *rxq = &il->rxq; local 1065 il3945_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq) argument 1121 il3945_rx_queue_free(struct il_priv *il, struct il_rx_queue *rxq) argument 1192 struct il_rx_queue *rxq = &il->rxq; local [all...] |
H A D | 4965-mac.c | 101 il4965_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq) argument 105 spin_lock_irqsave(&rxq->lock, flags); 106 INIT_LIST_HEAD(&rxq->rx_free); 107 INIT_LIST_HEAD(&rxq->rx_used); 112 if (rxq->pool[i].page != NULL) { 113 pci_unmap_page(il->pci_dev, rxq->pool[i].page_dma, 116 __il_free_pages(il, rxq->pool[i].page); 117 rxq->pool[i].page = NULL; 119 list_add_tail(&rxq->pool[i].list, &rxq 134 il4965_rx_init(struct il_priv *il, struct il_rx_queue *rxq) argument 199 struct il_rx_queue *rxq = &il->rxq; local 268 struct il_rx_queue *rxq = &il->rxq; local 318 struct il_rx_queue *rxq = &il->rxq; local 420 il4965_rx_queue_free(struct il_priv *il, struct il_rx_queue *rxq) argument 4185 struct il_rx_queue *rxq = &il->rxq; local [all...] |
H A D | 4965.h | 58 void il4965_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq); 59 int il4965_rx_init(struct il_priv *il, struct il_rx_queue *rxq); 69 void il4965_rx_queue_free(struct il_priv *il, struct il_rx_queue *rxq);
|
H A D | 3945.c | 769 il3945_rx_init(struct il_priv *il, struct il_rx_queue *rxq) argument 771 il_wr(il, FH39_RCSR_RBD_BASE(0), rxq->bd_dma); 772 il_wr(il, FH39_RCSR_RPTR_ADDR(0), rxq->rb_stts_dma); 948 struct il_rx_queue *rxq = &il->rxq; local 958 if (!rxq->bd) { 965 il3945_rx_queue_reset(il, rxq); 969 il3945_rx_init(il, rxq); 972 rxq->need_update = 1; 973 il_rx_queue_update_write_ptr(il, rxq); [all...] |
H A D | common.c | 2463 * + A list of pre-allocated SKBs is stored in iwl->rxq->rx_free. When 2464 * iwl->rxq->free_count drops to or below RX_LOW_WATERMARK, work is scheduled 2465 * to replenish the iwl->rxq->rx_free. 2467 * iwl->rxq is replenished and the READ IDX is updated (updating the 2470 * detached from the iwl->rxq. The driver 'processed' idx is updated. 2471 * + The Host/Firmware iwl->rxq is replenished at tasklet time from the rx_free 2472 * list. If there are no allocated buffers in iwl->rxq->rx_free, the READ 2561 struct il_rx_queue *rxq = &il->rxq; local 2565 spin_lock_init(&rxq [all...] |
/drivers/net/ethernet/sfc/ |
H A D | vfdi.h | 151 * @u.mac_filter.rxq: Insert MAC filter at VF local address/VLAN targetting 198 u32 rxq; member in struct:vfdi_req::__anon2918::__anon2922
|
/drivers/net/ethernet/brocade/bna/ |
H A D | bnad_ethtool.c | 648 sprintf(string, "rxq%d_packets", q_num); 650 sprintf(string, "rxq%d_bytes", q_num); 652 sprintf(string, "rxq%d_packets_with_error", 655 sprintf(string, "rxq%d_allocbuf_failed", q_num); 657 sprintf(string, "rxq%d_producer_index", q_num); 659 sprintf(string, "rxq%d_consumer_index", q_num); 666 rcb[1]->rxq) { 667 sprintf(string, "rxq%d_packets", q_num); 669 sprintf(string, "rxq%d_bytes", q_num); 672 "rxq [all...] |
H A D | bna.h | 311 (q0) = rxp->rxq.single.only; \ 315 (q0) = rxp->rxq.slr.large; \ 316 (q1) = rxp->rxq.slr.small; \ 319 (q0) = rxp->rxq.hds.data; \ 320 (q1) = rxp->rxq.hds.hdr; \
|
H A D | bna_tx_rx.c | 1793 struct bna_rxq *rxq = NULL; local 1798 rxq = (struct bna_rxq *)qe; 1799 bfa_q_qe_init(&rxq->qe); 1801 return rxq; 1805 bna_rxq_put(struct bna_rx_mod *rx_mod, struct bna_rxq *rxq) argument 1807 bfa_q_qe_init(&rxq->qe); 1808 list_add_tail(&rxq->qe, &rx_mod->rxq_free_q); 1892 rxp->rxq.single.only = q0; 1893 rxp->rxq.single.reserved = NULL; 1896 rxp->rxq 1909 bna_rxq_qpt_setup(struct bna_rxq *rxq, struct bna_rxp *rxp, u32 page_count, u32 page_size, struct bna_mem_descr *qpt_mem, struct bna_mem_descr *swqpt_mem, struct bna_mem_descr *page_mem) argument [all...] |
H A D | bna_types.h | 606 struct bna_rxq *rxq; member in struct:bna_rcb 732 union bna_rxq_u rxq; member in struct:bna_rxp 879 struct bna_rxq *rxq; /* BFI_MAX_RXQ entries */ member in struct:bna_rx_mod
|
/drivers/net/ethernet/ibm/ |
H A D | ibmveth.h | 51 #define h_register_logical_lan(ua, buflst, rxq, fltlst, mac) \ 52 plpar_hcall_norets(H_REGISTER_LOGICAL_LAN, ua, buflst, rxq, fltlst, mac)
|
/drivers/net/ |
H A D | macvtap.c | 163 * Select a queue based on the rxq of the device on which this packet 175 __u32 rxq; local 181 rxq = skb_get_rxhash(skb); 182 if (rxq) { 183 tap = rcu_dereference(vlan->taps[rxq % numvtaps]); 189 rxq = skb_get_rx_queue(skb); 191 while (unlikely(rxq >= numvtaps)) 192 rxq -= numvtaps; 194 tap = rcu_dereference(vlan->taps[rxq]); 200 for (rxq [all...] |
/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 1514 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, argument 1520 skb = napi_get_frags(&rxq->rspq.napi); 1523 rxq->stats.rx_drops++; 1532 skb_record_rx_queue(skb, rxq->rspq.idx); 1533 if (rxq->rspq.netdev->features & NETIF_F_RXHASH) 1538 rxq->stats.vlan_ex++; 1540 ret = napi_gro_frags(&rxq->rspq.napi); 1542 rxq->stats.lro_pkts++; 1544 rxq->stats.lro_merged++; 1545 rxq 1563 struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq); local 1693 struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq); local 1895 struct sge_eth_rxq *rxq; local [all...] |
/drivers/net/wireless/ |
H A D | mwl8k.c | 229 struct mwl8k_rx_queue rxq[MWL8K_RX_QUEUES]; member in struct:mwl8k_priv 1117 struct mwl8k_rx_queue *rxq = priv->rxq + index; local 1121 rxq->rxd_count = 0; 1122 rxq->head = 0; 1123 rxq->tail = 0; 1127 rxq->rxd = pci_alloc_consistent(priv->pdev, size, &rxq->rxd_dma); 1128 if (rxq->rxd == NULL) { 1132 memset(rxq 1164 struct mwl8k_rx_queue *rxq = priv->rxq + index; local 1201 struct mwl8k_rx_queue *rxq = priv->rxq + index; local 1278 struct mwl8k_rx_queue *rxq = priv->rxq + index; local [all...] |
/drivers/net/usb/ |
H A D | usbnet.c | 365 spin_lock_irqsave (&dev->rxq.lock, lockflags); 391 __usbnet_queue_skb(&dev->rxq, skb, rx_start); 397 spin_unlock_irqrestore (&dev->rxq.lock, lockflags); 508 state = defer_bh(dev, skb, &dev->rxq, state); 648 (void) unlink_urbs (dev, &dev->rxq); 668 unlink_urbs(dev, &dev->rxq); 671 while (!skb_queue_empty(&dev->rxq) 951 unlink_urbs (dev, &dev->rxq); 1234 if ((dev->txq.qlen + dev->rxq.qlen + dev->done.qlen) == 0) { 1243 int temp = dev->rxq [all...] |
/drivers/net/ethernet/xscale/ |
H A D | ixp4xx_eth.c | 662 qmgr_disable_irq(port->plat->rxq); 670 unsigned int rxq = port->plat->rxq, rxfreeq = RXFREE_QUEUE(port->id); local 686 if ((n = queue_get_desc(rxq, port, 0)) < 0) { 692 qmgr_enable_irq(rxq); 693 if (!qmgr_stat_below_low_watermark(rxq) && 700 qmgr_disable_irq(rxq); 1023 err = qmgr_request_queue(port->plat->rxq, RX_DESCS, 0, 0, 1052 qmgr_release_queue(port->plat->rxq); 1063 qmgr_release_queue(port->plat->rxq); [all...] |
/drivers/net/wireless/ipw2x00/ |
H A D | ipw2200.c | 3442 struct ipw_rx_queue *rxq) 3447 spin_lock_irqsave(&rxq->lock, flags); 3449 INIT_LIST_HEAD(&rxq->rx_free); 3450 INIT_LIST_HEAD(&rxq->rx_used); 3456 if (rxq->pool[i].skb != NULL) { 3457 pci_unmap_single(priv->pci_dev, rxq->pool[i].dma_addr, 3459 dev_kfree_skb(rxq->pool[i].skb); 3460 rxq->pool[i].skb = NULL; 3462 list_add_tail(&rxq->pool[i].list, &rxq 3441 ipw_rx_queue_reset(struct ipw_priv *priv, struct ipw_rx_queue *rxq) argument 5171 struct ipw_rx_queue *rxq = priv->rxq; local 5211 struct ipw_rx_queue *rxq = priv->rxq; local 5257 ipw_rx_queue_free(struct ipw_priv *priv, struct ipw_rx_queue *rxq) argument 5277 struct ipw_rx_queue *rxq; local [all...] |
/drivers/net/ethernet/emulex/benet/ |
H A D | be_main.c | 1110 struct be_queue_info *rxq = &rxo->q; local 1122 atomic_dec(&rxq->used); 1130 struct be_queue_info *rxq = &rxo->q; local 1138 index_inc(&rxcp->rxq_idx, rxq->len); 1149 struct be_queue_info *rxq = &rxo->q; local 1189 index_inc(&rxcp->rxq_idx, rxq->len); 1213 index_inc(&rxcp->rxq_idx, rxq->len); 1259 struct be_queue_info *rxq = &rxo->q; local 1289 index_inc(&rxcp->rxq_idx, rxq->len); 1428 struct be_queue_info *rxq local 1577 struct be_queue_info *rxq = &rxo->q; local [all...] |