Lines Matching defs:rds_ring

49 		struct nx_host_rds_ring *rds_ring);
110 struct nx_host_rds_ring *rds_ring;
116 rds_ring = &recv_ctx->rds_rings[ring];
117 for (i = 0; i < rds_ring->num_desc; ++i) {
118 rx_buf = &(rds_ring->rx_buf_arr[i]);
123 rds_ring->dma_size,
168 struct nx_host_rds_ring *rds_ring;
178 rds_ring = &recv_ctx->rds_rings[ring];
179 vfree(rds_ring->rx_buf_arr);
180 rds_ring->rx_buf_arr = NULL;
197 struct nx_host_rds_ring *rds_ring;
223 rds_ring = kcalloc(adapter->max_rds_rings,
225 if (rds_ring == NULL)
228 recv_ctx->rds_rings = rds_ring;
231 rds_ring = &recv_ctx->rds_rings[ring];
234 rds_ring->num_desc = adapter->num_rxd;
236 rds_ring->dma_size =
238 rds_ring->skb_size =
242 rds_ring->dma_size =
245 rds_ring->dma_size =
247 rds_ring->skb_size =
248 rds_ring->dma_size + NET_IP_ALIGN;
253 rds_ring->num_desc = adapter->num_jumbo_rxd;
255 rds_ring->dma_size =
258 rds_ring->dma_size =
262 rds_ring->dma_size += NX_LRO_BUFFER_EXTRA;
264 rds_ring->skb_size =
265 rds_ring->dma_size + NET_IP_ALIGN;
269 rds_ring->num_desc = adapter->num_lro_rxd;
270 rds_ring->dma_size = NX_RX_LRO_BUFFER_LENGTH;
271 rds_ring->skb_size = rds_ring->dma_size + NET_IP_ALIGN;
275 rds_ring->rx_buf_arr = vzalloc(RCV_BUFF_RINGSIZE(rds_ring));
276 if (rds_ring->rx_buf_arr == NULL)
280 INIT_LIST_HEAD(&rds_ring->free_list);
285 rx_buf = rds_ring->rx_buf_arr;
286 for (i = 0; i < rds_ring->num_desc; i++) {
288 &rds_ring->free_list);
293 spin_lock_init(&rds_ring->lock);
1474 struct nx_host_rds_ring *rds_ring,
1481 buffer->skb = netdev_alloc_skb(adapter->netdev, rds_ring->skb_size);
1491 rds_ring->dma_size, PCI_DMA_FROMDEVICE);
1507 struct nx_host_rds_ring *rds_ring, u16 index, u16 cksum)
1512 buffer = &rds_ring->rx_buf_arr[index];
1514 pci_unmap_single(adapter->pdev, buffer->dma, rds_ring->dma_size,
1543 struct nx_host_rds_ring *rds_ring;
1549 rds_ring = &recv_ctx->rds_rings[ring];
1552 if (unlikely(index >= rds_ring->num_desc))
1555 buffer = &rds_ring->rx_buf_arr[index];
1561 skb = netxen_process_rxbuf(adapter, rds_ring, index, cksum);
1565 if (length > rds_ring->skb_size)
1566 skb_put(skb, rds_ring->skb_size);
1597 struct nx_host_rds_ring *rds_ring;
1610 rds_ring = &recv_ctx->rds_rings[ring];
1613 if (unlikely(index >= rds_ring->num_desc))
1616 buffer = &rds_ring->rx_buf_arr[index];
1625 skb = netxen_process_rxbuf(adapter, rds_ring, index, STATUS_CKSUM_OK);
1729 struct nx_host_rds_ring *rds_ring =
1736 netxen_alloc_rx_skb(adapter, rds_ring, rxbuf);
1738 spin_lock(&rds_ring->lock);
1740 &rds_ring->free_list);
1741 spin_unlock(&rds_ring->lock);
1744 netxen_post_rx_buffers_nodb(adapter, rds_ring);
1829 struct nx_host_rds_ring *rds_ring)
1837 producer = rds_ring->producer;
1839 head = &rds_ring->free_list;
1845 if (netxen_alloc_rx_skb(adapter, rds_ring, buffer))
1853 pdesc = &rds_ring->desc_head[producer];
1856 pdesc->buffer_length = cpu_to_le32(rds_ring->dma_size);
1858 producer = get_next_index(producer, rds_ring->num_desc);
1862 rds_ring->producer = producer;
1863 NXWRIO(adapter, rds_ring->crb_rcv_producer,
1864 (producer-1) & (rds_ring->num_desc-1));
1876 (rds_ring->num_desc - 1)));
1887 struct nx_host_rds_ring *rds_ring)
1894 if (!spin_trylock(&rds_ring->lock))
1897 producer = rds_ring->producer;
1899 head = &rds_ring->free_list;
1905 if (netxen_alloc_rx_skb(adapter, rds_ring, buffer))
1913 pdesc = &rds_ring->desc_head[producer];
1915 pdesc->buffer_length = cpu_to_le32(rds_ring->dma_size);
1918 producer = get_next_index(producer, rds_ring->num_desc);
1922 rds_ring->producer = producer;
1923 NXWRIO(adapter, rds_ring->crb_rcv_producer,
1924 (producer - 1) & (rds_ring->num_desc - 1));
1926 spin_unlock(&rds_ring->lock);