/drivers/net/ethernet/intel/fm10k/ |
H A D | fm10k_main.c | 120 union fm10k_rx_desc *rx_desc; local 128 rx_desc = FM10K_RX_DESC(rx_ring, i); 139 rx_desc->q.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); 141 rx_desc++; 145 rx_desc = FM10K_RX_DESC(rx_ring, 0); 151 rx_desc->q.hdr_addr = 0; 244 * @rx_desc: descriptor containing length of buffer written by hardware 257 union fm10k_rx_desc *rx_desc, 261 unsigned int size = le16_to_cpu(rx_desc->w.length); 289 union fm10k_rx_desc *rx_desc, 255 fm10k_add_rx_frag(struct fm10k_ring *rx_ring, struct fm10k_rx_buffer *rx_buffer, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 288 fm10k_fetch_rx_buffer(struct fm10k_ring *rx_ring, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 348 fm10k_rx_checksum(struct fm10k_ring *ring, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 383 fm10k_rx_hash(struct fm10k_ring *ring, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 401 fm10k_rx_hwtstamp(struct fm10k_ring *rx_ring, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 414 fm10k_type_trans(struct fm10k_ring *rx_ring, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 453 fm10k_process_skb_fields(struct fm10k_ring *rx_ring, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 493 fm10k_is_non_eop(struct fm10k_ring *rx_ring, union fm10k_rx_desc *rx_desc) argument 523 fm10k_pull_tail(struct fm10k_ring *rx_ring, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 566 fm10k_cleanup_headers(struct fm10k_ring *rx_ring, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 613 union fm10k_rx_desc *rx_desc; local [all...] |
/drivers/infiniband/ulp/iser/ |
H A D | iser_initiator.c | 260 struct iser_rx_desc *rx_desc; local 281 rx_desc = iser_conn->rx_descs; 283 for (i = 0; i < iser_conn->qp_max_recv_dtos; i++, rx_desc++) { 284 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc, 289 rx_desc->dma_addr = dma_addr; 291 rx_sg = &rx_desc->rx_sg; 292 rx_sg->addr = rx_desc->dma_addr; 301 rx_desc = iser_conn->rx_descs; 302 for (j = 0; j < i; j++, rx_desc++) 303 ib_dma_unmap_single(device->ib_device, rx_desc 319 struct iser_rx_desc *rx_desc; local 577 iser_rcv_completion(struct iser_rx_desc *rx_desc, unsigned long rx_xfer_len, struct ib_conn *ib_conn) argument [all...] |
H A D | iser_verbs.c | 1072 struct iser_rx_desc *rx_desc; local 1075 rx_desc = &iser_conn->rx_descs[my_rx_head]; 1076 rx_wr->wr_id = (unsigned long)rx_desc; 1077 rx_wr->sg_list = &rx_desc->rx_sg; 1193 struct iser_rx_desc *rx_desc; local 1198 rx_desc = (struct iser_rx_desc *)wc->wr_id; 1199 iser_rcv_completion(rx_desc, wc->byte_len,
|
/drivers/net/ethernet/seeq/ |
H A D | sgiseeq.c | 58 (unsigned long)((sp)->rx_desc))) 94 struct sgiseeq_rx_desc *rx_desc; member in struct:sgiseeq_private 197 if (!sp->rx_desc[i].skb) { 207 sp->rx_desc[i].skb = skb; 208 sp->rx_desc[i].rdma.pbuf = dma_addr; 210 sp->rx_desc[i].rdma.cntinfo = RCNTINFO_INIT; 211 dma_sync_desc_dev(dev, &sp->rx_desc[i]); 213 sp->rx_desc[i - 1].rdma.cntinfo |= HPCDMA_EOR; 214 dma_sync_desc_dev(dev, &sp->rx_desc[i - 1]); 233 if (sp->rx_desc[ [all...] |
/drivers/net/wireless/ti/wl1251/ |
H A D | rx.c | 220 struct wl1251_rx_descriptor *rx_desc; local 225 rx_desc = wl->rx_descriptor; 228 wl1251_rx_header(wl, rx_desc); 231 wl1251_rx_body(wl, rx_desc);
|
/drivers/net/ethernet/ibm/emac/ |
H A D | debug.c | 58 i, p->rx_desc[i].data_ptr, p->rx_skb[i] ? 'V' : ' ', 59 p->rx_desc[i].ctrl, p->rx_desc[i].data_len, 61 p->rx_desc[NUM_RX_BUFF / 2 + i].data_ptr, 63 p->rx_desc[NUM_RX_BUFF / 2 + i].ctrl, 64 p->rx_desc[NUM_RX_BUFF / 2 + i].data_len);
|
/drivers/spi/ |
H A D | spi-pxa2xx-dma.c | 266 struct dma_async_tx_descriptor *tx_desc, *rx_desc; local 275 rx_desc = pxa2xx_spi_dma_prepare_one(drv_data, DMA_DEV_TO_MEM); 276 if (!rx_desc) { 283 rx_desc->callback = pxa2xx_spi_dma_callback; 284 rx_desc->callback_param = drv_data; 286 dmaengine_submit(rx_desc);
|
/drivers/infiniband/hw/amso1100/ |
H A D | c2.c | 163 struct c2_rx_desc *rx_desc; local 173 rx_desc = vaddr; 175 for (i = 0; i < rx_ring->count; i++, elem++, rx_desc++, rxp_desc++) { 176 rx_desc->len = 0; 177 rx_desc->status = 0; 190 elem->ht_desc = rx_desc; 195 rx_desc->next_offset = base; 198 rx_desc->next_offset = 199 base + (i + 1) * sizeof(*rx_desc); 212 struct c2_rx_desc *rx_desc local 282 struct c2_rx_desc *rx_desc; local 422 struct c2_rx_desc *rx_desc = elem->ht_desc; local 467 struct c2_rx_desc *rx_desc; local [all...] |
/drivers/net/wireless/ti/wlcore/ |
H A D | hw_ops.h | 59 wlcore_hw_get_rx_buf_align(struct wl1271 *wl, u32 rx_desc) argument 65 return wl->ops->get_rx_buf_align(wl, rx_desc); 69 wlcore_hw_prepare_read(struct wl1271 *wl, u32 rx_desc, u32 len) argument 72 return wl->ops->prepare_read(wl, rx_desc, len);
|
/drivers/staging/rtl8192e/rtl8192e/ |
H A D | r8192E_dev.h | 58 struct rx_desc *pdesc,
|
/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_txrx.c | 448 * @rx_desc: the Rx descriptor for programming Status, not a packet descriptor. 455 union i40e_rx_desc *rx_desc, u8 prog_id) 463 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); 468 if ((rx_desc->wb.qword0.hi_dword.fd_id != 0) || 471 rx_desc->wb.qword0.hi_dword.fd_id); 500 rx_desc->wb.qword0.hi_dword.fd_id); 908 * @rx_desc: the rx descriptor written back by HW 916 union i40e_rx_desc *rx_desc) 921 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); 926 i40e_fd_handle_status(rx_ring, rx_desc, i 454 i40e_fd_handle_status(struct i40e_ring *rx_ring, union i40e_rx_desc *rx_desc, u8 prog_id) argument 915 i40e_clean_programming_status(struct i40e_ring *rx_ring, union i40e_rx_desc *rx_desc) argument 1115 union i40e_rx_desc *rx_desc; local 1330 i40e_rx_hash(struct i40e_ring *ring, union i40e_rx_desc *rx_desc) argument 1382 union i40e_rx_desc *rx_desc; local [all...] |
/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_rx.c | 87 struct mlx4_en_rx_desc *rx_desc, 115 rx_desc->data[i].addr = cpu_to_be64(dma); 207 struct mlx4_en_rx_desc *rx_desc = ring->buf + ring->stride * index; local 213 rx_desc->data[i].byte_count = 215 rx_desc->data[i].lkey = cpu_to_be32(priv->mdev->mr.key); 223 rx_desc->data[i].byte_count = 0; 224 rx_desc->data[i].lkey = cpu_to_be32(MLX4_EN_MEMTYPE_PAD); 225 rx_desc->data[i].addr = 0; 233 struct mlx4_en_rx_desc *rx_desc = ring->buf + (index * ring->stride); local 237 return mlx4_en_alloc_frags(priv, rx_desc, frag 86 mlx4_en_alloc_frags(struct mlx4_en_priv *priv, struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_alloc *frags, struct mlx4_en_rx_alloc *ring_alloc, gfp_t gfp) argument 515 mlx4_en_complete_rx_desc(struct mlx4_en_priv *priv, struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_alloc *frags, struct sk_buff *skb, int length) argument 560 mlx4_en_rx_skb(struct mlx4_en_priv *priv, struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_alloc *frags, unsigned int length) argument 653 struct mlx4_en_rx_desc *rx_desc; local [all...] |
/drivers/net/ethernet/alteon/ |
H A D | acenic.h | 476 #define RX_STD_RING_SIZE (RX_STD_RING_ENTRIES * sizeof(struct rx_desc)) 479 #define RX_JUMBO_RING_SIZE (RX_JUMBO_RING_ENTRIES *sizeof(struct rx_desc)) 482 #define RX_MINI_RING_SIZE (RX_MINI_RING_ENTRIES *sizeof(struct rx_desc)) 486 sizeof(struct rx_desc)) 488 struct rx_desc{ struct 663 struct rx_desc *rx_std_ring; 664 struct rx_desc *rx_jumbo_ring; 665 struct rx_desc *rx_mini_ring; 666 struct rx_desc *rx_return_ring;
|
/drivers/net/ethernet/ |
H A D | ec_bhf.c | 96 struct rx_desc { struct 148 struct rx_desc *rx_descs; 203 static int ec_bhf_pkt_received(struct rx_desc *desc) 208 static void ec_bhf_add_rx_desc(struct ec_bhf_priv *priv, struct rx_desc *desc) 216 struct rx_desc *desc = &priv->rx_descs[priv->rx_dnext]; 382 priv->rx_dcount = priv->rx_buf.len / sizeof(struct rx_desc); 383 priv->rx_descs = (struct rx_desc *)priv->rx_buf.buf; 387 struct rx_desc *desc = &priv->rx_descs[i]; 410 FIFO_SIZE * sizeof(struct rx_desc));
|
/drivers/net/ethernet/intel/i40evf/ |
H A D | i40e_txrx.c | 620 union i40e_rx_desc *rx_desc; local 629 rx_desc = I40E_RX_DESC(rx_ring, i); 685 rx_desc->read.pkt_addr = cpu_to_le64(bi->page_dma); 686 rx_desc->read.hdr_addr = cpu_to_le64(bi->dma); 688 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma); 689 rx_desc->read.hdr_addr = 0; 833 * @rx_desc: specific descriptor 836 union i40e_rx_desc *rx_desc) 843 (rx_desc->wb.qword1.status_error_len & rss_mask) == rss_mask) 844 return le32_to_cpu(rx_desc 835 i40e_rx_hash(struct i40e_ring *ring, union i40e_rx_desc *rx_desc) argument 887 union i40e_rx_desc *rx_desc; local [all...] |
/drivers/net/ethernet/marvell/ |
H A D | pxa168_eth.c | 181 struct rx_desc { struct 209 struct rx_desc *p_rx_desc_area; 332 struct rx_desc *p_used_rx_desc; 675 (u32) (pep->rx_desc_dma + rx_curr_desc * sizeof(struct rx_desc))); 678 (u32) (pep->rx_desc_dma + rx_curr_desc * sizeof(struct rx_desc))); 803 struct rx_desc *rx_desc; local 811 rx_desc = &pep->p_rx_desc_area[rx_curr_desc]; 812 cmd_sts = rx_desc->cmd_sts; 827 dma_unmap_single(NULL, rx_desc [all...] |
H A D | mvneta.c | 635 int rx_desc = rxq->next_desc_to_proc; local 637 rxq->next_desc_to_proc = MVNETA_QUEUE_NEXT_DESC(rxq, rx_desc); 639 return rxq->descs + rx_desc; 1141 static void mvneta_rx_desc_fill(struct mvneta_rx_desc *rx_desc, argument 1144 rx_desc->buf_cookie = cookie; 1145 rx_desc->buf_phys_addr = phys_addr; 1229 struct mvneta_rx_desc *rx_desc) 1231 u32 status = rx_desc->status; 1236 status, rx_desc->data_size); 1243 status, rx_desc 1228 mvneta_rx_error(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc) argument 1347 mvneta_rx_refill(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc) argument 1420 struct mvneta_rx_desc *rx_desc = rxq->descs + i; local 1452 struct mvneta_rx_desc *rx_desc = mvneta_rxq_next_desc_get(rxq); local [all...] |
/drivers/net/ethernet/intel/igbvf/ |
H A D | igbvf.h | 144 union e1000_adv_rx_desc rx_desc; member in union:igbvf_desc 300 (&((((R).desc))[i].rx_desc))
|
/drivers/infiniband/ulp/isert/ |
H A D | ib_isert.c | 158 struct iser_rx_desc *rx_desc; local 168 rx_desc = isert_conn->conn_rx_descs; 170 for (i = 0; i < ISERT_QP_MAX_RECV_DTOS; i++, rx_desc++) { 171 dma_addr = ib_dma_map_single(ib_dev, (void *)rx_desc, 176 rx_desc->dma_addr = dma_addr; 178 rx_sg = &rx_desc->rx_sg; 179 rx_sg->addr = rx_desc->dma_addr; 188 rx_desc = isert_conn->conn_rx_descs; 189 for (j = 0; j < i; j++, rx_desc++) { 190 ib_dma_unmap_single(ib_dev, rx_desc 203 struct iser_rx_desc *rx_desc; local 875 struct iser_rx_desc *rx_desc; local 1112 isert_rx_login_req(struct iser_rx_desc *rx_desc, int rx_buflen, struct isert_conn *isert_conn) argument 1180 isert_handle_scsi_cmd(struct isert_conn *isert_conn, struct isert_cmd *isert_cmd, struct iscsi_cmd *cmd, struct iser_rx_desc *rx_desc, unsigned char *buf) argument 1238 isert_handle_iscsi_dataout(struct isert_conn *isert_conn, struct iser_rx_desc *rx_desc, unsigned char *buf) argument 1291 isert_handle_nop_out(struct isert_conn *isert_conn, struct isert_cmd *isert_cmd, struct iscsi_cmd *cmd, struct iser_rx_desc *rx_desc, unsigned char *buf) argument 1310 isert_handle_text_cmd(struct isert_conn *isert_conn, struct isert_cmd *isert_cmd, struct iscsi_cmd *cmd, struct iser_rx_desc *rx_desc, struct iscsi_text *hdr) argument 1337 isert_rx_opcode(struct isert_conn *isert_conn, struct iser_rx_desc *rx_desc, uint32_t read_stag, uint64_t read_va, uint32_t write_stag, uint64_t write_va) argument 1422 isert_rx_do_work(struct iser_rx_desc *rx_desc, struct isert_conn *isert_conn) argument 2110 struct iser_rx_desc *rx_desc; local [all...] |
/drivers/atm/ |
H A D | horizon.c | 743 rx_ch_desc * rx_desc = &memmap->rx_descs[channel]; local 748 channel_type = rd_mem (dev, &rx_desc->wr_buf_type) & BUFFER_PTR_MASK; 776 wr_mem (dev, &rx_desc->wr_buf_type, 779 wr_mem (dev, &rx_desc->rd_buf_type, buf_ptr); 817 rx_ch_desc * rx_desc = &memmap->rx_descs[vc]; local 822 value = rd_mem (dev, &rx_desc->wr_buf_type) & BUFFER_PTR_MASK; 836 wr_mem (dev, &rx_desc->wr_buf_type, RX_CHANNEL_DISABLED); 838 if ((rd_mem (dev, &rx_desc->wr_buf_type) & BUFFER_PTR_MASK) == RX_CHANNEL_DISABLED) 882 r1 = rd_mem (dev, &rx_desc->rd_buf_type); 900 r2 = rd_mem (dev, &rx_desc 1859 cell_buf * rx_desc; local 1929 rx_ch_desc * rx_desc = &memmap->rx_descs[chan]; local [all...] |
/drivers/net/wireless/ath/ath10k/ |
H A D | htt_rx.c | 130 struct htt_rx_desc *rx_desc; local 148 /* Clear rx_desc attention word before posting to Rx ring */ 149 rx_desc = (struct htt_rx_desc *)skb->data; 150 rx_desc->attention.flags = __cpu_to_le32(0); 318 struct htt_rx_desc *rx_desc; local 339 rx_desc = (struct htt_rx_desc *)msdu->data; 354 if (!(__le32_to_cpu(rx_desc->attention.flags) 364 *attention |= __le32_to_cpu(rx_desc->attention.flags) & 384 rx_desc->fw_desc.info0 = **fw_desc; 408 rx_desc [all...] |
/drivers/net/ethernet/intel/e1000/ |
H A D | e1000_main.c | 3441 struct e1000_rx_desc *rx_desc = E1000_RX_DESC(*rx_ring, i); local 3444 struct my_u *u = (struct my_u *)rx_desc; 4130 struct e1000_rx_desc *rx_desc, *next_rxd; local 4139 rx_desc = E1000_RX_DESC(*rx_ring, i); 4142 while (rx_desc->status & E1000_RXD_STAT_DD) { 4151 status = rx_desc->status; 4165 length = le16_to_cpu(rx_desc->length); 4169 (rx_desc->errors & E1000_RXD_ERR_FRAME_ERR_MASK))) { 4173 rx_desc->errors, 4244 status | rx_desc 4350 struct e1000_rx_desc *rx_desc, *next_rxd; local 4495 struct e1000_rx_desc *rx_desc; local 4559 struct e1000_rx_desc *rx_desc; local [all...] |
/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | adapter.h | 100 struct rx_desc; 115 struct rx_desc *desc; /* address of HW Rx descriptor ring */
|
/drivers/net/ethernet/intel/ixgb/ |
H A D | ixgb_main.c | 1931 * @rx_desc: receive descriptor 1937 struct ixgb_rx_desc *rx_desc, 1943 if ((rx_desc->status & IXGB_RX_DESC_STATUS_IXSM) || 1944 (!(rx_desc->status & IXGB_RX_DESC_STATUS_TCPCS))) { 1951 if (rx_desc->errors & IXGB_RX_DESC_ERRORS_TCPE) { 1998 struct ixgb_rx_desc *rx_desc, *next_rxd; local 2006 rx_desc = IXGB_RX_DESC(*rx_ring, i); 2009 while (rx_desc->status & IXGB_RX_DESC_STATUS_DD) { 2018 status = rx_desc->status; 2046 length = le16_to_cpu(rx_desc 1936 ixgb_rx_checksum(struct ixgb_adapter *adapter, struct ixgb_rx_desc *rx_desc, struct sk_buff *skb) argument 2117 struct ixgb_rx_desc *rx_desc; local [all...] |
/drivers/scsi/mvsas/ |
H A D | mv_sas.c | 985 static void mvs_slot_free(struct mvs_info *mvi, u32 rx_desc) argument 987 u32 slot_idx = rx_desc & RXQ_SLOT_MASK; 1814 int mvs_slot_complete(struct mvs_info *mvi, u32 rx_desc, u32 flags) argument 1816 u32 slot_idx = rx_desc & RXQ_SLOT_MASK; 1868 if (unlikely((rx_desc & RXQ_ERR) 1871 mv_dprintk("port %d slot %d rx_desc %X has error info" 1873 rx_desc, get_unaligned_le64(slot->response)); 1882 if (rx_desc & RXQ_GOOD) { 1887 else if (rx_desc & RXQ_RSP) { 2165 u32 rx_prod_idx, rx_desc; local [all...] |