/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_qp.c | 42 static int build_rdma_send(union t3_wr *wqe, struct ib_send_wr *wr, argument 51 wqe->send.rdmaop = T3_SEND_WITH_SE; 53 wqe->send.rdmaop = T3_SEND; 54 wqe->send.rem_stag = 0; 58 wqe->send.rdmaop = T3_SEND_WITH_SE_INV; 60 wqe->send.rdmaop = T3_SEND_WITH_INV; 61 wqe->send.rem_stag = cpu_to_be32(wr->ex.invalidate_rkey); 68 wqe->send.reserved[0] = 0; 69 wqe->send.reserved[1] = 0; 70 wqe 87 build_rdma_write(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument 128 build_rdma_read(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument 149 build_fastreg(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt, int *wr_cnt, struct t3_wq *wq) argument 191 build_inv_stag(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument 247 build_rdma_recv(struct iwch_qp *qhp, union t3_wr *wqe, struct ib_recv_wr *wr) argument 287 build_zero_stag_recv(struct iwch_qp *qhp, union t3_wr *wqe, struct ib_recv_wr *wr) argument 359 union t3_wr *wqe; local 471 union t3_wr *wqe; local 535 union t3_wr *wqe; local 743 union t3_wr *wqe; local 775 union t3_wr *wqe; local 883 union t3_wr *wqe = qhp->wq.queue; local [all...] |
H A D | cxio_hal.c | 140 struct t3_modify_qp_wr *wqe; local 141 struct sk_buff *skb = alloc_skb(sizeof(*wqe), GFP_KERNEL); 146 wqe = (struct t3_modify_qp_wr *) skb_put(skb, sizeof(*wqe)); 147 memset(wqe, 0, sizeof(*wqe)); 148 build_fw_riwrh((struct fw_riwrh *) wqe, T3_WR_QP_MOD, 151 wqe->flags = cpu_to_be32(MODQP_WRITE_EC); 153 wqe->sge_cmd = cpu_to_be64(sge_cmd); 519 struct t3_modify_qp_wr *wqe; local 604 __be64 *wqe; local 838 struct t3_rdma_init_wr *wqe; local [all...] |
H A D | cxio_dbg.c | 111 void cxio_dump_wqe(union t3_wr *wqe) argument 113 __be64 *data = (__be64 *)wqe;
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_ruc.c | 123 int ipath_init_sge(struct ipath_qp *qp, struct ipath_rwqe *wqe, argument 130 for (i = j = 0; i < wqe->num_sge; i++) { 131 if (wqe->sg_list[i].length == 0) 135 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) 137 *lengthp += wqe->sg_list[i].length; 146 wc.wr_id = wqe->wr_id; 172 struct ipath_rwqe *wqe; local 205 wqe = get_rwqe_ptr(rq, tail); 211 } while (!ipath_init_sge(qp, wqe, &qp->r_len, &qp->r_sge)); 212 qp->r_wr_id = wqe 264 struct ipath_swqe *wqe; local 699 ipath_send_complete(struct ipath_qp *qp, struct ipath_swqe *wqe, enum ib_wc_status status) argument [all...] |
H A D | ipath_rc.c | 42 static u32 restart_sge(struct ipath_sge_state *ss, struct ipath_swqe *wqe, argument 47 len = ((psn - wqe->psn) & IPATH_PSN_MASK) * pmtu; 48 ss->sge = wqe->sg_list[0]; 49 ss->sg_list = wqe->sg_list + 1; 50 ss->num_sge = wqe->wr.num_sge; 52 return wqe->length - len; 58 * @wqe: the work queue to initialize the QP's SGE from 62 static void ipath_init_restart(struct ipath_qp *qp, struct ipath_swqe *wqe) argument 66 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, 218 struct ipath_swqe *wqe; local 727 struct ipath_swqe *wqe = get_swqe_ptr(qp, n); local 807 struct ipath_swqe *wqe = get_swqe_ptr(qp, qp->s_last); local 863 struct ipath_swqe *wqe; local 1125 struct ipath_swqe *wqe; local [all...] |
H A D | ipath_uc.c | 49 struct ipath_swqe *wqe; local 70 wqe = get_swqe_ptr(qp, qp->s_last); 71 ipath_send_complete(qp, wqe, IB_WC_WR_FLUSH_ERR); 84 wqe = get_swqe_ptr(qp, qp->s_cur); 97 qp->s_psn = wqe->psn = qp->s_next_psn; 98 qp->s_sge.sge = wqe->sg_list[0]; 99 qp->s_sge.sg_list = wqe->sg_list + 1; 100 qp->s_sge.num_sge = wqe->wr.num_sge; 101 qp->s_len = len = wqe->length; 102 switch (wqe [all...] |
H A D | ipath_ud.c | 61 struct ipath_rwqe *wqe; local 133 wqe = get_rwqe_ptr(rq, tail); 135 if (!ipath_init_sge(qp, wqe, &rlen, &rsge)) { 149 wc.wr_id = wqe->wr_id; 246 struct ipath_swqe *wqe; local 269 wqe = get_swqe_ptr(qp, qp->s_last); 270 ipath_send_complete(qp, wqe, IB_WC_WR_FLUSH_ERR); 277 wqe = get_swqe_ptr(qp, qp->s_cur); 283 ah_attr = &to_iah(wqe->wr.wr.ud.ah)->attr; 306 ipath_ud_loopback(qp, wqe); [all...] |
H A D | ipath_srq.c | 57 struct ipath_rwqe *wqe; local 79 wqe = get_rwqe_ptr(&srq->rq, wq->head); 80 wqe->wr_id = wr->wr_id; 81 wqe->num_sge = wr->num_sge; 83 wqe->sg_list[i] = wr->sg_list[i]; 286 struct ipath_rwqe *wqe; local 289 wqe = get_rwqe_ptr(&srq->rq, tail); 290 p->wr_id = wqe->wr_id; 291 p->num_sge = wqe->num_sge; 292 for (i = 0; i < wqe [all...] |
/drivers/infiniband/hw/cxgb4/ |
H A D | qp.c | 359 static int build_rdma_send(struct t4_sq *sq, union t4_wr *wqe, argument 371 wqe->send.sendop_pkd = cpu_to_be32( 374 wqe->send.sendop_pkd = cpu_to_be32( 376 wqe->send.stag_inv = 0; 380 wqe->send.sendop_pkd = cpu_to_be32( 383 wqe->send.sendop_pkd = cpu_to_be32( 385 wqe->send.stag_inv = cpu_to_be32(wr->ex.invalidate_rkey); 395 ret = build_immd(sq, wqe->send.u.immd_src, wr, 399 size = sizeof wqe->send + sizeof(struct fw_ri_immd) + 404 wqe 424 build_rdma_write(struct t4_sq *sq, union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 467 build_rdma_read(union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 496 build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, struct ib_recv_wr *wr, u8 *len16) argument 511 build_fastreg(struct t4_sq *sq, union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 559 build_inv_stag(union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 589 union t4_wr *wqe; local 700 union t4_recv_wr *wqe; local 901 struct fw_ri_wr *wqe; local 1001 struct fw_ri_wr *wqe; local 1062 struct fw_ri_wr *wqe; local [all...] |
H A D | t4.h | 108 static inline void init_wr_hdr(union t4_wr *wqe, u16 wrid, argument 111 wqe->send.opcode = (u8)opcode; 112 wqe->send.flags = flags; 113 wqe->send.wrid = wrid; 114 wqe->send.r1[0] = 0; 115 wqe->send.r1[1] = 0; 116 wqe->send.r1[2] = 0; 117 wqe->send.len16 = len16;
|
/drivers/infiniband/hw/qib/ |
H A D | qib_rc.c | 43 static u32 restart_sge(struct qib_sge_state *ss, struct qib_swqe *wqe, argument 48 len = ((psn - wqe->psn) & QIB_PSN_MASK) * pmtu; 49 ss->sge = wqe->sg_list[0]; 50 ss->sg_list = wqe->sg_list + 1; 51 ss->num_sge = wqe->wr.num_sge; 52 ss->total_len = wqe->length; 54 return wqe->length - len; 236 struct qib_swqe *wqe; local 273 wqe = get_swqe_ptr(qp, qp->s_last); 274 qib_send_complete(qp, wqe, q 786 struct qib_swqe *wqe = get_swqe_ptr(qp, n); local 870 struct qib_swqe *wqe = get_swqe_ptr(qp, qp->s_acked); local 946 struct qib_swqe *wqe; local 972 struct qib_swqe *wqe; local 1054 do_rc_completion(struct qib_qp *qp, struct qib_swqe *wqe, struct qib_ibport *ibp) argument 1132 struct qib_swqe *wqe; local 1353 struct qib_swqe *wqe; local 1405 struct qib_swqe *wqe; local [all...] |
H A D | qib_ruc.c | 81 static int qib_init_sge(struct qib_qp *qp, struct qib_rwqe *wqe) argument 94 for (i = j = 0; i < wqe->num_sge; i++) { 95 if (wqe->sg_list[i].length == 0) 99 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) 101 qp->r_len += wqe->sg_list[i].length; 117 wc.wr_id = wqe->wr_id; 144 struct qib_rwqe *wqe; local 176 wqe = get_rwqe_ptr(rq, tail); 185 if (!wr_id_only && !qib_init_sge(qp, wqe)) { 189 qp->r_wr_id = wqe 359 struct qib_swqe *wqe; local 773 qib_send_complete(struct qib_qp *qp, struct qib_swqe *wqe, enum ib_wc_status status) argument [all...] |
H A D | qib_uc.c | 49 struct qib_swqe *wqe; local 70 wqe = get_swqe_ptr(qp, qp->s_last); 71 qib_send_complete(qp, wqe, IB_WC_WR_FLUSH_ERR); 84 wqe = get_swqe_ptr(qp, qp->s_cur); 97 wqe->psn = qp->s_next_psn; 99 qp->s_sge.sge = wqe->sg_list[0]; 100 qp->s_sge.sg_list = wqe->sg_list + 1; 101 qp->s_sge.num_sge = wqe->wr.num_sge; 102 qp->s_sge.total_len = wqe->length; 103 len = wqe [all...] |
H A D | qib_ud.c | 238 struct qib_swqe *wqe; local 261 wqe = get_swqe_ptr(qp, qp->s_last); 262 qib_send_complete(qp, wqe, IB_WC_WR_FLUSH_ERR); 269 wqe = get_swqe_ptr(qp, qp->s_cur); 277 ah_attr = &to_iah(wqe->wr.wr.ud.ah)->attr; 300 qib_ud_loopback(qp, wqe); 302 qib_send_complete(qp, wqe, IB_WC_SUCCESS); 308 extra_bytes = -wqe->length & 3; 309 nwords = (wqe->length + extra_bytes) >> 2; 313 qp->s_cur_size = wqe [all...] |
H A D | qib_srq.c | 57 struct qib_rwqe *wqe; local 79 wqe = get_rwqe_ptr(&srq->rq, wq->head); 80 wqe->wr_id = wr->wr_id; 81 wqe->num_sge = wr->num_sge; 83 wqe->sg_list[i] = wr->sg_list[i]; 282 struct qib_rwqe *wqe; local 285 wqe = get_rwqe_ptr(&srq->rq, tail); 286 p->wr_id = wqe->wr_id; 287 p->num_sge = wqe->num_sge; 288 for (i = 0; i < wqe [all...] |
/drivers/infiniband/hw/mthca/ |
H A D | mthca_srq.c | 90 static inline int *wqe_to_link(void *wqe) argument 92 return (int *) (wqe + offsetof(struct mthca_next_seg, imm)); 151 void *wqe; local 178 next = wqe = get_wqe(srq, i); 181 *wqe_to_link(wqe) = i + 1; 184 *wqe_to_link(wqe) = -1; 188 for (scatter = wqe + sizeof (struct mthca_next_seg); 189 (void *) scatter < wqe + (1 << srq->wqe_shift); 487 void *wqe; local 496 wqe 586 void *wqe; local [all...] |
H A D | mthca_qp.c | 1604 void *wqe; local 1640 wqe = get_send_wqe(qp, ind); 1642 qp->sq.last = wqe; 1644 ((struct mthca_next_seg *) wqe)->nda_op = 0; 1645 ((struct mthca_next_seg *) wqe)->ee_nds = 0; 1646 ((struct mthca_next_seg *) wqe)->flags = 1654 ((struct mthca_next_seg *) wqe)->imm = wr->ex.imm_data; 1656 wqe += sizeof (struct mthca_next_seg); 1664 set_raddr_seg(wqe, wr->wr.atomic.remote_addr, 1666 wqe 1821 void *wqe; local 1919 void *wqe; local 2163 void *wqe; local [all...] |
H A D | mthca_cq.c | 126 __be32 wqe; member in struct:mthca_cqe 140 __be32 wqe; member in struct:mthca_err_cqe 312 mthca_free_srq_wqe(srq, be32_to_cpu(cqe->wqe)); 388 be32_to_cpu(cqe->my_qpn), be32_to_cpu(cqe->wqe), 477 cqe->wqe = new_wqe; 511 be32_to_cpu(cqe->wqe)); 540 wqe_index = ((be32_to_cpu(cqe->wqe) - (*cur_qp)->send_wqe_offset) 546 u32 wqe = be32_to_cpu(cqe->wqe); local 548 wqe_index = wqe >> sr 552 s32 wqe; local [all...] |
/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 146 __be32 *wqe; local 161 wqe = buf + (i & ((1 << qp->sq.wqe_shift) - 1)); 162 *wqe = stamp; 168 wqe = buf + i; 169 *wqe = cpu_to_be32(0xffffffff); 178 void *wqe; local 181 ctrl = wqe = get_send_wqe(qp, n & (qp->sq.wqe_cnt - 1)); 185 struct mlx4_wqe_datagram_seg *dgram = wqe + sizeof *ctrl; 194 inl = wqe + s; 1351 void *wqe, unsigne 1350 build_mlx_header(struct mlx4_ib_sqp *sqp, struct ib_send_wr *wr, void *wqe, unsigned *mlx_seg_len) argument 1658 build_lso_seg(struct mlx4_wqe_lso_seg *wqe, struct ib_send_wr *wr, struct mlx4_ib_qp *qp, unsigned *lso_seg_len, __be32 *lso_hdr_sz, __be32 *blh) argument 1698 void *wqe; local [all...] |
/drivers/scsi/bfa/ |
H A D | bfa_svc.c | 643 struct bfa_fcxp_wqe_s *wqe; local 645 bfa_q_deq(&mod->wait_q, &wqe); 646 if (wqe) { 649 bfa_fcxp_init(fcxp, wqe->caller, wqe->bfa, wqe->nreq_sgles, 650 wqe->nrsp_sgles, wqe->req_sga_cbfn, 651 wqe->req_sglen_cbfn, wqe 1074 bfa_fcxp_alloc_wait(struct bfa_s *bfa, struct bfa_fcxp_wqe_s *wqe, bfa_fcxp_alloc_cbfn_t alloc_cbfn, void *alloc_cbarg, void *caller, int nreq_sgles, int nrsp_sgles, bfa_fcxp_get_sgaddr_t req_sga_cbfn, bfa_fcxp_get_sglen_t req_sglen_cbfn, bfa_fcxp_get_sgaddr_t rsp_sga_cbfn, bfa_fcxp_get_sglen_t rsp_sglen_cbfn) argument 1101 bfa_fcxp_walloc_cancel(struct bfa_s *bfa, struct bfa_fcxp_wqe_s *wqe) argument 4919 struct bfa_sgpg_wqe_s *wqe; local 4948 bfa_sgpg_wait(struct bfa_s *bfa, struct bfa_sgpg_wqe_s *wqe, int nsgpg) argument 4974 bfa_sgpg_wcancel(struct bfa_s *bfa, struct bfa_sgpg_wqe_s *wqe) argument 4987 bfa_sgpg_winit(struct bfa_sgpg_wqe_s *wqe, void (*cbfn) (void *cbarg), void *cbarg) argument [all...] |
H A D | bfa_svc.h | 77 void bfa_sgpg_winit(struct bfa_sgpg_wqe_s *wqe, 79 void bfa_sgpg_wait(struct bfa_s *bfa, struct bfa_sgpg_wqe_s *wqe, int nsgpgs); 80 void bfa_sgpg_wcancel(struct bfa_s *bfa, struct bfa_sgpg_wqe_s *wqe); 415 struct bfa_reqq_wait_s wqe; /* request wait queue element */ member in struct:bfa_lps_s 595 void bfa_fcxp_alloc_wait(struct bfa_s *bfa, struct bfa_fcxp_wqe_s *wqe, 604 struct bfa_fcxp_wqe_s *wqe);
|
/drivers/infiniband/hw/nes/ |
H A D | nes_verbs.c | 215 struct nes_hw_qp_wqe *wqe; local 235 wqe = &nesqp->hwqp.sq_vbase[head]; 236 /* nes_debug(NES_DBG_MR, "processing sq wqe at %p, head = %u.\n", wqe, head); */ 237 nes_fill_init_qp_wqe(wqe, nesqp, head); 239 set_wqe_64bit_value(wqe->wqe_words, NES_IWARP_SQ_WQE_COMP_SCRATCH_LOW_IDX, u64temp); 254 set_wqe_32bit_value(wqe->wqe_words, NES_IWARP_SQ_WQE_MISC_IDX, wqe_misc); 255 set_wqe_32bit_value(wqe->wqe_words, NES_IWARP_SQ_BIND_WQE_MR_IDX, ibmw_bind->mr->lkey); 256 set_wqe_32bit_value(wqe->wqe_words, NES_IWARP_SQ_BIND_WQE_MW_IDX, ibmw->rkey); 257 set_wqe_32bit_value(wqe 3230 fill_wqe_sg_send(struct nes_hw_qp_wqe *wqe, struct ib_send_wr *ib_wr, u32 uselkey) argument 3264 struct nes_hw_qp_wqe *wqe; local 3544 struct nes_hw_qp_wqe *wqe; local [all...] |
/drivers/scsi/bnx2fc/ |
H A D | 57xx_hsi_bnx2fc.h | 626 __le16 wqe; member in struct:fcoe_sqe 769 __le16 wqe; member in struct:fcoe_xfrqe 821 __le16 wqe; member in struct:fcoe_cqe 885 __le32 wqe; member in struct:fcoe_lcqe 900 __le16 wqe; member in struct:fcoe_pend_wq_cqe 979 __le16 wqe; member in struct:fcoe_unsolicited_cqe
|
/drivers/scsi/lpfc/ |
H A D | lpfc_sli.c | 80 * @wqe: The work Queue Entry to put on the Work queue. 82 * This routine will copy the contents of @wqe to the next available entry on 90 lpfc_sli4_wq_put(struct lpfc_queue *q, union lpfc_wqe *wqe) argument 99 temp_wqe = q->qe[q->host_index].wqe; 106 bf_set(wqe_wqec, &wqe->generic.wqe_com, 1); 108 bf_set(wqe_wqid, &wqe->generic.wqe_com, q->queue_id); 109 lpfc_sli_pcimem_bcopy(wqe, temp_wqe, q->entry_size); 158 * @wqe: The Mailbox Queue Entry to put on the Work queue. 387 * @wqe: The Receive Queue Entry to put on the Receive queue. 389 * This routine will copy the contents of @wqe t 7651 lpfc_sli4_iocb2wqe(struct lpfc_hba *phba, struct lpfc_iocbq *iocbq, union lpfc_wqe *wqe) argument 8078 union lpfc_wqe wqe; local 15981 union lpfc_wqe wqe; local [all...] |
/drivers/infiniband/hw/ehca/ |
H A D | ehca_reqs.c | 84 /* clear wqe header until sglist */ 102 ehca_dmp(wqe_p, 16*(6 + wqe_p->nr_of_data_seg), "recv wqe"); 174 /* clear wqe header until sglist */ 309 ehca_dmp( wqe_p, 16*(6 + wqe_p->nr_of_data_seg), "send wqe"); 678 * ignore this to avoid double cqes of bad wqe 803 struct ehca_wqe *wqe; local 821 wqe = (struct ehca_wqe *)ipz_qeit_calc(ipz_queue, offset); 822 if (!wqe) { 823 ehca_err(cq->device, "Invalid wqe offset=%#llx on " 828 wc->wr_id = replace_wr_id(wqe [all...] |