/drivers/infiniband/hw/cxgb4/ |
H A D | ev.c | 39 struct c4iw_qp *qhp, 47 if ((qhp->attr.state == C4IW_QP_STATE_ERROR) || 48 (qhp->attr.state == C4IW_QP_STATE_TERMINATE)) { 51 qhp->attr.state, qhp->wq.sq.qid, CQE_STATUS(err_cqe)); 61 if (qhp->attr.state == C4IW_QP_STATE_RTS) { 63 c4iw_modify_qp(qhp->rhp, qhp, C4IW_QP_ATTR_NEXT_STATE, 72 event.element.qp = &qhp->ibqp; 73 if (qhp 38 post_qp_event(struct c4iw_dev *dev, struct c4iw_cq *chp, struct c4iw_qp *qhp, struct t4_cqe *err_cqe, enum ib_event_type ib_event) argument 84 struct c4iw_qp *qhp; local [all...] |
H A D | qp.c | 41 static void set_state(struct c4iw_qp *qhp, enum c4iw_qp_state state) argument 44 spin_lock_irqsave(&qhp->lock, flag); 45 qhp->attr.state = state; 46 spin_unlock_irqrestore(&qhp->lock, flag); 496 static int build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, argument 501 ret = build_isgl((__be64 *)qhp->wq.rq.queue, 502 (__be64 *)&qhp->wq.rq.queue[qhp->wq.rq.size], 588 struct c4iw_qp *qhp; local 595 qhp 699 struct c4iw_qp *qhp; local 898 post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe, gfp_t gfp) argument 934 __flush_qp(struct c4iw_qp *qhp, struct c4iw_cq *rchp, struct c4iw_cq *schp) argument 972 flush_qp(struct c4iw_qp *qhp) argument 998 rdma_fini(struct c4iw_dev *rhp, struct c4iw_qp *qhp, struct c4iw_ep *ep) argument 1060 rdma_init(struct c4iw_dev *rhp, struct c4iw_qp *qhp) argument 1131 c4iw_modify_qp(struct c4iw_dev *rhp, struct c4iw_qp *qhp, enum c4iw_qp_attr_mask mask, struct c4iw_qp_attributes *attrs, int internal) argument 1358 struct c4iw_qp *qhp; local 1390 struct c4iw_qp *qhp; local 1584 struct c4iw_qp *qhp; local [all...] |
H A D | cq.c | 567 struct c4iw_qp *qhp = NULL; local 580 qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); 581 if (!qhp) 584 spin_lock(&qhp->lock); 585 wq = &(qhp->wq); 592 wc->qp = &qhp->ibqp; 705 spin_unlock(&qhp->lock);
|
H A D | iw_cxgb4.h | 450 struct c4iw_qp *qhp, 787 u16 c4iw_rqes_posted(struct c4iw_qp *qhp); 788 int c4iw_post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe);
|
/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_qp.c | 247 static int build_rdma_recv(struct iwch_qp *qhp, union t3_wr *wqe, argument 254 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr, 280 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr, 281 qhp->wq.rq_size_log2)].wr_id = wr->wr_id; 282 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr, 283 qhp->wq.rq_size_log2)].pbl_addr = 0; 287 static int build_zero_stag_recv(struct iwch_qp *qhp, union t3_wr *wqe, argument 301 pbl_addr = cxio_hal_pblpool_alloc(&qhp 357 struct iwch_qp *qhp; local 469 struct iwch_qp *qhp; local 534 struct iwch_qp *qhp; local 773 iwch_post_terminate(struct iwch_qp *qhp, struct respQ_msg_t *rsp_msg) argument 805 __flush_qp(struct iwch_qp *qhp, struct iwch_cq *rchp, struct iwch_cq *schp) argument 852 flush_qp(struct iwch_qp *qhp) argument 881 iwch_rqes_posted(struct iwch_qp *qhp) argument 893 rdma_init(struct iwch_dev *rhp, struct iwch_qp *qhp, enum iwch_qp_attr_mask mask, struct iwch_qp_attributes *attrs) argument 945 iwch_modify_qp(struct iwch_dev *rhp, struct iwch_qp *qhp, enum iwch_qp_attr_mask mask, struct iwch_qp_attributes *attrs, int internal) argument [all...] |
H A D | iwch_ev.c | 48 struct iwch_qp *qhp; local 52 qhp = get_qhp(rnicp, CQE_QPID(rsp_msg->cqe)); 54 if (!qhp) { 62 if ((qhp->attr.state == IWCH_QP_STATE_ERROR) || 63 (qhp->attr.state == IWCH_QP_STATE_TERMINATE)) { 66 qhp->attr.state, qhp->wq.qpid, CQE_STATUS(rsp_msg->cqe)); 77 atomic_inc(&qhp->refcnt); 80 if (qhp->attr.state == IWCH_QP_STATE_RTS) { 82 iwch_modify_qp(qhp 111 struct iwch_qp *qhp; local [all...] |
H A D | iwch_cq.c | 47 struct iwch_qp *qhp = NULL; local 60 qhp = get_qhp(rhp, CQE_QPID(*rd_cqe)); 61 if (!qhp) 64 spin_lock(&qhp->lock); 65 wq = &(qhp->wq); 82 wc->qp = &qhp->ibqp; 191 spin_unlock(&qhp->lock);
|
H A D | iwch_provider.c | 862 struct iwch_qp *qhp; local 866 qhp = to_iwch_qp(ib_qp); 867 rhp = qhp->rhp; 870 iwch_modify_qp(rhp, qhp, IWCH_QP_ATTR_NEXT_STATE, &attrs, 0); 871 wait_event(qhp->wait, !qhp->ep); 873 remove_handle(rhp, &rhp->qpidr, qhp->wq.qpid); 875 atomic_dec(&qhp->refcnt); 876 wait_event(qhp->wait, !atomic_read(&qhp 894 struct iwch_qp *qhp; local 1051 struct iwch_qp *qhp; local [all...] |
H A D | iwch.c | 68 struct iwch_qp *qhp = p; local 70 cxio_disable_wq_db(&qhp->wq); 76 struct iwch_qp *qhp = p; local 79 ring_doorbell(qhp->rhp->rdev.ctrl_qp.doorbell, qhp->wq.qpid); 80 cxio_enable_wq_db(&qhp->wq);
|
H A D | iwch_provider.h | 172 static inline int qp_quiesced(struct iwch_qp *qhp) argument 174 return qhp->flags & QP_QUIESCED; 259 struct iwch_qp *qhp, 326 u16 iwch_rqes_posted(struct iwch_qp *qhp); 335 int iwch_post_terminate(struct iwch_qp *qhp, struct respQ_msg_t *rsp_msg); 339 void stop_read_rep_timer(struct iwch_qp *qhp);
|