/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_qp.c | 247 static int build_rdma_recv(struct iwch_qp *qhp, union t3_wr *wqe, argument 254 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr, 280 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr, 281 qhp->wq.rq_size_log2)].wr_id = wr->wr_id; 282 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr, 283 qhp->wq.rq_size_log2)].pbl_addr = 0; 287 static int build_zero_stag_recv(struct iwch_qp *qhp, union t3_wr *wqe, argument 301 pbl_addr = cxio_hal_pblpool_alloc(&qhp 357 struct iwch_qp *qhp; local 469 struct iwch_qp *qhp; local 534 struct iwch_qp *qhp; local 774 iwch_post_terminate(struct iwch_qp *qhp, struct respQ_msg_t *rsp_msg) argument 806 __flush_qp(struct iwch_qp *qhp, struct iwch_cq *rchp, struct iwch_cq *schp) argument 853 flush_qp(struct iwch_qp *qhp) argument 882 iwch_rqes_posted(struct iwch_qp *qhp) argument 895 rdma_init(struct iwch_dev *rhp, struct iwch_qp *qhp, enum iwch_qp_attr_mask mask, struct iwch_qp_attributes *attrs) argument 947 iwch_modify_qp(struct iwch_dev *rhp, struct iwch_qp *qhp, enum iwch_qp_attr_mask mask, struct iwch_qp_attributes *attrs, int internal) argument [all...] |
H A D | iwch_ev.c | 48 struct iwch_qp *qhp; local 52 qhp = get_qhp(rnicp, CQE_QPID(rsp_msg->cqe)); 54 if (!qhp) { 62 if ((qhp->attr.state == IWCH_QP_STATE_ERROR) || 63 (qhp->attr.state == IWCH_QP_STATE_TERMINATE)) { 66 qhp->attr.state, qhp->wq.qpid, CQE_STATUS(rsp_msg->cqe)); 77 atomic_inc(&qhp->refcnt); 80 if (qhp->attr.state == IWCH_QP_STATE_RTS) { 82 iwch_modify_qp(qhp 111 struct iwch_qp *qhp; local [all...] |
H A D | iwch_cq.c | 47 struct iwch_qp *qhp = NULL; local 60 qhp = get_qhp(rhp, CQE_QPID(*rd_cqe)); 61 if (!qhp) 64 spin_lock(&qhp->lock); 65 wq = &(qhp->wq); 82 wc->qp = &qhp->ibqp; 191 spin_unlock(&qhp->lock);
|
H A D | iwch_provider.c | 861 struct iwch_qp *qhp; local 865 qhp = to_iwch_qp(ib_qp); 866 rhp = qhp->rhp; 869 iwch_modify_qp(rhp, qhp, IWCH_QP_ATTR_NEXT_STATE, &attrs, 0); 870 wait_event(qhp->wait, !qhp->ep); 872 remove_handle(rhp, &rhp->qpidr, qhp->wq.qpid); 874 atomic_dec(&qhp->refcnt); 875 wait_event(qhp->wait, !atomic_read(&qhp 893 struct iwch_qp *qhp; local 1050 struct iwch_qp *qhp; local [all...] |
H A D | iwch.c | 68 struct iwch_qp *qhp = p; local 70 cxio_disable_wq_db(&qhp->wq); 76 struct iwch_qp *qhp = p; local 79 ring_doorbell(qhp->rhp->rdev.ctrl_qp.doorbell, qhp->wq.qpid); 80 cxio_enable_wq_db(&qhp->wq);
|
H A D | iwch_provider.h | 172 static inline int qp_quiesced(struct iwch_qp *qhp) argument 174 return qhp->flags & QP_QUIESCED; 259 struct iwch_qp *qhp, 326 u16 iwch_rqes_posted(struct iwch_qp *qhp); 335 int iwch_post_terminate(struct iwch_qp *qhp, struct respQ_msg_t *rsp_msg); 339 void stop_read_rep_timer(struct iwch_qp *qhp);
|
/drivers/infiniband/hw/cxgb4/ |
H A D | qp.c | 86 static void set_state(struct c4iw_qp *qhp, enum c4iw_qp_state state) argument 89 spin_lock_irqsave(&qhp->lock, flag); 90 qhp->attr.state = state; 91 spin_unlock_irqrestore(&qhp->lock, flag); 573 static int build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, argument 578 ret = build_isgl((__be64 *)qhp->wq.rq.queue, 579 (__be64 *)&qhp->wq.rq.queue[qhp->wq.rq.size], 688 static int ring_kernel_sq_db(struct c4iw_qp *qhp, u16 inc) argument 692 spin_lock_irqsave(&qhp 706 ring_kernel_rq_db(struct c4iw_qp *qhp, u16 inc) argument 731 struct c4iw_qp *qhp; local 857 struct c4iw_qp *qhp; local 1068 post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe, gfp_t gfp) argument 1104 __flush_qp(struct c4iw_qp *qhp, struct c4iw_cq *rchp, struct c4iw_cq *schp) argument 1163 flush_qp(struct c4iw_qp *qhp) argument 1189 rdma_fini(struct c4iw_dev *rhp, struct c4iw_qp *qhp, struct c4iw_ep *ep) argument 1251 rdma_init(struct c4iw_dev *rhp, struct c4iw_qp *qhp) argument 1334 c4iw_modify_qp(struct c4iw_dev *rhp, struct c4iw_qp *qhp, enum c4iw_qp_attr_mask mask, struct c4iw_qp_attributes *attrs, int internal) argument 1574 struct c4iw_qp *qhp; local 1612 struct c4iw_qp *qhp; local 1817 struct c4iw_qp *qhp; local 1874 struct c4iw_qp *qhp = to_c4iw_qp(ibqp); local [all...] |
H A D | ev.c | 88 struct c4iw_qp *qhp, 98 if (qhp->attr.state == C4IW_QP_STATE_RTS) { 100 c4iw_modify_qp(qhp->rhp, qhp, C4IW_QP_ATTR_NEXT_STATE, 109 event.element.qp = &qhp->ibqp; 110 if (qhp->ibqp.event_handler) 111 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); 121 struct c4iw_qp *qhp; local 125 qhp 87 post_qp_event(struct c4iw_dev *dev, struct c4iw_cq *chp, struct c4iw_qp *qhp, struct t4_cqe *err_cqe, enum ib_event_type ib_event) argument [all...] |
H A D | cq.c | 231 int c4iw_flush_sq(struct c4iw_qp *qhp) argument 234 struct t4_wq *wq = &qhp->wq; 235 struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq); 337 struct c4iw_qp *qhp; local 350 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); 355 if (qhp == NULL) 378 if (!qhp->wq.sq.oldest_read->signaled) { 379 advance_oldest_read(&qhp->wq); 387 create_read_req_cqe(&qhp->wq, hw_cqe, &read_cqe); 389 advance_oldest_read(&qhp 679 struct c4iw_qp *qhp = NULL; local [all...] |
H A D | iw_cxgb4.h | 574 struct c4iw_qp *qhp, 1015 int c4iw_flush_sq(struct c4iw_qp *qhp); 1017 u16 c4iw_rqes_posted(struct c4iw_qp *qhp); 1018 int c4iw_post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe);
|