Searched refs:qhp (Results 1 - 10 of 10) sorted by relevance

/drivers/infiniband/hw/cxgb3/
H A Diwch_qp.c247 static int build_rdma_recv(struct iwch_qp *qhp, union t3_wr *wqe, argument
254 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr,
280 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr,
281 qhp->wq.rq_size_log2)].wr_id = wr->wr_id;
282 qhp->wq.rq[Q_PTR2IDX(qhp->wq.rq_wptr,
283 qhp->wq.rq_size_log2)].pbl_addr = 0;
287 static int build_zero_stag_recv(struct iwch_qp *qhp, union t3_wr *wqe, argument
301 pbl_addr = cxio_hal_pblpool_alloc(&qhp
357 struct iwch_qp *qhp; local
469 struct iwch_qp *qhp; local
534 struct iwch_qp *qhp; local
774 iwch_post_terminate(struct iwch_qp *qhp, struct respQ_msg_t *rsp_msg) argument
806 __flush_qp(struct iwch_qp *qhp, struct iwch_cq *rchp, struct iwch_cq *schp) argument
853 flush_qp(struct iwch_qp *qhp) argument
882 iwch_rqes_posted(struct iwch_qp *qhp) argument
895 rdma_init(struct iwch_dev *rhp, struct iwch_qp *qhp, enum iwch_qp_attr_mask mask, struct iwch_qp_attributes *attrs) argument
947 iwch_modify_qp(struct iwch_dev *rhp, struct iwch_qp *qhp, enum iwch_qp_attr_mask mask, struct iwch_qp_attributes *attrs, int internal) argument
[all...]
H A Diwch_ev.c48 struct iwch_qp *qhp; local
52 qhp = get_qhp(rnicp, CQE_QPID(rsp_msg->cqe));
54 if (!qhp) {
62 if ((qhp->attr.state == IWCH_QP_STATE_ERROR) ||
63 (qhp->attr.state == IWCH_QP_STATE_TERMINATE)) {
66 qhp->attr.state, qhp->wq.qpid, CQE_STATUS(rsp_msg->cqe));
77 atomic_inc(&qhp->refcnt);
80 if (qhp->attr.state == IWCH_QP_STATE_RTS) {
82 iwch_modify_qp(qhp
111 struct iwch_qp *qhp; local
[all...]
H A Diwch_cq.c47 struct iwch_qp *qhp = NULL; local
60 qhp = get_qhp(rhp, CQE_QPID(*rd_cqe));
61 if (!qhp)
64 spin_lock(&qhp->lock);
65 wq = &(qhp->wq);
82 wc->qp = &qhp->ibqp;
191 spin_unlock(&qhp->lock);
H A Diwch_provider.c861 struct iwch_qp *qhp; local
865 qhp = to_iwch_qp(ib_qp);
866 rhp = qhp->rhp;
869 iwch_modify_qp(rhp, qhp, IWCH_QP_ATTR_NEXT_STATE, &attrs, 0);
870 wait_event(qhp->wait, !qhp->ep);
872 remove_handle(rhp, &rhp->qpidr, qhp->wq.qpid);
874 atomic_dec(&qhp->refcnt);
875 wait_event(qhp->wait, !atomic_read(&qhp
893 struct iwch_qp *qhp; local
1050 struct iwch_qp *qhp; local
[all...]
H A Diwch.c68 struct iwch_qp *qhp = p; local
70 cxio_disable_wq_db(&qhp->wq);
76 struct iwch_qp *qhp = p; local
79 ring_doorbell(qhp->rhp->rdev.ctrl_qp.doorbell, qhp->wq.qpid);
80 cxio_enable_wq_db(&qhp->wq);
H A Diwch_provider.h172 static inline int qp_quiesced(struct iwch_qp *qhp) argument
174 return qhp->flags & QP_QUIESCED;
259 struct iwch_qp *qhp,
326 u16 iwch_rqes_posted(struct iwch_qp *qhp);
335 int iwch_post_terminate(struct iwch_qp *qhp, struct respQ_msg_t *rsp_msg);
339 void stop_read_rep_timer(struct iwch_qp *qhp);
/drivers/infiniband/hw/cxgb4/
H A Dqp.c86 static void set_state(struct c4iw_qp *qhp, enum c4iw_qp_state state) argument
89 spin_lock_irqsave(&qhp->lock, flag);
90 qhp->attr.state = state;
91 spin_unlock_irqrestore(&qhp->lock, flag);
573 static int build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, argument
578 ret = build_isgl((__be64 *)qhp->wq.rq.queue,
579 (__be64 *)&qhp->wq.rq.queue[qhp->wq.rq.size],
688 static int ring_kernel_sq_db(struct c4iw_qp *qhp, u16 inc) argument
692 spin_lock_irqsave(&qhp
706 ring_kernel_rq_db(struct c4iw_qp *qhp, u16 inc) argument
731 struct c4iw_qp *qhp; local
857 struct c4iw_qp *qhp; local
1068 post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe, gfp_t gfp) argument
1104 __flush_qp(struct c4iw_qp *qhp, struct c4iw_cq *rchp, struct c4iw_cq *schp) argument
1163 flush_qp(struct c4iw_qp *qhp) argument
1189 rdma_fini(struct c4iw_dev *rhp, struct c4iw_qp *qhp, struct c4iw_ep *ep) argument
1251 rdma_init(struct c4iw_dev *rhp, struct c4iw_qp *qhp) argument
1334 c4iw_modify_qp(struct c4iw_dev *rhp, struct c4iw_qp *qhp, enum c4iw_qp_attr_mask mask, struct c4iw_qp_attributes *attrs, int internal) argument
1574 struct c4iw_qp *qhp; local
1612 struct c4iw_qp *qhp; local
1817 struct c4iw_qp *qhp; local
1874 struct c4iw_qp *qhp = to_c4iw_qp(ibqp); local
[all...]
H A Dev.c88 struct c4iw_qp *qhp,
98 if (qhp->attr.state == C4IW_QP_STATE_RTS) {
100 c4iw_modify_qp(qhp->rhp, qhp, C4IW_QP_ATTR_NEXT_STATE,
109 event.element.qp = &qhp->ibqp;
110 if (qhp->ibqp.event_handler)
111 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context);
121 struct c4iw_qp *qhp; local
125 qhp
87 post_qp_event(struct c4iw_dev *dev, struct c4iw_cq *chp, struct c4iw_qp *qhp, struct t4_cqe *err_cqe, enum ib_event_type ib_event) argument
[all...]
H A Dcq.c231 int c4iw_flush_sq(struct c4iw_qp *qhp) argument
234 struct t4_wq *wq = &qhp->wq;
235 struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq);
337 struct c4iw_qp *qhp; local
350 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe));
355 if (qhp == NULL)
378 if (!qhp->wq.sq.oldest_read->signaled) {
379 advance_oldest_read(&qhp->wq);
387 create_read_req_cqe(&qhp->wq, hw_cqe, &read_cqe);
389 advance_oldest_read(&qhp
679 struct c4iw_qp *qhp = NULL; local
[all...]
H A Diw_cxgb4.h574 struct c4iw_qp *qhp,
1015 int c4iw_flush_sq(struct c4iw_qp *qhp);
1017 u16 c4iw_rqes_posted(struct c4iw_qp *qhp);
1018 int c4iw_post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe);

Completed in 74 milliseconds