Lines Matching defs:send_cq
717 qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn);
819 if (qp->ibqp.send_cq != qp->ibqp.recv_cq)
820 mthca_cq_clean(dev, to_mcq(qp->ibqp.send_cq), qp->qpn, NULL);
1140 struct mthca_cq *send_cq,
1268 struct mthca_cq *send_cq,
1295 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
1310 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq)
1311 __acquires(&send_cq->lock) __acquires(&recv_cq->lock)
1313 if (send_cq == recv_cq) {
1314 spin_lock_irq(&send_cq->lock);
1316 } else if (send_cq->cqn < recv_cq->cqn) {
1317 spin_lock_irq(&send_cq->lock);
1321 spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING);
1325 static void mthca_unlock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq)
1326 __releases(&send_cq->lock) __releases(&recv_cq->lock)
1328 if (send_cq == recv_cq) {
1330 spin_unlock_irq(&send_cq->lock);
1331 } else if (send_cq->cqn < recv_cq->cqn) {
1333 spin_unlock_irq(&send_cq->lock);
1335 spin_unlock(&send_cq->lock);
1342 struct mthca_cq *send_cq,
1378 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
1392 mthca_lock_cqs(send_cq, recv_cq);
1398 mthca_unlock_cqs(send_cq, recv_cq);
1421 struct mthca_cq *send_cq;
1424 send_cq = to_mcq(qp->ibqp.send_cq);
1431 mthca_lock_cqs(send_cq, recv_cq);
1439 mthca_unlock_cqs(send_cq, recv_cq);
1455 if (send_cq != recv_cq)
1456 mthca_cq_clean(dev, send_cq, qp->qpn, NULL);
1630 if (mthca_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) {
1971 if (mthca_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) {