Searched defs:recv_cq (Results 1 - 7 of 7) sorted by relevance

/drivers/infiniband/hw/amso1100/
H A Dc2_qp.c421 struct c2_cq *recv_cq = to_c2cq(qp_attrs->recv_cq); local
461 wr.rq_cq_handle = recv_cq->adapter_handle;
572 static inline void c2_lock_cqs(struct c2_cq *send_cq, struct c2_cq *recv_cq) argument
574 if (send_cq == recv_cq)
576 else if (send_cq > recv_cq) {
578 spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING);
580 spin_lock_irq(&recv_cq->lock);
585 static inline void c2_unlock_cqs(struct c2_cq *send_cq, struct c2_cq *recv_cq) argument
587 if (send_cq == recv_cq)
601 struct c2_cq *recv_cq; local
[all...]
/drivers/infiniband/ulp/srp/
H A Dib_srp.h136 struct ib_cq *recv_cq; member in struct:srp_target_port
/drivers/infiniband/hw/ehca/
H A Dehca_classes.h208 struct ehca_cq *recv_cq; member in struct:ehca_qp
/drivers/infiniband/hw/mlx4/
H A Dqp.c643 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
644 __acquires(&send_cq->lock) __acquires(&recv_cq->lock)
646 if (send_cq == recv_cq) {
648 __acquire(&recv_cq->lock);
649 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
651 spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING);
653 spin_lock_irq(&recv_cq->lock);
658 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
659 __releases(&send_cq->lock) __releases(&recv_cq->lock)
661 if (send_cq == recv_cq) {
691 get_cqs(struct mlx4_ib_qp *qp, struct mlx4_ib_cq **send_cq, struct mlx4_ib_cq **recv_cq) argument
713 struct mlx4_ib_cq *send_cq, *recv_cq; local
1013 struct mlx4_ib_cq *send_cq, *recv_cq; local
[all...]
/drivers/infiniband/hw/mthca/
H A Dmthca_qp.c756 qp_context->cqn_rcv = cpu_to_be32(to_mcq(ibqp->recv_cq)->cqn);
817 mthca_cq_clean(dev, to_mcq(qp->ibqp.recv_cq), qp->qpn,
819 if (qp->ibqp.send_cq != qp->ibqp.recv_cq)
1141 struct mthca_cq *recv_cq,
1269 struct mthca_cq *recv_cq,
1295 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
1310 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq)
1311 __acquires(&send_cq->lock) __acquires(&recv_cq->lock)
1313 if (send_cq == recv_cq) {
1315 __acquire(&recv_cq
1138 mthca_alloc_qp_common(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct mthca_qp *qp) argument
1266 mthca_alloc_qp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_qp_type type, enum ib_sig_type send_policy, struct ib_qp_cap *cap, struct mthca_qp *qp) argument
1340 mthca_alloc_sqp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct ib_qp_cap *cap, int qpn, int port, struct mthca_sqp *sqp) argument
1422 struct mthca_cq *recv_cq; local
[all...]
/drivers/infiniband/ulp/ipoib/
H A Dipoib.h301 struct ib_cq *recv_cq; member in struct:ipoib_dev_priv
/drivers/net/ethernet/ibm/ehea/
H A Dehea.h366 struct ehea_cq *recv_cq; member in struct:ehea_port_res

Completed in 453 milliseconds