Searched refs:send_cq (Results 1 - 25 of 39) sorted by relevance

12

/drivers/infiniband/hw/amso1100/
H A Dc2_qp.c422 struct c2_cq *send_cq = to_c2cq(qp_attrs->send_cq); local
462 wr.sq_cq_handle = send_cq->adapter_handle;
574 static inline void c2_lock_cqs(struct c2_cq *send_cq, struct c2_cq *recv_cq) argument
576 if (send_cq == recv_cq)
577 spin_lock_irq(&send_cq->lock);
578 else if (send_cq > recv_cq) {
579 spin_lock_irq(&send_cq->lock);
583 spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING);
587 static inline void c2_unlock_cqs(struct c2_cq *send_cq, struc argument
602 struct c2_cq *send_cq; local
[all...]
/drivers/infiniband/ulp/ipoib/
H A Dipoib_verbs.c176 priv->send_cq = ib_create_cq(priv->ca, ipoib_send_comp_handler, NULL,
178 if (IS_ERR(priv->send_cq)) {
186 init_attr.send_cq = priv->send_cq;
234 ib_destroy_cq(priv->send_cq);
260 if (ib_destroy_cq(priv->send_cq))
H A Dipoib_cm.c254 .send_cq = priv->recv_cq, /* For drain WR */
765 rc = ib_req_notify_cq(priv->send_cq,
770 ipoib_send_comp_handler(priv->send_cq, dev);
1026 .send_cq = priv->recv_cq,
H A Dipoib_ib.c420 n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc);
598 if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP))
H A Dipoib.h335 struct ib_cq *send_cq; member in struct:ipoib_dev_priv
/drivers/infiniband/hw/mthca/
H A Dmthca_qp.c719 qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn);
821 if (qp->ibqp.send_cq != qp->ibqp.recv_cq)
822 mthca_cq_clean(dev, to_mcq(qp->ibqp.send_cq), qp->qpn, NULL);
1143 struct mthca_cq *send_cq,
1271 struct mthca_cq *send_cq,
1298 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
1313 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq)
1314 __acquires(&send_cq->lock) __acquires(&recv_cq->lock)
1316 if (send_cq == recv_cq) {
1317 spin_lock_irq(&send_cq
1141 mthca_alloc_qp_common(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct mthca_qp *qp) argument
1269 mthca_alloc_qp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_qp_type type, enum ib_sig_type send_policy, struct ib_qp_cap *cap, struct mthca_qp *qp) argument
1343 mthca_alloc_sqp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct ib_qp_cap *cap, int qpn, int port, struct mthca_sqp *sqp) argument
1424 struct mthca_cq *send_cq; local
[all...]
H A Dmthca_dev.h545 struct mthca_cq *send_cq,
553 struct mthca_cq *send_cq,
H A Dmthca_provider.c566 to_mcq(init_attr->send_cq),
601 to_mcq(init_attr->send_cq),
/drivers/infiniband/hw/mlx5/
H A Dqp.c910 scqe_sz = mlx5_ib_get_cqe_size(dev, init_attr->send_cq);
960 if (init_attr->send_cq)
961 in->ctx.cqn_send = cpu_to_be32(to_mcq(init_attr->send_cq)->mcq.cqn);
995 static void mlx5_ib_lock_cqs(struct mlx5_ib_cq *send_cq, struct mlx5_ib_cq *recv_cq)
996 __acquires(&send_cq->lock) __acquires(&recv_cq->lock)
998 if (send_cq) {
1000 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
1001 spin_lock_irq(&send_cq->lock);
1004 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) {
1005 spin_lock_irq(&send_cq
1048 get_cqs(struct mlx5_ib_qp *qp, struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq) argument
1084 struct mlx5_ib_cq *send_cq, *recv_cq; local
1497 struct mlx5_ib_cq *send_cq, *recv_cq; local
[all...]
/drivers/infiniband/hw/mlx4/
H A Dqp.c890 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
891 __acquires(&send_cq->lock) __acquires(&recv_cq->lock)
893 if (send_cq == recv_cq) {
894 spin_lock_irq(&send_cq->lock);
896 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
897 spin_lock_irq(&send_cq->lock);
901 spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING);
905 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
906 __releases(&send_cq->lock) __releases(&recv_cq->lock)
908 if (send_cq
938 get_cqs(struct mlx4_ib_qp *qp, struct mlx4_ib_cq **send_cq, struct mlx4_ib_cq **recv_cq) argument
960 struct mlx4_ib_cq *send_cq, *recv_cq; local
1422 struct mlx4_ib_cq *send_cq, *recv_cq; local
[all...]
/drivers/infiniband/hw/ehca/
H A Dehca_qp.c409 list = &qp->send_cq->sqp_err_list;
635 if (init_attr->send_cq)
636 my_qp->send_cq =
637 container_of(init_attr->send_cq, struct ehca_cq, ib_cq);
683 if (my_qp->send_cq)
684 parms.send_cq_handle = my_qp->send_cq->ipz_cq_handle;
829 my_qp->ib_qp.send_cq = init_attr->send_cq;
878 if (my_qp->send_cq) {
879 ret = ehca_cq_assign_qp(my_qp->send_cq, my_q
[all...]
H A Dehca_classes.h207 struct ehca_cq *send_cq; member in struct:ehca_qp
H A Dehca_main.c558 qp_init_attr.send_cq = ibcq;
/drivers/infiniband/ulp/srp/
H A Dib_srp.h140 struct ib_cq *send_cq ____cacheline_aligned_in_smp;
H A Dib_srp.c450 struct ib_cq *recv_cq, *send_cq; local
468 send_cq = ib_create_cq(dev->dev, srp_send_completion, NULL, target,
470 if (IS_ERR(send_cq)) {
471 ret = PTR_ERR(send_cq);
484 init_attr->send_cq = send_cq;
525 if (target->send_cq)
526 ib_destroy_cq(target->send_cq);
530 target->send_cq = send_cq;
[all...]
/drivers/net/ethernet/ibm/ehea/
H A Dehea_main.c209 arr[i++].fwh = pr->send_cq->fw_handle;
816 struct ehea_cq *send_cq = pr->send_cq; local
825 cqe = ehea_poll_cq(send_cq);
827 ehea_inc_cq(send_cq);
867 cqe = ehea_poll_cq(send_cq);
870 ehea_update_feca(send_cq, cqe_counter);
905 ehea_reset_cq_ep(pr->send_cq);
907 ehea_reset_cq_n1(pr->send_cq);
910 cqe_skb = ehea_poll_cq(pr->send_cq);
[all...]
H A Dehea.h364 struct ehea_cq *send_cq; member in struct:ehea_port_res
/drivers/infiniband/core/
H A Dverbs.c462 qp->send_cq = qp->recv_cq = NULL;
490 qp->send_cq = qp_init_attr->send_cq;
494 atomic_inc(&qp_init_attr->send_cq->usecnt);
989 scq = qp->send_cq;
H A Duverbs_cmd.c1661 attr.send_cq = scq;
1693 qp->send_cq = attr.send_cq;
1701 atomic_inc(&attr.send_cq->usecnt);
/drivers/infiniband/hw/ipath/
H A Dipath_qp.c658 init_attr->send_cq = qp->ibqp.send_cq;
H A Dipath_ruc.c720 ipath_cq_enter(to_icq(qp->ibqp.send_cq), &wc,
/drivers/infiniband/hw/qib/
H A Dqib_qp.c891 init_attr->send_cq = qp->ibqp.send_cq;
/drivers/infiniband/hw/cxgb4/
H A Dqp.c1169 schp = to_c4iw_cq(qhp->ibqp.send_cq);
1629 schp = get_chp(rhp, ((struct c4iw_cq *)attrs->send_cq)->cq.cqid);
1680 qhp->attr.scq = ((struct c4iw_cq *) attrs->send_cq)->cq.cqid;
H A Dcq.c235 struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq);
/drivers/infiniband/hw/cxgb3/
H A Diwch_provider.c906 schp = get_chp(rhp, ((struct iwch_cq *) attrs->send_cq)->cq.cqid);
962 qhp->attr.scq = ((struct iwch_cq *) attrs->send_cq)->cq.cqid;

Completed in 6264 milliseconds

12