/drivers/infiniband/hw/amso1100/ |
H A D | c2_qp.c | 422 struct c2_cq *send_cq = to_c2cq(qp_attrs->send_cq); local 462 wr.sq_cq_handle = send_cq->adapter_handle; 574 static inline void c2_lock_cqs(struct c2_cq *send_cq, struct c2_cq *recv_cq) argument 576 if (send_cq == recv_cq) 577 spin_lock_irq(&send_cq->lock); 578 else if (send_cq > recv_cq) { 579 spin_lock_irq(&send_cq->lock); 583 spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING); 587 static inline void c2_unlock_cqs(struct c2_cq *send_cq, struc argument 602 struct c2_cq *send_cq; local [all...] |
/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_verbs.c | 176 priv->send_cq = ib_create_cq(priv->ca, ipoib_send_comp_handler, NULL, 178 if (IS_ERR(priv->send_cq)) { 186 init_attr.send_cq = priv->send_cq; 234 ib_destroy_cq(priv->send_cq); 260 if (ib_destroy_cq(priv->send_cq))
|
H A D | ipoib_cm.c | 254 .send_cq = priv->recv_cq, /* For drain WR */ 765 rc = ib_req_notify_cq(priv->send_cq, 770 ipoib_send_comp_handler(priv->send_cq, dev); 1026 .send_cq = priv->recv_cq,
|
H A D | ipoib_ib.c | 420 n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc); 598 if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP))
|
H A D | ipoib.h | 335 struct ib_cq *send_cq; member in struct:ipoib_dev_priv
|
/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 719 qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn); 821 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) 822 mthca_cq_clean(dev, to_mcq(qp->ibqp.send_cq), qp->qpn, NULL); 1143 struct mthca_cq *send_cq, 1271 struct mthca_cq *send_cq, 1298 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, 1313 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) 1314 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) 1316 if (send_cq == recv_cq) { 1317 spin_lock_irq(&send_cq 1141 mthca_alloc_qp_common(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct mthca_qp *qp) argument 1269 mthca_alloc_qp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_qp_type type, enum ib_sig_type send_policy, struct ib_qp_cap *cap, struct mthca_qp *qp) argument 1343 mthca_alloc_sqp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct ib_qp_cap *cap, int qpn, int port, struct mthca_sqp *sqp) argument 1424 struct mthca_cq *send_cq; local [all...] |
H A D | mthca_dev.h | 545 struct mthca_cq *send_cq, 553 struct mthca_cq *send_cq,
|
H A D | mthca_provider.c | 566 to_mcq(init_attr->send_cq), 601 to_mcq(init_attr->send_cq),
|
/drivers/infiniband/hw/mlx5/ |
H A D | qp.c | 910 scqe_sz = mlx5_ib_get_cqe_size(dev, init_attr->send_cq); 960 if (init_attr->send_cq) 961 in->ctx.cqn_send = cpu_to_be32(to_mcq(init_attr->send_cq)->mcq.cqn); 995 static void mlx5_ib_lock_cqs(struct mlx5_ib_cq *send_cq, struct mlx5_ib_cq *recv_cq) 996 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) 998 if (send_cq) { 1000 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { 1001 spin_lock_irq(&send_cq->lock); 1004 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { 1005 spin_lock_irq(&send_cq 1048 get_cqs(struct mlx5_ib_qp *qp, struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq) argument 1084 struct mlx5_ib_cq *send_cq, *recv_cq; local 1497 struct mlx5_ib_cq *send_cq, *recv_cq; local [all...] |
/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 890 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) 891 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) 893 if (send_cq == recv_cq) { 894 spin_lock_irq(&send_cq->lock); 896 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { 897 spin_lock_irq(&send_cq->lock); 901 spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING); 905 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) 906 __releases(&send_cq->lock) __releases(&recv_cq->lock) 908 if (send_cq 938 get_cqs(struct mlx4_ib_qp *qp, struct mlx4_ib_cq **send_cq, struct mlx4_ib_cq **recv_cq) argument 960 struct mlx4_ib_cq *send_cq, *recv_cq; local 1422 struct mlx4_ib_cq *send_cq, *recv_cq; local [all...] |
/drivers/infiniband/hw/ehca/ |
H A D | ehca_qp.c | 409 list = &qp->send_cq->sqp_err_list; 635 if (init_attr->send_cq) 636 my_qp->send_cq = 637 container_of(init_attr->send_cq, struct ehca_cq, ib_cq); 683 if (my_qp->send_cq) 684 parms.send_cq_handle = my_qp->send_cq->ipz_cq_handle; 829 my_qp->ib_qp.send_cq = init_attr->send_cq; 878 if (my_qp->send_cq) { 879 ret = ehca_cq_assign_qp(my_qp->send_cq, my_q [all...] |
H A D | ehca_classes.h | 207 struct ehca_cq *send_cq; member in struct:ehca_qp
|
H A D | ehca_main.c | 558 qp_init_attr.send_cq = ibcq;
|
/drivers/infiniband/ulp/srp/ |
H A D | ib_srp.h | 140 struct ib_cq *send_cq ____cacheline_aligned_in_smp;
|
H A D | ib_srp.c | 450 struct ib_cq *recv_cq, *send_cq; local 468 send_cq = ib_create_cq(dev->dev, srp_send_completion, NULL, target, 470 if (IS_ERR(send_cq)) { 471 ret = PTR_ERR(send_cq); 484 init_attr->send_cq = send_cq; 525 if (target->send_cq) 526 ib_destroy_cq(target->send_cq); 530 target->send_cq = send_cq; [all...] |
/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_main.c | 209 arr[i++].fwh = pr->send_cq->fw_handle; 816 struct ehea_cq *send_cq = pr->send_cq; local 825 cqe = ehea_poll_cq(send_cq); 827 ehea_inc_cq(send_cq); 867 cqe = ehea_poll_cq(send_cq); 870 ehea_update_feca(send_cq, cqe_counter); 905 ehea_reset_cq_ep(pr->send_cq); 907 ehea_reset_cq_n1(pr->send_cq); 910 cqe_skb = ehea_poll_cq(pr->send_cq); [all...] |
H A D | ehea.h | 364 struct ehea_cq *send_cq; member in struct:ehea_port_res
|
/drivers/infiniband/core/ |
H A D | verbs.c | 462 qp->send_cq = qp->recv_cq = NULL; 490 qp->send_cq = qp_init_attr->send_cq; 494 atomic_inc(&qp_init_attr->send_cq->usecnt); 989 scq = qp->send_cq;
|
H A D | uverbs_cmd.c | 1661 attr.send_cq = scq; 1693 qp->send_cq = attr.send_cq; 1701 atomic_inc(&attr.send_cq->usecnt);
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_qp.c | 658 init_attr->send_cq = qp->ibqp.send_cq;
|
H A D | ipath_ruc.c | 720 ipath_cq_enter(to_icq(qp->ibqp.send_cq), &wc,
|
/drivers/infiniband/hw/qib/ |
H A D | qib_qp.c | 891 init_attr->send_cq = qp->ibqp.send_cq;
|
/drivers/infiniband/hw/cxgb4/ |
H A D | qp.c | 1169 schp = to_c4iw_cq(qhp->ibqp.send_cq); 1629 schp = get_chp(rhp, ((struct c4iw_cq *)attrs->send_cq)->cq.cqid); 1680 qhp->attr.scq = ((struct c4iw_cq *) attrs->send_cq)->cq.cqid;
|
H A D | cq.c | 235 struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq);
|
/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_provider.c | 906 schp = get_chp(rhp, ((struct iwch_cq *) attrs->send_cq)->cq.cqid); 962 qhp->attr.scq = ((struct iwch_cq *) attrs->send_cq)->cq.cqid;
|