/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_verbs.c | 170 priv->recv_cq = ib_create_cq(priv->ca, ipoib_ib_completion, NULL, dev, size, 0); 171 if (IS_ERR(priv->recv_cq)) { 183 if (ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP)) 187 init_attr.recv_cq = priv->recv_cq; 234 ib_destroy_cq(priv->recv_cq); 260 if (ib_destroy_cq(priv->recv_cq))
|
H A D | ipoib_ethtool.c | 70 ret = ib_modify_cq(priv->recv_cq, coal->rx_max_coalesced_frames,
|
H A D | ipoib_ib.c | 438 n = ib_poll_cq(priv->recv_cq, t, priv->ibwc); 459 if (unlikely(ib_req_notify_cq(priv->recv_cq, 784 n = ib_poll_cq(priv->recv_cq, IPOIB_NUM_WC, priv->ibwc); 899 ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP);
|
H A D | ipoib_cm.c | 253 .send_cq = priv->recv_cq, /* For drain WR */ 254 .recv_cq = priv->recv_cq, 1019 .send_cq = priv->recv_cq, 1020 .recv_cq = priv->recv_cq,
|
H A D | ipoib.h | 301 struct ib_cq *recv_cq; member in struct:ipoib_dev_priv
|
/drivers/infiniband/hw/amso1100/ |
H A D | c2_qp.c | 421 struct c2_cq *recv_cq = to_c2cq(qp_attrs->recv_cq); local 461 wr.rq_cq_handle = recv_cq->adapter_handle; 572 static inline void c2_lock_cqs(struct c2_cq *send_cq, struct c2_cq *recv_cq) argument 574 if (send_cq == recv_cq) 576 else if (send_cq > recv_cq) { 578 spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING); 580 spin_lock_irq(&recv_cq->lock); 585 static inline void c2_unlock_cqs(struct c2_cq *send_cq, struct c2_cq *recv_cq) argument 587 if (send_cq == recv_cq) 601 struct c2_cq *recv_cq; local [all...] |
/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 756 qp_context->cqn_rcv = cpu_to_be32(to_mcq(ibqp->recv_cq)->cqn); 817 mthca_cq_clean(dev, to_mcq(qp->ibqp.recv_cq), qp->qpn, 819 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) 1141 struct mthca_cq *recv_cq, 1269 struct mthca_cq *recv_cq, 1295 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, 1310 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) 1311 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) 1313 if (send_cq == recv_cq) { 1315 __acquire(&recv_cq 1138 mthca_alloc_qp_common(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct mthca_qp *qp) argument 1266 mthca_alloc_qp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_qp_type type, enum ib_sig_type send_policy, struct ib_qp_cap *cap, struct mthca_qp *qp) argument 1340 mthca_alloc_sqp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct ib_qp_cap *cap, int qpn, int port, struct mthca_sqp *sqp) argument 1422 struct mthca_cq *recv_cq; local [all...] |
H A D | mthca_dev.h | 546 struct mthca_cq *recv_cq, 554 struct mthca_cq *recv_cq,
|
/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 643 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) 644 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) 646 if (send_cq == recv_cq) { 648 __acquire(&recv_cq->lock); 649 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { 651 spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING); 653 spin_lock_irq(&recv_cq->lock); 658 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) 659 __releases(&send_cq->lock) __releases(&recv_cq->lock) 661 if (send_cq == recv_cq) { 691 get_cqs(struct mlx4_ib_qp *qp, struct mlx4_ib_cq **send_cq, struct mlx4_ib_cq **recv_cq) argument 713 struct mlx4_ib_cq *send_cq, *recv_cq; local 1013 struct mlx4_ib_cq *send_cq, *recv_cq; local [all...] |
/drivers/infiniband/ulp/srp/ |
H A D | ib_srp.h | 136 struct ib_cq *recv_cq; member in struct:srp_target_port
|
H A D | ib_srp.c | 231 target->recv_cq = ib_create_cq(target->srp_host->srp_dev->dev, 233 if (IS_ERR(target->recv_cq)) { 234 ret = PTR_ERR(target->recv_cq); 245 ib_req_notify_cq(target->recv_cq, IB_CQ_NEXT_COMP); 255 init_attr->recv_cq = target->recv_cq; 277 ib_destroy_cq(target->recv_cq); 290 ib_destroy_cq(target->recv_cq); 636 while (ib_poll_cq(target->recv_cq, 1, &wc) > 0)
|
/drivers/infiniband/hw/ehca/ |
H A D | ehca_qp.c | 412 list = &qp->recv_cq->rqp_err_list; 632 if (init_attr->recv_cq) 633 my_qp->recv_cq = 634 container_of(init_attr->recv_cq, struct ehca_cq, ib_cq); 689 if (my_qp->recv_cq) 690 parms.recv_cq_handle = my_qp->recv_cq->ipz_cq_handle; 832 my_qp->ib_qp.recv_cq = init_attr->recv_cq; 1197 spin_lock_irqsave(&my_qp->recv_cq->spinlock, flags); 1200 spin_unlock_irqrestore(&my_qp->recv_cq [all...] |
H A D | ehca_classes.h | 208 struct ehca_cq *recv_cq; member in struct:ehca_qp
|
/drivers/infiniband/core/ |
H A D | verbs.c | 429 qp->send_cq = qp->recv_cq = NULL; 446 qp->recv_cq = NULL; 449 qp->recv_cq = qp_init_attr->recv_cq; 450 atomic_inc(&qp_init_attr->recv_cq->usecnt); 885 rcq = qp->recv_cq;
|
H A D | uverbs_cmd.c | 1451 attr.recv_cq = rcq; 1483 qp->recv_cq = attr.recv_cq; 1491 if (attr.recv_cq) 1492 atomic_inc(&attr.recv_cq->usecnt);
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_qp.c | 406 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 429 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 659 init_attr->recv_cq = qp->ibqp.recv_cq;
|
H A D | ipath_ruc.c | 151 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 445 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
|
H A D | ipath_ud.c | 227 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 575 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
|
H A D | ipath_uc.c | 420 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
|
/drivers/infiniband/hw/qib/ |
H A D | qib_ud.c | 216 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 584 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
|
H A D | qib_qp.c | 520 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 543 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 888 init_attr->recv_cq = qp->ibqp.recv_cq;
|
H A D | qib_ruc.c | 122 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 567 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
|
H A D | qib_uc.c | 426 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
|
/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea.h | 366 struct ehea_cq *recv_cq; member in struct:ehea_port_res
|
H A D | ehea_main.c | 199 arr[i++].fwh = pr->recv_cq->fw_handle; 896 ehea_reset_cq_ep(pr->recv_cq); 898 ehea_reset_cq_n1(pr->recv_cq); 1486 pr->recv_cq = ehea_create_cq(adapter, pr_cfg->max_entries_rcq, 1489 if (!pr->recv_cq) { 1505 pr->recv_cq->attr.act_nr_of_cqes); 1530 init_attr->recv_cq_handle = pr->recv_cq->fw_handle; 1580 ehea_destroy_cq(pr->recv_cq); 1597 ehea_destroy_cq(pr->recv_cq);
|