Searched refs:recv_cq (Results 1 - 25 of 35) sorted by relevance

12

/drivers/infiniband/ulp/ipoib/
H A Dipoib_verbs.c170 priv->recv_cq = ib_create_cq(priv->ca, ipoib_ib_completion, NULL, dev, size, 0);
171 if (IS_ERR(priv->recv_cq)) {
183 if (ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP))
187 init_attr.recv_cq = priv->recv_cq;
234 ib_destroy_cq(priv->recv_cq);
260 if (ib_destroy_cq(priv->recv_cq))
H A Dipoib_ethtool.c70 ret = ib_modify_cq(priv->recv_cq, coal->rx_max_coalesced_frames,
H A Dipoib_ib.c438 n = ib_poll_cq(priv->recv_cq, t, priv->ibwc);
459 if (unlikely(ib_req_notify_cq(priv->recv_cq,
784 n = ib_poll_cq(priv->recv_cq, IPOIB_NUM_WC, priv->ibwc);
899 ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP);
H A Dipoib_cm.c253 .send_cq = priv->recv_cq, /* For drain WR */
254 .recv_cq = priv->recv_cq,
1019 .send_cq = priv->recv_cq,
1020 .recv_cq = priv->recv_cq,
H A Dipoib.h301 struct ib_cq *recv_cq; member in struct:ipoib_dev_priv
/drivers/infiniband/hw/amso1100/
H A Dc2_qp.c421 struct c2_cq *recv_cq = to_c2cq(qp_attrs->recv_cq); local
461 wr.rq_cq_handle = recv_cq->adapter_handle;
572 static inline void c2_lock_cqs(struct c2_cq *send_cq, struct c2_cq *recv_cq) argument
574 if (send_cq == recv_cq)
576 else if (send_cq > recv_cq) {
578 spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING);
580 spin_lock_irq(&recv_cq->lock);
585 static inline void c2_unlock_cqs(struct c2_cq *send_cq, struct c2_cq *recv_cq) argument
587 if (send_cq == recv_cq)
601 struct c2_cq *recv_cq; local
[all...]
/drivers/infiniband/hw/mthca/
H A Dmthca_qp.c756 qp_context->cqn_rcv = cpu_to_be32(to_mcq(ibqp->recv_cq)->cqn);
817 mthca_cq_clean(dev, to_mcq(qp->ibqp.recv_cq), qp->qpn,
819 if (qp->ibqp.send_cq != qp->ibqp.recv_cq)
1141 struct mthca_cq *recv_cq,
1269 struct mthca_cq *recv_cq,
1295 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
1310 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq)
1311 __acquires(&send_cq->lock) __acquires(&recv_cq->lock)
1313 if (send_cq == recv_cq) {
1315 __acquire(&recv_cq
1138 mthca_alloc_qp_common(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct mthca_qp *qp) argument
1266 mthca_alloc_qp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_qp_type type, enum ib_sig_type send_policy, struct ib_qp_cap *cap, struct mthca_qp *qp) argument
1340 mthca_alloc_sqp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct ib_qp_cap *cap, int qpn, int port, struct mthca_sqp *sqp) argument
1422 struct mthca_cq *recv_cq; local
[all...]
H A Dmthca_dev.h546 struct mthca_cq *recv_cq,
554 struct mthca_cq *recv_cq,
/drivers/infiniband/hw/mlx4/
H A Dqp.c643 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
644 __acquires(&send_cq->lock) __acquires(&recv_cq->lock)
646 if (send_cq == recv_cq) {
648 __acquire(&recv_cq->lock);
649 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
651 spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING);
653 spin_lock_irq(&recv_cq->lock);
658 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq)
659 __releases(&send_cq->lock) __releases(&recv_cq->lock)
661 if (send_cq == recv_cq) {
691 get_cqs(struct mlx4_ib_qp *qp, struct mlx4_ib_cq **send_cq, struct mlx4_ib_cq **recv_cq) argument
713 struct mlx4_ib_cq *send_cq, *recv_cq; local
1013 struct mlx4_ib_cq *send_cq, *recv_cq; local
[all...]
/drivers/infiniband/ulp/srp/
H A Dib_srp.h136 struct ib_cq *recv_cq; member in struct:srp_target_port
H A Dib_srp.c231 target->recv_cq = ib_create_cq(target->srp_host->srp_dev->dev,
233 if (IS_ERR(target->recv_cq)) {
234 ret = PTR_ERR(target->recv_cq);
245 ib_req_notify_cq(target->recv_cq, IB_CQ_NEXT_COMP);
255 init_attr->recv_cq = target->recv_cq;
277 ib_destroy_cq(target->recv_cq);
290 ib_destroy_cq(target->recv_cq);
636 while (ib_poll_cq(target->recv_cq, 1, &wc) > 0)
/drivers/infiniband/hw/ehca/
H A Dehca_qp.c412 list = &qp->recv_cq->rqp_err_list;
632 if (init_attr->recv_cq)
633 my_qp->recv_cq =
634 container_of(init_attr->recv_cq, struct ehca_cq, ib_cq);
689 if (my_qp->recv_cq)
690 parms.recv_cq_handle = my_qp->recv_cq->ipz_cq_handle;
832 my_qp->ib_qp.recv_cq = init_attr->recv_cq;
1197 spin_lock_irqsave(&my_qp->recv_cq->spinlock, flags);
1200 spin_unlock_irqrestore(&my_qp->recv_cq
[all...]
H A Dehca_classes.h208 struct ehca_cq *recv_cq; member in struct:ehca_qp
/drivers/infiniband/core/
H A Dverbs.c429 qp->send_cq = qp->recv_cq = NULL;
446 qp->recv_cq = NULL;
449 qp->recv_cq = qp_init_attr->recv_cq;
450 atomic_inc(&qp_init_attr->recv_cq->usecnt);
885 rcq = qp->recv_cq;
H A Duverbs_cmd.c1451 attr.recv_cq = rcq;
1483 qp->recv_cq = attr.recv_cq;
1491 if (attr.recv_cq)
1492 atomic_inc(&attr.recv_cq->usecnt);
/drivers/infiniband/hw/ipath/
H A Dipath_qp.c406 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1);
429 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1);
659 init_attr->recv_cq = qp->ibqp.recv_cq;
H A Dipath_ruc.c151 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1);
445 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
H A Dipath_ud.c227 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
575 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
H A Dipath_uc.c420 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
/drivers/infiniband/hw/qib/
H A Dqib_ud.c216 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
584 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
H A Dqib_qp.c520 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1);
543 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1);
888 init_attr->recv_cq = qp->ibqp.recv_cq;
H A Dqib_ruc.c122 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1);
567 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
H A Dqib_uc.c426 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc,
/drivers/net/ethernet/ibm/ehea/
H A Dehea.h366 struct ehea_cq *recv_cq; member in struct:ehea_port_res
H A Dehea_main.c199 arr[i++].fwh = pr->recv_cq->fw_handle;
896 ehea_reset_cq_ep(pr->recv_cq);
898 ehea_reset_cq_n1(pr->recv_cq);
1486 pr->recv_cq = ehea_create_cq(adapter, pr_cfg->max_entries_rcq,
1489 if (!pr->recv_cq) {
1505 pr->recv_cq->attr.act_nr_of_cqes);
1530 init_attr->recv_cq_handle = pr->recv_cq->fw_handle;
1580 ehea_destroy_cq(pr->recv_cq);
1597 ehea_destroy_cq(pr->recv_cq);

Completed in 1546 milliseconds

12