/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_cq.c | 195 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) argument 203 chp = to_iwch_cq(ibcq);
|
H A D | iwch_provider.c | 191 chp->ibcq.cqe = 1 << chp->cq.size_log2; 207 iwch_destroy_cq(&chp->ibcq); 233 iwch_destroy_cq(&chp->ibcq); 241 return &chp->ibcq; 292 chp->ibcq.cqe = (1<<chp->cq.size_log2) - 1; 312 static int iwch_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) argument 321 chp = to_iwch_cq(ibcq);
|
H A D | iwch_provider.h | 102 struct ib_cq ibcq; member in struct:iwch_cq 112 static inline struct iwch_cq *to_iwch_cq(struct ib_cq *ibcq) argument 114 return container_of(ibcq, struct iwch_cq, ibcq); 334 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_cq.c | 63 if (head >= (unsigned) cq->ibcq.cqe) { 64 head = cq->ibcq.cqe; 70 if (cq->ibcq.event_handler) { 73 ev.device = cq->ibcq.device; 74 ev.element.cq = &cq->ibcq; 76 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); 115 to_idev(cq->ibcq.device)->n_wqe_errs++; 120 * @ibcq: the completion queue to poll 129 int ipath_poll_cq(struct ib_cq *ibcq, in argument 314 ipath_destroy_cq(struct ib_cq *ibcq) argument 342 ipath_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) argument 371 ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) argument [all...] |
H A D | ipath_verbs.h | 205 /* these are actually size ibcq.cqe + 1 */ 215 struct ib_cq ibcq; member in struct:ipath_cq 670 static inline struct ipath_cq *to_icq(struct ib_cq *ibcq) argument 672 return container_of(ibcq, struct ipath_cq, ibcq); 808 int ipath_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry); 814 int ipath_destroy_cq(struct ib_cq *ibcq); 816 int ipath_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags); 818 int ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
|
/drivers/infiniband/hw/qib/ |
H A D | qib_cq.c | 63 if (head >= (unsigned) cq->ibcq.cqe) { 64 head = cq->ibcq.cqe; 70 if (cq->ibcq.event_handler) { 73 ev.device = cq->ibcq.device; 74 ev.element.cq = &cq->ibcq; 76 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); 119 * @ibcq: the completion queue to poll 128 int qib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) argument 130 struct qib_cq *cq = to_icq(ibcq); 321 qib_destroy_cq(struct ib_cq *ibcq) argument 349 qib_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) argument 378 qib_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) argument [all...] |
H A D | qib_verbs.h | 258 /* these are actually size ibcq.cqe + 1 */ 268 struct ib_cq ibcq; member in struct:qib_cq 780 static inline struct qib_cq *to_icq(struct ib_cq *ibcq) argument 782 return container_of(ibcq, struct qib_cq, ibcq); 955 int qib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry); 961 int qib_destroy_cq(struct ib_cq *ibcq); 963 int qib_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags); 965 int qib_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
|
/drivers/infiniband/hw/amso1100/ |
H A D | c2_cq.c | 78 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); 201 int c2_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) argument 203 struct c2_dev *c2dev = to_c2dev(ibcq->device); 204 struct c2_cq *cq = to_c2cq(ibcq); 222 int c2_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) argument 229 cq = to_c2cq(ibcq); 297 cq->ibcq.cqe = entries - 1;
|
H A D | c2_provider.h | 92 struct ib_cq ibcq; member in struct:c2_cq 153 static inline struct c2_cq *to_c2cq(struct ib_cq *ibcq) argument 155 return container_of(ibcq, struct c2_cq, ibcq);
|
/drivers/infiniband/hw/cxgb4/ |
H A D | cq.c | 709 int c4iw_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) argument 716 chp = to_c4iw_cq(ibcq); 819 chp->ibcq.cqe = entries - 2; 864 return &chp->ibcq; 884 int c4iw_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) argument 890 chp = to_c4iw_cq(ibcq);
|
H A D | iw_cxgb4.h | 308 struct ib_cq ibcq; member in struct:c4iw_cq 317 static inline struct c4iw_cq *to_c4iw_cq(struct ib_cq *ibcq) argument 319 return container_of(ibcq, struct c4iw_cq, ibcq); 722 int c4iw_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc); 765 int c4iw_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags);
|
/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 43 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; local 44 ibcq->comp_handler(ibcq, ibcq->cq_context); 50 struct ib_cq *ibcq; local 58 ibcq = &to_mibcq(cq)->ibcq; 59 if (ibcq->event_handler) { 60 event.device = ibcq 345 mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) argument 735 mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) argument 761 mlx4_ib_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) argument [all...] |
H A D | mlx4_ib.h | 77 struct ib_cq ibcq; member in struct:mlx4_ib_cq 227 static inline struct mlx4_ib_cq *to_mcq(struct ib_cq *ibcq) argument 229 return container_of(ibcq, struct mlx4_ib_cq, ibcq); 294 int mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata); 299 int mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
|
/drivers/infiniband/hw/ehca/ |
H A D | ehca_main.c | 533 struct ib_cq *ibcq; local 543 ibcq = ib_create_cq(&shca->ib_device, NULL, NULL, (void *)(-1), 10, 0); 544 if (IS_ERR(ibcq)) { 546 return PTR_ERR(ibcq); 548 sport->ibcq_aqp1 = ibcq; 557 qp_init_attr.send_cq = ibcq; 558 qp_init_attr.recv_cq = ibcq;
|
/drivers/infiniband/hw/mthca/ |
H A D | mthca_provider.c | 703 return &cq->ibcq; 766 static int mthca_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) argument 768 struct mthca_dev *dev = to_mdev(ibcq->device); 769 struct mthca_cq *cq = to_mcq(ibcq); 780 if (entries == ibcq->cqe + 1) { 820 tcqe = cq->ibcq.cqe; 822 cq->ibcq.cqe = cq->resize_buf->cqe; 834 ibcq->cqe = entries - 1;
|
H A D | mthca_cq.c | 181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); 235 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); 259 event.element.cq = &cq->ibcq; 260 if (cq->ibcq.event_handler) 261 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); 295 cqe_sw(get_cqe(cq, prod_index & cq->ibcq.cqe)); 297 if (prod_index == cq->cons_index + cq->ibcq 662 mthca_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) argument 744 mthca_arbel_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) argument [all...] |
H A D | mthca_provider.h | 202 struct ib_cq ibcq; member in struct:mthca_cq 324 static inline struct mthca_cq *to_mcq(struct ib_cq *ibcq) argument 326 return container_of(ibcq, struct mthca_cq, ibcq);
|
/drivers/infiniband/hw/nes/ |
H A D | nes_verbs.h | 112 struct ib_cq ibcq; member in struct:nes_cq
|
H A D | nes.h | 494 static inline struct nes_cq *to_nescq(struct ib_cq *ibcq) argument 496 return container_of(ibcq, struct nes_cq, ibcq);
|
H A D | nes_verbs.c | 1563 nescq->ibcq.cqe = nescq->hw_cq.cq_size - 1; 1772 return &nescq->ibcq; 3629 static int nes_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) argument 3634 struct nes_vnic *nesvnic = to_nesvnic(ibcq->device); 3636 struct nes_cq *nescq = to_nescq(ibcq); 3804 static int nes_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) argument 3806 struct nes_vnic *nesvnic = to_nesvnic(ibcq->device); 3808 struct nes_cq *nescq = to_nescq(ibcq);
|