Searched defs:ibcq (Results 1 - 20 of 20) sorted by relevance

/drivers/infiniband/hw/cxgb3/
H A Diwch_cq.c195 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) argument
203 chp = to_iwch_cq(ibcq);
H A Diwch_provider.c191 chp->ibcq.cqe = 1 << chp->cq.size_log2;
207 iwch_destroy_cq(&chp->ibcq);
233 iwch_destroy_cq(&chp->ibcq);
241 return &chp->ibcq;
292 chp->ibcq.cqe = (1<<chp->cq.size_log2) - 1;
312 static int iwch_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) argument
321 chp = to_iwch_cq(ibcq);
H A Diwch_provider.h102 struct ib_cq ibcq; member in struct:iwch_cq
112 static inline struct iwch_cq *to_iwch_cq(struct ib_cq *ibcq) argument
114 return container_of(ibcq, struct iwch_cq, ibcq);
334 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
/drivers/infiniband/hw/ipath/
H A Dipath_cq.c63 if (head >= (unsigned) cq->ibcq.cqe) {
64 head = cq->ibcq.cqe;
70 if (cq->ibcq.event_handler) {
73 ev.device = cq->ibcq.device;
74 ev.element.cq = &cq->ibcq;
76 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context);
115 to_idev(cq->ibcq.device)->n_wqe_errs++;
120 * @ibcq: the completion queue to poll
129 int ipath_poll_cq(struct ib_cq *ibcq, in argument
314 ipath_destroy_cq(struct ib_cq *ibcq) argument
342 ipath_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) argument
371 ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) argument
[all...]
H A Dipath_verbs.h205 /* these are actually size ibcq.cqe + 1 */
215 struct ib_cq ibcq; member in struct:ipath_cq
670 static inline struct ipath_cq *to_icq(struct ib_cq *ibcq) argument
672 return container_of(ibcq, struct ipath_cq, ibcq);
808 int ipath_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry);
814 int ipath_destroy_cq(struct ib_cq *ibcq);
816 int ipath_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags);
818 int ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
/drivers/infiniband/hw/qib/
H A Dqib_cq.c63 if (head >= (unsigned) cq->ibcq.cqe) {
64 head = cq->ibcq.cqe;
70 if (cq->ibcq.event_handler) {
73 ev.device = cq->ibcq.device;
74 ev.element.cq = &cq->ibcq;
76 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context);
119 * @ibcq: the completion queue to poll
128 int qib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) argument
130 struct qib_cq *cq = to_icq(ibcq);
321 qib_destroy_cq(struct ib_cq *ibcq) argument
349 qib_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) argument
378 qib_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) argument
[all...]
H A Dqib_verbs.h258 /* these are actually size ibcq.cqe + 1 */
268 struct ib_cq ibcq; member in struct:qib_cq
780 static inline struct qib_cq *to_icq(struct ib_cq *ibcq) argument
782 return container_of(ibcq, struct qib_cq, ibcq);
955 int qib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry);
961 int qib_destroy_cq(struct ib_cq *ibcq);
963 int qib_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags);
965 int qib_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
/drivers/infiniband/hw/amso1100/
H A Dc2_cq.c78 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context);
201 int c2_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) argument
203 struct c2_dev *c2dev = to_c2dev(ibcq->device);
204 struct c2_cq *cq = to_c2cq(ibcq);
222 int c2_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) argument
229 cq = to_c2cq(ibcq);
297 cq->ibcq.cqe = entries - 1;
H A Dc2_provider.h92 struct ib_cq ibcq; member in struct:c2_cq
153 static inline struct c2_cq *to_c2cq(struct ib_cq *ibcq) argument
155 return container_of(ibcq, struct c2_cq, ibcq);
/drivers/infiniband/hw/cxgb4/
H A Dcq.c709 int c4iw_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) argument
716 chp = to_c4iw_cq(ibcq);
819 chp->ibcq.cqe = entries - 2;
864 return &chp->ibcq;
884 int c4iw_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) argument
890 chp = to_c4iw_cq(ibcq);
H A Diw_cxgb4.h308 struct ib_cq ibcq; member in struct:c4iw_cq
317 static inline struct c4iw_cq *to_c4iw_cq(struct ib_cq *ibcq) argument
319 return container_of(ibcq, struct c4iw_cq, ibcq);
722 int c4iw_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
765 int c4iw_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags);
/drivers/infiniband/hw/mlx4/
H A Dcq.c43 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; local
44 ibcq->comp_handler(ibcq, ibcq->cq_context);
50 struct ib_cq *ibcq; local
58 ibcq = &to_mibcq(cq)->ibcq;
59 if (ibcq->event_handler) {
60 event.device = ibcq
345 mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) argument
735 mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) argument
761 mlx4_ib_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) argument
[all...]
H A Dmlx4_ib.h77 struct ib_cq ibcq; member in struct:mlx4_ib_cq
227 static inline struct mlx4_ib_cq *to_mcq(struct ib_cq *ibcq) argument
229 return container_of(ibcq, struct mlx4_ib_cq, ibcq);
294 int mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata);
299 int mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
/drivers/infiniband/hw/ehca/
H A Dehca_main.c533 struct ib_cq *ibcq; local
543 ibcq = ib_create_cq(&shca->ib_device, NULL, NULL, (void *)(-1), 10, 0);
544 if (IS_ERR(ibcq)) {
546 return PTR_ERR(ibcq);
548 sport->ibcq_aqp1 = ibcq;
557 qp_init_attr.send_cq = ibcq;
558 qp_init_attr.recv_cq = ibcq;
/drivers/infiniband/hw/mthca/
H A Dmthca_provider.c703 return &cq->ibcq;
766 static int mthca_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) argument
768 struct mthca_dev *dev = to_mdev(ibcq->device);
769 struct mthca_cq *cq = to_mcq(ibcq);
780 if (entries == ibcq->cqe + 1) {
820 tcqe = cq->ibcq.cqe;
822 cq->ibcq.cqe = cq->resize_buf->cqe;
834 ibcq->cqe = entries - 1;
H A Dmthca_cq.c181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe));
235 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context);
259 event.element.cq = &cq->ibcq;
260 if (cq->ibcq.event_handler)
261 cq->ibcq.event_handler(&event, cq->ibcq.cq_context);
295 cqe_sw(get_cqe(cq, prod_index & cq->ibcq.cqe));
297 if (prod_index == cq->cons_index + cq->ibcq
662 mthca_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) argument
744 mthca_arbel_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) argument
[all...]
H A Dmthca_provider.h202 struct ib_cq ibcq; member in struct:mthca_cq
324 static inline struct mthca_cq *to_mcq(struct ib_cq *ibcq) argument
326 return container_of(ibcq, struct mthca_cq, ibcq);
/drivers/infiniband/hw/nes/
H A Dnes_verbs.h112 struct ib_cq ibcq; member in struct:nes_cq
H A Dnes.h494 static inline struct nes_cq *to_nescq(struct ib_cq *ibcq) argument
496 return container_of(ibcq, struct nes_cq, ibcq);
H A Dnes_verbs.c1563 nescq->ibcq.cqe = nescq->hw_cq.cq_size - 1;
1772 return &nescq->ibcq;
3629 static int nes_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) argument
3634 struct nes_vnic *nesvnic = to_nesvnic(ibcq->device);
3636 struct nes_cq *nescq = to_nescq(ibcq);
3804 static int nes_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) argument
3806 struct nes_vnic *nesvnic = to_nesvnic(ibcq->device);
3808 struct nes_cq *nescq = to_nescq(ibcq);

Completed in 284 milliseconds