/drivers/infiniband/hw/ehca/ |
H A D | ehca_mcast.c | 59 int ehca_attach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 61 struct ehca_qp *my_qp = container_of(ibqp, struct ehca_qp, ib_qp); 62 struct ehca_shca *shca = container_of(ibqp->device, struct ehca_shca, 67 if (ibqp->qp_type != IB_QPT_UD) { 68 ehca_err(ibqp->device, "invalid qp_type=%x", ibqp->qp_type); 73 ehca_err(ibqp->device, "invalid mulitcast gid"); 76 ehca_err(ibqp->device, "invalid mulitcast lid=%x", lid); 89 ehca_err(ibqp->device, 91 "h_ret=%lli", my_qp, ibqp 96 ehca_detach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument [all...] |
H A D | ehca_qp.c | 1241 static int internal_modify_qp(struct ib_qp *ibqp, argument 1249 struct ehca_qp *my_qp = container_of(ibqp, struct ehca_qp, ib_qp); 1251 container_of(ibqp->pd->device, struct ehca_shca, ib_device); 1262 ehca_err(ibqp->device, "Could not get zeroed page for mqpcb " 1263 "ehca_qp=%p qp_num=%x ", my_qp, ibqp->qp_num); 1272 ehca_err(ibqp->device, "hipz_h_query_qp() failed " 1274 my_qp, ibqp->qp_num, h_ret); 1278 if (ibqp->uobject) 1285 ehca_err(ibqp->device, "Invalid current ehca_qp_state=%x " 1287 mqpcb->qp_state, my_qp, ibqp 1782 ehca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) argument [all...] |
/drivers/infiniband/hw/qib/ |
H A D | qib_ud.c | 51 struct qib_ibport *ibp = to_iport(sqp->ibqp.device, sqp->port_num); 66 if (qp->ibqp.qp_type != sqp->ibqp.qp_type || 75 if (qp->ibqp.qp_num > 1) { 87 sqp->ibqp.qp_num, qp->ibqp.qp_num, 99 if (qp->ibqp.qp_num) { 111 sqp->ibqp.qp_num, qp->ibqp.qp_num, 147 if (qp->ibqp [all...] |
H A D | qib_verbs_mcast.c | 225 int qib_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 227 struct qib_qp *qp = to_iqp(ibqp); 228 struct qib_ibdev *dev = to_idev(ibqp->device); 234 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) { 254 ibp = to_iport(ibqp->device, qp->port_num); 283 int qib_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 285 struct qib_qp *qp = to_iqp(ibqp); 286 struct qib_ibdev *dev = to_idev(ibqp->device); 287 struct qib_ibport *ibp = to_iport(ibqp->device, qp->port_num); 294 if (ibqp [all...] |
H A D | qib_qp.c | 221 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); 223 unsigned n = qpn_hash(dev, qp->ibqp.qp_num); 228 if (qp->ibqp.qp_num == 0) 230 else if (qp->ibqp.qp_num == 1) 247 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); 248 unsigned n = qpn_hash(dev, qp->ibqp.qp_num); 341 if (qp->ibqp.qp_num == qpn) 432 if (qp->ibqp.qp_type == IB_QPT_UD || 433 qp->ibqp.qp_type == IB_QPT_SMI || 434 qp->ibqp 564 qib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) argument 850 qib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) argument 1178 qib_destroy_qp(struct ib_qp *ibqp) argument [all...] |
H A D | qib_ruc.c | 89 rkt = &to_idev(qp->ibqp.device)->lk_table; 90 pd = to_ipd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd); 120 wc.qp = &qp->ibqp; 122 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 149 if (qp->ibqp.srq) { 150 srq = to_isrq(qp->ibqp.srq); 212 ev.device = qp->ibqp.device; 213 ev.element.srq = qp->ibqp [all...] |
H A D | qib_keys.c | 229 struct qib_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table; 241 struct qib_pd *pd = to_ipd(qp->ibqp.pd); 261 if (unlikely(mr == NULL || mr->lkey != rkey || qp->ibqp.pd != mr->pd)) 314 struct qib_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table; 315 struct qib_pd *pd = to_ipd(qp->ibqp.pd); 329 if (unlikely(mr == NULL || qp->ibqp.pd != mr->pd))
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_ruc.c | 89 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); 149 wc.qp = &qp->ibqp; 151 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 177 if (qp->ibqp.srq) { 178 srq = to_isrq(qp->ibqp.srq); 236 ev.device = qp->ibqp.device; 237 ev.element.srq = qp->ibqp.srq; 262 struct ipath_ibdev *dev = to_idev(sqp->ibqp.device); 318 if (sqp->ibqp.qp_type == IB_QPT_RC) 439 wc.qp = &qp->ibqp; [all...] |
H A D | ipath_qp.c | 218 qp->ibqp.qp_num = ret; 251 qpp = &qpt->table[qp->ibqp.qp_num % qpt->max]; 309 if (qp->ibqp.qp_num == qpn) { 379 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); 400 wc.qp = &qp->ibqp; 406 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 429 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); 434 } else if (qp->ibqp.event_handler) 443 * @ibqp: the queue pair who's attributes we're modifying 450 int ipath_modify_qp(struct ib_qp *ibqp, struc argument 621 ipath_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) argument 962 ipath_destroy_qp(struct ib_qp *ibqp) argument [all...] |
H A D | ipath_ud.c | 52 struct ipath_ibdev *dev = to_idev(sqp->ibqp.device); 79 if (unlikely(qp->ibqp.qp_num && 107 if (qp->ibqp.srq) { 108 srq = to_isrq(qp->ibqp.srq); 169 ev.device = qp->ibqp.device; 170 ev.element.srq = qp->ibqp.srq; 215 wc.qp = &qp->ibqp; 216 wc.src_qp = sqp->ibqp.qp_num; 227 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 243 struct ipath_ibdev *dev = to_idev(qp->ibqp [all...] |
H A D | ipath_keys.c | 124 struct ipath_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table; 136 struct ipath_pd *pd = to_ipd(qp->ibqp.pd); 151 qp->ibqp.pd != mr->pd)) { 202 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); 216 struct ipath_pd *pd = to_ipd(qp->ibqp.pd); 234 qp->ibqp.pd != mr->pd)) {
|
H A D | ipath_verbs_mcast.c | 236 int ipath_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 238 struct ipath_qp *qp = to_iqp(ibqp); 239 struct ipath_ibdev *dev = to_idev(ibqp->device); 285 int ipath_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 287 struct ipath_qp *qp = to_iqp(ibqp); 288 struct ipath_ibdev *dev = to_idev(ibqp->device);
|
/drivers/infiniband/hw/cxgb4/ |
H A D | ev.c | 72 event.element.qp = &qhp->ibqp; 73 if (qhp->ibqp.event_handler) 74 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); 116 c4iw_qp_add_ref(&qhp->ibqp); 181 c4iw_qp_rem_ref(&qhp->ibqp);
|
/drivers/infiniband/hw/amso1100/ |
H A D | c2_provider.h | 108 struct ib_qp ibqp; member in struct:c2_qp 158 static inline struct c2_qp *to_c2qp(struct ib_qp *ibqp) argument 160 return container_of(ibqp, struct c2_qp, ibqp);
|
H A D | c2_cm.c | 44 struct ib_qp *ibqp; local 50 ibqp = c2_get_qp(cm_id->device, iw_param->qpn); 51 if (!ibqp) 53 qp = to_c2qp(ibqp); 287 struct ib_qp *ibqp; local 293 ibqp = c2_get_qp(cm_id->device, iw_param->qpn); 294 if (!ibqp) 296 qp = to_c2qp(ibqp);
|
H A D | c2_ae.c | 233 ib_event.element.qp = &qp->ibqp; 236 if (qp->ibqp.event_handler) 237 qp->ibqp.event_handler(&ib_event, 238 qp->ibqp.
|
H A D | c2_provider.c | 202 static void c2_add_ref(struct ib_qp *ibqp) argument 205 BUG_ON(!ibqp); 206 qp = to_c2qp(ibqp); 210 static void c2_rem_ref(struct ib_qp *ibqp) argument 213 BUG_ON(!ibqp); 214 qp = to_c2qp(ibqp); 229 return (qp?&qp->ibqp:NULL); 275 return &qp->ibqp; 563 static int c2_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, argument 569 c2_qp_modify(to_c2dev(ibqp 575 c2_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 581 c2_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument [all...] |
/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 184 if (qp->ibqp.qp_type == IB_QPT_UD) { 188 av->port_pd = cpu_to_be32((qp->port << 24) | to_mpd(qp->ibqp.pd)->pdn); 225 struct ib_qp *ibqp = &to_mibqp(qp)->ibqp; local 230 if (ibqp->event_handler) { 231 event.device = ibqp->device; 232 event.element.qp = ibqp; 264 ibqp->event_handler(&event, ibqp->qp_context); 685 if (qp->ibqp 1006 __mlx4_ib_modify_qp(struct ib_qp *ibqp, const struct ib_qp_attr *attr, int attr_mask, enum ib_qp_state cur_state, enum ib_qp_state new_state) argument 1301 mlx4_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) argument 1694 mlx4_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, struct ib_send_wr **bad_wr) argument 1943 mlx4_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr, struct ib_recv_wr **bad_wr) argument 2080 mlx4_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) argument [all...] |
H A D | mlx4_ib.h | 130 struct ib_qp ibqp; member in struct:mlx4_ib_qp 251 static inline struct mlx4_ib_qp *to_mqp(struct ib_qp *ibqp) argument 253 return container_of(ibqp, struct mlx4_ib_qp, ibqp); 323 int mlx4_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 325 int mlx4_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, 327 int mlx4_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, 329 int mlx4_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr,
|
/drivers/infiniband/hw/mthca/ |
H A D | mthca_mcg.c | 120 int mthca_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 122 struct mthca_dev *dev = to_mdev(ibqp->device); 165 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) { 167 ibqp->qp_num); 171 mgm->qp[i] = cpu_to_be32(ibqp->qp_num | (1 << 31)); 214 int mthca_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) argument 216 struct mthca_dev *dev = to_mdev(ibqp->device); 242 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) 249 mthca_err(dev, "QP %06x not found in MGM\n", ibqp->qp_num);
|
H A D | mthca_provider.h | 262 struct ib_qp ibqp; member in struct:mthca_qp 334 static inline struct mthca_qp *to_mqp(struct ib_qp *ibqp) argument 336 return container_of(ibqp, struct mthca_qp, ibqp);
|
H A D | mthca_qp.c | 259 event.element.qp = &qp->ibqp; 260 if (qp->ibqp.event_handler) 261 qp->ibqp.event_handler(&event, qp->ibqp.qp_context); 423 int mthca_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, argument 426 struct mthca_dev *dev = to_mdev(ibqp->device); 427 struct mthca_qp *qp = to_mqp(ibqp); 541 static int __mthca_modify_qp(struct ib_qp *ibqp, argument 545 struct mthca_dev *dev = to_mdev(ibqp->device); 546 struct mthca_qp *qp = to_mqp(ibqp); 840 mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) argument 1599 mthca_tavor_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, struct ib_send_wr **bad_wr) argument 1802 mthca_tavor_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, struct ib_recv_wr **bad_wr) argument 1913 mthca_arbel_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, struct ib_send_wr **bad_wr) argument 2153 mthca_arbel_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, struct ib_recv_wr **bad_wr) argument [all...] |
H A D | mthca_dev.h | 529 int mthca_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, 531 int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, 533 int mthca_tavor_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, 535 int mthca_tavor_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, 537 int mthca_arbel_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, 539 int mthca_arbel_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, 573 int mthca_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid); 574 int mthca_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
|
/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_ev.c | 93 event.element.qp = &qhp->ibqp; 95 if (qhp->ibqp.event_handler) 96 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); 129 iwch_qp_add_ref(&qhp->ibqp); 229 iwch_qp_rem_ref(&qhp->ibqp);
|
H A D | iwch_provider.h | 160 struct ib_qp ibqp; member in struct:iwch_qp 177 static inline struct iwch_qp *to_iwch_qp(struct ib_qp *ibqp) argument 179 return container_of(ibqp, struct iwch_qp, ibqp); 327 int iwch_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, 329 int iwch_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr,
|