/drivers/infiniband/hw/ipath/ |
H A D | ipath_ruc.c | 251 * @sqp: the sending QP 260 static void ipath_ruc_loopback(struct ipath_qp *sqp) argument 262 struct ipath_ibdev *dev = to_idev(sqp->ibqp.device); 276 qp = ipath_lookup_qpn(&dev->qp_table, sqp->remote_qpn); 278 spin_lock_irqsave(&sqp->s_lock, flags); 281 if ((sqp->s_flags & (IPATH_S_BUSY | IPATH_S_ANY_WAIT)) || 282 !(ib_ipath_state_ops[sqp->state] & IPATH_PROCESS_OR_FLUSH_SEND)) 285 sqp->s_flags |= IPATH_S_BUSY; 288 if (sqp->s_last == sqp [all...] |
H A D | ipath_ud.c | 42 * @sqp: the sending QP 50 static void ipath_ud_loopback(struct ipath_qp *sqp, struct ipath_swqe *swqe) argument 52 struct ipath_ibdev *dev = to_idev(sqp->ibqp.device); 81 sqp->qkey : swqe->wr.wr.ud.remote_qkey) != qp->qkey)) { 216 wc.src_qp = sqp->ibqp.qp_num;
|
/drivers/infiniband/hw/qib/ |
H A D | qib_ruc.c | 346 * @sqp: the sending QP 355 static void qib_ruc_loopback(struct qib_qp *sqp) argument 357 struct qib_ibport *ibp = to_iport(sqp->ibqp.device, sqp->port_num); 373 qp = qib_lookup_qpn(ibp, sqp->remote_qpn); 375 spin_lock_irqsave(&sqp->s_lock, flags); 378 if ((sqp->s_flags & (QIB_S_BUSY | QIB_S_ANY_WAIT)) || 379 !(ib_qib_state_ops[sqp->state] & QIB_PROCESS_OR_FLUSH_SEND)) 382 sqp->s_flags |= QIB_S_BUSY; 385 if (sqp [all...] |
H A D | qib_ud.c | 41 * @sqp: the sending QP 49 static void qib_ud_loopback(struct qib_qp *sqp, struct qib_swqe *swqe) argument 51 struct qib_ibport *ibp = to_iport(sqp->ibqp.device, sqp->port_num); 66 if (qp->ibqp.qp_type != sqp->ibqp.qp_type || 80 pkey1 = qib_get_pkey(ibp, sqp->s_pkey_index); 87 sqp->ibqp.qp_num, qp->ibqp.qp_num, 103 sqp->qkey : swqe->wr.wr.ud.remote_qkey; 111 sqp->ibqp.qp_num, qp->ibqp.qp_num, 208 wc.src_qp = sqp [all...] |
/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_qp.c | 362 struct t3_swsq *sqp; local 391 sqp = qhp->wq.sq + 413 sqp->read_len = wqe->read.local_len; 415 qhp->wq.oldest_read = sqp; 436 sqp->wr_id = wr->wr_id; 437 sqp->opcode = wr2opcode(t3_wr_opcode); 438 sqp->sq_wptr = qhp->wq.sq_wptr; 439 sqp->complete = 0; 440 sqp->signaled = (wr->send_flags & IB_SEND_SIGNALED); 449 sqp 544 struct t3_swsq *sqp; local [all...] |
H A D | cxio_hal.c | 387 struct t3_swsq *sqp) 395 V_CQE_OPCODE(sqp->opcode) | 401 cqe.u.scqe.wrid_hi = sqp->sq_wptr; 411 struct t3_swsq *sqp = wq->sq + Q_PTR2IDX(wq->sq_rptr, wq->sq_size_log2); local 414 sqp = wq->sq + Q_PTR2IDX(ptr, wq->sq_size_log2); 416 sqp->signaled = 0; 417 insert_sq_cqe(wq, cq, sqp); 419 sqp = wq->sq + Q_PTR2IDX(ptr, wq->sq_size_log2); 1077 struct t3_swsq *sqp; local 1081 sqp 386 insert_sq_cqe(struct t3_wq *wq, struct t3_cq *cq, struct t3_swsq *sqp) argument 1281 struct t3_swsq *sqp; local [all...] |
/drivers/infiniband/hw/ehca/ |
H A D | ehca_qp.c | 1847 void ehca_recover_sqp(struct ib_qp *sqp) argument 1849 struct ehca_qp *my_sqp = container_of(sqp, struct ehca_qp, ib_qp); 1858 ehca_dbg(sqp->device, "SQP port=%x qp_num=%x", port, sqp->qp_num); 1864 ret = internal_modify_qp(sqp, &attr, qp_parm[i].mask, 0); 1866 ehca_err(sqp->device, "Could not modify SQP port=%x " 1867 "qp_num=%x ret=%x", port, sqp->qp_num, ret); 1870 ehca_dbg(sqp->device, "SQP port=%x qp_num=%x in state=%x", 1871 port, sqp->qp_num, attr.qp_state); 1881 ehca_dbg(sqp [all...] |
/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 763 struct mlx4_ib_sqp *sqp; local 817 sqp = kzalloc(sizeof *sqp, GFP_KERNEL); 818 if (!sqp) 821 qp = &sqp->qp; 829 kfree(sqp); 910 static void store_sqp_attrs(struct mlx4_ib_sqp *sqp, const struct ib_qp_attr *attr, argument 914 sqp->pkey_index = attr->pkey_index; 916 sqp->qkey = attr->qkey; 918 sqp 1350 build_mlx_header(struct mlx4_ib_sqp *sqp, struct ib_send_wr *wr, void *wqe, unsigned *mlx_seg_len) argument [all...] |
/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 297 static void store_attrs(struct mthca_sqp *sqp, const struct ib_qp_attr *attr, argument 301 sqp->pkey_index = attr->pkey_index; 303 sqp->qkey = attr->qkey; 305 sqp->send_psn = attr->sq_psn; 1348 struct mthca_sqp *sqp) 1353 sqp->qp.transport = MLX; 1354 err = mthca_set_qp_size(dev, cap, pd, &sqp->qp); 1358 sqp->header_buf_size = sqp->qp.sq.max * MTHCA_UD_HEADER_SIZE; 1359 sqp 1340 mthca_alloc_sqp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct ib_qp_cap *cap, int qpn, int port, struct mthca_sqp *sqp) argument 1475 build_mlx_header(struct mthca_dev *dev, struct mthca_sqp *sqp, int ind, struct ib_send_wr *wr, struct mthca_mlx_seg *mlx, struct mthca_data_seg *data) argument [all...] |