/drivers/infiniband/hw/ipath/ |
H A D | ipath_rc.c | 56 * ipath_init_restart- initialize the qp->s_sge after a restart 57 * @qp: the QP who's SGE we're restarting 62 static void ipath_init_restart(struct ipath_qp *qp, struct ipath_swqe *wqe) argument 66 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, 67 ib_mtu_enum_to_int(qp->path_mtu)); 68 dev = to_idev(qp->ibqp.device); 70 if (list_empty(&qp->timerwait)) 71 list_add_tail(&qp 86 ipath_make_rc_ack(struct ipath_ibdev *dev, struct ipath_qp *qp, struct ipath_other_headers *ohdr, u32 pmtu) argument 213 ipath_make_rc_req(struct ipath_qp *qp) argument 612 send_rc_ack(struct ipath_qp *qp) argument 724 reset_psn(struct ipath_qp *qp, u32 psn) argument 805 ipath_restart_rc(struct ipath_qp *qp, u32 psn) argument 841 update_last_psn(struct ipath_qp *qp, u32 psn) argument 857 do_rc_ack(struct ipath_qp *qp, u32 aeth, u32 psn, int opcode, u64 val) argument 1117 ipath_rc_rcv_resp(struct ipath_ibdev *dev, struct ipath_other_headers *ohdr, void *data, u32 tlen, struct ipath_qp *qp, u32 opcode, u32 psn, u32 hdrsize, u32 pmtu, int header_in_data) argument 1335 ipath_rc_rcv_error(struct ipath_ibdev *dev, struct ipath_other_headers *ohdr, void *data, struct ipath_qp *qp, u32 opcode, u32 psn, int diff, int header_in_data) argument 1525 ipath_rc_error(struct ipath_qp *qp, enum ib_wc_status err) argument 1544 ipath_update_ack_queue(struct ipath_qp *qp, unsigned n) argument 1570 ipath_rc_rcv(struct ipath_ibdev *dev, struct ipath_ib_header *hdr, int has_grh, void *data, u32 tlen, struct ipath_qp *qp) argument [all...] |
H A D | ipath_qp.c | 203 * @qp: the QP 209 static int ipath_alloc_qpn(struct ipath_qp_table *qpt, struct ipath_qp *qp, argument 218 qp->ibqp.qp_num = ret; 224 qp->next = qpt->table[ret]; 225 qpt->table[ret] = qp; 226 atomic_inc(&qp->refcount); 238 * @qp: the QP to remove 243 static void ipath_free_qp(struct ipath_qp_table *qpt, struct ipath_qp *qp) argument 251 qpp = &qpt->table[qp->ibqp.qp_num % qpt->max]; 253 if (q == qp) { 274 struct ipath_qp *qp; local 304 struct ipath_qp *qp; local 324 ipath_reset_qp(struct ipath_qp *qp, enum ib_qp_type type) argument 377 ipath_error_qp(struct ipath_qp *qp, enum ib_wc_status err) argument 454 struct ipath_qp *qp = to_iqp(ibqp); local 624 struct ipath_qp *qp = to_iqp(ibqp); local 677 ipath_compute_aeth(struct ipath_qp *qp) argument 745 struct ipath_qp *qp; local 964 struct ipath_qp *qp = to_iqp(ibqp); local 1055 ipath_get_credit(struct ipath_qp *qp, u32 aeth) argument [all...] |
H A D | ipath_ruc.c | 80 * @qp: the QP 87 void ipath_insert_rnr_queue(struct ipath_qp *qp) argument 89 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); 94 list_add(&qp->timerwait, &dev->rnrwait); 100 while (qp->s_rnr_timeout >= nqp->s_rnr_timeout) { 101 qp->s_rnr_timeout -= nqp->s_rnr_timeout; 111 nqp->s_rnr_timeout -= qp->s_rnr_timeout; 112 list_add(&qp->timerwait, l); 119 * @qp: the QP 123 int ipath_init_sge(struct ipath_qp *qp, struc argument 166 ipath_get_rwqe(struct ipath_qp *qp, int wr_id_only) argument 263 struct ipath_qp *qp; local 517 want_buffer(struct ipath_devdata *dd, struct ipath_qp *qp) argument 541 ipath_no_bufs_available(struct ipath_qp *qp, struct ipath_ibdev *dev) argument 600 ipath_make_ruc_header(struct ipath_ibdev *dev, struct ipath_qp *qp, struct ipath_other_headers *ohdr, u32 bth0, u32 bth2) argument 641 struct ipath_qp *qp = (struct ipath_qp *)data; local 699 ipath_send_complete(struct ipath_qp *qp, struct ipath_swqe *wqe, enum ib_wc_status status) argument [all...] |
H A D | ipath_uc.c | 42 * @qp: a pointer to the QP 46 int ipath_make_uc_req(struct ipath_qp *qp) argument 54 u32 pmtu = ib_mtu_enum_to_int(qp->path_mtu); 57 spin_lock_irqsave(&qp->s_lock, flags); 59 if (!(ib_ipath_state_ops[qp->state] & IPATH_PROCESS_SEND_OK)) { 60 if (!(ib_ipath_state_ops[qp->state] & IPATH_FLUSH_SEND)) 63 if (qp->s_last == qp->s_head) 66 if (atomic_read(&qp->s_dma_busy)) { 67 qp 240 ipath_uc_rcv(struct ipath_ibdev *dev, struct ipath_ib_header *hdr, int has_grh, void *data, u32 tlen, struct ipath_qp *qp) argument [all...] |
H A D | ipath_ud.c | 53 struct ipath_qp *qp; local 68 qp = ipath_lookup_qpn(&dev->qp_table, swqe->wr.wr.ud.remote_qpn); 69 if (!qp || !(ib_ipath_state_ops[qp->state] & IPATH_PROCESS_RECV_OK)) { 79 if (unlikely(qp->ibqp.qp_num && 81 sqp->qkey : swqe->wr.wr.ud.remote_qkey) != qp->qkey)) { 107 if (qp->ibqp.srq) { 108 srq = to_isrq(qp->ibqp.srq); 114 rq = &qp->r_rq; 134 rsge.sg_list = qp 241 ipath_make_ud_req(struct ipath_qp *qp) argument 409 ipath_ud_rcv(struct ipath_ibdev *dev, struct ipath_ib_header *hdr, int has_grh, void *data, u32 tlen, struct ipath_qp *qp) argument [all...] |
H A D | ipath_verbs.c | 333 * @qp: the QP to post on 336 static int ipath_post_one_send(struct ipath_qp *qp, struct ib_send_wr *wr) argument 345 struct ipath_devdata *dd = to_idev(qp->ibqp.device)->dd; 347 spin_lock_irqsave(&qp->s_lock, flags); 349 if (qp->ibqp.qp_type != IB_QPT_SMI && 356 if (unlikely(!(ib_ipath_state_ops[qp->state] & IPATH_POST_SEND_OK))) 360 if (wr->num_sge > qp->s_max_sge) 368 if (qp->ibqp.qp_type == IB_QPT_UC) { 371 } else if (qp->ibqp.qp_type == IB_QPT_UD) { 377 if (qp 448 struct ipath_qp *qp = to_iqp(ibqp); local 477 struct ipath_qp *qp = to_iqp(ibqp); local 540 ipath_qp_rcv(struct ipath_ibdev *dev, struct ipath_ib_header *hdr, int has_grh, void *data, u32 tlen, struct ipath_qp *qp) argument 588 struct ipath_qp *qp; local 682 struct ipath_qp *qp; local 1031 struct ipath_qp *qp = tx->qp; local 1061 decrement_dma_busy(struct ipath_qp *qp) argument 1097 ipath_verbs_send_dma(struct ipath_qp *qp, struct ipath_ib_header *hdr, u32 hdrwords, struct ipath_sge_state *ss, u32 len, u32 plen, u32 dwords) argument 1222 ipath_verbs_send_pio(struct ipath_qp *qp, struct ipath_ib_header *ibhdr, u32 hdrwords, struct ipath_sge_state *ss, u32 len, u32 plen, u32 dwords) argument 1322 ipath_verbs_send(struct ipath_qp *qp, struct ipath_ib_header *hdr, u32 hdrwords, struct ipath_sge_state *ss, u32 len) argument 1461 struct ipath_qp *qp; local [all...] |
/drivers/infiniband/hw/qib/ |
H A D | qib_qp.c | 219 static void insert_qp(struct qib_ibdev *dev, struct qib_qp *qp) argument 221 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); 223 unsigned n = qpn_hash(dev, qp->ibqp.qp_num); 226 atomic_inc(&qp->refcount); 228 if (qp->ibqp.qp_num == 0) 229 rcu_assign_pointer(ibp->qp0, qp); 230 else if (qp->ibqp.qp_num == 1) 231 rcu_assign_pointer(ibp->qp1, qp); 233 qp 245 remove_qp(struct qib_ibdev *dev, struct qib_qp *qp) argument 287 struct qib_qp *qp; local 327 struct qib_qp *qp = NULL; local 357 qib_reset_qp(struct qib_qp *qp, enum ib_qp_type type) argument 404 clear_mr_refs(struct qib_qp *qp, int clr_sends) argument 469 qib_error_qp(struct qib_qp *qp, enum ib_wc_status err) argument 568 struct qib_qp *qp = to_iqp(ibqp); local 853 struct qib_qp *qp = to_iqp(ibqp); local 906 qib_compute_aeth(struct qib_qp *qp) argument 974 struct qib_qp *qp; local 1180 struct qib_qp *qp = to_iqp(ibqp); local 1253 qib_get_credit(struct qib_qp *qp, u32 aeth) argument [all...] |
H A D | qib_rc.c | 57 static void start_timer(struct qib_qp *qp) argument 59 qp->s_flags |= QIB_S_TIMER; 60 qp->s_timer.function = rc_timeout; 61 /* 4.096 usec. * (1 << qp->timeout) */ 62 qp->s_timer.expires = jiffies + qp->timeout_jiffies; 63 add_timer(&qp->s_timer); 69 * @qp: a pointer to the QP 77 static int qib_make_rc_ack(struct qib_ibdev *dev, struct qib_qp *qp, argument 87 if (!(ib_qib_state_ops[qp 231 qib_make_rc_req(struct qib_qp *qp) argument 648 qib_send_rc_ack(struct qib_qp *qp) argument 783 reset_psn(struct qib_qp *qp, u32 psn) argument 868 qib_restart_rc(struct qib_qp *qp, u32 psn, int wait) argument 905 struct qib_qp *qp = (struct qib_qp *)arg; local 928 struct qib_qp *qp = (struct qib_qp *)arg; local 944 reset_sending_psn(struct qib_qp *qp, u32 psn) argument 969 qib_rc_send_complete(struct qib_qp *qp, struct qib_ib_header *hdr) argument 1044 update_last_psn(struct qib_qp *qp, u32 psn) argument 1054 do_rc_completion(struct qib_qp *qp, struct qib_swqe *wqe, struct qib_ibport *ibp) argument 1127 do_rc_ack(struct qib_qp *qp, u32 aeth, u32 psn, int opcode, u64 val, struct qib_ctxtdata *rcd) argument 1350 rdma_seq_err(struct qib_qp *qp, struct qib_ibport *ibp, u32 psn, struct qib_ctxtdata *rcd) argument 1397 qib_rc_rcv_resp(struct qib_ibport *ibp, struct qib_other_headers *ohdr, void *data, u32 tlen, struct qib_qp *qp, u32 opcode, u32 psn, u32 hdrsize, u32 pmtu, struct qib_ctxtdata *rcd) argument 1622 qib_rc_rcv_error(struct qib_other_headers *ohdr, void *data, struct qib_qp *qp, u32 opcode, u32 psn, int diff, struct qib_ctxtdata *rcd) argument 1819 qib_rc_error(struct qib_qp *qp, enum ib_wc_status err) argument 1838 qib_update_ack_queue(struct qib_qp *qp, unsigned n) argument 1862 qib_rc_rcv(struct qib_ctxtdata *rcd, struct qib_ib_header *hdr, int has_grh, void *data, u32 tlen, struct qib_qp *qp) argument [all...] |
H A D | qib_uc.c | 42 * @qp: a pointer to the QP 46 int qib_make_uc_req(struct qib_qp *qp) argument 54 u32 pmtu = qp->pmtu; 57 spin_lock_irqsave(&qp->s_lock, flags); 59 if (!(ib_qib_state_ops[qp->state] & QIB_PROCESS_SEND_OK)) { 60 if (!(ib_qib_state_ops[qp->state] & QIB_FLUSH_SEND)) 63 if (qp->s_last == qp->s_head) 66 if (atomic_read(&qp->s_dma_busy)) { 67 qp 242 qib_uc_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr, int has_grh, void *data, u32 tlen, struct qib_qp *qp) argument [all...] |
H A D | qib_ud.c | 53 struct qib_qp *qp; local 61 qp = qib_lookup_qpn(ibp, swqe->wr.wr.ud.remote_qpn); 62 if (!qp) { 66 if (qp->ibqp.qp_type != sqp->ibqp.qp_type || 67 !(ib_qib_state_ops[qp->state] & QIB_PROCESS_RECV_OK)) { 75 if (qp->ibqp.qp_num > 1) { 81 pkey2 = qib_get_pkey(ibp, qp->s_pkey_index); 87 sqp->ibqp.qp_num, qp->ibqp.qp_num, 99 if (qp->ibqp.qp_num) { 104 if (unlikely(qkey != qp 232 qib_make_ud_req(struct qib_qp *qp) argument 425 qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr, int has_grh, void *data, u32 tlen, struct qib_qp *qp) argument [all...] |
H A D | qib_ruc.c | 81 static int qib_init_sge(struct qib_qp *qp, struct qib_rwqe *wqe) argument 89 rkt = &to_idev(qp->ibqp.device)->lk_table; 90 pd = to_ipd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd); 91 ss = &qp->r_sge; 92 ss->sg_list = qp->r_sg_list; 93 qp->r_len = 0; 101 qp->r_len += wqe->sg_list[i].length; 105 ss->total_len = qp 138 qib_get_rwqe(struct qib_qp *qp, int wr_id_only) argument 229 qib_migrate_qp(struct qib_qp *qp) argument 267 qib_ruc_check_hdr(struct qib_ibport *ibp, struct qib_ib_header *hdr, int has_grh, struct qib_qp *qp, u32 bth0) argument 358 struct qib_qp *qp; local 678 qib_make_ruc_header(struct qib_qp *qp, struct qib_other_headers *ohdr, u32 bth0, u32 bth2) argument 722 struct qib_qp *qp = container_of(work, struct qib_qp, s_work); local 773 qib_send_complete(struct qib_qp *qp, struct qib_swqe *wqe, enum ib_wc_status status) argument [all...] |
H A D | qib_verbs_mcast.c | 40 * @qp: the QP to link 42 static struct qib_mcast_qp *qib_mcast_qp_alloc(struct qib_qp *qp) argument 50 mqp->qp = qp; 51 atomic_inc(&qp->refcount); 59 struct qib_qp *qp = mqp->qp; local 62 if (atomic_dec_and_test(&qp->refcount)) 63 wake_up(&qp->wait); 182 if (p->qp 227 struct qib_qp *qp = to_iqp(ibqp); local 285 struct qib_qp *qp = to_iqp(ibqp); local [all...] |
H A D | qib_verbs.c | 333 * @qp: the QP to post on 336 static int qib_post_one_send(struct qib_qp *qp, struct ib_send_wr *wr) argument 348 spin_lock_irqsave(&qp->s_lock, flags); 351 if (unlikely(!(ib_qib_state_ops[qp->state] & QIB_POST_SEND_OK))) 355 if (wr->num_sge > qp->s_max_sge) 364 if (qib_fast_reg_mr(qp, wr)) 366 } else if (qp->ibqp.qp_type == IB_QPT_UC) { 369 } else if (qp->ibqp.qp_type != IB_QPT_RC) { 375 if (qp->ibqp.pd != wr->wr.ud.ah->pd) 384 else if (wr->opcode >= IB_WR_RDMA_READ && !qp 458 struct qib_qp *qp = to_iqp(ibqp); local 487 struct qib_qp *qp = to_iqp(ibqp); local 550 qib_qp_rcv(struct qib_ctxtdata *rcd, struct qib_ib_header *hdr, int has_grh, void *data, u32 tlen, struct qib_qp *qp) argument 605 struct qib_qp *qp; local 698 struct qib_qp *qp = NULL; local 916 __get_txreq(struct qib_ibdev *dev, struct qib_qp *qp) argument 947 get_txreq(struct qib_ibdev *dev, struct qib_qp *qp) argument 972 struct qib_qp *qp; local 1025 struct qib_qp *qp, *nqp; local 1070 struct qib_qp *qp = tx->qp; local 1100 wait_kmem(struct qib_ibdev *dev, struct qib_qp *qp) argument 1123 qib_verbs_send_dma(struct qib_qp *qp, struct qib_ib_header *hdr, u32 hdrwords, struct qib_sge_state *ss, u32 len, u32 plen, u32 dwords) argument 1227 no_bufs_available(struct qib_qp *qp) argument 1258 qib_verbs_send_pio(struct qib_qp *qp, struct qib_ib_header *ibhdr, u32 hdrwords, struct qib_sge_state *ss, u32 len, u32 plen, u32 dwords) argument 1365 qib_verbs_send(struct qib_qp *qp, struct qib_ib_header *hdr, u32 hdrwords, struct qib_sge_state *ss, u32 len) argument 1499 struct qib_qp *qp; local [all...] |
/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 195 static int is_sqp(struct mthca_dev *dev, struct mthca_qp *qp) argument 197 return qp->qpn >= dev->qp_table.sqp_start && 198 qp->qpn <= dev->qp_table.sqp_start + 3; 201 static int is_qp0(struct mthca_dev *dev, struct mthca_qp *qp) argument 203 return qp->qpn >= dev->qp_table.sqp_start && 204 qp->qpn <= dev->qp_table.sqp_start + 1; 207 static void *get_recv_wqe(struct mthca_qp *qp, int n) argument 209 if (qp->is_direct) 210 return qp->queue.direct.buf + (n << qp 216 get_send_wqe(struct mthca_qp *qp, int n) argument 240 struct mthca_qp *qp; local 326 get_hw_access_flags(struct mthca_qp *qp, const struct ib_qp_attr *attr, int attr_mask) argument 427 struct mthca_qp *qp = to_mqp(ibqp); local 546 struct mthca_qp *qp = to_mqp(ibqp); local 844 struct mthca_qp *qp = to_mqp(ibqp); local 908 mthca_max_data_size(struct mthca_dev *dev, struct mthca_qp *qp, int desc_sz) argument 942 mthca_adjust_qp_caps(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_qp *qp) argument 967 mthca_alloc_wqe_buf(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_qp *qp) argument 1059 mthca_free_wqe_buf(struct mthca_dev *dev, struct mthca_qp *qp) argument 1068 mthca_map_memfree(struct mthca_dev *dev, struct mthca_qp *qp) argument 1100 mthca_unmap_memfree(struct mthca_dev *dev, struct mthca_qp *qp) argument 1109 mthca_alloc_memfree(struct mthca_dev *dev, struct mthca_qp *qp) argument 1129 mthca_free_memfree(struct mthca_dev *dev, struct mthca_qp *qp) argument 1138 mthca_alloc_qp_common(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct mthca_qp *qp) argument 1227 mthca_set_qp_size(struct mthca_dev *dev, struct ib_qp_cap *cap, struct mthca_pd *pd, struct mthca_qp *qp) argument 1266 mthca_alloc_qp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_qp_type type, enum ib_sig_type send_policy, struct ib_qp_cap *cap, struct mthca_qp *qp) argument 1407 get_qp_refcount(struct mthca_dev *dev, struct mthca_qp *qp) argument 1418 mthca_free_qp(struct mthca_dev *dev, struct mthca_qp *qp) argument 1603 struct mthca_qp *qp = to_mqp(ibqp); local 1806 struct mthca_qp *qp = to_mqp(ibqp); local 1917 struct mthca_qp *qp = to_mqp(ibqp); local 2157 struct mthca_qp *qp = to_mqp(ibqp); local 2224 mthca_free_err_wqe(struct mthca_dev *dev, struct mthca_qp *qp, int is_send, int index, int *dbd, __be32 *new_wqe) argument [all...] |
/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 42 #include <linux/mlx4/qp.h> 74 struct mlx4_ib_qp qp; member in struct:mlx4_ib_sqp 105 return container_of(mqp, struct mlx4_ib_sqp, qp); 108 static int is_sqp(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp) argument 110 return qp->mqp.qpn >= dev->dev->caps.sqp_start && 111 qp->mqp.qpn <= dev->dev->caps.sqp_start + 3; 114 static int is_qp0(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp) argument 116 return qp->mqp.qpn >= dev->dev->caps.sqp_start && 117 qp->mqp.qpn <= dev->dev->caps.sqp_start + 1; 120 static void *get_wqe(struct mlx4_ib_qp *qp, in argument 125 get_recv_wqe(struct mlx4_ib_qp *qp, int n) argument 130 get_send_wqe(struct mlx4_ib_qp *qp, int n) argument 144 stamp_send_wqe(struct mlx4_ib_qp *qp, int n, int size) argument 174 post_nop_wqe(struct mlx4_ib_qp *qp, int n, int size) argument 212 pad_wraparound(struct mlx4_ib_qp *qp, int ind) argument 222 mlx4_ib_qp_event(struct mlx4_qp *qp, enum mlx4_event type) argument 304 set_rq_size(struct mlx4_ib_dev *dev, struct ib_qp_cap *cap, int is_user, int has_rq, struct mlx4_ib_qp *qp) argument 333 set_kernel_sq_size(struct mlx4_ib_dev *dev, struct ib_qp_cap *cap, enum ib_qp_type type, struct mlx4_ib_qp *qp) argument 445 set_user_sq_size(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp, struct mlx4_ib_create_qp *ucmd) argument 473 create_qp_common(struct mlx4_ib_dev *dev, struct ib_pd *pd, struct ib_qp_init_attr *init_attr, struct ib_udata *udata, int sqpn, struct mlx4_ib_qp *qp) argument 673 del_gid_entries(struct mlx4_ib_qp *qp) argument 683 get_pd(struct mlx4_ib_qp *qp) argument 691 get_cqs(struct mlx4_ib_qp *qp, struct mlx4_ib_cq **send_cq, struct mlx4_ib_cq **recv_cq) argument 710 destroy_qp_common(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp, int is_user) argument 764 struct mlx4_ib_qp *qp; local 846 mlx4_ib_destroy_qp(struct ib_qp *qp) argument 880 to_mlx4_access_flags(struct mlx4_ib_qp *qp, const struct ib_qp_attr *attr, int attr_mask) argument 994 update_mcg_macs(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp) argument 1011 struct mlx4_ib_qp *qp = to_mqp(ibqp); local 1305 struct mlx4_ib_qp *qp = to_mqp(ibqp); local 1658 build_lso_seg(struct mlx4_wqe_lso_seg *wqe, struct ib_send_wr *wr, struct mlx4_ib_qp *qp, unsigned *lso_seg_len, __be32 *lso_hdr_sz, __be32 *blh) argument 1697 struct mlx4_ib_qp *qp = to_mqp(ibqp); local 1946 struct mlx4_ib_qp *qp = to_mqp(ibqp); local 2084 struct mlx4_ib_qp *qp = to_mqp(ibqp); local [all...] |
H A D | Makefile | 3 mlx4_ib-y := ah.o cq.o doorbell.o mad.o main.o mr.o qp.o srq.o
|
/drivers/infiniband/hw/ehca/ |
H A D | ehca_uverbs.c | 198 static int ehca_mmap_qp(struct vm_area_struct *vma, struct ehca_qp *qp, argument 205 ehca_dbg(qp->ib_qp.device, "qp_num=%x fw", qp->ib_qp.qp_num); 206 ret = ehca_mmap_fw(vma, &qp->galpas, &qp->mm_count_galpa); 208 ehca_err(qp->ib_qp.device, 210 ret, qp->ib_qp.qp_num); 215 case 1: /* qp rqueue_addr */ 216 ehca_dbg(qp->ib_qp.device, "qp_num=%x rq", qp 256 struct ehca_qp *qp; local [all...] |
H A D | hipz_fns_core.h | 61 static inline void hipz_update_sqa(struct ehca_qp *qp, u16 nr_wqes) argument 64 hipz_galpa_store_qp(qp->galpas.kernel, qpx_sqa, 68 static inline void hipz_update_rqa(struct ehca_qp *qp, u16 nr_wqes) argument 71 hipz_galpa_store_qp(qp->galpas.kernel, qpx_rqa,
|
/drivers/infiniband/hw/amso1100/ |
H A D | c2_qp.c | 120 void c2_set_qp_state(struct c2_qp *qp, int c2_state) argument 124 pr_debug("%s: qp[%p] state modify %s --> %s\n", 126 qp, 127 to_ib_state_str(qp->state), 129 qp->state = new_state; 134 int c2_qp_modify(struct c2_dev *c2dev, struct c2_qp *qp, argument 144 pr_debug("%s:%d qp=%p, %s --> %s\n", 146 qp, 147 to_ib_state_str(qp->state), 157 wr.qp_handle = qp 253 c2_qp_set_read_limits(struct c2_dev *c2dev, struct c2_qp *qp, int ord, int ird) argument 302 destroy_qp(struct c2_dev *c2dev, struct c2_qp *qp) argument 381 c2_alloc_qpn(struct c2_dev *c2dev, struct c2_qp *qp) argument 405 struct c2_qp *qp; local 413 c2_alloc_qp(struct c2_dev *c2dev, struct c2_pd *pd, struct ib_qp_init_attr *qp_attrs, struct c2_qp *qp) argument 598 c2_free_qp(struct c2_dev *c2dev, struct c2_qp *qp) argument 760 qp_wr_post(struct c2_mq *q, union c2wr * wr, struct c2_qp *qp, u32 size) argument 792 struct c2_qp *qp = to_c2qp(ibqp); local 946 struct c2_qp *qp = to_c2qp(ibqp); local [all...] |
H A D | c2_ae.c | 184 struct c2_qp *qp = (struct c2_qp *)resource_user_context; local 185 struct iw_cm_id *cm_id = qp->cm_id; 189 pr_debug("event received, but cm_id is <nul>, qp=%p!\n", 190 qp); 203 c2_set_qp_state(qp, be32_to_cpu(wr->ae.ae_generic.qp_state)); 218 spin_lock_irqsave(&qp->lock, flags); 219 if (qp->cm_id) { 220 qp->cm_id->rem_ref(qp->cm_id); 221 qp [all...] |
/drivers/infiniband/core/ |
H A D | verbs.c | 348 struct ib_qp *qp = context; local 350 list_for_each_entry(event->element.qp, &qp->open_list, open_list) 351 event->element.qp->event_handler(event, event->element.qp->qp_context); 354 static void __ib_insert_xrcd_qp(struct ib_xrcd *xrcd, struct ib_qp *qp) argument 357 list_add(&qp->xrcd_list, &xrcd->tgt_qp_list); 365 struct ib_qp *qp; local 368 qp = kzalloc(sizeof *qp, GFP_KERNE 390 struct ib_qp *qp, *real_qp; local 412 struct ib_qp *qp, *real_qp; local 803 ib_modify_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask) argument 811 ib_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) argument 822 ib_close_qp(struct ib_qp *qp) argument 842 __ib_destroy_shared_qp(struct ib_qp *qp) argument 870 ib_destroy_qp(struct ib_qp *qp) argument 1184 ib_attach_mcast(struct ib_qp *qp, union ib_gid *gid, u16 lid) argument 1195 ib_detach_mcast(struct ib_qp *qp, union ib_gid *gid, u16 lid) argument 1228 struct ib_qp *qp; local [all...] |
H A D | iwcm.h | 50 struct ib_qp *qp; member in struct:iwcm_id_private
|
/drivers/scsi/bnx2i/ |
H A D | bnx2i_hwi.c | 151 cq_db = (struct bnx2i_5771x_cq_db *) ep->qp.cq_pgtbl_virt; 168 cq_index = ep->qp.cqe_exp_seq_sn + next_index - 1; 169 if (cq_index > ep->qp.cqe_size * 2) 170 cq_index -= ep->qp.cqe_size * 2; 193 if (!bnx2i_conn->ep->qp.rqe_left) 196 bnx2i_conn->ep->qp.rqe_left--; 197 memcpy(ptr, (u8 *) bnx2i_conn->ep->qp.rq_cons_qe, len); 198 if (bnx2i_conn->ep->qp.rq_cons_qe == bnx2i_conn->ep->qp.rq_last_qe) { 199 bnx2i_conn->ep->qp 1964 struct qp_info *qp; local [all...] |
/drivers/net/ethernet/mellanox/mlx4/ |
H A D | qp.c | 41 #include <linux/mlx4/qp.h> 49 struct mlx4_qp *qp; local 53 qp = __mlx4_qp_lookup(dev, qpn); 54 if (qp) 55 atomic_inc(&qp->refcount); 59 if (!qp) { 64 qp->event(qp, event_type); 66 if (atomic_dec_and_test(&qp->refcount)) 67 complete(&qp 70 is_qp0(struct mlx4_dev *dev, struct mlx4_qp *qp) argument 76 __mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, enum mlx4_qp_state cur_state, enum mlx4_qp_state new_state, struct mlx4_qp_context *context, enum mlx4_qp_optpar optpar, int sqd_event, struct mlx4_qp *qp, int native) argument 169 mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, enum mlx4_qp_state cur_state, enum mlx4_qp_state new_state, struct mlx4_qp_context *context, enum mlx4_qp_optpar optpar, int sqd_event, struct mlx4_qp *qp) argument 329 mlx4_qp_alloc(struct mlx4_dev *dev, int qpn, struct mlx4_qp *qp) argument 363 mlx4_qp_remove(struct mlx4_dev *dev, struct mlx4_qp *qp) argument 374 mlx4_qp_free(struct mlx4_dev *dev, struct mlx4_qp *qp) argument 458 mlx4_qp_query(struct mlx4_dev *dev, struct mlx4_qp *qp, struct mlx4_qp_context *context) argument 479 mlx4_qp_to_ready(struct mlx4_dev *dev, struct mlx4_mtt *mtt, struct mlx4_qp_context *context, struct mlx4_qp *qp, enum mlx4_qp_state *qp_state) argument [all...] |
/drivers/infiniband/hw/cxgb4/ |
H A D | Makefile | 5 iw_cxgb4-y := device.o cm.o provider.o mem.o cq.o qp.o resource.o ev.o
|