Searched refs:qp (Results 1 - 25 of 137) sorted by relevance

123456

/drivers/infiniband/hw/ipath/
H A Dipath_rc.c56 * ipath_init_restart- initialize the qp->s_sge after a restart
57 * @qp: the QP who's SGE we're restarting
62 static void ipath_init_restart(struct ipath_qp *qp, struct ipath_swqe *wqe) argument
66 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn,
67 ib_mtu_enum_to_int(qp->path_mtu));
68 dev = to_idev(qp->ibqp.device);
70 if (list_empty(&qp->timerwait))
71 list_add_tail(&qp
86 ipath_make_rc_ack(struct ipath_ibdev *dev, struct ipath_qp *qp, struct ipath_other_headers *ohdr, u32 pmtu) argument
213 ipath_make_rc_req(struct ipath_qp *qp) argument
612 send_rc_ack(struct ipath_qp *qp) argument
724 reset_psn(struct ipath_qp *qp, u32 psn) argument
805 ipath_restart_rc(struct ipath_qp *qp, u32 psn) argument
841 update_last_psn(struct ipath_qp *qp, u32 psn) argument
857 do_rc_ack(struct ipath_qp *qp, u32 aeth, u32 psn, int opcode, u64 val) argument
1117 ipath_rc_rcv_resp(struct ipath_ibdev *dev, struct ipath_other_headers *ohdr, void *data, u32 tlen, struct ipath_qp *qp, u32 opcode, u32 psn, u32 hdrsize, u32 pmtu, int header_in_data) argument
1335 ipath_rc_rcv_error(struct ipath_ibdev *dev, struct ipath_other_headers *ohdr, void *data, struct ipath_qp *qp, u32 opcode, u32 psn, int diff, int header_in_data) argument
1525 ipath_rc_error(struct ipath_qp *qp, enum ib_wc_status err) argument
1544 ipath_update_ack_queue(struct ipath_qp *qp, unsigned n) argument
1570 ipath_rc_rcv(struct ipath_ibdev *dev, struct ipath_ib_header *hdr, int has_grh, void *data, u32 tlen, struct ipath_qp *qp) argument
[all...]
H A Dipath_qp.c203 * @qp: the QP
209 static int ipath_alloc_qpn(struct ipath_qp_table *qpt, struct ipath_qp *qp, argument
218 qp->ibqp.qp_num = ret;
224 qp->next = qpt->table[ret];
225 qpt->table[ret] = qp;
226 atomic_inc(&qp->refcount);
238 * @qp: the QP to remove
243 static void ipath_free_qp(struct ipath_qp_table *qpt, struct ipath_qp *qp) argument
251 qpp = &qpt->table[qp->ibqp.qp_num % qpt->max];
253 if (q == qp) {
274 struct ipath_qp *qp; local
304 struct ipath_qp *qp; local
324 ipath_reset_qp(struct ipath_qp *qp, enum ib_qp_type type) argument
377 ipath_error_qp(struct ipath_qp *qp, enum ib_wc_status err) argument
454 struct ipath_qp *qp = to_iqp(ibqp); local
624 struct ipath_qp *qp = to_iqp(ibqp); local
677 ipath_compute_aeth(struct ipath_qp *qp) argument
745 struct ipath_qp *qp; local
964 struct ipath_qp *qp = to_iqp(ibqp); local
1055 ipath_get_credit(struct ipath_qp *qp, u32 aeth) argument
[all...]
H A Dipath_ruc.c80 * @qp: the QP
87 void ipath_insert_rnr_queue(struct ipath_qp *qp) argument
89 struct ipath_ibdev *dev = to_idev(qp->ibqp.device);
94 list_add(&qp->timerwait, &dev->rnrwait);
100 while (qp->s_rnr_timeout >= nqp->s_rnr_timeout) {
101 qp->s_rnr_timeout -= nqp->s_rnr_timeout;
111 nqp->s_rnr_timeout -= qp->s_rnr_timeout;
112 list_add(&qp->timerwait, l);
119 * @qp: the QP
123 int ipath_init_sge(struct ipath_qp *qp, struc argument
166 ipath_get_rwqe(struct ipath_qp *qp, int wr_id_only) argument
263 struct ipath_qp *qp; local
517 want_buffer(struct ipath_devdata *dd, struct ipath_qp *qp) argument
541 ipath_no_bufs_available(struct ipath_qp *qp, struct ipath_ibdev *dev) argument
600 ipath_make_ruc_header(struct ipath_ibdev *dev, struct ipath_qp *qp, struct ipath_other_headers *ohdr, u32 bth0, u32 bth2) argument
641 struct ipath_qp *qp = (struct ipath_qp *)data; local
699 ipath_send_complete(struct ipath_qp *qp, struct ipath_swqe *wqe, enum ib_wc_status status) argument
[all...]
H A Dipath_uc.c42 * @qp: a pointer to the QP
46 int ipath_make_uc_req(struct ipath_qp *qp) argument
54 u32 pmtu = ib_mtu_enum_to_int(qp->path_mtu);
57 spin_lock_irqsave(&qp->s_lock, flags);
59 if (!(ib_ipath_state_ops[qp->state] & IPATH_PROCESS_SEND_OK)) {
60 if (!(ib_ipath_state_ops[qp->state] & IPATH_FLUSH_SEND))
63 if (qp->s_last == qp->s_head)
66 if (atomic_read(&qp->s_dma_busy)) {
67 qp
240 ipath_uc_rcv(struct ipath_ibdev *dev, struct ipath_ib_header *hdr, int has_grh, void *data, u32 tlen, struct ipath_qp *qp) argument
[all...]
H A Dipath_ud.c53 struct ipath_qp *qp; local
68 qp = ipath_lookup_qpn(&dev->qp_table, swqe->wr.wr.ud.remote_qpn);
69 if (!qp || !(ib_ipath_state_ops[qp->state] & IPATH_PROCESS_RECV_OK)) {
79 if (unlikely(qp->ibqp.qp_num &&
81 sqp->qkey : swqe->wr.wr.ud.remote_qkey) != qp->qkey)) {
107 if (qp->ibqp.srq) {
108 srq = to_isrq(qp->ibqp.srq);
114 rq = &qp->r_rq;
134 rsge.sg_list = qp
241 ipath_make_ud_req(struct ipath_qp *qp) argument
409 ipath_ud_rcv(struct ipath_ibdev *dev, struct ipath_ib_header *hdr, int has_grh, void *data, u32 tlen, struct ipath_qp *qp) argument
[all...]
H A Dipath_verbs.c333 * @qp: the QP to post on
336 static int ipath_post_one_send(struct ipath_qp *qp, struct ib_send_wr *wr) argument
345 struct ipath_devdata *dd = to_idev(qp->ibqp.device)->dd;
347 spin_lock_irqsave(&qp->s_lock, flags);
349 if (qp->ibqp.qp_type != IB_QPT_SMI &&
356 if (unlikely(!(ib_ipath_state_ops[qp->state] & IPATH_POST_SEND_OK)))
360 if (wr->num_sge > qp->s_max_sge)
368 if (qp->ibqp.qp_type == IB_QPT_UC) {
371 } else if (qp->ibqp.qp_type == IB_QPT_UD) {
377 if (qp
448 struct ipath_qp *qp = to_iqp(ibqp); local
477 struct ipath_qp *qp = to_iqp(ibqp); local
540 ipath_qp_rcv(struct ipath_ibdev *dev, struct ipath_ib_header *hdr, int has_grh, void *data, u32 tlen, struct ipath_qp *qp) argument
588 struct ipath_qp *qp; local
682 struct ipath_qp *qp; local
1031 struct ipath_qp *qp = tx->qp; local
1061 decrement_dma_busy(struct ipath_qp *qp) argument
1097 ipath_verbs_send_dma(struct ipath_qp *qp, struct ipath_ib_header *hdr, u32 hdrwords, struct ipath_sge_state *ss, u32 len, u32 plen, u32 dwords) argument
1222 ipath_verbs_send_pio(struct ipath_qp *qp, struct ipath_ib_header *ibhdr, u32 hdrwords, struct ipath_sge_state *ss, u32 len, u32 plen, u32 dwords) argument
1322 ipath_verbs_send(struct ipath_qp *qp, struct ipath_ib_header *hdr, u32 hdrwords, struct ipath_sge_state *ss, u32 len) argument
1461 struct ipath_qp *qp; local
[all...]
/drivers/infiniband/hw/qib/
H A Dqib_qp.c219 static void insert_qp(struct qib_ibdev *dev, struct qib_qp *qp) argument
221 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num);
223 unsigned n = qpn_hash(dev, qp->ibqp.qp_num);
226 atomic_inc(&qp->refcount);
228 if (qp->ibqp.qp_num == 0)
229 rcu_assign_pointer(ibp->qp0, qp);
230 else if (qp->ibqp.qp_num == 1)
231 rcu_assign_pointer(ibp->qp1, qp);
233 qp
245 remove_qp(struct qib_ibdev *dev, struct qib_qp *qp) argument
287 struct qib_qp *qp; local
327 struct qib_qp *qp = NULL; local
357 qib_reset_qp(struct qib_qp *qp, enum ib_qp_type type) argument
404 clear_mr_refs(struct qib_qp *qp, int clr_sends) argument
469 qib_error_qp(struct qib_qp *qp, enum ib_wc_status err) argument
568 struct qib_qp *qp = to_iqp(ibqp); local
853 struct qib_qp *qp = to_iqp(ibqp); local
906 qib_compute_aeth(struct qib_qp *qp) argument
974 struct qib_qp *qp; local
1180 struct qib_qp *qp = to_iqp(ibqp); local
1253 qib_get_credit(struct qib_qp *qp, u32 aeth) argument
[all...]
H A Dqib_rc.c57 static void start_timer(struct qib_qp *qp) argument
59 qp->s_flags |= QIB_S_TIMER;
60 qp->s_timer.function = rc_timeout;
61 /* 4.096 usec. * (1 << qp->timeout) */
62 qp->s_timer.expires = jiffies + qp->timeout_jiffies;
63 add_timer(&qp->s_timer);
69 * @qp: a pointer to the QP
77 static int qib_make_rc_ack(struct qib_ibdev *dev, struct qib_qp *qp, argument
87 if (!(ib_qib_state_ops[qp
231 qib_make_rc_req(struct qib_qp *qp) argument
648 qib_send_rc_ack(struct qib_qp *qp) argument
783 reset_psn(struct qib_qp *qp, u32 psn) argument
868 qib_restart_rc(struct qib_qp *qp, u32 psn, int wait) argument
905 struct qib_qp *qp = (struct qib_qp *)arg; local
928 struct qib_qp *qp = (struct qib_qp *)arg; local
944 reset_sending_psn(struct qib_qp *qp, u32 psn) argument
969 qib_rc_send_complete(struct qib_qp *qp, struct qib_ib_header *hdr) argument
1044 update_last_psn(struct qib_qp *qp, u32 psn) argument
1054 do_rc_completion(struct qib_qp *qp, struct qib_swqe *wqe, struct qib_ibport *ibp) argument
1127 do_rc_ack(struct qib_qp *qp, u32 aeth, u32 psn, int opcode, u64 val, struct qib_ctxtdata *rcd) argument
1350 rdma_seq_err(struct qib_qp *qp, struct qib_ibport *ibp, u32 psn, struct qib_ctxtdata *rcd) argument
1397 qib_rc_rcv_resp(struct qib_ibport *ibp, struct qib_other_headers *ohdr, void *data, u32 tlen, struct qib_qp *qp, u32 opcode, u32 psn, u32 hdrsize, u32 pmtu, struct qib_ctxtdata *rcd) argument
1622 qib_rc_rcv_error(struct qib_other_headers *ohdr, void *data, struct qib_qp *qp, u32 opcode, u32 psn, int diff, struct qib_ctxtdata *rcd) argument
1819 qib_rc_error(struct qib_qp *qp, enum ib_wc_status err) argument
1838 qib_update_ack_queue(struct qib_qp *qp, unsigned n) argument
1862 qib_rc_rcv(struct qib_ctxtdata *rcd, struct qib_ib_header *hdr, int has_grh, void *data, u32 tlen, struct qib_qp *qp) argument
[all...]
H A Dqib_uc.c42 * @qp: a pointer to the QP
46 int qib_make_uc_req(struct qib_qp *qp) argument
54 u32 pmtu = qp->pmtu;
57 spin_lock_irqsave(&qp->s_lock, flags);
59 if (!(ib_qib_state_ops[qp->state] & QIB_PROCESS_SEND_OK)) {
60 if (!(ib_qib_state_ops[qp->state] & QIB_FLUSH_SEND))
63 if (qp->s_last == qp->s_head)
66 if (atomic_read(&qp->s_dma_busy)) {
67 qp
242 qib_uc_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr, int has_grh, void *data, u32 tlen, struct qib_qp *qp) argument
[all...]
H A Dqib_ud.c53 struct qib_qp *qp; local
61 qp = qib_lookup_qpn(ibp, swqe->wr.wr.ud.remote_qpn);
62 if (!qp) {
66 if (qp->ibqp.qp_type != sqp->ibqp.qp_type ||
67 !(ib_qib_state_ops[qp->state] & QIB_PROCESS_RECV_OK)) {
75 if (qp->ibqp.qp_num > 1) {
81 pkey2 = qib_get_pkey(ibp, qp->s_pkey_index);
87 sqp->ibqp.qp_num, qp->ibqp.qp_num,
99 if (qp->ibqp.qp_num) {
104 if (unlikely(qkey != qp
232 qib_make_ud_req(struct qib_qp *qp) argument
425 qib_ud_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr, int has_grh, void *data, u32 tlen, struct qib_qp *qp) argument
[all...]
H A Dqib_ruc.c81 static int qib_init_sge(struct qib_qp *qp, struct qib_rwqe *wqe) argument
89 rkt = &to_idev(qp->ibqp.device)->lk_table;
90 pd = to_ipd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd);
91 ss = &qp->r_sge;
92 ss->sg_list = qp->r_sg_list;
93 qp->r_len = 0;
101 qp->r_len += wqe->sg_list[i].length;
105 ss->total_len = qp
138 qib_get_rwqe(struct qib_qp *qp, int wr_id_only) argument
229 qib_migrate_qp(struct qib_qp *qp) argument
267 qib_ruc_check_hdr(struct qib_ibport *ibp, struct qib_ib_header *hdr, int has_grh, struct qib_qp *qp, u32 bth0) argument
358 struct qib_qp *qp; local
678 qib_make_ruc_header(struct qib_qp *qp, struct qib_other_headers *ohdr, u32 bth0, u32 bth2) argument
722 struct qib_qp *qp = container_of(work, struct qib_qp, s_work); local
773 qib_send_complete(struct qib_qp *qp, struct qib_swqe *wqe, enum ib_wc_status status) argument
[all...]
H A Dqib_verbs_mcast.c40 * @qp: the QP to link
42 static struct qib_mcast_qp *qib_mcast_qp_alloc(struct qib_qp *qp) argument
50 mqp->qp = qp;
51 atomic_inc(&qp->refcount);
59 struct qib_qp *qp = mqp->qp; local
62 if (atomic_dec_and_test(&qp->refcount))
63 wake_up(&qp->wait);
182 if (p->qp
227 struct qib_qp *qp = to_iqp(ibqp); local
285 struct qib_qp *qp = to_iqp(ibqp); local
[all...]
H A Dqib_verbs.c333 * @qp: the QP to post on
336 static int qib_post_one_send(struct qib_qp *qp, struct ib_send_wr *wr) argument
348 spin_lock_irqsave(&qp->s_lock, flags);
351 if (unlikely(!(ib_qib_state_ops[qp->state] & QIB_POST_SEND_OK)))
355 if (wr->num_sge > qp->s_max_sge)
364 if (qib_fast_reg_mr(qp, wr))
366 } else if (qp->ibqp.qp_type == IB_QPT_UC) {
369 } else if (qp->ibqp.qp_type != IB_QPT_RC) {
375 if (qp->ibqp.pd != wr->wr.ud.ah->pd)
384 else if (wr->opcode >= IB_WR_RDMA_READ && !qp
458 struct qib_qp *qp = to_iqp(ibqp); local
487 struct qib_qp *qp = to_iqp(ibqp); local
550 qib_qp_rcv(struct qib_ctxtdata *rcd, struct qib_ib_header *hdr, int has_grh, void *data, u32 tlen, struct qib_qp *qp) argument
605 struct qib_qp *qp; local
698 struct qib_qp *qp = NULL; local
916 __get_txreq(struct qib_ibdev *dev, struct qib_qp *qp) argument
947 get_txreq(struct qib_ibdev *dev, struct qib_qp *qp) argument
972 struct qib_qp *qp; local
1025 struct qib_qp *qp, *nqp; local
1070 struct qib_qp *qp = tx->qp; local
1100 wait_kmem(struct qib_ibdev *dev, struct qib_qp *qp) argument
1123 qib_verbs_send_dma(struct qib_qp *qp, struct qib_ib_header *hdr, u32 hdrwords, struct qib_sge_state *ss, u32 len, u32 plen, u32 dwords) argument
1227 no_bufs_available(struct qib_qp *qp) argument
1258 qib_verbs_send_pio(struct qib_qp *qp, struct qib_ib_header *ibhdr, u32 hdrwords, struct qib_sge_state *ss, u32 len, u32 plen, u32 dwords) argument
1365 qib_verbs_send(struct qib_qp *qp, struct qib_ib_header *hdr, u32 hdrwords, struct qib_sge_state *ss, u32 len) argument
1499 struct qib_qp *qp; local
[all...]
/drivers/infiniband/hw/mthca/
H A Dmthca_qp.c195 static int is_sqp(struct mthca_dev *dev, struct mthca_qp *qp) argument
197 return qp->qpn >= dev->qp_table.sqp_start &&
198 qp->qpn <= dev->qp_table.sqp_start + 3;
201 static int is_qp0(struct mthca_dev *dev, struct mthca_qp *qp) argument
203 return qp->qpn >= dev->qp_table.sqp_start &&
204 qp->qpn <= dev->qp_table.sqp_start + 1;
207 static void *get_recv_wqe(struct mthca_qp *qp, int n) argument
209 if (qp->is_direct)
210 return qp->queue.direct.buf + (n << qp
216 get_send_wqe(struct mthca_qp *qp, int n) argument
240 struct mthca_qp *qp; local
326 get_hw_access_flags(struct mthca_qp *qp, const struct ib_qp_attr *attr, int attr_mask) argument
427 struct mthca_qp *qp = to_mqp(ibqp); local
546 struct mthca_qp *qp = to_mqp(ibqp); local
844 struct mthca_qp *qp = to_mqp(ibqp); local
908 mthca_max_data_size(struct mthca_dev *dev, struct mthca_qp *qp, int desc_sz) argument
942 mthca_adjust_qp_caps(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_qp *qp) argument
967 mthca_alloc_wqe_buf(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_qp *qp) argument
1059 mthca_free_wqe_buf(struct mthca_dev *dev, struct mthca_qp *qp) argument
1068 mthca_map_memfree(struct mthca_dev *dev, struct mthca_qp *qp) argument
1100 mthca_unmap_memfree(struct mthca_dev *dev, struct mthca_qp *qp) argument
1109 mthca_alloc_memfree(struct mthca_dev *dev, struct mthca_qp *qp) argument
1129 mthca_free_memfree(struct mthca_dev *dev, struct mthca_qp *qp) argument
1138 mthca_alloc_qp_common(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct mthca_qp *qp) argument
1227 mthca_set_qp_size(struct mthca_dev *dev, struct ib_qp_cap *cap, struct mthca_pd *pd, struct mthca_qp *qp) argument
1266 mthca_alloc_qp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_qp_type type, enum ib_sig_type send_policy, struct ib_qp_cap *cap, struct mthca_qp *qp) argument
1407 get_qp_refcount(struct mthca_dev *dev, struct mthca_qp *qp) argument
1418 mthca_free_qp(struct mthca_dev *dev, struct mthca_qp *qp) argument
1603 struct mthca_qp *qp = to_mqp(ibqp); local
1806 struct mthca_qp *qp = to_mqp(ibqp); local
1917 struct mthca_qp *qp = to_mqp(ibqp); local
2157 struct mthca_qp *qp = to_mqp(ibqp); local
2224 mthca_free_err_wqe(struct mthca_dev *dev, struct mthca_qp *qp, int is_send, int index, int *dbd, __be32 *new_wqe) argument
[all...]
/drivers/infiniband/hw/mlx4/
H A Dqp.c42 #include <linux/mlx4/qp.h>
74 struct mlx4_ib_qp qp; member in struct:mlx4_ib_sqp
105 return container_of(mqp, struct mlx4_ib_sqp, qp);
108 static int is_sqp(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp) argument
110 return qp->mqp.qpn >= dev->dev->caps.sqp_start &&
111 qp->mqp.qpn <= dev->dev->caps.sqp_start + 3;
114 static int is_qp0(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp) argument
116 return qp->mqp.qpn >= dev->dev->caps.sqp_start &&
117 qp->mqp.qpn <= dev->dev->caps.sqp_start + 1;
120 static void *get_wqe(struct mlx4_ib_qp *qp, in argument
125 get_recv_wqe(struct mlx4_ib_qp *qp, int n) argument
130 get_send_wqe(struct mlx4_ib_qp *qp, int n) argument
144 stamp_send_wqe(struct mlx4_ib_qp *qp, int n, int size) argument
174 post_nop_wqe(struct mlx4_ib_qp *qp, int n, int size) argument
212 pad_wraparound(struct mlx4_ib_qp *qp, int ind) argument
222 mlx4_ib_qp_event(struct mlx4_qp *qp, enum mlx4_event type) argument
304 set_rq_size(struct mlx4_ib_dev *dev, struct ib_qp_cap *cap, int is_user, int has_rq, struct mlx4_ib_qp *qp) argument
333 set_kernel_sq_size(struct mlx4_ib_dev *dev, struct ib_qp_cap *cap, enum ib_qp_type type, struct mlx4_ib_qp *qp) argument
445 set_user_sq_size(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp, struct mlx4_ib_create_qp *ucmd) argument
473 create_qp_common(struct mlx4_ib_dev *dev, struct ib_pd *pd, struct ib_qp_init_attr *init_attr, struct ib_udata *udata, int sqpn, struct mlx4_ib_qp *qp) argument
673 del_gid_entries(struct mlx4_ib_qp *qp) argument
683 get_pd(struct mlx4_ib_qp *qp) argument
691 get_cqs(struct mlx4_ib_qp *qp, struct mlx4_ib_cq **send_cq, struct mlx4_ib_cq **recv_cq) argument
710 destroy_qp_common(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp, int is_user) argument
764 struct mlx4_ib_qp *qp; local
846 mlx4_ib_destroy_qp(struct ib_qp *qp) argument
880 to_mlx4_access_flags(struct mlx4_ib_qp *qp, const struct ib_qp_attr *attr, int attr_mask) argument
994 update_mcg_macs(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp) argument
1011 struct mlx4_ib_qp *qp = to_mqp(ibqp); local
1305 struct mlx4_ib_qp *qp = to_mqp(ibqp); local
1658 build_lso_seg(struct mlx4_wqe_lso_seg *wqe, struct ib_send_wr *wr, struct mlx4_ib_qp *qp, unsigned *lso_seg_len, __be32 *lso_hdr_sz, __be32 *blh) argument
1697 struct mlx4_ib_qp *qp = to_mqp(ibqp); local
1946 struct mlx4_ib_qp *qp = to_mqp(ibqp); local
2084 struct mlx4_ib_qp *qp = to_mqp(ibqp); local
[all...]
H A DMakefile3 mlx4_ib-y := ah.o cq.o doorbell.o mad.o main.o mr.o qp.o srq.o
/drivers/infiniband/hw/ehca/
H A Dehca_uverbs.c198 static int ehca_mmap_qp(struct vm_area_struct *vma, struct ehca_qp *qp, argument
205 ehca_dbg(qp->ib_qp.device, "qp_num=%x fw", qp->ib_qp.qp_num);
206 ret = ehca_mmap_fw(vma, &qp->galpas, &qp->mm_count_galpa);
208 ehca_err(qp->ib_qp.device,
210 ret, qp->ib_qp.qp_num);
215 case 1: /* qp rqueue_addr */
216 ehca_dbg(qp->ib_qp.device, "qp_num=%x rq", qp
256 struct ehca_qp *qp; local
[all...]
H A Dhipz_fns_core.h61 static inline void hipz_update_sqa(struct ehca_qp *qp, u16 nr_wqes) argument
64 hipz_galpa_store_qp(qp->galpas.kernel, qpx_sqa,
68 static inline void hipz_update_rqa(struct ehca_qp *qp, u16 nr_wqes) argument
71 hipz_galpa_store_qp(qp->galpas.kernel, qpx_rqa,
/drivers/infiniband/hw/amso1100/
H A Dc2_qp.c120 void c2_set_qp_state(struct c2_qp *qp, int c2_state) argument
124 pr_debug("%s: qp[%p] state modify %s --> %s\n",
126 qp,
127 to_ib_state_str(qp->state),
129 qp->state = new_state;
134 int c2_qp_modify(struct c2_dev *c2dev, struct c2_qp *qp, argument
144 pr_debug("%s:%d qp=%p, %s --> %s\n",
146 qp,
147 to_ib_state_str(qp->state),
157 wr.qp_handle = qp
253 c2_qp_set_read_limits(struct c2_dev *c2dev, struct c2_qp *qp, int ord, int ird) argument
302 destroy_qp(struct c2_dev *c2dev, struct c2_qp *qp) argument
381 c2_alloc_qpn(struct c2_dev *c2dev, struct c2_qp *qp) argument
405 struct c2_qp *qp; local
413 c2_alloc_qp(struct c2_dev *c2dev, struct c2_pd *pd, struct ib_qp_init_attr *qp_attrs, struct c2_qp *qp) argument
598 c2_free_qp(struct c2_dev *c2dev, struct c2_qp *qp) argument
760 qp_wr_post(struct c2_mq *q, union c2wr * wr, struct c2_qp *qp, u32 size) argument
792 struct c2_qp *qp = to_c2qp(ibqp); local
946 struct c2_qp *qp = to_c2qp(ibqp); local
[all...]
H A Dc2_ae.c184 struct c2_qp *qp = (struct c2_qp *)resource_user_context; local
185 struct iw_cm_id *cm_id = qp->cm_id;
189 pr_debug("event received, but cm_id is <nul>, qp=%p!\n",
190 qp);
203 c2_set_qp_state(qp, be32_to_cpu(wr->ae.ae_generic.qp_state));
218 spin_lock_irqsave(&qp->lock, flags);
219 if (qp->cm_id) {
220 qp->cm_id->rem_ref(qp->cm_id);
221 qp
[all...]
/drivers/infiniband/core/
H A Dverbs.c348 struct ib_qp *qp = context; local
350 list_for_each_entry(event->element.qp, &qp->open_list, open_list)
351 event->element.qp->event_handler(event, event->element.qp->qp_context);
354 static void __ib_insert_xrcd_qp(struct ib_xrcd *xrcd, struct ib_qp *qp) argument
357 list_add(&qp->xrcd_list, &xrcd->tgt_qp_list);
365 struct ib_qp *qp; local
368 qp = kzalloc(sizeof *qp, GFP_KERNE
390 struct ib_qp *qp, *real_qp; local
412 struct ib_qp *qp, *real_qp; local
803 ib_modify_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask) argument
811 ib_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr) argument
822 ib_close_qp(struct ib_qp *qp) argument
842 __ib_destroy_shared_qp(struct ib_qp *qp) argument
870 ib_destroy_qp(struct ib_qp *qp) argument
1184 ib_attach_mcast(struct ib_qp *qp, union ib_gid *gid, u16 lid) argument
1195 ib_detach_mcast(struct ib_qp *qp, union ib_gid *gid, u16 lid) argument
1228 struct ib_qp *qp; local
[all...]
H A Diwcm.h50 struct ib_qp *qp; member in struct:iwcm_id_private
/drivers/scsi/bnx2i/
H A Dbnx2i_hwi.c151 cq_db = (struct bnx2i_5771x_cq_db *) ep->qp.cq_pgtbl_virt;
168 cq_index = ep->qp.cqe_exp_seq_sn + next_index - 1;
169 if (cq_index > ep->qp.cqe_size * 2)
170 cq_index -= ep->qp.cqe_size * 2;
193 if (!bnx2i_conn->ep->qp.rqe_left)
196 bnx2i_conn->ep->qp.rqe_left--;
197 memcpy(ptr, (u8 *) bnx2i_conn->ep->qp.rq_cons_qe, len);
198 if (bnx2i_conn->ep->qp.rq_cons_qe == bnx2i_conn->ep->qp.rq_last_qe) {
199 bnx2i_conn->ep->qp
1964 struct qp_info *qp; local
[all...]
/drivers/net/ethernet/mellanox/mlx4/
H A Dqp.c41 #include <linux/mlx4/qp.h>
49 struct mlx4_qp *qp; local
53 qp = __mlx4_qp_lookup(dev, qpn);
54 if (qp)
55 atomic_inc(&qp->refcount);
59 if (!qp) {
64 qp->event(qp, event_type);
66 if (atomic_dec_and_test(&qp->refcount))
67 complete(&qp
70 is_qp0(struct mlx4_dev *dev, struct mlx4_qp *qp) argument
76 __mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, enum mlx4_qp_state cur_state, enum mlx4_qp_state new_state, struct mlx4_qp_context *context, enum mlx4_qp_optpar optpar, int sqd_event, struct mlx4_qp *qp, int native) argument
169 mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, enum mlx4_qp_state cur_state, enum mlx4_qp_state new_state, struct mlx4_qp_context *context, enum mlx4_qp_optpar optpar, int sqd_event, struct mlx4_qp *qp) argument
329 mlx4_qp_alloc(struct mlx4_dev *dev, int qpn, struct mlx4_qp *qp) argument
363 mlx4_qp_remove(struct mlx4_dev *dev, struct mlx4_qp *qp) argument
374 mlx4_qp_free(struct mlx4_dev *dev, struct mlx4_qp *qp) argument
458 mlx4_qp_query(struct mlx4_dev *dev, struct mlx4_qp *qp, struct mlx4_qp_context *context) argument
479 mlx4_qp_to_ready(struct mlx4_dev *dev, struct mlx4_mtt *mtt, struct mlx4_qp_context *context, struct mlx4_qp *qp, enum mlx4_qp_state *qp_state) argument
[all...]
/drivers/infiniband/hw/cxgb4/
H A DMakefile5 iw_cxgb4-y := device.o cm.o provider.o mem.o cq.o qp.o resource.o ev.o

Completed in 364 milliseconds

123456