/drivers/net/ethernet/mellanox/mlx4/ |
H A D | qp.c | 46 void mlx4_qp_event(struct mlx4_dev *dev, u32 qpn, int event_type) argument 53 qp = __mlx4_qp_lookup(dev, qpn); 60 mlx4_dbg(dev, "Async event for none existent QP %08x\n", qpn); 72 return qp->qpn >= dev->caps.sqp_start && 73 qp->qpn <= dev->caps.sqp_start + 1; 132 ret = mlx4_cmd(dev, 0, qp->qpn, 2, 137 port = (qp->qpn & 1) + 1; 158 cpu_to_be32(qp->qpn); 161 qp->qpn | (!!sqd_event << 31), 246 int __mlx4_qp_alloc_icm(struct mlx4_dev *dev, int qpn) argument 290 mlx4_qp_alloc_icm(struct mlx4_dev *dev, int qpn) argument 303 __mlx4_qp_free_icm(struct mlx4_dev *dev, int qpn) argument 315 mlx4_qp_free_icm(struct mlx4_dev *dev, int qpn) argument 329 mlx4_qp_alloc(struct mlx4_dev *dev, int qpn, struct mlx4_qp *qp) argument [all...] |
H A D | mcg.c | 108 u32 qpn) 114 if (pqp->qpn == qpn) 127 unsigned int index, u32 qpn) 148 /* If the given qpn is also a promisc qp, 151 pqp = get_promisc_qp(dev, 0, steer, qpn); 158 dqp->qpn = qpn; 183 /* don't add already existing qpn */ 184 if (pqp->qpn 106 get_promisc_qp(struct mlx4_dev *dev, u8 pf_num, enum mlx4_steer_type steer, u32 qpn) argument 125 new_steering_entry(struct mlx4_dev *dev, u8 port, enum mlx4_steer_type steer, unsigned int index, u32 qpn) argument 214 existing_steering_entry(struct mlx4_dev *dev, u8 port, enum mlx4_steer_type steer, unsigned int index, u32 qpn) argument 260 check_duplicate_entry(struct mlx4_dev *dev, u8 port, enum mlx4_steer_type steer, unsigned int index, u32 qpn) argument 304 u32 qpn; local 347 add_promisc_qp(struct mlx4_dev *dev, u8 port, enum mlx4_steer_type steer, u32 qpn) argument 453 remove_promisc_qp(struct mlx4_dev *dev, u8 port, enum mlx4_steer_type steer, u32 qpn) argument 847 int qpn; local 949 u32 qpn = (u32) vhcr->in_param & 0xffffffff; local 963 mlx4_PROMISC(struct mlx4_dev *dev, u32 qpn, enum mlx4_steer_type steer, u8 add, u8 port) argument 971 mlx4_multicast_promisc_add(struct mlx4_dev *dev, u32 qpn, u8 port) argument 983 mlx4_multicast_promisc_remove(struct mlx4_dev *dev, u32 qpn, u8 port) argument 995 mlx4_unicast_promisc_add(struct mlx4_dev *dev, u32 qpn, u8 port) argument 1007 mlx4_unicast_promisc_remove(struct mlx4_dev *dev, u32 qpn, u8 port) argument [all...] |
H A D | port.c | 78 static int mlx4_uc_steer_add(struct mlx4_dev *dev, u8 port, u64 mac, int *qpn) argument 85 qp.qpn = *qpn; 100 u64 mac, int qpn) 106 qp.qpn = qpn; 141 int mlx4_get_eth_qp(struct mlx4_dev *dev, u8 port, u64 mac, int *qpn) argument 159 *qpn = info->base_qpn + index; 163 err = mlx4_qp_reserve_range(dev, 1, 1, qpn); 164 mlx4_dbg(dev, "Reserved qp %d\n", *qpn); 99 mlx4_uc_steer_release(struct mlx4_dev *dev, u8 port, u64 mac, int qpn) argument 200 mlx4_put_eth_qp(struct mlx4_dev *dev, u8 port, u64 mac, int qpn) argument 355 mlx4_replace_mac(struct mlx4_dev *dev, u8 port, int qpn, u64 new_mac) argument [all...] |
H A D | en_resources.c | 41 int is_tx, int rss, int qpn, int cqn, 57 context->local_qpn = cpu_to_be32(qpn); 40 mlx4_en_fill_qp_context(struct mlx4_en_priv *priv, int size, int stride, int is_tx, int rss, int qpn, int cqn, struct mlx4_qp_context *context) argument
|
H A D | resource_tracker.c | 710 static int qp_res_start_move_to(struct mlx4_dev *dev, int slave, int qpn, argument 720 r = radix_tree_lookup(&tracker->res_tree[RES_QP], qpn); 1008 static int valid_reserved(struct mlx4_dev *dev, int slave, int qpn) argument 1010 return mlx4_is_qp_reserved(dev, qpn); 1020 int qpn; local 1038 qpn = get_param_l(&in_param) & 0x7fffff; 1039 if (valid_reserved(dev, slave, qpn)) { 1040 err = add_res_range(dev, slave, qpn, 1, RES_QP, 0); 1045 err = qp_res_start_move_to(dev, slave, qpn, RES_QP_MAPPED, 1050 if (!valid_reserved(dev, slave, qpn)) { 1331 int qpn; local 1749 int qpn = vhcr->in_modifier & 0x7fffff; local 2409 int qpn = vhcr->in_modifier & 0x7fffff; local 2446 int qpn = vhcr->in_modifier & 0x7fffff; local 2540 int qpn; local 2685 int qpn; local [all...] |
H A D | en_tx.c | 57 struct mlx4_en_tx_ring *ring, int qpn, u32 size, 106 ring->qpn = qpn; 107 err = mlx4_qp_alloc(mdev->dev, ring->qpn, &ring->qp); 109 en_err(priv, "Failed allocating qp %d\n", ring->qpn); 142 en_dbg(DRV, priv, "Destroying tx ring, qpn: %d\n", ring->qpn); 148 mlx4_qp_release_range(mdev->dev, ring->qpn, 1); 174 ring->doorbell_qpn = ring->qp.qpn << 8; 176 mlx4_en_fill_qp_context(priv, ring->size, ring->stride, 1, 0, ring->qpn, 56 mlx4_en_create_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, int qpn, u32 size, u16 stride) argument [all...] |
H A D | en_rx.c | 802 static int mlx4_en_config_rss_qp(struct mlx4_en_priv *priv, int qpn, argument 817 err = mlx4_qp_alloc(mdev->dev, qpn, qp); 819 en_err(priv, "Failed to allocate qp #%x\n", qpn); 826 qpn, ring->cqn, context); 858 int i, qpn; local 875 qpn = rss_map->base_qpn + i; 876 err = mlx4_en_config_rss_qp(priv, qpn, &priv->rx_ring[i],
|
H A D | mlx4.h | 236 __be32 qpn; member in struct:mlx4_mpt_entry 335 __be32 qpn; member in struct:mlx4_eqe::__anon2677::__anon2680 446 u32 qpn; member in struct:mlx4_promisc_qp 798 int __mlx4_qp_alloc_icm(struct mlx4_dev *dev, int qpn); 799 void __mlx4_qp_free_icm(struct mlx4_dev *dev, int qpn); 851 int __mlx4_replace_mac(struct mlx4_dev *dev, u8 port, int qpn, u64 new_mac); 980 void mlx4_qp_event(struct mlx4_dev *dev, u32 qpn, int event_type);
|
/drivers/infiniband/core/ |
H A D | agent.h | 49 int port_num, int qpn);
|
H A D | cm_msgs.h | 127 static inline void cm_req_set_local_qpn(struct cm_req_msg *req_msg, __be32 qpn) argument 129 req_msg->offset32 = cpu_to_be32((be32_to_cpu(qpn) << 8) | 535 static inline void cm_rep_set_local_qpn(struct cm_rep_msg *rep_msg, __be32 qpn) argument 537 rep_msg->offset12 = cpu_to_be32((be32_to_cpu(qpn) << 8) | 655 static inline void cm_dreq_set_remote_qpn(struct cm_dreq_msg *dreq_msg, __be32 qpn) argument 657 dreq_msg->offset8 = cpu_to_be32((be32_to_cpu(qpn) << 8) | 704 static inline void cm_lap_set_remote_qpn(struct cm_lap_msg *lap_msg, __be32 qpn) argument 706 lap_msg->offset12 = cpu_to_be32((be32_to_cpu(qpn) << 8) | 841 __be32 qpn) 843 sidr_rep_msg->offset8 = cpu_to_be32((be32_to_cpu(qpn) << 840 cm_sidr_rep_set_qpn(struct cm_sidr_rep_msg *sidr_rep_msg, __be32 qpn) argument [all...] |
H A D | agent.c | 83 int port_num, int qpn) 101 agent = port_priv->agent[qpn]; 81 agent_send_response(struct ib_mad *mad, struct ib_grh *grh, struct ib_wc *wc, struct ib_device *device, int port_num, int qpn) argument
|
/drivers/infiniband/hw/qib/ |
H A D | qib_qp.c | 123 u32 i, offset, max_scan, qpn; local 141 qpn = qpt->last + 2; 142 if (qpn >= QPN_MAX) 143 qpn = 2; 144 if (qpt->mask && ((qpn & qpt->mask) >> 1) >= dd->n_krcv_queues) 145 qpn = (qpn | qpt->mask) + 2; 146 offset = qpn & BITS_PER_PAGE_MASK; 147 map = &qpt->map[qpn / BITS_PER_PAGE]; 157 qpt->last = qpn; 199 free_qpn(struct qib_qpn_table *qpt, u32 qpn) argument 208 qpn_hash(struct qib_ibdev *dev, u32 qpn) argument 325 qib_lookup_qpn(struct qib_ibport *ibp, u32 qpn) argument [all...] |
/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 197 return qp->qpn >= dev->qp_table.sqp_start && 198 qp->qpn <= dev->qp_table.sqp_start + 3; 203 return qp->qpn >= dev->qp_table.sqp_start && 204 qp->qpn <= dev->qp_table.sqp_start + 1; 237 void mthca_qp_event(struct mthca_dev *dev, u32 qpn, argument 244 qp = mthca_array_get(&dev->qp_table.qp, qpn & (dev->limits.num_qps - 1)); 250 mthca_warn(dev, "Async event for bogus QP %08x\n", qpn); 447 err = mthca_QUERY_QP(dev, qp->qpn, 0, mailbox); 612 qp_context->local_qpn = cpu_to_be32(qp->qpn); 753 ((qp->qpn 1340 mthca_alloc_sqp(struct mthca_dev *dev, struct mthca_pd *pd, struct mthca_cq *send_cq, struct mthca_cq *recv_cq, enum ib_sig_type send_policy, struct ib_qp_cap *cap, int qpn, int port, struct mthca_sqp *sqp) argument [all...] |
H A D | mthca_mad.c | 165 int qpn = mad->mad_hdr.mgmt_class != IB_MGMT_CLASS_SUBN_LID_ROUTED; local 167 struct ib_mad_agent *agent = dev->send_agent[port_num - 1][qpn]; 172 send_buf = ib_create_send_mad(agent, qpn, 0, 0, IB_MGMT_MAD_HDR,
|
H A D | mthca_eq.c | 143 __be32 qpn; member in struct:mthca_eqe::__anon1012::__anon1015 282 mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, 287 mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, 292 mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, 297 mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, 307 mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, 312 mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, 317 mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, 322 mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff,
|
H A D | mthca_provider.h | 264 u32 qpn; member in struct:mthca_qp
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_qp.c | 107 u32 i, offset, max_scan, qpn; local 132 qpn = qpt->last + 1; 133 if (qpn >= QPN_MAX) 134 qpn = 2; 135 offset = qpn & BITS_PER_PAGE_MASK; 136 map = &qpt->map[qpn / BITS_PER_PAGE]; 148 qpt->last = qpn; 149 ret = qpn; 153 qpn = mk_qpn(qpt, map, offset); 162 } while (offset < BITS_PER_PAGE && qpn < QPN_MA 190 free_qpn(struct ipath_qp_table *qpt, u32 qpn) argument 301 ipath_lookup_qpn(struct ipath_qp_table *qpt, u32 qpn) argument [all...] |
/drivers/infiniband/hw/amso1100/ |
H A D | c2_provider.h | 113 int qpn; member in struct:c2_qp
|
H A D | c2_qp.c | 388 c2dev->qp_table.last++, &qp->qpn); 395 static void c2_free_qpn(struct c2_dev *c2dev, int qpn) argument 398 idr_remove(&c2dev->qp_table.idr, qpn); 402 struct c2_qp *c2_find_qpn(struct c2_dev *c2dev, int qpn) argument 408 qp = idr_find(&c2dev->qp_table.idr, qpn); 430 qp->ibqp.qp_num = qp->qpn; 568 c2_free_qpn(c2dev, qp->qpn); 611 c2_free_qpn(c2dev, qp->qpn);
|
H A D | c2_cm.c | 50 ibqp = c2_get_qp(cm_id->device, iw_param->qpn); 293 ibqp = c2_get_qp(cm_id->device, iw_param->qpn);
|
/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 110 return qp->mqp.qpn >= dev->dev->caps.sqp_start && 111 qp->mqp.qpn <= dev->dev->caps.sqp_start + 3; 116 return qp->mqp.qpn >= dev->dev->caps.sqp_start && 117 qp->mqp.qpn <= dev->dev->caps.sqp_start + 1; 260 "on QP %06x\n", type, qp->qpn); 477 int qpn; local 574 qpn = sqpn; 576 err = mlx4_qp_reserve_range(dev->dev, 1, 1, &qpn); 581 err = mlx4_qp_alloc(dev->dev, qpn, &qp->mqp); 586 qp->mqp.qpn | [all...] |
H A D | mad.c | 208 int qpn = mad->mad_hdr.mgmt_class != IB_MGMT_CLASS_SUBN_LID_ROUTED; local 210 struct ib_mad_agent *agent = dev->send_agent[port_num - 1][qpn]; 214 send_buf = ib_create_send_mad(agent, qpn, 0, 0, IB_MGMT_MAD_HDR,
|
H A D | mlx4_ib.h | 301 void __mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq); 302 void mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq);
|
H A D | cq.c | 600 (be32_to_cpu(cqe->vlan_my_qpn) & MLX4_CQE_QPN_MASK) != (*cur_qp)->mqp.qpn) { 772 void __mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq) argument 796 if ((be32_to_cpu(cqe->vlan_my_qpn) & MLX4_CQE_QPN_MASK) == qpn) { 820 void mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq) argument 823 __mlx4_ib_cq_clean(cq, qpn, srq);
|
/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_cm.c | 425 data.qpn = cpu_to_be32(priv->qp->qp_num); 1034 u32 qpn, 1041 data.qpn = cpu_to_be32(priv->qp->qp_num); 1046 req.service_id = cpu_to_be64(IPOIB_CM_IETF_ID | qpn); 1094 static int ipoib_cm_tx_init(struct ipoib_cm_tx *p, u32 qpn, argument 1127 ret = ipoib_cm_send_req(p->dev, p->id, p->qp, qpn, pathrec); 1133 ipoib_dbg(priv, "Request connection 0x%x for gid %pI6 qpn 0x%x\n", 1134 p->qp->qp_num, pathrec->dgid.raw, qpn); 1298 u32 qpn; local 1307 qpn 1032 ipoib_cm_send_req(struct net_device *dev, struct ib_cm_id *id, struct ib_qp *qp, u32 qpn, struct ib_sa_path_rec *pathrec) argument [all...] |