/drivers/infiniband/hw/amso1100/ |
H A D | c2_cm.c | 41 int c2_llp_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument 43 struct c2_dev *c2dev = to_c2dev(cm_id->device); 49 struct sockaddr_in *raddr = (struct sockaddr_in *)&cm_id->remote_addr; 51 if (cm_id->remote_addr.ss_family != AF_INET) 54 ibqp = c2_get_qp(cm_id->device, iw_param->qpn); 60 cm_id->provider_data = qp; 61 cm_id->add_ref(cm_id); 62 qp->cm_id = cm_id; 135 c2_llp_service_create(struct iw_cm_id *cm_id, int backlog) argument 224 c2_llp_service_destroy(struct iw_cm_id *cm_id) argument 291 c2_llp_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument 397 c2_llp_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument [all...] |
H A D | c2_ae.c | 187 struct iw_cm_id *cm_id = qp->cm_id; local 190 if (!cm_id) { 191 pr_debug("event received, but cm_id is <nul>, qp=%p!\n", 221 if (qp->cm_id) { 222 qp->cm_id->rem_ref(qp->cm_id); 223 qp->cm_id = NULL; 229 if (cm_id->event_handler) 230 cm_id 275 struct iw_cm_id *cm_id = local [all...] |
H A D | c2_intr.c | 200 cm_event.local_addr = req->cm_id->local_addr; 201 cm_event.remote_addr = req->cm_id->remote_addr; 204 req->cm_id->event_handler(req->cm_id, &cm_event);
|
H A D | c2_vq.h | 46 struct iw_cm_id *cm_id; member in struct:c2_vq_req
|
H A D | c2_qp.c | 174 if (qp->cm_id && qp->state == IB_QPS_RTS) { 176 "qp=%p, cm_id=%p\n",qp,qp->cm_id); 178 vq_req->cm_id = qp->cm_id; 236 if (vq_req->event==IW_CM_EVENT_CLOSE && qp->cm_id) { 237 qp->cm_id->rem_ref(qp->cm_id); 238 qp->cm_id = NULL; 332 if (qp->cm_id [all...] |
H A D | c2.h | 521 extern int c2_llp_connect(struct iw_cm_id *cm_id, 523 extern int c2_llp_accept(struct iw_cm_id *cm_id, 525 extern int c2_llp_reject(struct iw_cm_id *cm_id, const void *pdata, 527 extern int c2_llp_service_create(struct iw_cm_id *cm_id, int backlog); 528 extern int c2_llp_service_destroy(struct iw_cm_id *cm_id);
|
H A D | c2_provider.c | 593 static int c2_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument 598 return c2_llp_connect(cm_id, iw_param); 601 static int c2_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument 606 return c2_llp_accept(cm_id, iw_param); 609 static int c2_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument 615 err = c2_llp_reject(cm_id, pdata, pdata_len); 619 static int c2_service_create(struct iw_cm_id *cm_id, int backlog) argument 624 err = c2_llp_service_create(cm_id, backlog); 631 static int c2_service_destroy(struct iw_cm_id *cm_id) argument 636 err = c2_llp_service_destroy(cm_id); [all...] |
H A D | c2_provider.h | 109 struct iw_cm_id *cm_id; member in struct:c2_qp
|
H A D | c2_vq.c | 112 r->cm_id = NULL;
|
/drivers/infiniband/core/ |
H A D | iwcm.c | 63 struct iwcm_id_private *cm_id; member in struct:iwcm_work 85 * elements. The design pre-allocates them based on the cm_id type: 95 * One exception is when creating the cm_id for incoming connection requests. 97 * 1) in the event upcall, cm_event_handler(), for a listening cm_id. If 102 * If work elements cannot be allocated for the new connect request cm_id, 121 list_add(&work->free_list, &work->cm_id->work_free_list); 143 work->cm_id = cm_id_priv; 173 * Release a reference on cm_id. If the last reference is being 189 static void add_ref(struct iw_cm_id *cm_id) argument 192 cm_id_priv = container_of(cm_id, struc 196 rem_ref(struct iw_cm_id *cm_id) argument 281 iw_cm_disconnect(struct iw_cm_id *cm_id, int abrupt) argument 348 destroy_cm_id(struct iw_cm_id *cm_id) argument 414 iw_destroy_cm_id(struct iw_cm_id *cm_id) argument 435 iw_cm_listen(struct iw_cm_id *cm_id, int backlog) argument 474 iw_cm_reject(struct iw_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 512 iw_cm_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument 568 iw_cm_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param) argument 641 struct iw_cm_id *cm_id; local 927 cm_event_handler(struct iw_cm_id *cm_id, struct iw_cm_event *iw_event) argument 1019 iw_cm_init_qp_attr(struct iw_cm_id *cm_id, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument [all...] |
H A D | ucma.c | 87 struct rdma_cm_id *cm_id; member in struct:ucma_context 108 struct rdma_cm_id *cm_id; member in struct:ucma_event 245 static int ucma_event_handler(struct rdma_cm_id *cm_id, argument 249 struct ucma_context *ctx = cm_id->context; 257 uevent->cm_id = cm_id; 261 if (cm_id->qp_type == IB_QPT_UD) 274 } else if (!ctx->uid || ctx->cm_id != cm_id) { 329 ctx->cm_id 755 ucma_query_device_addr(struct rdma_cm_id *cm_id, struct rdma_ucm_query_addr_resp *resp) argument [all...] |
H A D | cm.c | 814 * The cm_id could be destroyed by the user before we exit timewait. 815 * To protect against this, we search for the cm_id after exiting 839 static void cm_destroy_id(struct ib_cm_id *cm_id, int err) argument 844 cm_id_priv = container_of(cm_id, struct cm_id_private, id); 847 switch (cm_id->state) { 849 cm_id->state = IB_CM_IDLE; 856 cm_id->state = IB_CM_IDLE; 867 ib_send_cm_rej(cm_id, IB_CM_REJ_TIMEOUT, 879 ib_send_cm_rej(cm_id, IB_CM_REJ_CONSUMER_DEFINED, 892 ib_send_cm_rej(cm_id, IB_CM_REJ_CONSUMER_DEFINE 925 ib_destroy_cm_id(struct ib_cm_id *cm_id) argument 931 ib_cm_listen(struct ib_cm_id *cm_id, __be64 service_id, __be64 service_mask, struct ib_cm_compare_data *compare_data) argument 1109 ib_send_cm_req(struct ib_cm_id *cm_id, struct ib_cm_req_param *param) argument 1523 struct ib_cm_id *cm_id; local 1641 ib_send_cm_rep(struct ib_cm_id *cm_id, struct ib_cm_rep_param *param) argument 1703 ib_send_cm_rtu(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 1992 ib_send_cm_dreq(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 2054 ib_send_cm_drep(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len) argument 2239 ib_send_cm_rej(struct ib_cm_id *cm_id, enum ib_cm_rej_reason reason, void *ari, u8 ari_length, const void *private_data, u8 private_data_len) argument 2413 ib_send_cm_mra(struct ib_cm_id *cm_id, u8 service_timeout, const void *private_data, u8 private_data_len) argument 2609 ib_send_cm_lap(struct ib_cm_id *cm_id, struct ib_sa_path_rec *alternate_path, const void *private_data, u8 private_data_len) argument 2782 ib_send_cm_apr(struct ib_cm_id *cm_id, enum ib_cm_apr_status status, void *info, u8 info_length, const void *private_data, u8 private_data_len) argument 2922 ib_send_cm_sidr_req(struct ib_cm_id *cm_id, struct ib_cm_sidr_req_param *param) argument 2988 struct ib_cm_id *cm_id; local 3066 ib_send_cm_sidr_rep(struct ib_cm_id *cm_id, struct ib_cm_sidr_rep_param *param) argument 3301 cm_establish(struct ib_cm_id *cm_id) argument 3350 cm_migrate(struct ib_cm_id *cm_id) argument 3370 ib_cm_notify(struct ib_cm_id *cm_id, enum ib_event_type event) argument 3625 ib_cm_init_qp_attr(struct ib_cm_id *cm_id, struct ib_qp_attr *qp_attr, int *qp_attr_mask) argument [all...] |
H A D | ucm.c | 83 struct ib_cm_id *cm_id; member in struct:ib_ucm_context 95 struct ib_cm_id *cm_id; member in struct:ib_ucm_event 168 ib_destroy_cm_id(uevent->cm_id); 347 static int ib_ucm_event_handler(struct ib_cm_id *cm_id, argument 354 ctx = cm_id->context; 361 uevent->cm_id = cm_id; 380 /* Destroy new cm_id's */ 422 ctx->cm_id = uevent->cm_id; 829 ib_ucm_send_private_data(struct ib_ucm_file *file, const char __user *inbuf, int in_len, int (*func)(struct ib_cm_id *cm_id, const void *private_data, u8 private_data_len)) argument 879 ib_ucm_send_info(struct ib_ucm_file *file, const char __user *inbuf, int in_len, int (*func)(struct ib_cm_id *cm_id, int status, const void *info, u8 info_len, const void *data, u8 data_len)) argument [all...] |
H A D | cma.c | 138 } cm_id; member in struct:rdma_id_private 740 if (!id_priv->cm_id.ib || (id_priv->id.qp_type == IB_QPT_UD)) 743 ret = ib_cm_init_qp_attr(id_priv->cm_id.ib, qp_attr, 750 if (!id_priv->cm_id.iw) { 754 ret = iw_cm_init_qp_attr(id_priv->cm_id.iw, qp_attr, 1042 if (id_priv->cm_id.ib) 1043 ib_destroy_cm_id(id_priv->cm_id.ib); 1046 if (id_priv->cm_id.iw) 1047 iw_destroy_cm_id(id_priv->cm_id.iw); 1080 ret = ib_send_cm_rtu(id_priv->cm_id 1106 cma_ib_handler(struct ib_cm_id *cm_id, struct ib_cm_event *ib_event) argument 1282 cma_req_handler(struct ib_cm_id *cm_id, struct ib_cm_event *ib_event) argument 1468 iw_conn_req_handler(struct iw_cm_id *cm_id, struct iw_cm_event *iw_event) argument 2646 cma_sidr_rep_handler(struct ib_cm_id *cm_id, struct ib_cm_event *ib_event) argument 2835 struct iw_cm_id *cm_id; local [all...] |
/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_cm.h | 155 struct iw_cm_id *cm_id; member in struct:iwch_ep_common 196 static inline struct iwch_ep *to_ep(struct iw_cm_id *cm_id) argument 198 return cm_id->provider_data; 201 static inline struct iwch_listen_ep *to_listen_ep(struct iw_cm_id *cm_id) argument 203 return cm_id->provider_data; 217 int iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 218 int iwch_create_listen(struct iw_cm_id *cm_id, int backlog); 219 int iwch_destroy_listen(struct iw_cm_id *cm_id); 220 int iwch_reject_cr(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len); 221 int iwch_accept_cr(struct iw_cm_id *cm_id, struc [all...] |
H A D | iwch_cm.c | 675 if (ep->com.cm_id) { 676 PDBG("close complete delivered ep %p cm_id %p tid %d\n", 677 ep, ep->com.cm_id, ep->hwtid); 678 ep->com.cm_id->event_handler(ep->com.cm_id, &event); 679 ep->com.cm_id->rem_ref(ep->com.cm_id); 680 ep->com.cm_id = NULL; 692 if (ep->com.cm_id) { 693 PDBG("peer close delivered ep %p cm_id 1773 iwch_reject_cr(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument 1794 iwch_accept_cr(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 1877 is_loopback_dst(struct iw_cm_id *cm_id) argument 1889 iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 1986 iwch_create_listen(struct iw_cm_id *cm_id, int backlog) argument 2046 iwch_destroy_listen(struct iw_cm_id *cm_id) argument [all...] |
/drivers/infiniband/hw/nes/ |
H A D | nes_cm.c | 178 if (!cm_node->cm_id) 193 event->cm_info.cm_id = cm_node->cm_id; 872 struct iw_cm_id *cm_id = cm_node->cm_id; local 883 if (cm_node->cm_id) 884 cm_id->rem_ref(cm_id); 897 struct iw_cm_id *cm_id = cm_node->cm_id; local 2578 struct iw_cm_id *cm_id = cm_node->cm_id; local 3001 struct iw_cm_id *cm_id; local 3180 nes_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 3435 nes_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument 3468 nes_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 3595 nes_create_listen(struct iw_cm_id *cm_id, int backlog) argument 3666 nes_destroy_listen(struct iw_cm_id *cm_id) argument 3736 struct iw_cm_id *cm_id; local 3855 struct iw_cm_id *cm_id; local 3914 struct iw_cm_id *cm_id; local 3970 struct iw_cm_id *cm_id; local 4019 struct iw_cm_id *cm_id; local [all...] |
H A D | nes_cm.h | 298 struct iw_cm_id *cm_id; member in struct:nes_cm_listener 343 struct iw_cm_id *cm_id; member in struct:nes_cm_node 361 struct iw_cm_id *cm_id; member in union:nes_cm_info::__anon1571
|
/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_cm.c | 77 static int ipoib_cm_tx_handler(struct ib_cm_id *cm_id, 273 struct ib_cm_id *cm_id, struct ib_qp *qp, 281 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); 292 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); 313 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); 346 static int ipoib_cm_nonsrq_init_rx(struct net_device *dev, struct ib_cm_id *cm_id, argument 376 ib_send_cm_rej(cm_id, IB_CM_REJ_NO_QP, NULL, 0, NULL, 0); 419 static int ipoib_cm_send_rep(struct net_device *dev, struct ib_cm_id *cm_id, argument 437 return ib_send_cm_rep(cm_id, &rep); 440 static int ipoib_cm_req_handler(struct ib_cm_id *cm_id, struc argument 272 ipoib_cm_modify_rx_qp(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp, unsigned psn) argument 501 ipoib_cm_rx_handler(struct ib_cm_id *cm_id, struct ib_cm_event *event) argument 954 ipoib_cm_rep_handler(struct ib_cm_id *cm_id, struct ib_cm_event *event) argument 1086 ipoib_cm_modify_tx_init(struct net_device *dev, struct ib_cm_id *cm_id, struct ib_qp *qp) argument 1219 ipoib_cm_tx_handler(struct ib_cm_id *cm_id, struct ib_cm_event *event) argument [all...] |
/drivers/infiniband/hw/cxgb4/ |
H A D | cm.c | 1089 if (ep->com.cm_id) { 1090 PDBG("close complete delivered ep %p cm_id %p tid %u\n", 1091 ep, ep->com.cm_id, ep->hwtid); 1092 ep->com.cm_id->event_handler(ep->com.cm_id, &event); 1093 ep->com.cm_id->rem_ref(ep->com.cm_id); 1094 ep->com.cm_id = NULL; 1114 if (ep->com.cm_id) { 1115 PDBG("peer close delivered ep %p cm_id 2733 c4iw_reject_cr(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) argument 2761 c4iw_accept_cr(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 2872 pick_local_ipaddrs(struct c4iw_dev *dev, struct iw_cm_id *cm_id) argument 2919 pick_local_ip6addrs(struct c4iw_dev *dev, struct iw_cm_id *cm_id) argument 2933 c4iw_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) argument 3170 c4iw_create_listen(struct iw_cm_id *cm_id, int backlog) argument 3264 c4iw_destroy_listen(struct iw_cm_id *cm_id) argument [all...] |
H A D | iw_cxgb4.h | 797 struct iw_cm_id *cm_id; member in struct:c4iw_ep_common 891 static inline struct c4iw_ep *to_ep(struct iw_cm_id *cm_id) argument 893 return cm_id->provider_data; 896 static inline struct c4iw_listen_ep *to_listen_ep(struct iw_cm_id *cm_id) argument 898 return cm_id->provider_data; 958 int c4iw_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 959 int c4iw_create_listen(struct iw_cm_id *cm_id, int backlog); 960 int c4iw_destroy_listen(struct iw_cm_id *cm_id); 961 int c4iw_accept_cr(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 962 int c4iw_reject_cr(struct iw_cm_id *cm_id, cons [all...] |
/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.c | 229 pr_debug("QP event %d on cm_id=%p sess_name=%s state=%d\n", 230 event->event, ch->cm_id, ch->sess_name, srpt_get_ch_state(ch)); 234 ib_cm_notify(ch->cm_id, event->event); 993 ret = ib_cm_init_qp_attr(ch->cm_id, &qp_attr, &attr_mask); 1023 ret = ib_cm_init_qp_attr(ch->cm_id, &qp_attr, &attr_mask); 1833 " cm_id %p sess %p\n", srp_tsk->tsk_mgmt_func, 1834 srp_tsk->task_tag, srp_tsk->tag, ch->cm_id, ch->sess); 2132 pr_debug("%s: max_cqe= %d max_sge= %d sq_size = %d cm_id= %p\n", 2134 qp_init->cap.max_send_wr, ch->cm_id); 2202 ib_send_cm_rej(ch->cm_id, IB_CM_REJ_NO_RESOURCE 2263 srpt_drain_channel(struct ib_cm_id *cm_id) argument 2305 srpt_find_channel(struct srpt_device *sdev, struct ib_cm_id *cm_id) argument 2413 srpt_cm_req_recv(struct ib_cm_id *cm_id, struct ib_cm_req_event_param *param, void *private_data) argument 2680 srpt_cm_rej_recv(struct ib_cm_id *cm_id) argument 2692 srpt_cm_rtu_recv(struct ib_cm_id *cm_id) argument 2715 srpt_cm_timewait_exit(struct ib_cm_id *cm_id) argument 2721 srpt_cm_rep_error(struct ib_cm_id *cm_id) argument 2730 srpt_cm_dreq_recv(struct ib_cm_id *cm_id) argument 2767 srpt_cm_drep_recv(struct ib_cm_id *cm_id) argument 2784 srpt_cm_handler(struct ib_cm_id *cm_id, struct ib_cm_event *event) argument [all...] |
H A D | ib_srpt.h | 274 * @cm_id: IB CM ID associated with the channel. 296 * against concurrent modification by the cm_id spinlock. 305 struct ib_cm_id *cm_id; member in struct:srpt_rdma_ch 382 * @cm_id: Connection identifier. 399 struct ib_cm_id *cm_id; member in struct:srpt_device
|
/drivers/infiniband/hw/mlx4/ |
H A D | cm.c | 75 static void set_local_comm_id(struct ib_mad *mad, u32 cm_id) argument 80 msg->request_id = cpu_to_be32(cm_id); 86 msg->local_comm_id = cpu_to_be32(cm_id); 105 static void set_remote_comm_id(struct ib_mad *mad, u32 cm_id) argument 110 msg->request_id = cpu_to_be32(cm_id); 116 msg->remote_comm_id = cpu_to_be32(cm_id); 220 pr_debug("overriding existing sl_id_map entry (cm_id = %x)\n",
|
/drivers/char/ |
H A D | mbcs.h | 79 union cm_id { union 145 uint64_t cm_id:2, // 1:0 member in struct:cm_control::__anon400 342 union cm_id id;
|