/drivers/infiniband/hw/ehca/ |
H A D | ehca_eq.c | 53 int ehca_create_eq(struct ehca_shca *shca, argument 62 struct ib_device *ib_dev = &shca->ib_device; 77 h_ret = hipz_h_alloc_resource_eq(shca->ipz_hca_handle, 105 h_ret = hipz_h_register_rpage_eq(shca->ipz_hca_handle, 125 tasklet_init(&eq->interrupt_task, ehca_tasklet_eq, (long)shca); 129 (void *)shca); 133 tasklet_init(&eq->interrupt_task, ehca_tasklet_neq, (long)shca); 137 (void *)shca); 150 hipz_h_destroy_eq(shca->ipz_hca_handle, eq); 155 void *ehca_poll_eq(struct ehca_shca *shca, struc argument 167 ehca_destroy_eq(struct ehca_shca *shca, struct ehca_eq *eq) argument [all...] |
H A D | ehca_mcast.c | 62 struct ehca_shca *shca = container_of(ibqp->device, struct ehca_shca, local 84 h_ret = hipz_h_attach_mcqp(shca->ipz_hca_handle, 99 struct ehca_shca *shca = container_of(ibqp->pd->device, local 121 h_ret = hipz_h_detach_mcqp(shca->ipz_hca_handle,
|
H A D | ehca_pd.c | 76 struct ehca_shca *shca = container_of(device, struct ehca_shca, local 78 pd->fw_pd.value = shca->pd->fw_pd.value;
|
H A D | ehca_av.c | 52 int ehca_calc_ipd(struct ehca_shca *shca, int port, argument 65 ehca_err(&shca->ib_device, "Invalid static rate! path_rate=%x", 70 ret = ehca_query_port(&shca->ib_device, port, &pa); 72 ehca_err(&shca->ib_device, "Failed to query port ret=%i", ret); 92 struct ehca_shca *shca = container_of(pd->device, struct ehca_shca, local 108 if (ehca_calc_ipd(shca, ah_attr->port_num, 154 av->av.pmtu = shca->max_mtu; 172 struct ehca_shca *shca = container_of(ah->pd->device, struct ehca_shca, local 221 new_ehca_av.pmtu = shca->max_mtu;
|
H A D | ehca_cq.c | 123 struct ehca_shca *shca = local 137 if (!atomic_add_unless(&shca->num_cqs, 1, shca->max_num_cqs)) { 139 "CQs reached.", shca->max_num_cqs); 149 atomic_dec(&shca->num_cqs); 163 adapter_handle = shca->ipz_hca_handle; 164 param.eq_handle = shca->eq.ipz_eq_handle; 322 atomic_dec(&shca->num_cqs); 332 struct ehca_shca *shca = container_of(device, struct ehca_shca, local 334 struct ipz_adapter_handle adapter_handle = shca [all...] |
H A D | ehca_hca.c | 56 struct ehca_shca *shca = container_of(ibdev, struct ehca_shca, local 76 ehca_err(&shca->ib_device, "Can't allocate rblock memory."); 80 if (hipz_h_query_hca(shca->ipz_hca_handle, rblock) != H_SUCCESS) { 81 ehca_err(&shca->ib_device, "Can't query device properties"); 87 props->page_size_cap = shca->hca_cap_mr_pgsize; 112 if (EHCA_BMASK_GET(HCA_CAP_SRQ, shca->hca_cap)) { 142 static enum ib_mtu map_mtu(struct ehca_shca *shca, u32 fw_mtu) argument 156 ehca_err(&shca->ib_device, "Unknown MTU size: %x.", 162 static u8 map_number_of_vls(struct ehca_shca *shca, u32 vl_cap) argument 176 ehca_err(&shca 187 struct ehca_shca *shca = container_of(ibdev, struct ehca_shca, local 245 ehca_query_sma_attr(struct ehca_shca *shca, u8 port, struct ehca_sma_attr *attr) argument 285 struct ehca_shca *shca; local 320 struct ehca_shca *shca = container_of(ibdev, struct ehca_shca, local 361 struct ehca_shca *shca; local [all...] |
H A D | ehca_sqp.c | 64 u64 ehca_define_sqp(struct ehca_shca *shca, argument 73 shca->sport[port - 1].port_state = IB_PORT_DOWN; 80 ret = hipz_h_define_aqp1(shca->ipz_hca_handle, 87 ehca_err(&shca->ib_device, 92 shca->sport[port - 1].pma_qp_nr = pma_qp_nr; 93 ehca_dbg(&shca->ib_device, "port=%x pma_qp_nr=%x", 97 ehca_err(&shca->ib_device, "invalid qp_type=%x", 106 shca->sport[port - 1].port_state != IB_PORT_ACTIVE && 109 ehca_dbg(&shca->ib_device, "... wait until port %x is active", 115 ehca_err(&shca 152 struct ehca_shca *shca = local [all...] |
H A D | ehca_irq.c | 88 static void print_error_data(struct ehca_shca *shca, void *data, argument 103 ehca_err(&shca->ib_device, 112 ehca_err(&shca->ib_device, 118 ehca_err(&shca->ib_device, 120 type, shca->ib_device.name); 124 ehca_err(&shca->ib_device, "Error data is available: %llx.", resource); 125 ehca_err(&shca->ib_device, "EHCA ----- error data begin " 128 ehca_err(&shca->ib_device, "EHCA ----- error data end " 134 int ehca_error_data(struct ehca_shca *shca, void *data, argument 144 ehca_err(&shca 178 dispatch_qp_event(struct ehca_shca *shca, struct ehca_qp *qp, enum ib_event_type event_type) argument 205 qp_event_callback(struct ehca_shca *shca, u64 eqe, enum ib_event_type event_type, int fatal) argument 239 cq_event_callback(struct ehca_shca *shca, u64 eqe) argument 262 parse_identifier(struct ehca_shca *shca, u64 eqe) argument 326 dispatch_port_event(struct ehca_shca *shca, int port_num, enum ib_event_type type, const char *msg) argument 338 notify_port_conf_change(struct ehca_shca *shca, int port_num) argument 384 parse_ec(struct ehca_shca *shca, u64 eqe) argument 471 struct ehca_shca *shca = (struct ehca_shca*)dev_id; local 480 struct ehca_shca *shca = (struct ehca_shca*)data; local 504 struct ehca_shca *shca = (struct ehca_shca*)dev_id; local 512 process_eqe(struct ehca_shca *shca, struct ehca_eqe *eqe) argument 548 ehca_process_eq(struct ehca_shca *shca, int is_irq) argument [all...] |
H A D | ehca_main.c | 278 static int ehca_sense_attributes(struct ehca_shca *shca) argument 294 shca->ofdev->dev.of_node->full_name); 295 loc_code = of_get_property(shca->ofdev->dev.of_node, "ibm,loc-code", 306 h_ret = hipz_h_query_hca(shca->ipz_hca_handle, rblock); 315 shca->num_ports = 1; 317 shca->num_ports = (u8)rblock->num_ports; 332 shca->hw_level = 0x10 | (revid + 1); 334 shca->hw_level = 0x14; 337 shca->hw_level = 0x21; 339 shca 409 init_node_guid(struct ehca_shca *shca) argument 433 ehca_init_device(struct ehca_shca *shca) argument 530 ehca_create_aqp1(struct ehca_shca *shca, u32 port) argument 686 struct ehca_shca *shca = dev_get_drvdata(dev); local 719 struct ehca_shca *shca; local 884 struct ehca_shca *shca = dev_get_drvdata(&dev->dev); local 953 struct ehca_shca *shca; local [all...] |
H A D | ehca_qp.c | 279 static inline int init_qp_queue(struct ehca_shca *shca, argument 291 struct ib_device *ib_dev = &shca->ib_device; 292 struct ipz_adapter_handle ipz_hca_handle = shca->ipz_hca_handle; 460 struct ehca_shca *shca = container_of(pd->device, struct ehca_shca, local 472 if (!atomic_add_unless(&shca->num_qps, 1, shca->max_num_qps)) { 474 "QPs reached.", shca->max_num_qps); 481 atomic_dec(&shca->num_qps); 492 atomic_dec(&shca->num_qps); 511 atomic_dec(&shca 975 struct ehca_shca *shca = container_of(pd->device, struct ehca_shca, local 1073 prepare_sqe_rts(struct ehca_qp *my_qp, struct ehca_shca *shca, int *bad_wqe_cnt) argument 1163 check_for_left_cqes(struct ehca_qp *my_qp, struct ehca_shca *shca) argument 1250 struct ehca_shca *shca = local 1787 struct ehca_shca *shca = container_of(ibqp->device, struct ehca_shca, local 1896 struct ehca_shca *shca = container_of(qp->device, struct ehca_shca, local 2041 struct ehca_shca *shca = local 2097 struct ehca_shca *shca = container_of(srq->device, struct ehca_shca, local 2138 struct ehca_shca *shca = container_of(dev, struct ehca_shca, ib_device); local [all...] |
H A D | ehca_mrmw.c | 113 static u64 ehca_get_max_hwpage_size(struct ehca_shca *shca) argument 115 return rounddown_pow_of_two(shca->hca_cap_mr_pgsize); 162 struct ehca_shca *shca = local 165 if (shca->maxmr) { 168 ehca_err(&shca->ib_device, "out of memory"); 173 ret = ehca_reg_maxmr(shca, e_maxmr, 185 ehca_err(&shca->ib_device, "no internal max-MR exist!"); 192 ehca_err(&shca->ib_device, "h_ret=%li pd=%p mr_access_flags=%x", 208 struct ehca_shca *shca = local 259 ret = ehca_reg_maxmr(shca, e_m 319 struct ehca_shca *shca = local 453 struct ehca_shca *shca = local 601 struct ehca_shca *shca = local 648 struct ehca_shca *shca = local 697 struct ehca_shca *shca = local 746 struct ehca_shca *shca = local 770 struct ehca_shca *shca = local 861 struct ehca_shca *shca = local 927 struct ehca_shca *shca = NULL; local 985 struct ehca_shca *shca = local 1021 ehca_reg_mr(struct ehca_shca *shca, struct ehca_mr *e_mr, u64 *iova_start, u64 size, int acl, struct ehca_pd *e_pd, struct ehca_mr_pginfo *pginfo, u32 *lkey, u32 *rkey, enum ehca_reg_type reg_type) argument 1099 ehca_reg_mr_rpages(struct ehca_shca *shca, struct ehca_mr *e_mr, struct ehca_mr_pginfo *pginfo) argument 1197 ehca_rereg_mr_rereg1(struct ehca_shca *shca, struct ehca_mr *e_mr, u64 *iova_start, u64 size, u32 acl, struct ehca_pd *e_pd, struct ehca_mr_pginfo *pginfo, u32 *lkey, u32 *rkey) argument 1288 ehca_rereg_mr(struct ehca_shca *shca, struct ehca_mr *e_mr, u64 *iova_start, u64 size, int acl, struct ehca_pd *e_pd, struct ehca_mr_pginfo *pginfo, u32 *lkey, u32 *rkey) argument 1383 ehca_unmap_one_fmr(struct ehca_shca *shca, struct ehca_mr *e_fmr) argument 1471 ehca_reg_smr(struct ehca_shca *shca, struct ehca_mr *e_origmr, struct ehca_mr *e_newmr, u64 *iova_start, int acl, struct ehca_pd *e_pd, u32 *lkey, u32 *rkey) argument 1534 ehca_reg_mr_section(int top, int dir, int idx, u64 *kpage, struct ehca_shca *shca, struct ehca_mr *mr, struct ehca_mr_pginfo *pginfo) argument 1571 ehca_reg_mr_sections(int top, int dir, u64 *kpage, struct ehca_shca *shca, struct ehca_mr *mr, struct ehca_mr_pginfo *pginfo) argument 1590 ehca_reg_mr_dir_sections(int top, u64 *kpage, struct ehca_shca *shca, struct ehca_mr *mr, struct ehca_mr_pginfo *pginfo) argument 1609 ehca_reg_internal_maxmr( struct ehca_shca *shca, struct ehca_pd *e_pd, struct ehca_mr **e_maxmr) argument 1687 ehca_reg_maxmr(struct ehca_shca *shca, struct ehca_mr *e_newmr, u64 *iova_start, int acl, struct ehca_pd *e_pd, u32 *lkey, u32 *rkey) argument 1729 ehca_dereg_internal_maxmr(struct ehca_shca *shca) argument 2467 ehca_reg_bmap_mr_rpages(struct ehca_shca *shca, struct ehca_mr *e_mr, struct ehca_mr_pginfo *pginfo) argument [all...] |