/drivers/infiniband/core/ |
H A D | agent.h | 43 extern int ib_agent_port_open(struct ib_device *device, int port_num); 45 extern int ib_agent_port_close(struct ib_device *device, int port_num); 48 struct ib_wc *wc, struct ib_device *device,
|
H A D | core_priv.h | 41 int ib_device_register_sysfs(struct ib_device *device, 42 int (*port_callback)(struct ib_device *, 44 void ib_device_unregister_sysfs(struct ib_device *device);
|
H A D | smi.h | 66 struct ib_device *device) 81 struct ib_device *device)
|
H A D | device.c | 70 static int ib_device_check_mandatory(struct ib_device *device) 72 #define IB_MANDATORY_FUNC(x) { offsetof(struct ib_device, x), #x } 110 static struct ib_device *__ib_device_get_by_name(const char *name) 112 struct ib_device *device; 126 struct ib_device *device; 154 static int start_port(struct ib_device *device) 160 static int end_port(struct ib_device *device) 171 * ib_device. @size is the size of the structure to be allocated, 176 struct ib_device *ib_alloc_device(size_t size) 178 BUG_ON(size < sizeof (struct ib_device)); [all...] |
H A D | sa.h | 52 struct ib_device *device, u8 port_num,
|
H A D | agent.c | 57 __ib_get_agent_port(struct ib_device *device, int port_num) 70 ib_get_agent_port(struct ib_device *device, int port_num) 82 struct ib_wc *wc, struct ib_device *device, 145 int ib_agent_port_open(struct ib_device *device, int port_num) 196 int ib_agent_port_close(struct ib_device *device, int port_num)
|
H A D | cache.c | 57 struct ib_device *device; 61 static inline int start_port(struct ib_device *device) 66 static inline int end_port(struct ib_device *device) 72 int ib_get_cached_gid(struct ib_device *device, 99 int ib_find_cached_gid(struct ib_device *device, 134 int ib_get_cached_pkey(struct ib_device *device, 161 int ib_find_cached_pkey(struct ib_device *device, 202 int ib_find_exact_cached_pkey(struct ib_device *device, 234 int ib_get_cached_lmc(struct ib_device *device, 252 static void ib_cache_update(struct ib_device *devic [all...] |
H A D | sysfs.c | 45 struct ib_device *ibdev; 457 struct ib_device *dev = container_of(device, struct ib_device, dev); 465 struct ib_device *dev = container_of(device, struct ib_device, dev); 520 static int add_port(struct ib_device *device, int port_num, 521 int (*port_callback)(struct ib_device *, 617 struct ib_device *dev = container_of(device, struct ib_device, dev); 633 struct ib_device *de [all...] |
/drivers/infiniband/hw/ipath/ |
H A D | ipath_dma.c | 50 static int ipath_mapping_error(struct ib_device *dev, u64 dma_addr) 55 static u64 ipath_dma_map_single(struct ib_device *dev, 63 static void ipath_dma_unmap_single(struct ib_device *dev, 70 static u64 ipath_dma_map_page(struct ib_device *dev, 94 static void ipath_dma_unmap_page(struct ib_device *dev, 101 static int ipath_map_sg(struct ib_device *dev, struct scatterlist *sgl, 126 static void ipath_unmap_sg(struct ib_device *dev, 133 static void ipath_sync_single_for_cpu(struct ib_device *dev, 140 static void ipath_sync_single_for_device(struct ib_device *dev, 147 static void *ipath_dma_alloc_coherent(struct ib_device *de [all...] |
/drivers/infiniband/hw/qib/ |
H A D | qib_dma.c | 48 static int qib_mapping_error(struct ib_device *dev, u64 dma_addr) 53 static u64 qib_dma_map_single(struct ib_device *dev, void *cpu_addr, 60 static void qib_dma_unmap_single(struct ib_device *dev, u64 addr, size_t size, 66 static u64 qib_dma_map_page(struct ib_device *dev, struct page *page, 88 static void qib_dma_unmap_page(struct ib_device *dev, u64 addr, size_t size, 94 static int qib_map_sg(struct ib_device *dev, struct scatterlist *sgl, 119 static void qib_unmap_sg(struct ib_device *dev, 126 static void qib_sync_single_for_cpu(struct ib_device *dev, u64 addr, 131 static void qib_sync_single_for_device(struct ib_device *dev, u64 addr, 137 static void *qib_dma_alloc_coherent(struct ib_device *de [all...] |
/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_verbs.h | 24 enum rdma_link_layer usnic_ib_port_link_layer(struct ib_device *device, 26 int usnic_ib_query_device(struct ib_device *ibdev, 28 int usnic_ib_query_port(struct ib_device *ibdev, u8 port, 33 int usnic_ib_query_gid(struct ib_device *ibdev, u8 port, int index, 35 int usnic_ib_query_pkey(struct ib_device *ibdev, u8 port, u16 index, 37 struct ib_pd *usnic_ib_alloc_pd(struct ib_device *ibdev, 47 struct ib_cq *usnic_ib_create_cq(struct ib_device *ibdev, int entries, 55 struct ib_ucontext *usnic_ib_alloc_ucontext(struct ib_device *ibdev,
|
H A D | usnic_ib.h | 55 struct ib_device ib_dev; 82 struct usnic_ib_dev *to_usdev(struct ib_device *ibdev)
|
/drivers/infiniband/hw/ehca/ |
H A D | ehca_main.c | 417 ehca_err(&shca->ib_device, "Can't allocate rblock memory."); 422 ehca_err(&shca->ib_device, "Can't query device properties"); 427 memcpy(&shca->ib_device.node_guid, &rblock->node_guid, sizeof(u64)); 442 strlcpy(shca->ib_device.name, "ehca%d", IB_DEVICE_NAME_MAX); 443 shca->ib_device.owner = THIS_MODULE; 445 shca->ib_device.uverbs_abi_ver = 8; 446 shca->ib_device.uverbs_cmd_mask = 464 shca->ib_device.node_type = RDMA_NODE_IB_CA; 465 shca->ib_device.phys_port_cnt = shca->num_ports; 466 shca->ib_device [all...] |
H A D | ehca_hca.c | 53 int ehca_query_device(struct ib_device *ibdev, struct ib_device_attr *props) 57 ib_device); 76 ehca_err(&shca->ib_device, "Can't allocate rblock memory."); 81 ehca_err(&shca->ib_device, "Can't query device properties"); 156 ehca_err(&shca->ib_device, "Unknown MTU size: %x.", 176 ehca_err(&shca->ib_device, "invalid Vl Capability: %x.", 182 int ehca_query_port(struct ib_device *ibdev, 188 ib_device); 193 ehca_err(&shca->ib_device, "Can't allocate rblock memory."); 199 ehca_err(&shca->ib_device, "Ca [all...] |
H A D | ehca_irq.c | 104 ehca_err(&shca->ib_device, 113 ehca_err(&shca->ib_device, 119 ehca_err(&shca->ib_device, 121 type, shca->ib_device.name); 125 ehca_err(&shca->ib_device, "Error data is available: %llx.", resource); 126 ehca_err(&shca->ib_device, "EHCA ----- error data begin " 129 ehca_err(&shca->ib_device, "EHCA ----- error data end " 145 ehca_err(&shca->ib_device, "Cannot allocate rblock memory."); 157 ehca_err(&shca->ib_device, 169 ehca_err(&shca->ib_device, [all...] |
H A D | ehca_iverbs.h | 47 int ehca_query_device(struct ib_device *ibdev, struct ib_device_attr *props); 49 int ehca_query_port(struct ib_device *ibdev, u8 port, 55 int ehca_query_pkey(struct ib_device *ibdev, u8 port, u16 index, u16 * pkey); 57 int ehca_query_gid(struct ib_device *ibdev, u8 port, int index, 60 int ehca_modify_port(struct ib_device *ibdev, u8 port, int port_modify_mask, 63 struct ib_pd *ehca_alloc_pd(struct ib_device *device, 129 struct ib_cq *ehca_create_cq(struct ib_device *device, int cqe, int comp_vector, 183 struct ib_ucontext *ehca_alloc_ucontext(struct ib_device *device, 190 int ehca_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num,
|
H A D | ehca_pd.c | 48 struct ib_pd *ehca_alloc_pd(struct ib_device *device, 77 ib_device);
|
H A D | ehca_sqp.c | 87 ehca_err(&shca->ib_device, 93 ehca_dbg(&shca->ib_device, "port=%x pma_qp_nr=%x", 97 ehca_err(&shca->ib_device, "invalid qp_type=%x", 109 ehca_dbg(&shca->ib_device, "... wait until port %x is active", 115 ehca_err(&shca->ib_device, "Port %x is not active.", port); 142 static int ehca_process_perf(struct ib_device *ibdev, u8 port_num, 153 container_of(ibdev, struct ehca_shca, ib_device); 219 int ehca_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num,
|
H A D | ehca_eq.c | 62 struct ib_device *ib_dev = &shca->ib_device; 183 ehca_err(&shca->ib_device, "Can't free EQ resources.");
|
H A D | ehca_mrmw.c | 163 container_of(pd->device, struct ehca_shca, ib_device); 168 ehca_err(&shca->ib_device, "out of memory"); 185 ehca_err(&shca->ib_device, "no internal max-MR exist!"); 192 ehca_err(&shca->ib_device, "h_ret=%li pd=%p mr_access_flags=%x", 209 container_of(pd->device, struct ehca_shca, ib_device); 320 container_of(pd->device, struct ehca_shca, ib_device); 451 container_of(mr->device, struct ehca_shca, ib_device); 599 container_of(mr->device, struct ehca_shca, ib_device); 646 container_of(mr->device, struct ehca_shca, ib_device); 695 container_of(pd->device, struct ehca_shca, ib_device); [all...] |
/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.h | 39 int ocrdma_query_device(struct ib_device *, struct ib_device_attr *props); 40 int ocrdma_query_port(struct ib_device *, u8 port, struct ib_port_attr *props); 41 int ocrdma_modify_port(struct ib_device *, u8 port, int mask, 45 int ocrdma_query_gid(struct ib_device *, u8 port, 47 int ocrdma_query_pkey(struct ib_device *, u8 port, u16 index, u16 *pkey); 49 struct ib_ucontext *ocrdma_alloc_ucontext(struct ib_device *, 55 struct ib_pd *ocrdma_alloc_pd(struct ib_device *, 59 struct ib_cq *ocrdma_create_cq(struct ib_device *, int entries, int vector, 94 struct ib_fast_reg_page_list *ocrdma_alloc_frmr_page_list(struct ib_device
|
H A D | ocrdma_ah.h | 36 int ocrdma_process_mad(struct ib_device *,
|
/drivers/infiniband/ulp/iser/ |
H A D | iser_initiator.c | 168 ib_dma_sync_single_for_cpu(device->ib_device, 190 ib_dma_unmap_single(device->ib_device, 195 ib_dma_unmap_single(device->ib_device, 223 iser_conn->login_req_dma = ib_dma_map_single(device->ib_device, 228 iser_conn->login_resp_dma = ib_dma_map_single(device->ib_device, 233 req_err = ib_dma_mapping_error(device->ib_device, 235 resp_err = ib_dma_mapping_error(device->ib_device, 284 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc, 286 if (ib_dma_mapping_error(device->ib_device, dma_addr)) 303 ib_dma_unmap_single(device->ib_device, rx_des [all...] |
H A D | iser_verbs.c | 81 ret = ib_query_device(device->ib_device, dev_attr); 83 pr_warn("Query device failed for %s\n", device->ib_device->name); 88 if (device->ib_device->alloc_fmr && device->ib_device->dealloc_fmr && 89 device->ib_device->map_phys_fmr && device->ib_device->unmap_fmr) { 108 device->ib_device->num_comp_vectors); 110 device->comps_used, device->ib_device->name, 111 device->ib_device->num_comp_vectors); 113 device->pd = ib_alloc_pd(device->ib_device); 264 iser_create_fastreg_desc(struct ib_device *ib_device, struct ib_pd *pd, bool pi_enable, struct fast_reg_descriptor *desc) argument [all...] |
/drivers/infiniband/hw/mlx4/ |
H A D | mlx4_ib.h | 405 struct ib_device *ib_dev; 415 struct ib_device *ib_dev; 498 struct ib_device ib_dev; 545 static inline struct mlx4_ib_dev *to_mdev(struct ib_device *ibdev) 645 struct ib_fast_reg_page_list *mlx4_ib_alloc_fast_reg_page_list(struct ib_device *ibdev, 651 struct ib_cq *mlx4_ib_create_cq(struct ib_device *ibdev, int entries, int vector, 691 int mlx4_ib_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num, 703 int __mlx4_ib_query_port(struct ib_device *ibdev, u8 port, 705 int __mlx4_ib_query_pkey(struct ib_device *ibdev, u8 port, u16 index, 708 int __mlx4_ib_query_gid(struct ib_device *ibde [all...] |