Searched refs:ib_device (Results 1 - 25 of 107) sorted by relevance

12345

/drivers/infiniband/core/
H A Dagent.h43 extern int ib_agent_port_open(struct ib_device *device, int port_num);
45 extern int ib_agent_port_close(struct ib_device *device, int port_num);
48 struct ib_wc *wc, struct ib_device *device,
H A Dcore_priv.h41 int ib_device_register_sysfs(struct ib_device *device,
42 int (*port_callback)(struct ib_device *,
44 void ib_device_unregister_sysfs(struct ib_device *device);
H A Dsmi.h66 struct ib_device *device)
81 struct ib_device *device)
H A Ddevice.c70 static int ib_device_check_mandatory(struct ib_device *device)
72 #define IB_MANDATORY_FUNC(x) { offsetof(struct ib_device, x), #x }
110 static struct ib_device *__ib_device_get_by_name(const char *name)
112 struct ib_device *device;
126 struct ib_device *device;
154 static int start_port(struct ib_device *device)
160 static int end_port(struct ib_device *device)
171 * ib_device. @size is the size of the structure to be allocated,
176 struct ib_device *ib_alloc_device(size_t size)
178 BUG_ON(size < sizeof (struct ib_device));
[all...]
H A Dsa.h52 struct ib_device *device, u8 port_num,
H A Dagent.c57 __ib_get_agent_port(struct ib_device *device, int port_num)
70 ib_get_agent_port(struct ib_device *device, int port_num)
82 struct ib_wc *wc, struct ib_device *device,
145 int ib_agent_port_open(struct ib_device *device, int port_num)
196 int ib_agent_port_close(struct ib_device *device, int port_num)
H A Dcache.c57 struct ib_device *device;
61 static inline int start_port(struct ib_device *device)
66 static inline int end_port(struct ib_device *device)
72 int ib_get_cached_gid(struct ib_device *device,
99 int ib_find_cached_gid(struct ib_device *device,
134 int ib_get_cached_pkey(struct ib_device *device,
161 int ib_find_cached_pkey(struct ib_device *device,
202 int ib_find_exact_cached_pkey(struct ib_device *device,
234 int ib_get_cached_lmc(struct ib_device *device,
252 static void ib_cache_update(struct ib_device *devic
[all...]
H A Dsysfs.c45 struct ib_device *ibdev;
457 struct ib_device *dev = container_of(device, struct ib_device, dev);
465 struct ib_device *dev = container_of(device, struct ib_device, dev);
520 static int add_port(struct ib_device *device, int port_num,
521 int (*port_callback)(struct ib_device *,
617 struct ib_device *dev = container_of(device, struct ib_device, dev);
633 struct ib_device *de
[all...]
/drivers/infiniband/hw/ipath/
H A Dipath_dma.c50 static int ipath_mapping_error(struct ib_device *dev, u64 dma_addr)
55 static u64 ipath_dma_map_single(struct ib_device *dev,
63 static void ipath_dma_unmap_single(struct ib_device *dev,
70 static u64 ipath_dma_map_page(struct ib_device *dev,
94 static void ipath_dma_unmap_page(struct ib_device *dev,
101 static int ipath_map_sg(struct ib_device *dev, struct scatterlist *sgl,
126 static void ipath_unmap_sg(struct ib_device *dev,
133 static void ipath_sync_single_for_cpu(struct ib_device *dev,
140 static void ipath_sync_single_for_device(struct ib_device *dev,
147 static void *ipath_dma_alloc_coherent(struct ib_device *de
[all...]
/drivers/infiniband/hw/qib/
H A Dqib_dma.c48 static int qib_mapping_error(struct ib_device *dev, u64 dma_addr)
53 static u64 qib_dma_map_single(struct ib_device *dev, void *cpu_addr,
60 static void qib_dma_unmap_single(struct ib_device *dev, u64 addr, size_t size,
66 static u64 qib_dma_map_page(struct ib_device *dev, struct page *page,
88 static void qib_dma_unmap_page(struct ib_device *dev, u64 addr, size_t size,
94 static int qib_map_sg(struct ib_device *dev, struct scatterlist *sgl,
119 static void qib_unmap_sg(struct ib_device *dev,
126 static void qib_sync_single_for_cpu(struct ib_device *dev, u64 addr,
131 static void qib_sync_single_for_device(struct ib_device *dev, u64 addr,
137 static void *qib_dma_alloc_coherent(struct ib_device *de
[all...]
/drivers/infiniband/hw/usnic/
H A Dusnic_ib_verbs.h24 enum rdma_link_layer usnic_ib_port_link_layer(struct ib_device *device,
26 int usnic_ib_query_device(struct ib_device *ibdev,
28 int usnic_ib_query_port(struct ib_device *ibdev, u8 port,
33 int usnic_ib_query_gid(struct ib_device *ibdev, u8 port, int index,
35 int usnic_ib_query_pkey(struct ib_device *ibdev, u8 port, u16 index,
37 struct ib_pd *usnic_ib_alloc_pd(struct ib_device *ibdev,
47 struct ib_cq *usnic_ib_create_cq(struct ib_device *ibdev, int entries,
55 struct ib_ucontext *usnic_ib_alloc_ucontext(struct ib_device *ibdev,
H A Dusnic_ib.h55 struct ib_device ib_dev;
82 struct usnic_ib_dev *to_usdev(struct ib_device *ibdev)
/drivers/infiniband/hw/ehca/
H A Dehca_main.c417 ehca_err(&shca->ib_device, "Can't allocate rblock memory.");
422 ehca_err(&shca->ib_device, "Can't query device properties");
427 memcpy(&shca->ib_device.node_guid, &rblock->node_guid, sizeof(u64));
442 strlcpy(shca->ib_device.name, "ehca%d", IB_DEVICE_NAME_MAX);
443 shca->ib_device.owner = THIS_MODULE;
445 shca->ib_device.uverbs_abi_ver = 8;
446 shca->ib_device.uverbs_cmd_mask =
464 shca->ib_device.node_type = RDMA_NODE_IB_CA;
465 shca->ib_device.phys_port_cnt = shca->num_ports;
466 shca->ib_device
[all...]
H A Dehca_hca.c53 int ehca_query_device(struct ib_device *ibdev, struct ib_device_attr *props)
57 ib_device);
76 ehca_err(&shca->ib_device, "Can't allocate rblock memory.");
81 ehca_err(&shca->ib_device, "Can't query device properties");
156 ehca_err(&shca->ib_device, "Unknown MTU size: %x.",
176 ehca_err(&shca->ib_device, "invalid Vl Capability: %x.",
182 int ehca_query_port(struct ib_device *ibdev,
188 ib_device);
193 ehca_err(&shca->ib_device, "Can't allocate rblock memory.");
199 ehca_err(&shca->ib_device, "Ca
[all...]
H A Dehca_irq.c104 ehca_err(&shca->ib_device,
113 ehca_err(&shca->ib_device,
119 ehca_err(&shca->ib_device,
121 type, shca->ib_device.name);
125 ehca_err(&shca->ib_device, "Error data is available: %llx.", resource);
126 ehca_err(&shca->ib_device, "EHCA ----- error data begin "
129 ehca_err(&shca->ib_device, "EHCA ----- error data end "
145 ehca_err(&shca->ib_device, "Cannot allocate rblock memory.");
157 ehca_err(&shca->ib_device,
169 ehca_err(&shca->ib_device,
[all...]
H A Dehca_iverbs.h47 int ehca_query_device(struct ib_device *ibdev, struct ib_device_attr *props);
49 int ehca_query_port(struct ib_device *ibdev, u8 port,
55 int ehca_query_pkey(struct ib_device *ibdev, u8 port, u16 index, u16 * pkey);
57 int ehca_query_gid(struct ib_device *ibdev, u8 port, int index,
60 int ehca_modify_port(struct ib_device *ibdev, u8 port, int port_modify_mask,
63 struct ib_pd *ehca_alloc_pd(struct ib_device *device,
129 struct ib_cq *ehca_create_cq(struct ib_device *device, int cqe, int comp_vector,
183 struct ib_ucontext *ehca_alloc_ucontext(struct ib_device *device,
190 int ehca_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num,
H A Dehca_pd.c48 struct ib_pd *ehca_alloc_pd(struct ib_device *device,
77 ib_device);
H A Dehca_sqp.c87 ehca_err(&shca->ib_device,
93 ehca_dbg(&shca->ib_device, "port=%x pma_qp_nr=%x",
97 ehca_err(&shca->ib_device, "invalid qp_type=%x",
109 ehca_dbg(&shca->ib_device, "... wait until port %x is active",
115 ehca_err(&shca->ib_device, "Port %x is not active.", port);
142 static int ehca_process_perf(struct ib_device *ibdev, u8 port_num,
153 container_of(ibdev, struct ehca_shca, ib_device);
219 int ehca_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num,
H A Dehca_eq.c62 struct ib_device *ib_dev = &shca->ib_device;
183 ehca_err(&shca->ib_device, "Can't free EQ resources.");
H A Dehca_mrmw.c163 container_of(pd->device, struct ehca_shca, ib_device);
168 ehca_err(&shca->ib_device, "out of memory");
185 ehca_err(&shca->ib_device, "no internal max-MR exist!");
192 ehca_err(&shca->ib_device, "h_ret=%li pd=%p mr_access_flags=%x",
209 container_of(pd->device, struct ehca_shca, ib_device);
320 container_of(pd->device, struct ehca_shca, ib_device);
451 container_of(mr->device, struct ehca_shca, ib_device);
599 container_of(mr->device, struct ehca_shca, ib_device);
646 container_of(mr->device, struct ehca_shca, ib_device);
695 container_of(pd->device, struct ehca_shca, ib_device);
[all...]
/drivers/infiniband/hw/ocrdma/
H A Docrdma_verbs.h39 int ocrdma_query_device(struct ib_device *, struct ib_device_attr *props);
40 int ocrdma_query_port(struct ib_device *, u8 port, struct ib_port_attr *props);
41 int ocrdma_modify_port(struct ib_device *, u8 port, int mask,
45 int ocrdma_query_gid(struct ib_device *, u8 port,
47 int ocrdma_query_pkey(struct ib_device *, u8 port, u16 index, u16 *pkey);
49 struct ib_ucontext *ocrdma_alloc_ucontext(struct ib_device *,
55 struct ib_pd *ocrdma_alloc_pd(struct ib_device *,
59 struct ib_cq *ocrdma_create_cq(struct ib_device *, int entries, int vector,
94 struct ib_fast_reg_page_list *ocrdma_alloc_frmr_page_list(struct ib_device
H A Docrdma_ah.h36 int ocrdma_process_mad(struct ib_device *,
/drivers/infiniband/ulp/iser/
H A Diser_initiator.c168 ib_dma_sync_single_for_cpu(device->ib_device,
190 ib_dma_unmap_single(device->ib_device,
195 ib_dma_unmap_single(device->ib_device,
223 iser_conn->login_req_dma = ib_dma_map_single(device->ib_device,
228 iser_conn->login_resp_dma = ib_dma_map_single(device->ib_device,
233 req_err = ib_dma_mapping_error(device->ib_device,
235 resp_err = ib_dma_mapping_error(device->ib_device,
284 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc,
286 if (ib_dma_mapping_error(device->ib_device, dma_addr))
303 ib_dma_unmap_single(device->ib_device, rx_des
[all...]
H A Diser_verbs.c81 ret = ib_query_device(device->ib_device, dev_attr);
83 pr_warn("Query device failed for %s\n", device->ib_device->name);
88 if (device->ib_device->alloc_fmr && device->ib_device->dealloc_fmr &&
89 device->ib_device->map_phys_fmr && device->ib_device->unmap_fmr) {
108 device->ib_device->num_comp_vectors);
110 device->comps_used, device->ib_device->name,
111 device->ib_device->num_comp_vectors);
113 device->pd = ib_alloc_pd(device->ib_device);
264 iser_create_fastreg_desc(struct ib_device *ib_device, struct ib_pd *pd, bool pi_enable, struct fast_reg_descriptor *desc) argument
[all...]
/drivers/infiniband/hw/mlx4/
H A Dmlx4_ib.h405 struct ib_device *ib_dev;
415 struct ib_device *ib_dev;
498 struct ib_device ib_dev;
545 static inline struct mlx4_ib_dev *to_mdev(struct ib_device *ibdev)
645 struct ib_fast_reg_page_list *mlx4_ib_alloc_fast_reg_page_list(struct ib_device *ibdev,
651 struct ib_cq *mlx4_ib_create_cq(struct ib_device *ibdev, int entries, int vector,
691 int mlx4_ib_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num,
703 int __mlx4_ib_query_port(struct ib_device *ibdev, u8 port,
705 int __mlx4_ib_query_pkey(struct ib_device *ibdev, u8 port, u16 index,
708 int __mlx4_ib_query_gid(struct ib_device *ibde
[all...]

Completed in 1289 milliseconds

12345