Lines Matching refs:mmr

67 			    ~0ull, convert_access(acc), 0, 0, &mr->mmr);
71 err = mlx4_mr_enable(to_mdev(pd->device)->dev, &mr->mmr);
75 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key;
81 (void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr);
160 convert_access(access_flags), n, shift, &mr->mmr);
164 err = mlx4_ib_umem_write_mtt(dev, &mr->mmr.mtt, mr->umem);
168 err = mlx4_mr_enable(dev->dev, &mr->mmr);
172 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key;
177 (void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr);
194 struct mlx4_ib_mr *mmr = to_mmr(mr);
203 err = mlx4_mr_hw_get_mpt(dev->dev, &mmr->mmr, &pmpt_entry);
229 mlx4_mr_rereg_mem_cleanup(dev->dev, &mmr->mmr);
230 ib_umem_release(mmr->umem);
231 mmr->umem = ib_umem_get(mr->uobject->context, start, length,
235 if (IS_ERR(mmr->umem)) {
236 err = PTR_ERR(mmr->umem);
238 mmr->umem = NULL;
241 n = ib_umem_page_count(mmr->umem);
242 shift = ilog2(mmr->umem->page_size);
244 err = mlx4_mr_rereg_mem_write(dev->dev, &mmr->mmr,
248 ib_umem_release(mmr->umem);
251 mmr->mmr.iova = virt_addr;
252 mmr->mmr.size = length;
254 err = mlx4_ib_umem_write_mtt(dev, &mmr->mmr.mtt, mmr->umem);
256 mlx4_mr_rereg_mem_cleanup(dev->dev, &mmr->mmr);
257 ib_umem_release(mmr->umem);
265 err = mlx4_mr_hw_write_mpt(dev->dev, &mmr->mmr, pmpt_entry);
267 mmr->mmr.access = mr_access_flags;
280 ret = mlx4_mr_free(to_mdev(ibmr->device)->dev, &mr->mmr);
366 max_page_list_len, 0, &mr->mmr);
370 err = mlx4_mr_enable(dev->dev, &mr->mmr);
374 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key;
380 (void) mlx4_mr_free(dev->dev, &mr->mmr);