Lines Matching refs:ops
42 dmabuf->ops->release(dmabuf);
61 return dmabuf->ops->mmap(dmabuf, vma);
80 * Also connect the allocator specific data and ops to the buffer.
83 * @ops: [in] Attach allocator-defined dma buf ops to the new buffer.
89 * ops, or error in allocating struct dma_buf, will return negative error.
92 struct dma_buf *dma_buf_export(void *priv, const struct dma_buf_ops *ops,
98 if (WARN_ON(!priv || !ops
99 || !ops->map_dma_buf
100 || !ops->unmap_dma_buf
101 || !ops->release
102 || !ops->kmap_atomic
103 || !ops->kmap
104 || !ops->mmap)) {
113 dmabuf->ops = ops;
222 if (dmabuf->ops->attach) {
223 ret = dmabuf->ops->attach(dmabuf, dev, attach);
253 if (dmabuf->ops->detach)
254 dmabuf->ops->detach(dmabuf, attach);
282 sg_table = attach->dmabuf->ops->map_dma_buf(attach, direction);
304 attach->dmabuf->ops->unmap_dma_buf(attach, sg_table,
330 if (dmabuf->ops->begin_cpu_access)
331 ret = dmabuf->ops->begin_cpu_access(dmabuf, start, len, direction);
354 if (dmabuf->ops->end_cpu_access)
355 dmabuf->ops->end_cpu_access(dmabuf, start, len, direction);
372 return dmabuf->ops->kmap_atomic(dmabuf, page_num);
389 if (dmabuf->ops->kunmap_atomic)
390 dmabuf->ops->kunmap_atomic(dmabuf, page_num, vaddr);
407 return dmabuf->ops->kmap(dmabuf, page_num);
424 if (dmabuf->ops->kunmap)
425 dmabuf->ops->kunmap(dmabuf, page_num, vaddr);
468 return dmabuf->ops->mmap(dmabuf, vma);