Lines Matching defs:ttm

14 	struct ttm_dma_tt ttm;
20 nouveau_sgdma_destroy(struct ttm_tt *ttm)
22 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
24 if (ttm) {
26 ttm_dma_tt_fini(&nvbe->ttm);
32 nv04_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem)
34 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
44 for (i = 0; i < ttm->num_pages; i++) {
45 dma_addr_t dma_offset = nvbe->ttm.dma_address[i];
58 nv04_sgdma_unbind(struct ttm_tt *ttm)
60 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
68 if (ttm->state != tt_bound)
72 for (i = 0; i < ttm->num_pages; i++) {
99 nv41_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem)
101 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
104 dma_addr_t *list = nvbe->ttm.dma_address;
106 u32 cnt = ttm->num_pages;
120 nv41_sgdma_unbind(struct ttm_tt *ttm)
122 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
126 u32 cnt = ttm->num_pages;
144 nv44_sgdma_flush(struct ttm_tt *ttm)
146 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
149 nv_wr32(dev, 0x100814, (ttm->num_pages - 1) << 12);
208 nv44_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem)
210 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
213 dma_addr_t *list = nvbe->ttm.dma_address;
215 u32 cnt = ttm->num_pages;
243 nv44_sgdma_flush(ttm);
248 nv44_sgdma_unbind(struct ttm_tt *ttm)
250 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
254 u32 cnt = ttm->num_pages;
276 nv44_sgdma_flush(ttm);
287 nv50_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem)
289 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
293 node->pages = nvbe->ttm.dma_address;
298 nv50_sgdma_unbind(struct ttm_tt *ttm)
324 nvbe->ttm.ttm.func = dev_priv->gart_info.func;
326 if (ttm_dma_tt_init(&nvbe->ttm, bdev, size, page_flags, dummy_read_page)) {
330 return &nvbe->ttm.ttm;