/drivers/gpu/drm/tegra/ |
H A D | gem.c | 20 #include "gem.h" 30 struct drm_device *drm = obj->gem.dev; 33 drm_gem_object_unreference(&obj->gem); 74 struct drm_device *drm = obj->gem.dev; 77 drm_gem_object_reference(&obj->gem); 96 dma_free_writecombine(drm->dev, bo->gem.size, bo->vaddr, bo->paddr); 121 err = drm_gem_object_init(drm, &bo->gem, size); 125 err = drm_gem_create_mmap_offset(&bo->gem); 138 drm_gem_object_release(&bo->gem); 160 ret = drm_gem_handle_create(file, &bo->gem, handl 241 tegra_bo_free_object(struct drm_gem_object *gem) argument 284 struct drm_gem_object *gem; local 314 struct drm_gem_object *gem; local 337 struct drm_gem_object *gem = attach->dmabuf->priv; local 398 struct drm_gem_object *gem = buf->priv; local 421 tegra_gem_prime_export(struct drm_device *drm, struct drm_gem_object *gem, int flags) argument 435 struct drm_gem_object *gem = buf->priv; local [all...] |
H A D | gem.h | 34 struct drm_gem_object gem; member in struct:tegra_bo 44 static inline struct tegra_bo *to_tegra_bo(struct drm_gem_object *gem) argument 46 return container_of(gem, struct tegra_bo, gem); 56 void tegra_bo_free_object(struct drm_gem_object *gem); 67 struct drm_gem_object *gem,
|
H A D | Makefile | 5 gem.o \
|
H A D | drm.c | 13 #include "gem.h" 120 struct drm_gem_object *gem; local 123 gem = drm_gem_object_lookup(drm, file, handle); 124 if (!gem) 128 drm_gem_object_unreference(gem); 131 bo = to_tegra_bo(gem); 312 struct drm_gem_object *gem; local 315 gem = drm_gem_object_lookup(drm, file, args->handle); 316 if (!gem) 319 bo = to_tegra_bo(gem); 490 struct drm_gem_object *gem; local 542 struct drm_gem_object *gem; local 582 struct drm_gem_object *gem; local 607 struct drm_gem_object *gem; local [all...] |
H A D | fb.c | 14 #include "gem.h" 69 drm_gem_object_unreference_unlocked(&bo->gem); 82 return drm_gem_handle_create(file, &fb->planes[0]->gem, handle); 134 struct drm_gem_object *gem; local 146 gem = drm_gem_object_lookup(drm, file, cmd->handles[i]); 147 if (!gem) { 157 if (gem->size < size) { 162 planes[i] = to_tegra_bo(gem); 175 drm_gem_object_unreference_unlocked(&planes[i]->gem); 226 tegra_bo_free_object(&bo->gem); [all...] |
/drivers/gpu/drm/nouveau/ |
H A D | nouveau_gem.h | 13 nouveau_gem_object(struct drm_gem_object *gem) argument 15 return gem ? container_of(gem, struct nouveau_bo, gem) : NULL;
|
H A D | nouveau_gem.c | 36 nouveau_gem_object_del(struct drm_gem_object *gem) argument 38 struct nouveau_bo *nvbo = nouveau_gem_object(gem); 41 if (gem->import_attach) 42 drm_prime_gem_destroy(gem, nvbo->bo.sg); 44 drm_gem_object_release(gem); 47 gem->filp = NULL; 52 nouveau_gem_object_open(struct drm_gem_object *gem, struct drm_file *file_priv) argument 55 struct nouveau_bo *nvbo = nouveau_gem_object(gem); 128 nouveau_gem_object_close(struct drm_gem_object *gem, struct drm_file *file_priv) argument 131 struct nouveau_bo *nvbo = nouveau_gem_object(gem); 195 nouveau_gem_info(struct drm_file *file_priv, struct drm_gem_object *gem, struct drm_nouveau_gem_info *rep) argument 258 nouveau_gem_set_domain(struct drm_gem_object *gem, uint32_t read_domains, uint32_t write_domains, uint32_t valid_domains) argument 357 struct drm_gem_object *gem; local 846 struct drm_gem_object *gem; local 887 struct drm_gem_object *gem; local [all...] |
H A D | nouveau_prime.c | 79 /* Initialize the embedded gem-object. We return a single gem-reference 81 ret = drm_gem_object_init(dev, &nvbo->gem, nvbo->bo.mem.size); 87 return &nvbo->gem;
|
H A D | nouveau_ttm.c | 389 drm->gem.vram_available = drm->device.info.ram_user; 392 drm->gem.vram_available >> PAGE_SHIFT); 403 drm->gem.gart_available = nvkm_vmmgr(&drm->device)->limit; 405 drm->gem.gart_available = drm->agp.size; 409 drm->gem.gart_available >> PAGE_SHIFT); 415 NV_INFO(drm, "VRAM: %d MiB\n", (u32)(drm->gem.vram_available >> 20)); 416 NV_INFO(drm, "GART: %d MiB\n", (u32)(drm->gem.gart_available >> 20));
|
H A D | nouveau_display.c | 222 drm_gem_object_unreference_unlocked(&fb->nvbo->gem); 235 return drm_gem_handle_create(file_priv, &fb->nvbo->gem, handle); 275 struct drm_gem_object *gem; local 278 gem = drm_gem_object_lookup(dev, file_priv, mode_cmd->handles[0]); 279 if (!gem) 286 ret = nouveau_framebuffer_init(dev, nouveau_fb, mode_cmd, nouveau_gem_object(gem)); 295 drm_gem_object_unreference(gem); 874 ret = drm_gem_handle_create(file_priv, &bo->gem, &args->handle); 875 drm_gem_object_unreference_unlocked(&bo->gem); 884 struct drm_gem_object *gem; local [all...] |
H A D | nouveau_bo.h | 33 * gem reference to it! For debugging, use gem.filp != NULL to test 35 struct drm_gem_object gem; member in struct:nouveau_bo
|
/drivers/gpu/drm/gma500/ |
H A D | gem.c | 34 struct gtt_range *gtt = container_of(obj, struct gtt_range, gem); 109 r = psb_gtt_alloc_range(dev, size, "gem", 0, PAGE_SIZE); 115 if (drm_gem_object_init(dev, &r->gem, size) != 0) { 122 mapping_set_gfp_mask(r->gem.filp->f_mapping, GFP_KERNEL | __GFP_DMA32); 124 ret = drm_gem_handle_create(file, &r->gem, &handle); 127 &r->gem, size); 128 drm_gem_object_release(&r->gem); 133 drm_gem_object_unreference(&r->gem); 188 r = container_of(obj, struct gtt_range, gem); /* Get the gtt range */
|
H A D | gtt.h | 47 struct drm_gem_object gem; /* GEM high level stuff */ member in struct:gtt_range
|
H A D | Makefile | 10 gem.o \
|
H A D | gtt.c | 209 pages = drm_gem_get_pages(>->gem); 213 gt->npage = gt->gem.size / PAGE_SIZE; 230 drm_gem_put_pages(>->gem, gt->pages, true, false); 247 struct drm_device *dev = gt->gem.dev; 285 struct drm_device *dev = gt->gem.dev; 359 gt->gem.dev = dev;
|
H A D | framebuffer.c | 324 drm_gem_private_object_init(dev, &backing->gem, aligned_size); 485 drm_gem_object_unreference(&backing->gem); 516 r = container_of(obj, struct gtt_range, gem); 588 drm_gem_object_unreference(&psbfb->gtt->gem); 654 return drm_gem_handle_create(file_priv, &r->gem, handle); 672 drm_gem_object_unreference_unlocked(&r->gem);
|
/drivers/gpu/drm/qxl/ |
H A D | qxl_gem.c | 62 mutex_lock(&qdev->gem.mutex); 63 list_add_tail(&qbo->list, &qdev->gem.objects); 64 mutex_unlock(&qdev->gem.mutex); 110 INIT_LIST_HEAD(&qdev->gem.objects);
|
H A D | qxl_object.c | 39 mutex_lock(&qdev->gem.mutex); 41 mutex_unlock(&qdev->gem.mutex); 274 if (list_empty(&qdev->gem.objects)) 277 list_for_each_entry_safe(bo, n, &qdev->gem.objects, list) { 282 mutex_lock(&qdev->gem.mutex); 284 mutex_unlock(&qdev->gem.mutex);
|
/drivers/net/ethernet/sun/ |
H A D | sungem.c | 86 #define GEM_MODULE_NAME "gem" 118 static u16 __sungem_phy_read(struct gem *gp, int phy_addr, int reg) 146 struct gem *gp = netdev_priv(dev); 150 static inline u16 sungem_phy_read(struct gem *gp, int reg) 155 static void __sungem_phy_write(struct gem *gp, int phy_addr, int reg, u16 val) 179 struct gem *gp = netdev_priv(dev); 183 static inline void sungem_phy_write(struct gem *gp, int reg, u16 val) 188 static inline void gem_enable_ints(struct gem *gp) 194 static inline void gem_disable_ints(struct gem *gp) 201 static void gem_get_cell(struct gem *g [all...] |
/drivers/gpu/drm/radeon/ |
H A D | radeon_prime.c | 76 mutex_lock(&rdev->gem.mutex); 77 list_add_tail(&bo->list, &rdev->gem.objects); 78 mutex_unlock(&rdev->gem.mutex);
|
H A D | radeon_gem.c | 85 mutex_lock(&rdev->gem.mutex); 86 list_add_tail(&robj->list, &rdev->gem.objects); 87 mutex_unlock(&rdev->gem.mutex); 127 INIT_LIST_HEAD(&rdev->gem.objects); 255 /* create a gem object to contain this object in */ 312 /* create a gem object to contain this object in */ 703 mutex_lock(&rdev->gem.mutex); 704 list_for_each_entry(rbo, &rdev->gem.objects, list) { 726 mutex_unlock(&rdev->gem.mutex);
|
H A D | radeon_object.c | 80 mutex_lock(&bo->rdev->gem.mutex); 82 mutex_unlock(&bo->rdev->gem.mutex); 397 if (list_empty(&rdev->gem.objects)) { 401 list_for_each_entry_safe(bo, n, &rdev->gem.objects, list) { 406 mutex_lock(&bo->rdev->gem.mutex); 408 mutex_unlock(&bo->rdev->gem.mutex);
|
/drivers/gpu/drm/bochs/ |
H A D | bochs.h | 103 struct drm_gem_object gem; member in struct:bochs_bo 113 static inline struct bochs_bo *gem_to_bochs_bo(struct drm_gem_object *gem) argument 115 return container_of(gem, struct bochs_bo, gem);
|
/drivers/gpu/drm/shmobile/ |
H A D | shmob_drm_plane.c | 49 struct drm_gem_cma_object *gem; local 53 gem = drm_fb_cma_get_gem_obj(fb, 0); 54 splane->dma[0] = gem->paddr + fb->offsets[0] 59 gem = drm_fb_cma_get_gem_obj(fb, 1); 60 splane->dma[1] = gem->paddr + fb->offsets[1]
|
/drivers/gpu/drm/rcar-du/ |
H A D | rcar_du_plane.c | 143 struct drm_gem_cma_object *gem; local 147 gem = drm_fb_cma_get_gem_obj(fb, 0); 148 plane->dma[0] = gem->paddr + fb->offsets[0]; 151 gem = drm_fb_cma_get_gem_obj(fb, 1); 152 plane->dma[1] = gem->paddr + fb->offsets[1];
|