/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_dma_buf.c | 51 struct xe_bo *bo = gem_to_xe_bo(obj); local 52 struct xe_device *xe = xe_bo_device(bo); 60 if (xe_bo_is_pinned(bo) && bo->ttm.resource->placement != XE_PL_TT) { 61 drm_dbg(&xe->drm, "Can't migrate pinned bo for dma-buf pin.\n"); 65 ret = xe_bo_migrate(bo, XE_PL_TT); 74 ret = xe_bo_pin_external(bo); 83 struct xe_bo *bo = gem_to_xe_bo(obj); local 85 xe_bo_unpin_external(bo); 93 struct xe_bo *bo local 148 struct xe_bo *bo = gem_to_xe_bo(dma_buf->priv); local 163 struct xe_bo *bo = gem_to_xe_bo(obj); local 194 struct xe_bo *bo = gem_to_xe_bo(obj); local 213 struct xe_bo *bo; local 236 struct xe_bo *bo = gem_to_xe_bo(obj); local 268 struct xe_bo *bo; local [all...] |
H A D | xe_guc_log.c | 58 xe_assert(xe, log->bo); 60 size = log->bo->size; 67 xe_map_memcpy_from(xe, read, &log->bo->vmap, i * sizeof(u32), 84 struct xe_bo *bo; local 86 bo = xe_managed_bo_create_pin_map(xe, tile, guc_log_size(), 89 if (IS_ERR(bo)) 90 return PTR_ERR(bo); 92 xe_map_memset(xe, &bo->vmap, 0, 0, guc_log_size()); 93 log->bo = bo; [all...] |
H A D | xe_bb_types.h | 14 struct drm_suballoc *bo; member in struct:xe_bb
|
H A D | xe_guc_ads_types.h | 17 /** @bo: XE BO for GuC ads blob */ 18 struct xe_bo *bo; member in struct:xe_guc_ads
|
H A D | xe_guc_log_types.h | 19 /** @bo: XE BO for GuC log */ 20 struct xe_bo *bo; member in struct:xe_guc_log
|
H A D | xe_sa_types.h | 14 struct xe_bo *bo; member in struct:xe_sa_manager
|
H A D | xe_drm_client.c | 66 * @bo: The xe BO ptr 74 struct xe_bo *bo) 76 XE_WARN_ON(bo->client); 77 XE_WARN_ON(!list_empty(&bo->client_link)); 80 bo->client = xe_drm_client_get(client); 81 list_add_tail_rcu(&bo->client_link, &client->bos_list); 87 * @bo: The xe BO ptr 94 void xe_drm_client_remove_bo(struct xe_bo *bo) argument 96 struct xe_drm_client *client = bo->client; 99 list_del_rcu(&bo 73 xe_drm_client_add_bo(struct xe_drm_client *client, struct xe_bo *bo) argument 105 bo_meminfo(struct xe_bo *bo, struct drm_memory_stats stats[TTM_NUM_MEM_TYPES]) argument 140 struct xe_bo *bo; local 149 struct xe_bo *bo = gem_to_xe_bo(obj); local [all...] |
H A D | xe_guc_hwconfig.c | 42 int ret = send_get_hwconfig(guc, xe_bo_ggtt_addr(guc->hwconfig.bo), 56 struct xe_bo *bo; local 61 if (guc->hwconfig.bo) 80 bo = xe_managed_bo_create_pin_map(xe, tile, PAGE_ALIGN(size), 83 if (IS_ERR(bo)) 84 return PTR_ERR(bo); 85 guc->hwconfig.bo = bo; 93 return !guc->hwconfig.bo ? 0 : guc->hwconfig.size; 100 XE_WARN_ON(!guc->hwconfig.bo); [all...] |
/linux-master/drivers/gpu/drm/ttm/ |
H A D | ttm_execbuf_util.c | 36 struct ttm_buffer_object *bo = entry->bo; local 38 dma_resv_unlock(bo->base.resv); 51 struct ttm_buffer_object *bo = entry->bo; local 53 ttm_bo_move_to_lru_tail_unlocked(bo); 54 dma_resv_unlock(bo->base.resv); 88 struct ttm_buffer_object *bo = entry->bo; local 91 ret = ttm_bo_reserve(bo, int 151 struct ttm_buffer_object *bo = entry->bo; local [all...] |
/linux-master/drivers/gpu/drm/ttm/tests/ |
H A D | ttm_tt_test.c | 55 struct ttm_buffer_object *bo; local 66 bo = ttm_bo_kunit_init(test, test->priv, params->size); 68 err = ttm_tt_init(tt, bo, page_flags, caching, extra_pages); 82 struct ttm_buffer_object *bo; local 92 bo = ttm_bo_kunit_init(test, test->priv, size); 95 bo->base.size += 1; 97 err = ttm_tt_init(tt, bo, 0, caching, 0); 105 struct ttm_buffer_object *bo; local 113 bo = ttm_bo_kunit_init(test, test->priv, BO_SIZE); 115 err = ttm_tt_init(tt, bo, 125 struct ttm_buffer_object *bo; local 145 struct ttm_buffer_object *bo; local 168 struct ttm_buffer_object *bo; local 187 struct ttm_buffer_object *bo; local 203 struct ttm_buffer_object *bo; local 226 ttm_tt_null_create(struct ttm_buffer_object *bo, uint32_t page_flags) argument 239 struct ttm_buffer_object *bo; local 257 struct ttm_buffer_object *bo; local [all...] |
H A D | ttm_bo_test.c | 56 struct ttm_buffer_object *bo; local 59 bo = ttm_bo_kunit_init(test, test->priv, BO_SIZE); 61 err = ttm_bo_reserve(bo, params->interruptible, params->no_wait, NULL); 64 dma_resv_unlock(bo->base.resv); 69 struct ttm_buffer_object *bo; local 74 bo = ttm_bo_kunit_init(test, test->priv, BO_SIZE); 77 dma_resv_lock(bo->base.resv, NULL); 79 err = ttm_bo_reserve(bo, interruptible, no_wait, NULL); 80 dma_resv_unlock(bo->base.resv); 87 struct ttm_buffer_object *bo; local 105 struct ttm_buffer_object *bo; local 180 struct ttm_buffer_object *bo = arg; local 207 struct ttm_buffer_object *bo; local 234 struct ttm_buffer_object *bo; local 277 struct ttm_buffer_object *bo; local 371 struct ttm_buffer_object *bo; local 415 struct ttm_buffer_object *bo; local 458 struct ttm_buffer_object *bo; local 486 struct ttm_buffer_object *bo; local 537 struct ttm_buffer_object *bo; local [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_dma_buf.c | 59 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); local 60 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); 91 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); local 92 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); 109 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); local 112 return amdgpu_bo_pin(bo, AMDGPU_GEM_DOMAIN_GTT); 125 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); local 127 amdgpu_bo_unpin(bo); 148 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); local 149 struct amdgpu_device *adev = amdgpu_ttm_adev(bo 242 struct amdgpu_bo *bo = gem_to_amdgpu_bo(dma_buf->priv); local 295 struct amdgpu_bo *bo = gem_to_amdgpu_bo(gobj); local 327 struct amdgpu_bo *bo; local 373 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); local 488 amdgpu_dmabuf_is_xgmi_accessible(struct amdgpu_device *adev, struct amdgpu_bo *bo) argument [all...] |
H A D | amdgpu_object.h | 45 #define to_amdgpu_bo_user(abo) container_of((abo), struct amdgpu_bo_user, bo) 46 #define to_amdgpu_bo_vm(abo) container_of((abo), struct amdgpu_bo_vm, bo) 58 void (*destroy)(struct ttm_buffer_object *bo); 63 /* bo virtual addresses in a vm */ 79 /* protected by bo being reserved */ 123 struct amdgpu_bo bo; member in struct:amdgpu_bo_user 132 struct amdgpu_bo bo; member in struct:amdgpu_bo_vm 200 * amdgpu_bo_reserve - reserve bo 201 * @bo: bo structur 208 amdgpu_bo_reserve(struct amdgpu_bo *bo, bool no_intr) argument 222 amdgpu_bo_unreserve(struct amdgpu_bo *bo) argument 227 amdgpu_bo_size(struct amdgpu_bo *bo) argument 232 amdgpu_bo_ngpu_pages(struct amdgpu_bo *bo) argument 237 amdgpu_bo_gpu_page_alignment(struct amdgpu_bo *bo) argument 248 amdgpu_bo_mmap_offset(struct amdgpu_bo *bo) argument 256 amdgpu_bo_explicit_sync(struct amdgpu_bo *bo) argument 267 amdgpu_bo_encrypted(struct amdgpu_bo *bo) argument 280 amdgpu_bo_shadowed(struct amdgpu_bo *bo) argument [all...] |
H A D | amdgpu_hmm.h | 41 int amdgpu_hmm_register(struct amdgpu_bo *bo, unsigned long addr); 42 void amdgpu_hmm_unregister(struct amdgpu_bo *bo); 44 static inline int amdgpu_hmm_register(struct amdgpu_bo *bo, unsigned long addr) argument 50 static inline void amdgpu_hmm_unregister(struct amdgpu_bo *bo) {} argument
|
H A D | amdgpu_csa.h | 32 int amdgpu_allocate_static_csa(struct amdgpu_device *adev, struct amdgpu_bo **bo, 35 struct amdgpu_bo *bo, struct amdgpu_bo_va **bo_va, 38 struct amdgpu_bo *bo, struct amdgpu_bo_va *bo_va, 40 void amdgpu_free_static_csa(struct amdgpu_bo **bo);
|
/linux-master/drivers/gpu/drm/tegra/ |
H A D | gem.c | 52 static void tegra_bo_put(struct host1x_bo *bo) argument 54 struct tegra_bo *obj = host1x_to_tegra_bo(bo); 59 static struct host1x_bo_mapping *tegra_bo_pin(struct device *dev, struct host1x_bo *bo, argument 62 struct tegra_bo *obj = host1x_to_tegra_bo(bo); 72 map->bo = host1x_bo_get(bo); 173 host1x_bo_put(map->bo); 177 static void *tegra_bo_mmap(struct host1x_bo *bo) argument 179 struct tegra_bo *obj = host1x_to_tegra_bo(bo); 203 static void tegra_bo_munmap(struct host1x_bo *bo, voi argument 217 tegra_bo_get(struct host1x_bo *bo) argument 235 tegra_bo_iommu_map(struct tegra_drm *tegra, struct tegra_bo *bo) argument 278 tegra_bo_iommu_unmap(struct tegra_drm *tegra, struct tegra_bo *bo) argument 302 struct tegra_bo *bo; local 331 tegra_bo_free(struct drm_device *drm, struct tegra_bo *bo) argument 343 tegra_bo_get_pages(struct drm_device *drm, struct tegra_bo *bo) argument 373 tegra_bo_alloc(struct drm_device *drm, struct tegra_bo *bo) argument 407 struct tegra_bo *bo; local 438 struct tegra_bo *bo; local 461 struct tegra_bo *bo; local 508 struct tegra_bo *bo = to_tegra_bo(gem); local 539 struct tegra_bo *bo; local 556 struct tegra_bo *bo = to_tegra_bo(gem); local 577 struct tegra_bo *bo = to_tegra_bo(gem); local 629 struct tegra_bo *bo = to_tegra_bo(gem); local 662 struct tegra_bo *bo = to_tegra_bo(gem); local 680 struct tegra_bo *bo = to_tegra_bo(gem); local 693 struct tegra_bo *bo = to_tegra_bo(gem); local 717 struct tegra_bo *bo = to_tegra_bo(gem); local 732 struct tegra_bo *bo = to_tegra_bo(gem); local 766 struct tegra_bo *bo; local 787 struct tegra_bo *bo; local [all...] |
/linux-master/drivers/gpu/drm/v3d/ |
H A D | v3d_bo.c | 34 struct v3d_bo *bo = to_v3d_bo(obj); local 36 if (bo->vaddr) 37 v3d_put_bo_vaddr(bo); 39 v3d_mmu_remove_ptes(bo); 47 drm_mm_remove_node(&bo->node); 51 bo->base.pages_mark_dirty_on_put = true; 53 drm_gem_shmem_free(&bo->base); 73 struct v3d_bo *bo; local 79 bo = kzalloc(sizeof(*bo), GFP_KERNE 95 struct v3d_bo *bo = to_v3d_bo(obj); local 133 struct v3d_bo *bo; local 174 v3d_get_bo_vaddr(struct v3d_bo *bo) argument 182 v3d_put_bo_vaddr(struct v3d_bo *bo) argument 192 struct v3d_bo *bo = NULL; local 240 struct v3d_bo *bo; local [all...] |
/linux-master/drivers/gpu/drm/panfrost/ |
H A D | panfrost_gem.c | 19 struct panfrost_gem_object *bo = to_panfrost_bo(obj); local 29 list_del_init(&bo->base.madv_list); 36 WARN_ON_ONCE(!list_empty(&bo->mappings.list)); 38 if (bo->sgts) { 40 int n_sgt = bo->base.base.size / SZ_2M; 43 if (bo->sgts[i].sgl) { 44 dma_unmap_sgtable(pfdev->dev, &bo->sgts[i], 46 sg_free_table(&bo->sgts[i]); 49 kvfree(bo->sgts); 52 drm_gem_shmem_free(&bo 56 panfrost_gem_mapping_get(struct panfrost_gem_object *bo, struct panfrost_file_priv *priv) argument 106 panfrost_gem_teardown_mappings_locked(struct panfrost_gem_object *bo) argument 119 struct panfrost_gem_object *bo = to_panfrost_bo(obj); local 172 struct panfrost_gem_object *bo = to_panfrost_bo(obj); local 190 struct panfrost_gem_object *bo = to_panfrost_bo(obj); local 200 struct panfrost_gem_object *bo = to_panfrost_bo(obj); local 214 struct panfrost_gem_object *bo = to_panfrost_bo(obj); local 271 struct panfrost_gem_object *bo; local 294 struct panfrost_gem_object *bo; local [all...] |
/linux-master/drivers/gpu/drm/lima/ |
H A D | lima_gem.c | 21 int lima_heap_alloc(struct lima_bo *bo, struct lima_vm *vm) argument 24 struct address_space *mapping = bo->base.base.filp->f_mapping; 25 struct device *dev = bo->base.base.dev->dev; 26 size_t old_size = bo->heap_size; 27 size_t new_size = bo->heap_size ? bo->heap_size * 2 : 32 if (bo->heap_size >= bo->base.base.size) 35 new_size = min(new_size, bo->base.base.size); 37 dma_resv_lock(bo 115 struct lima_bo *bo; local 155 struct lima_bo *bo = to_lima_bo(obj); local 165 struct lima_bo *bo = to_lima_bo(obj); local 174 struct lima_bo *bo = to_lima_bo(obj); local 183 struct lima_bo *bo = to_lima_bo(obj); local 193 struct lima_bo *bo = to_lima_bo(obj); local 203 struct lima_bo *bo = to_lima_bo(obj); local 227 struct lima_bo *bo; local 244 struct lima_bo *bo; local 262 lima_gem_sync_bo(struct lima_sched_task *task, struct lima_bo *bo, bool write, bool explicit) argument 315 struct lima_bo *bo; local [all...] |
/linux-master/drivers/staging/media/atomisp/include/hmm/ |
H A D | hmm_bo.h | 41 #define check_bo_status_yes_goto(bo, _status, label) \ 42 var_not_equal_goto((bo->status & (_status)), (_status), \ 47 #define check_bo_status_no_goto(bo, _status, label) \ 48 var_equal_goto((bo->status & (_status)), (_status), \ 62 #define check_bo_null_return(bo, exp) \ 63 check_null_return(bo, exp, "NULL hmm buffer object.\n") 65 #define check_bo_null_return_void(bo) \ 66 check_null_return_void(bo, "NULL hmm buffer object.\n") 93 /* start/pgnr/size is used to record the virtual memory of this bo */ 131 * When insert a bo whic [all...] |
/linux-master/drivers/gpu/drm/xe/tests/ |
H A D | xe_bo.c | 17 static int ccs_test_migrate(struct xe_tile *tile, struct xe_bo *bo, argument 30 /* Move bo to VRAM if not already there. */ 31 ret = xe_bo_validate(bo, NULL, false); 33 KUNIT_FAIL(test, "Failed to validate bo.\n"); 37 /* Optionally clear bo *and* CCS data in VRAM. */ 39 fence = xe_migrate_clear(tile->migrate, bo, bo->ttm.resource); 41 KUNIT_FAIL(test, "Failed to submit bo clear.\n"); 48 ret = xe_bo_evict(bo, true); 50 KUNIT_FAIL(test, "Failed to evict bo 114 struct xe_bo *bo; local 188 struct xe_bo *bo, *external; local [all...] |
/linux-master/drivers/gpu/drm/xe/display/ |
H A D | intel_fb_bo.h | 15 void intel_fb_bo_framebuffer_fini(struct xe_bo *bo); 17 struct xe_bo *bo,
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nv50_fence.c | 40 struct ttm_resource *reg = priv->bo->bo.resource; 42 u32 limit = start + priv->bo->bo.base.size - 1; 86 0, 0x0000, NULL, NULL, &priv->bo); 88 ret = nouveau_bo_pin(priv->bo, NOUVEAU_GEM_DOMAIN_VRAM, false); 90 ret = nouveau_bo_map(priv->bo); 92 nouveau_bo_unpin(priv->bo); 95 nouveau_bo_ref(NULL, &priv->bo); 103 nouveau_bo_wr32(priv->bo, [all...] |
/linux-master/drivers/gpu/drm/ |
H A D | drm_gem_ttm_helper.c | 42 const struct ttm_buffer_object *bo = drm_gem_ttm_of_gem(gem); local 45 drm_print_bits(p, bo->resource->placement, plname, ARRAY_SIZE(plname)); 48 if (bo->resource->bus.is_iomem) 50 (unsigned long)bo->resource->bus.offset); 68 struct ttm_buffer_object *bo = drm_gem_ttm_of_gem(gem); local 70 return ttm_bo_vmap(bo, map); 85 struct ttm_buffer_object *bo = drm_gem_ttm_of_gem(gem); local 87 ttm_bo_vunmap(bo, map); 102 struct ttm_buffer_object *bo = drm_gem_ttm_of_gem(gem); local 105 ret = ttm_bo_mmap_obj(vma, bo); [all...] |
/linux-master/drivers/gpu/drm/virtio/ |
H A D | virtgpu_object.c | 64 void virtio_gpu_cleanup_object(struct virtio_gpu_object *bo) argument 66 struct virtio_gpu_device *vgdev = bo->base.base.dev->dev_private; 68 virtio_gpu_resource_id_put(vgdev, bo->hw_res_handle); 69 if (virtio_gpu_is_shmem(bo)) { 70 drm_gem_shmem_free(&bo->base); 71 } else if (virtio_gpu_is_vram(bo)) { 72 struct virtio_gpu_object_vram *vram = to_virtio_gpu_vram(bo); 88 struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(obj); local 89 struct virtio_gpu_device *vgdev = bo->base.base.dev->dev_private; 91 if (bo 115 virtio_gpu_is_shmem(struct virtio_gpu_object *bo) argument 135 virtio_gpu_object_shmem_init(struct virtio_gpu_device *vgdev, struct virtio_gpu_object *bo, struct virtio_gpu_mem_entry **ents, unsigned int *nents) argument 186 struct virtio_gpu_object *bo; local [all...] |