/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_vm_types.h | 138 u64 size; 158 u64 composite_fence_ctx; 261 u64 tlb_flush_seqno; 291 u64 start; 293 u64 range;
|
H A D | xe_vm.c | 758 u64 bo_offset_or_userptr, 759 u64 start, u64 end, 832 u64 size = end - start + 1; 983 xe_vm_find_overlapping_vma(struct xe_vm *vm, u64 start, u64 range) 1046 static u64 pde_encode_pat_index(struct xe_device *xe, u16 pat_index) 1048 u64 pte = 0; 1059 static u64 pte_encode_pat_index(struct xe_device *xe, u16 pat_index, 1062 u64 pt [all...] |
H A D | xe_uc_fw.c | 794 u64 src_offset;
|
H A D | xe_ttm_vram_mgr.c | 28 u64 start, size; 55 u64 size, remaining_size, min_page_size; 94 min_page_size = (u64)tbo->page_alignment << PAGE_SHIFT; 138 u64 alloc_size = min_t(u64, remaining_size, SZ_2G); 140 err = drm_buddy_alloc_blocks(mm, (u64)place->fpfn << PAGE_SHIFT, 141 (u64)lpfn << PAGE_SHIFT, 163 u64 start = drm_buddy_block_offset(block); 166 u64 end = start + drm_buddy_block_size(mm, block); 236 (u64)mg [all...] |
H A D | xe_ttm_vram_mgr.h | 16 u32 mem_type, u64 size, u64 io_size, 17 u64 default_page_size); 21 u64 offset, u64 length, 28 u64 xe_ttm_vram_get_avail(struct ttm_resource_manager *man); 29 u64 xe_ttm_vram_get_cpu_visible_size(struct ttm_resource_manager *man); 31 u64 *used, u64 *used_visible);
|
H A D | xe_ttm_stolen_mgr.c | 64 u64 stolen_size; 65 u64 tile_offset; 66 u64 tile_size; 94 u64 val; 161 u64 gscpsmi_base = xe_mmio_read64_2x32(media_gt, GSCPSMI_BASE) 189 static u64 detect_stolen(struct xe_device *xe, struct xe_ttm_stolen_mgr *mgr) 207 u64 stolen_size, io_size; 252 u64 xe_ttm_stolen_io_offset(struct xe_bo *bo, u32 offset) 333 u64 xe_ttm_stolen_gpu_offset(struct xe_device *xe)
|
H A D | xe_ttm_sys_mgr.c | 108 u64 gtt_size; 111 gtt_size = (u64)si.totalram * si.mem_unit;
|
H A D | xe_trace.h | 262 __field(u64, batch_addr) 273 __entry->batch_addr = (u64)job->batch_addr[0]; 351 __field(u64, ctx) 393 __field(u64, start) 394 __field(u64, end) 395 __field(u64, ptr)
|
H A D | xe_sync.c | 28 u64 __user *addr; 29 u64 value; 52 static struct xe_user_fence *user_fence_create(struct xe_device *xe, u64 addr, 53 u64 value)
|
H A D | xe_sched_job.c | 32 sizeof(u64), 0, 40 sizeof(u64) * 83 u64 *batch_addr) 316 size_t len = sizeof(*snapshot) + (sizeof(u64) * q->width);
|
H A D | xe_sched_job.h | 22 u64 *batch_addr);
|
H A D | xe_ring_ops.c | 95 static int emit_bb_start(u64 batch_addr, u32 ppgtt_flag, u32 *dw, int i) 152 static int emit_store_imm_ppgtt_posted(u64 addr, u64 value, 219 u64 batch_addr, u32 seqno) 266 u64 batch_addr, u32 seqno) 312 u64 batch_addr, u32 seqno)
|
H A D | xe_pt.c | 36 static const u64 xe_normal_pt_shifts[] = {12, 21, 30, 39, 48}; 37 static const u64 xe_compact_pt_shifts[] = {16, 21, 30, 39, 48}; 51 static u64 __xe_pt_empty_pte(struct xe_tile *tile, struct xe_vm *vm, 149 u64 empty; 257 u64 default_pte; 259 u64 dma_offset; 274 u64 va_curs_start; 291 u64 l0_end_addr; 293 u64 addr_64K; 347 pgoff_t offset, struct xe_pt *xe_child, u64 pt 807 sizeof(u64), u64, ptes[i].pte); local 1502 sizeof(u64), u64, empty); local [all...] |
H A D | xe_query.c | 61 typedef u64 (*__ktime_func_t)(void); 88 u64 *engine_ts, 89 u64 *cpu_ts, 90 u64 *cpu_delta, 105 *engine_ts = (u64)upper << 32 | lower; 317 sizeof(struct drm_xe_query_config) + num_params * sizeof(u64);
|
H A D | xe_pm.c | 631 u64 vram_used;
|
H A D | xe_mmio.c | 57 u64 force_vram_bar_size = xe_modparam.force_vram_bar_size; 85 (u64)rebar_size >> 20, bar_size_mask, (u64)current_size >> 20); 102 (u64)current_size >> 20, (u64)rebar_size >> 20); 109 (u64)root_res->start > 0x100000000ul) 166 static inline u64 get_flat_ccs_offset(struct xe_gt *gt, u64 tile_size) 169 u64 offset; 173 u64 ccs_siz [all...] |
H A D | xe_mmio.h | 35 u64 xe_mmio_read64_2x32(struct xe_gt *gt, struct xe_reg reg);
|
H A D | xe_migrate.c | 52 u64 batch_base_ofs; 54 u64 usm_batch_base_ofs; 56 u64 cleared_mem_ofs; 68 u64 min_chunk_size; 117 static u64 xe_migrate_vm_addr(u64 slot, u32 level) 125 static u64 xe_migrate_vram_ofs(struct xe_device *xe, u64 addr) 144 u64 entry; 174 xe_map_wr(xe, &bo->vmap, map_ofs + level * 8, u64, entr [all...] |
H A D | xe_lrc.h | 38 u64 xe_lrc_descriptor(struct xe_lrc *lrc);
|
H A D | xe_lrc_types.h | 40 u64 desc;
|
H A D | xe_lrc.c | 716 u64 desc = xe_vm_pdp4_descriptor(vm, lrc->tile); 894 u64 xe_lrc_descriptor(struct xe_lrc *lrc) 1006 dw[1], ((u64)(dw[3]) << 32 | (u64)(dw[2])));
|
H A D | xe_lmtt.c | 51 static u64 lmtt_page_size(struct xe_lmtt *lmtt) 220 u64 pte, unsigned int idx) 231 case sizeof(u64): 232 xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u64), u64, pte); 275 u64 start, u64 end) 277 u64 pte_addr_shift = BIT_ULL(lmtt->ops->lmtt_pte_shift(pd->level)); 278 u64 offset; 286 u64 nex [all...] |
H A D | xe_hwmon.c | 125 enum xe_hwmon_reg_operation operation, u64 *value, 162 u64 reg_val, min, max; 183 *value = clamp_t(u64, *value, min, max); 191 u64 reg_val; 209 reg_val = DIV_ROUND_CLOSEST_ULL((u64)value << hwmon->scl_shift_power, SF_POWER); 221 u64 reg_val; 252 u64 reg_val; 274 u64 r, tau4, out; 301 tau4 = (u64)((1 << x_w) | x) << y; 315 u64 tau [all...] |
H A D | xe_hw_engine_class_sysfs.h | 15 bool xe_hw_engine_timeout_in_range(u64 timeout, u64 min, u64 max);
|
H A D | xe_hw_engine_class_sysfs.c | 30 bool xe_hw_engine_timeout_in_range(u64 timeout, u64 min, u64 max)
|