Searched refs:u64 (Results 251 - 275 of 10442) sorted by last modified time

<<11121314151617181920>>

/linux-master/drivers/gpu/drm/xe/
H A Dxe_vm_types.h138 u64 size;
158 u64 composite_fence_ctx;
261 u64 tlb_flush_seqno;
291 u64 start;
293 u64 range;
H A Dxe_vm.c758 u64 bo_offset_or_userptr,
759 u64 start, u64 end,
832 u64 size = end - start + 1;
983 xe_vm_find_overlapping_vma(struct xe_vm *vm, u64 start, u64 range)
1046 static u64 pde_encode_pat_index(struct xe_device *xe, u16 pat_index)
1048 u64 pte = 0;
1059 static u64 pte_encode_pat_index(struct xe_device *xe, u16 pat_index,
1062 u64 pt
[all...]
H A Dxe_uc_fw.c794 u64 src_offset;
H A Dxe_ttm_vram_mgr.c28 u64 start, size;
55 u64 size, remaining_size, min_page_size;
94 min_page_size = (u64)tbo->page_alignment << PAGE_SHIFT;
138 u64 alloc_size = min_t(u64, remaining_size, SZ_2G);
140 err = drm_buddy_alloc_blocks(mm, (u64)place->fpfn << PAGE_SHIFT,
141 (u64)lpfn << PAGE_SHIFT,
163 u64 start = drm_buddy_block_offset(block);
166 u64 end = start + drm_buddy_block_size(mm, block);
236 (u64)mg
[all...]
H A Dxe_ttm_vram_mgr.h16 u32 mem_type, u64 size, u64 io_size,
17 u64 default_page_size);
21 u64 offset, u64 length,
28 u64 xe_ttm_vram_get_avail(struct ttm_resource_manager *man);
29 u64 xe_ttm_vram_get_cpu_visible_size(struct ttm_resource_manager *man);
31 u64 *used, u64 *used_visible);
H A Dxe_ttm_stolen_mgr.c64 u64 stolen_size;
65 u64 tile_offset;
66 u64 tile_size;
94 u64 val;
161 u64 gscpsmi_base = xe_mmio_read64_2x32(media_gt, GSCPSMI_BASE)
189 static u64 detect_stolen(struct xe_device *xe, struct xe_ttm_stolen_mgr *mgr)
207 u64 stolen_size, io_size;
252 u64 xe_ttm_stolen_io_offset(struct xe_bo *bo, u32 offset)
333 u64 xe_ttm_stolen_gpu_offset(struct xe_device *xe)
H A Dxe_ttm_sys_mgr.c108 u64 gtt_size;
111 gtt_size = (u64)si.totalram * si.mem_unit;
H A Dxe_trace.h262 __field(u64, batch_addr)
273 __entry->batch_addr = (u64)job->batch_addr[0];
351 __field(u64, ctx)
393 __field(u64, start)
394 __field(u64, end)
395 __field(u64, ptr)
H A Dxe_sync.c28 u64 __user *addr;
29 u64 value;
52 static struct xe_user_fence *user_fence_create(struct xe_device *xe, u64 addr,
53 u64 value)
H A Dxe_sched_job.c32 sizeof(u64), 0,
40 sizeof(u64) *
83 u64 *batch_addr)
316 size_t len = sizeof(*snapshot) + (sizeof(u64) * q->width);
H A Dxe_sched_job.h22 u64 *batch_addr);
H A Dxe_ring_ops.c95 static int emit_bb_start(u64 batch_addr, u32 ppgtt_flag, u32 *dw, int i)
152 static int emit_store_imm_ppgtt_posted(u64 addr, u64 value,
219 u64 batch_addr, u32 seqno)
266 u64 batch_addr, u32 seqno)
312 u64 batch_addr, u32 seqno)
H A Dxe_pt.c36 static const u64 xe_normal_pt_shifts[] = {12, 21, 30, 39, 48};
37 static const u64 xe_compact_pt_shifts[] = {16, 21, 30, 39, 48};
51 static u64 __xe_pt_empty_pte(struct xe_tile *tile, struct xe_vm *vm,
149 u64 empty;
257 u64 default_pte;
259 u64 dma_offset;
274 u64 va_curs_start;
291 u64 l0_end_addr;
293 u64 addr_64K;
347 pgoff_t offset, struct xe_pt *xe_child, u64 pt
807 sizeof(u64), u64, ptes[i].pte); local
1502 sizeof(u64), u64, empty); local
[all...]
H A Dxe_query.c61 typedef u64 (*__ktime_func_t)(void);
88 u64 *engine_ts,
89 u64 *cpu_ts,
90 u64 *cpu_delta,
105 *engine_ts = (u64)upper << 32 | lower;
317 sizeof(struct drm_xe_query_config) + num_params * sizeof(u64);
H A Dxe_pm.c631 u64 vram_used;
H A Dxe_mmio.c57 u64 force_vram_bar_size = xe_modparam.force_vram_bar_size;
85 (u64)rebar_size >> 20, bar_size_mask, (u64)current_size >> 20);
102 (u64)current_size >> 20, (u64)rebar_size >> 20);
109 (u64)root_res->start > 0x100000000ul)
166 static inline u64 get_flat_ccs_offset(struct xe_gt *gt, u64 tile_size)
169 u64 offset;
173 u64 ccs_siz
[all...]
H A Dxe_mmio.h35 u64 xe_mmio_read64_2x32(struct xe_gt *gt, struct xe_reg reg);
H A Dxe_migrate.c52 u64 batch_base_ofs;
54 u64 usm_batch_base_ofs;
56 u64 cleared_mem_ofs;
68 u64 min_chunk_size;
117 static u64 xe_migrate_vm_addr(u64 slot, u32 level)
125 static u64 xe_migrate_vram_ofs(struct xe_device *xe, u64 addr)
144 u64 entry;
174 xe_map_wr(xe, &bo->vmap, map_ofs + level * 8, u64, entr
[all...]
H A Dxe_lrc.h38 u64 xe_lrc_descriptor(struct xe_lrc *lrc);
H A Dxe_lrc_types.h40 u64 desc;
H A Dxe_lrc.c716 u64 desc = xe_vm_pdp4_descriptor(vm, lrc->tile);
894 u64 xe_lrc_descriptor(struct xe_lrc *lrc)
1006 dw[1], ((u64)(dw[3]) << 32 | (u64)(dw[2])));
H A Dxe_lmtt.c51 static u64 lmtt_page_size(struct xe_lmtt *lmtt)
220 u64 pte, unsigned int idx)
231 case sizeof(u64):
232 xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u64), u64, pte);
275 u64 start, u64 end)
277 u64 pte_addr_shift = BIT_ULL(lmtt->ops->lmtt_pte_shift(pd->level));
278 u64 offset;
286 u64 nex
[all...]
H A Dxe_hwmon.c125 enum xe_hwmon_reg_operation operation, u64 *value,
162 u64 reg_val, min, max;
183 *value = clamp_t(u64, *value, min, max);
191 u64 reg_val;
209 reg_val = DIV_ROUND_CLOSEST_ULL((u64)value << hwmon->scl_shift_power, SF_POWER);
221 u64 reg_val;
252 u64 reg_val;
274 u64 r, tau4, out;
301 tau4 = (u64)((1 << x_w) | x) << y;
315 u64 tau
[all...]
H A Dxe_hw_engine_class_sysfs.h15 bool xe_hw_engine_timeout_in_range(u64 timeout, u64 min, u64 max);
H A Dxe_hw_engine_class_sysfs.c30 bool xe_hw_engine_timeout_in_range(u64 timeout, u64 min, u64 max)

Completed in 226 milliseconds

<<11121314151617181920>>