/linux-master/include/drm/ |
H A D | drm_exec.h | 56 * @exec: Pointer to the drm_exec context 63 drm_exec_obj(struct drm_exec *exec, unsigned long index) argument 65 return index < exec->num_objects ? exec->objects[index] : NULL; 70 * @exec: drm_exec object 76 #define drm_exec_for_each_locked_object(exec, index, obj) \ 77 for ((index) = 0; ((obj) = drm_exec_obj(exec, index)); ++(index)) 82 * @exec: drm_exec object 90 #define drm_exec_for_each_locked_object_reverse(exec, index, obj) \ 91 for ((index) = (exec) 133 drm_exec_is_contended(struct drm_exec *exec) argument [all...] |
/linux-master/drivers/gpu/drm/ |
H A D | drm_exec.c | 24 * struct drm_exec exec; 28 * drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT); 29 * drm_exec_until_all_locked(&exec) { 30 * ret = drm_exec_prepare_obj(&exec, boA, 1); 31 * drm_exec_retry_on_contention(&exec); 35 * ret = drm_exec_prepare_obj(&exec, boB, 1); 36 * drm_exec_retry_on_contention(&exec); 41 * drm_exec_for_each_locked_object(&exec, index, obj) { 45 * drm_exec_fini(&exec); 54 static void drm_exec_unlock_all(struct drm_exec *exec) argument 79 drm_exec_init(struct drm_exec *exec, u32 flags, unsigned nr) argument 102 drm_exec_fini(struct drm_exec *exec) argument 121 drm_exec_cleanup(struct drm_exec *exec) argument 141 drm_exec_obj_locked(struct drm_exec *exec, struct drm_gem_object *obj) argument 163 drm_exec_lock_contended(struct drm_exec *exec) argument 208 drm_exec_lock_obj(struct drm_exec *exec, struct drm_gem_object *obj) argument 261 drm_exec_unlock_obj(struct drm_exec *exec, struct drm_gem_object *obj) argument 290 drm_exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj, unsigned int num_fences) argument 322 drm_exec_prepare_array(struct drm_exec *exec, struct drm_gem_object **objects, unsigned int num_objects, unsigned int num_fences) argument [all...] |
H A D | drm_gpuvm.c | 1089 exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj, argument 1092 return num_fences ? drm_exec_prepare_obj(exec, obj, num_fences) : 1093 drm_exec_lock_obj(exec, obj); 1099 * @exec: the &drm_exec context 1112 struct drm_exec *exec, 1115 return exec_prepare_obj(exec, gpuvm->r_obj, num_fences); 1121 struct drm_exec *exec, 1129 ret = exec_prepare_obj(exec, vm_bo->obj, num_fences); 1142 struct drm_exec *exec, 1150 ret = exec_prepare_obj(exec, vm_b 1111 drm_gpuvm_prepare_vm(struct drm_gpuvm *gpuvm, struct drm_exec *exec, unsigned int num_fences) argument 1120 __drm_gpuvm_prepare_objects(struct drm_gpuvm *gpuvm, struct drm_exec *exec, unsigned int num_fences) argument 1141 drm_gpuvm_prepare_objects_locked(struct drm_gpuvm *gpuvm, struct drm_exec *exec, unsigned int num_fences) argument 1185 drm_gpuvm_prepare_objects(struct drm_gpuvm *gpuvm, struct drm_exec *exec, unsigned int num_fences) argument 1212 drm_gpuvm_prepare_range(struct drm_gpuvm *gpuvm, struct drm_exec *exec, u64 addr, u64 range, unsigned int num_fences) argument 1249 struct drm_exec *exec = &vm_exec->exec; local 1341 struct drm_exec *exec = &vm_exec->exec; local 1363 __drm_gpuvm_validate(struct drm_gpuvm *gpuvm, struct drm_exec *exec) argument 1383 drm_gpuvm_validate_locked(struct drm_gpuvm *gpuvm, struct drm_exec *exec) argument 1416 drm_gpuvm_validate(struct drm_gpuvm *gpuvm, struct drm_exec *exec) argument 1440 drm_gpuvm_resv_add_fence(struct drm_gpuvm *gpuvm, struct drm_exec *exec, struct dma_fence *fence, enum dma_resv_usage private_usage, enum dma_resv_usage extobj_usage) argument [all...] |
/linux-master/drivers/gpu/drm/tests/ |
H A D | drm_exec_test.c | 47 struct drm_exec exec; local 49 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0); 50 drm_exec_fini(&exec); 58 struct drm_exec exec; local 63 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0); 64 drm_exec_until_all_locked(&exec) { 65 ret = drm_exec_lock_obj(&exec, &gobj); 66 drm_exec_retry_on_contention(&exec); 71 drm_exec_fini(&exec); 78 struct drm_exec exec; local 105 struct drm_exec exec; local 132 struct drm_exec exec; local 156 struct drm_exec exec; local 175 struct drm_exec exec; local [all...] |
/linux-master/tools/perf/util/ |
H A D | comm.h | 15 bool exec; member in struct:comm 23 struct comm *comm__new(const char *str, u64 timestamp, bool exec); 26 bool exec);
|
H A D | comm.c | 108 struct comm *comm__new(const char *str, u64 timestamp, bool exec) argument 116 comm->exec = exec; 127 int comm__override(struct comm *comm, const char *str, u64 timestamp, bool exec) argument 138 if (exec) 139 comm->exec = true;
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_csa.c | 69 struct drm_exec exec; local 72 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0); 73 drm_exec_until_all_locked(&exec) { 74 r = amdgpu_vm_lock_pd(vm, &exec, 0); 76 r = drm_exec_lock_obj(&exec, &bo->tbo.base); 77 drm_exec_retry_on_contention(&exec); 101 drm_exec_fini(&exec); 109 struct drm_exec exec; local 112 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0); 113 drm_exec_until_all_locked(&exec) { [all...] |
H A D | amdgpu_seq64.c | 67 struct drm_exec exec; local 75 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0); 76 drm_exec_until_all_locked(&exec) { 77 r = amdgpu_vm_lock_pd(vm, &exec, 0); 79 r = drm_exec_lock_obj(&exec, &bo->tbo.base); 80 drm_exec_retry_on_contention(&exec); 108 drm_exec_fini(&exec); 124 struct drm_exec exec; local 136 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0); 137 drm_exec_until_all_locked(&exec) { [all...] |
/linux-master/include/uapi/linux/ |
H A D | a.out.h | 44 #define N_MAGIC(exec) ((exec).a_info & 0xffff) 46 #define N_MACHTYPE(exec) ((enum machine_type)(((exec).a_info >> 16) & 0xff)) 47 #define N_FLAGS(exec) (((exec).a_info >> 24) & 0xff) 48 #define N_SET_INFO(exec, magic, type, flags) \ 49 ((exec).a_info = ((magic) & 0xffff) \ 52 #define N_SET_MAGIC(exec, magic) \ 53 ((exec) [all...] |
/linux-master/drivers/gpu/drm/vc4/ |
H A D | vc4_gem.c | 157 struct vc4_exec_info *exec[2]; local 169 exec[0] = vc4_first_bin_job(vc4); 170 exec[1] = vc4_first_render_job(vc4); 171 if (!exec[0] && !exec[1]) { 179 if (!exec[i]) 183 list_for_each_entry(bo, &exec[i]->unref_list, unref_head) 185 state->bo_count += exec[i]->bo_count + unref_list_count; 198 if (!exec[i]) 201 for (j = 0; j < exec[ 475 struct vc4_exec_info *exec; local 520 struct vc4_exec_info *exec = vc4_first_render_job(vc4); local 541 vc4_move_job_to_render(struct drm_device *dev, struct vc4_exec_info *exec) argument 555 vc4_update_bo_seqnos(struct vc4_exec_info *exec, uint64_t seqno) argument 582 vc4_unlock_bo_reservations(struct drm_device *dev, struct vc4_exec_info *exec, struct ww_acquire_ctx *acquire_ctx) argument 602 vc4_lock_bo_reservations(struct drm_device *dev, struct vc4_exec_info *exec, struct ww_acquire_ctx *acquire_ctx) argument 681 vc4_queue_submit(struct drm_device *dev, struct vc4_exec_info *exec, struct ww_acquire_ctx *acquire_ctx, struct drm_syncobj *out_sync) argument 744 vc4_cl_lookup_bos(struct drm_device *dev, struct drm_file *file_priv, struct vc4_exec_info *exec) argument 799 vc4_get_bcl(struct drm_device *dev, struct vc4_exec_info *exec) argument 919 vc4_complete_exec(struct drm_device *dev, struct vc4_exec_info *exec) argument 978 struct vc4_exec_info *exec = local 1125 struct vc4_exec_info *exec; local [all...] |
H A D | vc4_irq.c | 67 struct vc4_exec_info *exec; local 92 exec = vc4_first_bin_job(vc4); 93 if (!exec) 94 exec = vc4_last_render_job(vc4); 95 if (exec) { 96 exec->bin_slots |= vc4->bin_alloc_overflow; 120 struct vc4_exec_info *next, *exec = vc4_first_bin_job(vc4); local 122 if (!exec) 125 trace_vc4_bcl_end_irq(dev, exec->seqno); 127 vc4_move_job_to_render(dev, exec); 142 struct vc4_exec_info *exec = vc4_first_bin_job(vc4); local 159 struct vc4_exec_info *exec = vc4_first_render_job(vc4); local [all...] |
H A D | vc4_validate.c | 51 struct vc4_exec_info *exec, \ 106 vc4_use_bo(struct vc4_exec_info *exec, uint32_t hindex) argument 108 struct vc4_dev *vc4 = exec->dev; 115 if (hindex >= exec->bo_count) { 117 hindex, exec->bo_count); 120 obj = to_drm_gem_dma_obj(exec->bo[hindex]); 133 vc4_use_handle(struct vc4_exec_info *exec, uint32_t gem_handles_packet_index) argument 135 return vc4_use_bo(exec, exec->bo_index[gem_handles_packet_index]); 139 validate_bin_pos(struct vc4_exec_info *exec, voi argument 163 vc4_check_tex_size(struct vc4_exec_info *exec, struct drm_gem_dma_object *fbo, uint32_t offset, uint8_t tiling_format, uint32_t width, uint32_t height, uint8_t cpp) argument 488 vc4_validate_bin_cl(struct drm_device *dev, void *validated, void *unvalidated, struct vc4_exec_info *exec) argument 573 reloc_tex(struct vc4_exec_info *exec, void *uniform_data_u, struct vc4_texture_sample_info *sample, uint32_t texture_handle_index, bool is_cs) argument 756 validate_gl_shader_rec(struct drm_device *dev, struct vc4_exec_info *exec, struct vc4_shader_state *state) argument 938 vc4_validate_shader_recs(struct drm_device *dev, struct vc4_exec_info *exec) argument [all...] |
H A D | vc4_render_cl.c | 99 static uint32_t vc4_full_res_offset(struct vc4_exec_info *exec, argument 105 (DIV_ROUND_UP(exec->args->width, 32) * y + x); 123 static void emit_tile(struct vc4_exec_info *exec, argument 127 struct drm_vc4_submit_cl *args = exec->args; 139 vc4_full_res_offset(exec, setup->color_read, 161 vc4_full_res_offset(exec, setup->zs_read, 185 rcl_u32(setup, (exec->tile_alloc_offset + 186 (y * exec->bin_tiles_x + x) * 32)); 201 vc4_full_res_offset(exec, setup->msaa_color_write, 219 vc4_full_res_offset(exec, setu 253 vc4_create_rcl_bo(struct drm_device *dev, struct vc4_exec_info *exec, struct vc4_rcl_setup *setup) argument 383 vc4_full_res_bounds_check(struct vc4_exec_info *exec, struct drm_gem_dma_object *obj, struct drm_vc4_submit_rcl_surface *surf) argument 409 vc4_rcl_msaa_surface_setup(struct vc4_exec_info *exec, struct drm_gem_dma_object **obj, struct drm_vc4_submit_rcl_surface *surf) argument 435 vc4_rcl_surface_setup(struct vc4_exec_info *exec, struct drm_gem_dma_object **obj, struct drm_vc4_submit_rcl_surface *surf, bool is_write) argument 534 vc4_rcl_render_config_surface_setup(struct vc4_exec_info *exec, struct vc4_rcl_setup *setup, struct drm_gem_dma_object **obj, struct drm_vc4_submit_rcl_surface *surf) argument 594 vc4_get_rcl(struct drm_device *dev, struct vc4_exec_info *exec) argument [all...] |
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | test_overhead.c | 13 int BPF_KPROBE(prog1, struct task_struct *tsk, const char *buf, bool exec) argument 31 int BPF_PROG(prog4, struct task_struct *tsk, const char *buf, bool exec) argument 37 int BPF_PROG(prog5, struct task_struct *tsk, const char *buf, bool exec) argument
|
/linux-master/arch/powerpc/boot/dts/fsl/ |
H A D | pq3-sec2.1-0.dtsi | 41 fsl,exec-units-mask = <0xfe>;
|
H A D | pq3-sec3.0-0.dtsi | 43 fsl,exec-units-mask = <0x9fe>;
|
H A D | pq3-sec3.1-0.dtsi | 43 fsl,exec-units-mask = <0xbfe>;
|
H A D | pq3-sec3.3-0.dtsi | 43 fsl,exec-units-mask = <0x97c>;
|
/linux-master/arch/x86/include/uapi/asm/ |
H A D | a.out.h | 5 struct exec struct
|
/linux-master/arch/m68k/include/uapi/asm/ |
H A D | a.out.h | 5 struct exec struct
|
/linux-master/tools/testing/selftests/exec/ |
H A D | null-argv.c | 10 #define FORK(exec) \ 15 exec; /* Some kind of exec */ \ 16 perror("# " #exec); \ 19 check_result(pid, #exec); \
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/timer/ |
H A D | base.c | 73 LIST_HEAD(exec); 90 list_add(&alarm->exec, &exec); 99 list_for_each_entry_safe(alarm, atemp, &exec, exec) { 100 list_del(&alarm->exec);
|
/linux-master/arch/sparc/mm/ |
H A D | tlb.c | 70 bool exec, unsigned int hugepage_shift) 76 if (exec) 159 bool exec = pte_exec(*pte); local 161 tlb_batch_add_one(mm, vaddr, exec, PAGE_SHIFT); 211 bool exec = pte_exec(orig_pte); local 213 tlb_batch_add_one(mm, addr, exec, REAL_HPAGE_SHIFT); 214 tlb_batch_add_one(mm, addr + REAL_HPAGE_SIZE, exec, 69 tlb_batch_add_one(struct mm_struct *mm, unsigned long vaddr, bool exec, unsigned int hugepage_shift) argument
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_exec.c | 30 * - Binding at exec time 31 * - Flow controlling the ring at exec time 34 * passed into an exec, using the dma-buf implicit sync uAPI, have binds as 39 * every exec rather than the kernel tracking dependencies of BO (e.g. if the 40 * user knows an exec writes to a BO and reads from the BO in the next exec, it 44 * implicit dependency uAPI (TODO: add link). To make this works each exec must 48 * We do not allow a user to trigger a bind at exec time rather we have a VM 49 * bind IOCTL which uses the same in / out fence interface as exec. In that 50 * sense, a VM bind is basically the same operation as an exec fro 120 struct drm_exec *exec = &vm_exec.exec; local [all...] |
/linux-master/drivers/leds/ |
H A D | leds-lp8501.c | 162 u8 exec; local 180 ret = lp55xx_read(chip, LP8501_REG_ENABLE, &exec); 187 exec = (exec & ~LP8501_EXEC_ENG1_M) | LP8501_RUN_ENG1; 192 exec = (exec & ~LP8501_EXEC_ENG2_M) | LP8501_RUN_ENG2; 197 exec = (exec & ~LP8501_EXEC_ENG3_M) | LP8501_RUN_ENG3; 203 lp55xx_update_bits(chip, LP8501_REG_ENABLE, LP8501_EXEC_M, exec);
|