Searched refs:exec (Results 1 - 25 of 147) sorted by relevance

123456

/linux-master/include/drm/
H A Ddrm_exec.h56 * @exec: Pointer to the drm_exec context
63 drm_exec_obj(struct drm_exec *exec, unsigned long index) argument
65 return index < exec->num_objects ? exec->objects[index] : NULL;
70 * @exec: drm_exec object
76 #define drm_exec_for_each_locked_object(exec, index, obj) \
77 for ((index) = 0; ((obj) = drm_exec_obj(exec, index)); ++(index))
82 * @exec: drm_exec object
90 #define drm_exec_for_each_locked_object_reverse(exec, index, obj) \
91 for ((index) = (exec)
133 drm_exec_is_contended(struct drm_exec *exec) argument
[all...]
/linux-master/drivers/gpu/drm/
H A Ddrm_exec.c24 * struct drm_exec exec;
28 * drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT);
29 * drm_exec_until_all_locked(&exec) {
30 * ret = drm_exec_prepare_obj(&exec, boA, 1);
31 * drm_exec_retry_on_contention(&exec);
35 * ret = drm_exec_prepare_obj(&exec, boB, 1);
36 * drm_exec_retry_on_contention(&exec);
41 * drm_exec_for_each_locked_object(&exec, index, obj) {
45 * drm_exec_fini(&exec);
54 static void drm_exec_unlock_all(struct drm_exec *exec) argument
79 drm_exec_init(struct drm_exec *exec, u32 flags, unsigned nr) argument
102 drm_exec_fini(struct drm_exec *exec) argument
121 drm_exec_cleanup(struct drm_exec *exec) argument
141 drm_exec_obj_locked(struct drm_exec *exec, struct drm_gem_object *obj) argument
163 drm_exec_lock_contended(struct drm_exec *exec) argument
208 drm_exec_lock_obj(struct drm_exec *exec, struct drm_gem_object *obj) argument
261 drm_exec_unlock_obj(struct drm_exec *exec, struct drm_gem_object *obj) argument
290 drm_exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj, unsigned int num_fences) argument
322 drm_exec_prepare_array(struct drm_exec *exec, struct drm_gem_object **objects, unsigned int num_objects, unsigned int num_fences) argument
[all...]
H A Ddrm_gpuvm.c1089 exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj, argument
1092 return num_fences ? drm_exec_prepare_obj(exec, obj, num_fences) :
1093 drm_exec_lock_obj(exec, obj);
1099 * @exec: the &drm_exec context
1112 struct drm_exec *exec,
1115 return exec_prepare_obj(exec, gpuvm->r_obj, num_fences);
1121 struct drm_exec *exec,
1129 ret = exec_prepare_obj(exec, vm_bo->obj, num_fences);
1142 struct drm_exec *exec,
1150 ret = exec_prepare_obj(exec, vm_b
1111 drm_gpuvm_prepare_vm(struct drm_gpuvm *gpuvm, struct drm_exec *exec, unsigned int num_fences) argument
1120 __drm_gpuvm_prepare_objects(struct drm_gpuvm *gpuvm, struct drm_exec *exec, unsigned int num_fences) argument
1141 drm_gpuvm_prepare_objects_locked(struct drm_gpuvm *gpuvm, struct drm_exec *exec, unsigned int num_fences) argument
1185 drm_gpuvm_prepare_objects(struct drm_gpuvm *gpuvm, struct drm_exec *exec, unsigned int num_fences) argument
1212 drm_gpuvm_prepare_range(struct drm_gpuvm *gpuvm, struct drm_exec *exec, u64 addr, u64 range, unsigned int num_fences) argument
1249 struct drm_exec *exec = &vm_exec->exec; local
1341 struct drm_exec *exec = &vm_exec->exec; local
1363 __drm_gpuvm_validate(struct drm_gpuvm *gpuvm, struct drm_exec *exec) argument
1383 drm_gpuvm_validate_locked(struct drm_gpuvm *gpuvm, struct drm_exec *exec) argument
1416 drm_gpuvm_validate(struct drm_gpuvm *gpuvm, struct drm_exec *exec) argument
1440 drm_gpuvm_resv_add_fence(struct drm_gpuvm *gpuvm, struct drm_exec *exec, struct dma_fence *fence, enum dma_resv_usage private_usage, enum dma_resv_usage extobj_usage) argument
[all...]
/linux-master/drivers/gpu/drm/tests/
H A Ddrm_exec_test.c47 struct drm_exec exec; local
49 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
50 drm_exec_fini(&exec);
58 struct drm_exec exec; local
63 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
64 drm_exec_until_all_locked(&exec) {
65 ret = drm_exec_lock_obj(&exec, &gobj);
66 drm_exec_retry_on_contention(&exec);
71 drm_exec_fini(&exec);
78 struct drm_exec exec; local
105 struct drm_exec exec; local
132 struct drm_exec exec; local
156 struct drm_exec exec; local
175 struct drm_exec exec; local
[all...]
/linux-master/tools/perf/util/
H A Dcomm.h15 bool exec; member in struct:comm
23 struct comm *comm__new(const char *str, u64 timestamp, bool exec);
26 bool exec);
H A Dcomm.c108 struct comm *comm__new(const char *str, u64 timestamp, bool exec) argument
116 comm->exec = exec;
127 int comm__override(struct comm *comm, const char *str, u64 timestamp, bool exec) argument
138 if (exec)
139 comm->exec = true;
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_csa.c69 struct drm_exec exec; local
72 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
73 drm_exec_until_all_locked(&exec) {
74 r = amdgpu_vm_lock_pd(vm, &exec, 0);
76 r = drm_exec_lock_obj(&exec, &bo->tbo.base);
77 drm_exec_retry_on_contention(&exec);
101 drm_exec_fini(&exec);
109 struct drm_exec exec; local
112 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
113 drm_exec_until_all_locked(&exec) {
[all...]
H A Damdgpu_seq64.c67 struct drm_exec exec; local
75 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
76 drm_exec_until_all_locked(&exec) {
77 r = amdgpu_vm_lock_pd(vm, &exec, 0);
79 r = drm_exec_lock_obj(&exec, &bo->tbo.base);
80 drm_exec_retry_on_contention(&exec);
108 drm_exec_fini(&exec);
124 struct drm_exec exec; local
136 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
137 drm_exec_until_all_locked(&exec) {
[all...]
/linux-master/include/uapi/linux/
H A Da.out.h44 #define N_MAGIC(exec) ((exec).a_info & 0xffff)
46 #define N_MACHTYPE(exec) ((enum machine_type)(((exec).a_info >> 16) & 0xff))
47 #define N_FLAGS(exec) (((exec).a_info >> 24) & 0xff)
48 #define N_SET_INFO(exec, magic, type, flags) \
49 ((exec).a_info = ((magic) & 0xffff) \
52 #define N_SET_MAGIC(exec, magic) \
53 ((exec)
[all...]
/linux-master/drivers/gpu/drm/vc4/
H A Dvc4_gem.c157 struct vc4_exec_info *exec[2]; local
169 exec[0] = vc4_first_bin_job(vc4);
170 exec[1] = vc4_first_render_job(vc4);
171 if (!exec[0] && !exec[1]) {
179 if (!exec[i])
183 list_for_each_entry(bo, &exec[i]->unref_list, unref_head)
185 state->bo_count += exec[i]->bo_count + unref_list_count;
198 if (!exec[i])
201 for (j = 0; j < exec[
475 struct vc4_exec_info *exec; local
520 struct vc4_exec_info *exec = vc4_first_render_job(vc4); local
541 vc4_move_job_to_render(struct drm_device *dev, struct vc4_exec_info *exec) argument
555 vc4_update_bo_seqnos(struct vc4_exec_info *exec, uint64_t seqno) argument
582 vc4_unlock_bo_reservations(struct drm_device *dev, struct vc4_exec_info *exec, struct ww_acquire_ctx *acquire_ctx) argument
602 vc4_lock_bo_reservations(struct drm_device *dev, struct vc4_exec_info *exec, struct ww_acquire_ctx *acquire_ctx) argument
681 vc4_queue_submit(struct drm_device *dev, struct vc4_exec_info *exec, struct ww_acquire_ctx *acquire_ctx, struct drm_syncobj *out_sync) argument
744 vc4_cl_lookup_bos(struct drm_device *dev, struct drm_file *file_priv, struct vc4_exec_info *exec) argument
799 vc4_get_bcl(struct drm_device *dev, struct vc4_exec_info *exec) argument
919 vc4_complete_exec(struct drm_device *dev, struct vc4_exec_info *exec) argument
978 struct vc4_exec_info *exec = local
1125 struct vc4_exec_info *exec; local
[all...]
H A Dvc4_irq.c67 struct vc4_exec_info *exec; local
92 exec = vc4_first_bin_job(vc4);
93 if (!exec)
94 exec = vc4_last_render_job(vc4);
95 if (exec) {
96 exec->bin_slots |= vc4->bin_alloc_overflow;
120 struct vc4_exec_info *next, *exec = vc4_first_bin_job(vc4); local
122 if (!exec)
125 trace_vc4_bcl_end_irq(dev, exec->seqno);
127 vc4_move_job_to_render(dev, exec);
142 struct vc4_exec_info *exec = vc4_first_bin_job(vc4); local
159 struct vc4_exec_info *exec = vc4_first_render_job(vc4); local
[all...]
H A Dvc4_validate.c51 struct vc4_exec_info *exec, \
106 vc4_use_bo(struct vc4_exec_info *exec, uint32_t hindex) argument
108 struct vc4_dev *vc4 = exec->dev;
115 if (hindex >= exec->bo_count) {
117 hindex, exec->bo_count);
120 obj = to_drm_gem_dma_obj(exec->bo[hindex]);
133 vc4_use_handle(struct vc4_exec_info *exec, uint32_t gem_handles_packet_index) argument
135 return vc4_use_bo(exec, exec->bo_index[gem_handles_packet_index]);
139 validate_bin_pos(struct vc4_exec_info *exec, voi argument
163 vc4_check_tex_size(struct vc4_exec_info *exec, struct drm_gem_dma_object *fbo, uint32_t offset, uint8_t tiling_format, uint32_t width, uint32_t height, uint8_t cpp) argument
488 vc4_validate_bin_cl(struct drm_device *dev, void *validated, void *unvalidated, struct vc4_exec_info *exec) argument
573 reloc_tex(struct vc4_exec_info *exec, void *uniform_data_u, struct vc4_texture_sample_info *sample, uint32_t texture_handle_index, bool is_cs) argument
756 validate_gl_shader_rec(struct drm_device *dev, struct vc4_exec_info *exec, struct vc4_shader_state *state) argument
938 vc4_validate_shader_recs(struct drm_device *dev, struct vc4_exec_info *exec) argument
[all...]
H A Dvc4_render_cl.c99 static uint32_t vc4_full_res_offset(struct vc4_exec_info *exec, argument
105 (DIV_ROUND_UP(exec->args->width, 32) * y + x);
123 static void emit_tile(struct vc4_exec_info *exec, argument
127 struct drm_vc4_submit_cl *args = exec->args;
139 vc4_full_res_offset(exec, setup->color_read,
161 vc4_full_res_offset(exec, setup->zs_read,
185 rcl_u32(setup, (exec->tile_alloc_offset +
186 (y * exec->bin_tiles_x + x) * 32));
201 vc4_full_res_offset(exec, setup->msaa_color_write,
219 vc4_full_res_offset(exec, setu
253 vc4_create_rcl_bo(struct drm_device *dev, struct vc4_exec_info *exec, struct vc4_rcl_setup *setup) argument
383 vc4_full_res_bounds_check(struct vc4_exec_info *exec, struct drm_gem_dma_object *obj, struct drm_vc4_submit_rcl_surface *surf) argument
409 vc4_rcl_msaa_surface_setup(struct vc4_exec_info *exec, struct drm_gem_dma_object **obj, struct drm_vc4_submit_rcl_surface *surf) argument
435 vc4_rcl_surface_setup(struct vc4_exec_info *exec, struct drm_gem_dma_object **obj, struct drm_vc4_submit_rcl_surface *surf, bool is_write) argument
534 vc4_rcl_render_config_surface_setup(struct vc4_exec_info *exec, struct vc4_rcl_setup *setup, struct drm_gem_dma_object **obj, struct drm_vc4_submit_rcl_surface *surf) argument
594 vc4_get_rcl(struct drm_device *dev, struct vc4_exec_info *exec) argument
[all...]
/linux-master/tools/testing/selftests/bpf/progs/
H A Dtest_overhead.c13 int BPF_KPROBE(prog1, struct task_struct *tsk, const char *buf, bool exec) argument
31 int BPF_PROG(prog4, struct task_struct *tsk, const char *buf, bool exec) argument
37 int BPF_PROG(prog5, struct task_struct *tsk, const char *buf, bool exec) argument
/linux-master/arch/powerpc/boot/dts/fsl/
H A Dpq3-sec2.1-0.dtsi41 fsl,exec-units-mask = <0xfe>;
H A Dpq3-sec3.0-0.dtsi43 fsl,exec-units-mask = <0x9fe>;
H A Dpq3-sec3.1-0.dtsi43 fsl,exec-units-mask = <0xbfe>;
H A Dpq3-sec3.3-0.dtsi43 fsl,exec-units-mask = <0x97c>;
/linux-master/arch/x86/include/uapi/asm/
H A Da.out.h5 struct exec struct
/linux-master/arch/m68k/include/uapi/asm/
H A Da.out.h5 struct exec struct
/linux-master/tools/testing/selftests/exec/
H A Dnull-argv.c10 #define FORK(exec) \
15 exec; /* Some kind of exec */ \
16 perror("# " #exec); \
19 check_result(pid, #exec); \
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/timer/
H A Dbase.c73 LIST_HEAD(exec);
90 list_add(&alarm->exec, &exec);
99 list_for_each_entry_safe(alarm, atemp, &exec, exec) {
100 list_del(&alarm->exec);
/linux-master/arch/sparc/mm/
H A Dtlb.c70 bool exec, unsigned int hugepage_shift)
76 if (exec)
159 bool exec = pte_exec(*pte); local
161 tlb_batch_add_one(mm, vaddr, exec, PAGE_SHIFT);
211 bool exec = pte_exec(orig_pte); local
213 tlb_batch_add_one(mm, addr, exec, REAL_HPAGE_SHIFT);
214 tlb_batch_add_one(mm, addr + REAL_HPAGE_SIZE, exec,
69 tlb_batch_add_one(struct mm_struct *mm, unsigned long vaddr, bool exec, unsigned int hugepage_shift) argument
/linux-master/drivers/gpu/drm/xe/
H A Dxe_exec.c30 * - Binding at exec time
31 * - Flow controlling the ring at exec time
34 * passed into an exec, using the dma-buf implicit sync uAPI, have binds as
39 * every exec rather than the kernel tracking dependencies of BO (e.g. if the
40 * user knows an exec writes to a BO and reads from the BO in the next exec, it
44 * implicit dependency uAPI (TODO: add link). To make this works each exec must
48 * We do not allow a user to trigger a bind at exec time rather we have a VM
49 * bind IOCTL which uses the same in / out fence interface as exec. In that
50 * sense, a VM bind is basically the same operation as an exec fro
120 struct drm_exec *exec = &vm_exec.exec; local
[all...]
/linux-master/drivers/leds/
H A Dleds-lp8501.c162 u8 exec; local
180 ret = lp55xx_read(chip, LP8501_REG_ENABLE, &exec);
187 exec = (exec & ~LP8501_EXEC_ENG1_M) | LP8501_RUN_ENG1;
192 exec = (exec & ~LP8501_EXEC_ENG2_M) | LP8501_RUN_ENG2;
197 exec = (exec & ~LP8501_EXEC_ENG3_M) | LP8501_RUN_ENG3;
203 lp55xx_update_bits(chip, LP8501_REG_ENABLE, LP8501_EXEC_M, exec);

Completed in 565 milliseconds

123456