Searched refs:job (Results 76 - 100 of 189) sorted by relevance

12345678

/linux-master/drivers/accel/habanalabs/common/
H A Dcommand_submission.c267 struct hl_cs_job *job = container_of(ref, struct hl_cs_job, refcount); local
269 kfree(job);
272 static void hl_cs_job_put(struct hl_cs_job *job) argument
274 kref_put(&job->refcount, cs_job_do_release);
299 static bool is_cb_patched(struct hl_device *hdev, struct hl_cs_job *job) argument
302 return (job->queue_type == QUEUE_TYPE_EXT);
309 * @job : pointer to the job that holds the command submission info
316 static int cs_parser(struct hl_fpriv *hpriv, struct hl_cs_job *job) argument
322 parser.ctx_id = job
361 hl_complete_job(struct hl_device *hdev, struct hl_cs_job *job) argument
698 struct hl_cs_job *job, *tmp; local
1028 struct hl_cs_job *job, *tmp; local
1168 struct hl_cs_job *job = container_of(work, struct hl_cs_job, local
1181 struct hl_cs_job *job, *tmp; local
1303 struct hl_cs_job *job; local
1499 struct hl_cs_job *job; local
1943 struct hl_cs_job *job; local
[all...]
/linux-master/drivers/media/platform/renesas/
H A Drcar_fdp1.c551 /* A job can only be on one list at a time */
628 * while preparing our job lists.
648 * list_remove_job: Take the first item off the specified job list
650 * Returns: pointer to a job, or NULL if the list is empty.
655 struct fdp1_job *job; local
659 job = list_first_entry_or_null(list, struct fdp1_job, list);
660 if (job)
661 list_del(&job->list);
664 return job;
668 * list_add_job: Add a job t
672 list_add_job(struct fdp1_dev *fdp1, struct list_head *list, struct fdp1_job *job) argument
688 fdp1_job_free(struct fdp1_dev *fdp1, struct fdp1_job *job) argument
696 queue_job(struct fdp1_dev *fdp1, struct fdp1_job *job) argument
706 queue_hw_job(struct fdp1_dev *fdp1, struct fdp1_job *job) argument
879 fdp1_configure_rpf(struct fdp1_ctx *ctx, struct fdp1_job *job) argument
938 fdp1_configure_wpf(struct fdp1_ctx *ctx, struct fdp1_job *job) argument
999 fdp1_configure_deint_mode(struct fdp1_ctx *ctx, struct fdp1_job *job) argument
1070 struct fdp1_job *job; local
1185 struct fdp1_job *job; local
1310 struct fdp1_job *job = get_hw_queued_job(fdp1); local
2003 struct fdp1_job *job; local
[all...]
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_sync.h58 int amdgpu_sync_push_to_job(struct amdgpu_sync *sync, struct amdgpu_job *job);
H A Dvcn_sw_ring.c44 void vcn_dec_sw_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, argument
47 uint32_t vmid = AMDGPU_JOB_GET_VMID(job);
H A Damdgpu_fence.c133 * @job: job the fence is embedded in
139 int amdgpu_fence_emit(struct amdgpu_ring *ring, struct dma_fence **f, struct amdgpu_job *job, argument
149 if (job == NULL) {
157 /* take use of job-embedded fence */
158 fence = &job->hw_fence;
162 if (job && job->job_run_counter) {
168 if (job) {
703 * amdgpu_fence_driver_clear_job_fences - clear job embedde
717 struct amdgpu_job *job; local
785 struct amdgpu_job *job = container_of(f, struct amdgpu_job, hw_fence); local
815 struct amdgpu_job *job = container_of(f, struct amdgpu_job, hw_fence); local
[all...]
H A Damdgpu_ring.h143 int amdgpu_fence_emit(struct amdgpu_ring *ring, struct dma_fence **fence, struct amdgpu_job *job,
180 struct amdgpu_job *job,
183 struct amdgpu_job *job,
190 struct amdgpu_job *job,
310 #define amdgpu_ring_parse_cs(r, p, job, ib) ((r)->funcs->parse_cs((p), (job), (ib)))
311 #define amdgpu_ring_patch_cs_in_place(r, p, job, ib) ((r)->funcs->patch_cs_in_place((p), (job), (ib)))
317 #define amdgpu_ring_emit_ib(r, job, ib, flags) ((r)->funcs->emit_ib((r), (job), (i
[all...]
H A Djpeg_v2_0.h52 void jpeg_v2_0_dec_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job,
H A Damdgpu_ids.h86 struct amdgpu_job *job, struct dma_fence **fence);
H A Damdgpu_uvd.h87 struct amdgpu_job *job,
/linux-master/include/linux/
H A Ddm-kcopyd.h58 * Submit a copy job to kcopyd. This is built on top of the
84 void dm_kcopyd_do_callback(void *job, int read_err, unsigned int long write_err);
H A Dhost1x.h240 int host1x_job_submit(struct host1x_job *job);
243 * host1x job
263 /* When refcount goes to zero, job can be freed */
269 /* Channel where job is submitted to */
272 /* client where the job originated */
294 /* Completion fence for job tracking */
298 /* Maximum time to wait for this job */
325 /* Fast-forward syncpoint increments on job timeout */
328 /* Callback called when job is freed */
329 void (*release)(struct host1x_job *job);
[all...]
/linux-master/drivers/gpu/drm/xe/
H A Dxe_guc_submit.h30 xe_guc_exec_queue_snapshot_capture(struct xe_sched_job *job);
H A Dxe_gt.c109 struct xe_sched_job *job; local
118 job = xe_bb_create_job(q, bb);
119 if (IS_ERR(job)) {
121 return PTR_ERR(job);
124 xe_sched_job_arm(job);
125 fence = dma_fence_get(&job->drm.s_fence->finished);
126 xe_sched_job_push(job);
153 struct xe_sched_job *job; local
204 job = xe_bb_create_job(q, bb);
205 if (IS_ERR(job)) {
[all...]
H A Dxe_guc_submit.c725 struct xe_sched_job *job = to_xe_sched_job(drm_job); local
726 struct xe_exec_queue *q = job->q;
734 trace_xe_sched_job_run(job);
736 if (!exec_queue_killed_or_banned(q) && !xe_sched_job_is_error(job)) {
740 q->ring_ops->emit_job(job);
745 xe_sched_job_set_error(job, -EOPNOTSUPP);
747 } else if (test_and_set_bit(JOB_FLAG_SUBMIT, &job->fence->flags)) {
748 return job->fence;
750 return dma_fence_get(job->fence);
756 struct xe_sched_job *job local
924 struct xe_sched_job *job = to_xe_sched_job(drm_job); local
1441 struct xe_sched_job *job = xe_sched_first_pending_job(sched); local
1794 xe_guc_exec_queue_snapshot_capture(struct xe_sched_job *job) argument
1945 struct xe_sched_job *job; local
[all...]
H A Dxe_sync.c211 int xe_sync_entry_add_deps(struct xe_sync_entry *sync, struct xe_sched_job *job) argument
216 err = drm_sched_job_add_dependency(&job->drm,
227 void xe_sync_entry_signal(struct xe_sync_entry *sync, struct xe_sched_job *job, argument
258 job->user_fence.used = true;
259 job->user_fence.addr = sync->addr;
260 job->user_fence.value = sync->timeline_value;
/linux-master/drivers/gpu/host1x/
H A Dcdma.h69 struct list_head sync_queue; /* job queue */
82 int host1x_cdma_begin(struct host1x_cdma *cdma, struct host1x_job *job);
86 void host1x_cdma_end(struct host1x_cdma *cdma, struct host1x_job *job);
H A Dchannel.c13 #include "job.h"
41 int host1x_job_submit(struct host1x_job *job) argument
43 struct host1x *host = dev_get_drvdata(job->channel->dev->parent);
45 return host1x_hw_channel_submit(host, job);
/linux-master/drivers/gpu/drm/imagination/
H A Dpvr_sync.c206 pvr_sync_add_dep_to_job(struct drm_sched_job *job, struct dma_fence *f) argument
220 return drm_sched_job_add_dependency(job, f);
236 err = drm_sched_job_add_dependency(job,
239 err = drm_sched_job_add_dependency(job, dma_fence_get(uf));
248 pvr_sync_add_deps_to_job(struct pvr_file *pvr_file, struct drm_sched_job *job, argument
283 err = pvr_sync_add_dep_to_job(job, fence);
/linux-master/drivers/scsi/libfc/
H A Dfc_lport.c135 * @job: The passthrough job
138 * @sg: job->reply_payload.sg_list
139 * @nents: job->reply_payload.sg_cnt
143 struct bsg_job *job; member in struct:fc_bsg_info
1958 struct bsg_job *job = info->job; local
1959 struct fc_bsg_reply *bsg_reply = job->reply;
1968 job->reply_len = sizeof(uint32_t);
1969 bsg_job_done(job, bsg_repl
2019 fc_lport_els_request(struct bsg_job *job, struct fc_lport *lport, u32 did, u32 tov) argument
2079 fc_lport_ct_request(struct bsg_job *job, struct fc_lport *lport, u32 did, u32 tov) argument
2137 fc_lport_bsg_request(struct bsg_job *job) argument
[all...]
/linux-master/drivers/gpu/drm/msm/disp/dpu1/
H A Ddpu_encoder_phys_wb.c552 struct drm_writeback_job *job)
560 if (!job->fb)
563 wb_enc->wb_job = job;
564 wb_enc->wb_conn = job->connector;
571 ret = msm_framebuffer_prepare(job->fb, aspace, false);
577 format = msm_framebuffer_format(job->fb);
580 format->pixel_format, job->fb->modifier);
587 ret = dpu_format_populate_layout(aspace, job->fb, &wb_cfg->dest);
593 wb_cfg->dest.width = job->fb->width;
594 wb_cfg->dest.height = job
551 dpu_encoder_phys_wb_prepare_wb_job(struct dpu_encoder_phys *phys_enc, struct drm_writeback_job *job) argument
610 dpu_encoder_phys_wb_cleanup_wb_job(struct dpu_encoder_phys *phys_enc, struct drm_writeback_job *job) argument
[all...]
/linux-master/drivers/gpu/drm/lima/
H A Dlima_sched.c103 static inline struct lima_sched_task *to_lima_task(struct drm_sched_job *job) argument
105 return container_of(job, struct lima_sched_task, base);
202 static struct dma_fence *lima_sched_run_job(struct drm_sched_job *job) argument
204 struct lima_sched_task *task = to_lima_task(job);
205 struct lima_sched_pipe *pipe = to_lima_pipe(job->sched);
211 if (job->s_fence->finished.error < 0)
399 static enum drm_gpu_sched_stat lima_sched_timedout_job(struct drm_sched_job *job) argument
401 struct lima_sched_pipe *pipe = to_lima_pipe(job->sched);
402 struct lima_sched_task *task = to_lima_task(job);
419 * In order to catch such cases and not report spurious Lima job
462 lima_sched_free_job(struct drm_sched_job *job) argument
[all...]
/linux-master/tools/testing/selftests/rcutorture/bin/
H A Dkvm-test-1-run-qemu.sh65 echo Monitoring qemu job at pid $qemu_pid `date`
68 echo Monitoring qemu job at yet-as-unknown pid `date`
132 echo Grace period for qemu job at pid $qemu_pid `date`
/linux-master/drivers/gpu/drm/tegra/
H A Ddrm.c181 struct host1x_job *job; local
197 job = host1x_job_alloc(context->channel, args->num_cmdbufs,
199 if (!job)
202 job->num_relocs = args->num_relocs;
203 job->client = client;
204 job->class = client->class;
205 job->serialize = true;
206 job->syncpt_recovery = true;
263 host1x_job_add_gather(job, bo, cmdbuf.words, cmdbuf.offset);
273 err = host1x_reloc_copy_from_user(&job
[all...]
/linux-master/tools/memory-model/scripts/
H A Dparseargs.sh115 --jobs|--job|-j)
/linux-master/drivers/scsi/libsas/
H A Dsas_internal.h85 void sas_smp_handler(struct bsg_job *job, struct Scsi_Host *shost,
114 extern void sas_smp_host_handler(struct bsg_job *job, struct Scsi_Host *shost);
116 static inline void sas_smp_host_handler(struct bsg_job *job, argument
121 bsg_job_done(job, -EINVAL, 0);

Completed in 518 milliseconds

12345678