/linux-master/drivers/accel/habanalabs/common/ |
H A D | command_submission.c | 267 struct hl_cs_job *job = container_of(ref, struct hl_cs_job, refcount); local 269 kfree(job); 272 static void hl_cs_job_put(struct hl_cs_job *job) argument 274 kref_put(&job->refcount, cs_job_do_release); 299 static bool is_cb_patched(struct hl_device *hdev, struct hl_cs_job *job) argument 302 return (job->queue_type == QUEUE_TYPE_EXT); 309 * @job : pointer to the job that holds the command submission info 316 static int cs_parser(struct hl_fpriv *hpriv, struct hl_cs_job *job) argument 322 parser.ctx_id = job 361 hl_complete_job(struct hl_device *hdev, struct hl_cs_job *job) argument 698 struct hl_cs_job *job, *tmp; local 1028 struct hl_cs_job *job, *tmp; local 1168 struct hl_cs_job *job = container_of(work, struct hl_cs_job, local 1181 struct hl_cs_job *job, *tmp; local 1303 struct hl_cs_job *job; local 1499 struct hl_cs_job *job; local 1943 struct hl_cs_job *job; local [all...] |
/linux-master/drivers/media/platform/renesas/ |
H A D | rcar_fdp1.c | 551 /* A job can only be on one list at a time */ 628 * while preparing our job lists. 648 * list_remove_job: Take the first item off the specified job list 650 * Returns: pointer to a job, or NULL if the list is empty. 655 struct fdp1_job *job; local 659 job = list_first_entry_or_null(list, struct fdp1_job, list); 660 if (job) 661 list_del(&job->list); 664 return job; 668 * list_add_job: Add a job t 672 list_add_job(struct fdp1_dev *fdp1, struct list_head *list, struct fdp1_job *job) argument 688 fdp1_job_free(struct fdp1_dev *fdp1, struct fdp1_job *job) argument 696 queue_job(struct fdp1_dev *fdp1, struct fdp1_job *job) argument 706 queue_hw_job(struct fdp1_dev *fdp1, struct fdp1_job *job) argument 879 fdp1_configure_rpf(struct fdp1_ctx *ctx, struct fdp1_job *job) argument 938 fdp1_configure_wpf(struct fdp1_ctx *ctx, struct fdp1_job *job) argument 999 fdp1_configure_deint_mode(struct fdp1_ctx *ctx, struct fdp1_job *job) argument 1070 struct fdp1_job *job; local 1185 struct fdp1_job *job; local 1310 struct fdp1_job *job = get_hw_queued_job(fdp1); local 2003 struct fdp1_job *job; local [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_sync.h | 58 int amdgpu_sync_push_to_job(struct amdgpu_sync *sync, struct amdgpu_job *job);
|
H A D | vcn_sw_ring.c | 44 void vcn_dec_sw_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, argument 47 uint32_t vmid = AMDGPU_JOB_GET_VMID(job);
|
H A D | amdgpu_fence.c | 133 * @job: job the fence is embedded in 139 int amdgpu_fence_emit(struct amdgpu_ring *ring, struct dma_fence **f, struct amdgpu_job *job, argument 149 if (job == NULL) { 157 /* take use of job-embedded fence */ 158 fence = &job->hw_fence; 162 if (job && job->job_run_counter) { 168 if (job) { 703 * amdgpu_fence_driver_clear_job_fences - clear job embedde 717 struct amdgpu_job *job; local 785 struct amdgpu_job *job = container_of(f, struct amdgpu_job, hw_fence); local 815 struct amdgpu_job *job = container_of(f, struct amdgpu_job, hw_fence); local [all...] |
H A D | amdgpu_ring.h | 143 int amdgpu_fence_emit(struct amdgpu_ring *ring, struct dma_fence **fence, struct amdgpu_job *job, 180 struct amdgpu_job *job, 183 struct amdgpu_job *job, 190 struct amdgpu_job *job, 310 #define amdgpu_ring_parse_cs(r, p, job, ib) ((r)->funcs->parse_cs((p), (job), (ib))) 311 #define amdgpu_ring_patch_cs_in_place(r, p, job, ib) ((r)->funcs->patch_cs_in_place((p), (job), (ib))) 317 #define amdgpu_ring_emit_ib(r, job, ib, flags) ((r)->funcs->emit_ib((r), (job), (i [all...] |
H A D | jpeg_v2_0.h | 52 void jpeg_v2_0_dec_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job,
|
H A D | amdgpu_ids.h | 86 struct amdgpu_job *job, struct dma_fence **fence);
|
H A D | amdgpu_uvd.h | 87 struct amdgpu_job *job,
|
/linux-master/include/linux/ |
H A D | dm-kcopyd.h | 58 * Submit a copy job to kcopyd. This is built on top of the 84 void dm_kcopyd_do_callback(void *job, int read_err, unsigned int long write_err);
|
H A D | host1x.h | 240 int host1x_job_submit(struct host1x_job *job); 243 * host1x job 263 /* When refcount goes to zero, job can be freed */ 269 /* Channel where job is submitted to */ 272 /* client where the job originated */ 294 /* Completion fence for job tracking */ 298 /* Maximum time to wait for this job */ 325 /* Fast-forward syncpoint increments on job timeout */ 328 /* Callback called when job is freed */ 329 void (*release)(struct host1x_job *job); [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_guc_submit.h | 30 xe_guc_exec_queue_snapshot_capture(struct xe_sched_job *job);
|
H A D | xe_gt.c | 109 struct xe_sched_job *job; local 118 job = xe_bb_create_job(q, bb); 119 if (IS_ERR(job)) { 121 return PTR_ERR(job); 124 xe_sched_job_arm(job); 125 fence = dma_fence_get(&job->drm.s_fence->finished); 126 xe_sched_job_push(job); 153 struct xe_sched_job *job; local 204 job = xe_bb_create_job(q, bb); 205 if (IS_ERR(job)) { [all...] |
H A D | xe_guc_submit.c | 725 struct xe_sched_job *job = to_xe_sched_job(drm_job); local 726 struct xe_exec_queue *q = job->q; 734 trace_xe_sched_job_run(job); 736 if (!exec_queue_killed_or_banned(q) && !xe_sched_job_is_error(job)) { 740 q->ring_ops->emit_job(job); 745 xe_sched_job_set_error(job, -EOPNOTSUPP); 747 } else if (test_and_set_bit(JOB_FLAG_SUBMIT, &job->fence->flags)) { 748 return job->fence; 750 return dma_fence_get(job->fence); 756 struct xe_sched_job *job local 924 struct xe_sched_job *job = to_xe_sched_job(drm_job); local 1441 struct xe_sched_job *job = xe_sched_first_pending_job(sched); local 1794 xe_guc_exec_queue_snapshot_capture(struct xe_sched_job *job) argument 1945 struct xe_sched_job *job; local [all...] |
H A D | xe_sync.c | 211 int xe_sync_entry_add_deps(struct xe_sync_entry *sync, struct xe_sched_job *job) argument 216 err = drm_sched_job_add_dependency(&job->drm, 227 void xe_sync_entry_signal(struct xe_sync_entry *sync, struct xe_sched_job *job, argument 258 job->user_fence.used = true; 259 job->user_fence.addr = sync->addr; 260 job->user_fence.value = sync->timeline_value;
|
/linux-master/drivers/gpu/host1x/ |
H A D | cdma.h | 69 struct list_head sync_queue; /* job queue */ 82 int host1x_cdma_begin(struct host1x_cdma *cdma, struct host1x_job *job); 86 void host1x_cdma_end(struct host1x_cdma *cdma, struct host1x_job *job);
|
H A D | channel.c | 13 #include "job.h" 41 int host1x_job_submit(struct host1x_job *job) argument 43 struct host1x *host = dev_get_drvdata(job->channel->dev->parent); 45 return host1x_hw_channel_submit(host, job);
|
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_sync.c | 206 pvr_sync_add_dep_to_job(struct drm_sched_job *job, struct dma_fence *f) argument 220 return drm_sched_job_add_dependency(job, f); 236 err = drm_sched_job_add_dependency(job, 239 err = drm_sched_job_add_dependency(job, dma_fence_get(uf)); 248 pvr_sync_add_deps_to_job(struct pvr_file *pvr_file, struct drm_sched_job *job, argument 283 err = pvr_sync_add_dep_to_job(job, fence);
|
/linux-master/drivers/scsi/libfc/ |
H A D | fc_lport.c | 135 * @job: The passthrough job 138 * @sg: job->reply_payload.sg_list 139 * @nents: job->reply_payload.sg_cnt 143 struct bsg_job *job; member in struct:fc_bsg_info 1958 struct bsg_job *job = info->job; local 1959 struct fc_bsg_reply *bsg_reply = job->reply; 1968 job->reply_len = sizeof(uint32_t); 1969 bsg_job_done(job, bsg_repl 2019 fc_lport_els_request(struct bsg_job *job, struct fc_lport *lport, u32 did, u32 tov) argument 2079 fc_lport_ct_request(struct bsg_job *job, struct fc_lport *lport, u32 did, u32 tov) argument 2137 fc_lport_bsg_request(struct bsg_job *job) argument [all...] |
/linux-master/drivers/gpu/drm/msm/disp/dpu1/ |
H A D | dpu_encoder_phys_wb.c | 552 struct drm_writeback_job *job) 560 if (!job->fb) 563 wb_enc->wb_job = job; 564 wb_enc->wb_conn = job->connector; 571 ret = msm_framebuffer_prepare(job->fb, aspace, false); 577 format = msm_framebuffer_format(job->fb); 580 format->pixel_format, job->fb->modifier); 587 ret = dpu_format_populate_layout(aspace, job->fb, &wb_cfg->dest); 593 wb_cfg->dest.width = job->fb->width; 594 wb_cfg->dest.height = job 551 dpu_encoder_phys_wb_prepare_wb_job(struct dpu_encoder_phys *phys_enc, struct drm_writeback_job *job) argument 610 dpu_encoder_phys_wb_cleanup_wb_job(struct dpu_encoder_phys *phys_enc, struct drm_writeback_job *job) argument [all...] |
/linux-master/drivers/gpu/drm/lima/ |
H A D | lima_sched.c | 103 static inline struct lima_sched_task *to_lima_task(struct drm_sched_job *job) argument 105 return container_of(job, struct lima_sched_task, base); 202 static struct dma_fence *lima_sched_run_job(struct drm_sched_job *job) argument 204 struct lima_sched_task *task = to_lima_task(job); 205 struct lima_sched_pipe *pipe = to_lima_pipe(job->sched); 211 if (job->s_fence->finished.error < 0) 399 static enum drm_gpu_sched_stat lima_sched_timedout_job(struct drm_sched_job *job) argument 401 struct lima_sched_pipe *pipe = to_lima_pipe(job->sched); 402 struct lima_sched_task *task = to_lima_task(job); 419 * In order to catch such cases and not report spurious Lima job 462 lima_sched_free_job(struct drm_sched_job *job) argument [all...] |
/linux-master/tools/testing/selftests/rcutorture/bin/ |
H A D | kvm-test-1-run-qemu.sh | 65 echo Monitoring qemu job at pid $qemu_pid `date` 68 echo Monitoring qemu job at yet-as-unknown pid `date` 132 echo Grace period for qemu job at pid $qemu_pid `date`
|
/linux-master/drivers/gpu/drm/tegra/ |
H A D | drm.c | 181 struct host1x_job *job; local 197 job = host1x_job_alloc(context->channel, args->num_cmdbufs, 199 if (!job) 202 job->num_relocs = args->num_relocs; 203 job->client = client; 204 job->class = client->class; 205 job->serialize = true; 206 job->syncpt_recovery = true; 263 host1x_job_add_gather(job, bo, cmdbuf.words, cmdbuf.offset); 273 err = host1x_reloc_copy_from_user(&job [all...] |
/linux-master/tools/memory-model/scripts/ |
H A D | parseargs.sh | 115 --jobs|--job|-j)
|
/linux-master/drivers/scsi/libsas/ |
H A D | sas_internal.h | 85 void sas_smp_handler(struct bsg_job *job, struct Scsi_Host *shost, 114 extern void sas_smp_host_handler(struct bsg_job *job, struct Scsi_Host *shost); 116 static inline void sas_smp_host_handler(struct bsg_job *job, argument 121 bsg_job_done(job, -EINVAL, 0);
|