Searched refs:tl (Results 1 - 25 of 115) sorted by last modified time

12345

/linux-master/fs/smb/client/
H A Dconnect.c474 static int reconnect_target_unlocked(struct TCP_Server_Info *server, struct dfs_cache_tgt_list *tl, argument
483 tit = dfs_cache_get_tgt_iterator(tl);
487 /* Otherwise, try every dfs target in @tl */
488 for (; tit; tit = dfs_cache_get_next_tgt(tl, tit)) {
502 DFS_CACHE_TGT_LIST(tl);
515 if (!dfs_cache_noreq_find(server->leaf_fullpath + 1, NULL, &tl))
516 num_targets = dfs_cache_get_nr_tgts(&tl);
537 rc = reconnect_target_unlocked(server, &tl, &target_hint);
564 dfs_cache_free_tgts(&tl);
H A Ddfs_cache.c790 static int get_targets(struct cache_entry *ce, struct dfs_cache_tgt_list *tl) argument
793 struct list_head *head = &tl->tl_list;
797 memset(tl, 0, sizeof(*tl));
821 tl->tl_numtgts = ce->numtgts;
H A Ddfs.c444 struct dfs_cache_tgt_list *tl)
454 tit = dfs_cache_get_tgt_iterator(tl);
461 for (; tit; tit = dfs_cache_get_next_tgt(tl, tit)) {
509 dfs_cache_free_tgts(tl);
513 list_replace_init(&ntl.tl_list, &tl->tl_list);
529 struct dfs_cache_tgt_list *tl)
537 rc = __tree_connect_dfs_target(xid, tcon, cifs_sb, tree, islink, tl);
548 dfs_cache_free_tgts(tl);
557 DFS_CACHE_TGT_LIST(tl);
607 dfs_cache_noreq_find(server->leaf_fullpath + 1, &ref, &tl)) {
442 __tree_connect_dfs_target(const unsigned int xid, struct cifs_tcon *tcon, struct cifs_sb_info *cifs_sb, char *tree, bool islink, struct dfs_cache_tgt_list *tl) argument
527 tree_connect_dfs_target(const unsigned int xid, struct cifs_tcon *tcon, struct cifs_sb_info *cifs_sb, char *tree, bool islink, struct dfs_cache_tgt_list *tl) argument
[all...]
H A Ddfs.h22 struct dfs_cache_tgt_list tl; member in struct:dfs_ref
40 #define ref_walk_tl(w) (&ref_walk_cur(w)->tl)
62 dfs_cache_free_tgts(&ref->tl);
93 tit = dfs_cache_get_tgt_iterator(&ref->tl);
95 tit = dfs_cache_get_next_tgt(&ref->tl, ref->tit);
129 struct dfs_info3_param *ref, struct dfs_cache_tgt_list *tl)
136 cifs_remap(cifs_sb), path, ref, tl);
128 dfs_get_referral(struct cifs_mount_ctx *mnt_ctx, const char *path, struct dfs_info3_param *ref, struct dfs_cache_tgt_list *tl) argument
/linux-master/drivers/gpu/drm/i915/gt/uc/
H A Dintel_guc_submission.c4293 struct intel_timeline *tl; local
4295 list_for_each_entry(tl, &engine->status_page.timelines, engine_link)
4296 intel_timeline_reset_seqno(tl);
/linux-master/drivers/gpu/drm/i915/gt/
H A Dintel_engine_cs.c2023 struct intel_timeline *tl; local
2034 tl = rcu_dereference(rq->timeline);
2035 if (!kref_get_unless_zero(&tl->kref))
2036 tl = NULL;
2039 return tl;
2047 struct intel_timeline *tl = get_timeline(rq); local
2052 tl ? tl->hwsp_offset : 0,
2057 if (tl)
2058 intel_timeline_put(tl);
2316 struct intel_timeline *tl = get_timeline(rq); local
2518 struct intel_timeline *tl = request->context->timeline; local
[all...]
H A Dintel_execlists_submission.c212 __active_request(const struct intel_timeline * const tl, argument
218 list_for_each_entry_from_reverse(rq, &tl->requests, link) {
233 active_request(const struct intel_timeline * const tl, struct i915_request *rq) argument
235 return __active_request(tl, rq, 0);
2851 struct intel_timeline *tl; local
2853 list_for_each_entry(tl, &engine->status_page.timelines, engine_link)
2854 intel_timeline_reset_seqno(tl);
H A Dintel_engine_pm.c110 struct intel_timeline *tl,
139 if (!atomic_fetch_inc(&tl->active_count))
140 list_add_tail(&tl->link, &timelines->active_list);
109 __queue_and_release_pm(struct i915_request *rq, struct intel_timeline *tl, struct intel_engine_cs *engine) argument
H A Dselftest_engine_heartbeat.c20 static int timeline_sync(struct intel_timeline *tl) argument
25 fence = i915_active_fence_get(&tl->last_request);
H A Dselftest_context.c17 struct intel_timeline *tl = i915_request_timeline(rq); local
21 intel_timeline_get(tl);
35 lockdep_unpin_lock(&tl->mutex, rq->cookie);
36 mutex_unlock(&tl->mutex);
39 intel_timeline_put(tl);
46 struct intel_timeline *tl = ce->timeline; local
49 mutex_lock(&tl->mutex);
54 if (list_empty(&tl->requests))
57 rq = list_last_entry(&tl->requests, typeof(*rq), link);
68 mutex_unlock(&tl
[all...]
H A Dintel_reset.c1041 struct intel_timeline *tl; local
1064 list_for_each_entry(tl, &timelines->active_list, link) {
1067 fence = i915_active_fence_get(&tl->last_request);
1085 tl = list_entry(&timelines->active_list, typeof(*tl), link);
H A Dgen8_engine_cs.c427 const struct intel_timeline *tl; local
430 tl = rcu_dereference_protected(rq->timeline,
434 return page_mask_bits(tl->hwsp_offset) + offset_in_page(rq->hwsp_seqno);
/linux-master/arch/x86/net/
H A Dbpf_jit_comp.c2600 struct bpf_tramp_links *tl, int stack_size,
2607 for (i = 0; i < tl->nr_links; i++) {
2608 if (invoke_bpf_prog(m, &prog, tl->links[i], stack_size,
2617 struct bpf_tramp_links *tl, int stack_size,
2629 for (i = 0; i < tl->nr_links; i++) {
2630 if (invoke_bpf_prog(m, &prog, tl->links[i], stack_size, run_ctx_off, true,
2599 invoke_bpf(const struct btf_func_model *m, u8 **pprog, struct bpf_tramp_links *tl, int stack_size, int run_ctx_off, bool save_ret, void *image, void *rw_image) argument
2616 invoke_bpf_mod_ret(const struct btf_func_model *m, u8 **pprog, struct bpf_tramp_links *tl, int stack_size, int run_ctx_off, u8 **branches, void *image, void *rw_image) argument
/linux-master/arch/arm64/net/
H A Dbpf_jit_comp.c1880 static void invoke_bpf_mod_ret(struct jit_ctx *ctx, struct bpf_tramp_links *tl, argument
1890 for (i = 0; i < tl->nr_links; i++) {
1891 invoke_bpf_prog(ctx, tl->links[i], args_off, retval_off,
/linux-master/drivers/nvme/target/
H A Dtrace.c251 u32 tl = get_unaligned_le32(spc + 4); local
253 trace_seq_printf(p, "spsp0=%02x, spsp1=%02x, secp=%02x, tl=%u",
254 spsp0, spsp1, secp, tl);
/linux-master/drivers/nvme/host/
H A Dtrace.c393 u32 tl = get_unaligned_le32(spc + 4); local
395 trace_seq_printf(p, "spsp0=%02x, spsp1=%02x, secp=%02x, tl=%u",
396 spsp0, spsp1, secp, tl);
/linux-master/drivers/crypto/allwinner/sun8i-ce/
H A Dsun8i-ce-hash.c386 cet = chan->tl;
/linux-master/arch/sparc/include/asm/
H A Dhypervisor.h1374 unsigned char tl; /* Trap level */ member in struct:hv_trap_trace_entry
/linux-master/arch/s390/kvm/
H A Dgaccess.c32 unsigned long tl : 2; /* Region- or Segment-Table Length */ member in struct:asce::__anon61
54 unsigned long tl : 2; /* Region-Second-Table Length */ member in struct:region1_table_entry::__anon63
69 unsigned long tl : 2; /* Region-Third-Table Length */ member in struct:region2_table_entry::__anon64
83 unsigned long tl : 2; /* Segment-Table Length */ member in struct:region3_table_entry_fc0
638 if (vaddr.rfx01 > asce.tl)
645 if (vaddr.rsx01 > asce.tl)
652 if (vaddr.rtx01 > asce.tl)
659 if (vaddr.sx01 > asce.tl)
676 if (vaddr.rsx01 < rfte.tf || vaddr.rsx01 > rfte.tl)
694 if (vaddr.rtx01 < rste.tf || vaddr.rtx01 > rste.tl)
[all...]
/linux-master/security/apparmor/
H A Dlsm.c1016 struct aa_label *cl, *tl; local
1020 tl = aa_get_newest_cred_label(tc);
1026 error = aa_may_signal(cred, cl, tc, tl, sig);
1030 error = aa_may_signal(current_cred(), cl, tc, tl, sig);
1033 aa_put_label(tl);
/linux-master/drivers/s390/net/
H A Dfsm.c135 fsm_timer *this = from_timer(this, t, tl);
151 timer_setup(&this->tl, fsm_expire_timer, 0);
161 del_timer(&this->tl);
173 timer_setup(&this->tl, fsm_expire_timer, 0);
176 this->tl.expires = jiffies + (millisec * HZ) / 1000;
177 add_timer(&this->tl);
191 del_timer(&this->tl);
192 timer_setup(&this->tl, fsm_expire_timer, 0);
195 this->tl.expires = jiffies + (millisec * HZ) / 1000;
196 add_timer(&this->tl);
[all...]
/linux-master/drivers/gpu/drm/radeon/
H A Dr600_dpm.c219 int r600_calculate_at(u32 t, u32 h, u32 fh, u32 fl, u32 *tl, u32 *th) argument
235 *tl = t + al;
H A Dr600_dpm.h141 int r600_calculate_at(u32 t, u32 h, u32 fh, u32 fl, u32 *tl, u32 *th);
/linux-master/drivers/gpu/drm/i915/
H A Di915_request.c420 struct intel_timeline * const tl = i915_request_timeline(rq); local
427 tmp = list_first_entry(&tl->requests, typeof(*tmp), link);
824 static void retire_requests(struct intel_timeline *tl) argument
828 list_for_each_entry_safe(rq, rn, &tl->requests, link)
834 request_alloc_slow(struct intel_timeline *tl, argument
849 if (list_empty(&tl->requests))
853 rq = list_first_entry(&tl->requests, typeof(*rq), link);
862 rq = list_last_entry(&tl->requests, typeof(*rq), link);
866 retire_requests(tl);
896 struct intel_timeline *tl local
1031 struct intel_timeline *tl; local
1248 intel_timeline_sync_has_start(struct intel_timeline *tl, struct dma_fence *fence) argument
1256 intel_timeline_sync_set_start(struct intel_timeline *tl, const struct dma_fence *fence) argument
1846 struct intel_timeline * const tl = i915_request_timeline(rq); local
[all...]
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_execbuffer.c2484 struct intel_timeline *tl = ce->timeline; local
2501 list_for_each_entry(rq, &tl->requests, link) {
2509 if (&rq->link == &tl->requests)
2518 struct intel_timeline *tl; local
2529 tl = intel_context_timeline_lock(ce);
2530 if (IS_ERR(tl))
2531 return PTR_ERR(tl);
2536 intel_context_timeline_unlock(tl);
3075 static void retire_requests(struct intel_timeline *tl, struct i915_request *end) argument
3079 list_for_each_entry_safe(rq, rn, &tl
3087 struct intel_timeline * const tl = i915_request_timeline(rq); local
[all...]

Completed in 297 milliseconds

12345