/linux-master/fs/smb/client/ |
H A D | connect.c | 474 static int reconnect_target_unlocked(struct TCP_Server_Info *server, struct dfs_cache_tgt_list *tl, argument 483 tit = dfs_cache_get_tgt_iterator(tl); 487 /* Otherwise, try every dfs target in @tl */ 488 for (; tit; tit = dfs_cache_get_next_tgt(tl, tit)) { 502 DFS_CACHE_TGT_LIST(tl); 515 if (!dfs_cache_noreq_find(server->leaf_fullpath + 1, NULL, &tl)) 516 num_targets = dfs_cache_get_nr_tgts(&tl); 537 rc = reconnect_target_unlocked(server, &tl, &target_hint); 564 dfs_cache_free_tgts(&tl);
|
H A D | dfs_cache.c | 790 static int get_targets(struct cache_entry *ce, struct dfs_cache_tgt_list *tl) argument 793 struct list_head *head = &tl->tl_list; 797 memset(tl, 0, sizeof(*tl)); 821 tl->tl_numtgts = ce->numtgts;
|
H A D | dfs.c | 444 struct dfs_cache_tgt_list *tl) 454 tit = dfs_cache_get_tgt_iterator(tl); 461 for (; tit; tit = dfs_cache_get_next_tgt(tl, tit)) { 509 dfs_cache_free_tgts(tl); 513 list_replace_init(&ntl.tl_list, &tl->tl_list); 529 struct dfs_cache_tgt_list *tl) 537 rc = __tree_connect_dfs_target(xid, tcon, cifs_sb, tree, islink, tl); 548 dfs_cache_free_tgts(tl); 557 DFS_CACHE_TGT_LIST(tl); 607 dfs_cache_noreq_find(server->leaf_fullpath + 1, &ref, &tl)) { 442 __tree_connect_dfs_target(const unsigned int xid, struct cifs_tcon *tcon, struct cifs_sb_info *cifs_sb, char *tree, bool islink, struct dfs_cache_tgt_list *tl) argument 527 tree_connect_dfs_target(const unsigned int xid, struct cifs_tcon *tcon, struct cifs_sb_info *cifs_sb, char *tree, bool islink, struct dfs_cache_tgt_list *tl) argument [all...] |
H A D | dfs.h | 22 struct dfs_cache_tgt_list tl; member in struct:dfs_ref 40 #define ref_walk_tl(w) (&ref_walk_cur(w)->tl) 62 dfs_cache_free_tgts(&ref->tl); 93 tit = dfs_cache_get_tgt_iterator(&ref->tl); 95 tit = dfs_cache_get_next_tgt(&ref->tl, ref->tit); 129 struct dfs_info3_param *ref, struct dfs_cache_tgt_list *tl) 136 cifs_remap(cifs_sb), path, ref, tl); 128 dfs_get_referral(struct cifs_mount_ctx *mnt_ctx, const char *path, struct dfs_info3_param *ref, struct dfs_cache_tgt_list *tl) argument
|
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 4293 struct intel_timeline *tl; local 4295 list_for_each_entry(tl, &engine->status_page.timelines, engine_link) 4296 intel_timeline_reset_seqno(tl);
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_engine_cs.c | 2023 struct intel_timeline *tl; local 2034 tl = rcu_dereference(rq->timeline); 2035 if (!kref_get_unless_zero(&tl->kref)) 2036 tl = NULL; 2039 return tl; 2047 struct intel_timeline *tl = get_timeline(rq); local 2052 tl ? tl->hwsp_offset : 0, 2057 if (tl) 2058 intel_timeline_put(tl); 2316 struct intel_timeline *tl = get_timeline(rq); local 2518 struct intel_timeline *tl = request->context->timeline; local [all...] |
H A D | intel_execlists_submission.c | 212 __active_request(const struct intel_timeline * const tl, argument 218 list_for_each_entry_from_reverse(rq, &tl->requests, link) { 233 active_request(const struct intel_timeline * const tl, struct i915_request *rq) argument 235 return __active_request(tl, rq, 0); 2851 struct intel_timeline *tl; local 2853 list_for_each_entry(tl, &engine->status_page.timelines, engine_link) 2854 intel_timeline_reset_seqno(tl);
|
H A D | intel_engine_pm.c | 110 struct intel_timeline *tl, 139 if (!atomic_fetch_inc(&tl->active_count)) 140 list_add_tail(&tl->link, &timelines->active_list); 109 __queue_and_release_pm(struct i915_request *rq, struct intel_timeline *tl, struct intel_engine_cs *engine) argument
|
H A D | selftest_engine_heartbeat.c | 20 static int timeline_sync(struct intel_timeline *tl) argument 25 fence = i915_active_fence_get(&tl->last_request);
|
H A D | selftest_context.c | 17 struct intel_timeline *tl = i915_request_timeline(rq); local 21 intel_timeline_get(tl); 35 lockdep_unpin_lock(&tl->mutex, rq->cookie); 36 mutex_unlock(&tl->mutex); 39 intel_timeline_put(tl); 46 struct intel_timeline *tl = ce->timeline; local 49 mutex_lock(&tl->mutex); 54 if (list_empty(&tl->requests)) 57 rq = list_last_entry(&tl->requests, typeof(*rq), link); 68 mutex_unlock(&tl [all...] |
H A D | intel_reset.c | 1041 struct intel_timeline *tl; local 1064 list_for_each_entry(tl, &timelines->active_list, link) { 1067 fence = i915_active_fence_get(&tl->last_request); 1085 tl = list_entry(&timelines->active_list, typeof(*tl), link);
|
H A D | gen8_engine_cs.c | 427 const struct intel_timeline *tl; local 430 tl = rcu_dereference_protected(rq->timeline, 434 return page_mask_bits(tl->hwsp_offset) + offset_in_page(rq->hwsp_seqno);
|
/linux-master/arch/x86/net/ |
H A D | bpf_jit_comp.c | 2600 struct bpf_tramp_links *tl, int stack_size, 2607 for (i = 0; i < tl->nr_links; i++) { 2608 if (invoke_bpf_prog(m, &prog, tl->links[i], stack_size, 2617 struct bpf_tramp_links *tl, int stack_size, 2629 for (i = 0; i < tl->nr_links; i++) { 2630 if (invoke_bpf_prog(m, &prog, tl->links[i], stack_size, run_ctx_off, true, 2599 invoke_bpf(const struct btf_func_model *m, u8 **pprog, struct bpf_tramp_links *tl, int stack_size, int run_ctx_off, bool save_ret, void *image, void *rw_image) argument 2616 invoke_bpf_mod_ret(const struct btf_func_model *m, u8 **pprog, struct bpf_tramp_links *tl, int stack_size, int run_ctx_off, u8 **branches, void *image, void *rw_image) argument
|
/linux-master/arch/arm64/net/ |
H A D | bpf_jit_comp.c | 1880 static void invoke_bpf_mod_ret(struct jit_ctx *ctx, struct bpf_tramp_links *tl, argument 1890 for (i = 0; i < tl->nr_links; i++) { 1891 invoke_bpf_prog(ctx, tl->links[i], args_off, retval_off,
|
/linux-master/drivers/nvme/target/ |
H A D | trace.c | 251 u32 tl = get_unaligned_le32(spc + 4); local 253 trace_seq_printf(p, "spsp0=%02x, spsp1=%02x, secp=%02x, tl=%u", 254 spsp0, spsp1, secp, tl);
|
/linux-master/drivers/nvme/host/ |
H A D | trace.c | 393 u32 tl = get_unaligned_le32(spc + 4); local 395 trace_seq_printf(p, "spsp0=%02x, spsp1=%02x, secp=%02x, tl=%u", 396 spsp0, spsp1, secp, tl);
|
/linux-master/drivers/crypto/allwinner/sun8i-ce/ |
H A D | sun8i-ce-hash.c | 386 cet = chan->tl;
|
/linux-master/arch/sparc/include/asm/ |
H A D | hypervisor.h | 1374 unsigned char tl; /* Trap level */ member in struct:hv_trap_trace_entry
|
/linux-master/arch/s390/kvm/ |
H A D | gaccess.c | 32 unsigned long tl : 2; /* Region- or Segment-Table Length */ member in struct:asce::__anon61 54 unsigned long tl : 2; /* Region-Second-Table Length */ member in struct:region1_table_entry::__anon63 69 unsigned long tl : 2; /* Region-Third-Table Length */ member in struct:region2_table_entry::__anon64 83 unsigned long tl : 2; /* Segment-Table Length */ member in struct:region3_table_entry_fc0 638 if (vaddr.rfx01 > asce.tl) 645 if (vaddr.rsx01 > asce.tl) 652 if (vaddr.rtx01 > asce.tl) 659 if (vaddr.sx01 > asce.tl) 676 if (vaddr.rsx01 < rfte.tf || vaddr.rsx01 > rfte.tl) 694 if (vaddr.rtx01 < rste.tf || vaddr.rtx01 > rste.tl) [all...] |
/linux-master/security/apparmor/ |
H A D | lsm.c | 1016 struct aa_label *cl, *tl; local 1020 tl = aa_get_newest_cred_label(tc); 1026 error = aa_may_signal(cred, cl, tc, tl, sig); 1030 error = aa_may_signal(current_cred(), cl, tc, tl, sig); 1033 aa_put_label(tl);
|
/linux-master/drivers/s390/net/ |
H A D | fsm.c | 135 fsm_timer *this = from_timer(this, t, tl); 151 timer_setup(&this->tl, fsm_expire_timer, 0); 161 del_timer(&this->tl); 173 timer_setup(&this->tl, fsm_expire_timer, 0); 176 this->tl.expires = jiffies + (millisec * HZ) / 1000; 177 add_timer(&this->tl); 191 del_timer(&this->tl); 192 timer_setup(&this->tl, fsm_expire_timer, 0); 195 this->tl.expires = jiffies + (millisec * HZ) / 1000; 196 add_timer(&this->tl); [all...] |
/linux-master/drivers/gpu/drm/radeon/ |
H A D | r600_dpm.c | 219 int r600_calculate_at(u32 t, u32 h, u32 fh, u32 fl, u32 *tl, u32 *th) argument 235 *tl = t + al;
|
H A D | r600_dpm.h | 141 int r600_calculate_at(u32 t, u32 h, u32 fh, u32 fl, u32 *tl, u32 *th);
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_request.c | 420 struct intel_timeline * const tl = i915_request_timeline(rq); local 427 tmp = list_first_entry(&tl->requests, typeof(*tmp), link); 824 static void retire_requests(struct intel_timeline *tl) argument 828 list_for_each_entry_safe(rq, rn, &tl->requests, link) 834 request_alloc_slow(struct intel_timeline *tl, argument 849 if (list_empty(&tl->requests)) 853 rq = list_first_entry(&tl->requests, typeof(*rq), link); 862 rq = list_last_entry(&tl->requests, typeof(*rq), link); 866 retire_requests(tl); 896 struct intel_timeline *tl local 1031 struct intel_timeline *tl; local 1248 intel_timeline_sync_has_start(struct intel_timeline *tl, struct dma_fence *fence) argument 1256 intel_timeline_sync_set_start(struct intel_timeline *tl, const struct dma_fence *fence) argument 1846 struct intel_timeline * const tl = i915_request_timeline(rq); local [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_execbuffer.c | 2484 struct intel_timeline *tl = ce->timeline; local 2501 list_for_each_entry(rq, &tl->requests, link) { 2509 if (&rq->link == &tl->requests) 2518 struct intel_timeline *tl; local 2529 tl = intel_context_timeline_lock(ce); 2530 if (IS_ERR(tl)) 2531 return PTR_ERR(tl); 2536 intel_context_timeline_unlock(tl); 3075 static void retire_requests(struct intel_timeline *tl, struct i915_request *end) argument 3079 list_for_each_entry_safe(rq, rn, &tl 3087 struct intel_timeline * const tl = i915_request_timeline(rq); local [all...] |