/linux-master/tools/testing/selftests/damon/ |
H A D | sysfs_update_schemes_tried_regions_wss_estimation.py | 14 contexts=[_damon_sysfs.DamonCtx( 21 )] # contexts 38 kdamonds.kdamonds[0].contexts[0].schemes[0].tried_bytes)
|
H A D | sysfs_update_schemes_tried_regions_hang.py | 12 contexts=[_damon_sysfs.DamonCtx( 18 )] # contexts
|
H A D | damos_quota.py | 18 contexts=[_damon_sysfs.DamonCtx( 28 )] # contexts 49 scheme = kdamonds.kdamonds[0].contexts[0].schemes[0]
|
H A D | _damon_sysfs.py | 261 return os.path.join(self.kdamond.sysfs_dir(), 'contexts', 303 contexts = None variable in class:Kdamond 307 def __init__(self, contexts=[]): 308 self.contexts = contexts 309 for idx, context in enumerate(self.contexts): 318 'contexts', 'nr_contexts') 322 if int(content) != len(self.contexts): 323 err = write_file(nr_contexts_file, '%d' % len(self.contexts)) 327 for context in self.contexts [all...] |
H A D | damos_apply_interval.py | 16 contexts=[_damon_sysfs.DamonCtx( 34 )] # contexts 50 schemes = kdamonds.kdamonds[0].contexts[0].schemes
|
/linux-master/mm/damon/ |
H A D | sysfs.c | 810 * contexts directory 824 static void damon_sysfs_contexts_rm_dirs(struct damon_sysfs_contexts *contexts) argument 826 struct damon_sysfs_context **contexts_arr = contexts->contexts_arr; 829 for (i = 0; i < contexts->nr; i++) { 833 contexts->nr = 0; 835 contexts->contexts_arr = NULL; 838 static int damon_sysfs_contexts_add_dirs(struct damon_sysfs_contexts *contexts, argument 844 damon_sysfs_contexts_rm_dirs(contexts); 852 contexts->contexts_arr = contexts_arr; 857 damon_sysfs_contexts_rm_dirs(contexts); 885 struct damon_sysfs_contexts *contexts = container_of(kobj, local 894 struct damon_sysfs_contexts *contexts; local 941 struct damon_sysfs_contexts *contexts; member in struct:damon_sysfs_kdamond 952 struct damon_sysfs_contexts *contexts; local [all...] |
/linux-master/drivers/md/dm-vdo/ |
H A D | vdo.c | 408 * @contexts: An array of queue_count contexts, one for each individual queue; may be NULL. 418 unsigned int queue_count, void *contexts[]) 436 type, queue_count, contexts, &thread->queue); 416 vdo_make_thread(struct vdo *vdo, thread_id_t thread_id, const struct vdo_work_queue_type *type, unsigned int queue_count, void *contexts[]) argument
|
H A D | vdo.h | 292 unsigned int queue_count, void *contexts[]);
|
H A D | dedupe.c | 94 * hash_lock's agent, added to the list of pending contexts, and then sent to the index. The 103 * pending list will be searched for any contexts in the pending state which have been running for 104 * too long. Those contexts are transitioned to the DEDUPE_CONTEXT_TIMED_OUT state and moved to the 106 * data_vios associated with timed out contexts are sent to continue processing their write 114 * be searched for any contexts which are timed out and complete. One of these will be used 2290 VDO_ASSERT_LOG_ONLY(READ_ONCE(zone->active) == 0, "all contexts inactive"); 2383 struct dedupe_context *context = &zone->contexts[i];
|
H A D | dedupe.h | 57 /* These fields are used to manage the dedupe contexts */ 66 /* The dedupe contexts for querying the index from this zone */ 67 struct dedupe_context contexts[MAXIMUM_VDO_USER_VIOS]; member in struct:hash_zone
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_perf.c | 77 * the metrics of other processes/contexts. For these use cases a read() based 345 * @single_context: Whether a single or all gpu contexts should be monitored 834 * contexts while filtering since we can't stop userspace 843 * associated with other contexts. Note that the hardware 849 * other, unrelated contexts) we simply elect to forward them 863 * leaking the IDs of other contexts. 1401 * 0 - (NUM_CONTEXT_TAG -1) are used by other contexts 2658 * configuration across all contexts. 2673 * - Ensure that all existing contexts will have the correct per-context 2675 * - Ensure any new contexts wil [all...] |
H A D | i915_drv.h | 338 } contexts; member in struct:drm_i915_private::__anon439
|
H A D | i915_gem.c | 1185 * default HW state for all contexts. 1285 drm_WARN_ON(&dev_priv->drm, !list_empty(&dev_priv->gem.contexts.list));
|
H A D | i915_sysfs.c | 79 spin_lock(&i915->gem.contexts.lock); 84 spin_unlock(&i915->gem.contexts.lock); 112 spin_lock(&i915->gem.contexts.lock); 125 list_for_each_entry(ctx, &i915->gem.contexts.list, link) 128 spin_unlock(&i915->gem.contexts.lock);
|
/linux-master/drivers/net/wireless/intel/iwlwifi/mvm/ |
H A D | ops.c | 310 * The handler can be one from three contexts, see &iwl_rx_handler_context 1564 * This function receives a bitmap of rx async handler contexts 1568 u8 contexts) 1575 * wanted contexts from this list, add them to a local one (lock free), 1580 if (!(BIT(entry->context) & contexts)) 1604 u8 contexts = BIT(RX_HANDLER_ASYNC_LOCKED_WIPHY); local 1606 iwl_mvm_async_handlers_by_context(mvm, contexts); 1613 u8 contexts = BIT(RX_HANDLER_ASYNC_LOCKED) | local 1616 iwl_mvm_async_handlers_by_context(mvm, contexts); 1567 iwl_mvm_async_handlers_by_context(struct iwl_mvm *mvm, u8 contexts) argument
|
/linux-master/drivers/net/wireless/intel/iwlwifi/dvm/ |
H A D | mac80211.c | 339 if (priv->contexts[IWL_RXON_CTX_BSS].vif != vif) 359 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; 423 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; 522 if (resume_data.valid && priv->contexts[IWL_RXON_CTX_BSS].vif) { 944 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS]; 1018 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
|
/linux-master/drivers/gpu/drm/tegra/ |
H A D | drm.c | 111 xa_init_flags(&fpriv->contexts, XA_FLAGS_ALLOC1);
|
/linux-master/tools/include/uapi/drm/ |
H A D | drm.h | 434 struct drm_ctx __user *contexts; member in struct:drm_ctx_res
|
/linux-master/include/uapi/drm/ |
H A D | drm.h | 434 struct drm_ctx __user *contexts; member in struct:drm_ctx_res
|
/linux-master/drivers/staging/media/omap4iss/ |
H A D | iss_csi2.c | 254 struct iss_csi2_ctx_cfg *ctx = &csi2->contexts[0]; 283 struct iss_csi2_ctx_cfg *ctx = &csi2->contexts[ctxnum]; 538 if (csi2->contexts[0].enabled || csi2->ctrl.if_enable) 568 csi2->contexts[0].format_id = csi2_ctx_map_format(csi2); 571 csi2->contexts[0].data_offset = 0; 573 csi2->contexts[0].data_offset = csi2->video_out.bpl_value; 582 csi2->contexts[0].eof_enabled = 1; 583 csi2->contexts[0].eol_enabled = 1; 592 csi2_ctx_config(csi2, &csi2->contexts[0]); 780 csi2_isr_ctx(csi2, &csi2->contexts[ [all...] |
/linux-master/drivers/media/platform/ti/omap3isp/ |
H A D | ispcsi2.c | 220 struct isp_csi2_ctx_cfg *ctx = &csi2->contexts[0]; 250 struct isp_csi2_ctx_cfg *ctx = &csi2->contexts[ctxnum]; 561 if (csi2->contexts[0].enabled || csi2->ctrl.if_enable) 599 csi2->contexts[0].format_id = csi2_ctx_map_format(csi2); 602 csi2->contexts[0].data_offset = 0; 604 csi2->contexts[0].data_offset = csi2->video_out.bpl_value; 613 csi2->contexts[0].eof_enabled = 1; 614 csi2->contexts[0].eol_enabled = 1; 623 csi2_ctx_config(isp, csi2, &csi2->contexts[0]); 785 csi2_isr_ctx(csi2, &csi2->contexts[ [all...] |
/linux-master/drivers/gpu/drm/i915/pxp/ |
H A D | intel_pxp.c | 41 * session are marked as invalid and all contexts marked as using protected 495 /* ban all contexts marked as protected */ 496 spin_lock_irq(&i915->gem.contexts.lock); 497 list_for_each_entry_safe(ctx, cn, &i915->gem.contexts.list, link) { 509 spin_unlock_irq(&i915->gem.contexts.lock); 535 spin_lock_irq(&i915->gem.contexts.lock); 539 spin_unlock_irq(&i915->gem.contexts.lock);
|
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | mock_context.c | 74 init_contexts(&i915->gem.contexts);
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_context.c | 13 * supports contexts for the render ring. 15 * In software, there is a distinction between contexts created by the user, 24 * swap out, and/or destroy other contexts. 26 * All other contexts are created as a request by the GPU client. These contexts 217 * Only contexts that are short-lived [that will expire or be 652 /* Create contexts / engines */ 1278 spin_lock(&ctx->i915->gem.contexts.lock); 1280 spin_unlock(&ctx->i915->gem.contexts.lock); 1556 * Only contexts tha [all...] |
/linux-master/include/linux/ |
H A D | perf_event.h | 76 short contexts; member in struct:perf_callchain_entry_ctx 949 * These fields let us detect when two contexts have both 971 * Number of contexts where an event can trigger: 1571 if (ctx->contexts < sysctl_perf_event_max_contexts_per_stack) { 1574 ++ctx->contexts;
|