/linux-master/arch/powerpc/include/asm/ |
H A D | mmu_context.h | 15 * Most if the context management is out of line 83 VM_WARN_ON(mm->context.extended_id[index]); 84 mm->context.extended_id[index] = context_id; 92 context_id = get_user_context(&mm->context, ea); 125 atomic_inc(&mm->context.active_cpus); 130 VM_WARN_ON_ONCE(atomic_read(&mm->context.active_cpus) <= 0); 131 atomic_dec(&mm->context.active_cpus); 141 if (atomic_inc_return(&mm->context.copros) == 1) 158 * same context, it can lead to over-flushing. It's hopefully 170 c = atomic_dec_if_positive(&mm->context [all...] |
/linux-master/include/linux/usb/ |
H A D | ljca.h | 23 * @context: the execution context of who registered this callback 28 * The callback function is called in interrupt context and the data payload is 32 typedef void (*ljca_event_cb_t)(void *context, u8 cmd, const void *evt_data, int len); 42 * @context: the execution context of the event callback 54 void *context; member in struct:ljca_client 100 * @context: execution context of event callback 104 int ljca_register_event_cb(struct ljca_client *client, ljca_event_cb_t event_cb, void *context); [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/hwss/dcn10/ |
H A D | dcn10_hwseq.h | 45 struct dc_state *context, 49 struct dc_state *context); 52 struct dc_state *context); 77 struct dc_state *context); 81 struct dc_state *context, 85 struct dc_state *context); 103 struct dc_state *context); 106 void dcn10_init_pipes(struct dc *dc, struct dc_state *context); 110 struct dc_state *context); 117 void dce110_enable_accelerated_mode(struct dc *dc, struct dc_state *context); [all...] |
/linux-master/arch/arm64/kernel/ |
H A D | suspend.c | 102 struct arm_cpuidle_irq_context context; local 138 arm_cpuidle_save_irq_context(&context); 162 arm_cpuidle_restore_irq_context(&context);
|
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_provider.c | 291 struct mthca_ucontext *context = to_mucontext(uctx); local 303 err = mthca_uar_alloc(to_mdev(ibdev), &context->uar); 307 context->db_tab = mthca_init_user_db_tab(to_mdev(ibdev)); 308 if (IS_ERR(context->db_tab)) { 309 err = PTR_ERR(context->db_tab); 310 mthca_uar_free(to_mdev(ibdev), &context->uar); 315 mthca_cleanup_user_db_tab(to_mdev(ibdev), &context->uar, context->db_tab); 316 mthca_uar_free(to_mdev(ibdev), &context->uar); 320 context 325 mthca_dealloc_ucontext(struct ib_ucontext *context) argument 332 mthca_mmap_uar(struct ib_ucontext *context, struct vm_area_struct *vma) argument 396 struct mthca_ucontext *context = rdma_udata_to_drv_context( local 440 struct mthca_ucontext *context = local 458 struct mthca_ucontext *context = rdma_udata_to_drv_context( local 555 struct mthca_ucontext *context = local 585 struct mthca_ucontext *context = rdma_udata_to_drv_context( local 775 struct mthca_ucontext *context = local 831 struct mthca_ucontext *context = rdma_udata_to_drv_context( local [all...] |
/linux-master/arch/sparc/kernel/ |
H A D | adi_64.c | 135 if (mm->context.tag_store) { 136 tag_desc = mm->context.tag_store; 137 spin_lock_irqsave(&mm->context.tag_lock, flags); 144 spin_unlock_irqrestore(&mm->context.tag_lock, flags); 174 spin_lock_irqsave(&mm->context.tag_lock, flags); 175 if (mm->context.tag_store) { 176 tag_desc = mm->context.tag_store; 205 mm->context.tag_store = kzalloc(size, GFP_NOWAIT|__GFP_NOWARN); 206 if (mm->context.tag_store == NULL) { 210 tag_desc = mm->context [all...] |
/linux-master/drivers/misc/vmw_vmci/ |
H A D | vmci_host.c | 85 struct vmci_ctx *context; member in struct:vmci_host_dev 88 struct mutex lock; /* Mutex lock for vmci context access */ 143 vmci_ctx_destroy(vmci_host_dev->context); 144 vmci_host_dev->context = NULL; 149 * a context is created through the IOCTL_VMCI_INIT_CONTEXT 168 struct vmci_ctx *context; local 173 * Read context only if ct_type == VMCIOBJ_CONTEXT to make 174 * sure that context is initialized 176 context = vmci_host_dev->context; 227 vmci_host_setup_notify(struct vmci_ctx *context, unsigned long uva) argument [all...] |
/linux-master/arch/arm64/mm/ |
H A D | context.c | 3 * Based on arch/arm/mm/context.c 129 * context-switch 161 u64 asid = atomic64_read(&mm->context.id); 179 if (refcount_read(&mm->context.pinned)) 224 asid = atomic64_read(&mm->context.id); 248 asid = atomic64_read(&mm->context.id); 251 atomic64_set(&mm->context.id, asid); 283 asid = atomic64_read(&mm->context.id); 285 if (refcount_inc_not_zero(&mm->context.pinned)) 299 atomic64_set(&mm->context [all...] |
/linux-master/drivers/nvmem/ |
H A D | imx-ocotp-scu.c | 75 static bool in_hole(void *context, u32 index) argument 77 struct ocotp_priv *priv = context; 92 static bool in_ecc(void *context, u32 index) argument 94 struct ocotp_priv *priv = context; 132 static int imx_scu_ocotp_read(void *context, unsigned int offset, argument 135 struct ocotp_priv *priv = context; 157 if (in_hole(context, i)) { 180 static int imx_scu_ocotp_write(void *context, unsigned int offset, argument 183 struct ocotp_priv *priv = context; 196 if (in_hole(context, inde [all...] |
H A D | stm32-romem.c | 43 static int stm32_romem_read(void *context, unsigned int offset, void *buf, argument 46 struct stm32_romem_priv *priv = context; 74 static int stm32_bsec_read(void *context, unsigned int offset, void *buf, argument 77 struct stm32_romem_priv *priv = context; 121 static int stm32_bsec_write(void *context, unsigned int offset, void *buf, argument 124 struct stm32_romem_priv *priv = context; 148 static int stm32_bsec_pta_read(void *context, unsigned int offset, void *buf, argument 151 struct stm32_romem_priv *priv = context; 156 static int stm32_bsec_pta_write(void *context, unsigned int offset, void *buf, argument 159 struct stm32_romem_priv *priv = context; [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/dml2/ |
H A D | dml2_mall_phantom.c | 32 unsigned int dml2_helper_calculate_num_ways_for_subvp(struct dml2_context *ctx, struct dc_state *context) argument 50 struct pipe_ctx *pipe = &context->res_ctx.pipe_ctx[i]; 54 ctx->config.svp_pstate.callbacks.get_pipe_subvp_type(context, pipe) == SUBVP_PHANTOM) { 101 static void merge_pipes_for_subvp(struct dml2_context *ctx, struct dc_state *context) argument 107 struct pipe_ctx *pipe = &context->res_ctx.pipe_ctx[i]; 125 ctx->config.svp_pstate.callbacks.release_dsc(&context->res_ctx, ctx->config.svp_pstate.callbacks.dc->res_pool, &pipe->stream_res.dsc); 146 static bool all_pipes_have_stream_and_plane(struct dml2_context *ctx, const struct dc_state *context) argument 151 const struct pipe_ctx *pipe = &context->res_ctx.pipe_ctx[i]; 162 static bool mpo_in_use(const struct dc_state *context) argument 166 for (i = 0; i < context 228 assign_subvp_pipe(struct dml2_context *ctx, struct dc_state *context, unsigned int *index) argument 357 subvp_subvp_schedulable(struct dml2_context *ctx, struct dc_state *context) argument 431 dml2_svp_drr_schedulable(struct dml2_context *ctx, struct dc_state *context, struct dc_crtc_timing *drr_timing) argument 507 subvp_vblank_schedulable(struct dml2_context *ctx, struct dc_state *context) argument 601 dml2_svp_validate_static_schedulability(struct dml2_context *ctx, struct dc_state *context, enum dml_dram_clock_change_support pstate_change_type) argument 810 remove_all_phantom_planes_for_stream(struct dml2_context *ctx, struct dc_stream_state *stream, struct dc_state *context) argument [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/inc/ |
H A D | resource.h | 103 struct dc_state *context, 114 struct dc_state *context); 153 struct dc_state *context, 314 struct dc_state *context, 365 * Get the OTG master pipe in resource context associated with the stream. 403 * Get the OTG master pipe for the input pipe context. 405 * context. 410 * Get the OPP head pipe for the input pipe context. 412 * context. 464 * Look for a free pipe in new resource context tha [all...] |
/linux-master/crypto/asymmetric_keys/ |
H A D | pkcs8_parser.c | 33 int pkcs8_note_OID(void *context, size_t hdrlen, argument 37 struct pkcs8_parse_context *ctx = context; 53 int pkcs8_note_version(void *context, size_t hdrlen, argument 67 int pkcs8_note_algo(void *context, size_t hdrlen, argument 71 struct pkcs8_parse_context *ctx = context; 83 int pkcs8_note_key(void *context, size_t hdrlen, argument 87 struct pkcs8_parse_context *ctx = context;
|
/linux-master/arch/mips/include/asm/ |
H A D | mmu_context.h | 2 * Switch a MMU context. 63 * into the context register. 109 return atomic64_read(&mm->context.mmid); 111 return mm->context.asid[cpu]; 118 atomic64_set(&mm->context.mmid, ctx); 120 mm->context.asid[cpu] = ctx; 132 * Initialize the context related info for a new mm_struct 148 mm->context.bd_emupage_allocmap = NULL; 149 spin_lock_init(&mm->context.bd_emupage_lock); 150 init_waitqueue_head(&mm->context [all...] |
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | slice.c | 91 if ((mm_ctx_slb_addr_limit(&mm->context) - len) < addr) 200 psize_mask = slice_mask_for_size(&mm->context, psize); 207 lpsizes = mm_ctx_low_slices(&mm->context); 217 old_mask = slice_mask_for_size(&mm->context, old_psize); 226 hpsizes = mm_ctx_high_slices(&mm->context); 227 for (i = 0; i < GET_HIGH_SLICE_INDEX(mm_ctx_slb_addr_limit(&mm->context)); i++) { 236 old_mask = slice_mask_for_size(&mm->context, old_psize); 246 (unsigned long)mm_ctx_low_slices(&mm->context), 247 (unsigned long)mm_ctx_high_slices(&mm->context)); 343 addr += mm_ctx_slb_addr_limit(&mm->context) [all...] |
/linux-master/drivers/acpi/acpica/ |
H A D | acevents.h | 106 void *context); 127 u32 level, void *context, void **return_value); 133 acpi_ev_walk_gpe_list(acpi_gpe_callback gpe_walk_callback, void *context); 137 struct acpi_gpe_block_info *gpe_block, void *context); 148 void *context); 168 acpi_adr_space_setup setup, void *context); 243 u32 ACPI_SYSTEM_XFACE acpi_ev_gpe_xrupt_handler(void *context);
|
/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | mmap.c | 32 struct rvt_dev_info *rdi = ib_to_rvt(ip->context->device); 63 * @context: the IB user context of the process making the mmap() call 68 int rvt_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) argument 70 struct rvt_dev_info *rdi = ib_to_rvt(context->device); 85 if (context != ip->context || (__u64)offset != ip->offset) 139 ip->context = 141 ->context;
|
/linux-master/drivers/gpu/drm/amd/display/dc/resource/dcn20/ |
H A D | dcn20_resource.h | 66 void dcn20_release_pipe(struct dc_state *context, 119 struct dc_state *context, 122 bool dcn20_validate_bandwidth(struct dc *dc, struct dc_state *context, bool fast_validate); 125 struct dc_state *context); 128 struct dc_state *context, 156 struct dc_state *context, 163 enum dc_status dcn20_build_mapped_resource(const struct dc *dc, struct dc_state *context, struct dc_stream_state *stream);
|
/linux-master/drivers/md/ |
H A D | dm-ps-round-robin.c | 77 ps->context = s; 83 struct selector *s = ps->context; 88 ps->context = NULL; 124 struct selector *s = ps->context; 149 *error = "round-robin ps: Error allocating path context"; 168 struct selector *s = ps->context; 179 struct selector *s = ps->context; 192 struct selector *s = ps->context;
|
/linux-master/drivers/gpu/drm/amd/display/dc/clk_mgr/dce100/ |
H A D | dce_clk_mgr.c | 162 /* unit: in_khz before mode set, get pixel clock from context. ASIC register 165 uint32_t dce_get_max_pixel_clock_for_all_paths(struct dc_state *context) argument 171 struct pipe_ctx *pipe_ctx = &context->res_ctx.pipe_ctx[i]; 196 struct dc_state *context) 201 int max_pix_clk = dce_get_max_pixel_clock_for_all_paths(context); 208 if (context->bw_ctx.bw.dce.dispclk_khz > 218 < context->bw_ctx.bw.dce.dispclk_khz) 385 struct dc_state *context) 387 struct dm_pp_display_configuration *pp_display_cfg = &context->pp_display_cfg; 389 pp_display_cfg->avail_mclk_switch_time_us = dce110_get_min_vblank_time_us(context); 194 dce_get_required_clocks_state( struct clk_mgr *clk_mgr_base, struct dc_state *context) argument 383 dce_pplib_apply_display_requirements( struct dc *dc, struct dc_state *context) argument 397 dce_update_clocks(struct clk_mgr *clk_mgr_base, struct dc_state *context, bool safe_to_lower) argument [all...] |
/linux-master/include/linux/ |
H A D | damon.h | 85 * Each monitoring context could have multiple targets. For example, a context 387 * monitoring context are inactive, DAMON stops monitoring either, and just 446 * @cleanup: Clean up the context. 484 void (*init)(struct damon_ctx *context); 485 void (*update)(struct damon_ctx *context); 486 void (*prepare_access_checks)(struct damon_ctx *context); 487 unsigned int (*check_accesses)(struct damon_ctx *context); 488 void (*reset_aggregated)(struct damon_ctx *context); 489 int (*get_scheme_score)(struct damon_ctx *context, [all...] |
/linux-master/tools/testing/selftests/damon/ |
H A D | _damon_sysfs.py | 120 context = None variable in class:Damos 135 self.context.sysfs_dir(), 'schemes', '%d' % self.idx) 169 context = None variable in class:DamonTarget 176 self.context.sysfs_dir(), 'targets', '%d' % self.idx) 192 context = None variable in class:DamonAttrs 203 return os.path.join(self.context.sysfs_dir(), 'monitoring_attrs', 207 return os.path.join(self.context.sysfs_dir(), 'monitoring_attrs', 248 self.monitoring_attrs.context = self 253 target.context = self 258 scheme.context [all...] |
/linux-master/drivers/md/dm-vdo/ |
H A D | dedupe.c | 93 * its hash_zone's pool. If one is available, that context is prepared, associated with the 95 * context's state will be transitioned from DEDUPE_CONTEXT_IDLE to DEDUPE_CONTEXT_PENDING. If all 96 * goes well, the dedupe callback will be called by the index which will change the context's state 98 * zone where the query results will be processed and the context will be put back in the idle 109 * When the dedupe callback is run for a context which is in the timed out state, that context is 113 * If a hash_lock needs a dedupe context, and the available list is empty, the timed_out list will 332 static inline bool change_context_state(struct dedupe_context *context, int old, int new) argument 334 return (atomic_cmpxchg(&context->state, old, new) == old); 571 * @context 721 release_context(struct dedupe_context *context) argument 731 struct dedupe_context *context = agent->dedupe_context; local 928 enter_forked_lock(struct vdo_waiter *waiter, void *context) argument 1601 decode_uds_advice(struct dedupe_context *context) argument 1649 struct dedupe_context *context = agent->dedupe_context; local 2135 start_expiration_timer(struct dedupe_context *context) argument 2226 struct dedupe_context *context = container_of(request, struct dedupe_context, local 2275 struct dedupe_context *context; local 2296 struct dedupe_context *context, *tmp; local 2383 struct dedupe_context *context = &zone->contexts[i]; local 2395 get_thread_id_for_zone(void *context, zone_count_t zone_number) argument 2516 suspend_index(void *context, struct vdo_completion *completion) argument 2540 drain_hash_zone(void *context, zone_count_t zone_number, struct vdo_completion *parent) argument 2579 resume_index(void *context, struct vdo_completion *parent) argument 2611 resume_hash_zone(void *context, zone_count_t zone_number, struct vdo_completion *parent) argument 2865 struct dedupe_context *context; local 2909 struct dedupe_context *context; local [all...] |
/linux-master/include/uapi/drm/ |
H A D | tegra_drm.h | 172 * @context: 174 * The application context of this channel. Set by the kernel upon 175 * successful completion of the IOCTL. This context needs to be passed 178 __u64 context; member in struct:drm_tegra_open_channel 186 * @context: 188 * The application context of this channel. This is obtained from the 191 __u64 context; member in struct:drm_tegra_close_channel 199 * @context: 201 * The application context identifying the channel for which to obtain 204 __u64 context; member in struct:drm_tegra_get_syncpt 232 __u64 context; member in struct:drm_tegra_get_syncpt_base 401 __u64 context; member in struct:drm_tegra_submit 691 __u32 context; member in struct:drm_tegra_channel_open 716 __u32 context; member in struct:drm_tegra_channel_close 737 __u32 context; member in struct:drm_tegra_channel_map 768 __u32 context; member in struct:drm_tegra_channel_unmap 925 __u32 context; member in struct:drm_tegra_channel_submit [all...] |
/linux-master/drivers/acpi/ |
H A D | sbshc.c | 28 void *context; member in struct:acpi_smb_hc 179 smbus_alarm_callback callback, void *context) 183 hc->context = context; 194 hc->context = NULL; 202 static inline void acpi_smbus_callback(void *context) argument 204 struct acpi_smb_hc *hc = context; 206 hc->callback(hc->context); 209 static int smbus_alarm(void *context) argument 211 struct acpi_smb_hc *hc = context; 178 acpi_smbus_register_callback(struct acpi_smb_hc *hc, smbus_alarm_callback callback, void *context) argument [all...] |