/linux-master/drivers/media/platform/samsung/s5p-mfc/ |
H A D | s5p_mfc_intr.c | 51 int s5p_mfc_wait_for_done_ctx(struct s5p_mfc_ctx *ctx, argument 57 ret = wait_event_interruptible_timeout(ctx->queue, 58 (ctx->int_cond && (ctx->int_type == command 59 || ctx->int_type == S5P_MFC_R2H_CMD_ERR_RET)), 62 ret = wait_event_timeout(ctx->queue, 63 (ctx->int_cond && (ctx->int_type == command 64 || ctx->int_type == S5P_MFC_R2H_CMD_ERR_RET)), 68 mfc_err("Interrupt (ctx 82 s5p_mfc_clean_ctx_int_flags(struct s5p_mfc_ctx *ctx) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_context.h | 22 static inline bool i915_gem_context_is_closed(const struct i915_gem_context *ctx) argument 24 return test_bit(CONTEXT_CLOSED, &ctx->flags); 27 static inline void i915_gem_context_set_closed(struct i915_gem_context *ctx) argument 29 GEM_BUG_ON(i915_gem_context_is_closed(ctx)); 30 set_bit(CONTEXT_CLOSED, &ctx->flags); 33 static inline bool i915_gem_context_no_error_capture(const struct i915_gem_context *ctx) argument 35 return test_bit(UCONTEXT_NO_ERROR_CAPTURE, &ctx->user_flags); 38 static inline void i915_gem_context_set_no_error_capture(struct i915_gem_context *ctx) argument 40 set_bit(UCONTEXT_NO_ERROR_CAPTURE, &ctx->user_flags); 43 static inline void i915_gem_context_clear_no_error_capture(struct i915_gem_context *ctx) argument 48 i915_gem_context_is_bannable(const struct i915_gem_context *ctx) argument 53 i915_gem_context_set_bannable(struct i915_gem_context *ctx) argument 58 i915_gem_context_clear_bannable(struct i915_gem_context *ctx) argument 63 i915_gem_context_is_recoverable(const struct i915_gem_context *ctx) argument 68 i915_gem_context_set_recoverable(struct i915_gem_context *ctx) argument 73 i915_gem_context_clear_recoverable(struct i915_gem_context *ctx) argument 78 i915_gem_context_is_persistent(const struct i915_gem_context *ctx) argument 83 i915_gem_context_set_persistence(struct i915_gem_context *ctx) argument 88 i915_gem_context_clear_persistence(struct i915_gem_context *ctx) argument 94 i915_gem_context_user_engines(const struct i915_gem_context *ctx) argument 100 i915_gem_context_set_user_engines(struct i915_gem_context *ctx) argument 106 i915_gem_context_clear_user_engines(struct i915_gem_context *ctx) argument 112 i915_gem_context_uses_protected_content(const struct i915_gem_context *ctx) argument 146 i915_gem_context_get(struct i915_gem_context *ctx) argument 152 i915_gem_context_put(struct i915_gem_context *ctx) argument 158 i915_gem_context_vm(struct i915_gem_context *ctx) argument 163 i915_gem_context_has_full_ppgtt(struct i915_gem_context *ctx) argument 171 i915_gem_context_get_eb_vm(struct i915_gem_context *ctx) argument 184 i915_gem_context_engines(struct i915_gem_context *ctx) argument 206 i915_gem_context_get_engine(struct i915_gem_context *ctx, unsigned int idx) argument [all...] |
/linux-master/drivers/media/platform/mediatek/vcodec/common/ |
H A D | mtk_vcodec_intr.c | 25 struct mtk_vcodec_dec_ctx *ctx; local 27 ctx = priv; 28 ctx_id = ctx->id; 29 ctx_type = ctx->type; 30 ctx_int_cond = ctx->int_cond; 31 ctx_int_type = ctx->int_type; 32 ctx_queue = ctx->queue; 33 pdev = ctx->dev->plat_dev; 35 struct mtk_vcodec_enc_ctx *ctx; local 37 ctx [all...] |
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | missed_tp_recursion.c | 19 int test1(struct pt_regs *ctx) argument 26 int test2(struct pt_regs *ctx) argument 32 int test3(struct pt_regs *ctx) argument 38 int test4(struct pt_regs *ctx) argument
|
H A D | priv_prog.c | 10 int kprobe_prog(void *ctx) argument
|
H A D | dev_cgroup.c | 13 int bpf_prog1(struct bpf_cgroup_dev_ctx *ctx) argument 15 short type = ctx->access_type & 0xFFFF; 17 short access = ctx->access_type >> 16; 41 bpf_trace_printk(fmt, sizeof(fmt), ctx->major, ctx->minor); 47 if (ctx->major != 1 || type != BPF_DEVCG_DEV_CHAR) 50 switch (ctx->minor) {
|
H A D | test_send_signal_kern.c | 9 static __always_inline int bpf_send_signal_test(void *ctx) argument 29 int send_signal_tp(void *ctx) argument 31 return bpf_send_signal_test(ctx); 35 int send_signal_tp_sched(void *ctx) argument 37 return bpf_send_signal_test(ctx); 41 int send_signal_perf(void *ctx) argument 43 return bpf_send_signal_test(ctx);
|
H A D | recvmsg4_prog.c | 17 int recvmsg4_prog(struct bpf_sock_addr *ctx) argument 21 sk = ctx->sk; 28 if (ctx->type != SOCK_STREAM && ctx->type != SOCK_DGRAM) 31 if (!get_set_sk_priority(ctx)) 34 ctx->user_ip4 = bpf_htonl(SERV4_IP); 35 ctx->user_port = bpf_htons(SERV4_PORT);
|
H A D | recvmsg6_prog.c | 20 int recvmsg6_prog(struct bpf_sock_addr *ctx) argument 24 sk = ctx->sk; 31 if (ctx->type != SOCK_STREAM && ctx->type != SOCK_DGRAM) 34 if (!get_set_sk_priority(ctx)) 37 ctx->user_ip6[0] = bpf_htonl(SERV6_IP_0); 38 ctx->user_ip6[1] = bpf_htonl(SERV6_IP_1); 39 ctx->user_ip6[2] = bpf_htonl(SERV6_IP_2); 40 ctx->user_ip6[3] = bpf_htonl(SERV6_IP_3); 41 ctx [all...] |
H A D | sendmsg4_prog.c | 22 int sendmsg_v4_prog(struct bpf_sock_addr *ctx) argument 24 if (ctx->type != SOCK_DGRAM) 27 if (!get_set_sk_priority(ctx)) 31 if (ctx->msg_src_ip4 == bpf_htonl(SRC1_IP4) || 32 ctx->msg_src_ip4 == bpf_htonl(SRC2_IP4)) { 33 ctx->msg_src_ip4 = bpf_htonl(SRC_REWRITE_IP4); 40 if ((ctx->user_ip4 >> 24) == (bpf_htonl(DST_IP4) >> 24) && 41 ctx->user_port == bpf_htons(DST_PORT)) { 42 ctx->user_ip4 = bpf_htonl(DST_REWRITE_IP4); 43 ctx [all...] |
H A D | connect4_prog.c | 36 int do_bind(struct bpf_sock_addr *ctx) argument 44 if (bpf_bind(ctx, (struct sockaddr *)&sa, sizeof(sa)) != 0) 50 static __inline int verify_cc(struct bpf_sock_addr *ctx, argument 56 if (bpf_getsockopt(ctx, SOL_TCP, TCP_CONGESTION, &buf, sizeof(buf))) 69 static __inline int set_cc(struct bpf_sock_addr *ctx) argument 74 if (bpf_setsockopt(ctx, SOL_TCP, TCP_CONGESTION, &reno, sizeof(reno))) 76 if (verify_cc(ctx, reno)) 79 if (bpf_setsockopt(ctx, SOL_TCP, TCP_CONGESTION, &cubic, sizeof(cubic))) 81 if (verify_cc(ctx, cubic)) 87 static __inline int bind_to_device(struct bpf_sock_addr *ctx) argument 110 set_keepalive(struct bpf_sock_addr *ctx) argument 134 set_notsent_lowat(struct bpf_sock_addr *ctx) argument 147 connect_v4_prog(struct bpf_sock_addr *ctx) argument [all...] |
/linux-master/drivers/gpu/drm/amd/display/dmub/src/ |
H A D | dmub_dcn351.h | 11 void dmub_srv_dcn351_regs_init(struct dmub_srv *dmub, struct dc_context *ctx);
|
/linux-master/drivers/media/platform/chips-media/coda/ |
H A D | coda-bit.c | 41 static void coda_free_bitstream_buffer(struct coda_ctx *ctx); 64 static void coda_command_async(struct coda_ctx *ctx, int cmd) argument 66 struct coda_dev *dev = ctx->dev; 72 coda_write(dev, ctx->bit_stream_param, 74 coda_write(dev, ctx->frm_dis_flg, 75 CODA_REG_BIT_FRM_DIS_FLG(ctx->reg_idx)); 76 coda_write(dev, ctx->frame_mem_ctrl, 78 coda_write(dev, ctx->workbuf.paddr, CODA_REG_BIT_WORK_BUF_ADDR); 88 coda_write(dev, ctx->idx, CODA_REG_BIT_RUN_INDEX); 89 coda_write(dev, ctx 97 coda_command_sync(struct coda_ctx *ctx, int cmd) argument 111 coda_hw_reset(struct coda_ctx *ctx) argument 149 coda_kfifo_sync_from_device(struct coda_ctx *ctx) argument 162 coda_kfifo_sync_to_device_full(struct coda_ctx *ctx) argument 174 coda_kfifo_sync_to_device_write(struct coda_ctx *ctx) argument 184 coda_h264_bitstream_pad(struct coda_ctx *ctx, u32 size) argument 203 coda_bitstream_flush(struct coda_ctx *ctx) argument 223 coda_bitstream_queue(struct coda_ctx *ctx, const u8 *buf, u32 size) argument 230 coda_buffer_parse_headers(struct coda_ctx *ctx, struct vb2_v4l2_buffer *src_buf, u32 payload) argument 251 coda_bitstream_try_queue(struct coda_ctx *ctx, struct vb2_v4l2_buffer *src_buf) argument 323 coda_fill_bitstream(struct coda_ctx *ctx, struct list_head *buffer_list) argument 435 coda_bit_stream_end_flag(struct coda_ctx *ctx) argument 450 coda_parabuf_write(struct coda_ctx *ctx, int index, u32 value) argument 461 coda_alloc_context_buf(struct coda_ctx *ctx, struct coda_aux_buf *buf, size_t size, const char *name) argument 469 coda_free_framebuffers(struct coda_ctx *ctx) argument 477 coda_alloc_framebuffers(struct coda_ctx *ctx, struct coda_q_data *q_data, u32 fourcc) argument 558 coda_free_context_buffers(struct coda_ctx *ctx) argument 569 coda_alloc_context_buffers(struct coda_ctx *ctx, struct coda_q_data *q_data) argument 622 coda_encode_header(struct coda_ctx *ctx, struct vb2_v4l2_buffer *buf, int header_code, u8 *header, int *size) argument 680 coda_slice_mode(struct coda_ctx *ctx) argument 703 coda_enc_param_change(struct coda_ctx *ctx) argument 791 coda_setup_iram(struct coda_ctx *ctx) argument 1010 coda9_set_frame_cache(struct coda_ctx *ctx, u32 fourcc) argument 1040 coda_encoder_reqbufs(struct coda_ctx *ctx, struct v4l2_requestbuffers *rb) argument 1061 coda_start_encoding(struct coda_ctx *ctx) argument 1478 coda_prepare_encode(struct coda_ctx *ctx) argument 1639 coda_finish_encode(struct coda_ctx *ctx) argument 1707 struct coda_ctx *ctx = container_of(work, struct coda_ctx, seq_end_work); local 1742 coda_bit_release(struct coda_ctx *ctx) argument 1765 coda_alloc_bitstream_buffer(struct coda_ctx *ctx, struct coda_q_data *q_data) argument 1785 coda_free_bitstream_buffer(struct coda_ctx *ctx) argument 1796 coda_decoder_reqbufs(struct coda_ctx *ctx, struct v4l2_requestbuffers *rb) argument 1823 coda_reorder_enable(struct coda_ctx *ctx) argument 1848 coda_decoder_drop_used_metas(struct coda_ctx *ctx) argument 1871 __coda_decoder_seq_init(struct coda_ctx *ctx) argument 2043 struct coda_ctx *ctx = container_of(work, local 2057 __coda_start_decoding(struct coda_ctx *ctx) argument 2152 coda_start_decoding(struct coda_ctx *ctx) argument 2164 coda_prepare_decode(struct coda_ctx *ctx) argument 2301 coda_finish_decode(struct coda_ctx *ctx) argument 2603 coda_decode_timeout(struct coda_ctx *ctx) argument 2637 struct coda_ctx *ctx; local [all...] |
/linux-master/drivers/ras/amd/atl/ |
H A D | dehash.c | 22 static inline bool map_bits_valid(struct addr_ctx *ctx, u8 bit1, u8 bit2, argument 25 if (!(ctx->map.intlv_bit_pos == bit1 || ctx->map.intlv_bit_pos == bit2)) { 26 pr_debug("Invalid interleave bit: %u", ctx->map.intlv_bit_pos); 30 if (ctx->map.num_intlv_dies > num_intlv_dies) { 31 pr_debug("Invalid number of interleave dies: %u", ctx->map.num_intlv_dies); 35 if (ctx->map.num_intlv_sockets > num_intlv_sockets) { 36 pr_debug("Invalid number of interleave sockets: %u", ctx->map.num_intlv_sockets); 43 static int df2_dehash_addr(struct addr_ctx *ctx) argument 47 if (!map_bits_valid(ctx, 65 df3_dehash_addr(struct addr_ctx *ctx) argument 120 df3_6chan_dehash_addr(struct addr_ctx *ctx) argument 169 df4_dehash_addr(struct addr_ctx *ctx) argument 244 df4p5_dehash_addr(struct addr_ctx *ctx) argument 357 mi300_dehash_addr(struct addr_ctx *ctx) argument 434 dehash_address(struct addr_ctx *ctx) argument [all...] |
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_modeset_lock.c | 11 void _intel_modeset_lock_begin(struct drm_modeset_acquire_ctx *ctx, argument 15 drm_modeset_acquire_init(ctx, flags); 18 state->base.acquire_ctx = ctx; 33 void _intel_modeset_lock_end(struct drm_modeset_acquire_ctx *ctx, argument 41 *ret = drm_modeset_backoff(ctx); 48 drm_modeset_drop_locks(ctx); 49 drm_modeset_acquire_fini(ctx);
|
H A D | intel_modeset_lock.h | 14 void _intel_modeset_lock_begin(struct drm_modeset_acquire_ctx *ctx, 19 void _intel_modeset_lock_end(struct drm_modeset_acquire_ctx *ctx, 28 #define intel_modeset_lock_ctx_retry(ctx, state, flags, ret) \ 29 for (_intel_modeset_lock_begin((ctx), (state), (flags), &(ret)); \ 31 _intel_modeset_lock_end((ctx), (state), &(ret)))
|
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | map_in_map.c | 17 static int wait_for_start_or_abort(struct thread_ctx *ctx) argument 19 while (!ctx->start && !ctx->abort) 21 return ctx->abort ? -1 : 0; 26 struct thread_ctx *ctx = data; local 27 int loop = ctx->loop, err = 0; 29 if (wait_for_start_or_abort(ctx) < 0) 31 pthread_barrier_wait(&ctx->barrier); 39 pthread_barrier_wait(&ctx->barrier); 44 if (bpf_map_update_elem(ctx 57 struct thread_ctx *ctx = data; local 78 struct thread_ctx ctx; local [all...] |
/linux-master/tools/testing/selftests/filesystems/epoll/ |
H A D | epoll_wakeup_test.c | 43 static void kill_timeout(struct epoll_mtcontext *ctx) argument 46 pthread_kill(ctx->main, SIGUSR1); 47 pthread_kill(ctx->waiter, SIGUSR1); 53 struct epoll_mtcontext *ctx = data; local 55 if (epoll_wait(ctx->efd[0], &e, 1, -1) > 0) 56 __sync_fetch_and_add(&ctx->count, 1); 65 struct epoll_mtcontext *ctx = data; local 67 pfd.fd = ctx->efd[0]; 70 if (epoll_wait(ctx->efd[0], &e, 1, 0) > 0) 71 __sync_fetch_and_add(&ctx 80 struct epoll_mtcontext *ctx = data; local 92 struct epoll_mtcontext *ctx = data; local 107 struct epoll_mtcontext *ctx = data; local 119 struct epoll_mtcontext *ctx = data; local 133 struct epoll_mtcontext *ctx = data; local 145 struct epoll_mtcontext *ctx = data; local 476 struct epoll_mtcontext ctx = { 0 }; local 519 struct epoll_mtcontext ctx = { 0 }; local 562 struct epoll_mtcontext ctx = { 0 }; local 611 struct epoll_mtcontext ctx = { 0 }; local 660 struct epoll_mtcontext ctx = { 0 }; local 703 struct epoll_mtcontext ctx = { 0 }; local 746 struct epoll_mtcontext ctx = { 0 }; local 795 struct epoll_mtcontext ctx = { 0 }; local 1198 struct epoll_mtcontext ctx = { 0 }; local 1250 struct epoll_mtcontext ctx = { 0 }; local 1302 struct epoll_mtcontext ctx = { 0 }; local 1354 struct epoll_mtcontext ctx = { 0 }; local 1406 struct epoll_mtcontext ctx = { 0 }; local 1457 struct epoll_mtcontext ctx = { 0 }; local 1508 struct epoll_mtcontext ctx = { 0 }; local 1559 struct epoll_mtcontext ctx = { 0 }; local 1610 struct epoll_mtcontext ctx = { 0 }; local 1662 struct epoll_mtcontext ctx = { 0 }; local 1714 struct epoll_mtcontext ctx = { 0 }; local 1766 struct epoll_mtcontext ctx = { 0 }; local 1819 struct epoll_mtcontext ctx = { 0 }; local 1876 struct epoll_mtcontext ctx = { 0 }; local 1933 struct epoll_mtcontext ctx = { 0 }; local 1990 struct epoll_mtcontext ctx = { 0 }; local 2046 struct epoll_mtcontext ctx = { 0 }; local 2098 struct epoll_mtcontext ctx = { 0 }; local 2150 struct epoll_mtcontext ctx = { 0 }; local 2202 struct epoll_mtcontext ctx = { 0 }; local 2255 struct epoll_mtcontext ctx = { 0 }; local 2311 struct epoll_mtcontext ctx = { 0 }; local 2364 struct epoll_mtcontext ctx = { 0 }; local 2420 struct epoll_mtcontext ctx = { 0 }; local 2704 struct epoll_mtcontext ctx = { 0 }; local 2769 struct epoll_mtcontext ctx = { 0 }; local 2834 struct epoll_mtcontext ctx = { 0 }; local 2899 struct epoll_mtcontext ctx = { 0 }; local 2965 struct epoll_mtcontext ctx = { 0 }; local 3035 struct epoll_mtcontext ctx = { 0 }; local 3093 struct epoll_mtcontext *ctx = ctx_; local 3123 struct epoll_mtcontext ctx = { 0 }; local 3169 struct epoll60_ctx *ctx = ctx_; local 3224 count_waiters(struct epoll60_ctx *ctx) argument 3231 struct epoll60_ctx ctx = { 0 }; local 3307 struct epoll61_ctx *ctx = ctx_; local 3317 struct epoll61_ctx *ctx = ctx_; local 3336 struct epoll61_ctx *ctx = ctx_; local 3345 struct epoll61_ctx ctx; local 3463 struct epoll_mtcontext ctx = { 0 }; local [all...] |
/linux-master/sound/soc/intel/atom/sst/ |
H A D | sst_acpi.c | 148 static int sst_platform_get_resources(struct intel_sst_drv *ctx) argument 151 struct platform_device *pdev = to_platform_device(ctx->dev); 156 ctx->pdata->res_info->acpi_lpe_res_index); 158 dev_err(ctx->dev, "Invalid SHIM base from IFWI\n"); 161 dev_info(ctx->dev, "LPE base: %#x size:%#x", (unsigned int) rsrc->start, 164 ctx->iram_base = rsrc->start + ctx->pdata->res_info->iram_offset; 165 ctx->iram_end = ctx->iram_base + ctx 238 struct intel_sst_drv *ctx; local 339 struct intel_sst_drv *ctx; local [all...] |
/linux-master/arch/mips/net/ |
H A D | bpf_jit_comp64.c | 98 static void emit_sext(struct jit_context *ctx, u8 dst, u8 src) argument 100 emit(ctx, sll, dst, src, 0); 101 clobber_reg(ctx, dst); 105 static void emit_zext(struct jit_context *ctx, u8 dst) argument 108 emit(ctx, dinsu, dst, MIPS_R_ZERO, 32, 32); 110 emit(ctx, and, dst, dst, bpf2mips64[JIT_REG_ZX]); 111 access_reg(ctx, JIT_REG_ZX); /* We need the ZX register */ 113 clobber_reg(ctx, dst); 117 static void emit_zext_ver(struct jit_context *ctx, u8 dst) argument 119 if (!ctx 124 emit_mov_i64(struct jit_context *ctx, u8 dst, u64 imm64) argument 158 emit_alu_i64(struct jit_context *ctx, u8 dst, s32 imm, u8 op) argument 201 emit_alu_r64(struct jit_context *ctx, u8 dst, u8 src, u8 op) argument 262 emit_swap_r64(struct jit_context *ctx, u8 dst, u8 mask, u32 bits) argument 274 emit_bswap_r64(struct jit_context *ctx, u8 dst, u32 width) argument 316 emit_trunc_r64(struct jit_context *ctx, u8 dst, u32 width) argument 334 emit_ldx(struct jit_context *ctx, u8 dst, u8 src, s16 off, u8 size) argument 358 emit_stx(struct jit_context *ctx, u8 dst, u8 src, s16 off, u8 size) argument 381 emit_atomic_r64(struct jit_context *ctx, u8 dst, u8 src, s16 off, u8 code) argument 421 emit_cmpxchg_r64(struct jit_context *ctx, u8 dst, u8 src, s16 off) argument 439 emit_call(struct jit_context *ctx, const struct bpf_insn *insn) argument 477 emit_tail_call(struct jit_context *ctx) argument 546 build_prologue(struct jit_context *ctx) argument 619 build_epilogue(struct jit_context *ctx, int dest_reg) argument 635 build_insn(const struct bpf_insn *insn, struct jit_context *ctx) argument [all...] |
/linux-master/samples/bpf/ |
H A D | xdp2skb_meta_kern.c | 32 int _xdp_mark(struct xdp_md *ctx) argument 41 ret = bpf_xdp_adjust_meta(ctx, -(int)sizeof(*meta)); 46 * ctx->data MUST happen _after_ helper bpf_xdp_adjust_meta(), 50 data = (void *)(unsigned long)ctx->data; 53 meta = (void *)(unsigned long)ctx->data_meta; 63 int _tc_mark(struct __sk_buff *ctx) argument 65 void *data = (void *)(unsigned long)ctx->data; 66 void *data_end = (void *)(unsigned long)ctx->data_end; 67 void *data_meta = (void *)(unsigned long)ctx->data_meta; 72 ctx [all...] |
/linux-master/drivers/gpu/drm/radeon/ |
H A D | atom-bits.h | 32 #define U8(ptr) get_u8(ctx->ctx->bios, (ptr)) 33 #define CU8(ptr) get_u8(ctx->bios, (ptr)) 38 #define U16(ptr) get_u16(ctx->ctx->bios, (ptr)) 39 #define CU16(ptr) get_u16(ctx->bios, (ptr)) 44 #define U32(ptr) get_u32(ctx->ctx->bios, (ptr)) 45 #define CU32(ptr) get_u32(ctx->bios, (ptr)) 46 #define CSTR(ptr) (((char *)(ctx [all...] |
/linux-master/drivers/gpu/drm/amd/include/ |
H A D | atom-bits.h | 32 #define U8(ptr) get_u8(ctx->ctx->bios, (ptr)) 33 #define CU8(ptr) get_u8(ctx->bios, (ptr)) 38 #define U16(ptr) get_u16(ctx->ctx->bios, (ptr)) 39 #define CU16(ptr) get_u16(ctx->bios, (ptr)) 44 #define U32(ptr) get_u32(ctx->ctx->bios, (ptr)) 45 #define CU32(ptr) get_u32(ctx->bios, (ptr)) 46 #define CSTR(ptr) (((char *)(ctx [all...] |
/linux-master/drivers/gpu/drm/amd/display/dc/dml2/ |
H A D | dml2_mall_phantom.h | 42 unsigned int dml2_helper_calculate_num_ways_for_subvp(struct dml2_context *ctx, struct dc_state *context); 44 bool dml2_svp_add_phantom_pipe_to_dc_state(struct dml2_context *ctx, struct dc_state *state, struct dml_mode_support_info_st *mode_support_info); 46 bool dml2_svp_remove_all_phantom_pipes(struct dml2_context *ctx, struct dc_state *state); 48 bool dml2_svp_validate_static_schedulability(struct dml2_context *ctx, struct dc_state *context, enum dml_dram_clock_change_support pstate_change_type); 50 bool dml2_svp_drr_schedulable(struct dml2_context *ctx, struct dc_state *context, struct dc_crtc_timing *drr_timing);
|
/linux-master/drivers/gpu/drm/panel/ |
H A D | panel-visionox-rm69299.c | 30 static int visionox_rm69299_power_on(struct visionox_rm69299 *ctx) argument 34 ret = regulator_bulk_enable(ARRAY_SIZE(ctx->supplies), ctx->supplies); 43 gpiod_set_value(ctx->reset_gpio, 1); 45 gpiod_set_value(ctx->reset_gpio, 0); 47 gpiod_set_value(ctx->reset_gpio, 1); 53 static int visionox_rm69299_power_off(struct visionox_rm69299 *ctx) argument 55 gpiod_set_value(ctx->reset_gpio, 0); 57 return regulator_bulk_disable(ARRAY_SIZE(ctx->supplies), ctx 62 struct visionox_rm69299 *ctx = panel_to_ctx(panel); local 86 struct visionox_rm69299 *ctx = panel_to_ctx(panel); local 160 struct visionox_rm69299 *ctx = panel_to_ctx(panel); local 187 struct visionox_rm69299 *ctx; local 253 struct visionox_rm69299 *ctx = mipi_dsi_get_drvdata(dsi); local [all...] |