/linux-master/fs/btrfs/ |
H A D | extent_io.c | 1839 static int submit_eb_page(struct page *page, struct btrfs_eb_write_context *ctx) argument 1841 struct writeback_control *wbc = ctx->wbc; 1870 if (eb == ctx->eb) { 1879 ctx->eb = eb; 1881 ret = btrfs_check_meta_write_pointer(eb->fs_info, ctx); 1894 if (ctx->zoned_bg) { 1896 btrfs_schedule_zone_finish_bg(ctx->zoned_bg, eb); 1897 ctx->zoned_bg->meta_write_pointer += eb->len; 1907 struct btrfs_eb_write_context ctx = { .wbc = wbc }; local 1949 ret = submit_eb_page(&folio->page, &ctx); [all...] |
/linux-master/drivers/pwm/ |
H A D | pwm-dwc.h | 56 struct dwc_pwm_ctx ctx[DWC_TIMERS_TOTAL]; member in struct:dwc_pwm
|
H A D | pwm-dwc.c | 121 dwc->ctx[i].cnt = dwc_pwm_readl(dwc, DWC_TIM_LD_CNT(i)); 122 dwc->ctx[i].cnt2 = dwc_pwm_readl(dwc, DWC_TIM_LD_CNT2(i)); 123 dwc->ctx[i].ctrl = dwc_pwm_readl(dwc, DWC_TIM_CTRL(i)); 141 dwc_pwm_writel(dwc, dwc->ctx[i].cnt, DWC_TIM_LD_CNT(i)); 142 dwc_pwm_writel(dwc, dwc->ctx[i].cnt2, DWC_TIM_LD_CNT2(i)); 143 dwc_pwm_writel(dwc, dwc->ctx[i].ctrl, DWC_TIM_CTRL(i));
|
/linux-master/include/linux/ |
H A D | virtio.h | 55 void *ctx, 76 void **ctx);
|
H A D | io_uring_types.h | 32 /* ctx state flags, for URING_CMD */ 282 * ->iopoll_list is protected by the ctx->uring_lock for 402 /* ctx exit and cancelation */ 492 /* ctx owns file */ 629 struct io_ring_ctx *ctx; member in struct:io_kiocb
|
/linux-master/drivers/vhost/ |
H A D | vhost.c | 353 call_ctx->ctx = NULL; 980 if (dev->vqs[i]->call_ctx.ctx) 981 eventfd_ctx_put(dev->vqs[i]->call_ctx.ctx); 1880 struct eventfd_ctx *ctx = NULL; local 1953 ctx = f.fd == VHOST_FILE_UNBIND ? NULL : eventfd_ctx_fdget(f.fd); 1954 if (IS_ERR(ctx)) { 1955 r = PTR_ERR(ctx); 1959 swap(ctx, vq->call_ctx.ctx); 1966 ctx 2045 struct eventfd_ctx *ctx; local [all...] |
/linux-master/io_uring/ |
H A D | net.c | 126 if (io_alloc_cache_put(&req->ctx->netmsg_cache, &hdr->cache)) { 135 struct io_ring_ctx *ctx = req->ctx; local 140 entry = io_alloc_cache_get(&ctx->netmsg_cache); 301 if (unlikely(req->ctx->compat)) { 403 if (req->ctx->compat) 586 if (unlikely(req->ctx->compat)) { 671 if (req->ctx->compat) 1040 struct io_ring_ctx *ctx = req->ctx; local [all...] |
H A D | io_uring.c | 143 static bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx, 177 static inline void io_submit_flush_completions(struct io_ring_ctx *ctx) argument 179 if (!wq_list_empty(&ctx->submit_state.compl_reqs) || 180 ctx->submit_state.cqes_count) 181 __io_submit_flush_completions(ctx); 184 static inline unsigned int __io_cqring_events(struct io_ring_ctx *ctx) argument 186 return ctx->cached_cq_tail - READ_ONCE(ctx->rings->cq.head); 189 static inline unsigned int __io_cqring_events_user(struct io_ring_ctx *ctx) argument 191 return READ_ONCE(ctx 220 struct io_ring_ctx *ctx = head->ctx; local 238 io_req_add_to_cache(struct io_kiocb *req, struct io_ring_ctx *ctx) argument 245 struct io_ring_ctx *ctx = container_of(ref, struct io_ring_ctx, refs); local 252 struct io_ring_ctx *ctx = container_of(work, struct io_ring_ctx, local 285 struct io_ring_ctx *ctx; local 359 io_account_cq_overflow(struct io_ring_ctx *ctx) argument 370 struct io_ring_ctx *ctx = req->ctx; local 454 struct io_ring_ctx *ctx = req->ctx; local 489 struct io_ring_ctx *ctx = req->ctx; local 528 io_queue_deferred(struct io_ring_ctx *ctx) argument 560 io_eventfd_signal(struct io_ring_ctx *ctx) argument 597 io_eventfd_flush_signal(struct io_ring_ctx *ctx) argument 620 __io_commit_cqring_flush(struct io_ring_ctx *ctx) argument 635 __io_cq_lock(struct io_ring_ctx *ctx) argument 647 __io_cq_unlock_post(struct io_ring_ctx *ctx) argument 663 io_commit_cqring(ctx); variable 665 io_cqring_wake(ctx); variable 666 io_commit_cqring_flush(ctx); variable 669 io_cqring_overflow_kill(struct io_ring_ctx *ctx) argument 686 __io_cqring_overflow_flush(struct io_ring_ctx *ctx) argument 717 io_cqring_do_overflow_flush(struct io_ring_ctx *ctx) argument 727 io_cqring_overflow_flush(struct io_ring_ctx *ctx) argument 780 io_cqring_event_overflow(struct io_ring_ctx *ctx, u64 user_data, s32 res, u32 cflags, u64 extra1, u64 extra2) argument 833 io_cqe_cache_refill(struct io_ring_ctx *ctx, bool overflow) argument 865 io_fill_cqe_aux(struct io_ring_ctx *ctx, u64 user_data, s32 res, u32 cflags) argument 918 __io_post_aux_cqe(struct io_ring_ctx *ctx, u64 user_data, s32 res, u32 cflags, bool allow_overflow) argument 932 io_post_aux_cqe(struct io_ring_ctx *ctx, u64 user_data, s32 res, u32 cflags) argument 943 struct io_ring_ctx *ctx = req->ctx; local 977 struct io_ring_ctx *ctx = req->ctx; local 1025 struct io_ring_ctx *ctx = req->ctx; local 1058 io_preinit_req(struct io_kiocb *req, struct io_ring_ctx *ctx) argument 1068 io_flush_cached_locked_reqs(struct io_ring_ctx *ctx, struct io_submit_state *state) argument 1136 struct io_ring_ctx *ctx = req->ctx; local 1160 ctx_flush_and_put(struct io_ring_ctx *ctx, struct io_tw_state *ts) argument 1183 struct io_ring_ctx *ctx = NULL; local 1295 struct io_ring_ctx *ctx = req->ctx; local 1364 struct io_ring_ctx *ctx = req->ctx; local 1399 io_move_task_work_from_local(struct io_ring_ctx *ctx) argument 1413 io_run_local_work_continue(struct io_ring_ctx *ctx, int events, int min_events) argument 1425 __io_run_local_work(struct io_ring_ctx *ctx, struct io_tw_state *ts, int min_events) argument 1466 io_run_local_work_locked(struct io_ring_ctx *ctx, int min_events) argument 1482 io_run_local_work(struct io_ring_ctx *ctx, int min_events) argument 1578 __io_cq_lock(ctx); variable 1581 __io_flush_post_cqes(ctx); variable 1597 __io_cq_unlock_post(ctx); variable 1605 io_cqring_events(struct io_ring_ctx *ctx) argument 1616 io_iopoll_try_reap_events(struct io_ring_ctx *ctx) argument 1640 io_iopoll_check(struct io_ring_ctx *ctx, long min) argument 1728 struct io_ring_ctx *ctx = req->ctx; local 1830 struct io_ring_ctx *ctx = req->ctx; variable in typeref:struct:io_ring_ctx 2046 struct io_ring_ctx *ctx = req->ctx; local 2152 io_check_restriction(struct io_ring_ctx *ctx, struct io_kiocb *req, unsigned int sqe_flags) argument 2172 struct io_ring_ctx *ctx = req->ctx; local 2294 struct io_ring_ctx *ctx = req->ctx; local 2384 io_submit_state_end(struct io_ring_ctx *ctx) argument 2409 io_commit_sqring(struct io_ring_ctx *ctx) argument 2429 io_get_sqe(struct io_ring_ctx *ctx, const struct io_uring_sqe **sqe) argument 2507 io_submit_state_end(ctx); variable 2509 io_commit_sqring(ctx); variable 2527 io_run_task_work_sig(struct io_ring_ctx *ctx) argument 2551 io_cqring_wait_schedule(struct io_ring_ctx *ctx, struct io_wait_queue *iowq) argument 2587 io_cqring_wait(struct io_ring_ctx *ctx, int min_events, const sigset_t __user *sig, size_t sigsz, struct __kernel_timespec __user *uts) argument 2784 io_rings_map(struct io_ring_ctx *ctx, unsigned long uaddr, size_t size) argument 2791 io_sqes_map(struct io_ring_ctx *ctx, unsigned long uaddr, size_t size) argument 2798 io_rings_free(struct io_ring_ctx *ctx) argument 2825 rings_size(struct io_ring_ctx *ctx, unsigned int sq_entries, unsigned int cq_entries, size_t *sq_offset) argument 2864 io_req_caches_free(struct io_ring_ctx *ctx) argument 2887 io_ring_ctx_free(struct io_ring_ctx *ctx) argument 2940 struct io_ring_ctx *ctx = container_of(cb, struct io_ring_ctx, local 2955 io_activate_pollwq(struct io_ring_ctx *ctx) argument 2979 struct io_ring_ctx *ctx = file->private_data; local 3017 struct io_ring_ctx *ctx; member in struct:io_tctx_exit 3046 struct io_ring_ctx *ctx = container_of(work, struct io_ring_ctx, exit_work); local 3139 io_ring_ctx_wait_and_kill(struct io_ring_ctx *ctx) argument 3173 struct io_ring_ctx *ctx = file->private_data; local 3193 io_cancel_defer_files(struct io_ring_ctx *ctx, struct task_struct *task, bool cancel_all) argument 3220 io_uring_try_cancel_iowq(struct io_ring_ctx *ctx) argument 3244 io_uring_try_cancel_uring_cmd(struct io_ring_ctx *ctx, struct task_struct *task, bool cancel_all) argument 3275 io_uring_try_cancel_requests(struct io_ring_ctx *ctx, struct task_struct *task, bool cancel_all) argument 3346 struct io_ring_ctx *ctx; local 3430 struct io_ring_ctx *ctx = file->private_data; local 3607 struct io_ring_ctx *ctx; local 3757 io_allocate_scq_urings(struct io_ring_ctx *ctx, struct io_uring_params *p) argument 3827 io_uring_get_file(struct io_ring_ctx *ctx) argument 3837 struct io_ring_ctx *ctx; local [all...] |
/linux-master/fs/zonefs/ |
H A D | super.c | 497 struct zonefs_context *ctx = fc->fs_private; local 507 ctx->s_mount_opts &= ~ZONEFS_MNTOPT_ERRORS_MASK; 508 ctx->s_mount_opts |= result.uint_32; 511 ctx->s_mount_opts |= ZONEFS_MNTOPT_EXPLICIT_OPEN; 747 static int zonefs_readdir_root(struct file *file, struct dir_context *ctx) argument 755 if (ctx->pos >= inode->i_size) 758 if (!dir_emit_dots(file, ctx)) 761 if (ctx->pos == 2) { 765 if (!dir_emit(ctx, zonefs_zgroup_name(ztype), 3, 768 ctx 782 zonefs_readdir_zgroup(struct file *file, struct dir_context *ctx) argument 824 zonefs_readdir(struct file *file, struct dir_context *ctx) argument 1238 struct zonefs_context *ctx = fc->fs_private; local 1362 struct zonefs_context *ctx = fc->fs_private; local 1374 struct zonefs_context *ctx = fc->fs_private; local 1397 struct zonefs_context *ctx; local [all...] |
/linux-master/fs/smb/client/ |
H A D | smb2ops.c | 438 smb2_negotiate_wsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx) argument 444 wsize = ctx->wsize ? ctx->wsize : CIFS_DEFAULT_IOSIZE; 453 smb3_negotiate_wsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx) argument 459 wsize = ctx->wsize ? ctx->wsize : SMB3_DEFAULT_IOSIZE; 485 smb2_negotiate_rsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx) argument 491 rsize = ctx->rsize ? ctx->rsize : CIFS_DEFAULT_IOSIZE; 501 smb3_negotiate_rsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx) argument [all...] |
H A D | misc.c | 644 if (uid_eq(cifs_sb->ctx->backupuid, current_fsuid())) 648 if (in_group_p(cifs_sb->ctx->backupgid)) 994 struct cifs_aio_ctx *ctx; local 997 * Must use kzalloc to initialize ctx->bv to NULL and ctx->direct_io 1001 ctx = kzalloc(sizeof(struct cifs_aio_ctx), GFP_KERNEL); 1002 if (!ctx) 1005 INIT_LIST_HEAD(&ctx->list); 1006 mutex_init(&ctx->aio_mutex); 1007 init_completion(&ctx 1015 struct cifs_aio_ctx *ctx = container_of(refcount, local [all...] |
H A D | inode.c | 303 fattr->cf_uid = cifs_sb->ctx->linux_uid; 313 fattr->cf_gid = cifs_sb->ctx->linux_gid; 342 fattr->cf_uid = cifs_sb->ctx->linux_uid; 343 fattr->cf_gid = cifs_sb->ctx->linux_gid; 761 fattr->cf_uid = cifs_sb->ctx->linux_uid; 762 fattr->cf_gid = cifs_sb->ctx->linux_gid; 764 fattr->cf_mode = cifs_sb->ctx->file_mode; 770 fattr->cf_mode = S_IFDIR | cifs_sb->ctx->dir_mode; 779 fattr->cf_mode = S_IFREG | cifs_sb->ctx->file_mode; 1514 inode->i_uid = cifs_sb->ctx [all...] |
H A D | fs_context.c | 202 cifs_parse_security_flavors(struct fs_context *fc, char *value, struct smb3_fs_context *ctx) argument 211 ctx->sectype = Unspecified; 212 ctx->sign = false; 219 ctx->sign = true; 222 ctx->sectype = Kerberos; 225 ctx->sign = true; 228 ctx->sectype = RawNTLMSSP; 231 ctx->sign = true; 234 ctx->sectype = NTLMv2; 237 ctx 259 cifs_parse_cache_flavor(struct fs_context *fc, char *value, struct smb3_fs_context *ctx) argument 308 parse_reparse_flavor(struct fs_context *fc, char *value, struct smb3_fs_context *ctx) argument 342 smb3_fs_context_dup(struct smb3_fs_context *new_ctx, struct smb3_fs_context *ctx) argument 375 cifs_parse_smb_version(struct fs_context *fc, char *value, struct smb3_fs_context *ctx, bool is_smb3) argument 526 smb3_fs_context_fullpath(const struct smb3_fs_context *ctx, char dirsep) argument 553 smb3_parse_devname(const char *devname, struct smb3_fs_context *ctx) argument 703 struct smb3_fs_context *ctx = smb3_fc2context(fc); local 766 struct smb3_fs_context *ctx = smb3_fc2context(fc); local 797 struct smb3_fs_context *ctx = smb3_fc2context(fc); local 883 struct smb3_fs_context *ctx = smb3_fc2context(fc); local 940 struct smb3_fs_context *ctx = smb3_fc2context(fc); local 1635 struct smb3_fs_context *ctx; local 1723 smb3_cleanup_fs_context_contents(struct smb3_fs_context *ctx) argument 1756 smb3_cleanup_fs_context(struct smb3_fs_context *ctx) argument 1766 struct smb3_fs_context *ctx = cifs_sb->ctx; local [all...] |
H A D | fs_context.h | 290 extern void smb3_cleanup_fs_context_contents(struct smb3_fs_context *ctx); 291 extern void smb3_cleanup_fs_context(struct smb3_fs_context *ctx); 298 extern int smb3_fs_context_dup(struct smb3_fs_context *new_ctx, struct smb3_fs_context *ctx);
|
H A D | connect.c | 1484 match_security(struct TCP_Server_Info *server, struct smb3_fs_context *ctx) argument 1487 * The select_sectype function should either return the ctx->sectype 1491 if (server->ops->select_sectype(server, ctx->sectype) 1500 if (ctx->sign && !server->sign) 1508 struct smb3_fs_context *ctx, 1511 struct sockaddr *addr = (struct sockaddr *)&ctx->dstaddr; 1515 if (ctx->nosharesock) 1523 if (strcmp(ctx->vals->version_string, SMB3ANY_VERSION_STRING) == 0) { 1526 } else if (strcmp(ctx->vals->version_string, 1530 } else if ((server->vals != ctx 1507 match_server(struct TCP_Server_Info *server, struct smb3_fs_context *ctx, bool match_super) argument 1599 cifs_find_tcp_session(struct smb3_fs_context *ctx) argument 1680 cifs_get_tcp_session(struct smb3_fs_context *ctx, struct TCP_Server_Info *primary_server) argument 1865 match_session(struct cifs_ses *ses, struct smb3_fs_context *ctx) argument 1923 cifs_setup_ipc(struct cifs_ses *ses, struct smb3_fs_context *ctx) argument 1978 cifs_find_smb_ses(struct TCP_Server_Info *server, struct smb3_fs_context *ctx) argument 2087 cifs_set_cifscreds(struct smb3_fs_context *ctx, struct cifs_ses *ses) argument 2245 cifs_get_smb_ses(struct TCP_Server_Info *server, struct smb3_fs_context *ctx) argument 2393 match_tcon(struct cifs_tcon *tcon, struct smb3_fs_context *ctx) argument 2423 cifs_find_tcon(struct cifs_ses *ses, struct smb3_fs_context *ctx) argument 2523 cifs_get_tcon(struct cifs_ses *ses, struct smb3_fs_context *ctx) argument 2815 struct smb3_fs_context *ctx = mnt_data->ctx; local 2839 struct smb3_fs_context *ctx; local 3156 reset_cifs_unix_caps(unsigned int xid, struct cifs_tcon *tcon, struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx) argument 3266 struct smb3_fs_context *ctx = cifs_sb->ctx; local 3337 struct smb3_fs_context *ctx; local 3385 struct smb3_fs_context *ctx; local 3565 struct smb3_fs_context *ctx = mnt_ctx->fs_ctx; local 3602 cifs_mount(struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx) argument 3642 cifs_mount(struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx) argument 3989 cifs_set_vol_auth(struct smb3_fs_context *ctx, struct cifs_ses *ses) argument 4007 struct smb3_fs_context *ctx; local [all...] |
H A D | cifsglob.h | 337 unsigned int (*negotiate_wsize)(struct cifs_tcon *tcon, struct smb3_fs_context *ctx); 339 unsigned int (*negotiate_rsize)(struct cifs_tcon *tcon, struct smb3_fs_context *ctx); 659 struct smb3_fs_context *ctx; member in struct:cifs_mnt_data 1499 struct cifs_aio_ctx *ctx; member in struct:cifs_readdata 1523 struct cifs_aio_ctx *ctx; member in struct:cifs_writedata 1984 * vol_info->ctx_lock vol_info->ctx
|
/linux-master/fs/ceph/ |
H A D | caps.c | 203 struct ceph_cap_reservation *ctx, int need) 216 doutc(cl, "ctx=%p need=%d\n", ctx, need); 276 pr_warn_client(cl, "ctx=%p ENOMEM need=%d got=%d\n", ctx, need, 284 ctx->count = need; 285 ctx->used = 0; 302 doutc(cl, "ctx=%p %d = %d used + %d resv + %d avail\n", ctx, 309 struct ceph_cap_reservation *ctx) 202 ceph_reserve_caps(struct ceph_mds_client *mdsc, struct ceph_cap_reservation *ctx, int need) argument 308 ceph_unreserve_caps(struct ceph_mds_client *mdsc, struct ceph_cap_reservation *ctx) argument 330 ceph_get_cap(struct ceph_mds_client *mdsc, struct ceph_cap_reservation *ctx) argument [all...] |
/linux-master/block/ |
H A D | blk-iocost.c | 1462 struct iocg_wake_ctx *ctx = key; local 1463 u64 cost = abs_cost_to_cost(wait->abs_cost, ctx->hw_inuse); 1465 ctx->vbudget -= cost; 1467 if (ctx->vbudget < 0) 1470 iocg_commit_bio(ctx->iocg, wait->bio, wait->abs_cost, cost); 1494 struct iocg_wake_ctx ctx = { .iocg = iocg }; local 1537 ctx.vbudget = vbudget; 1538 current_hweight(iocg, NULL, &ctx.hw_inuse); 1540 __wake_up_locked_key(&iocg->waitq, TASK_NORMAL, &ctx); 1553 if (WARN_ON_ONCE(ctx 3090 struct blkg_conf_ctx ctx; local 3209 struct blkg_conf_ctx ctx; local 3403 struct blkg_conf_ctx ctx; local [all...] |
H A D | blk-cgroup.h | 218 void blkg_conf_init(struct blkg_conf_ctx *ctx, char *input); 219 int blkg_conf_open_bdev(struct blkg_conf_ctx *ctx); 221 struct blkg_conf_ctx *ctx); 222 void blkg_conf_exit(struct blkg_conf_ctx *ctx);
|
H A D | blk-cgroup.c | 733 * @ctx: blkg_conf_ctx to initialize 736 * Initialize @ctx which can be used to parse blkg config input string @input. 737 * Once initialized, @ctx can be used with blkg_conf_open_bdev() and 740 void blkg_conf_init(struct blkg_conf_ctx *ctx, char *input) argument 742 *ctx = (struct blkg_conf_ctx){ .input = input }; 748 * @ctx: blkg_conf_ctx initialized with blkg_conf_init() 751 * @ctx->input and get and store the matching bdev in @ctx->bdev. @ctx->body is 754 * This function may be called multiple times on @ctx an 759 blkg_conf_open_bdev(struct blkg_conf_ctx *ctx) argument [all...] |
/linux-master/tools/testing/cxl/test/ |
H A D | cxl.c | 477 struct cxl_cedt_context *ctx = arg; local 478 struct device *dev = ctx->dev; 657 struct target_map_ctx *ctx = data; local 659 ctx->target_map[ctx->index++] = pdev->id; 661 if (ctx->index > ctx->target_count) { 866 struct target_map_ctx ctx = { local 899 rc = device_for_each_child(port->uport_dev, &ctx,
|
/linux-master/net/unix/ |
H A D | af_unix.c | 3413 struct bpf_iter__unix ctx; local 3416 ctx.meta = meta; 3417 ctx.unix_sk = unix_sk; 3418 ctx.uid = uid; 3419 return bpf_iter_run_prog(prog, &ctx);
|
/linux-master/net/ipv6/ |
H A D | ip6_fib.c | 2713 struct bpf_iter__ipv6_route ctx; local 2715 ctx.meta = meta; 2716 ctx.rt = v; 2717 return bpf_iter_run_prog(prog, &ctx);
|
H A D | addrconf.c | 738 } *ctx = (void *)cb->ctx; local 759 for_each_netdev_dump(net, dev, ctx->ifindex) { 773 if (ctx->all_default == 0) { 782 ctx->all_default++; 784 if (ctx->all_default == 1) { 793 ctx->all_default++; 5406 } *ctx = (void *)cb->ctx; local 5428 &ctx 6121 } *ctx = (void *)cb->ctx; local [all...] |
/linux-master/fs/tracefs/ |
H A D | event_inode.c | 125 static int eventfs_iterate(struct file *file, struct dir_context *ctx); 576 static int eventfs_iterate(struct file *file, struct dir_context *ctx) argument 591 if (!dir_emit_dots(file, ctx)) 598 c = ctx->pos - 2; 618 for (i = c; i < ei->nr_entries; i++, ctx->pos++) { 637 if (!dir_emit(ctx, name, strlen(name), ino, DT_REG)) 652 ctx->pos++; 661 if (!dir_emit(ctx, name, strlen(name), ino, DT_DIR)) 671 /* Incremented ctx->pos without adding something, reset it */ 672 ctx [all...] |