Lines Matching refs:fd

330 	 * rather to the map fd instead, so they have no meaning upon
373 * disappears - even if someone holds an fd to them they are unusable,
830 static fmode_t map_get_sys_perms(struct bpf_map *map, struct fd f)
1278 * on fd availability and access to bpffs, so are dependent on
1399 /* failed to allocate fd.
1421 /* if error is returned, fd is released.
1422 * On success caller should complete fd access with matching fdput()
1424 struct bpf_map *__bpf_map_get(struct fd f)
1451 struct fd f = fdget(ufd);
1467 struct fd f = fdget(ufd);
1544 struct fd f;
1619 struct fd f;
1674 struct fd f;
1732 struct fd f;
2018 struct fd f;
2107 struct fd f;
2229 * disappears - even if someone grabs an fd to them they are unusable,
2410 static struct bpf_prog *____bpf_prog_get(struct fd f)
2477 struct fd f = fdget(ufd);
2748 * on fd availability and access to bpffs, so are dependent on
2771 /* attach_prog_fd/attach_btf_obj_fd can specify fd of either bpf_prog
2927 * one reference and any parallel close on that fd could
3033 put_unused_fd(primer->fd);
3192 int fd, id;
3194 fd = get_unused_fd_flags(O_CLOEXEC);
3195 if (fd < 0)
3196 return fd;
3201 put_unused_fd(fd);
3208 put_unused_fd(fd);
3214 primer->fd = fd;
3226 fd_install(primer->fd, primer->file);
3228 return primer->fd;
3238 struct fd f = fdget(ufd);
3895 int fd;
3906 fd = bpf_raw_tp_link_attach(prog, tp_name, cookie);
3907 if (fd < 0)
3909 return fd;
4387 int fd;
4399 fd = bpf_prog_new_fd(prog);
4400 if (fd < 0)
4403 return fd;
4413 int fd;
4437 fd = bpf_map_new_fd(map, f_flags);
4438 if (fd < 0)
4441 return fd;
4957 struct fd f;
5083 u32 fd = attr->task_fd_query.fd;
5105 file = fget_task(task, fd);
5171 struct fd f;
5472 int fd;
5484 fd = bpf_link_new_fd(link);
5485 if (fd < 0)
5488 return fd;
5507 int fd;
5517 fd = anon_inode_getfd("bpf-stats", &bpf_stats_fops, NULL, O_CLOEXEC);
5518 if (fd >= 0)
5522 return fd;
5894 BPF_CALL_1(bpf_sys_close, u32, fd)
5901 return close_fd(fd);