/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | exceptions.c | 37 &(int){bpf_program__fd(skel->progs.exception_tail_call_target)}, BPF_ANY), 43 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs._prog), &ropts); \ 124 prog = eskel->progs.pfentry; 127 bpf_program__fd(skel->progs.exception_ext_mod_cb_runtime), 134 prog = eskel->progs.throwing_fentry; 137 bpf_program__fd(skel->progs.exception_ext_mod_cb_runtime), 144 prog = eskel->progs.pfexit; 147 bpf_program__fd(skel->progs.exception_ext_mod_cb_runtime), 154 prog = eskel->progs.throwing_fexit; 157 bpf_program__fd(skel->progs [all...] |
H A D | autoattach.c | 16 bpf_program__set_autoattach(skel->progs.prog2, false); 17 ASSERT_TRUE(bpf_program__autoattach(skel->progs.prog1), "autoattach_prog1"); 18 ASSERT_FALSE(bpf_program__autoattach(skel->progs.prog2), "autoattach_prog2");
|
H A D | missed.c | 28 prog_fd = bpf_program__fd(skel->progs.trigger); 78 prog_fd = bpf_program__fd(skel->progs.trigger); 83 ASSERT_EQ(get_missed_count(bpf_program__fd(skel->progs.test1)), 0, "test1_recursion_misses"); 84 ASSERT_GE(get_missed_count(bpf_program__fd(skel->progs.test2)), 1, "test2_recursion_misses"); 85 ASSERT_GE(get_missed_count(bpf_program__fd(skel->progs.test3)), 1, "test3_recursion_misses"); 86 ASSERT_GE(get_missed_count(bpf_program__fd(skel->progs.test4)), 1, "test4_recursion_misses"); 87 ASSERT_GE(get_missed_count(bpf_program__fd(skel->progs.test5)), 1, "test5_recursion_misses"); 116 prog_fd = bpf_program__fd(skel->progs.trigger); 121 ASSERT_EQ(get_missed_count(bpf_program__fd(skel->progs.test1)), 0, "test1_recursion_misses"); 122 ASSERT_EQ(get_missed_count(bpf_program__fd(skel->progs [all...] |
H A D | rcu_read_lock.c | 26 bpf_program__set_autoload(skel->progs.get_cgroup_id, true); 27 bpf_program__set_autoload(skel->progs.task_succ, true); 28 bpf_program__set_autoload(skel->progs.two_regions, true); 29 bpf_program__set_autoload(skel->progs.non_sleepable_1, true); 30 bpf_program__set_autoload(skel->progs.non_sleepable_2, true); 31 bpf_program__set_autoload(skel->progs.task_trusted_non_rcuptr, true); 32 bpf_program__set_autoload(skel->progs.rcu_read_lock_subprog, true); 33 bpf_program__set_autoload(skel->progs.rcu_read_lock_global_subprog, true); 34 bpf_program__set_autoload(skel->progs.rcu_read_lock_subprog_lock, true); 35 bpf_program__set_autoload(skel->progs [all...] |
H A D | tracing_failure.c | 16 bpf_program__set_autoload(skel->progs.test_spin_lock, true); 18 bpf_program__set_autoload(skel->progs.test_spin_unlock, true);
|
H A D | btf_tag.c | 123 bpf_program__set_autoload(skel->progs.test_sys_getsockname, false); 125 bpf_program__set_autoload(skel->progs.test_user2, false); 127 bpf_program__set_autoload(skel->progs.test_user1, false); 152 bpf_program__set_autoload(skel->progs.test_user2, false); 153 bpf_program__set_autoload(skel->progs.test_user1, false); 177 bpf_program__set_autoload(skel->progs.test_percpu_load, false); 178 bpf_program__set_autoload(skel->progs.test_percpu_helper, false); 180 bpf_program__set_autoload(skel->progs.test_percpu2, false); 182 bpf_program__set_autoload(skel->progs.test_percpu1, false); 207 bpf_program__set_autoload(skel->progs [all...] |
H A D | log_fixup.c | 24 bpf_program__set_autoload(skel->progs.bad_relo, true); 26 bpf_program__set_log_buf(skel->progs.bad_relo, log_buf, log_buf_size ?: sizeof(log_buf)); 27 bpf_program__set_log_level(skel->progs.bad_relo, 1 | 8); /* BPF_LOG_FIXED to force truncation */ 82 bpf_program__set_autoload(skel->progs.bad_relo_subprog, true); 83 bpf_program__set_log_buf(skel->progs.bad_relo_subprog, log_buf, sizeof(log_buf)); 116 bpf_program__set_autoload(skel->progs.use_missing_map, true); 117 bpf_program__set_log_buf(skel->progs.use_missing_map, log_buf, sizeof(log_buf)); 148 bpf_program__set_autoload(skel->progs.use_missing_kfunc, true); 149 bpf_program__set_log_buf(skel->progs.use_missing_kfunc, log_buf, sizeof(log_buf));
|
H A D | global_func_dead_code.c | 20 bpf_program__set_autoload(tgt_skel->progs.chained_global_func_calls_success, true); 26 tgt_fd = bpf_program__fd(tgt_skel->progs.chained_global_func_calls_success); 33 err = bpf_program__set_attach_target(skel->progs.freplace_prog, tgt_fd, "global_good"); 47 bpf_program__set_log_buf(skel->progs.freplace_prog, log_buf, sizeof(log_buf)); 48 err = bpf_program__set_attach_target(skel->progs.freplace_prog, tgt_fd, "global_dead");
|
H A D | attach_probe.c | 57 kprobe_link = bpf_program__attach_kprobe_opts(skel->progs.handle_kprobe, 65 kretprobe_link = bpf_program__attach_kprobe_opts(skel->progs.handle_kretprobe, 76 uprobe_link = bpf_program__attach_uprobe_opts(skel->progs.handle_uprobe, 86 uretprobe_link = bpf_program__attach_uprobe_opts(skel->progs.handle_uretprobe, 99 bpf_program__attach_uprobe_opts(skel->progs.handle_uprobe_byname, 130 skel->links.handle_kprobe_auto = bpf_program__attach(skel->progs.handle_kprobe_auto); 133 skel->links.handle_kretprobe_auto = bpf_program__attach(skel->progs.handle_kretprobe_auto); 137 uprobe_err_link = bpf_program__attach(skel->progs.handle_uprobe_byname); 144 bpf_program__attach(skel->progs.handle_uretprobe_byname); 170 bpf_program__attach_uprobe_opts(skel->progs [all...] |
H A D | refcounted_kptr.c | 34 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.rbtree_wrong_owner_remove_fail_a1), &opts); 38 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.rbtree_wrong_owner_remove_fail_b), &opts); 42 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.rbtree_wrong_owner_remove_fail_a2), &opts);
|
H A D | recursive_attach.c | 9 /* Test recursive attachment of tracing progs with more than one nesting level 13 * - Recursive loading of tracing progs, without attaching (attach = false, 19 * - Recursive attach of tracing progs (attach = true, detach = false). The 26 * - Recursive attach and detach of tracing progs (attach = true, detach = 47 /* Create an attachment chain with two fentry progs */ 56 prog = tracing_chain[i]->progs.recursive_attach; 58 prev_fd = bpf_program__fd(target_skel->progs.test1); 61 prev_fd = bpf_program__fd(tracing_chain[i-1]->progs.recursive_attach); 128 prog = tracing_skel->progs.recursive_attach; 129 tgt_prog_fd = bpf_program__fd(target_skel->progs [all...] |
H A D | ptr_untrusted.c | 21 skel->links.lsm_run = bpf_program__attach_lsm(skel->progs.lsm_run); 26 skel->links.raw_tp_run = bpf_program__attach_raw_tracepoint(skel->progs.raw_tp_run,
|
H A D | percpu_alloc.c | 17 bpf_program__set_autoload(skel->progs.test_array_map_1, true); 18 bpf_program__set_autoload(skel->progs.test_array_map_2, true); 19 bpf_program__set_autoload(skel->progs.test_array_map_3, true); 20 bpf_program__set_autoload(skel->progs.test_array_map_4, true); 33 prog_fd = bpf_program__fd(skel->progs.test_array_map_1); 53 bpf_program__set_autoload(skel->progs.test_array_map_10, true); 66 prog_fd = bpf_program__fd(skel->progs.test_array_map_10); 101 prog_fd = bpf_program__fd(skel->progs.test_cgrp_local_storage_1);
|
H A D | local_kptr_stash.c | 23 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.stash_rb_nodes), &opts); 44 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.stash_plain), &opts); 65 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.stash_local_with_root), &opts); 86 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.stash_rb_nodes), &opts); 90 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.unstash_rb_node), &opts); 111 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.refcount_acquire_without_unstash), 116 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.stash_refcounted_node), &opts); 120 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.refcount_acquire_without_unstash),
|
H A D | test_ldsx_insn.c | 22 bpf_program__set_autoload(skel->progs.rdonly_map_prog, true); 23 bpf_program__set_autoload(skel->progs.map_val_prog, true); 24 bpf_program__set_autoload(skel->progs.test_ptr_struct_arg, true); 66 bpf_program__set_autoload(skel->progs._getsockopt, true); 73 bpf_program__attach_cgroup(skel->progs._getsockopt, cgroup_fd); 115 bpf_program__set_autoload(skel->progs._tc, true); 121 prog_fd = bpf_program__fd(skel->progs._tc);
|
H A D | cgroup1_hierarchy.c | 16 lsm_link = bpf_program__attach_lsm(skel->progs.lsm_run); 21 fentry_link = bpf_program__attach_trace(skel->progs.fentry_run); 34 lsm_link = bpf_program__attach_lsm(skel->progs.lsm_s_run); 39 fentry_link = bpf_program__attach_trace(skel->progs.fentry_run); 52 lsm_link = bpf_program__attach_lsm(skel->progs.lsm_run); 57 fentry_link = bpf_program__attach_trace(skel->progs.fentry_run); 81 err = bpf_program__set_attach_target(skel->progs.fentry_run, 0, "bpf_fentry_test1");
|
H A D | tc_links.c | 28 pid1 = id_from_prog_fd(bpf_program__fd(skel->progs.tc1)); 29 pid2 = id_from_prog_fd(bpf_program__fd(skel->progs.tc2)); 39 link = bpf_program__attach_tcx(skel->progs.tc1, loopback, &optl); 74 link = bpf_program__attach_tcx(skel->progs.tc2, loopback, &optl); 128 ASSERT_EQ(bpf_program__set_expected_attach_type(skel->progs.tc1, target), 130 ASSERT_EQ(bpf_program__set_expected_attach_type(skel->progs.tc2, target), 132 ASSERT_EQ(bpf_program__set_expected_attach_type(skel->progs.tc3, target), 134 ASSERT_EQ(bpf_program__set_expected_attach_type(skel->progs.tc4, target), 141 pid1 = id_from_prog_fd(bpf_program__fd(skel->progs.tc1)); 142 pid2 = id_from_prog_fd(bpf_program__fd(skel->progs [all...] |
H A D | sk_lookup.c | 532 link[0] = attach_lookup_prog(skel->progs.lookup_pass); 535 link[1] = attach_lookup_prog(skel->progs.lookup_pass); 538 link[2] = attach_lookup_prog(skel->progs.lookup_drop); 676 .lookup_prog = skel->progs.redir_port, 684 .lookup_prog = skel->progs.redir_ip4, 692 .lookup_prog = skel->progs.select_sock_a, 693 .reuseport_prog = skel->progs.select_sock_b, 702 .lookup_prog = skel->progs.select_sock_a_no_reuseport, 703 .reuseport_prog = skel->progs.select_sock_b, 712 .lookup_prog = skel->progs [all...] |
H A D | struct_ops_autocreate.c | 65 ASSERT_TRUE(bpf_program__autoload(skel->progs.test_1), "test_1 default autoload"); 66 ASSERT_TRUE(bpf_program__autoload(skel->progs.test_2), "test_2 default autoload"); 72 ASSERT_TRUE(bpf_program__autoload(skel->progs.test_1), "test_1 actual autoload"); 73 ASSERT_FALSE(bpf_program__autoload(skel->progs.test_2), "test_2 actual autoload"); 124 ASSERT_FALSE(bpf_program__autoload(skel->progs.foo), "foo default autoload"); 125 ASSERT_FALSE(bpf_program__autoload(skel->progs.bar), "bar default autoload"); 128 skel->struct_ops.testmod_1->test_1 = skel->progs.foo; 134 ASSERT_TRUE(bpf_program__autoload(skel->progs.foo), "foo actual autoload"); 135 ASSERT_FALSE(bpf_program__autoload(skel->progs.bar), "bar actual autoload");
|
H A D | map_kptr.c | 25 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref1), &opts); 28 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref2), &opts); 32 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_ls_map_kptr_ref1), &lopts); 36 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_ls_map_kptr_ref2), &lopts); 55 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts); 63 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts); 70 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts); 77 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts); 84 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref3), &opts); 91 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs [all...] |
H A D | custom_sec_handlers.c | 133 ASSERT_EQ(bpf_program__type(skel->progs.abc1), BPF_PROG_TYPE_RAW_TRACEPOINT, "abc1_type"); 134 ASSERT_FALSE(bpf_program__autoload(skel->progs.abc1), "abc1_autoload"); 136 ASSERT_EQ(bpf_program__type(skel->progs.abc2), BPF_PROG_TYPE_RAW_TRACEPOINT, "abc2_type"); 137 ASSERT_EQ(bpf_program__type(skel->progs.custom1), BPF_PROG_TYPE_TRACEPOINT, "custom1_type"); 138 ASSERT_EQ(bpf_program__type(skel->progs.custom2), BPF_PROG_TYPE_TRACEPOINT, "custom2_type"); 139 ASSERT_EQ(bpf_program__type(skel->progs.kprobe1), BPF_PROG_TYPE_TRACEPOINT, "kprobe1_type"); 140 ASSERT_EQ(bpf_program__type(skel->progs.xyz), BPF_PROG_TYPE_SYSCALL, "xyz_type"); 154 skel->links.xyz = bpf_program__attach(skel->progs.kprobe1);
|
H A D | linked_funcs.c | 20 bpf_program__set_autoload(skel->progs.handler1, true); 21 bpf_program__set_autoload(skel->progs.handler2, true);
|
H A D | xdp_link.c | 24 prog_fd1 = bpf_program__fd(skel1->progs.xdp_handler); 29 prog_fd2 = bpf_program__fd(skel2->progs.xdp_handler); 54 link = bpf_program__attach_xdp(skel1->progs.xdp_handler, IFINDEX_LO); 70 link = bpf_program__attach_xdp(skel1->progs.xdp_handler, IFINDEX_LO); 97 link = bpf_program__attach_xdp(skel2->progs.xdp_handler, IFINDEX_LO); 107 link = bpf_program__attach_xdp(skel2->progs.xdp_handler, IFINDEX_LO); 117 err = bpf_link__update_program(link, skel1->progs.xdp_handler); 132 err = bpf_link__update_program(link, skel1->progs.tc_handler);
|
/linux-master/include/net/netns/ |
H A D | bpf.h | 22 /* Array of programs to run compiled from progs or links */ 24 struct bpf_prog *progs[MAX_NETNS_BPF_ATTACH_TYPE]; member in struct:netns_bpf
|
/linux-master/drivers/media/rc/ |
H A D | bpf-lirc.c | 157 old_array = lirc_rcu_dereference(raw->progs); 167 rcu_assign_pointer(raw->progs, new_array); 195 old_array = lirc_rcu_dereference(raw->progs); 205 rcu_assign_pointer(raw->progs, new_array); 219 if (raw->progs) { 221 bpf_prog_run_array(rcu_dereference(raw->progs), 238 array = lirc_rcu_dereference(rcdev->raw->progs); 298 struct bpf_prog_array *progs; local 319 progs = lirc_rcu_dereference(rcdev->raw->progs); [all...] |