/linux-master/tools/testing/selftests/netfilter/ |
H A D | conntrack_tcp_unreplied.sh | 96 type filter hook input priority 0; policy accept; 120 type nat hook prerouting priority 0; policy accept;
|
H A D | nft_conntrack_helper.sh | 79 type filter hook prerouting priority 0; policy accept; 83 type filter hook output priority 0; policy accept;
|
H A D | bridge_netfilter.sh | 133 type filter hook input priority 1; policy accept 141 type filter hook forward priority 0; policy accept
|
H A D | nft_nat_zones.sh | 28 # NAT Gateway uses policy routing and conntrack zones to isolate client 147 # gw: policy routing 186 type filter hook prerouting priority -64000; policy accept; 196 type nat hook postrouting priority 0; policy accept; 201 type filter hook prerouting priority -100; policy accept;
|
H A D | nft_zones_many.sh | 54 type filter hook output priority -64000; policy accept;
|
/linux-master/net/netfilter/ |
H A D | nft_counter.c | 192 .policy = nft_counter_policy, 305 .policy = nft_counter_policy,
|
H A D | nft_quota.c | 198 .policy = nft_quota_policy, 268 .policy = nft_quota_policy,
|
H A D | nfnetlink_cttimeout.c | 133 /* You cannot replace one timeout policy by another of 547 .policy = cttimeout_nla_policy 553 .policy = cttimeout_nla_policy 559 .policy = cttimeout_nla_policy 565 .policy = cttimeout_nla_policy 571 .policy = cttimeout_nla_policy
|
H A D | nfnetlink_acct.c | 386 .policy = nfnl_acct_policy 392 .policy = nfnl_acct_policy 398 .policy = nfnl_acct_policy 404 .policy = nfnl_acct_policy
|
H A D | nft_reject_netdev.c | 169 .policy = nft_reject_policy,
|
/linux-master/include/linux/ |
H A D | mempolicy.h | 25 * Describe a memory policy. 28 * For VMA related allocations the VMA policy is preferred, otherwise 29 * the process policy is used. Interrupts ignore the memory policy 32 * Locking policy for interleave: 37 * Freeing policy: 41 * Duplicating policy objects: 61 * The default fast path of a NULL MPOL_DEFAULT policy is always inlined. 118 struct mempolicy *policy; member in struct:sp_node 178 extern bool apply_policy_zone(struct mempolicy *policy, enu [all...] |
/linux-master/security/apparmor/ |
H A D | net.c | 16 #include "include/policy.h" 130 state = aa_dfa_match_len(rules->policy->dfa, state, (char *) &buffer, 132 perms = *aa_lookup_perms(rules->policy, state);
|
/linux-master/arch/arm64/kernel/ |
H A D | topology.c | 216 struct cpufreq_policy *policy = data; local 219 amu_fie_setup(policy->related_cpus);
|
/linux-master/drivers/gpu/drm/amd/display/dc/dml2/ |
H A D | dml2_wrapper.c | 236 build_unoptimized_policy_settings(dml2->v20.dml_core_ctx.project, &dml2->v20.dml_core_ctx.policy); 341 build_unoptimized_policy_settings(dml2->v20.dml_core_ctx.project, &dml2->v20.dml_core_ctx.policy); 378 s->cur_policy = dml2->v20.dml_core_ctx.policy; 389 dml2->v20.dml_core_ctx.policy = s->new_policy; 402 // Else, restore policy to current 403 dml2->v20.dml_core_ctx.policy = s->cur_policy; 579 memset(&dml2->v20.dml_core_ctx.policy, 0, sizeof(struct dml_mode_eval_policy_st)); 659 memset(&dml2->v20.dml_core_ctx.policy, 0, sizeof(struct dml_mode_eval_policy_st)); 663 build_unoptimized_policy_settings(dml2->v20.dml_core_ctx.project, &dml2->v20.dml_core_ctx.policy);
|
/linux-master/tools/testing/selftests/net/ |
H A D | srv6_end_next_csid_l3vpn_test.sh | 104 # SRv6 policy enforcement consists of encapsulating the received traffic into a 646 # build and install the SRv6 policy into the ingress SRv6 router as well as the 662 local policy 684 # single SID policy 688 policy="$(build_ipv6_addr "${container}")" 694 # double SID policy 697 policy="$(build_ipv6_addr "${container}"),${decapsid}" 700 # apply encap policy 704 encap seg6 mode "${HEADEND_ENCAP}" segs "${policy}" \ 715 encap seg6 mode "${HEADEND_ENCAP}" segs "${policy}" \ [all...] |
H A D | srv6_end_x_next_csid_l3vpn_test.sh | 107 # SRv6 policy enforcement consists of encapsulating the received traffic into a 248 # H.Encap.Red behavior with policy i) described above. Since there is only one 249 # SID, the SRH header is omitted and the policy is inserted directly into the DA 706 # build and install the SRv6 policy into the ingress SRv6 router as well as the 722 local policy 744 # single SID policy 748 policy="$(build_ipv6_addr "${container}")" 754 # double SID policy 757 policy="$(build_ipv6_addr "${container}"),${decapsid}" 760 # apply encap policy [all...] |
/linux-master/drivers/acpi/ |
H A D | processor_perflib.c | 30 * _PPC support is implemented as a CPUfreq policy notifier: 32 * the ACPI core is asked to change the speed policy, the maximum 36 * policy is adjusted accordingly. 173 void acpi_processor_ppc_init(struct cpufreq_policy *policy) argument 177 for_each_cpu(cpu, policy->related_cpus) { 191 ret = freq_qos_add_request(&policy->constraints, 200 void acpi_processor_ppc_exit(struct cpufreq_policy *policy) argument 204 for_each_cpu(cpu, policy->related_cpus) {
|
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | plpks.c | 441 if (!(var->policy & PLPKS_SIGNEDUPDATE)) 461 label->size, var->policy, flags, 495 if (var.policy & PLPKS_SIGNEDUPDATE) 509 virt_to_phys(label), label->size, var.policy, 605 var->policy = retbuf[1];
|
/linux-master/drivers/base/ |
H A D | arch_topology.c | 431 struct cpufreq_policy *policy = data; local 438 cpumask_pr_args(policy->related_cpus), 441 cpumask_andnot(cpus_to_visit, cpus_to_visit, policy->related_cpus); 443 for_each_cpu(cpu, policy->related_cpus) { 444 per_cpu(capacity_freq_ref, cpu) = policy->cpuinfo.max_freq;
|
/linux-master/arch/arm/mm/ |
H A D | mmu.c | 73 const char policy[16]; member in struct:cachepolicy 81 .policy = "uncached", 86 .policy = "buffered", 91 .policy = "writethrough", 96 .policy = "writeback", 101 .policy = "writealloc", 114 * the C code sets the page tables up with the same policy as the head 133 pr_err("ERROR: could not find cache policy\n"); 146 int len = strlen(cache_policies[i].policy); 148 if (memcmp(p, cache_policies[i].policy, le [all...] |
/linux-master/kernel/sched/ |
H A D | sched.h | 172 static inline int idle_policy(int policy) argument 174 return policy == SCHED_IDLE; 176 static inline int fair_policy(int policy) argument 178 return policy == SCHED_NORMAL || policy == SCHED_BATCH; 181 static inline int rt_policy(int policy) argument 183 return policy == SCHED_FIFO || policy == SCHED_RR; 186 static inline int dl_policy(int policy) argument 188 return policy 190 valid_policy(int policy) argument [all...] |
/linux-master/drivers/clk/bcm/ |
H A D | clk-kona.c | 17 * "Normal", and "Turbo".) A lower policy number has lower power 18 * consumption, and policy 2 is the default. 208 struct bcm_policy_ctl *control = &ccu->policy.control; 214 /* If we don't need to control policy for this CCU, we're done. */ 224 pr_err("%s: ccu %s policy engine wouldn't go idle\n", 254 pr_err("%s: ccu %s policy engine never started\n", 262 struct bcm_lvm_en *enable = &ccu->policy.enable; 267 /* If we don't need to control policy for this CCU, we're done. */ 276 pr_err("%s: ccu %s policy engine already stopped\n", 287 pr_err("%s: ccu %s policy engin 302 policy_init(struct ccu_data *ccu, struct bcm_clk_policy *policy) argument [all...] |
/linux-master/drivers/thermal/ |
H A D | imx_thermal.c | 204 struct cpufreq_policy *policy; member in struct:imx_thermal_data 563 data->policy = cpufreq_cpu_get(0); 564 if (!data->policy) { 565 pr_debug("%s: CPUFreq policy not found\n", __func__); 569 np = of_get_cpu_node(data->policy->cpu, NULL); 572 data->cdev = cpufreq_cooling_register(data->policy); 575 cpufreq_cpu_put(data->policy); 587 cpufreq_cpu_put(data->policy);
|
/linux-master/drivers/cpufreq/ |
H A D | pmac64-cpufreq.c | 312 static int g5_cpufreq_target(struct cpufreq_policy *policy, unsigned int index) argument 322 static int g5_cpufreq_cpu_init(struct cpufreq_policy *policy) argument 324 cpufreq_generic_init(policy, g5_cpu_freqs, transition_latency);
|
/linux-master/net/ieee802154/ |
H A D | netlink.c | 131 .policy = ieee802154_policy,
|