/linux-master/arch/x86/kvm/mmu/ |
H A D | tdp_iter.h | 99 int min_level; member in struct:tdp_iter 123 #define for_each_tdp_pte_min_level(iter, root, min_level, start, end) \ 124 for (tdp_iter_start(&iter, root, min_level, start); \ 134 int min_level, gfn_t next_last_level_gfn);
|
H A D | tdp_iter.c | 40 int min_level, gfn_t next_last_level_gfn) 50 iter->min_level = min_level; 82 if (iter->level == iter->min_level) 39 tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root, int min_level, gfn_t next_last_level_gfn) argument
|
H A D | tdp_mmu.h | 37 const struct kvm_memory_slot *slot, int min_level); 49 int min_level);
|
H A D | tdp_mmu.c | 1308 * Remove write access from all SPTEs at or above min_level that map GFNs 1313 gfn_t start, gfn_t end, int min_level) 1321 BUG_ON(min_level > KVM_MAX_HUGEPAGE_LEVEL); 1323 for_each_tdp_pte_min_level(iter, root, min_level, start, end) { 1347 * only affect leaf SPTEs down to min_level. 1351 const struct kvm_memory_slot *slot, int min_level) 1360 slot->base_gfn + slot->npages, min_level); 1743 gfn_t gfn, int min_level) 1749 BUG_ON(min_level > KVM_MAX_HUGEPAGE_LEVEL); 1753 for_each_tdp_pte_min_level(iter, root, min_level, gf 1312 wrprot_gfn_range(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t start, gfn_t end, int min_level) argument 1350 kvm_tdp_mmu_wrprot_slot(struct kvm *kvm, const struct kvm_memory_slot *slot, int min_level) argument 1742 write_protect_gfn(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t gfn, int min_level) argument 1778 kvm_tdp_mmu_write_protect_gfn(struct kvm *kvm, struct kvm_memory_slot *slot, gfn_t gfn, int min_level) argument [all...] |
H A D | mmu_internal.h | 173 int min_level);
|
H A D | mmu.c | 1416 int min_level) 1423 for (i = min_level; i <= KVM_MAX_HUGEPAGE_LEVEL; ++i) { 1431 kvm_tdp_mmu_write_protect_gfn(kvm, slot, gfn, min_level); 1414 kvm_mmu_slot_gfn_write_protect(struct kvm *kvm, struct kvm_memory_slot *slot, u64 gfn, int min_level) argument
|
/linux-master/net/rxrpc/ |
H A D | protocol.h | 152 __be32 min_level; /* minimum security level */ member in struct:rxkad_challenge
|
H A D | rxkad.c | 642 challenge.min_level = htonl(0); 795 u32 version, nonce, min_level; local 816 min_level = ntohl(challenge.min_level); 818 trace_rxrpc_rx_challenge(conn, sp->hdr.serial, version, nonce, min_level); 824 if (conn->security_level < min_level)
|
/linux-master/kernel/ |
H A D | params.c | 122 s16 min_level, 132 if (params[i].level < min_level 165 s16 min_level, 187 min_level, max_level, arg, unknown); 117 parse_one(char *param, char *val, const char *doing, const struct kernel_param *params, unsigned num_params, s16 min_level, s16 max_level, void *arg, parse_unknown_fn handle_unknown) argument 161 parse_args(const char *doing, char *args, const struct kernel_param *params, unsigned num, s16 min_level, s16 max_level, void *arg, parse_unknown_fn unknown) argument
|
/linux-master/drivers/firmware/arm_scmi/ |
H A D | perf.c | 93 __le32 min_level; member in struct:scmi_perf_set_limits 98 __le32 min_level; member in struct:scmi_perf_get_limits 530 limits->min_level = cpu_to_le32(min_perf); 614 *min_perf = le32_to_cpu(limits->min_level);
|
/linux-master/include/trace/events/ |
H A D | rxrpc.h | 960 u32 version, u32 nonce, u32 min_level), 962 TP_ARGS(conn, serial, version, nonce, min_level), 969 __field(u32, min_level) 977 __entry->min_level = min_level; 985 __entry->min_level)
|
/linux-master/drivers/gpu/drm/amd/pm/powerplay/hwmgr/ |
H A D | vega12_hwmgr.c | 779 uint32_t min_level; 788 min_level = dpm_table->count - 2; 790 min_level = 1; 792 min_level = 0; 795 dpm_table->dpm_levels[min_level].value;
|
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_bios.c | 1067 u16 min_level; local 1071 min_level = backlight_data->brightness_min_level[panel_type].level; 1079 min_level = min_level / 255; 1081 if (min_level > 255) { 1085 panel->vbt.backlight.min_brightness = min_level;
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | fs_core.c | 1072 int min_level = INT_MAX; local 1077 min_level = root->root_ft->level; 1079 if (ft->level >= min_level)
|