/linux-master/arch/powerpc/kvm/ |
H A D | e500_mmu_host.c | 327 struct kvm_memory_slot *slot; local 350 * have gotten this far. Eventually we should just pass the slot 353 slot = gfn_to_memslot(vcpu_e500->vcpu.kvm, gfn); 354 hva = gfn_to_hva_memslot(slot, gfn); 381 slot_start = pfn - (gfn - slot->base_gfn); 382 slot_end = slot_start + slot->npages; 449 pfn = gfn_to_pfn_memslot(slot, gfn);
|
H A D | book3s_pr.c | 1863 * Get (and clear) the dirty memory log for a memory slot. 1921 static void kvmppc_core_free_memslot_pr(struct kvm_memory_slot *slot) argument
|
H A D | book3s_hv.c | 5093 * Get (and clear) the dirty memory log for a memory slot. 5108 if (log->slot >= KVM_USER_MEM_SLOTS) 5112 memslot = id_to_memslot(slots, log->slot); 5161 static void kvmppc_core_free_memslot_hv(struct kvm_memory_slot *slot) argument 5163 vfree(slot->arch.rmap); 5164 slot->arch.rmap = NULL; 5343 /* Look up the VMA for the start of this memory slot */
|
H A D | book3s_64_mmu_hv.c | 863 kvm_unmap_radix(kvm, range->slot, gfn); 866 kvm_unmap_rmapp(kvm, range->slot, gfn); 961 ret |= kvm_age_radix(kvm, range->slot, gfn); 964 ret |= kvm_age_rmapp(kvm, range->slot, gfn); 1008 return kvm_test_age_radix(kvm, range->slot, range->start); 1010 return kvm_test_age_rmapp(kvm, range->slot, range->start);
|
H A D | book3s.c | 857 void kvmppc_core_free_memslot(struct kvm *kvm, struct kvm_memory_slot *slot) argument 859 kvm->arch.kvm_ops->free_memslot(slot);
|
/linux-master/arch/powerpc/kernel/ |
H A D | eeh_driver.c | 360 * eeh_report_reset - Tell device that slot has been reset 613 * eeh_reset_device - Perform actual reset of a pci slot 616 * @bus: PCI bus corresponding to the isolcated slot 749 * eeh_check_slot_presence - Check if a device is still present in a slot 752 * This function may return a false positive if we can't determine the slot's 763 struct pci_slot *slot; local 773 slot = pdev->slot; 774 if (!slot || !slot 795 struct pci_slot *slot; local [all...] |
/linux-master/arch/powerpc/include/asm/ |
H A D | kvm_ppc.h | 205 struct kvm_memory_slot *slot); 290 void (*free_memslot)(struct kvm_memory_slot *slot);
|
/linux-master/arch/parisc/mm/ |
H A D | init.c | 726 static void alloc_btlb(unsigned long start, unsigned long end, int *slot, argument 738 while (start < end && *slot < slot_max && size >= PAGE_SIZE) { 754 size >> PAGE_SHIFT, entry_info, *slot); 755 (*slot)++; 767 int slot; local 783 slot = 0; 784 alloc_btlb(s, e, &slot, 0x13800000); 794 alloc_btlb(s, e, &slot, 0x11800000);
|
/linux-master/arch/mips/kvm/ |
H A D | mmu.c | 406 * @slot: The memory slot associated with mask 407 * @gfn_offset: The gfn offset in memory slot 409 * slot to be write protected 415 struct kvm_memory_slot *slot, 418 gfn_t base_gfn = slot->base_gfn + gfn_offset; 414 kvm_arch_mmu_enable_log_dirty_pt_masked(struct kvm *kvm, struct kvm_memory_slot *slot, gfn_t gfn_offset, unsigned long mask) argument
|
/linux-master/arch/loongarch/kvm/ |
H A D | mmu.c | 16 static inline bool kvm_hugepage_capable(struct kvm_memory_slot *slot) argument 18 return slot->arch.flags & KVM_MEM_HUGEPAGE_CAPABLE; 21 static inline bool kvm_hugepage_incapable(struct kvm_memory_slot *slot) argument 23 return slot->arch.flags & KVM_MEM_HUGEPAGE_INCAPABLE; 353 * @slot: The memory slot associated with mask 354 * @gfn_offset: The gfn offset in memory slot 356 * slot to be write protected 362 struct kvm_memory_slot *slot, gfn_t gfn_offset, unsigned long mask) 365 gfn_t base_gfn = slot 361 kvm_arch_mmu_enable_log_dirty_pt_masked(struct kvm *kvm, struct kvm_memory_slot *slot, gfn_t gfn_offset, unsigned long mask) argument 475 kvm_arch_flush_shadow_memslot(struct kvm *kvm, struct kvm_memory_slot *slot) argument 542 struct kvm_memory_slot *slot; local 663 host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn, const struct kvm_memory_slot *slot) argument [all...] |
/linux-master/arch/loongarch/include/asm/ |
H A D | kvm_host.h | 268 static inline void kvm_arch_free_memslot(struct kvm *kvm, struct kvm_memory_slot *slot) {} argument
|
/linux-master/arch/arm64/kvm/ |
H A D | mmu.c | 1116 * kvm_mmu_wp_memory_region() - write protect stage 2 entries for memory slot 1118 * @slot: The memory slot to write protect 1128 static void kvm_mmu_wp_memory_region(struct kvm *kvm, int slot) argument 1131 struct kvm_memory_slot *memslot = id_to_memslot(slots, slot); 1148 * pages for memory slot 1150 * @slot: The memory slot to split 1155 static void kvm_mmu_split_memory_region(struct kvm *kvm, int slot) argument 1164 memslot = id_to_memslot(slots, slot); 1185 kvm_arch_mmu_enable_log_dirty_pt_masked(struct kvm *kvm, struct kvm_memory_slot *slot, gfn_t gfn_offset, unsigned long mask) argument 2017 kvm_arch_free_memslot(struct kvm *kvm, struct kvm_memory_slot *slot) argument 2030 kvm_arch_flush_shadow_memslot(struct kvm *kvm, struct kvm_memory_slot *slot) argument [all...] |
H A D | arm.c | 1934 /* A lookup table holding the hypervisor VA for each vector slot */ 1937 static void kvm_init_vector_slot(void *base, enum arm64_hyp_spectre_vector slot) argument 1939 hyp_spectre_vector_selector[slot] = __kvm_vector_slot2addr(base, slot); 2062 * - If the CPU also has the ARM64_SPECTRE_V3A cap, the slot 2067 * empty slot is selected, mapped next to the idmap page, and 2077 void *vector = hyp_spectre_vector_selector[data->slot]; 2082 kvm_call_hyp_nvhe(__pkvm_cpu_set_vector, data->slot);
|
/linux-master/arch/arm64/kvm/hyp/nvhe/ |
H A D | hyp-main.c | 221 DECLARE_REG(enum arm64_hyp_spectre_vector, slot, host_ctxt, 1); 223 cpu_reg(host_ctxt, 1) = pkvm_cpu_set_vector(slot);
|
/linux-master/arch/arm64/kernel/probes/ |
H A D | kprobes.c | 49 * Prepare insn slot, Mark Rutland points out it depends on a coupe of 59 * fetch thesingle-step slot instructions *after* this, ensuring that 63 * is observed on all CPUS, however, single-step slot is installed in 116 case INSN_GOOD: /* instruction uses slot */ 197 unsigned long slot; local 210 slot = (unsigned long)p->ainsn.api.insn; 213 instruction_pointer_set(regs, slot);
|
/linux-master/arch/arm/kernel/ |
H A D | hw_breakpoint.c | 287 pr_warn("unknown slot type: %d\n", type); 329 struct perf_event **slot, **slots; local 351 slot = &slots[i]; 353 if (!*slot) { 354 *slot = bp; 360 pr_warn("Can't find any breakpoint slot\n"); 386 struct perf_event **slot, **slots; local 403 slot = &slots[i]; 405 if (*slot == bp) { 406 *slot [all...] |
/linux-master/tools/testing/selftests/bpf/benchs/ |
H A D | bench_trigger.c | 55 /* adjust slot shift in inc_hits() if changing */ 72 unsigned slot; local 78 slot = 2654435769U * tid; 79 slot >>= 24; 81 atomic_inc(&base_hits[slot].value); /* use highest byte as an index */
|
/linux-master/tools/lib/bpf/ |
H A D | bpf_helpers.h | 138 * Helper function to perform a tail call with a constant/immediate map slot. 143 bpf_tail_call_static(void *ctx, const void *map, const __u32 slot) argument 145 if (!__builtin_constant_p(slot)) 163 "r3 = %[slot]\n\t" 165 :: [ctx]"r"(ctx), [map]"r"(map), [slot]"i"(slot)
|
/linux-master/net/sched/ |
H A D | sch_sfq.c | 104 sfq_index next; /* next slot in sfq RR chain */ 107 short allot; /* credit for this slot */ 121 u8 cur_depth; /* depth of longest slot */ 131 struct sfq_slot *tail; /* current slot in round */ 201 * x : slot number [0 .. SFQ_MAX_FLOWS - 1] 206 struct sfq_slot *slot = &q->slots[x]; local 207 int qlen = slot->qlen; 212 slot->dep.next = n; 213 slot->dep.prev = p; 256 /* remove one skb from tail of slot queu 257 slot_dequeue_tail(struct sfq_slot *slot) argument 268 slot_dequeue_head(struct sfq_slot *slot) argument 278 slot_queue_init(struct sfq_slot *slot) argument 285 slot_queue_add(struct sfq_slot *slot, struct sk_buff *skb) argument 299 struct sfq_slot *slot; local 351 struct sfq_slot *slot; local 485 struct sfq_slot *slot; local 540 struct sfq_slot *slot; local 879 const struct sfq_slot *slot = &q->slots[idx]; local [all...] |
H A D | sch_api.c | 572 int pkt_len, slot; local 578 slot = pkt_len + stab->szopts.cell_align; 579 if (unlikely(slot < 0)) 580 slot = 0; 582 slot >>= stab->szopts.cell_log; 583 if (likely(slot < stab->szopts.tsize)) 584 pkt_len = stab->data[slot]; 587 (slot / stab->szopts.tsize) + 588 stab->data[slot % stab->szopts.tsize];
|
/linux-master/net/ipv6/ |
H A D | inet6_hashtables.c | 64 unsigned int slot = hash & hashinfo->ehash_mask; local 65 struct inet_ehash_bucket *head = &hashinfo->ehash[slot]; 83 if (get_nulls_value(node) != slot)
|
/linux-master/net/ipv4/ |
H A D | udp.c | 2504 unsigned int slot; local 2506 slot = udp_hashfn(net, hnum, udptable->mask); 2507 hslot = &udptable->hash[slot]; 3432 21, /* one slot per 2 MB */
|
H A D | inet_hashtables.c | 505 unsigned int slot = hash & hashinfo->ehash_mask; local 506 struct inet_ehash_bucket *head = &hashinfo->ehash[slot]; 528 if (get_nulls_value(node) != slot)
|
H A D | inet_timewait_sock.c | 272 unsigned int slot; local 275 for (slot = 0; slot <= ehash_mask; slot++, head++) { 318 if (get_nulls_value(node) != slot)
|
H A D | fib_semantics.c | 1267 u32 slot = hash_32(net_hash_mix(net) ^ (__force u32)val, local 1270 return &fib_info_laddrhash[slot];
|