/linux-master/kernel/ |
H A D | static_call.c | 2 #include <linux/static_call.h>
|
/linux-master/arch/arm64/include/asm/ |
H A D | paravirt.h | 18 return static_call(pv_steal_clock)(cpu);
|
/linux-master/arch/arm/include/asm/ |
H A D | paravirt.h | 18 return static_call(pv_steal_clock)(cpu);
|
/linux-master/arch/arm/kernel/ |
H A D | paravirt.c | 12 #include <linux/static_call.h>
|
/linux-master/arch/riscv/include/asm/ |
H A D | paravirt.h | 18 return static_call(pv_steal_clock)(cpu);
|
/linux-master/tools/objtool/include/objtool/ |
H A D | builtin.h | 25 bool static_call; member in struct:opts
|
/linux-master/include/linux/ |
H A D | static_call_types.h | 90 #define static_call(name) __static_call(name) macro 98 #define static_call(name) \ macro
|
H A D | static_call.h | 24 * static_call(name)(args...); 41 * static_call(my_name)(arg1, arg2); 47 * static_call(my_name)(arg1, arg2); 55 * Each static_call() site calls into a trampoline associated with the name. 62 * compiler plugin to detect all the static_call() sites and annotate them 72 * static_call()s (although when HAVE_STATIC_CALL it might not be immediately 73 * fatal). A NULL static_call can be the result of: 95 * When calling a static_call that can be NULL, use: 129 * trampoline with the result that a module can use static_call{,_cond}() but 139 #include <asm/static_call [all...] |
H A D | entry-common.h | 510 #define irqentry_exit_cond_resched() static_call(irqentry_exit_cond_resched)()
|
/linux-master/tools/include/linux/ |
H A D | static_call_types.h | 90 #define static_call(name) __static_call(name) macro 98 #define static_call(name) \ macro
|
/linux-master/arch/powerpc/kernel/ |
H A D | static_call.c | 3 #include <linux/static_call.h>
|
/linux-master/arch/x86/include/asm/ |
H A D | apic.h | 6 #include <linux/static_call.h> 404 return static_call(apic_call_read)(reg); 409 static_call(apic_call_write)(reg, val); 414 static_call(apic_call_eoi)(); 419 static_call(apic_call_native_eoi)(); 424 return static_call(apic_call_icr_read)(); 429 static_call(apic_call_icr_write)(low, high); 434 static_call(apic_call_send_IPI)(cpu, vector); 444 static_call(apic_call_send_IPI_mask_allbutself)(mask, vector); 449 static_call(apic_call_send_IPI_allbutsel [all...] |
/linux-master/arch/x86/kvm/ |
H A D | smm.c | 203 static_call(kvm_x86_get_gdt)(vcpu, &dt); 207 static_call(kvm_x86_get_idt)(vcpu, &dt); 223 smram->int_shadow = static_call(kvm_x86_get_interrupt_shadow)(vcpu); 253 static_call(kvm_x86_get_idt)(vcpu, &dt); 259 static_call(kvm_x86_get_gdt)(vcpu, &dt); 270 smram->int_shadow = static_call(kvm_x86_get_interrupt_shadow)(vcpu); 300 if (static_call(kvm_x86_enter_smm)(vcpu, &smram)) 308 if (static_call(kvm_x86_get_nmi_mask)(vcpu)) 311 static_call(kvm_x86_set_nmi_mask)(vcpu, true); 316 static_call(kvm_x86_set_interrupt_shado [all...] |
H A D | kvm_cache_regs.h | 101 static_call(kvm_x86_cache_reg)(vcpu, reg); 141 static_call(kvm_x86_cache_reg)(vcpu, VCPU_EXREG_PDPTR); 156 static_call(kvm_x86_cache_reg)(vcpu, VCPU_EXREG_CR0); 178 static_call(kvm_x86_cache_reg)(vcpu, VCPU_EXREG_CR4); 193 static_call(kvm_x86_cache_reg)(vcpu, VCPU_EXREG_CR3);
|
H A D | pmu.c | 545 return static_call(kvm_x86_pmu_check_rdpmc_early)(vcpu, idx); 594 pmc = static_call(kvm_x86_pmu_rdpmc_ecx_to_pmc)(vcpu, idx, &mask); 599 (static_call(kvm_x86_get_cpl)(vcpu) != 0) && 625 return static_call(kvm_x86_pmu_msr_idx_to_pmc)(vcpu, msr) || 626 static_call(kvm_x86_pmu_is_valid_msr)(vcpu, msr); 632 struct kvm_pmc *pmc = static_call(kvm_x86_pmu_msr_idx_to_pmc)(vcpu, msr); 657 return static_call(kvm_x86_pmu_get_msr)(vcpu, msr_info); 716 return static_call(kvm_x86_pmu_set_msr)(vcpu, msr_info); 781 static_call(kvm_x86_pmu_refresh)(vcpu); 799 static_call(kvm_x86_pmu_ini [all...] |
H A D | x86.c | 842 if (static_call(kvm_x86_get_cpl)(vcpu) <= required_cpl) 926 return static_call(kvm_x86_is_valid_cr0)(vcpu, cr0); 990 static_call(kvm_x86_get_cs_db_l_bits)(vcpu, &cs_db, &cs_l); 1004 static_call(kvm_x86_set_cr0)(vcpu, cr0); 1122 if (static_call(kvm_x86_get_cpl)(vcpu) != 0 || 1147 static_call(kvm_x86_is_valid_cr4)(vcpu, cr4); 1215 static_call(kvm_x86_set_cr4)(vcpu, cr4); 1354 static_call(kvm_x86_set_dr7)(vcpu, dr7); 1697 return static_call(kvm_x86_get_msr_feature)(msr); 1771 r = static_call(kvm_x86_set_efe [all...] |
H A D | mmu.h | 164 static_call(kvm_x86_load_mmu_pgd)(vcpu, root_hpa, 200 unsigned long rflags = static_call(kvm_x86_get_rflags)(vcpu);
|
/linux-master/security/keys/trusted-keys/ |
H A D | trusted_core.c | 23 #include <linux/static_call.h> 181 ret = static_call(trusted_key_unseal)(payload, datablob); 188 ret = static_call(trusted_key_get_random)(payload->key, 199 ret = static_call(trusted_key_seal)(payload, datablob); 268 ret = static_call(trusted_key_seal)(new_p, datablob);
|
/linux-master/tools/objtool/ |
H A D | builtin-check.c | 81 OPT_BOOLEAN('t', "static-call", &opts.static_call, "annotate static calls"), 144 opts.static_call ||
|
/linux-master/arch/arm64/kernel/ |
H A D | paravirt.c | 21 #include <linux/static_call.h>
|
/linux-master/arch/riscv/kernel/ |
H A D | paravirt.c | 17 #include <linux/static_call.h>
|
/linux-master/drivers/xen/ |
H A D | time.c | 10 #include <linux/static_call.h>
|
/linux-master/kernel/bpf/ |
H A D | dispatcher.c | 7 #include <linux/static_call.h>
|
/linux-master/arch/x86/kernel/ |
H A D | static_call.c | 2 #include <linux/static_call.h> 138 pr_err("unexpected static_call insn opcode 0x%x at %pS\n", opcode, insn); 181 * The problem is that static_call() is available before determining 191 * Not all .return_sites are a static_call trampoline (most are not).
|
/linux-master/arch/x86/events/ |
H A D | core.c | 31 #include <linux/static_call.h> 60 * This here uses DEFINE_STATIC_CALL_NULL() to get a static_call defined 700 return static_call(x86_pmu_guest_get_msrs)(nr, data); 731 static_call(x86_pmu_disable_all)(); 1013 c = static_call(x86_pmu_get_event_constraints)(cpuc, i, cpuc->event_list[i]); 1356 static_call(x86_pmu_enable_all)(added); 1466 ret = static_call(x86_pmu_schedule_events)(cpuc, n, assign); 1508 static_call(x86_pmu_set_period)(event); 1515 static_call(x86_pmu_enable)(event); 1588 static_call(x86_pmu_disabl [all...] |