Searched refs:__always_inline (Results 1 - 25 of 730) sorted by last modified time
1234567891011>>
/linux-master/virt/kvm/ |
H A D | kvm_main.c | 594 static __always_inline kvm_mn_ret_t __kvm_handle_hva_range(struct kvm *kvm, 670 static __always_inline int kvm_handle_hva_range(struct mmu_notifier *mn, 690 static __always_inline int kvm_handle_hva_range_no_flush(struct mmu_notifier *mn, 2478 static __always_inline void kvm_handle_gfn_range(struct kvm *kvm,
|
/linux-master/kernel/sched/ |
H A D | sched.h | 1899 static __always_inline bool sched_asym_cpucap_active(void) 2102 static __always_inline bool static_branch_##name(struct static_key *key) \ 2621 static __always_inline 2638 static __always_inline
|
/linux-master/arch/x86/kvm/vmx/ |
H A D | vmx.h | 602 static __always_inline void lname##_controls_setbit(struct vcpu_vmx *vmx, u##bits val) \ 607 static __always_inline void lname##_controls_clearbit(struct vcpu_vmx *vmx, u##bits val) \ 653 static __always_inline struct kvm_vmx *to_kvm_vmx(struct kvm *kvm) 658 static __always_inline struct vcpu_vmx *to_vmx(struct kvm_vcpu *vcpu) 682 static __always_inline unsigned long vmx_get_exit_qual(struct kvm_vcpu *vcpu) 692 static __always_inline u32 vmx_get_intr_info(struct kvm_vcpu *vcpu)
|
H A D | vmx.c | 371 static __always_inline void vmx_disable_fb_clear(struct vcpu_vmx *vmx) 385 static __always_inline void vmx_enable_fb_clear(struct vcpu_vmx *vmx) 958 static __always_inline void clear_atomic_switch_msr_special(struct vcpu_vmx *vmx, 1016 static __always_inline void add_atomic_switch_msr_special(struct vcpu_vmx *vmx, 5340 static __always_inline int handle_external_interrupt(struct kvm_vcpu *vcpu)
|
H A D | pmu_intel.c | 430 static __always_inline u64 intel_get_fixed_pmc_eventsel(unsigned int index)
|
/linux-master/arch/x86/kvm/svm/ |
H A D | svm.h | 316 static __always_inline struct kvm_svm *to_kvm_svm(struct kvm *kvm) 321 static __always_inline bool sev_guest(struct kvm *kvm) 332 static __always_inline bool sev_es_guest(struct kvm *kvm) 364 static __always_inline struct vcpu_svm *to_svm(struct kvm_vcpu *vcpu) 706 static __always_inline bool kvm_ghcb_##field##_is_valid(const struct vcpu_svm *svm) \ 712 static __always_inline u64 kvm_ghcb_get_##field##_if_valid(struct vcpu_svm *svm, struct ghcb *ghcb) \
|
/linux-master/arch/x86/kvm/mmu/ |
H A D | tdp_mmu.c | 22 static __always_inline bool kvm_lockdep_assert_mmu_lock_held(struct kvm *kvm, 1179 static __always_inline bool kvm_tdp_mmu_handle_gfn(struct kvm *kvm,
|
H A D | mmu.c | 1567 static __always_inline bool kvm_handle_gfn_range(struct kvm *kvm, 6056 static __always_inline bool __walk_slot_rmaps(struct kvm *kvm, 6085 static __always_inline bool walk_slot_rmaps(struct kvm *kvm, 6096 static __always_inline bool walk_slot_rmaps_4k(struct kvm *kvm,
|
/linux-master/arch/x86/kvm/ |
H A D | lapic.c | 82 static __always_inline u64 __kvm_lapic_get_reg64(char *regs, int reg) 88 static __always_inline u64 kvm_lapic_get_reg64(struct kvm_lapic *apic, int reg) 93 static __always_inline void __kvm_lapic_set_reg64(char *regs, int reg, u64 val) 99 static __always_inline void kvm_lapic_set_reg64(struct kvm_lapic *apic,
|
H A D | cpuid.c | 561 static __always_inline void __kvm_cpu_cap_mask(unsigned int leaf) 574 static __always_inline 585 static __always_inline void kvm_cpu_cap_mask(enum cpuid_leafs leaf, u32 mask)
|
H A D | cpuid.h | 61 static __always_inline void cpuid_entry_override(struct kvm_cpuid_entry2 *entry, 70 static __always_inline u32 *guest_cpuid_get_register(struct kvm_vcpu *vcpu, 83 static __always_inline bool guest_cpuid_has(struct kvm_vcpu *vcpu, 95 static __always_inline void guest_cpuid_clear(struct kvm_vcpu *vcpu, 197 static __always_inline void kvm_cpu_cap_clear(unsigned int x86_feature) 205 static __always_inline void kvm_cpu_cap_set(unsigned int x86_feature) 213 static __always_inline u32 kvm_cpu_cap_get(unsigned int x86_feature) 221 static __always_inline bool kvm_cpu_cap_has(unsigned int x86_feature) 226 static __always_inline void kvm_cpu_cap_check_and_set(unsigned int x86_feature) 232 static __always_inline boo [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | perf_event.h | 618 static __always_inline void perf_lopwr_cb(bool lopwr_in)
|
H A D | barrier.h | 36 static __always_inline unsigned long array_index_mask_nospec(unsigned long index,
|
/linux-master/arch/x86/events/intel/ |
H A D | lbr.c | 107 static __always_inline bool is_lbr_call_stack_bit_set(u64 config) 281 static __always_inline void wrlbr_from(unsigned int idx, u64 val) 287 static __always_inline void wrlbr_to(unsigned int idx, u64 val) 292 static __always_inline void wrlbr_info(unsigned int idx, u64 val) 297 static __always_inline u64 rdlbr_from(unsigned int idx, struct lbr_entry *lbr) 309 static __always_inline u64 rdlbr_to(unsigned int idx, struct lbr_entry *lbr) 321 static __always_inline u64 rdlbr_info(unsigned int idx, struct lbr_entry *lbr) 416 static __always_inline bool lbr_is_reset_in_cstate(void *ctx) 856 static __always_inline int get_lbr_br_type(u64 info) 866 static __always_inline boo [all...] |
/linux-master/arch/x86/entry/ |
H A D | entry_fred.c | 78 static __always_inline void fred_other(struct pt_regs *regs)
|
H A D | common.c | 42 static __always_inline bool do_syscall_x64(struct pt_regs *regs, int nr) 58 static __always_inline bool do_syscall_x32(struct pt_regs *regs, int nr) 134 static __always_inline int syscall_32_enter(struct pt_regs *regs) 155 static __always_inline void do_syscall_32_irqs_on(struct pt_regs *regs, int nr) 172 static __always_inline bool int80_is_external(void) 477 static __always_inline bool get_and_clear_inhcall(void) 485 static __always_inline void restore_inhcall(bool inhcall) 490 static __always_inline bool get_and_clear_inhcall(void) { return false; } 491 static __always_inline void restore_inhcall(bool inhcall) { }
|
/linux-master/tools/include/asm-generic/bitops/ |
H A D | fls.h | 13 static __always_inline int generic_fls(unsigned int x)
|
H A D | __fls.h | 13 static __always_inline unsigned long generic___fls(unsigned long word)
|
/linux-master/net/core/ |
H A D | dev.c | 3953 static __always_inline enum tcx_action_base 3974 static __always_inline struct sk_buff * 4031 static __always_inline struct sk_buff * 4075 static __always_inline struct sk_buff * 4082 static __always_inline struct sk_buff *
|
/linux-master/mm/ |
H A D | page_owner.c | 89 static __always_inline depot_stack_handle_t create_dummy_stack(void)
|
H A D | internal.h | 1215 static __always_inline void vma_set_range(struct vm_area_struct *vma,
|
H A D | gup.c | 1482 static __always_inline long __get_user_pages_locked(struct mm_struct *mm,
|
/linux-master/kernel/ |
H A D | fork.c | 1231 static __always_inline void mm_clear_owner(struct mm_struct *mm, 2087 static __always_inline void delayed_free_task(struct task_struct *tsk)
|
/linux-master/include/linux/ |
H A D | shmem_fs.h | 117 static __always_inline bool shmem_is_huge(struct inode *inode, pgoff_t index, bool shmem_huge_force,
|
H A D | rwsem.h | 165 static __always_inline int rwsem_is_locked(const struct rw_semaphore *sem) 170 static __always_inline void rwsem_assert_held_nolockdep(const struct rw_semaphore *sem) 175 static __always_inline void rwsem_assert_held_write_nolockdep(const struct rw_semaphore *sem) 180 static __always_inline int rwsem_is_contended(struct rw_semaphore *sem)
|
Completed in 416 milliseconds
1234567891011>>