/linux-master/arch/sparc/kernel/ |
H A D | head_64.S | 183 ldx [%sp + 2047 + 128 + 0x20], %l4 ! prom root node
|
H A D | ioport.c | 322 struct resource *root = m->private, *r; local 325 for (r = root->child; r != NULL; r = r->sibling) {
|
H A D | of_device_32.c | 381 dev_set_name(&op->dev, "root"); 411 struct device_node *root = of_find_node_by_path("/"); local 414 parent = scan_one_device(root, NULL); 418 scan_tree(root->child, &parent->dev);
|
H A D | of_device_64.c | 167 * #size-cells properties, and if you walk to the root on such 290 * the translation up to the root. 557 /* Something more complicated. Walk up to the root, applying 675 dev_set_name(&op->dev, "root"); 704 struct device_node *root = of_find_node_by_path("/"); local 707 parent = scan_one_device(root, NULL); 711 scan_tree(root->child, &parent->dev);
|
H A D | pci.c | 625 struct resource *p, *root, *conflict; local 641 root = pci_find_parent_resource(dev, p); 642 if (!root) { 647 conflict = request_resource_conflict(root, p);
|
H A D | vio.c | 506 u64 root; local 519 root = mdesc_node_by_name(hp, MDESC_NODE_NULL, channel_devices_node); 520 if (root == MDESC_NODE_NULL) { 533 compat = mdesc_get_property(hp, root, "compatible", &len); 545 cfg_handle = mdesc_get_property(hp, root, cfg_handle_prop, NULL); 554 root_vdev = vio_create_one(hp, root, NULL, NULL); 557 printk(KERN_ERR "VIO: Could not create root device.\n");
|
/linux-master/arch/x86/entry/vsyscall/ |
H A D | vsyscall_64.c | 334 void __init set_vsyscall_pgtable_user_bits(pgd_t *root) argument 341 pgd = pgd_offset_pgd(root, VSYSCALL_ADDR);
|
/linux-master/arch/x86/include/asm/ |
H A D | amd_nb.h | 62 struct pci_dev *root; member in struct:amd_northbridge
|
H A D | kvm_host.h | 460 struct kvm_mmu_root_info root; member in struct:kvm_mmu 1449 * count to zero should removed the root from the list and clean 1450 * it up, freeing the root after an RCU grace period. 1471 * If set, at least one shadow root has been allocated. This flag
|
H A D | vsyscall.h | 11 extern void set_vsyscall_pgtable_user_bits(pgd_t *root);
|
/linux-master/arch/x86/kernel/ |
H A D | amd_nb.c | 185 struct pci_dev *root; local 191 root = node_to_amd_nb(node)->root; 192 if (!root) 197 err = pci_write_config_dword(root, 0x60, address); 203 err = (write ? pci_write_config_dword(root, 0x64, *value) 204 : pci_read_config_dword(root, 0x64, value)); 234 struct pci_dev *root, *misc, *link; local 257 root = NULL; 258 while ((root [all...] |
/linux-master/arch/x86/kernel/cpu/resctrl/ |
H A D | rdtgroup.c | 52 /* Kernel fs node for "info" directory under root */ 55 /* Kernel fs node for "mon_groups" directory under root */ 58 /* Kernel fs node for "mon_data" directory under root */ 863 * Task is part of the root resctrl control group, and it is not associated 869 * Task is part of the root resctrl control group and monitor group mon0. 2415 * except the root directory. Thus when looking for the rdtgroup 2903 * Forcibly remove all of subdirectories under root. 2916 /* Remove each rdtgroup other than root */ 3509 * These are rdtgroups created under the root directory. Can be used 3598 * If the parent directory is the root director [all...] |
/linux-master/arch/x86/kvm/ |
H A D | kvm_onhyperv.c | 44 hpa_t root; local 53 * on a common root, in which case future flushes can skip the 54 * loop and flush the common root. 57 root = vcpu->arch.hv_root_tdp; 58 if (!VALID_PAGE(root) || root == kvm_arch->hv_root_tdp) 62 * Set the tracked root to the first valid root. Keep 63 * this root for the entirety of the loop even if more 65 * to avoid flushing the same (first) root agai [all...] |
H A D | mmu.h | 131 if (likely(vcpu->arch.mmu->root.hpa != INVALID_PAGE)) 161 u64 root_hpa = vcpu->arch.mmu->root.hpa;
|
/linux-master/arch/x86/kvm/mmu/ |
H A D | mmu.c | 1883 * level tracks the root level 2324 struct kvm_vcpu *vcpu, hpa_t root, 2328 iterator->shadow_addr = root; 2341 BUG_ON(root != vcpu->arch.mmu->root.hpa); 2355 shadow_walk_init_using_root(iterator, vcpu, vcpu->arch.mmu->root.hpa, 2564 * Remove the active root from the active page list, the root 2583 * Make the request to free obsolete roots after marking the root 2640 * Don't zap active root page 2323 shadow_walk_init_using_root(struct kvm_shadow_walk_iterator *iterator, struct kvm_vcpu *vcpu, hpa_t root, u64 addr) argument 3672 hpa_t root; local 3782 hpa_t root; local 3966 is_unsync_root(hpa_t root) argument 4015 hpa_t root = vcpu->arch.mmu->root.hpa; local 4031 hpa_t root = vcpu->arch.mmu->pae_root[i]; local 4109 int root, leaf, level; local 4663 is_root_usable(struct kvm_mmu_root_info *root, gpa_t pgd, union kvm_mmu_page_role role) argument [all...] |
H A D | paging_tmpl.h | 644 * is a root page, it is either write protected (and cannot be 651 if (WARN_ON_ONCE(!VALID_PAGE(vcpu->arch.mmu->root.hpa))) 655 * Load a new root and retry the faulting instruction in the extremely 656 * unlikely scenario that the guest root gfn became visible between 657 * loading a dummy root and handling the resulting page fault, e.g. if 660 if (unlikely(kvm_mmu_is_dummy_root(vcpu->arch.mmu->root.hpa))) {
|
H A D | spte.h | 256 static inline struct kvm_mmu_page *root_to_sp(hpa_t root) argument 258 if (kvm_mmu_is_dummy_root(root)) 262 * The "root" may be a special root, e.g. a PAE entry, treat it as a 265 return spte_to_child_sp(root);
|
H A D | tdp_iter.c | 20 * Return the TDP iterator to the root PT and allow it to continue its 39 void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root, argument 42 if (WARN_ON_ONCE(!root || (root->role.level < 1) || 43 (root->role.level > PT64_ROOT_MAX_LEVEL))) { 49 iter->root_level = root->role.level; 51 iter->pt_path[iter->root_level - 1] = (tdp_ptep_t)root->spt; 52 iter->as_id = kvm_mmu_page_as_id(root); 131 * successful step up, false if already in the root page.
|
H A D | tdp_iter.h | 96 /* The level of the root page given to the iterator */ 113 * which case tdp_iter_next() needs to restart the walk at the root 123 #define for_each_tdp_pte_min_level(iter, root, min_level, start, end) \ 124 for (tdp_iter_start(&iter, root, min_level, start); \ 128 #define for_each_tdp_pte(iter, root, start, end) \ 129 for_each_tdp_pte_min_level(iter, root, PG_LEVEL_4K, start, end) 133 void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root,
|
H A D | tdp_mmu.c | 37 * for zapping and thus puts the TDP MMU's reference to each root, i.e. 76 void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root) argument 78 if (!refcount_dec_and_test(&root->tdp_mmu_root_count)) 82 * The TDP MMU itself holds a reference to each root until the root is 84 * put for a valid root. 86 KVM_BUG_ON(!is_tdp_mmu_page(root) || !root->role.invalid, kvm); 89 list_del_rcu(&root->link); 91 call_rcu(&root 232 struct kvm_mmu_page *root; local 746 __tdp_mmu_zap_root(struct kvm *kvm, struct kvm_mmu_page *root, bool shared, int zap_level) argument 772 tdp_mmu_zap_root(struct kvm *kvm, struct kvm_mmu_page *root, bool shared) argument 847 tdp_mmu_zap_leafs(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t start, gfn_t end, bool can_yield, bool flush) argument 895 struct kvm_mmu_page *root; local 906 struct kvm_mmu_page *root; local 931 struct kvm_mmu_page *root; local 976 struct kvm_mmu_page *root; local 1190 struct kvm_mmu_page *root; local 1206 struct kvm_mmu_page *root; local 1289 wrprot_gfn_range(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t start, gfn_t end, int min_level) argument 1330 struct kvm_mmu_page *root; local 1441 tdp_mmu_split_huge_pages_root(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t start, gfn_t end, int target_level, bool shared) argument 1515 struct kvm_mmu_page *root; local 1538 clear_dirty_gfn_range(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t start, gfn_t end) argument 1581 struct kvm_mmu_page *root; local 1592 clear_dirty_pt_masked(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t gfn, unsigned long mask, bool wrprot) argument 1643 struct kvm_mmu_page *root; local 1649 zap_collapsible_spte_range(struct kvm *kvm, struct kvm_mmu_page *root, const struct kvm_memory_slot *slot) argument 1707 struct kvm_mmu_page *root; local 1719 write_protect_gfn(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t gfn, int min_level) argument 1759 struct kvm_mmu_page *root; local [all...] |
H A D | tdp_mmu.h | 15 __must_check static inline bool kvm_tdp_mmu_get_root(struct kvm_mmu_page *root) argument 17 return refcount_inc_not_zero(&root->tdp_mmu_root_count); 20 void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root);
|
/linux-master/arch/x86/kvm/svm/ |
H A D | svm.c | 3914 hpa_t root_tdp = vcpu->arch.mmu->root.hpa;
|
/linux-master/arch/x86/kvm/vmx/ |
H A D | nested.c | 3444 * NVMX_VMENTRY_SUCCESS: Entered VMX non-root mode 3561 * when it executed VMLAUNCH/VMRESUME, as entering non-root mode can 4881 * Bare metal isolates VMX root (host) from VMX non-root (guest), but 5239 * The CPL is checked for "not in VMX operation" and for "in VMX root", 5241 * i.e. VMXON #GPs outside of VMX non-root if CPL!=0. In VMX non-root, 5244 * VMX non-root. 5421 * In VMX non-root operation, when the VMCS-link pointer is INVALID_GPA, 5531 * In VMX non-root operatio [all...] |
H A D | vmx.c | 3175 * associated with the root EPT structure and not any particular VPID 3200 u64 root_hpa = mmu->root.hpa; 4178 * Case 1: vCPU stays in non-root mode. Sending a notification 4181 * Case 2: vCPU exits to root mode and is still runnable. The 4186 * Case 3: vCPU exits to root mode and is blocked. vcpu_block() 4240 * 1. If target vcpu is running(non-root mode), send posted interrupt 4242 * 2. If target vcpu isn't running(root mode), kick it to pick up the 6465 * mode as if vcpus is in root mode, the PML buffer must has been
|
/linux-master/arch/x86/pci/ |
H A D | acpi.c | 277 struct acpi_pci_root *root = ci->root; local 281 info->start_bus = (u8)root->secondary.start; 282 info->end_bus = (u8)root->secondary.end; 287 &root->secondary, &root->mcfg_addr); 297 root->mcfg_addr); 332 static int pci_acpi_root_get_node(struct acpi_pci_root *root) argument 334 int busnum = root->secondary.start; 335 struct acpi_device *device = root 414 pci_acpi_scan_root(struct acpi_pci_root *root) argument [all...] |