Searched refs:root (Results 101 - 125 of 1078) sorted by path

1234567891011>>

/linux-master/arch/sparc/kernel/
H A Dhead_64.S183 ldx [%sp + 2047 + 128 + 0x20], %l4 ! prom root node
H A Dioport.c322 struct resource *root = m->private, *r; local
325 for (r = root->child; r != NULL; r = r->sibling) {
H A Dof_device_32.c381 dev_set_name(&op->dev, "root");
411 struct device_node *root = of_find_node_by_path("/"); local
414 parent = scan_one_device(root, NULL);
418 scan_tree(root->child, &parent->dev);
H A Dof_device_64.c167 * #size-cells properties, and if you walk to the root on such
290 * the translation up to the root.
557 /* Something more complicated. Walk up to the root, applying
675 dev_set_name(&op->dev, "root");
704 struct device_node *root = of_find_node_by_path("/"); local
707 parent = scan_one_device(root, NULL);
711 scan_tree(root->child, &parent->dev);
H A Dpci.c625 struct resource *p, *root, *conflict; local
641 root = pci_find_parent_resource(dev, p);
642 if (!root) {
647 conflict = request_resource_conflict(root, p);
H A Dvio.c506 u64 root; local
519 root = mdesc_node_by_name(hp, MDESC_NODE_NULL, channel_devices_node);
520 if (root == MDESC_NODE_NULL) {
533 compat = mdesc_get_property(hp, root, "compatible", &len);
545 cfg_handle = mdesc_get_property(hp, root, cfg_handle_prop, NULL);
554 root_vdev = vio_create_one(hp, root, NULL, NULL);
557 printk(KERN_ERR "VIO: Could not create root device.\n");
/linux-master/arch/x86/entry/vsyscall/
H A Dvsyscall_64.c334 void __init set_vsyscall_pgtable_user_bits(pgd_t *root) argument
341 pgd = pgd_offset_pgd(root, VSYSCALL_ADDR);
/linux-master/arch/x86/include/asm/
H A Damd_nb.h62 struct pci_dev *root; member in struct:amd_northbridge
H A Dkvm_host.h460 struct kvm_mmu_root_info root; member in struct:kvm_mmu
1449 * count to zero should removed the root from the list and clean
1450 * it up, freeing the root after an RCU grace period.
1471 * If set, at least one shadow root has been allocated. This flag
H A Dvsyscall.h11 extern void set_vsyscall_pgtable_user_bits(pgd_t *root);
/linux-master/arch/x86/kernel/
H A Damd_nb.c185 struct pci_dev *root; local
191 root = node_to_amd_nb(node)->root;
192 if (!root)
197 err = pci_write_config_dword(root, 0x60, address);
203 err = (write ? pci_write_config_dword(root, 0x64, *value)
204 : pci_read_config_dword(root, 0x64, value));
234 struct pci_dev *root, *misc, *link; local
257 root = NULL;
258 while ((root
[all...]
/linux-master/arch/x86/kernel/cpu/resctrl/
H A Drdtgroup.c52 /* Kernel fs node for "info" directory under root */
55 /* Kernel fs node for "mon_groups" directory under root */
58 /* Kernel fs node for "mon_data" directory under root */
863 * Task is part of the root resctrl control group, and it is not associated
869 * Task is part of the root resctrl control group and monitor group mon0.
2415 * except the root directory. Thus when looking for the rdtgroup
2903 * Forcibly remove all of subdirectories under root.
2916 /* Remove each rdtgroup other than root */
3509 * These are rdtgroups created under the root directory. Can be used
3598 * If the parent directory is the root director
[all...]
/linux-master/arch/x86/kvm/
H A Dkvm_onhyperv.c44 hpa_t root; local
53 * on a common root, in which case future flushes can skip the
54 * loop and flush the common root.
57 root = vcpu->arch.hv_root_tdp;
58 if (!VALID_PAGE(root) || root == kvm_arch->hv_root_tdp)
62 * Set the tracked root to the first valid root. Keep
63 * this root for the entirety of the loop even if more
65 * to avoid flushing the same (first) root agai
[all...]
H A Dmmu.h131 if (likely(vcpu->arch.mmu->root.hpa != INVALID_PAGE))
161 u64 root_hpa = vcpu->arch.mmu->root.hpa;
/linux-master/arch/x86/kvm/mmu/
H A Dmmu.c1883 * level tracks the root level
2324 struct kvm_vcpu *vcpu, hpa_t root,
2328 iterator->shadow_addr = root;
2341 BUG_ON(root != vcpu->arch.mmu->root.hpa);
2355 shadow_walk_init_using_root(iterator, vcpu, vcpu->arch.mmu->root.hpa,
2564 * Remove the active root from the active page list, the root
2583 * Make the request to free obsolete roots after marking the root
2640 * Don't zap active root page
2323 shadow_walk_init_using_root(struct kvm_shadow_walk_iterator *iterator, struct kvm_vcpu *vcpu, hpa_t root, u64 addr) argument
3672 hpa_t root; local
3782 hpa_t root; local
3966 is_unsync_root(hpa_t root) argument
4015 hpa_t root = vcpu->arch.mmu->root.hpa; local
4031 hpa_t root = vcpu->arch.mmu->pae_root[i]; local
4109 int root, leaf, level; local
4663 is_root_usable(struct kvm_mmu_root_info *root, gpa_t pgd, union kvm_mmu_page_role role) argument
[all...]
H A Dpaging_tmpl.h644 * is a root page, it is either write protected (and cannot be
651 if (WARN_ON_ONCE(!VALID_PAGE(vcpu->arch.mmu->root.hpa)))
655 * Load a new root and retry the faulting instruction in the extremely
656 * unlikely scenario that the guest root gfn became visible between
657 * loading a dummy root and handling the resulting page fault, e.g. if
660 if (unlikely(kvm_mmu_is_dummy_root(vcpu->arch.mmu->root.hpa))) {
H A Dspte.h256 static inline struct kvm_mmu_page *root_to_sp(hpa_t root) argument
258 if (kvm_mmu_is_dummy_root(root))
262 * The "root" may be a special root, e.g. a PAE entry, treat it as a
265 return spte_to_child_sp(root);
H A Dtdp_iter.c20 * Return the TDP iterator to the root PT and allow it to continue its
39 void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root, argument
42 if (WARN_ON_ONCE(!root || (root->role.level < 1) ||
43 (root->role.level > PT64_ROOT_MAX_LEVEL))) {
49 iter->root_level = root->role.level;
51 iter->pt_path[iter->root_level - 1] = (tdp_ptep_t)root->spt;
52 iter->as_id = kvm_mmu_page_as_id(root);
131 * successful step up, false if already in the root page.
H A Dtdp_iter.h96 /* The level of the root page given to the iterator */
113 * which case tdp_iter_next() needs to restart the walk at the root
123 #define for_each_tdp_pte_min_level(iter, root, min_level, start, end) \
124 for (tdp_iter_start(&iter, root, min_level, start); \
128 #define for_each_tdp_pte(iter, root, start, end) \
129 for_each_tdp_pte_min_level(iter, root, PG_LEVEL_4K, start, end)
133 void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root,
H A Dtdp_mmu.c37 * for zapping and thus puts the TDP MMU's reference to each root, i.e.
76 void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root) argument
78 if (!refcount_dec_and_test(&root->tdp_mmu_root_count))
82 * The TDP MMU itself holds a reference to each root until the root is
84 * put for a valid root.
86 KVM_BUG_ON(!is_tdp_mmu_page(root) || !root->role.invalid, kvm);
89 list_del_rcu(&root->link);
91 call_rcu(&root
232 struct kvm_mmu_page *root; local
746 __tdp_mmu_zap_root(struct kvm *kvm, struct kvm_mmu_page *root, bool shared, int zap_level) argument
772 tdp_mmu_zap_root(struct kvm *kvm, struct kvm_mmu_page *root, bool shared) argument
847 tdp_mmu_zap_leafs(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t start, gfn_t end, bool can_yield, bool flush) argument
895 struct kvm_mmu_page *root; local
906 struct kvm_mmu_page *root; local
931 struct kvm_mmu_page *root; local
976 struct kvm_mmu_page *root; local
1190 struct kvm_mmu_page *root; local
1206 struct kvm_mmu_page *root; local
1289 wrprot_gfn_range(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t start, gfn_t end, int min_level) argument
1330 struct kvm_mmu_page *root; local
1441 tdp_mmu_split_huge_pages_root(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t start, gfn_t end, int target_level, bool shared) argument
1515 struct kvm_mmu_page *root; local
1538 clear_dirty_gfn_range(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t start, gfn_t end) argument
1581 struct kvm_mmu_page *root; local
1592 clear_dirty_pt_masked(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t gfn, unsigned long mask, bool wrprot) argument
1643 struct kvm_mmu_page *root; local
1649 zap_collapsible_spte_range(struct kvm *kvm, struct kvm_mmu_page *root, const struct kvm_memory_slot *slot) argument
1707 struct kvm_mmu_page *root; local
1719 write_protect_gfn(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t gfn, int min_level) argument
1759 struct kvm_mmu_page *root; local
[all...]
H A Dtdp_mmu.h15 __must_check static inline bool kvm_tdp_mmu_get_root(struct kvm_mmu_page *root) argument
17 return refcount_inc_not_zero(&root->tdp_mmu_root_count);
20 void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root);
/linux-master/arch/x86/kvm/svm/
H A Dsvm.c3914 hpa_t root_tdp = vcpu->arch.mmu->root.hpa;
/linux-master/arch/x86/kvm/vmx/
H A Dnested.c3444 * NVMX_VMENTRY_SUCCESS: Entered VMX non-root mode
3561 * when it executed VMLAUNCH/VMRESUME, as entering non-root mode can
4881 * Bare metal isolates VMX root (host) from VMX non-root (guest), but
5239 * The CPL is checked for "not in VMX operation" and for "in VMX root",
5241 * i.e. VMXON #GPs outside of VMX non-root if CPL!=0. In VMX non-root,
5244 * VMX non-root.
5421 * In VMX non-root operation, when the VMCS-link pointer is INVALID_GPA,
5531 * In VMX non-root operatio
[all...]
H A Dvmx.c3175 * associated with the root EPT structure and not any particular VPID
3200 u64 root_hpa = mmu->root.hpa;
4178 * Case 1: vCPU stays in non-root mode. Sending a notification
4181 * Case 2: vCPU exits to root mode and is still runnable. The
4186 * Case 3: vCPU exits to root mode and is blocked. vcpu_block()
4240 * 1. If target vcpu is running(non-root mode), send posted interrupt
4242 * 2. If target vcpu isn't running(root mode), kick it to pick up the
6465 * mode as if vcpus is in root mode, the PML buffer must has been
/linux-master/arch/x86/pci/
H A Dacpi.c277 struct acpi_pci_root *root = ci->root; local
281 info->start_bus = (u8)root->secondary.start;
282 info->end_bus = (u8)root->secondary.end;
287 &root->secondary, &root->mcfg_addr);
297 root->mcfg_addr);
332 static int pci_acpi_root_get_node(struct acpi_pci_root *root) argument
334 int busnum = root->secondary.start;
335 struct acpi_device *device = root
414 pci_acpi_scan_root(struct acpi_pci_root *root) argument
[all...]

Completed in 355 milliseconds

1234567891011>>