Lines Matching refs:pte_index

69 			     unsigned long *rmap, long pte_index, int realmode)
84 tail->forw = pte_index;
85 head->back = pte_index;
87 rev->forw = rev->back = pte_index;
89 pte_index | KVMPPC_RMAP_PRESENT | KVMPPC_RMAP_HPT;
147 static void remove_revmap_chain(struct kvm *kvm, long pte_index,
170 if (head == pte_index) {
172 if (head == pte_index)
185 long pte_index, unsigned long pteh, unsigned long ptel,
303 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
306 pte_index &= ~7UL;
307 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
336 pte_index += i;
338 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
355 rev = &kvm->arch.hpt.rev[pte_index];
376 kvmppc_add_revmap_chain(kvm, rev, rmap, pte_index,
396 *pte_idx_ret = pte_index;
402 long pte_index, unsigned long pteh, unsigned long ptel)
404 return kvmppc_do_h_enter(vcpu->kvm, flags, pte_index, pteh, ptel,
486 unsigned long pte_index, unsigned long avpn,
496 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
498 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
514 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
518 rb = compute_tlbie_rb(v, pte_r, pte_index);
529 remove_revmap_chain(kvm, pte_index, rev, v,
548 unsigned long pte_index, unsigned long avpn)
550 return kvmppc_do_h_remove(vcpu->kvm, flags, pte_index, avpn,
562 unsigned long flags, req, pte_index, rcbits;
575 pte_index = args[j];
576 flags = pte_index >> 56;
577 pte_index &= ((1ul << 56) - 1);
585 pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) {
587 args[j] = ((0xa0 | flags) << 56) + pte_index;
591 hp = (__be64 *) (kvm->arch.hpt.virt + (pte_index << 4));
623 args[j] = ((0x90 | flags) << 56) + pte_index;
627 args[j] = ((0x80 | flags) << 56) + pte_index;
628 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
643 tlbrb[n] = compute_tlbie_rb(hp0, hp1, pte_index);
659 pte_index = args[j] & ((1ul << 56) - 1);
662 remove_revmap_chain(kvm, pte_index, rev,
675 unsigned long pte_index, unsigned long avpn)
685 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
688 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
708 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
727 rb = compute_tlbie_rb(v, r, pte_index);
746 unsigned long pte_index)
756 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
759 pte_index &= ~3;
762 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
763 for (i = 0; i < n; ++i, ++pte_index) {
764 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
787 unsigned long pte_index)
798 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
801 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
802 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
818 kvmppc_clear_ref_hpte(kvm, hpte, pte_index);
836 unsigned long pte_index)
846 if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
849 rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
850 hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
866 kvmppc_invalidate_hpte(kvm, hpte, pte_index);
1017 unsigned long pte_index)
1029 rb = compute_tlbie_rb(hp0, hp1, pte_index);
1035 unsigned long pte_index)
1047 rb = compute_tlbie_rb(hp0, hp1, pte_index);
1211 index = cache_entry->pte_index;
1287 cache_entry->pte_index = index;