Lines Matching refs:gtlbe

180 				struct kvm_book3e_206_tlb_entry *gtlbe)
184 size = get_tlb_bytes(gtlbe);
185 start = get_tlb_eaddr(gtlbe) & ~(size - 1);
194 struct kvm_book3e_206_tlb_entry *gtlbe)
199 if (!get_tlb_v(gtlbe))
202 size = get_tlb_bytes(gtlbe);
203 start = get_tlb_eaddr(gtlbe) & ~(size - 1);
214 struct kvm_book3e_206_tlb_entry *gtlbe =
217 if (unlikely(get_tlb_iprot(gtlbe)))
220 if (tlbsel == 1 && kvmppc_need_recalc_tlb1map_range(vcpu_e500, gtlbe))
223 gtlbe->mas1 = 0;
326 struct kvm_book3e_206_tlb_entry *gtlbe;
331 gtlbe = get_entry(vcpu_e500, tlbsel, esel);
334 vcpu->arch.shared->mas1 = gtlbe->mas1;
335 vcpu->arch.shared->mas2 = gtlbe->mas2;
336 vcpu->arch.shared->mas7_3 = gtlbe->mas7_3;
347 struct kvm_book3e_206_tlb_entry *gtlbe = NULL;
352 gtlbe = get_entry(vcpu_e500, tlbsel, esel);
357 if (gtlbe) {
362 vcpu->arch.shared->mas1 = gtlbe->mas1;
363 vcpu->arch.shared->mas2 = gtlbe->mas2;
364 vcpu->arch.shared->mas7_3 = gtlbe->mas7_3;
393 struct kvm_book3e_206_tlb_entry *gtlbe;
401 gtlbe = get_entry(vcpu_e500, tlbsel, esel);
403 if (get_tlb_v(gtlbe)) {
406 kvmppc_need_recalc_tlb1map_range(vcpu_e500, gtlbe))
410 gtlbe->mas1 = vcpu->arch.shared->mas1;
411 gtlbe->mas2 = vcpu->arch.shared->mas2;
413 gtlbe->mas2 &= 0xffffffffUL;
414 gtlbe->mas7_3 = vcpu->arch.shared->mas7_3;
416 trace_kvm_booke206_gtlb_write(vcpu->arch.shared->mas0, gtlbe->mas1,
417 gtlbe->mas2, gtlbe->mas7_3);
428 kvmppc_set_tlb1map_range(vcpu, gtlbe);
434 if (tlbe_is_host_safe(vcpu, gtlbe)) {
435 u64 eaddr = get_tlb_eaddr(gtlbe);
436 u64 raddr = get_tlb_raddr(gtlbe);
439 gtlbe->mas1 &= ~MAS1_TSIZE(~0);
440 gtlbe->mas1 |= MAS1_TSIZE(BOOK3E_PAGESZ_4K);
527 struct kvm_book3e_206_tlb_entry *gtlbe;
530 gtlbe = get_entry(vcpu_e500, tlbsel_of(index), esel_of(index));
531 pgmask = get_tlb_bytes(gtlbe) - 1;
533 return get_tlb_raddr(gtlbe) | (eaddr & pgmask);