• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/arch/ia64/kvm/

Lines Matching refs:arch

74 	if (vcpu->kvm->arch.is_sn2)
291 struct vpd *vpd = to_host(vcpu->kvm, vcpu->arch.vpd);
294 vcpu->arch.irq_new_pending = 1;
354 if (!target_vcpu->arch.launched) {
357 regs->cr_iip = vcpu->kvm->arch.rdv_sal_data.boot_ip;
358 regs->r1 = vcpu->kvm->arch.rdv_sal_data.boot_gp;
360 target_vcpu->arch.mp_state = KVM_MP_STATE_RUNNABLE;
386 if (vcpu->arch.ptc_g_count < MAX_PTC_G_NUM) {
387 vcpu->arch.ptc_g_data[vcpu->arch.ptc_g_count++] =
391 vcpu->arch.ptc_g_count = 0;
407 if (vcpui->arch.mp_state == KVM_MP_STATE_UNINITIALIZED ||
439 vcpu->arch.sn_rtc_tr_slot = slot;
454 struct hrtimer *p_ht = &vcpu->arch.hlt_timer;
456 struct vpd *vpd = to_host(vcpu->kvm, vcpu->arch.vpd);
460 vcpu_now_itc = kvm_get_itc(vcpu) + vcpu->arch.itc_offset;
463 vcpu->arch.timer_check = 1;
473 vcpu->arch.ht_active = 1;
476 vcpu->arch.mp_state = KVM_MP_STATE_HALTED;
479 vcpu->arch.ht_active = 0;
483 if (vcpu->arch.mp_state == KVM_MP_STATE_HALTED)
484 vcpu->arch.mp_state = KVM_MP_STATE_RUNNABLE;
486 if (vcpu->arch.mp_state != KVM_MP_STATE_RUNNABLE)
511 printk("VMM: %s", vcpu->arch.log_buf);
548 vcpu->arch.last_exit = exit_reason;
577 vcpu->arch.vmm_tr_slot = r;
579 pte = pte_val(mk_pte_phys(__pa(kvm->arch.vm_base), PAGE_KERNEL));
584 vcpu->arch.vm_tr_slot = r;
587 if (kvm->arch.is_sn2) {
602 ia64_ptr_entry(0x3, vcpu->arch.vmm_tr_slot);
603 ia64_ptr_entry(0x3, vcpu->arch.vm_tr_slot);
605 if (kvm->arch.is_sn2)
606 ia64_ptr_entry(0x3, vcpu->arch.sn_rtc_tr_slot);
616 if (vcpu->arch.last_run_cpu != cpu ||
619 vcpu->arch.last_run_cpu = cpu;
623 vcpu->arch.host_rr6 = ia64_get_rr(RR6);
624 vti_set_rr6(vcpu->arch.vmm_rr);
634 vti_set_rr6(vcpu->arch.host_rr6);
674 vcpu->arch.launched = 1;
731 if (unlikely(vcpu->arch.mp_state == KVM_MP_STATE_UNINITIALIZED)) {
768 kvm->arch.vm_base = vm_base;
813 kvm->arch.metaphysical_rr0 = GUEST_PHYSICAL_RR0;
814 kvm->arch.metaphysical_rr4 = GUEST_PHYSICAL_RR4;
815 kvm->arch.vmm_init_rr = VMM_INIT_RR;
822 INIT_LIST_HEAD(&kvm->arch.assigned_dev_head);
825 set_bit(KVM_USERSPACE_IRQ_SOURCE_ID, &kvm->arch.irq_sources_bitmap);
835 kvm->arch.is_sn2 = ia64_platform_is("sn2");
876 #define RESTORE_REGS(_x) vcpu->arch._x = regs->_x
880 struct vpd *vpd = to_host(vcpu->kvm, vcpu->arch.vpd);
896 memcpy(&vcpu->arch.guest, &regs->saved_guest, sizeof(union context));
900 memcpy(vcpu->arch.itrs, regs->itrs, sizeof(struct thash_data) * NITRS);
901 memcpy(vcpu->arch.dtrs, regs->dtrs, sizeof(struct thash_data) * NDTRS);
911 vcpu->arch.vrr[i] = regs->vrr[i];
912 vcpu->arch.ibr[i] = regs->ibr[i];
913 vcpu->arch.dbr[i] = regs->dbr[i];
916 vcpu->arch.insvc[i] = regs->insvc[i];
925 vcpu->arch.irq_new_pending = 1;
926 vcpu->arch.itc_offset = regs->saved_itc - kvm_get_itc(vcpu);
1087 struct vpd *vpd = to_host(vcpu->kvm, vcpu->arch.vpd);
1120 struct vpd *vpd = vcpu->arch.vpd;
1141 vcpu->arch.ptce_base = ptce.base;
1142 vcpu->arch.ptce_count[0] = ptce.count[0];
1143 vcpu->arch.ptce_count[1] = ptce.count[1];
1144 vcpu->arch.ptce_stride[0] = ptce.stride[0];
1145 vcpu->arch.ptce_stride[1] = ptce.stride[1];
1150 struct hrtimer *p_ht = &vcpu->arch.hlt_timer;
1161 vcpu = container_of(data, struct kvm_vcpu, arch.hlt_timer);
1164 if (vcpu->arch.mp_state != KVM_MP_STATE_HALTED)
1171 vcpu->arch.timer_fired = 1;
1172 vcpu->arch.timer_check = 1;
1187 union context *p_ctx = &vcpu->arch.guest;
1195 vcpu->arch.mp_state = KVM_MP_STATE_RUNNABLE;
1205 v->arch.itc_offset = itc_offset;
1206 v->arch.last_itc = 0;
1209 vcpu->arch.mp_state = KVM_MP_STATE_UNINITIALIZED;
1212 vcpu->arch.apic = kzalloc(sizeof(struct kvm_lapic), GFP_KERNEL);
1213 if (!vcpu->arch.apic)
1215 vcpu->arch.apic->vcpu = vcpu;
1245 vcpu->arch.vmm_rr = kvm->arch.vmm_init_rr;
1246 vcpu->arch.metaphysical_rr0 = kvm->arch.metaphysical_rr0;
1247 vcpu->arch.metaphysical_rr4 = kvm->arch.metaphysical_rr4;
1249 hrtimer_init(&vcpu->arch.hlt_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
1250 vcpu->arch.hlt_timer.function = hlt_timer_fn;
1252 vcpu->arch.last_run_cpu = -1;
1253 vcpu->arch.vpd = (struct vpd *)VPD_BASE(vcpu->vcpu_id);
1254 vcpu->arch.vsa_base = kvm_vsa_base;
1255 vcpu->arch.__gp = kvm_vmm_gp;
1256 vcpu->arch.dirty_log_lock_pa = __pa(&kvm->arch.dirty_log_lock);
1257 vcpu->arch.vhpt.hash = (struct thash_data *)VHPT_BASE(vcpu->vcpu_id);
1258 vcpu->arch.vtlb.hash = (struct thash_data *)VTLB_BASE(vcpu->vcpu_id);
1303 unsigned long vm_base = kvm->arch.vm_base;
1362 unsigned long vm_base = kvm->arch.vm_base;
1400 kfree(kvm->arch.vioapic);
1415 if (vcpu->arch.ht_active)
1420 #define SAVE_REGS(_x) regs->_x = vcpu->arch._x
1424 struct vpd *vpd = to_host(vcpu->kvm, vcpu->arch.vpd);
1441 memcpy(&regs->saved_guest, &vcpu->arch.guest, sizeof(union context));
1445 memcpy(regs->itrs, vcpu->arch.itrs, sizeof(struct thash_data) * NITRS);
1446 memcpy(regs->dtrs, vcpu->arch.dtrs, sizeof(struct thash_data) * NDTRS);
1456 regs->vrr[i] = vcpu->arch.vrr[i];
1457 regs->ibr[i] = vcpu->arch.ibr[i];
1458 regs->dbr[i] = vcpu->arch.dbr[i];
1461 regs->insvc[i] = vcpu->arch.insvc[i];
1462 regs->saved_itc = vcpu->arch.itc_offset + kvm_get_itc(vcpu);
1488 vcpu->arch.exit_data = ((struct kvm_vcpu *)stack)->arch.exit_data;
1495 hrtimer_cancel(&vcpu->arch.hlt_timer);
1496 kfree(vcpu->arch.apic);
1804 unsigned long *dirty_bitmap = (unsigned long *)(kvm->arch.vm_base +
1810 spin_lock(&kvm->arch.dirty_log_lock);
1815 spin_unlock(&kvm->arch.dirty_log_lock);
1895 return vcpu1->arch.xtp - vcpu2->arch.xtp;
1901 struct kvm_lapic *target = vcpu->arch.apic;
1926 struct vpd *vpd = to_host(vcpu->kvm, vcpu->arch.vpd);
1938 return vcpu->arch.timer_fired;
1943 return (vcpu->arch.mp_state == KVM_MP_STATE_RUNNABLE) ||
1950 mp_state->mp_state = vcpu->arch.mp_state;
1964 vcpu->arch.launched = 0;
1981 vcpu->arch.mp_state = mp_state->mp_state;
1982 if (vcpu->arch.mp_state == KVM_MP_STATE_UNINITIALIZED)