/linux-master/arch/s390/kvm/ |
H A D | intercept.c | 27 struct kvm_s390_sie_block *sie_block = vcpu->arch.sie_block; local 30 switch (vcpu->arch.sie_block->icptcode) { 37 ilen = insn_length(vcpu->arch.sie_block->ipa >> 8); 39 if (sie_block->icptstatus & 1) { 40 ilen = (sie_block->icptstatus >> 4) & 0x6; 47 ilen = vcpu->arch.sie_block->pgmilc & 0x6; 93 int viwhy = vcpu->arch.sie_block->ipb >> 16; 110 vcpu->arch.sie_block->ipa, 111 vcpu->arch.sie_block [all...] |
H A D | guestdbg.c | 62 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; 63 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; 64 u64 *cr11 = &vcpu->arch.sie_block->gcr[11]; 102 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; 103 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; 104 u64 *cr11 = &vcpu->arch.sie_block->gcr[11]; 132 vcpu->arch.guestdbg.cr0 = vcpu->arch.sie_block->gcr[0]; 133 vcpu->arch.guestdbg.cr9 = vcpu->arch.sie_block->gcr[9]; 134 vcpu->arch.guestdbg.cr10 = vcpu->arch.sie_block->gcr[10]; 135 vcpu->arch.guestdbg.cr11 = vcpu->arch.sie_block [all...] |
H A D | kvm-s390.h | 42 #define IS_TE_ENABLED(vcpu) ((vcpu->arch.sie_block->ecb & ECB_TE)) 45 ((*(char *)phys_to_virt((vcpu)->arch.sie_block->itdba) == TDB_FORMAT1)) 75 d_vcpu->arch.sie_block->gpsw.mask, d_vcpu->arch.sie_block->gpsw.addr,\ 81 atomic_or(flags, &vcpu->arch.sie_block->cpuflags); 86 atomic_andnot(flags, &vcpu->arch.sie_block->cpuflags); 91 return (atomic_read(&vcpu->arch.sie_block->cpuflags) & flags) == flags; 118 return vcpu->arch.sie_block->prefix << GUEST_PREFIX_SHIFT; 125 vcpu->arch.sie_block->prefix = prefix >> GUEST_PREFIX_SHIFT; 132 u32 base2 = vcpu->arch.sie_block 332 struct kvm_s390_sie_block *sie_block = vcpu->arch.sie_block; local [all...] |
H A D | priv.c | 38 vcpu->arch.sie_block->ecb3 |= ECB3_RI; 47 if ((vcpu->arch.sie_block->ipa & 0xf) <= 4) 64 vcpu->arch.sie_block->ecb |= ECB_GS; 65 vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT; 75 int code = vcpu->arch.sie_block->ipb & 0xff; 92 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) 131 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) 169 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) 199 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) 233 vcpu->arch.sie_block [all...] |
H A D | kvm-s390.c | 309 kvm_clock_sync_scb(vcpu->arch.sie_block, *delta); 311 kvm->arch.epoch = vcpu->arch.sie_block->epoch; 312 kvm->arch.epdx = vcpu->arch.sie_block->epdx; 1096 vcpu->arch.sie_block->ecb2 |= ECB2_ZPCI_LSI; 1097 vcpu->arch.sie_block->ecb3 |= ECB3_AISII + ECB3_AISI; 3452 free_page((unsigned long)(vcpu->arch.sie_block)); 3523 vcpu->arch.sie_block->scaoh = sca_phys >> 32; 3524 vcpu->arch.sie_block->scaol = sca_phys; 3532 sca->cpu[vcpu->vcpu_id].sda = virt_to_phys(vcpu->arch.sie_block); 3533 vcpu->arch.sie_block [all...] |
H A D | trace-s390.h | 44 struct kvm_s390_sie_block *sie_block), 45 TP_ARGS(id, vcpu, sie_block), 50 __field(struct kvm_s390_sie_block *, sie_block) 56 __entry->sie_block = sie_block; 60 __entry->id, __entry->vcpu, __entry->sie_block)
|
H A D | interrupt.c | 152 return !(vcpu->arch.sie_block->gpsw.mask & PSW_MASK_EXT); 157 return !(vcpu->arch.sie_block->gpsw.mask & PSW_MASK_IO); 162 return !(vcpu->arch.sie_block->gpsw.mask & PSW_MASK_MCHECK); 175 !(vcpu->arch.sie_block->gcr[0] & CR0_CLOCK_COMPARATOR_SUBMASK)) 186 const u64 ckc = vcpu->arch.sie_block->ckc; 188 if (vcpu->arch.sie_block->gcr[0] & CR0_CLOCK_COMPARATOR_SIGN) { 200 (vcpu->arch.sie_block->gcr[0] & CR0_CPU_TIMER_SUBMASK); 357 if (!(vcpu->arch.sie_block->gcr[6] & isc_to_isc_bits(i))) 377 if (!(vcpu->arch.sie_block->gcr[0] & CR0_EXTERNAL_CALL_SUBMASK)) 379 if (!(vcpu->arch.sie_block [all...] |
H A D | diag.c | 25 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; 26 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; 72 u16 rx = (vcpu->arch.sie_block->ipa & 0xf0) >> 4; 73 u16 ry = (vcpu->arch.sie_block->ipa & 0x0f); 172 tid = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; 214 unsigned int reg = vcpu->arch.sie_block->ipa & 0xf; 286 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
|
H A D | gaccess.h | 89 return _kvm_s390_logical_to_effective(&vcpu->arch.sie_block->gpsw, ga); 281 u8 access_key = psw_bits(vcpu->arch.sie_block->gpsw).key; 325 u8 access_key = psw_bits(vcpu->arch.sie_block->gpsw).key; 348 u8 access_key = psw_bits(vcpu->arch.sie_block->gpsw).key;
|
H A D | sigp.c | 76 psw = &dst_vcpu->arch.sie_block->gpsw; 77 p_asn = dst_vcpu->arch.sie_block->gcr[4] & 0xffff; /* Primary ASN */ 78 s_asn = dst_vcpu->arch.sie_block->gcr[3] & 0xffff; /* Secondary ASN */ 427 int r1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4; 428 int r3 = vcpu->arch.sie_block->ipa & 0x000f; 435 if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE) 478 int r3 = vcpu->arch.sie_block->ipa & 0x000f;
|
H A D | gaccess.c | 399 asce->val = vcpu->arch.sie_block->gcr[1]; 402 asce->val = vcpu->arch.sie_block->gcr[7]; 410 ald_addr = vcpu->arch.sie_block->gcr[5]; 412 ald_addr = vcpu->arch.sie_block->gcr[2]; 445 eax = (vcpu->arch.sie_block->gcr[8] >> 16) & 0xffff; 532 teid->as = psw_bits(vcpu->arch.sie_block->gpsw).as; 560 struct psw_bits psw = psw_bits(vcpu->arch.sie_block->gpsw); 573 asce->val = vcpu->arch.sie_block->gcr[1]; 576 asce->val = vcpu->arch.sie_block->gcr[7]; 579 asce->val = vcpu->arch.sie_block [all...] |
H A D | pv.c | 84 free_page((unsigned long)sida_addr(vcpu->arch.sie_block)); 85 vcpu->arch.sie_block->pv_handle_cpu = 0; 86 vcpu->arch.sie_block->pv_handle_config = 0; 88 vcpu->arch.sie_block->sdf = 0; 94 vcpu->arch.sie_block->gbea = 1; 119 uvcb.num = vcpu->arch.sie_block->icpua; 120 uvcb.state_origin = virt_to_phys(vcpu->arch.sie_block); 130 vcpu->arch.sie_block->sidad = virt_to_phys(sida_addr); 149 vcpu->arch.sie_block->pv_handle_cpu = uvcb.cpu_handle; 150 vcpu->arch.sie_block [all...] |
H A D | vsie.c | 307 int fmt_h = vcpu->arch.sie_block->crycbd & CRYCB_FORMAT_MASK; 312 apie_h = vcpu->arch.sie_block->eca & ECA_APIE; 335 ecb3_flags = scb_o->ecb3 & vcpu->arch.sie_block->ecb3 & 337 ecd_flags = scb_o->ecd & vcpu->arch.sie_block->ecd & ECD_ECC; 1042 cr0.val = vcpu->arch.sie_block->gcr[0]; 1136 vcpu->arch.sie_block->fpf & FPF_BPBC) 1149 vcpu->arch.sie_block->prog0c |= PROG_IN_SIE; 1154 vcpu->arch.sie_block->prog0c &= ~PROG_IN_SIE; 1215 asce = vcpu->arch.sie_block->gcr[1]; 1216 cr0.val = vcpu->arch.sie_block [all...] |
H A D | trace.h | 27 __entry->pswmask = vcpu->arch.sie_block->gpsw.mask; \ 28 __entry->pswaddr = vcpu->arch.sie_block->gpsw.addr; \
|
/linux-master/arch/s390/kernel/ |
H A D | nmi.c | 272 struct kvm_s390_sie_block *sie_block = phys_to_virt(regs->gprs[14]); local 274 if (sie_block == NULL) 278 sie_page = container_of(sie_block, struct sie_page, sie_block);
|
H A D | perf_event.c | 29 static struct kvm_s390_sie_block *sie_block(struct pt_regs *regs) function 52 return sie_block(regs)->gpsw.mask & PSW_MASK_PSTATE; 57 return sie_block(regs)->gpsw.addr;
|
/linux-master/arch/s390/include/asm/ |
H A D | kvm_host.h | 145 #define sida_addr(sie_block) phys_to_virt((sie_block)->sidad & PAGE_MASK) 146 #define sida_size(sie_block) \ 147 ((((sie_block)->sidad & SIDAD_SIZE_MASK) + 1) * PAGE_SIZE) 374 struct kvm_s390_sie_block sie_block; member in struct:sie_page 741 struct kvm_s390_sie_block *sie_block; member in struct:kvm_vcpu_arch 758 * The seqcount protects updates to cputm_start and sie_block.cputm, 1032 int __sie64a(phys_addr_t sie_block_phys, struct kvm_s390_sie_block *sie_block, u64 *rsa); 1034 static inline int sie64a(struct kvm_s390_sie_block *sie_block, u64 *rsa) argument 1036 return __sie64a(virt_to_phys(sie_block), sie_bloc [all...] |
/linux-master/drivers/s390/crypto/ |
H A D | vfio_ap_ops.c | 582 if (!(vcpu->arch.sie_block->eca & ECA_AIV)) { 584 __func__, apqn, vcpu->arch.sie_block->eca);
|