/linux-master/tools/perf/arch/arc/annotate/ |
H A D | instructions.c | 4 static int arc__annotate_init(struct arch *arch, char *cpuid __maybe_unused) argument 6 arch->initialized = true; 7 arch->objdump.comment_char = ';';
|
/linux-master/tools/perf/arch/riscv64/annotate/ |
H A D | instructions.c | 4 struct ins_ops *riscv64__associate_ins_ops(struct arch *arch, const char *name) argument 19 arch__associate_ins_ops(arch, name, ops); 25 int riscv64__annotate_init(struct arch *arch, char *cpuid __maybe_unused) argument 27 if (!arch->initialized) { 28 arch->associate_instruction_ops = riscv64__associate_ins_ops; 29 arch->initialized = true; 30 arch->objdump.comment_char = '#';
|
/linux-master/tools/perf/util/ |
H A D | perf_regs.c | 33 const char *perf_reg_name(int id, const char *arch) argument 37 if (!strcmp(arch, "csky")) 39 else if (!strcmp(arch, "loongarch")) 41 else if (!strcmp(arch, "mips")) 43 else if (!strcmp(arch, "powerpc")) 45 else if (!strcmp(arch, "riscv")) 47 else if (!strcmp(arch, "s390")) 49 else if (!strcmp(arch, "x86")) 51 else if (!strcmp(arch, "arm")) 53 else if (!strcmp(arch, "arm6 86 perf_arch_reg_ip(const char *arch) argument 111 perf_arch_reg_sp(const char *arch) argument [all...] |
/linux-master/tools/perf/trace/beauty/ |
H A D | arch_errno_names.sh | 20 arch="$1" 22 header="$toolsdir/arch/$arch/include/uapi/asm/errno.h" 32 arch=$(arch_string "$1") 34 printf "static const char *errno_to_name__%s(int err)\n{\n\tswitch (err) {\n" $arch 45 arch="$1" 46 asm_errno=$(asm_errno_file "$arch") 52 |IFS=, create_errno_lookup_func "$arch" 60 printf 'arch_syscalls__strerrno_t *arch_syscalls__strerrno_function(const char *arch)\n' 62 for arch i [all...] |
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_hv_tm.c | 19 u64 msr = vcpu->arch.shregs.msr; 21 tfiar = vcpu->arch.regs.nip & ~0x3ull; 23 if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) 29 vcpu->arch.tfiar = tfiar; 31 vcpu->arch.texasr = (vcpu->arch.texasr & 0x3ffffff) | texasr; 37 * instruction image is in vcpu->arch.emul_inst. If the guest was in 44 u32 instr = vcpu->arch.emul_inst; 45 u64 msr = vcpu->arch.shregs.msr; 56 vcpu->arch [all...] |
H A D | book3s_hv_tm_builtin.c | 22 u32 instr = vcpu->arch.emul_inst; 40 newmsr = vcpu->arch.shregs.srr1; 45 vcpu->arch.shregs.msr = newmsr; 46 vcpu->arch.cfar = vcpu->arch.regs.nip - 4; 47 vcpu->arch.regs.nip = vcpu->arch.shregs.srr0; 51 /* check for PR=1 and arch 2.06 bit set in PCR */ 52 msr = vcpu->arch.shregs.msr; 53 if ((msr & MSR_PR) && (vcpu->arch [all...] |
H A D | booke_emulate.c | 26 vcpu->arch.regs.nip = vcpu->arch.shared->srr0; 27 kvmppc_set_msr(vcpu, vcpu->arch.shared->srr1); 32 vcpu->arch.regs.nip = vcpu->arch.dsrr0; 33 kvmppc_set_msr(vcpu, vcpu->arch.dsrr1); 38 vcpu->arch.regs.nip = vcpu->arch.csrr0; 39 kvmppc_set_msr(vcpu, vcpu->arch.csrr1); 80 kvmppc_set_gpr(vcpu, rt, vcpu->arch [all...] |
H A D | book3s_hv_p9_entry.c | 14 mtspr(SPRN_TAR, vcpu->arch.tar); 18 current->thread.vrsave != vcpu->arch.vrsave) 19 mtspr(SPRN_VRSAVE, vcpu->arch.vrsave); 22 if (vcpu->arch.hfscr & HFSCR_EBB) { 23 if (current->thread.ebbhr != vcpu->arch.ebbhr) 24 mtspr(SPRN_EBBHR, vcpu->arch.ebbhr); 25 if (current->thread.ebbrr != vcpu->arch.ebbrr) 26 mtspr(SPRN_EBBRR, vcpu->arch.ebbrr); 27 if (current->thread.bescr != vcpu->arch.bescr) 28 mtspr(SPRN_BESCR, vcpu->arch [all...] |
H A D | book3s_hv_p9_perf.c | 41 lp = vcpu->arch.vpa.pinned_addr; 87 if (load_pmu || (vcpu->arch.hfscr & HFSCR_PM)) { 88 mtspr(SPRN_PMC1, vcpu->arch.pmc[0]); 89 mtspr(SPRN_PMC2, vcpu->arch.pmc[1]); 90 mtspr(SPRN_PMC3, vcpu->arch.pmc[2]); 91 mtspr(SPRN_PMC4, vcpu->arch.pmc[3]); 92 mtspr(SPRN_PMC5, vcpu->arch.pmc[4]); 93 mtspr(SPRN_PMC6, vcpu->arch.pmc[5]); 94 mtspr(SPRN_MMCR1, vcpu->arch.mmcr[1]); 95 mtspr(SPRN_MMCR2, vcpu->arch [all...] |
H A D | emulate_loadstore.c | 85 vcpu->arch.mmio_vsx_copy_nums = 0; 86 vcpu->arch.mmio_vsx_offset = 0; 87 vcpu->arch.mmio_copy_type = KVMPPC_VSX_COPY_NONE; 88 vcpu->arch.mmio_sp64_extend = 0; 89 vcpu->arch.mmio_sign_extend = 0; 90 vcpu->arch.mmio_vmx_copy_nums = 0; 91 vcpu->arch.mmio_vmx_offset = 0; 92 vcpu->arch.mmio_host_swabbed = 0; 95 vcpu->arch.regs.msr = kvmppc_get_msr(vcpu); 96 if (analyse_instr(&op, &vcpu->arch [all...] |
H A D | timing.c | 27 mutex_lock(&vcpu->arch.exit_timing_lock); 29 vcpu->arch.last_exit_type = 0xDEAD; 31 vcpu->arch.timing_count_type[i] = 0; 32 vcpu->arch.timing_max_duration[i] = 0; 33 vcpu->arch.timing_min_duration[i] = 0xFFFFFFFF; 34 vcpu->arch.timing_sum_duration[i] = 0; 35 vcpu->arch.timing_sum_quad_duration[i] = 0; 37 vcpu->arch.timing_last_exit = 0; 38 vcpu->arch.timing_exit.tv64 = 0; 39 vcpu->arch [all...] |
H A D | e500_emulate.c | 8 * This file is derived from arch/powerpc/kvm/44x_emulate.c, 53 ulong param = vcpu->arch.regs.gpr[rb]; 59 clear_bit(prio, &vcpu->arch.pending_exceptions); 65 ulong param = vcpu->arch.regs.gpr[rb]; 75 int cpir = cvcpu->arch.shared->pir; 77 set_bit(prio, &cvcpu->arch.pending_exceptions); 94 vcpu->run->debug.arch.address = vcpu->arch.regs.nip; 95 vcpu->run->debug.arch.status = 0; 225 vcpu->arch [all...] |
/linux-master/tools/testing/selftests/kvm/include/x86_64/ |
H A D | kvm_util_arch.h | 15 static inline bool __vm_arch_has_protected_memory(struct kvm_vm_arch *arch) argument 17 return arch->c_bit || arch->s_bit; 21 __vm_arch_has_protected_memory(&(vm)->arch)
|
/linux-master/tools/perf/arch/powerpc/annotate/ |
H A D | instructions.c | 4 static struct ins_ops *powerpc__associate_instruction_ops(struct arch *arch, const char *name) argument 48 arch__associate_ins_ops(arch, name, ops); 52 static int powerpc__annotate_init(struct arch *arch, char *cpuid __maybe_unused) argument 54 if (!arch->initialized) { 55 arch->initialized = true; 56 arch->associate_instruction_ops = powerpc__associate_instruction_ops; 57 arch->objdump.comment_char = '#';
|
/linux-master/tools/perf/arch/mips/annotate/ |
H A D | instructions.c | 4 struct ins_ops *mips__associate_ins_ops(struct arch *arch, const char *name) argument 31 arch__associate_ins_ops(arch, name, ops); 37 int mips__annotate_init(struct arch *arch, char *cpuid __maybe_unused) argument 39 if (!arch->initialized) { 40 arch->associate_instruction_ops = mips__associate_ins_ops; 41 arch->initialized = true; 42 arch->objdump.comment_char = '#';
|
/linux-master/tools/perf/arch/csky/annotate/ |
H A D | instructions.c | 6 static struct ins_ops *csky__associate_ins_ops(struct arch *arch, argument 37 arch__associate_ins_ops(arch, name, ops); 41 static int csky__annotate_init(struct arch *arch, char *cpuid __maybe_unused) argument 43 arch->initialized = true; 44 arch->objdump.comment_char = '/'; 45 arch->associate_instruction_ops = csky__associate_ins_ops;
|
/linux-master/arch/arm/mach-npcm/ |
H A D | wpcm450.c | 4 #include <asm/mach/arch.h>
|
/linux-master/arch/arm/mach-ixp4xx/ |
H A D | ixp4xx-of.c | 5 #include <asm/mach/arch.h>
|
/linux-master/arch/arm/mach-bcm/ |
H A D | board_bcm23550.c | 4 #include <asm/mach/arch.h>
|
/linux-master/tools/perf/arch/arm/annotate/ |
H A D | instructions.c | 13 static struct ins_ops *arm__associate_instruction_ops(struct arch *arch, const char *name) argument 15 struct arm_annotate *arm = arch->priv; 26 arch__associate_ins_ops(arch, name, ops); 30 static int arm__annotate_init(struct arch *arch, char *cpuid __maybe_unused) argument 35 if (arch->initialized) 51 arch->initialized = true; 52 arch->priv = arm; 53 arch [all...] |
/linux-master/arch/powerpc/include/asm/ |
H A D | kvm_booke.h | 28 vcpu->arch.regs.gpr[num] = val; 33 return vcpu->arch.regs.gpr[num]; 38 vcpu->arch.regs.ccr = val; 43 return vcpu->arch.regs.ccr; 48 vcpu->arch.regs.xer = val; 53 return vcpu->arch.regs.xer; 64 vcpu->arch.regs.ctr = val; 69 return vcpu->arch.regs.ctr; 74 vcpu->arch.regs.link = val; 79 return vcpu->arch [all...] |
/linux-master/arch/x86/tools/ |
H A D | Makefile | 16 reformatter = $(srctree)/arch/x86/tools/objdump_reformat.awk 17 chkobjdump = $(srctree)/arch/x86/tools/chkobjdump.awk 32 HOSTCFLAGS_insn_decoder_test.o := -Wall -I$(srctree)/tools/arch/x86/lib/ -I$(srctree)/tools/arch/x86/include/ -I$(objtree)/arch/x86/lib/ 34 HOSTCFLAGS_insn_sanity.o := -Wall -I$(srctree)/tools/arch/x86/lib/ -I$(srctree)/tools/arch/x86/include/ -I$(objtree)/arch/x86/lib/ 37 $(obj)/insn_decoder_test.o: $(srctree)/tools/arch/x86/lib/insn.c $(srctree)/tools/arch/x8 [all...] |
/linux-master/arch/mips/kvm/ |
H A D | hypcall.c | 22 kvm_debug("[%#lx] HYPCALL %#03x\n", vcpu->arch.pc, code); 45 num = vcpu->arch.gprs[2]; /* v0 */ 46 args[0] = vcpu->arch.gprs[4]; /* a0 */ 47 args[1] = vcpu->arch.gprs[5]; /* a1 */ 48 args[2] = vcpu->arch.gprs[6]; /* a2 */ 49 args[3] = vcpu->arch.gprs[7]; /* a3 */ 52 args, &vcpu->arch.gprs[2] /* v0 */);
|
H A D | emulate.c | 45 struct kvm_vcpu_arch *arch = &vcpu->arch; local 65 arch->gprs[insn.r_format.rd] = epc + 8; 68 nextpc = arch->gprs[insn.r_format.rs]; 84 if ((long)arch->gprs[insn.i_format.rs] < 0) 93 if ((long)arch->gprs[insn.i_format.rs] >= 0) 102 arch->gprs[31] = epc + 8; 103 if ((long)arch->gprs[insn.i_format.rs] < 0) 112 arch->gprs[31] = epc + 8; 113 if ((long)arch [all...] |
/linux-master/arch/s390/kvm/ |
H A D | guestdbg.c | 62 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; 63 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; 64 u64 *cr11 = &vcpu->arch.sie_block->gcr[11]; 67 if (vcpu->arch.guestdbg.nr_hw_bp <= 0 || 68 vcpu->arch.guestdbg.hw_bp_info == NULL) 79 for (i = 0; i < vcpu->arch.guestdbg.nr_hw_bp; i++) { 80 start = vcpu->arch.guestdbg.hw_bp_info[i].addr; 81 len = vcpu->arch.guestdbg.hw_bp_info[i].len; 102 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; 103 u64 *cr10 = &vcpu->arch [all...] |