| /xen/xen/arch/x86/cpu/mtrr/ |
| A D | generic.c | 34 rdmsrl(MSR_IA32_MTRR_PHYSBASE(index), vr->base); in get_mtrr_var_range() 35 rdmsrl(MSR_IA32_MTRR_PHYSMASK(index), vr->mask); in get_mtrr_var_range() 51 rdmsrl(block->base_msr + i, *p); in get_fixed_ranges() 107 rdmsrl(MSR_MTRRcap, msr_content); in get_mtrr_state() 114 rdmsrl(MSR_MTRRdefType, msr_content); in get_mtrr_state() 120 rdmsrl(MSR_MTRRcap, mtrr_state.mtrr_cap); in get_mtrr_state() 226 rdmsrl(MSR_K8_SYSCFG, syscfg); in print_mtrr_state() 228 rdmsrl(MSR_K8_TOP_MEM2, tom2); in print_mtrr_state() 281 rdmsrl(msr, msr_content); in set_fixed_range() 472 rdmsrl(MSR_MTRRdefType, deftype); in prepare_set() [all …]
|
| /xen/xen/arch/x86/oprofile/ |
| A D | op_model_athlon.c | 36 #define CTR_READ(msr_content,msrs,c) do {rdmsrl(msrs->counters[(c)].addr, (msr_content));} while (0) 262 rdmsrl(MSR_AMD64_IBSFETCHCTL, ctl); in handle_ibs() 264 rdmsrl(MSR_AMD64_IBSFETCHLINAD, val); in handle_ibs() 271 rdmsrl(MSR_AMD64_IBSFETCHPHYSAD, val); in handle_ibs() 282 rdmsrl(MSR_AMD64_IBSOPCTL, ctl); in handle_ibs() 285 rdmsrl(MSR_AMD64_IBSOPRIP, val); in handle_ibs() 291 rdmsrl(MSR_AMD64_IBSOPDATA, val); in handle_ibs() 293 rdmsrl(MSR_AMD64_IBSOPDATA2, val); in handle_ibs() 295 rdmsrl(MSR_AMD64_IBSOPDATA3, val); in handle_ibs() 297 rdmsrl(MSR_AMD64_IBSDCLINAD, val); in handle_ibs() [all …]
|
| A D | op_model_p4.c | 354 #define ESCR_READ(escr,ev,i) do {rdmsrl(ev->bindings[(i)].escr_address, (escr));} while (0) 365 #define CCCR_READ(msr_content, i) do {rdmsrl(p4_counters[(i)].cccr_address, (msr_content));} while … 370 #define CTR_READ(msr_content,i) do {rdmsrl(p4_counters[(i)].counter_address, (msr_content));} while… 542 rdmsrl(MSR_IA32_MISC_ENABLE, msr_content); in p4_setup_ctrs() 550 rdmsrl(p4_counters[VIRT_CTR(stag, i)].cccr_address, msr_content); in p4_setup_ctrs() 558 rdmsrl(p4_unused_cccr[i], msr_content); in p4_setup_ctrs()
|
| A D | nmi_int.c | 110 rdmsrl(counters[i].addr, counters[i].value); in nmi_cpu_save_registers() 114 rdmsrl(controls[i].addr, controls[i].value); in nmi_cpu_save_registers()
|
| A D | op_model_ppro.c | 51 #define CTRL_READ(msr_content,msrs,c) do {rdmsrl((msrs->controls[(c)].addr), (msr_content));} while… 145 rdmsrl(msrs->counters[i].addr, val); in ppro_check_ctrs()
|
| /xen/xen/arch/x86/guest/hyperv/ |
| A D | hyperv.c | 103 rdmsrl(HV_X64_MSR_GUEST_OS_ID, guest_id.raw); in setup_hypercall_page() 110 rdmsrl(HV_X64_MSR_HYPERCALL, hypercall_msr.as_uint64); in setup_hypercall_page() 121 rdmsrl(HV_X64_MSR_HYPERCALL, hypercall_msr.as_uint64); in setup_hypercall_page() 144 rdmsrl(HV_X64_MSR_VP_INDEX, vp_index_msr); in setup_hypercall_pcpu_arg() 170 rdmsrl(HV_X64_MSR_VP_ASSIST_PAGE, msr.raw); in setup_vp_assist()
|
| /xen/xen/arch/x86/cpu/ |
| A D | vpmu_intel.c | 130 rdmsrl(MSR_P6_PERFCTR(i), cnt); in handle_pmc_quirk() 142 rdmsrl(MSR_CORE_PERF_FIXED_CTR0 + i, cnt); in handle_pmc_quirk() 283 rdmsrl(MSR_CORE_PERF_FIXED_CTR0 + i, fixed_counters[i]); in __core2_vpmu_save() 285 rdmsrl(MSR_IA32_PERFCTR0 + i, xen_pmu_cntr_pair[i].counter); in __core2_vpmu_save() 288 rdmsrl(MSR_CORE_PERF_GLOBAL_STATUS, core2_vpmu_cxt->global_status); in __core2_vpmu_save() 597 rdmsrl(MSR_CORE_PERF_GLOBAL_CTRL, core2_vpmu_cxt->global_ctrl); in core2_vpmu_do_wrmsr() 718 rdmsrl(MSR_CORE_PERF_GLOBAL_CTRL, *msr_content); in core2_vpmu_do_rdmsr() 721 rdmsrl(msr, *msr_content); in core2_vpmu_do_rdmsr() 788 rdmsrl(MSR_CORE_PERF_GLOBAL_STATUS, msr_content); in core2_vpmu_do_interrupt() 861 rdmsrl(MSR_IA32_MISC_ENABLE, msr_content); in vmx_vpmu_initialise() [all …]
|
| A D | centaur.c | 27 rdmsrl(MSR_VIA_FCR, msr_content); in init_c3() 35 rdmsrl(MSR_VIA_RNG, msr_content); in init_c3()
|
| A D | amd.c | 380 rdmsrl(MSR_AMD_OSVW_ID_LENGTH, osvw_len); in cpu_has_amd_erratum() 385 rdmsrl(MSR_AMD_OSVW_STATUS + (osvw_id >> 6), in cpu_has_amd_erratum() 470 rdmsrl(MSR_K8_SYSCFG, syscfg); in check_syscfg_dram_mod_en() 560 rdmsrl(0xC001100C, val); in amd_log_freq() 568 rdmsrl(MSR_AMD64_NB_CFG, nbcfg); in amd_log_freq() 658 rdmsrl(MSR_K7_HWCR, value); in init_amd() 796 rdmsrl(MSR_K8_EXT_FEATURE_MASK, value); in init_amd() 852 rdmsrl(MSR_AMD64_LS_CFG, value); in init_amd() 864 rdmsrl(MSR_AMD64_DE_CFG, value); in init_amd() 894 rdmsrl(MSR_F10_BU_CFG2, value); in init_amd()
|
| A D | mwait-idle.c | 846 rdmsrl(MSR_NHM_SNB_PKG_CST_CFG_CTL, msr_bits); in auto_demotion_disable() 861 rdmsrl(MSR_IA32_POWER_CTL, msr_bits); in c1e_promotion_disable() 1057 rdmsrl(MSR_PKGC6_IRTL, msr); in bxt_idle_state_table_update() 1064 rdmsrl(MSR_PKGC7_IRTL, msr); in bxt_idle_state_table_update() 1071 rdmsrl(MSR_PKGC8_IRTL, msr); in bxt_idle_state_table_update() 1078 rdmsrl(MSR_PKGC9_IRTL, msr); in bxt_idle_state_table_update() 1085 rdmsrl(MSR_PKGC10_IRTL, msr); in bxt_idle_state_table_update() 1111 rdmsrl(MSR_NHM_SNB_PKG_CST_CFG_CTL, msr); in sklh_idle_state_table_update() 1119 rdmsrl(MSR_IA32_FEATURE_CONTROL, msr); in sklh_idle_state_table_update()
|
| A D | vpmu_amd.c | 285 rdmsrl(counters[i], counter_regs[i]); in context_save() 427 rdmsrl(msr, *msr_content); in amd_vpmu_do_rdmsr() 482 rdmsrl(ctrls[i], ctrl); in amd_vpmu_dump() 483 rdmsrl(counters[i], cntr); in amd_vpmu_dump() 564 rdmsrl(ctrls[i], ctrl_rsvd[i]); in common_init()
|
| A D | hygon.c | 98 rdmsrl(MSR_K7_HWCR, value); in init_hygon()
|
| /xen/xen/include/asm-x86/ |
| A D | msr.h | 21 #define rdmsrl(msr,val) do { unsigned long a__,b__; \ macro 166 rdmsrl(MSR_FS_BASE, base); in rdfsbase() 178 rdmsrl(MSR_GS_BASE, base); in rdgsbase() 194 rdmsrl(MSR_SHADOW_GS_BASE, base); in rdgsshadow()
|
| /xen/xen/arch/x86/ |
| A D | tsx.c | 45 rdmsrl(MSR_ARCH_CAPABILITIES, caps); in tsx_init() 54 rdmsrl(MSR_TSX_CTRL, val); in tsx_init()
|
| A D | nmi.c | 358 rdmsrl(MSR_IA32_MISC_ENABLE, misc_enable); in setup_p4_watchdog() 527 rdmsrl(MSR_P4_IQ_CCCR0, msr_content); in nmi_watchdog_tick() 543 rdmsrl(MSR_P6_PERFCTR(0), msr_content); in nmi_watchdog_tick() 555 rdmsrl(MSR_K7_PERFCTR0, msr_content); in nmi_watchdog_tick()
|
| A D | msr.c | 142 rdmsrl(MSR_ARCH_CAPABILITIES, val); in init_domain_msr_policy() 213 rdmsrl(MSR_IA32_PLATFORM_ID, *val); in guest_rdmsr() 327 rdmsrl(msr, *val); in guest_rdmsr()
|
| A D | apic.c | 310 rdmsrl(MSR_APIC_BASE, msr_content); in disable_local_APIC() 318 rdmsrl(MSR_APIC_BASE, msr_content); in disable_local_APIC() 486 rdmsrl(MSR_APIC_BASE, msr_content); in __enable_x2apic() 727 rdmsrl(MSR_APIC_BASE, msr_content); in lapic_resume() 1530 rdmsrl(MSR_APIC_BASE, msr_contents); in current_local_apic_mode()
|
| A D | e820.c | 454 rdmsrl(MSR_MTRRcap, mtrr_cap); in mtrr_top_of_ram() 455 rdmsrl(MSR_MTRRdefType, mtrr_def); in mtrr_top_of_ram() 471 rdmsrl(MSR_IA32_MTRR_PHYSBASE(i), base); in mtrr_top_of_ram() 472 rdmsrl(MSR_IA32_MTRR_PHYSMASK(i), mask); in mtrr_top_of_ram()
|
| /xen/xen/arch/x86/cpu/mcheck/ |
| A D | mce_intel.c | 71 rdmsrl(MSR_IA32_THERM_STATUS, msr_content); in intel_thermal_interrupt() 124 rdmsrl(MSR_IA32_MISC_ENABLE, msr_content); in intel_init_thermal() 167 rdmsrl(MSR_IA32_THERM_INTERRUPT, msr_content); in intel_init_thermal() 170 rdmsrl(MSR_IA32_MISC_ENABLE, msr_content); in intel_init_thermal() 186 rdmsrl(msr, ext->mc_msr[ext->mc_msrs].value); in intel_get_extended_msr() 497 rdmsrl(msr, val); in do_cmci_discover() 508 rdmsrl(msr, val); in do_cmci_discover() 629 rdmsrl(msr, val); in clear_cmci() 753 rdmsrl(MSR_IA32_MCG_CAP, msr_content); in intel_init_mca() 834 rdmsrl(MSR_IA32_MCx_CTL(i), msr_content); in intel_init_mce()
|
| A D | mce_amd.c | 323 rdmsrl(MSR_AMD_PPIN_CTL, val); in amd_mcheck_init() 329 rdmsrl(MSR_AMD_PPIN_CTL, val); in amd_mcheck_init()
|
| A D | amd_nonfatal.c | 218 rdmsrl(MSR_IA32_MCx_MISC(4), value); in amd_nonfatal_mcheck_init()
|
| /xen/xen/arch/x86/x86_64/ |
| A D | mmconf-fam10h.c | 69 rdmsrl(address, val); in get_fam10h_pci_mmconf_base() 77 rdmsrl(address, val); in get_fam10h_pci_mmconf_base() 150 rdmsrl(MSR_FAM10H_MMIO_CONF_BASE, val); in fam10h_check_enable_mmcfg()
|
| /xen/xen/arch/x86/cpu/microcode/ |
| A D | intel.c | 126 rdmsrl(MSR_IA32_PLATFORM_ID, msr_content); in collect_cpu_info() 134 rdmsrl(MSR_IA32_UCODE_REV, msr_content); in collect_cpu_info() 281 rdmsrl(MSR_IA32_UCODE_REV, msr_content); in apply_microcode()
|
| A D | amd.c | 99 rdmsrl(MSR_AMD_PATCHLEVEL, csig->rev); in collect_cpu_info() 233 rdmsrl(MSR_AMD_PATCHLEVEL, rev); in apply_microcode()
|
| /xen/xen/arch/x86/acpi/cpufreq/ |
| A D | cpufreq.c | 140 rdmsrl(cmd->addr.msr.reg, cmd->val); in do_drv_read() 160 rdmsrl(cmd->addr.msr.reg, msr_content); in do_drv_write() 251 rdmsrl(MSR_IA32_APERF, readin->aperf.whole); in read_measured_perf_ctrs() 252 rdmsrl(MSR_IA32_MPERF, readin->mperf.whole); in read_measured_perf_ctrs()
|