|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v5 04/17] intel/VPMU: Clean up Intel VPMU code
Am Montag 17 Februar 2014, 12:55:51 schrieb Boris Ostrovsky:
> Remove struct pmumsr and core2_pmu_enable. Replace static MSR structures with
> fields in core2_vpmu_context.
>
> Call core2_get_pmc_count() once, during initialization.
>
> Properly clean up when core2_vpmu_alloc_resource() fails and add routines
> to remove MSRs from VMCS.
Only a minor remark below.
>
> Signed-off-by: Boris Ostrovsky <boris.ostrovsky@xxxxxxxxxx>
> ---
> xen/arch/x86/hvm/vmx/vmcs.c | 55 ++++++
> xen/arch/x86/hvm/vmx/vpmu_core2.c | 310
> ++++++++++++++-----------------
> xen/include/asm-x86/hvm/vmx/vmcs.h | 2 +
> xen/include/asm-x86/hvm/vmx/vpmu_core2.h | 19 --
> 4 files changed, 199 insertions(+), 187 deletions(-)
>
> diff --git a/xen/arch/x86/hvm/vmx/vmcs.c b/xen/arch/x86/hvm/vmx/vmcs.c
> index 44f33cb..5f86b17 100644
> --- a/xen/arch/x86/hvm/vmx/vmcs.c
> +++ b/xen/arch/x86/hvm/vmx/vmcs.c
> @@ -1205,6 +1205,34 @@ int vmx_add_guest_msr(u32 msr)
> return 0;
> }
>
> +void vmx_rm_guest_msr(u32 msr)
> +{
> + struct vcpu *curr = current;
> + unsigned int idx, msr_count = curr->arch.hvm_vmx.msr_count;
> + struct vmx_msr_entry *msr_area = curr->arch.hvm_vmx.msr_area;
> +
> + if ( msr_area == NULL )
> + return;
> +
> + for ( idx = 0; idx < msr_count; idx++ )
> + if ( msr_area[idx].index == msr )
> + break;
> +
> + if ( idx == msr_count )
> + return;
> + for ( ; idx < msr_count - 1; idx++ )
> + {
> + msr_area[idx].index = msr_area[idx + 1].index;
> + msr_area[idx].data = msr_area[idx + 1].data;
> + }
> + msr_area[msr_count - 1].index = 0;
> +
> + curr->arch.hvm_vmx.msr_count = --msr_count;
> + __vmwrite(VM_EXIT_MSR_STORE_COUNT, msr_count);
> + __vmwrite(VM_ENTRY_MSR_LOAD_COUNT, msr_count);
> +}
> +
> int vmx_add_host_load_msr(u32 msr)
> {
> struct vcpu *curr = current;
> @@ -1235,6 +1263,33 @@ int vmx_add_host_load_msr(u32 msr)
> return 0;
> }
>
> +void vmx_rm_host_load_msr(u32 msr)
> +{
> + struct vcpu *curr = current;
> + unsigned int idx, msr_count = curr->arch.hvm_vmx.host_msr_count;
> + struct vmx_msr_entry *msr_area = curr->arch.hvm_vmx.host_msr_area;
> +
> + if ( msr_area == NULL )
> + return;
> +
> + for ( idx = 0; idx < msr_count; idx++ )
> + if ( msr_area[idx].index == msr )
> + break;
> +
> + if ( idx == msr_count )
> + return;
> +
> + for ( ; idx < msr_count - 1; idx++ )
> + {
> + msr_area[idx].index = msr_area[idx + 1].index;
> + msr_area[idx].data = msr_area[idx + 1].data;
> + }
> + msr_area[msr_count - 1].index = 0;
> +
> + curr->arch.hvm_vmx.host_msr_count = --msr_count;
> + __vmwrite(VM_EXIT_MSR_LOAD_COUNT, msr_count);
> +}
> +
> void vmx_set_eoi_exit_bitmap(struct vcpu *v, u8 vector)
> {
> if ( !test_and_set_bit(vector, v->arch.hvm_vmx.eoi_exit_bitmap) )
> diff --git a/xen/arch/x86/hvm/vmx/vpmu_core2.c
> b/xen/arch/x86/hvm/vmx/vpmu_core2.c
> index 1e32ff3..513eca4 100644
> --- a/xen/arch/x86/hvm/vmx/vpmu_core2.c
> +++ b/xen/arch/x86/hvm/vmx/vpmu_core2.c
> @@ -69,6 +69,26 @@
> static bool_t __read_mostly full_width_write;
>
> /*
> + * MSR_CORE_PERF_FIXED_CTR_CTRL contains the configuration of all fixed
> + * counters. 4 bits for every counter.
> + */
> +#define FIXED_CTR_CTRL_BITS 4
> +#define FIXED_CTR_CTRL_MASK ((1 << FIXED_CTR_CTRL_BITS) - 1)
> +
> +#define VPMU_CORE2_MAX_FIXED_PMCS 4
> +struct core2_vpmu_context {
> + u64 fixed_ctrl;
> + u64 ds_area;
> + u64 pebs_enable;
> + u64 global_ovf_status;
> + u64 fix_counters[VPMU_CORE2_MAX_FIXED_PMCS];
> + struct arch_msr_pair arch_msr_pair[1];
> +};
> +
> +/* Number of general-purpose and fixed performance counters */
> +static unsigned int __read_mostly arch_pmc_cnt, fixed_pmc_cnt;
> +
> +/*
> * QUIRK to workaround an issue on various family 6 cpus.
> * The issue leads to endless PMC interrupt loops on the processor.
> * If the interrupt handler is running and a pmc reaches the value 0, this
> @@ -88,11 +108,8 @@ static void check_pmc_quirk(void)
> is_pmc_quirk = 0;
> }
>
> -static int core2_get_pmc_count(void);
> static void handle_pmc_quirk(u64 msr_content)
> {
> - int num_gen_pmc = core2_get_pmc_count();
> - int num_fix_pmc = 3;
> int i;
> u64 val;
>
> @@ -100,7 +117,7 @@ static void handle_pmc_quirk(u64 msr_content)
> return;
>
> val = msr_content;
> - for ( i = 0; i < num_gen_pmc; i++ )
> + for ( i = 0; i < arch_pmc_cnt; i++ )
> {
> if ( val & 0x1 )
> {
> @@ -112,7 +129,7 @@ static void handle_pmc_quirk(u64 msr_content)
> val >>= 1;
> }
> val = msr_content >> 32;
> - for ( i = 0; i < num_fix_pmc; i++ )
> + for ( i = 0; i < fixed_pmc_cnt; i++ )
> {
> if ( val & 0x1 )
> {
> @@ -125,75 +142,42 @@ static void handle_pmc_quirk(u64 msr_content)
> }
> }
>
> -static const u32 core2_fix_counters_msr[] = {
> - MSR_CORE_PERF_FIXED_CTR0,
> - MSR_CORE_PERF_FIXED_CTR1,
> - MSR_CORE_PERF_FIXED_CTR2
> -};
> -
> /*
> - * MSR_CORE_PERF_FIXED_CTR_CTRL contains the configuration of all fixed
> - * counters. 4 bits for every counter.
> + * Read the number of general counters via CPUID.EAX[0xa].EAX[8..15]
> */
> -#define FIXED_CTR_CTRL_BITS 4
> -#define FIXED_CTR_CTRL_MASK ((1 << FIXED_CTR_CTRL_BITS) - 1)
> -
> -/* The index into the core2_ctrls_msr[] of this MSR used in
> core2_vpmu_dump() */
> -#define MSR_CORE_PERF_FIXED_CTR_CTRL_IDX 0
> -
> -/* Core 2 Non-architectual Performance Control MSRs. */
> -static const u32 core2_ctrls_msr[] = {
> - MSR_CORE_PERF_FIXED_CTR_CTRL,
> - MSR_IA32_PEBS_ENABLE,
> - MSR_IA32_DS_AREA
> -};
> -
> -struct pmumsr {
> - unsigned int num;
> - const u32 *msr;
> -};
> -
> -static const struct pmumsr core2_fix_counters = {
> - VPMU_CORE2_NUM_FIXED,
> - core2_fix_counters_msr
> -};
> +static int core2_get_arch_pmc_count(void)
> +{
> + u32 eax;
>
> -static const struct pmumsr core2_ctrls = {
> - VPMU_CORE2_NUM_CTRLS,
> - core2_ctrls_msr
> -};
> -static int arch_pmc_cnt;
> + eax = cpuid_eax(0xa);
> + return ( (eax & PMU_GENERAL_NR_MASK) >> PMU_GENERAL_NR_SHIFT );
> +}
>
> /*
> - * Read the number of general counters via CPUID.EAX[0xa].EAX[8..15]
> + * Read the number of fixed counters via CPUID.EDX[0xa].EDX[0..4]
> */
> -static int core2_get_pmc_count(void)
> +static int core2_get_fixed_pmc_count(void)
> {
> - u32 eax, ebx, ecx, edx;
> -
> - if ( arch_pmc_cnt == 0 )
> - {
> - cpuid(0xa, &eax, &ebx, &ecx, &edx);
> - arch_pmc_cnt = (eax & PMU_GENERAL_NR_MASK) >> PMU_GENERAL_NR_SHIFT;
> - }
> + u32 eax;
>
> - return arch_pmc_cnt;
> + eax = cpuid_eax(0xa);
> + return ( (eax & PMU_FIXED_NR_MASK) >> PMU_FIXED_NR_SHIFT );
> }
>
> static u64 core2_calc_intial_glb_ctrl_msr(void)
> {
> - int arch_pmc_bits = (1 << core2_get_pmc_count()) - 1;
> - u64 fix_pmc_bits = (1 << 3) - 1;
> - return ((fix_pmc_bits << 32) | arch_pmc_bits);
> + int arch_pmc_bits = (1 << arch_pmc_cnt) - 1;
> + u64 fix_pmc_bits = (1 << fixed_pmc_cnt) - 1;
> + return ( (fix_pmc_bits << 32) | arch_pmc_bits );
> }
>
> /* edx bits 5-12: Bit width of fixed-function performance counters */
> static int core2_get_bitwidth_fix_count(void)
> {
> - u32 eax, ebx, ecx, edx;
> + u32 edx;
>
> - cpuid(0xa, &eax, &ebx, &ecx, &edx);
> - return ((edx & PMU_FIXED_WIDTH_MASK) >> PMU_FIXED_WIDTH_SHIFT);
> + edx = cpuid_edx(0xa);
> + return ( (edx & PMU_FIXED_WIDTH_MASK) >> PMU_FIXED_WIDTH_SHIFT );
> }
>
> static int is_core2_vpmu_msr(u32 msr_index, int *type, int *index)
> @@ -201,9 +185,9 @@ static int is_core2_vpmu_msr(u32 msr_index, int *type,
> int *index)
> int i;
> u32 msr_index_pmc;
>
> - for ( i = 0; i < core2_fix_counters.num; i++ )
> + for ( i = 0; i < fixed_pmc_cnt; i++ )
> {
> - if ( core2_fix_counters.msr[i] == msr_index )
> + if ( msr_index == MSR_CORE_PERF_FIXED_CTR0 + i )
> {
> *type = MSR_TYPE_COUNTER;
> *index = i;
> @@ -211,14 +195,12 @@ static int is_core2_vpmu_msr(u32 msr_index, int *type,
> int *index)
> }
> }
>
> - for ( i = 0; i < core2_ctrls.num; i++ )
> + if ( (msr_index == MSR_CORE_PERF_FIXED_CTR_CTRL ) ||
> + (msr_index == MSR_IA32_DS_AREA) ||
> + (msr_index == MSR_IA32_PEBS_ENABLE) )
> {
> - if ( core2_ctrls.msr[i] == msr_index )
> - {
> - *type = MSR_TYPE_CTRL;
> - *index = i;
> - return 1;
> - }
> + *type = MSR_TYPE_CTRL;
> + return 1;
> }
>
> if ( (msr_index == MSR_CORE_PERF_GLOBAL_CTRL) ||
> @@ -231,7 +213,7 @@ static int is_core2_vpmu_msr(u32 msr_index, int *type,
> int *index)
>
> msr_index_pmc = msr_index & MSR_PMC_ALIAS_MASK;
> if ( (msr_index_pmc >= MSR_IA32_PERFCTR0) &&
> - (msr_index_pmc < (MSR_IA32_PERFCTR0 + core2_get_pmc_count())) )
> + (msr_index_pmc < (MSR_IA32_PERFCTR0 + arch_pmc_cnt)) )
> {
> *type = MSR_TYPE_ARCH_COUNTER;
> *index = msr_index_pmc - MSR_IA32_PERFCTR0;
> @@ -239,7 +221,7 @@ static int is_core2_vpmu_msr(u32 msr_index, int *type,
> int *index)
> }
>
> if ( (msr_index >= MSR_P6_EVNTSEL0) &&
> - (msr_index < (MSR_P6_EVNTSEL0 + core2_get_pmc_count())) )
> + (msr_index < (MSR_P6_EVNTSEL0 + arch_pmc_cnt)) )
> {
> *type = MSR_TYPE_ARCH_CTRL;
> *index = msr_index - MSR_P6_EVNTSEL0;
> @@ -254,13 +236,13 @@ static void core2_vpmu_set_msr_bitmap(unsigned long
> *msr_bitmap)
> int i;
>
> /* Allow Read/Write PMU Counters MSR Directly. */
> - for ( i = 0; i < core2_fix_counters.num; i++ )
> + for ( i = 0; i < fixed_pmc_cnt; i++ )
> {
> - clear_bit(msraddr_to_bitpos(core2_fix_counters.msr[i]), msr_bitmap);
> - clear_bit(msraddr_to_bitpos(core2_fix_counters.msr[i]),
> + clear_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR0 + i),
> msr_bitmap);
> + clear_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR0 + i),
> msr_bitmap + 0x800/BYTES_PER_LONG);
> }
> - for ( i = 0; i < core2_get_pmc_count(); i++ )
> + for ( i = 0; i < arch_pmc_cnt; i++ )
> {
> clear_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0+i), msr_bitmap);
> clear_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0+i),
> @@ -275,26 +257,28 @@ static void core2_vpmu_set_msr_bitmap(unsigned long
> *msr_bitmap)
> }
>
> /* Allow Read PMU Non-global Controls Directly. */
> - for ( i = 0; i < core2_ctrls.num; i++ )
> - clear_bit(msraddr_to_bitpos(core2_ctrls.msr[i]), msr_bitmap);
> - for ( i = 0; i < core2_get_pmc_count(); i++ )
> - clear_bit(msraddr_to_bitpos(MSR_P6_EVNTSEL0+i), msr_bitmap);
> + for ( i = 0; i < arch_pmc_cnt; i++ )
> + clear_bit(msraddr_to_bitpos(MSR_P6_EVNTSEL0 + i), msr_bitmap);
> +
> + clear_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR_CTRL), msr_bitmap);
> + clear_bit(msraddr_to_bitpos(MSR_IA32_PEBS_ENABLE), msr_bitmap);
> + clear_bit(msraddr_to_bitpos(MSR_IA32_DS_AREA), msr_bitmap);
> }
>
> static void core2_vpmu_unset_msr_bitmap(unsigned long *msr_bitmap)
> {
> int i;
>
> - for ( i = 0; i < core2_fix_counters.num; i++ )
> + for ( i = 0; i < fixed_pmc_cnt; i++ )
> {
> - set_bit(msraddr_to_bitpos(core2_fix_counters.msr[i]), msr_bitmap);
> - set_bit(msraddr_to_bitpos(core2_fix_counters.msr[i]),
> + set_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR0 + i), msr_bitmap);
> + set_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR0 + i),
> msr_bitmap + 0x800/BYTES_PER_LONG);
> }
> - for ( i = 0; i < core2_get_pmc_count(); i++ )
> + for ( i = 0; i < arch_pmc_cnt; i++ )
> {
> - set_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0+i), msr_bitmap);
> - set_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0+i),
> + set_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0 + i), msr_bitmap);
> + set_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0 + i),
> msr_bitmap + 0x800/BYTES_PER_LONG);
>
> if ( full_width_write )
> @@ -305,10 +289,12 @@ static void core2_vpmu_unset_msr_bitmap(unsigned long
> *msr_bitmap)
> }
> }
>
> - for ( i = 0; i < core2_ctrls.num; i++ )
> - set_bit(msraddr_to_bitpos(core2_ctrls.msr[i]), msr_bitmap);
> - for ( i = 0; i < core2_get_pmc_count(); i++ )
> - set_bit(msraddr_to_bitpos(MSR_P6_EVNTSEL0+i), msr_bitmap);
> + for ( i = 0; i < arch_pmc_cnt; i++ )
> + set_bit(msraddr_to_bitpos(MSR_P6_EVNTSEL0 + i), msr_bitmap);
> +
> + set_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR_CTRL), msr_bitmap);
> + set_bit(msraddr_to_bitpos(MSR_IA32_PEBS_ENABLE), msr_bitmap);
> + set_bit(msraddr_to_bitpos(MSR_IA32_DS_AREA), msr_bitmap);
> }
>
> static inline void __core2_vpmu_save(struct vcpu *v)
> @@ -316,10 +302,10 @@ static inline void __core2_vpmu_save(struct vcpu *v)
> int i;
> struct core2_vpmu_context *core2_vpmu_cxt = vcpu_vpmu(v)->context;
>
> - for ( i = 0; i < core2_fix_counters.num; i++ )
> - rdmsrl(core2_fix_counters.msr[i], core2_vpmu_cxt->fix_counters[i]);
> - for ( i = 0; i < core2_get_pmc_count(); i++ )
> - rdmsrl(MSR_IA32_PERFCTR0+i,
> core2_vpmu_cxt->arch_msr_pair[i].counter);
> + for ( i = 0; i < fixed_pmc_cnt; i++ )
> + rdmsrl(MSR_CORE_PERF_FIXED_CTR0 + i,
> core2_vpmu_cxt->fix_counters[i]);
> + for ( i = 0; i < arch_pmc_cnt; i++ )
> + rdmsrl(MSR_IA32_PERFCTR0 + i,
> core2_vpmu_cxt->arch_msr_pair[i].counter);
> }
>
> static int core2_vpmu_save(struct vcpu *v)
> @@ -343,20 +329,22 @@ static inline void __core2_vpmu_load(struct vcpu *v)
> unsigned int i, pmc_start;
> struct core2_vpmu_context *core2_vpmu_cxt = vcpu_vpmu(v)->context;
>
> - for ( i = 0; i < core2_fix_counters.num; i++ )
> - wrmsrl(core2_fix_counters.msr[i], core2_vpmu_cxt->fix_counters[i]);
> + for ( i = 0; i < fixed_pmc_cnt; i++ )
> + wrmsrl(MSR_CORE_PERF_FIXED_CTR0 + i,
> core2_vpmu_cxt->fix_counters[i]);
>
> if ( full_width_write )
> pmc_start = MSR_IA32_A_PERFCTR0;
> else
> pmc_start = MSR_IA32_PERFCTR0;
> - for ( i = 0; i < core2_get_pmc_count(); i++ )
> + for ( i = 0; i < arch_pmc_cnt; i++ )
> + {
> wrmsrl(pmc_start + i, core2_vpmu_cxt->arch_msr_pair[i].counter);
> + wrmsrl(MSR_P6_EVNTSEL0 + i,
> core2_vpmu_cxt->arch_msr_pair[i].control);
> + }
>
> - for ( i = 0; i < core2_ctrls.num; i++ )
> - wrmsrl(core2_ctrls.msr[i], core2_vpmu_cxt->ctrls[i]);
> - for ( i = 0; i < core2_get_pmc_count(); i++ )
> - wrmsrl(MSR_P6_EVNTSEL0+i, core2_vpmu_cxt->arch_msr_pair[i].control);
> + wrmsrl(MSR_CORE_PERF_FIXED_CTR_CTRL, core2_vpmu_cxt->fixed_ctrl);
> + wrmsrl(MSR_IA32_DS_AREA, core2_vpmu_cxt->ds_area);
> + wrmsrl(MSR_IA32_PEBS_ENABLE, core2_vpmu_cxt->pebs_enable);
> }
>
> static void core2_vpmu_load(struct vcpu *v)
> @@ -375,56 +363,39 @@ static int core2_vpmu_alloc_resource(struct vcpu *v)
> {
> struct vpmu_struct *vpmu = vcpu_vpmu(v);
> struct core2_vpmu_context *core2_vpmu_cxt;
> - struct core2_pmu_enable *pmu_enable;
>
> if ( !acquire_pmu_ownership(PMU_OWNER_HVM) )
> return 0;
>
> wrmsrl(MSR_CORE_PERF_GLOBAL_CTRL, 0);
> if ( vmx_add_host_load_msr(MSR_CORE_PERF_GLOBAL_CTRL) )
> - return 0;
> + goto out_err;
>
> if ( vmx_add_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL) )
> - return 0;
> + goto out_err;
> vmx_write_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL,
> core2_calc_intial_glb_ctrl_msr());
>
> - pmu_enable = xzalloc_bytes(sizeof(struct core2_pmu_enable) +
> - core2_get_pmc_count() - 1);
> - if ( !pmu_enable )
> - goto out1;
> -
> core2_vpmu_cxt = xzalloc_bytes(sizeof(struct core2_vpmu_context) +
> - (core2_get_pmc_count()-1)*sizeof(struct arch_msr_pair));
> + (arch_pmc_cnt-1)*sizeof(struct arch_msr_pair));
> if ( !core2_vpmu_cxt )
> - goto out2;
> - core2_vpmu_cxt->pmu_enable = pmu_enable;
> + goto out_err;
> +
> vpmu->context = (void *)core2_vpmu_cxt;
>
> + vpmu_set(vpmu, VPMU_CONTEXT_ALLOCATED);
> +
> return 1;
> - out2:
> - xfree(pmu_enable);
> - out1:
> - gdprintk(XENLOG_WARNING, "Insufficient memory for PMU, PMU feature is "
> - "unavailable on domain %d vcpu %d.\n",
> - v->vcpu_id, v->domain->domain_id);
> - return 0;
> -}
>
> -static void core2_vpmu_save_msr_context(struct vcpu *v, int type,
> - int index, u64 msr_data)
> -{
> - struct core2_vpmu_context *core2_vpmu_cxt = vcpu_vpmu(v)->context;
> +out_err:
> + vmx_rm_host_load_msr(MSR_CORE_PERF_GLOBAL_CTRL);
> + vmx_rm_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL);
> + release_pmu_ownship(PMU_OWNER_HVM);
>
> - switch ( type )
> - {
> - case MSR_TYPE_CTRL:
> - core2_vpmu_cxt->ctrls[index] = msr_data;
> - break;
> - case MSR_TYPE_ARCH_CTRL:
> - core2_vpmu_cxt->arch_msr_pair[index].control = msr_data;
> - break;
> - }
> + printk("Failed to allocate VPMU resources for domain %u vcpu %u\n",
> + v->vcpu_id, v->domain->domain_id);
> +
> + return 0;
> }
>
> static int core2_vpmu_msr_common_check(u32 msr_index, int *type, int *index)
> @@ -435,10 +406,8 @@ static int core2_vpmu_msr_common_check(u32 msr_index,
> int *type, int *index)
> return 0;
>
> if ( unlikely(!vpmu_is_set(vpmu, VPMU_CONTEXT_ALLOCATED)) &&
> - (vpmu->context != NULL ||
> - !core2_vpmu_alloc_resource(current)) )
> + !core2_vpmu_alloc_resource(current) )
> return 0;
> - vpmu_set(vpmu, VPMU_CONTEXT_ALLOCATED);
>
> /* Do the lazy load staff. */
> if ( !vpmu_is_set(vpmu, VPMU_CONTEXT_LOADED) )
> @@ -454,7 +423,7 @@ static int core2_vpmu_msr_common_check(u32 msr_index, int
> *type, int *index)
> static int core2_vpmu_do_wrmsr(unsigned int msr, uint64_t msr_content)
> {
> u64 global_ctrl, non_global_ctrl;
> - char pmu_enable = 0;
> + unsigned pmu_enable = 0;
> int i, tmp;
> int type = -1, index = -1;
> struct vcpu *v = current;
> @@ -499,6 +468,7 @@ static int core2_vpmu_do_wrmsr(unsigned int msr, uint64_t
> msr_content)
> if ( msr_content & 1 )
> gdprintk(XENLOG_WARNING, "Guest is trying to enable PEBS, "
> "which is not supported.\n");
> + core2_vpmu_cxt->pebs_enable = msr_content;
> return 1;
> case MSR_IA32_DS_AREA:
> if ( vpmu_is_set(vpmu, VPMU_CPU_HAS_DS) )
> @@ -511,27 +481,25 @@ static int core2_vpmu_do_wrmsr(unsigned int msr,
> uint64_t msr_content)
> hvm_inject_hw_exception(TRAP_gp_fault, 0);
> return 1;
> }
> - core2_vpmu_cxt->pmu_enable->ds_area_enable = msr_content ? 1 : 0;
> + core2_vpmu_cxt->ds_area = msr_content;
> break;
> }
> gdprintk(XENLOG_WARNING, "Guest setting of DTS is ignored.\n");
> return 1;
> case MSR_CORE_PERF_GLOBAL_CTRL:
> global_ctrl = msr_content;
> - for ( i = 0; i < core2_get_pmc_count(); i++ )
> + for ( i = 0; i < arch_pmc_cnt; i++ )
> {
> rdmsrl(MSR_P6_EVNTSEL0+i, non_global_ctrl);
> - core2_vpmu_cxt->pmu_enable->arch_pmc_enable[i] =
> - global_ctrl & (non_global_ctrl >> 22) & 1;
> + pmu_enable += global_ctrl & (non_global_ctrl >> 22) & 1;
> global_ctrl >>= 1;
> }
>
> rdmsrl(MSR_CORE_PERF_FIXED_CTR_CTRL, non_global_ctrl);
> global_ctrl = msr_content >> 32;
> - for ( i = 0; i < core2_fix_counters.num; i++ )
> + for ( i = 0; i < fixed_pmc_cnt; i++ )
> {
> - core2_vpmu_cxt->pmu_enable->fixed_ctr_enable[i] =
> - (global_ctrl & 1) & ((non_global_ctrl & 0x3)? 1: 0);
> + pmu_enable += (global_ctrl & 1) & ((non_global_ctrl & 0x3)? 1 :
> 0);
> non_global_ctrl >>= FIXED_CTR_CTRL_BITS;
> global_ctrl >>= 1;
> }
> @@ -540,27 +508,27 @@ static int core2_vpmu_do_wrmsr(unsigned int msr,
> uint64_t msr_content)
> non_global_ctrl = msr_content;
> vmx_read_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL, &global_ctrl);
> global_ctrl >>= 32;
> - for ( i = 0; i < core2_fix_counters.num; i++ )
> + for ( i = 0; i < fixed_pmc_cnt; i++ )
> {
> - core2_vpmu_cxt->pmu_enable->fixed_ctr_enable[i] =
> - (global_ctrl & 1) & ((non_global_ctrl & 0x3)? 1: 0);
> + pmu_enable += (global_ctrl & 1) & ((non_global_ctrl & 0x3)? 1 :
> 0);
> non_global_ctrl >>= 4;
> global_ctrl >>= 1;
> }
> + core2_vpmu_cxt->fixed_ctrl = msr_content;
> break;
> default:
> tmp = msr - MSR_P6_EVNTSEL0;
> - vmx_read_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL, &global_ctrl);
> - if ( tmp >= 0 && tmp < core2_get_pmc_count() )
> - core2_vpmu_cxt->pmu_enable->arch_pmc_enable[tmp] =
> - (global_ctrl >> tmp) & (msr_content >> 22) & 1;
> + if ( tmp >= 0 && tmp < arch_pmc_cnt )
> + {
> + vmx_read_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL, &global_ctrl);
> + core2_vpmu_cxt->arch_msr_pair[tmp].control = msr_content;
> + for ( i = 0; i < arch_pmc_cnt && !pmu_enable; i++ )
> + pmu_enable += (global_ctrl >> i) &
> + (core2_vpmu_cxt->arch_msr_pair[i].control >> 22) & 1;
> + }
> }
>
> - for ( i = 0; i < core2_fix_counters.num; i++ )
> - pmu_enable |= core2_vpmu_cxt->pmu_enable->fixed_ctr_enable[i];
> - for ( i = 0; i < core2_get_pmc_count(); i++ )
> - pmu_enable |= core2_vpmu_cxt->pmu_enable->arch_pmc_enable[i];
> - pmu_enable |= core2_vpmu_cxt->pmu_enable->ds_area_enable;
> + pmu_enable += (core2_vpmu_cxt->ds_area != 0);
Maybe it would be better to move this incrementing of pmu_enable to the
"case MSR_IA32_DS_AREA:" above where the ds_area gets handled ?
Dietmar.
> if ( pmu_enable )
> vpmu_set(vpmu, VPMU_RUNNING);
> else
> @@ -579,7 +547,6 @@ static int core2_vpmu_do_wrmsr(unsigned int msr, uint64_t
> msr_content)
> vpmu->hw_lapic_lvtpc = PMU_APIC_VECTOR | APIC_LVT_MASKED;
> }
>
> - core2_vpmu_save_msr_context(v, type, index, msr_content);
> if ( type != MSR_TYPE_GLOBAL )
> {
> u64 mask;
> @@ -595,7 +562,7 @@ static int core2_vpmu_do_wrmsr(unsigned int msr, uint64_t
> msr_content)
> if ( msr == MSR_IA32_DS_AREA )
> break;
> /* 4 bits per counter, currently 3 fixed counters implemented. */
> - mask = ~((1ull << (VPMU_CORE2_NUM_FIXED * FIXED_CTR_CTRL_BITS))
> - 1);
> + mask = ~((1ull << (fixed_pmc_cnt * FIXED_CTR_CTRL_BITS)) - 1);
> if (msr_content & mask)
> inject_gp = 1;
> break;
> @@ -680,7 +647,7 @@ static void core2_vpmu_do_cpuid(unsigned int input,
> static void core2_vpmu_dump(const struct vcpu *v)
> {
> const struct vpmu_struct *vpmu = vcpu_vpmu(v);
> - int i, num;
> + int i;
> const struct core2_vpmu_context *core2_vpmu_cxt = NULL;
> u64 val;
>
> @@ -698,27 +665,25 @@ static void core2_vpmu_dump(const struct vcpu *v)
>
> printk(" vPMU running\n");
> core2_vpmu_cxt = vpmu->context;
> - num = core2_get_pmc_count();
> +
> /* Print the contents of the counter and its configuration msr. */
> - for ( i = 0; i < num; i++ )
> + for ( i = 0; i < arch_pmc_cnt; i++ )
> {
> const struct arch_msr_pair *msr_pair = core2_vpmu_cxt->arch_msr_pair;
>
> - if ( core2_vpmu_cxt->pmu_enable->arch_pmc_enable[i] )
> - printk(" general_%d: 0x%016lx ctrl: 0x%016lx\n",
> - i, msr_pair[i].counter, msr_pair[i].control);
> + printk(" general_%d: 0x%016lx ctrl: 0x%016lx\n",
> + i, msr_pair[i].counter, msr_pair[i].control);
> }
> /*
> * The configuration of the fixed counter is 4 bits each in the
> * MSR_CORE_PERF_FIXED_CTR_CTRL.
> */
> - val = core2_vpmu_cxt->ctrls[MSR_CORE_PERF_FIXED_CTR_CTRL_IDX];
> - for ( i = 0; i < core2_fix_counters.num; i++ )
> + val = core2_vpmu_cxt->fixed_ctrl;
> + for ( i = 0; i < fixed_pmc_cnt; i++ )
> {
> - if ( core2_vpmu_cxt->pmu_enable->fixed_ctr_enable[i] )
> - printk(" fixed_%d: 0x%016lx ctrl: %#lx\n",
> - i, core2_vpmu_cxt->fix_counters[i],
> - val & FIXED_CTR_CTRL_MASK);
> + printk(" fixed_%d: 0x%016lx ctrl: %#lx\n",
> + i, core2_vpmu_cxt->fix_counters[i],
> + val & FIXED_CTR_CTRL_MASK);
> val >>= FIXED_CTR_CTRL_BITS;
> }
> }
> @@ -736,7 +701,7 @@ static int core2_vpmu_do_interrupt(struct cpu_user_regs
> *regs)
> if ( is_pmc_quirk )
> handle_pmc_quirk(msr_content);
> core2_vpmu_cxt->global_ovf_status |= msr_content;
> - msr_content = 0xC000000700000000 | ((1 << core2_get_pmc_count()) -
> 1);
> + msr_content = 0xC000000700000000 | ((1 << arch_pmc_cnt) - 1);
> wrmsrl(MSR_CORE_PERF_GLOBAL_OVF_CTRL, msr_content);
> }
> else
> @@ -799,18 +764,27 @@ static int core2_vpmu_initialise(struct vcpu *v,
> unsigned int vpmu_flags)
> }
> }
> func_out:
> +
> + arch_pmc_cnt = core2_get_arch_pmc_count();
> + fixed_pmc_cnt = core2_get_fixed_pmc_count();
> + if ( fixed_pmc_cnt > VPMU_CORE2_MAX_FIXED_PMCS )
> + {
> + fixed_pmc_cnt = VPMU_CORE2_MAX_FIXED_PMCS;
> + printk(XENLOG_G_WARNING "Limiting number of fixed counters to %d\n",
> + fixed_pmc_cnt);
> + }
> check_pmc_quirk();
> +
> return 0;
> }
>
> static void core2_vpmu_destroy(struct vcpu *v)
> {
> struct vpmu_struct *vpmu = vcpu_vpmu(v);
> - struct core2_vpmu_context *core2_vpmu_cxt = vpmu->context;
>
> if ( !vpmu_is_set(vpmu, VPMU_CONTEXT_ALLOCATED) )
> return;
> - xfree(core2_vpmu_cxt->pmu_enable);
> +
> xfree(vpmu->context);
> if ( cpu_has_vmx_msr_bitmap && is_hvm_domain(v->domain) )
> core2_vpmu_unset_msr_bitmap(v->arch.hvm_vmx.msr_bitmap);
> diff --git a/xen/include/asm-x86/hvm/vmx/vmcs.h
> b/xen/include/asm-x86/hvm/vmx/vmcs.h
> index ebaba5c..ed81cfb 100644
> --- a/xen/include/asm-x86/hvm/vmx/vmcs.h
> +++ b/xen/include/asm-x86/hvm/vmx/vmcs.h
> @@ -473,7 +473,9 @@ void vmx_enable_intercept_for_msr(struct vcpu *v, u32
> msr, int type);
> int vmx_read_guest_msr(u32 msr, u64 *val);
> int vmx_write_guest_msr(u32 msr, u64 val);
> int vmx_add_guest_msr(u32 msr);
> +void vmx_rm_guest_msr(u32 msr);
> int vmx_add_host_load_msr(u32 msr);
> +void vmx_rm_host_load_msr(u32 msr);
> void vmx_vmcs_switch(struct vmcs_struct *from, struct vmcs_struct *to);
> void vmx_set_eoi_exit_bitmap(struct vcpu *v, u8 vector);
> void vmx_clear_eoi_exit_bitmap(struct vcpu *v, u8 vector);
> diff --git a/xen/include/asm-x86/hvm/vmx/vpmu_core2.h
> b/xen/include/asm-x86/hvm/vmx/vpmu_core2.h
> index 60b05fd..410372d 100644
> --- a/xen/include/asm-x86/hvm/vmx/vpmu_core2.h
> +++ b/xen/include/asm-x86/hvm/vmx/vpmu_core2.h
> @@ -23,29 +23,10 @@
> #ifndef __ASM_X86_HVM_VPMU_CORE_H_
> #define __ASM_X86_HVM_VPMU_CORE_H_
>
> -/* Currently only 3 fixed counters are supported. */
> -#define VPMU_CORE2_NUM_FIXED 3
> -/* Currently only 3 Non-architectual Performance Control MSRs */
> -#define VPMU_CORE2_NUM_CTRLS 3
> -
> struct arch_msr_pair {
> u64 counter;
> u64 control;
> };
>
> -struct core2_pmu_enable {
> - char ds_area_enable;
> - char fixed_ctr_enable[VPMU_CORE2_NUM_FIXED];
> - char arch_pmc_enable[1];
> -};
> -
> -struct core2_vpmu_context {
> - struct core2_pmu_enable *pmu_enable;
> - u64 fix_counters[VPMU_CORE2_NUM_FIXED];
> - u64 ctrls[VPMU_CORE2_NUM_CTRLS];
> - u64 global_ovf_status;
> - struct arch_msr_pair arch_msr_pair[1];
> -};
> -
> #endif /* __ASM_X86_HVM_VPMU_CORE_H_ */
>
>
--
Company details: http://ts.fujitsu.com/imprint.html
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |