[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH RFC v3] x86/domctl: Don't pause the whole domain if only getting vcpu state
This patch adds the hvm_save_one_cpu_ctxt() function. It optimizes by only pausing the vcpu on all HVMSR_PER_VCPU save callbacks where only data for one VCPU is required. Signed-off-by: Alexandru Isaila <aisaila@xxxxxxxxxxxxxxx> --- Changes since V2: - Added functions for all the save_*_one cases - Added unpause to the error case in the hvm_save_one(). --- tools/tests/vhpet/emul.h | 3 +- tools/tests/vhpet/main.c | 2 +- xen/arch/x86/cpu/mcheck/vmce.c | 29 +++- xen/arch/x86/domctl.c | 2 - xen/arch/x86/hvm/hpet.c | 2 +- xen/arch/x86/hvm/hvm.c | 329 +++++++++++++++++++++++++---------------- xen/arch/x86/hvm/i8254.c | 2 +- xen/arch/x86/hvm/irq.c | 6 +- xen/arch/x86/hvm/mtrr.c | 58 +++++--- xen/arch/x86/hvm/pmtimer.c | 2 +- xen/arch/x86/hvm/rtc.c | 2 +- xen/arch/x86/hvm/save.c | 78 +++++++--- xen/arch/x86/hvm/vioapic.c | 2 +- xen/arch/x86/hvm/viridian.c | 30 +++- xen/arch/x86/hvm/vlapic.c | 23 ++- xen/arch/x86/hvm/vpic.c | 2 +- xen/include/asm-x86/hvm/hvm.h | 2 + xen/include/asm-x86/hvm/save.h | 5 +- 18 files changed, 381 insertions(+), 198 deletions(-) diff --git a/tools/tests/vhpet/emul.h b/tools/tests/vhpet/emul.h index 383acff..99d5bbd 100644 --- a/tools/tests/vhpet/emul.h +++ b/tools/tests/vhpet/emul.h @@ -296,7 +296,8 @@ struct hvm_hw_hpet }; typedef int (*hvm_save_handler)(struct domain *d, - hvm_domain_context_t *h); + hvm_domain_context_t *h, + unsigned int instance); typedef int (*hvm_load_handler)(struct domain *d, hvm_domain_context_t *h); diff --git a/tools/tests/vhpet/main.c b/tools/tests/vhpet/main.c index 6fe65ea..3d8e7f5 100644 --- a/tools/tests/vhpet/main.c +++ b/tools/tests/vhpet/main.c @@ -177,7 +177,7 @@ void __init hvm_register_savevm(uint16_t typecode, int do_save(uint16_t typecode, struct domain *d, hvm_domain_context_t *h) { - return hvm_sr_handlers[typecode].save(d, h); + return hvm_sr_handlers[typecode].save(d, h, d->max_vcpus); } int do_load(uint16_t typecode, struct domain *d, hvm_domain_context_t *h) diff --git a/xen/arch/x86/cpu/mcheck/vmce.c b/xen/arch/x86/cpu/mcheck/vmce.c index e07cd2f..98b12e0 100644 --- a/xen/arch/x86/cpu/mcheck/vmce.c +++ b/xen/arch/x86/cpu/mcheck/vmce.c @@ -349,20 +349,33 @@ int vmce_wrmsr(uint32_t msr, uint64_t val) return ret; } -static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) +void vmce_save_vcpu_ctxt_one(struct vcpu *v, struct hvm_vmce_vcpu *ctxt) +{ + ctxt->caps = v->arch.vmce.mcg_cap; + ctxt->mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2; + ctxt->mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2; + ctxt->mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl; +} + +static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct vcpu *v; int err = 0; - for_each_vcpu ( d, v ) + if( instance < d->max_vcpus ) + { + struct hvm_vmce_vcpu ctxt; + + v = d->vcpu[instance]; + vmce_save_vcpu_ctxt_one(v, &ctxt); + + err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt); + } + else for_each_vcpu ( d, v ) { - struct hvm_vmce_vcpu ctxt = { - .caps = v->arch.vmce.mcg_cap, - .mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2, - .mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2, - .mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl, - }; + struct hvm_vmce_vcpu ctxt; + vmce_save_vcpu_ctxt_one(v, &ctxt); err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt); if ( err ) break; diff --git a/xen/arch/x86/domctl.c b/xen/arch/x86/domctl.c index 540ba08..d3c4e14 100644 --- a/xen/arch/x86/domctl.c +++ b/xen/arch/x86/domctl.c @@ -624,12 +624,10 @@ long arch_do_domctl( !is_hvm_domain(d) ) break; - domain_pause(d); ret = hvm_save_one(d, domctl->u.hvmcontext_partial.type, domctl->u.hvmcontext_partial.instance, domctl->u.hvmcontext_partial.buffer, &domctl->u.hvmcontext_partial.bufsz); - domain_unpause(d); if ( !ret ) copyback = true; diff --git a/xen/arch/x86/hvm/hpet.c b/xen/arch/x86/hvm/hpet.c index 3ea895a..56f4691 100644 --- a/xen/arch/x86/hvm/hpet.c +++ b/xen/arch/x86/hvm/hpet.c @@ -509,7 +509,7 @@ static const struct hvm_mmio_ops hpet_mmio_ops = { }; -static int hpet_save(struct domain *d, hvm_domain_context_t *h) +static int hpet_save(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { HPETState *hp = domain_vhpet(d); struct vcpu *v = pt_global_vcpu_target(d); diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index 205b4cb..c7f38eb 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -728,15 +728,26 @@ void hvm_domain_destroy(struct domain *d) } } -static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t *h) +void hvm_save_tsc_adjust_one(struct vcpu *v, struct hvm_tsc_adjust *ctxt) +{ + ctxt->tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust; +} + +static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct vcpu *v; struct hvm_tsc_adjust ctxt; int err = 0; - for_each_vcpu ( d, v ) + if( instance < d->max_vcpus ) { - ctxt.tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust; + v = d->vcpu[instance]; + hvm_save_tsc_adjust_one(v, &ctxt); + err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt); + } + else for_each_vcpu ( d, v ) + { + hvm_save_tsc_adjust_one(v, &ctxt); err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt); if ( err ) break; @@ -768,117 +779,135 @@ static int hvm_load_tsc_adjust(struct domain *d, hvm_domain_context_t *h) HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust, hvm_load_tsc_adjust, 1, HVMSR_PER_VCPU); -static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t *h) +void hvm_save_cpu_ctxt_one(struct vcpu *v, struct hvm_hw_cpu *ctxt) +{ + struct segment_register seg; + + /* Architecture-specific vmcs/vmcb bits */ + hvm_funcs.save_cpu_ctxt(v, ctxt); + + ctxt->tsc = hvm_get_guest_tsc_fixed(v, v->domain->arch.hvm_domain.sync_tsc); + + ctxt->msr_tsc_aux = hvm_msr_tsc_aux(v); + + hvm_get_segment_register(v, x86_seg_idtr, &seg); + ctxt->idtr_limit = seg.limit; + ctxt->idtr_base = seg.base; + + hvm_get_segment_register(v, x86_seg_gdtr, &seg); + ctxt->gdtr_limit = seg.limit; + ctxt->gdtr_base = seg.base; + + hvm_get_segment_register(v, x86_seg_cs, &seg); + ctxt->cs_sel = seg.sel; + ctxt->cs_limit = seg.limit; + ctxt->cs_base = seg.base; + ctxt->cs_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_ds, &seg); + ctxt->ds_sel = seg.sel; + ctxt->ds_limit = seg.limit; + ctxt->ds_base = seg.base; + ctxt->ds_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_es, &seg); + ctxt->es_sel = seg.sel; + ctxt->es_limit = seg.limit; + ctxt->es_base = seg.base; + ctxt->es_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_ss, &seg); + ctxt->ss_sel = seg.sel; + ctxt->ss_limit = seg.limit; + ctxt->ss_base = seg.base; + ctxt->ss_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_fs, &seg); + ctxt->fs_sel = seg.sel; + ctxt->fs_limit = seg.limit; + ctxt->fs_base = seg.base; + ctxt->fs_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_gs, &seg); + ctxt->gs_sel = seg.sel; + ctxt->gs_limit = seg.limit; + ctxt->gs_base = seg.base; + ctxt->gs_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_tr, &seg); + ctxt->tr_sel = seg.sel; + ctxt->tr_limit = seg.limit; + ctxt->tr_base = seg.base; + ctxt->tr_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_ldtr, &seg); + ctxt->ldtr_sel = seg.sel; + ctxt->ldtr_limit = seg.limit; + ctxt->ldtr_base = seg.base; + ctxt->ldtr_arbytes = seg.attr; + + if ( v->fpu_initialised ) + { + memcpy(ctxt->fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt->fpu_regs)); + ctxt->flags = XEN_X86_FPU_INITIALISED; + } + + ctxt->rax = v->arch.user_regs.rax; + ctxt->rbx = v->arch.user_regs.rbx; + ctxt->rcx = v->arch.user_regs.rcx; + ctxt->rdx = v->arch.user_regs.rdx; + ctxt->rbp = v->arch.user_regs.rbp; + ctxt->rsi = v->arch.user_regs.rsi; + ctxt->rdi = v->arch.user_regs.rdi; + ctxt->rsp = v->arch.user_regs.rsp; + ctxt->rip = v->arch.user_regs.rip; + ctxt->rflags = v->arch.user_regs.rflags; + ctxt->r8 = v->arch.user_regs.r8; + ctxt->r9 = v->arch.user_regs.r9; + ctxt->r10 = v->arch.user_regs.r10; + ctxt->r11 = v->arch.user_regs.r11; + ctxt->r12 = v->arch.user_regs.r12; + ctxt->r13 = v->arch.user_regs.r13; + ctxt->r14 = v->arch.user_regs.r14; + ctxt->r15 = v->arch.user_regs.r15; + ctxt->dr0 = v->arch.debugreg[0]; + ctxt->dr1 = v->arch.debugreg[1]; + ctxt->dr2 = v->arch.debugreg[2]; + ctxt->dr3 = v->arch.debugreg[3]; + ctxt->dr6 = v->arch.debugreg[6]; + ctxt->dr7 = v->arch.debugreg[7]; +} + +static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct vcpu *v; struct hvm_hw_cpu ctxt; - struct segment_register seg; - for_each_vcpu ( d, v ) + if( instance < d->max_vcpus) { - /* We don't need to save state for a vcpu that is down; the restore + v = d->vcpu[instance]; + if ( v->pause_flags & VPF_down ) + return 1; + memset(&ctxt, 0, sizeof(ctxt)); + + hvm_save_cpu_ctxt_one(v, &ctxt); + + if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 ) + return 1; + } + else for_each_vcpu ( d, v ) + { + /* We don't need to save state for a vcpu that is down; the restore * code will leave it down if there is nothing saved. */ if ( v->pause_flags & VPF_down ) continue; memset(&ctxt, 0, sizeof(ctxt)); - /* Architecture-specific vmcs/vmcb bits */ - hvm_funcs.save_cpu_ctxt(v, &ctxt); - - ctxt.tsc = hvm_get_guest_tsc_fixed(v, d->arch.hvm_domain.sync_tsc); - - ctxt.msr_tsc_aux = hvm_msr_tsc_aux(v); - - hvm_get_segment_register(v, x86_seg_idtr, &seg); - ctxt.idtr_limit = seg.limit; - ctxt.idtr_base = seg.base; - - hvm_get_segment_register(v, x86_seg_gdtr, &seg); - ctxt.gdtr_limit = seg.limit; - ctxt.gdtr_base = seg.base; - - hvm_get_segment_register(v, x86_seg_cs, &seg); - ctxt.cs_sel = seg.sel; - ctxt.cs_limit = seg.limit; - ctxt.cs_base = seg.base; - ctxt.cs_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_ds, &seg); - ctxt.ds_sel = seg.sel; - ctxt.ds_limit = seg.limit; - ctxt.ds_base = seg.base; - ctxt.ds_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_es, &seg); - ctxt.es_sel = seg.sel; - ctxt.es_limit = seg.limit; - ctxt.es_base = seg.base; - ctxt.es_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_ss, &seg); - ctxt.ss_sel = seg.sel; - ctxt.ss_limit = seg.limit; - ctxt.ss_base = seg.base; - ctxt.ss_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_fs, &seg); - ctxt.fs_sel = seg.sel; - ctxt.fs_limit = seg.limit; - ctxt.fs_base = seg.base; - ctxt.fs_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_gs, &seg); - ctxt.gs_sel = seg.sel; - ctxt.gs_limit = seg.limit; - ctxt.gs_base = seg.base; - ctxt.gs_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_tr, &seg); - ctxt.tr_sel = seg.sel; - ctxt.tr_limit = seg.limit; - ctxt.tr_base = seg.base; - ctxt.tr_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_ldtr, &seg); - ctxt.ldtr_sel = seg.sel; - ctxt.ldtr_limit = seg.limit; - ctxt.ldtr_base = seg.base; - ctxt.ldtr_arbytes = seg.attr; - - if ( v->fpu_initialised ) - { - memcpy(ctxt.fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt.fpu_regs)); - ctxt.flags = XEN_X86_FPU_INITIALISED; - } - - ctxt.rax = v->arch.user_regs.rax; - ctxt.rbx = v->arch.user_regs.rbx; - ctxt.rcx = v->arch.user_regs.rcx; - ctxt.rdx = v->arch.user_regs.rdx; - ctxt.rbp = v->arch.user_regs.rbp; - ctxt.rsi = v->arch.user_regs.rsi; - ctxt.rdi = v->arch.user_regs.rdi; - ctxt.rsp = v->arch.user_regs.rsp; - ctxt.rip = v->arch.user_regs.rip; - ctxt.rflags = v->arch.user_regs.rflags; - ctxt.r8 = v->arch.user_regs.r8; - ctxt.r9 = v->arch.user_regs.r9; - ctxt.r10 = v->arch.user_regs.r10; - ctxt.r11 = v->arch.user_regs.r11; - ctxt.r12 = v->arch.user_regs.r12; - ctxt.r13 = v->arch.user_regs.r13; - ctxt.r14 = v->arch.user_regs.r14; - ctxt.r15 = v->arch.user_regs.r15; - ctxt.dr0 = v->arch.debugreg[0]; - ctxt.dr1 = v->arch.debugreg[1]; - ctxt.dr2 = v->arch.debugreg[2]; - ctxt.dr3 = v->arch.debugreg[3]; - ctxt.dr6 = v->arch.debugreg[6]; - ctxt.dr7 = v->arch.debugreg[7]; + hvm_save_cpu_ctxt_one(v, &ctxt); if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 ) - return 1; + return 1; } return 0; } @@ -1162,15 +1191,44 @@ HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, hvm_load_cpu_ctxt, save_area) + \ xstate_ctxt_size(xcr0)) -static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h) +void hvm_save_cpu_xsave_states_one(struct vcpu *v, struct hvm_hw_cpu_xsave *ctxt, hvm_domain_context_t *h) +{ + unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum); + + ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur]; + h->cur += size; + + ctxt->xfeature_mask = xfeature_mask; + ctxt->xcr0 = v->arch.xcr0; + ctxt->xcr0_accum = v->arch.xcr0_accum; +} + +static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h, + unsigned int instance) { struct vcpu *v; - struct hvm_hw_cpu_xsave *ctxt; + struct hvm_hw_cpu_xsave ctxt; if ( !cpu_has_xsave ) return 0; /* do nothing */ - for_each_vcpu ( d, v ) + if( instance < d->max_vcpus ) + { + unsigned int size; + + v = d->vcpu[instance]; + size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum); + + if ( !xsave_enabled(v) ) + return 1; + if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) ) + return 1; + hvm_save_cpu_xsave_states_one(v, &ctxt, h); + + expand_xsave_states(v, &ctxt.save_area, + size - offsetof(typeof(ctxt), save_area)); + } + else for_each_vcpu ( d, v ) { unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum); @@ -1178,14 +1236,10 @@ static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h) continue; if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) ) return 1; - ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur]; - h->cur += size; + hvm_save_cpu_xsave_states_one(v, &ctxt, h); - ctxt->xfeature_mask = xfeature_mask; - ctxt->xcr0 = v->arch.xcr0; - ctxt->xcr0_accum = v->arch.xcr0_accum; - expand_xsave_states(v, &ctxt->save_area, - size - offsetof(typeof(*ctxt), save_area)); + expand_xsave_states(v, &ctxt.save_area, + size - offsetof(typeof(ctxt), save_area)); } return 0; @@ -1324,31 +1378,58 @@ static int hvm_load_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h) #define HVM_CPU_MSR_SIZE(cnt) offsetof(struct hvm_msr, msr[cnt]) static unsigned int __read_mostly msr_count_max; -static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t *h) +void hvm_save_cpu_msrs_one(struct vcpu *v, struct hvm_msr *ctxt, hvm_domain_context_t *h) +{ + unsigned int i; + + ctxt = (struct hvm_msr *)&h->data[h->cur]; + ctxt->count = 0; + + if ( hvm_funcs.save_msr ) + hvm_funcs.save_msr(v, ctxt); + + ASSERT(ctxt->count <= msr_count_max); + + for ( i = 0; i < ctxt->count; ++i ) + ctxt->msr[i]._rsvd = 0; + +} + +static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t *h, + unsigned int instance) { struct vcpu *v; - for_each_vcpu ( d, v ) + if( instance < d->max_vcpus ) { - struct hvm_msr *ctxt; - unsigned int i; + struct hvm_msr ctxt; + + v = d->vcpu[instance]; if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id, HVM_CPU_MSR_SIZE(msr_count_max)) ) return 1; - ctxt = (struct hvm_msr *)&h->data[h->cur]; - ctxt->count = 0; - if ( hvm_funcs.save_msr ) - hvm_funcs.save_msr(v, ctxt); + hvm_save_cpu_msrs_one(v, &ctxt, h); - ASSERT(ctxt->count <= msr_count_max); + if ( ctxt.count ) + h->cur += HVM_CPU_MSR_SIZE(ctxt.count); + else + h->cur -= sizeof(struct hvm_save_descriptor); + } + + for_each_vcpu ( d, v ) + { + struct hvm_msr ctxt; + + if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id, + HVM_CPU_MSR_SIZE(msr_count_max)) ) + return 1; - for ( i = 0; i < ctxt->count; ++i ) - ctxt->msr[i]._rsvd = 0; + hvm_save_cpu_msrs_one(v, &ctxt, h); - if ( ctxt->count ) - h->cur += HVM_CPU_MSR_SIZE(ctxt->count); + if ( ctxt.count ) + h->cur += HVM_CPU_MSR_SIZE(ctxt.count); else h->cur -= sizeof(struct hvm_save_descriptor); } diff --git a/xen/arch/x86/hvm/i8254.c b/xen/arch/x86/hvm/i8254.c index 992f08d..143b64d 100644 --- a/xen/arch/x86/hvm/i8254.c +++ b/xen/arch/x86/hvm/i8254.c @@ -390,7 +390,7 @@ void pit_stop_channel0_irq(PITState *pit) spin_unlock(&pit->lock); } -static int pit_save(struct domain *d, hvm_domain_context_t *h) +static int pit_save(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { PITState *pit = domain_vpit(d); int rc; diff --git a/xen/arch/x86/hvm/irq.c b/xen/arch/x86/hvm/irq.c index e425df9..dbbf769 100644 --- a/xen/arch/x86/hvm/irq.c +++ b/xen/arch/x86/hvm/irq.c @@ -598,7 +598,7 @@ static int __init dump_irq_info_key_init(void) } __initcall(dump_irq_info_key_init); -static int irq_save_pci(struct domain *d, hvm_domain_context_t *h) +static int irq_save_pci(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct hvm_irq *hvm_irq = hvm_domain_irq(d); unsigned int asserted, pdev, pintx; @@ -630,7 +630,7 @@ static int irq_save_pci(struct domain *d, hvm_domain_context_t *h) return rc; } -static int irq_save_isa(struct domain *d, hvm_domain_context_t *h) +static int irq_save_isa(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct hvm_irq *hvm_irq = hvm_domain_irq(d); @@ -638,7 +638,7 @@ static int irq_save_isa(struct domain *d, hvm_domain_context_t *h) return ( hvm_save_entry(ISA_IRQ, 0, h, &hvm_irq->isa_irq) ); } -static int irq_save_link(struct domain *d, hvm_domain_context_t *h) +static int irq_save_link(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct hvm_irq *hvm_irq = hvm_domain_irq(d); diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c index b721c63..6dabe8b 100644 --- a/xen/arch/x86/hvm/mtrr.c +++ b/xen/arch/x86/hvm/mtrr.c @@ -666,36 +666,50 @@ int hvm_set_mem_pinned_cacheattr(struct domain *d, uint64_t gfn_start, return 0; } -static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t *h) +void hvm_save_mtrr_msr_one(struct vcpu *v, struct hvm_hw_mtrr *hw_mtrr) { + struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr; int i; + + hvm_get_guest_pat(v, &hw_mtrr->msr_pat_cr); + + hw_mtrr->msr_mtrr_def_type = mtrr_state->def_type + | (mtrr_state->enabled << 10); + hw_mtrr->msr_mtrr_cap = mtrr_state->mtrr_cap; + + for ( i = 0; i < MTRR_VCNT; i++ ) + { + /* save physbase */ + hw_mtrr->msr_mtrr_var[i*2] = + ((uint64_t*)mtrr_state->var_ranges)[i*2]; + /* save physmask */ + hw_mtrr->msr_mtrr_var[i*2+1] = + ((uint64_t*)mtrr_state->var_ranges)[i*2+1]; + } + + for ( i = 0; i < NUM_FIXED_MSR; i++ ) + hw_mtrr->msr_mtrr_fixed[i] = + ((uint64_t*)mtrr_state->fixed_ranges)[i]; + +} + +static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t *h, unsigned int instance) +{ struct vcpu *v; struct hvm_hw_mtrr hw_mtrr; - struct mtrr_state *mtrr_state; /* save mtrr&pat */ - for_each_vcpu(d, v) + if( instance < d->max_vcpus ) { - mtrr_state = &v->arch.hvm_vcpu.mtrr; - - hvm_get_guest_pat(v, &hw_mtrr.msr_pat_cr); + v = d->vcpu[instance]; - hw_mtrr.msr_mtrr_def_type = mtrr_state->def_type - | (mtrr_state->enabled << 10); - hw_mtrr.msr_mtrr_cap = mtrr_state->mtrr_cap; - - for ( i = 0; i < MTRR_VCNT; i++ ) - { - /* save physbase */ - hw_mtrr.msr_mtrr_var[i*2] = - ((uint64_t*)mtrr_state->var_ranges)[i*2]; - /* save physmask */ - hw_mtrr.msr_mtrr_var[i*2+1] = - ((uint64_t*)mtrr_state->var_ranges)[i*2+1]; - } + hvm_save_mtrr_msr_one(v, &hw_mtrr); - for ( i = 0; i < NUM_FIXED_MSR; i++ ) - hw_mtrr.msr_mtrr_fixed[i] = - ((uint64_t*)mtrr_state->fixed_ranges)[i]; + if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 ) + return 1; + } + else for_each_vcpu(d, v) + { + hvm_save_mtrr_msr_one(v, &hw_mtrr); if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 ) return 1; diff --git a/xen/arch/x86/hvm/pmtimer.c b/xen/arch/x86/hvm/pmtimer.c index b70c299..21dcdeb 100644 --- a/xen/arch/x86/hvm/pmtimer.c +++ b/xen/arch/x86/hvm/pmtimer.c @@ -249,7 +249,7 @@ static int handle_pmt_io( return X86EMUL_OKAY; } -static int acpi_save(struct domain *d, hvm_domain_context_t *h) +static int acpi_save(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct hvm_hw_acpi *acpi = &d->arch.hvm_domain.acpi; PMTState *s = &d->arch.hvm_domain.pl_time->vpmt; diff --git a/xen/arch/x86/hvm/rtc.c b/xen/arch/x86/hvm/rtc.c index bcfa169..83f339d 100644 --- a/xen/arch/x86/hvm/rtc.c +++ b/xen/arch/x86/hvm/rtc.c @@ -737,7 +737,7 @@ void rtc_migrate_timers(struct vcpu *v) } /* Save RTC hardware state */ -static int rtc_save(struct domain *d, hvm_domain_context_t *h) +static int rtc_save(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { RTCState *s = domain_vrtc(d); int rc; diff --git a/xen/arch/x86/hvm/save.c b/xen/arch/x86/hvm/save.c index 8984a23..a963d5f 100644 --- a/xen/arch/x86/hvm/save.c +++ b/xen/arch/x86/hvm/save.c @@ -138,6 +138,7 @@ int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int instance, int rv; hvm_domain_context_t ctxt = { }; const struct hvm_save_descriptor *desc; + bool is_single_instance = false; if ( d->is_dying || typecode > HVM_SAVE_CODE_MAX || @@ -145,41 +146,82 @@ int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int instance, !hvm_sr_handlers[typecode].save ) return -EINVAL; + if( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU && + instance < d->max_vcpus ) + is_single_instance = true; + ctxt.size = hvm_sr_handlers[typecode].size; - if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU ) + if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU && + instance == d->max_vcpus ) ctxt.size *= d->max_vcpus; + ctxt.data = xmalloc_bytes(ctxt.size); if ( !ctxt.data ) return -ENOMEM; - if ( (rv = hvm_sr_handlers[typecode].save(d, &ctxt)) != 0 ) + if( is_single_instance ) + vcpu_pause(d->vcpu[instance]); + else + domain_pause(d); + + if ( (rv = hvm_sr_handlers[typecode].save(d, &ctxt, instance)) != 0 ) + { printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16" (%d)\n", d->domain_id, typecode, rv); + + if( is_single_instance ) + vcpu_unpause(d->vcpu[instance]); + else + domain_unpause(d); + } else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) ) { uint32_t off; - for ( off = 0; off <= (ctxt.cur - sizeof(*desc)); off += desc->length ) + if( is_single_instance ) { - desc = (void *)(ctxt.data + off); + desc = (void *)(ctxt.data); /* Move past header */ - off += sizeof(*desc); + off = sizeof(*desc); if ( ctxt.cur < desc->length || off > ctxt.cur - desc->length ) - break; - if ( instance == desc->instance ) + rv = -EFAULT; + rv = 0; + if ( guest_handle_is_null(handle) ) + *bufsz = desc->length; + else if ( *bufsz < desc->length ) + rv = -ENOBUFS; + else if ( copy_to_guest(handle, ctxt.data + off, desc->length) ) + rv = -EFAULT; + else + *bufsz = desc->length; + vcpu_unpause(d->vcpu[instance]); + } + else + { + for ( off = 0; off <= (ctxt.cur - sizeof(*desc)); off += desc->length ) { - rv = 0; - if ( guest_handle_is_null(handle) ) - *bufsz = desc->length; - else if ( *bufsz < desc->length ) - rv = -ENOBUFS; - else if ( copy_to_guest(handle, ctxt.data + off, desc->length) ) - rv = -EFAULT; - else - *bufsz = desc->length; - break; + desc = (void *)(ctxt.data + off); + /* Move past header */ + off += sizeof(*desc); + if ( ctxt.cur < desc->length || + off > ctxt.cur - desc->length ) + break; + if ( instance == desc->instance ) + { + rv = 0; + if ( guest_handle_is_null(handle) ) + *bufsz = desc->length; + else if ( *bufsz < desc->length ) + rv = -ENOBUFS; + else if ( copy_to_guest(handle, ctxt.data + off, desc->length) ) + rv = -EFAULT; + else + *bufsz = desc->length; + break; + } } + domain_unpause(d); } } @@ -225,7 +267,7 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h) { printk(XENLOG_G_INFO "HVM%d save: %s\n", d->domain_id, hvm_sr_handlers[i].name); - if ( handler(d, h) != 0 ) + if ( handler(d, h, d->max_vcpus) != 0 ) { printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16"\n", diff --git a/xen/arch/x86/hvm/vioapic.c b/xen/arch/x86/hvm/vioapic.c index 97b419f..34d6907 100644 --- a/xen/arch/x86/hvm/vioapic.c +++ b/xen/arch/x86/hvm/vioapic.c @@ -569,7 +569,7 @@ int vioapic_get_trigger_mode(const struct domain *d, unsigned int gsi) return vioapic->redirtbl[pin].fields.trig_mode; } -static int ioapic_save(struct domain *d, hvm_domain_context_t *h) +static int ioapic_save(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct hvm_vioapic *s; diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c index f0fa59d..6415954 100644 --- a/xen/arch/x86/hvm/viridian.c +++ b/xen/arch/x86/hvm/viridian.c @@ -994,7 +994,7 @@ out: return HVM_HCALL_completed; } -static int viridian_save_domain_ctxt(struct domain *d, hvm_domain_context_t *h) +static int viridian_save_domain_ctxt(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct hvm_viridian_domain_context ctxt = { .time_ref_count = d->arch.hvm_domain.viridian.time_ref_count.val, @@ -1030,18 +1030,34 @@ static int viridian_load_domain_ctxt(struct domain *d, hvm_domain_context_t *h) HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN, viridian_save_domain_ctxt, viridian_load_domain_ctxt, 1, HVMSR_PER_DOM); -static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) +void viridian_save_vcpu_ctxt_one(struct vcpu *v, struct hvm_viridian_vcpu_context *ctxt) +{ + ctxt->vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw; + ctxt->vp_assist_vector = v->arch.hvm_vcpu.viridian.vp_assist.vector; +} + +static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct vcpu *v; if ( !is_viridian_domain(d) ) return 0; - for_each_vcpu( d, v ) { - struct hvm_viridian_vcpu_context ctxt = { - .vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw, - .vp_assist_vector = v->arch.hvm_vcpu.viridian.vp_assist.vector, - }; + if( instance < d->max_vcpus ) + { + struct hvm_viridian_vcpu_context ctxt; + + v = d->vcpu[instance]; + + viridian_save_vcpu_ctxt_one(v, &ctxt); + + if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 ) + return 1; + } + else for_each_vcpu( d, v ) { + struct hvm_viridian_vcpu_context ctxt; + + viridian_save_vcpu_ctxt_one(v, &ctxt); if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 ) return 1; diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c index 4bfc53e..591631a 100644 --- a/xen/arch/x86/hvm/vlapic.c +++ b/xen/arch/x86/hvm/vlapic.c @@ -1387,7 +1387,7 @@ static void lapic_rearm(struct vlapic *s) s->timer_last_update = s->pt.last_plt_gtime; } -static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h) +static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct vcpu *v; struct vlapic *s; @@ -1396,7 +1396,13 @@ static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h) if ( !has_vlapic(d) ) return 0; - for_each_vcpu ( d, v ) + if( instance < d->max_vcpus ) + { + v = d->vcpu[instance]; + s = vcpu_vlapic(v); + rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw); + } + else for_each_vcpu ( d, v ) { s = vcpu_vlapic(v); if ( (rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw)) != 0 ) @@ -1406,7 +1412,7 @@ static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h) return rc; } -static int lapic_save_regs(struct domain *d, hvm_domain_context_t *h) +static int lapic_save_regs(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct vcpu *v; struct vlapic *s; @@ -1415,7 +1421,16 @@ static int lapic_save_regs(struct domain *d, hvm_domain_context_t *h) if ( !has_vlapic(d) ) return 0; - for_each_vcpu ( d, v ) + if( instance < d->max_vcpus ) + { + v = d->vcpu[instance]; + if ( hvm_funcs.sync_pir_to_irr ) + hvm_funcs.sync_pir_to_irr(v); + + s = vcpu_vlapic(v); + rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs); + } + else for_each_vcpu ( d, v ) { if ( hvm_funcs.sync_pir_to_irr ) hvm_funcs.sync_pir_to_irr(v); diff --git a/xen/arch/x86/hvm/vpic.c b/xen/arch/x86/hvm/vpic.c index e160bbd..6b77f3c 100644 --- a/xen/arch/x86/hvm/vpic.c +++ b/xen/arch/x86/hvm/vpic.c @@ -371,7 +371,7 @@ static int vpic_intercept_elcr_io( return X86EMUL_OKAY; } -static int vpic_save(struct domain *d, hvm_domain_context_t *h) +static int vpic_save(struct domain *d, hvm_domain_context_t *h, unsigned int instance) { struct hvm_hw_vpic *s; int i; diff --git a/xen/include/asm-x86/hvm/hvm.h b/xen/include/asm-x86/hvm/hvm.h index b687e03..9f13112 100644 --- a/xen/include/asm-x86/hvm/hvm.h +++ b/xen/include/asm-x86/hvm/hvm.h @@ -609,6 +609,8 @@ static inline bool altp2m_vcpu_emulate_ve(struct vcpu *v) return false; } +void hvm_save_cpu_ctxt_one(struct vcpu *v, struct hvm_hw_cpu *ctxt); + /* Check CR4/EFER values */ const char *hvm_efer_valid(const struct vcpu *v, uint64_t value, signed int cr0_pg); diff --git a/xen/include/asm-x86/hvm/save.h b/xen/include/asm-x86/hvm/save.h index f889e8f..a2c39c4 100644 --- a/xen/include/asm-x86/hvm/save.h +++ b/xen/include/asm-x86/hvm/save.h @@ -95,8 +95,9 @@ static inline uint16_t hvm_load_instance(struct hvm_domain_context *h) * The save handler may save multiple instances of a type into the buffer; * the load handler will be called once for each instance found when * restoring. Both return non-zero on error. */ -typedef int (*hvm_save_handler) (struct domain *d, - hvm_domain_context_t *h); +typedef int (*hvm_save_handler) (struct domain *d, + hvm_domain_context_t *h, + unsigned int instance); typedef int (*hvm_load_handler) (struct domain *d, hvm_domain_context_t *h); -- 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |