[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v19 12/13] x86/hvm: Remove redundant save functions
This patch removes the redundant save functions and renames the save_one* to save. It then changes the domain param to vcpu in the save funcs and adapts print messages in order to match the format of the other save related messages. Signed-off-by: Alexandru Isaila <aisaila@xxxxxxxxxxxxxxx> --- Changes since V18: - Add const struct domain to rtc_save and hpet_save - Latched the vCPU into a local variable in hvm_save_one() - Add HVMSR_PER_VCPU kind check to the bounds if. --- xen/arch/x86/cpu/mcheck/vmce.c | 18 +------- xen/arch/x86/emul-i8254.c | 5 ++- xen/arch/x86/hvm/hpet.c | 7 ++-- xen/arch/x86/hvm/hvm.c | 75 +++------------------------------- xen/arch/x86/hvm/irq.c | 15 ++++--- xen/arch/x86/hvm/mtrr.c | 22 ++-------- xen/arch/x86/hvm/pmtimer.c | 5 ++- xen/arch/x86/hvm/rtc.c | 5 ++- xen/arch/x86/hvm/save.c | 28 +++++++------ xen/arch/x86/hvm/vioapic.c | 5 ++- xen/arch/x86/hvm/viridian.c | 23 ++--------- xen/arch/x86/hvm/vlapic.c | 38 ++--------------- xen/arch/x86/hvm/vpic.c | 5 ++- xen/include/asm-x86/hvm/save.h | 8 +--- 14 files changed, 63 insertions(+), 196 deletions(-) diff --git a/xen/arch/x86/cpu/mcheck/vmce.c b/xen/arch/x86/cpu/mcheck/vmce.c index 71afc06f9a..f15835e9f6 100644 --- a/xen/arch/x86/cpu/mcheck/vmce.c +++ b/xen/arch/x86/cpu/mcheck/vmce.c @@ -350,7 +350,7 @@ int vmce_wrmsr(uint32_t msr, uint64_t val) } #if CONFIG_HVM -static int vmce_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) +static int vmce_save_vcpu_ctxt(struct vcpu *v, hvm_domain_context_t *h) { struct hvm_vmce_vcpu ctxt = { .caps = v->arch.vmce.mcg_cap, @@ -362,21 +362,6 @@ static int vmce_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) return hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt); } -static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - int err = 0; - - for_each_vcpu ( d, v ) - { - err = vmce_save_vcpu_ctxt_one(v, h); - if ( err ) - break; - } - - return err; -} - static int vmce_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) { unsigned int vcpuid = hvm_load_instance(h); @@ -397,7 +382,6 @@ static int vmce_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) } HVM_REGISTER_SAVE_RESTORE(VMCE_VCPU, vmce_save_vcpu_ctxt, - vmce_save_vcpu_ctxt_one, vmce_load_vcpu_ctxt, 1, HVMSR_PER_VCPU); #endif diff --git a/xen/arch/x86/emul-i8254.c b/xen/arch/x86/emul-i8254.c index a85dfcccbc..73be4188ad 100644 --- a/xen/arch/x86/emul-i8254.c +++ b/xen/arch/x86/emul-i8254.c @@ -391,8 +391,9 @@ void pit_stop_channel0_irq(PITState *pit) spin_unlock(&pit->lock); } -static int pit_save(struct domain *d, hvm_domain_context_t *h) +static int pit_save(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; PITState *pit = domain_vpit(d); int rc; @@ -438,7 +439,7 @@ static int pit_load(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(PIT, pit_save, NULL, pit_load, 1, HVMSR_PER_DOM); +HVM_REGISTER_SAVE_RESTORE(PIT, pit_save, pit_load, 1, HVMSR_PER_DOM); #endif void pit_reset(struct domain *d) diff --git a/xen/arch/x86/hvm/hpet.c b/xen/arch/x86/hvm/hpet.c index 4d8f6da2d9..be371ecc0b 100644 --- a/xen/arch/x86/hvm/hpet.c +++ b/xen/arch/x86/hvm/hpet.c @@ -570,16 +570,17 @@ static const struct hvm_mmio_ops hpet_mmio_ops = { }; -static int hpet_save(struct domain *d, hvm_domain_context_t *h) +static int hpet_save(struct vcpu *v, hvm_domain_context_t *h) { + const struct domain *d = v->domain; HPETState *hp = domain_vhpet(d); - struct vcpu *v = pt_global_vcpu_target(d); int rc; uint64_t guest_time; if ( !has_vhpet(d) ) return 0; + v = pt_global_vcpu_target(d); write_lock(&hp->lock); guest_time = (v->arch.hvm.guest_time ?: hvm_get_guest_time(v)) / STIME_PER_HPET_TICK; @@ -695,7 +696,7 @@ static int hpet_load(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(HPET, hpet_save, NULL, hpet_load, 1, HVMSR_PER_DOM); +HVM_REGISTER_SAVE_RESTORE(HPET, hpet_save, hpet_load, 1, HVMSR_PER_DOM); static void hpet_set(HPETState *h) { diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index 58c03bed15..43145586c5 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -731,7 +731,7 @@ void hvm_domain_destroy(struct domain *d) destroy_vpci_mmcfg(d); } -static int hvm_save_tsc_adjust_one(struct vcpu *v, hvm_domain_context_t *h) +static int hvm_save_tsc_adjust(struct vcpu *v, hvm_domain_context_t *h) { struct hvm_tsc_adjust ctxt = { .tsc_adjust = v->arch.hvm.msr_tsc_adjust, @@ -740,21 +740,6 @@ static int hvm_save_tsc_adjust_one(struct vcpu *v, hvm_domain_context_t *h) return hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt); } -static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - int err = 0; - - for_each_vcpu ( d, v ) - { - err = hvm_save_tsc_adjust_one(v, h); - if ( err ) - break; - } - - return err; -} - static int hvm_load_tsc_adjust(struct domain *d, hvm_domain_context_t *h) { unsigned int vcpuid = hvm_load_instance(h); @@ -776,10 +761,9 @@ static int hvm_load_tsc_adjust(struct domain *d, hvm_domain_context_t *h) } HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust, - hvm_save_tsc_adjust_one, hvm_load_tsc_adjust, 1, HVMSR_PER_VCPU); -static int hvm_save_cpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) +static int hvm_save_cpu_ctxt(struct vcpu *v, hvm_domain_context_t *h) { struct segment_register seg; struct hvm_hw_cpu ctxt = { @@ -886,21 +870,6 @@ static int hvm_save_cpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) return hvm_save_entry(CPU, v->vcpu_id, h, &ctxt); } -static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - int err = 0; - - for_each_vcpu ( d, v ) - { - err = hvm_save_cpu_ctxt_one(v, h); - if ( err ) - break; - } - - return err; -} - /* Return a string indicating the error, or NULL for valid. */ const char *hvm_efer_valid(const struct vcpu *v, uint64_t value, signed int cr0_pg) @@ -1157,14 +1126,14 @@ static int hvm_load_cpu_ctxt(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, hvm_save_cpu_ctxt_one, - hvm_load_cpu_ctxt, 1, HVMSR_PER_VCPU); +HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, hvm_load_cpu_ctxt, 1, + HVMSR_PER_VCPU); #define HVM_CPU_XSAVE_SIZE(xcr0) (offsetof(struct hvm_hw_cpu_xsave, \ save_area) + \ xstate_ctxt_size(xcr0)) -static int hvm_save_cpu_xsave_states_one(struct vcpu *v, hvm_domain_context_t *h) +static int hvm_save_cpu_xsave_states(struct vcpu *v, hvm_domain_context_t *h) { struct hvm_hw_cpu_xsave *ctxt; unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum); @@ -1189,21 +1158,6 @@ static int hvm_save_cpu_xsave_states_one(struct vcpu *v, hvm_domain_context_t *h return 0; } -static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - int err = 0; - - for_each_vcpu ( d, v ) - { - err = hvm_save_cpu_xsave_states_one(v, h); - if ( err ) - break; - } - - return err; -} - /* * Structure layout conformity checks, documenting correctness of the cast in * the invocation of validate_xstate() below. @@ -1340,7 +1294,7 @@ static const uint32_t msrs_to_send[] = { }; static unsigned int __read_mostly msr_count_max = ARRAY_SIZE(msrs_to_send); -static int hvm_save_cpu_msrs_one(struct vcpu *v, hvm_domain_context_t *h) +static int hvm_save_cpu_msrs(struct vcpu *v, hvm_domain_context_t *h) { struct hvm_save_descriptor *desc = _p(&h->data[h->cur]); struct hvm_msr *ctxt; @@ -1402,21 +1356,6 @@ static int hvm_save_cpu_msrs_one(struct vcpu *v, hvm_domain_context_t *h) return 0; } -static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - int err = 0; - - for_each_vcpu ( d, v ) - { - err = hvm_save_cpu_msrs_one(v, h); - if ( err ) - break; - } - - return err; -} - static int hvm_load_cpu_msrs(struct domain *d, hvm_domain_context_t *h) { unsigned int i, vcpuid = hvm_load_instance(h); @@ -1509,7 +1448,6 @@ static int __init hvm_register_CPU_save_and_restore(void) hvm_register_savevm(CPU_XSAVE_CODE, "CPU_XSAVE", hvm_save_cpu_xsave_states, - hvm_save_cpu_xsave_states_one, hvm_load_cpu_xsave_states, HVM_CPU_XSAVE_SIZE(xfeature_mask) + sizeof(struct hvm_save_descriptor), @@ -1522,7 +1460,6 @@ static int __init hvm_register_CPU_save_and_restore(void) hvm_register_savevm(CPU_MSR_CODE, "CPU_MSR", hvm_save_cpu_msrs, - hvm_save_cpu_msrs_one, hvm_load_cpu_msrs, HVM_CPU_MSR_SIZE(msr_count_max) + sizeof(struct hvm_save_descriptor), diff --git a/xen/arch/x86/hvm/irq.c b/xen/arch/x86/hvm/irq.c index 9502bae645..37040d1d47 100644 --- a/xen/arch/x86/hvm/irq.c +++ b/xen/arch/x86/hvm/irq.c @@ -639,8 +639,9 @@ static int __init dump_irq_info_key_init(void) } __initcall(dump_irq_info_key_init); -static int irq_save_pci(struct domain *d, hvm_domain_context_t *h) +static int irq_save_pci(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_irq *hvm_irq = hvm_domain_irq(d); unsigned int asserted, pdev, pintx; int rc; @@ -671,16 +672,18 @@ static int irq_save_pci(struct domain *d, hvm_domain_context_t *h) return rc; } -static int irq_save_isa(struct domain *d, hvm_domain_context_t *h) +static int irq_save_isa(struct vcpu *v, hvm_domain_context_t *h) { + const struct domain *d = v->domain; struct hvm_irq *hvm_irq = hvm_domain_irq(d); /* Save ISA IRQ lines */ return ( hvm_save_entry(ISA_IRQ, 0, h, &hvm_irq->isa_irq) ); } -static int irq_save_link(struct domain *d, hvm_domain_context_t *h) +static int irq_save_link(struct vcpu *v, hvm_domain_context_t *h) { + const struct domain *d = v->domain; struct hvm_irq *hvm_irq = hvm_domain_irq(d); /* Save PCI-ISA link state */ @@ -773,9 +776,9 @@ static int irq_load_link(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(PCI_IRQ, irq_save_pci, NULL, irq_load_pci, +HVM_REGISTER_SAVE_RESTORE(PCI_IRQ, irq_save_pci, irq_load_pci, 1, HVMSR_PER_DOM); -HVM_REGISTER_SAVE_RESTORE(ISA_IRQ, irq_save_isa, NULL, irq_load_isa, +HVM_REGISTER_SAVE_RESTORE(ISA_IRQ, irq_save_isa, irq_load_isa, 1, HVMSR_PER_DOM); -HVM_REGISTER_SAVE_RESTORE(PCI_LINK, irq_save_link, NULL, irq_load_link, +HVM_REGISTER_SAVE_RESTORE(PCI_LINK, irq_save_link, irq_load_link, 1, HVMSR_PER_DOM); diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c index 2ddf5074cb..4f2f195f7d 100644 --- a/xen/arch/x86/hvm/mtrr.c +++ b/xen/arch/x86/hvm/mtrr.c @@ -690,7 +690,7 @@ int hvm_set_mem_pinned_cacheattr(struct domain *d, uint64_t gfn_start, return 0; } -static int hvm_save_mtrr_msr_one(struct vcpu *v, hvm_domain_context_t *h) +static int hvm_save_mtrr_msr(struct vcpu *v, hvm_domain_context_t *h) { const struct mtrr_state *mtrr_state = &v->arch.hvm.mtrr; struct hvm_hw_mtrr hw_mtrr = { @@ -728,22 +728,6 @@ static int hvm_save_mtrr_msr_one(struct vcpu *v, hvm_domain_context_t *h) return hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr); } -static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - int err = 0; - - /* save mtrr&pat */ - for_each_vcpu(d, v) - { - err = hvm_save_mtrr_msr_one(v, h); - if ( err ) - break; - } - - return err; -} - static int hvm_load_mtrr_msr(struct domain *d, hvm_domain_context_t *h) { unsigned int vcpuid, i; @@ -794,8 +778,8 @@ static int hvm_load_mtrr_msr(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr, hvm_save_mtrr_msr_one, - hvm_load_mtrr_msr, 1, HVMSR_PER_VCPU); +HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr, hvm_load_mtrr_msr, 1, + HVMSR_PER_VCPU); void memory_type_changed(struct domain *d) { diff --git a/xen/arch/x86/hvm/pmtimer.c b/xen/arch/x86/hvm/pmtimer.c index bd2fd35308..1f0d834bae 100644 --- a/xen/arch/x86/hvm/pmtimer.c +++ b/xen/arch/x86/hvm/pmtimer.c @@ -249,8 +249,9 @@ static int handle_pmt_io( return X86EMUL_OKAY; } -static int acpi_save(struct domain *d, hvm_domain_context_t *h) +static int acpi_save(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_hw_acpi *acpi = &d->arch.hvm.acpi; PMTState *s = &d->arch.hvm.pl_time->vpmt; uint32_t x, msb = acpi->tmr_val & TMR_VAL_MSB; @@ -309,7 +310,7 @@ static int acpi_load(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(PMTIMER, acpi_save, NULL, acpi_load, +HVM_REGISTER_SAVE_RESTORE(PMTIMER, acpi_save, acpi_load, 1, HVMSR_PER_DOM); int pmtimer_change_ioport(struct domain *d, unsigned int version) diff --git a/xen/arch/x86/hvm/rtc.c b/xen/arch/x86/hvm/rtc.c index eb67ce01a2..42339682e8 100644 --- a/xen/arch/x86/hvm/rtc.c +++ b/xen/arch/x86/hvm/rtc.c @@ -737,8 +737,9 @@ void rtc_migrate_timers(struct vcpu *v) } /* Save RTC hardware state */ -static int rtc_save(struct domain *d, hvm_domain_context_t *h) +static int rtc_save(struct vcpu *v, hvm_domain_context_t *h) { + const struct domain *d = v->domain; RTCState *s = domain_vrtc(d); int rc; @@ -783,7 +784,7 @@ static int rtc_load(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(RTC, rtc_save, NULL, rtc_load, 1, HVMSR_PER_DOM); +HVM_REGISTER_SAVE_RESTORE(RTC, rtc_save, rtc_load, 1, HVMSR_PER_DOM); void rtc_reset(struct domain *d) { diff --git a/xen/arch/x86/hvm/save.c b/xen/arch/x86/hvm/save.c index e059ab4e13..c7e2ecdb9f 100644 --- a/xen/arch/x86/hvm/save.c +++ b/xen/arch/x86/hvm/save.c @@ -85,7 +85,6 @@ int arch_hvm_load(struct domain *d, struct hvm_save_header *hdr) /* List of handlers for various HVM save and restore types */ static struct { hvm_save_handler save; - hvm_save_vcpu_handler save_one; hvm_load_handler load; const char *name; size_t size; @@ -96,7 +95,6 @@ static struct { void __init hvm_register_savevm(uint16_t typecode, const char *name, hvm_save_handler save_state, - hvm_save_vcpu_handler save_one, hvm_load_handler load_state, size_t size, int kind) { @@ -104,7 +102,6 @@ void __init hvm_register_savevm(uint16_t typecode, ASSERT(hvm_sr_handlers[typecode].save == NULL); ASSERT(hvm_sr_handlers[typecode].load == NULL); hvm_sr_handlers[typecode].save = save_state; - hvm_sr_handlers[typecode].save_one = save_one; hvm_sr_handlers[typecode].load = load_state; hvm_sr_handlers[typecode].name = name; hvm_sr_handlers[typecode].size = size; @@ -141,6 +138,8 @@ int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int instance, int rv; hvm_domain_context_t ctxt = { }; const struct hvm_save_descriptor *desc; + struct vcpu *v = (hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU) ? + d->vcpu[instance] : d->vcpu[0]; if ( d->is_dying || typecode > HVM_SAVE_CODE_MAX || @@ -148,14 +147,15 @@ int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int instance, !hvm_sr_handlers[typecode].save ) return -EINVAL; + if ( instance >= d->max_vcpus && + hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU ) + return -ENOENT; ctxt.size = hvm_sr_handlers[typecode].size; - if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU ) - ctxt.size *= d->max_vcpus; ctxt.data = xmalloc_bytes(ctxt.size); if ( !ctxt.data ) return -ENOMEM; - if ( (rv = hvm_sr_handlers[typecode].save(d, &ctxt)) != 0 ) + if ( (rv = hvm_sr_handlers[typecode].save(v, &ctxt)) != 0 ) printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16" (%d)\n", d->domain_id, typecode, rv); else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) ) @@ -222,10 +222,12 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h) /* Save all available kinds of state */ for ( i = 0; i <= HVM_SAVE_CODE_MAX; i++ ) { - hvm_save_vcpu_handler save_one_handler = hvm_sr_handlers[i].save_one; hvm_save_handler handler = hvm_sr_handlers[i].save; - if ( save_one_handler ) + if ( !handler ) + continue; + + if ( hvm_sr_handlers[i].kind == HVMSR_PER_VCPU ) { struct vcpu *v; @@ -233,7 +235,7 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h) { printk(XENLOG_G_INFO "HVM %pv save: %s\n", v, hvm_sr_handlers[i].name); - if ( save_one_handler(v, h) != 0 ) + if ( handler(v, h) != 0 ) { printk(XENLOG_G_ERR "HVM %pv save: failed to save type %"PRIu16"\n", @@ -242,14 +244,14 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h) } } } - else if ( handler ) + else { - printk(XENLOG_G_INFO "HVM%d save: %s\n", + printk(XENLOG_G_INFO "HVM d%d save: %s\n", d->domain_id, hvm_sr_handlers[i].name); - if ( handler(d, h) != 0 ) + if ( handler(d->vcpu[0], h) != 0 ) { printk(XENLOG_G_ERR - "HVM%d save: failed to save type %"PRIu16"\n", + "HVM d%d save: failed to save type %"PRIu16"\n", d->domain_id, i); return -ENODATA; } diff --git a/xen/arch/x86/hvm/vioapic.c b/xen/arch/x86/hvm/vioapic.c index b2cffe14d6..2b74f92d51 100644 --- a/xen/arch/x86/hvm/vioapic.c +++ b/xen/arch/x86/hvm/vioapic.c @@ -569,8 +569,9 @@ int vioapic_get_trigger_mode(const struct domain *d, unsigned int gsi) return vioapic->redirtbl[pin].fields.trig_mode; } -static int ioapic_save(struct domain *d, hvm_domain_context_t *h) +static int ioapic_save(struct vcpu *v, hvm_domain_context_t *h) { + const struct domain *d = v->domain; struct hvm_vioapic *s; if ( !has_vioapic(d) ) @@ -601,7 +602,7 @@ static int ioapic_load(struct domain *d, hvm_domain_context_t *h) return hvm_load_entry(IOAPIC, h, &s->domU); } -HVM_REGISTER_SAVE_RESTORE(IOAPIC, ioapic_save, NULL, ioapic_load, 1, HVMSR_PER_DOM); +HVM_REGISTER_SAVE_RESTORE(IOAPIC, ioapic_save, ioapic_load, 1, HVMSR_PER_DOM); void vioapic_reset(struct domain *d) { diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c index 3859333645..d7ab94bbab 100644 --- a/xen/arch/x86/hvm/viridian.c +++ b/xen/arch/x86/hvm/viridian.c @@ -994,8 +994,9 @@ out: return HVM_HCALL_completed; } -static int viridian_save_domain_ctxt(struct domain *d, hvm_domain_context_t *h) +static int viridian_save_domain_ctxt(struct vcpu *v, hvm_domain_context_t *h) { + const struct domain *d = v->domain; struct hvm_viridian_domain_context ctxt = { .time_ref_count = d->arch.hvm.viridian.time_ref_count.val, .hypercall_gpa = d->arch.hvm.viridian.hypercall_gpa.raw, @@ -1027,10 +1028,10 @@ static int viridian_load_domain_ctxt(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN, viridian_save_domain_ctxt, NULL, +HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN, viridian_save_domain_ctxt, viridian_load_domain_ctxt, 1, HVMSR_PER_DOM); -static int viridian_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) +static int viridian_save_vcpu_ctxt(struct vcpu *v, hvm_domain_context_t *h) { struct hvm_viridian_vcpu_context ctxt = { .vp_assist_msr = v->arch.hvm.viridian.vp_assist.msr.raw, @@ -1043,21 +1044,6 @@ static int viridian_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) return hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt); } -static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - int err = 0; - - for_each_vcpu ( d, v ) - { - err = viridian_save_vcpu_ctxt_one(v, h); - if ( err ) - break; - } - - return err; -} - static int viridian_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) { unsigned int vcpuid = hvm_load_instance(h); @@ -1088,7 +1074,6 @@ static int viridian_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) } HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_VCPU, viridian_save_vcpu_ctxt, - viridian_save_vcpu_ctxt_one, viridian_load_vcpu_ctxt, 1, HVMSR_PER_VCPU); static int __init parse_viridian_version(const char *arg) diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c index ee2c66ef9c..8a4a17311f 100644 --- a/xen/arch/x86/hvm/vlapic.c +++ b/xen/arch/x86/hvm/vlapic.c @@ -1399,7 +1399,7 @@ static void lapic_rearm(struct vlapic *s) s->timer_last_update = s->pt.last_plt_gtime; } -static int lapic_save_hidden_one(struct vcpu *v, hvm_domain_context_t *h) +static int lapic_save_hidden(struct vcpu *v, hvm_domain_context_t *h) { if ( !has_vlapic(v->domain) ) return 0; @@ -1407,22 +1407,7 @@ static int lapic_save_hidden_one(struct vcpu *v, hvm_domain_context_t *h) return hvm_save_entry(LAPIC, v->vcpu_id, h, &vcpu_vlapic(v)->hw); } -static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - int err = 0; - - for_each_vcpu ( d, v ) - { - err = lapic_save_hidden_one(v, h); - if ( err ) - break; - } - - return err; -} - -static int lapic_save_regs_one(struct vcpu *v, hvm_domain_context_t *h) +static int lapic_save_regs(struct vcpu *v, hvm_domain_context_t *h) { if ( !has_vlapic(v->domain) ) return 0; @@ -1433,21 +1418,6 @@ static int lapic_save_regs_one(struct vcpu *v, hvm_domain_context_t *h) return hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, vcpu_vlapic(v)->regs); } -static int lapic_save_regs(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - int err = 0; - - for_each_vcpu ( d, v ) - { - err = lapic_save_regs_one(v, h); - if ( err ) - break; - } - - return err; -} - /* * Following lapic_load_hidden()/lapic_load_regs() we may need to * correct ID and LDR when they come from an old, broken hypervisor. @@ -1546,9 +1516,9 @@ static int lapic_load_regs(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(LAPIC, lapic_save_hidden, lapic_save_hidden_one, +HVM_REGISTER_SAVE_RESTORE(LAPIC, lapic_save_hidden, lapic_load_hidden, 1, HVMSR_PER_VCPU); -HVM_REGISTER_SAVE_RESTORE(LAPIC_REGS, lapic_save_regs, lapic_save_regs_one, +HVM_REGISTER_SAVE_RESTORE(LAPIC_REGS, lapic_save_regs, lapic_load_regs, 1, HVMSR_PER_VCPU); int vlapic_init(struct vcpu *v) diff --git a/xen/arch/x86/hvm/vpic.c b/xen/arch/x86/hvm/vpic.c index 57eeae5c4d..3f3fb7a4ff 100644 --- a/xen/arch/x86/hvm/vpic.c +++ b/xen/arch/x86/hvm/vpic.c @@ -372,8 +372,9 @@ static int vpic_intercept_elcr_io( return X86EMUL_OKAY; } -static int vpic_save(struct domain *d, hvm_domain_context_t *h) +static int vpic_save(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_hw_vpic *s; int i; @@ -411,7 +412,7 @@ static int vpic_load(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(PIC, vpic_save, NULL, vpic_load, 2, HVMSR_PER_DOM); +HVM_REGISTER_SAVE_RESTORE(PIC, vpic_save, vpic_load, 2, HVMSR_PER_DOM); void vpic_reset(struct domain *d) { diff --git a/xen/include/asm-x86/hvm/save.h b/xen/include/asm-x86/hvm/save.h index 0dca46d770..4efc535055 100644 --- a/xen/include/asm-x86/hvm/save.h +++ b/xen/include/asm-x86/hvm/save.h @@ -95,10 +95,8 @@ static inline unsigned int hvm_load_instance(const struct hvm_domain_context *h) * The save handler may save multiple instances of a type into the buffer; * the load handler will be called once for each instance found when * restoring. Both return non-zero on error. */ -typedef int (*hvm_save_handler) (struct domain *d, +typedef int (*hvm_save_handler) (struct vcpu *v, hvm_domain_context_t *h); -typedef int (*hvm_save_vcpu_handler)(struct vcpu *v, - hvm_domain_context_t *h); typedef int (*hvm_load_handler) (struct domain *d, hvm_domain_context_t *h); @@ -107,7 +105,6 @@ typedef int (*hvm_load_handler) (struct domain *d, void hvm_register_savevm(uint16_t typecode, const char *name, hvm_save_handler save_state, - hvm_save_vcpu_handler save_one, hvm_load_handler load_state, size_t size, int kind); @@ -117,13 +114,12 @@ void hvm_register_savevm(uint16_t typecode, /* Syntactic sugar around that function: specify the max number of * saves, and this calculates the size of buffer needed */ -#define HVM_REGISTER_SAVE_RESTORE(_x, _save, _save_one, _load, _num, _k) \ +#define HVM_REGISTER_SAVE_RESTORE(_x, _save, _load, _num, _k) \ static int __init __hvm_register_##_x##_save_and_restore(void) \ { \ hvm_register_savevm(HVM_SAVE_CODE(_x), \ #_x, \ &_save, \ - _save_one, \ &_load, \ (_num) * (HVM_SAVE_LENGTH(_x) \ + sizeof (struct hvm_save_descriptor)), \ -- 2.17.1 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |