[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen-unstable] OOS cleanup: Fixup arrays instead of fixup tables.
# HG changeset patch # User Keir Fraser <keir.fraser@xxxxxxxxxx> # Date 1215262887 -3600 # Node ID 9a1d98a9d21b4eb795289ccb95c7242235b5e35b # Parent 30c20e467f0b1226f833eabbfa35c261f5c8d713 OOS cleanup: Fixup arrays instead of fixup tables. This patch avoids compromises with evil concepts (reverse maps) in the shadow code and limits to a maximum value the number of writable mappings of OOS pages. Code is simpler and performance do not degrade. Signed-off-by: Gianluca Guida <gianluca.guida@xxxxxxxxxxxxx> --- xen/arch/x86/mm/shadow/common.c | 262 ++++++++++++++------------------------- xen/arch/x86/mm/shadow/multi.c | 24 +-- xen/arch/x86/mm/shadow/private.h | 11 - xen/include/asm-x86/domain.h | 10 - xen/include/asm-x86/mm.h | 6 xen/include/asm-x86/perfc_defn.h | 9 - 6 files changed, 118 insertions(+), 204 deletions(-) diff -r 30c20e467f0b -r 9a1d98a9d21b xen/arch/x86/mm/shadow/common.c --- a/xen/arch/x86/mm/shadow/common.c Sat Jul 05 14:00:19 2008 +0100 +++ b/xen/arch/x86/mm/shadow/common.c Sat Jul 05 14:01:27 2008 +0100 @@ -69,12 +69,14 @@ void shadow_vcpu_init(struct vcpu *v) void shadow_vcpu_init(struct vcpu *v) { #if (SHADOW_OPTIMIZATIONS & SHOPT_OUT_OF_SYNC) - int i; + int i, j; for ( i = 0; i < SHADOW_OOS_PAGES; i++ ) { v->arch.paging.shadow.oos[i] = _mfn(INVALID_MFN); v->arch.paging.shadow.oos_snapshot[i] = _mfn(INVALID_MFN); + for ( j = 0; j < SHADOW_OOS_FIXUPS; j++ ) + v->arch.paging.shadow.oos_fixup[i].smfn[j] = _mfn(INVALID_MFN); } #endif @@ -579,132 +581,86 @@ static inline void _sh_resync_l1(struct #endif } -#define _FIXUP_IDX(_b, _i) ((_b) * SHADOW_OOS_FT_HASH + (_i)) + +/* + * Fixup arrays: We limit the maximum number of writable mappings to + * SHADOW_OOS_FIXUPS and store enough information to remove them + * quickly on resync. + */ + +static inline int oos_fixup_flush_gmfn(struct vcpu *v, mfn_t gmfn, + struct oos_fixup *fixup) +{ + int i; + for ( i = 0; i < SHADOW_OOS_FIXUPS; i++ ) + { + if ( mfn_x(fixup->smfn[i]) != INVALID_MFN ) + { + sh_remove_write_access_from_sl1p(v, gmfn, + fixup->smfn[i], + fixup->off[i]); + fixup->smfn[i] = _mfn(INVALID_MFN); + } + } + + /* Always flush the TLBs. See comment on oos_fixup_add(). */ + return 1; +} void oos_fixup_add(struct vcpu *v, mfn_t gmfn, - mfn_t smfn, unsigned long off) -{ - int idx, i, free = 0, free_slot = 0; - struct oos_fixup *fixups = v->arch.paging.shadow.oos_fixups; - - idx = mfn_x(gmfn) % SHADOW_OOS_FT_HASH; - for ( i = 0; i < SHADOW_OOS_FT_ENTRIES; i++ ) - { - if ( !mfn_valid(fixups[_FIXUP_IDX(idx, i)].gmfn) - || !mfn_is_out_of_sync(fixups[_FIXUP_IDX(idx, i)].gmfn) ) - { - free = 1; - free_slot = _FIXUP_IDX(idx, i); - } - else if ( (mfn_x(fixups[_FIXUP_IDX(idx, i)].gmfn) == mfn_x(gmfn)) - && (mfn_x(fixups[_FIXUP_IDX(idx, i)].smfn) == mfn_x(smfn)) - && (fixups[_FIXUP_IDX(idx, i)].off == off) ) - { - perfc_incr(shadow_oos_fixup_no_add); + mfn_t smfn, unsigned long off) +{ + int idx, next; + mfn_t *oos; + struct oos_fixup *oos_fixup; + struct domain *d = v->domain; + + perfc_incr(shadow_oos_fixup_add); + + for_each_vcpu(d, v) + { + oos = v->arch.paging.shadow.oos; + oos_fixup = v->arch.paging.shadow.oos_fixup; + idx = mfn_x(gmfn) % SHADOW_OOS_PAGES; + if ( mfn_x(oos[idx]) != mfn_x(gmfn) ) + idx = (idx + 1) % SHADOW_OOS_PAGES; + if ( mfn_x(oos[idx]) == mfn_x(gmfn) ) + { + next = oos_fixup[idx].next; + + if ( mfn_x(oos_fixup[idx].smfn[next]) != INVALID_MFN ) + { + /* Reuse this slot and remove current writable mapping. */ + sh_remove_write_access_from_sl1p(v, gmfn, + oos_fixup[idx].smfn[next], + oos_fixup[idx].off[next]); + perfc_incr(shadow_oos_fixup_evict); + /* We should flush the TLBs now, because we removed a + writable mapping, but since the shadow is already + OOS we have no problem if another vcpu write to + this page table. We just have to be very careful to + *always* flush the tlbs on resync. */ + } + + oos_fixup[idx].smfn[next] = smfn; + oos_fixup[idx].off[next] = off; + oos_fixup[idx].next = (next + 1) % SHADOW_OOS_FIXUPS; return; } } - if ( free ) - { - if ( !v->arch.paging.shadow.oos_fixup_used ) - v->arch.paging.shadow.oos_fixup_used = 1; - fixups[free_slot].gmfn = gmfn; - fixups[free_slot].smfn = smfn; - fixups[free_slot].off = off; - perfc_incr(shadow_oos_fixup_add_ok); - return; - } - - - perfc_incr(shadow_oos_fixup_add_fail); -} - -void oos_fixup_remove(struct vcpu *v, mfn_t gmfn) -{ - int idx, i; - struct domain *d = v->domain; - - perfc_incr(shadow_oos_fixup_remove); - - /* If the domain is dying we might get called when deallocating - * the shadows. Fixup tables are already freed so exit now. */ - if ( d->is_dying ) - return; - - idx = mfn_x(gmfn) % SHADOW_OOS_FT_HASH; - for_each_vcpu(d, v) - { - struct oos_fixup *fixups = v->arch.paging.shadow.oos_fixups; - for ( i = 0; i < SHADOW_OOS_FT_ENTRIES; i++ ) - if ( mfn_x(fixups[_FIXUP_IDX(idx, i)].gmfn) == mfn_x(gmfn) ) - fixups[_FIXUP_IDX(idx, i)].gmfn = _mfn(INVALID_MFN); - } -} - -int oos_fixup_flush(struct vcpu *v) -{ - int i, rc = 0; - struct oos_fixup *fixups = v->arch.paging.shadow.oos_fixups; - - perfc_incr(shadow_oos_fixup_flush); - - if ( !v->arch.paging.shadow.oos_fixup_used ) - return 0; - - for ( i = 0; i < SHADOW_OOS_FT_HASH * SHADOW_OOS_FT_ENTRIES; i++ ) - { - if ( mfn_valid(fixups[i].gmfn) ) - { - if ( mfn_is_out_of_sync(fixups[i].gmfn) ) - rc |= sh_remove_write_access_from_sl1p(v, fixups[i].gmfn, - fixups[i].smfn, - fixups[i].off); - fixups[i].gmfn = _mfn(INVALID_MFN); - } - } - - v->arch.paging.shadow.oos_fixup_used = 0; - - return rc; -} - -int oos_fixup_flush_gmfn(struct vcpu *v, mfn_t gmfn) -{ - int idx, i, rc = 0; - struct domain *d = v->domain; - - perfc_incr(shadow_oos_fixup_flush_gmfn); - - idx = mfn_x(gmfn) % SHADOW_OOS_FT_HASH; - for_each_vcpu(d, v) - { - struct oos_fixup *fixups = v->arch.paging.shadow.oos_fixups; - - for ( i = 0; i < SHADOW_OOS_FT_ENTRIES; i++ ) - { - if ( mfn_x(fixups[_FIXUP_IDX(idx, i)].gmfn) != mfn_x(gmfn) ) - continue; - - rc |= sh_remove_write_access_from_sl1p(v, - fixups[_FIXUP_IDX(idx,i)].gmfn, - fixups[_FIXUP_IDX(idx,i)].smfn, - fixups[_FIXUP_IDX(idx,i)].off); - - fixups[_FIXUP_IDX(idx,i)].gmfn = _mfn(INVALID_MFN); - } - } - - return rc; -} - -static int oos_remove_write_access(struct vcpu *v, mfn_t gmfn, unsigned long va) + SHADOW_ERROR("gmfn %lx was OOS but not in hash table\n", mfn_x(gmfn)); + BUG(); +} + +static int oos_remove_write_access(struct vcpu *v, mfn_t gmfn, + struct oos_fixup *fixup) { int ftlb = 0; - ftlb |= oos_fixup_flush_gmfn(v, gmfn); - - switch ( sh_remove_write_access(v, gmfn, 0, va) ) + ftlb |= oos_fixup_flush_gmfn(v, gmfn, fixup); + + switch ( sh_remove_write_access(v, gmfn, 0, 0) ) { default: case 0: @@ -732,7 +688,8 @@ static int oos_remove_write_access(struc /* Pull all the entries on an out-of-sync page back into sync. */ -static void _sh_resync(struct vcpu *v, mfn_t gmfn, unsigned long va, mfn_t snp) +static void _sh_resync(struct vcpu *v, mfn_t gmfn, + struct oos_fixup *fixup, mfn_t snp) { struct page_info *pg = mfn_to_page(gmfn); @@ -747,7 +704,7 @@ static void _sh_resync(struct vcpu *v, m v->domain->domain_id, v->vcpu_id, mfn_x(gmfn), va); /* Need to pull write access so the page *stays* in sync. */ - if ( oos_remove_write_access(v, gmfn, va) ) + if ( oos_remove_write_access(v, gmfn, fixup) ) { /* Page has been unshadowed. */ return; @@ -766,13 +723,17 @@ static void _sh_resync(struct vcpu *v, m /* Add an MFN to the list of out-of-sync guest pagetables */ -static void oos_hash_add(struct vcpu *v, mfn_t gmfn, unsigned long va) -{ - int idx, oidx, swap = 0; +static void oos_hash_add(struct vcpu *v, mfn_t gmfn) +{ + int i, idx, oidx, swap = 0; void *gptr, *gsnpptr; mfn_t *oos = v->arch.paging.shadow.oos; - unsigned long *oos_va = v->arch.paging.shadow.oos_va; mfn_t *oos_snapshot = v->arch.paging.shadow.oos_snapshot; + struct oos_fixup *oos_fixup = v->arch.paging.shadow.oos_fixup; + struct oos_fixup fixup = { .next = 0 }; + + for (i = 0; i < SHADOW_OOS_FIXUPS; i++ ) + fixup.smfn[i] = _mfn(INVALID_MFN); idx = mfn_x(gmfn) % SHADOW_OOS_PAGES; oidx = idx; @@ -782,18 +743,18 @@ static void oos_hash_add(struct vcpu *v, { /* Punt the current occupant into the next slot */ SWAP(oos[idx], gmfn); - SWAP(oos_va[idx], va); + SWAP(oos_fixup[idx], fixup); swap = 1; idx = (idx + 1) % SHADOW_OOS_PAGES; } if ( mfn_valid(oos[idx]) ) { /* Crush the current occupant. */ - _sh_resync(v, oos[idx], oos_va[idx], oos_snapshot[idx]); + _sh_resync(v, oos[idx], &oos_fixup[idx], oos_snapshot[idx]); perfc_incr(shadow_unsync_evict); } oos[idx] = gmfn; - oos_va[idx] = va; + oos_fixup[idx] = fixup; if ( swap ) SWAP(oos_snapshot[idx], oos_snapshot[oidx]); @@ -862,14 +823,14 @@ void sh_resync(struct vcpu *v, mfn_t gmf { int idx; mfn_t *oos; - unsigned long *oos_va; mfn_t *oos_snapshot; + struct oos_fixup *oos_fixup; struct domain *d = v->domain; for_each_vcpu(d, v) { oos = v->arch.paging.shadow.oos; - oos_va = v->arch.paging.shadow.oos_va; + oos_fixup = v->arch.paging.shadow.oos_fixup; oos_snapshot = v->arch.paging.shadow.oos_snapshot; idx = mfn_x(gmfn) % SHADOW_OOS_PAGES; if ( mfn_x(oos[idx]) != mfn_x(gmfn) ) @@ -877,7 +838,7 @@ void sh_resync(struct vcpu *v, mfn_t gmf if ( mfn_x(oos[idx]) == mfn_x(gmfn) ) { - _sh_resync(v, gmfn, oos_va[idx], oos_snapshot[idx]); + _sh_resync(v, gmfn, &oos_fixup[idx], oos_snapshot[idx]); oos[idx] = _mfn(INVALID_MFN); return; } @@ -917,8 +878,8 @@ void sh_resync_all(struct vcpu *v, int s int idx; struct vcpu *other; mfn_t *oos = v->arch.paging.shadow.oos; - unsigned long *oos_va = v->arch.paging.shadow.oos_va; mfn_t *oos_snapshot = v->arch.paging.shadow.oos_snapshot; + struct oos_fixup *oos_fixup = v->arch.paging.shadow.oos_fixup; SHADOW_PRINTK("d=%d, v=%d\n", v->domain->domain_id, v->vcpu_id); @@ -929,16 +890,13 @@ void sh_resync_all(struct vcpu *v, int s if ( do_locking ) shadow_lock(v->domain); - - if ( oos_fixup_flush(v) ) - flush_tlb_mask(v->domain->domain_dirty_cpumask); /* First: resync all of this vcpu's oos pages */ for ( idx = 0; idx < SHADOW_OOS_PAGES; idx++ ) if ( mfn_valid(oos[idx]) ) { /* Write-protect and sync contents */ - _sh_resync(v, oos[idx], oos_va[idx], oos_snapshot[idx]); + _sh_resync(v, oos[idx], &oos_fixup[idx], oos_snapshot[idx]); oos[idx] = _mfn(INVALID_MFN); } @@ -959,8 +917,9 @@ void sh_resync_all(struct vcpu *v, int s shadow_lock(v->domain); oos = other->arch.paging.shadow.oos; - oos_va = other->arch.paging.shadow.oos_va; + oos_fixup = other->arch.paging.shadow.oos_fixup; oos_snapshot = other->arch.paging.shadow.oos_snapshot; + for ( idx = 0; idx < SHADOW_OOS_PAGES; idx++ ) { if ( !mfn_valid(oos[idx]) ) @@ -976,7 +935,7 @@ void sh_resync_all(struct vcpu *v, int s else { /* Write-protect and sync contents */ - _sh_resync(other, oos[idx], oos_va[idx], oos_snapshot[idx]); + _sh_resync(other, oos[idx], &oos_fixup[idx], oos_snapshot[idx]); oos[idx] = _mfn(INVALID_MFN); } } @@ -987,7 +946,7 @@ void sh_resync_all(struct vcpu *v, int s } /* Allow a shadowed page to go out of sync */ -int sh_unsync(struct vcpu *v, mfn_t gmfn, unsigned long va) +int sh_unsync(struct vcpu *v, mfn_t gmfn) { struct page_info *pg; @@ -1009,7 +968,7 @@ int sh_unsync(struct vcpu *v, mfn_t gmfn return 0; pg->shadow_flags |= SHF_out_of_sync|SHF_oos_may_write; - oos_hash_add(v, gmfn, va); + oos_hash_add(v, gmfn); perfc_incr(shadow_unsync); return 1; } @@ -1064,7 +1023,6 @@ void shadow_demote(struct vcpu *v, mfn_t if ( page_is_out_of_sync(page) ) { oos_hash_remove(v, gmfn); - oos_fixup_remove(v, gmfn); } #endif clear_bit(_PGC_page_table, &page->count_info); @@ -2804,23 +2762,6 @@ static void sh_update_paging_modes(struc #endif /* (SHADOW_OPTIMIZATIONS & SHOPT_VIRTUAL_TLB) */ #if (SHADOW_OPTIMIZATIONS & SHOPT_OUT_OF_SYNC) - if ( v->arch.paging.shadow.oos_fixups == NULL ) - { - int i; - v->arch.paging.shadow.oos_fixups = - alloc_xenheap_pages(SHADOW_OOS_FT_ORDER); - if ( v->arch.paging.shadow.oos_fixups == NULL ) - { - SHADOW_ERROR("Could not allocate OOS fixup table" - " for dom %u vcpu %u\n", - v->domain->domain_id, v->vcpu_id); - domain_crash(v->domain); - return; - } - for ( i = 0; i < SHADOW_OOS_FT_HASH * SHADOW_OOS_FT_ENTRIES; i++ ) - v->arch.paging.shadow.oos_fixups[i].gmfn = _mfn(INVALID_MFN); - } - if ( mfn_x(v->arch.paging.shadow.oos_snapshot[0]) == INVALID_MFN ) { int i; @@ -3173,13 +3114,6 @@ void shadow_teardown(struct domain *d) #endif /* (SHADOW_OPTIMIZATIONS & SHOPT_VIRTUAL_TLB) */ #if (SHADOW_OPTIMIZATIONS & SHOPT_OUT_OF_SYNC) - if ( v->arch.paging.shadow.oos_fixups ) - { - free_xenheap_pages(v->arch.paging.shadow.oos_fixups, - SHADOW_OOS_FT_ORDER); - v->arch.paging.shadow.oos_fixups = NULL; - } - { int i; mfn_t *oos_snapshot = v->arch.paging.shadow.oos_snapshot; diff -r 30c20e467f0b -r 9a1d98a9d21b xen/arch/x86/mm/shadow/multi.c --- a/xen/arch/x86/mm/shadow/multi.c Sat Jul 05 14:00:19 2008 +0100 +++ b/xen/arch/x86/mm/shadow/multi.c Sat Jul 05 14:01:27 2008 +0100 @@ -916,7 +916,10 @@ _sh_propagate(struct vcpu *v, if ( unlikely((level == 1) && sh_mfn_is_a_page_table(target_mfn) #if (SHADOW_OPTIMIZATIONS & SHOPT_OUT_OF_SYNC ) - && !mfn_oos_may_write(target_mfn) + /* Unless the page is out of sync and the guest is + writing to it. */ + && !(mfn_oos_may_write(target_mfn) + && (ft == ft_demand_write)) #endif /* OOS */ ) ) { @@ -3291,7 +3294,7 @@ static int sh_page_fault(struct vcpu *v, /* Always unsync when writing to L1 page tables. */ if ( sh_mfn_is_a_page_table(gmfn) && ft == ft_demand_write ) - sh_unsync(v, gmfn, va); + sh_unsync(v, gmfn); #endif /* OOS */ /* Calculate the shadow entry and write it */ @@ -3322,8 +3325,10 @@ static int sh_page_fault(struct vcpu *v, /* Need to emulate accesses to page tables */ if ( sh_mfn_is_a_page_table(gmfn) #if (SHADOW_OPTIMIZATIONS & SHOPT_OUT_OF_SYNC) - /* Unless they've been allowed to go out of sync with their shadows */ - && !mfn_is_out_of_sync(gmfn) + /* Unless they've been allowed to go out of sync with their + shadows and we don't need to unshadow it. */ + && !(mfn_is_out_of_sync(gmfn) + && !(regs->error_code & PFEC_user_mode)) #endif ) { @@ -4350,15 +4355,8 @@ int sh_rm_write_access_from_sl1p(struct sp = mfn_to_shadow_page(smfn); - if ( sp->mbz != 0 || -#if GUEST_PAGING_LEVELS == 4 - (sp->type != SH_type_l1_64_shadow) -#elif GUEST_PAGING_LEVELS == 3 - (sp->type != SH_type_l1_pae_shadow) -#elif GUEST_PAGING_LEVELS == 2 - (sp->type != SH_type_l1_32_shadow) -#endif - ) + if ( sp->mbz != 0 + || (sp->type != SH_type_l1_shadow) ) goto fail; sl1p = sh_map_domain_page(smfn); diff -r 30c20e467f0b -r 9a1d98a9d21b xen/arch/x86/mm/shadow/private.h --- a/xen/arch/x86/mm/shadow/private.h Sat Jul 05 14:00:19 2008 +0100 +++ b/xen/arch/x86/mm/shadow/private.h Sat Jul 05 14:01:27 2008 +0100 @@ -323,15 +323,6 @@ static inline int sh_type_is_pinnable(st #define SHF_out_of_sync (1u<<30) #define SHF_oos_may_write (1u<<29) -/* Fixup tables are a non-complete writable-mappings reverse map for - OOS pages. This let us quickly resync pages (avoiding brute-force - search of the shadows) when the va hint is not sufficient (i.e., - the pagetable is mapped in multiple places and in multiple - shadows.) */ -#define SHADOW_OOS_FT_ENTRIES \ - ((PAGE_SIZE << SHADOW_OOS_FT_ORDER) \ - / (SHADOW_OOS_FT_HASH * sizeof(struct oos_fixup))) - #endif /* (SHADOW_OPTIMIZATIONS & SHOPT_OUT_OF_SYNC) */ static inline int sh_page_has_multiple_shadows(struct page_info *pg) @@ -421,7 +412,7 @@ int shadow_cmpxchg_guest_entry(struct vc #if (SHADOW_OPTIMIZATIONS & SHOPT_OUT_OF_SYNC) /* Allow a shadowed page to go out of sync */ -int sh_unsync(struct vcpu *v, mfn_t gmfn, unsigned long va); +int sh_unsync(struct vcpu *v, mfn_t gmfn); /* Pull an out-of-sync page back into sync. */ void sh_resync(struct vcpu *v, mfn_t gmfn); diff -r 30c20e467f0b -r 9a1d98a9d21b xen/include/asm-x86/domain.h --- a/xen/include/asm-x86/domain.h Sat Jul 05 14:00:19 2008 +0100 +++ b/xen/include/asm-x86/domain.h Sat Jul 05 14:01:27 2008 +0100 @@ -128,14 +128,12 @@ struct shadow_vcpu { /* Shadow out-of-sync: pages that this vcpu has let go out of sync */ mfn_t oos[SHADOW_OOS_PAGES]; - unsigned long oos_va[SHADOW_OOS_PAGES]; mfn_t oos_snapshot[SHADOW_OOS_PAGES]; struct oos_fixup { - mfn_t gmfn; - mfn_t smfn; - unsigned long off; - } *oos_fixups; - int oos_fixup_used; + int next; + mfn_t smfn[SHADOW_OOS_FIXUPS]; + unsigned long off[SHADOW_OOS_FIXUPS]; + } oos_fixup[SHADOW_OOS_PAGES]; }; /************************************************/ diff -r 30c20e467f0b -r 9a1d98a9d21b xen/include/asm-x86/mm.h --- a/xen/include/asm-x86/mm.h Sat Jul 05 14:00:19 2008 +0100 +++ b/xen/include/asm-x86/mm.h Sat Jul 05 14:01:27 2008 +0100 @@ -133,10 +133,8 @@ static inline u32 pickle_domptr(struct d /* The number of out-of-sync shadows we allow per vcpu (prime, please) */ #define SHADOW_OOS_PAGES 3 -/* The order OOS fixup tables per vcpu */ -#define SHADOW_OOS_FT_ORDER 1 -/* OOS fixup tables hash entries */ -#define SHADOW_OOS_FT_HASH 13 +/* OOS fixup entries */ +#define SHADOW_OOS_FIXUPS 2 #define page_get_owner(_p) (unpickle_domptr((_p)->u.inuse._domain)) #define page_set_owner(_p,_d) ((_p)->u.inuse._domain = pickle_domptr(_d)) diff -r 30c20e467f0b -r 9a1d98a9d21b xen/include/asm-x86/perfc_defn.h --- a/xen/include/asm-x86/perfc_defn.h Sat Jul 05 14:00:19 2008 +0100 +++ b/xen/include/asm-x86/perfc_defn.h Sat Jul 05 14:01:27 2008 +0100 @@ -105,13 +105,8 @@ PERFCOUNTER(shadow_em_ex_non_pt, "shad PERFCOUNTER(shadow_em_ex_non_pt, "shadow extra non-pt-write op") PERFCOUNTER(shadow_em_ex_fail, "shadow extra emulation failed") -PERFCOUNTER(shadow_oos_fixup_add_ok, "shadow OOS fixups adds") -PERFCOUNTER(shadow_oos_fixup_no_add, "shadow OOS fixups no adds") -PERFCOUNTER(shadow_oos_fixup_add_fail, "shadow OOS fixups adds failed") -PERFCOUNTER(shadow_oos_fixup_remove, "shadow OOS fixups removes") -PERFCOUNTER(shadow_oos_fixup_flush, "shadow OOS fixups flushes") -PERFCOUNTER(shadow_oos_fixup_flush_gmfn,"shadow OOS fixups gmfn flushes") - +PERFCOUNTER(shadow_oos_fixup_add, "shadow OOS fixup adds") +PERFCOUNTER(shadow_oos_fixup_evict,"shadow OOS fixup evictions") PERFCOUNTER(shadow_unsync, "shadow OOS unsyncs") PERFCOUNTER(shadow_unsync_evict, "shadow OOS evictions") PERFCOUNTER(shadow_resync, "shadow OOS resyncs") _______________________________________________ Xen-changelog mailing list Xen-changelog@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-changelog
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |