[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen-unstable] EPT: Clean up some code
# HG changeset patch # User Keir Fraser <keir.fraser@xxxxxxxxxx> # Date 1253002476 -3600 # Node ID dbe555567793bf2f135ce2ba8e5afa84d19fd95f # Parent 582970a2d2dcd1d041ab29237ea8f8a2e31b8989 EPT: Clean up some code Clean up and reorganize some code in preparation for adding populate-on-demand functionality. Should be no functional difference. Signed-off-by: George Dunlap <george.dunlap@xxxxxxxxxxxxx> --- xen/arch/x86/mm/hap/p2m-ept.c | 123 +++++++++++++++++++------------------- xen/include/asm-x86/hvm/vmx/vmx.h | 6 - 2 files changed, 67 insertions(+), 62 deletions(-) diff -r 582970a2d2dc -r dbe555567793 xen/arch/x86/mm/hap/p2m-ept.c --- a/xen/arch/x86/mm/hap/p2m-ept.c Tue Sep 15 09:13:38 2009 +0100 +++ b/xen/arch/x86/mm/hap/p2m-ept.c Tue Sep 15 09:14:36 2009 +0100 @@ -58,10 +58,11 @@ static void ept_p2m_type_to_flags(ept_en } } +#define GUEST_TABLE_MAP_FAILED 0 #define GUEST_TABLE_NORMAL_PAGE 1 #define GUEST_TABLE_SUPER_PAGE 2 -#define GUEST_TABLE_SPLIT_PAGE 3 - + +/* Fill in middle levels of ept table */ static int ept_set_middle_entry(struct domain *d, ept_entry_t *ept_entry) { struct page_info *pg; @@ -87,9 +88,21 @@ static int ept_set_middle_entry(struct d return 1; } +/* Take the currently mapped table, find the corresponding gfn entry, + * and map the next table, if available. If the entry is empty + * and read_only is set, + * Return values: + * 0: Failed to map. Either read_only was set and the entry was + * empty, or allocating a new page failed. + * GUEST_TABLE_NORMAL_PAGE: next level mapped normally + * GUEST_TABLE_SUPER_PAGE: + * The next entry points to a superpage, and caller indicates + * that they are going to the superpage level, or are only doing + * a read. + */ static int ept_next_level(struct domain *d, bool_t read_only, ept_entry_t **table, unsigned long *gfn_remainder, - u32 shift, int order) + u32 shift) { ept_entry_t *ept_entry; ept_entry_t *next; @@ -102,26 +115,22 @@ static int ept_next_level(struct domain if ( !(ept_entry->epte & 0x7) ) { if ( read_only ) - return 0; + return GUEST_TABLE_MAP_FAILED; if ( !ept_set_middle_entry(d, ept_entry) ) - return 0; - } - - if ( !ept_entry->sp_avail ) + return GUEST_TABLE_MAP_FAILED; + } + + /* The only time sp_avail would be set here is if we had hit a superpage */ + if ( ept_entry->sp_avail ) + return GUEST_TABLE_SUPER_PAGE; + else { *gfn_remainder &= (1UL << shift) - 1; next = map_domain_page(ept_entry->mfn); unmap_domain_page(*table); *table = next; return GUEST_TABLE_NORMAL_PAGE; - } - else - { - if ( order == shift || read_only ) - return GUEST_TABLE_SUPER_PAGE; - else - return GUEST_TABLE_SPLIT_PAGE; } } @@ -159,28 +168,32 @@ ept_set_entry(struct domain *d, unsigned for ( i = EPT_DEFAULT_GAW; i > walk_level; i-- ) { - ret = ept_next_level(d, 0, &table, &gfn_remainder, - i * EPT_TABLE_ORDER, order); + ret = ept_next_level(d, 0, &table, &gfn_remainder, i * EPT_TABLE_ORDER); if ( !ret ) goto out; else if ( ret != GUEST_TABLE_NORMAL_PAGE ) break; } + /* If order == 9, we should never get SUPERPAGE or PoD. + * If i > walk_level, we need to split the page; otherwise, + * just behave as normal. */ + ASSERT(order == 0 || ret == GUEST_TABLE_NORMAL_PAGE); + index = gfn_remainder >> ( i ? (i * EPT_TABLE_ORDER): order); - walk_level = ( i ? ( i * EPT_TABLE_ORDER) : order) / EPT_TABLE_ORDER; offset = (gfn_remainder & ( ((1 << (i*EPT_TABLE_ORDER)) - 1))); ept_entry = table + index; - if ( ret != GUEST_TABLE_SPLIT_PAGE ) - { + if ( i == walk_level ) + { + /* We reached the level we're looking for */ if ( mfn_valid(mfn_x(mfn)) || direct_mmio ) { ept_entry->emt = epte_get_entry_emt(d, gfn, mfn, &igmt, direct_mmio); ept_entry->igmt = igmt; - ept_entry->sp_avail = walk_level ? 1 : 0; + ept_entry->sp_avail = order ? 1 : 0; if ( ret == GUEST_TABLE_SUPER_PAGE ) { @@ -189,8 +202,8 @@ ept_set_entry(struct domain *d, unsigned else ept_entry->mfn = mfn_x(mfn) - offset; - if ( (ept_entry->avail1 == p2m_ram_logdirty) && - (p2mt == p2m_ram_rw) ) + if ( (ept_entry->avail1 == p2m_ram_logdirty) + && (p2mt == p2m_ram_rw) ) for ( i = 0; i < 512; i++ ) paging_mark_dirty(d, mfn_x(mfn) - offset + i); } @@ -217,37 +230,40 @@ ept_set_entry(struct domain *d, unsigned * It's super page before, now set one of the 4k pages, so * we should split the 2m page to 4k pages now. */ + /* Pointers to / into new (split) middle-level table */ ept_entry_t *split_table = NULL; ept_entry_t *split_ept_entry = NULL; - unsigned long split_mfn = ept_entry->mfn; - p2m_type_t split_p2mt = ept_entry->avail1; - ept_entry_t new_ept_entry; + /* Info about old (superpage) table */ + unsigned long super_mfn = ept_entry->mfn; + p2m_type_t super_p2mt = ept_entry->avail1; + /* The new l2 entry which we'll write after we've build the new l1 table */ + ept_entry_t l2_ept_entry; /* * Allocate new page for new ept middle level entry which is * before a leaf super entry */ - if ( !ept_set_middle_entry(d, &new_ept_entry) ) + if ( !ept_set_middle_entry(d, &l2_ept_entry) ) goto out; /* Split the super page before to 4k pages */ - split_table = map_domain_page(new_ept_entry.mfn); + split_table = map_domain_page(l2_ept_entry.mfn); offset = gfn & ((1 << EPT_TABLE_ORDER) - 1); for ( i = 0; i < 512; i++ ) { split_ept_entry = split_table + i; split_ept_entry->emt = epte_get_entry_emt(d, gfn - offset + i, - _mfn(split_mfn + i), + _mfn(super_mfn + i), &igmt, direct_mmio); split_ept_entry->igmt = igmt; split_ept_entry->sp_avail = 0; - split_ept_entry->mfn = split_mfn + i; - split_ept_entry->avail1 = split_p2mt; + split_ept_entry->mfn = super_mfn + i; + split_ept_entry->avail1 = super_p2mt; split_ept_entry->rsvd = 0; split_ept_entry->avail2 = 0; - ept_p2m_type_to_flags(split_ept_entry, split_p2mt); + ept_p2m_type_to_flags(split_ept_entry, super_p2mt); } /* Set the destinated 4k page as normal */ @@ -265,7 +281,7 @@ ept_set_entry(struct domain *d, unsigned ept_p2m_type_to_flags(split_ept_entry, p2mt); unmap_domain_page(split_table); - *ept_entry = new_ept_entry; + *ept_entry = l2_ept_entry; } /* Track the highest gfn for which we have ever had a valid mapping */ @@ -333,7 +349,7 @@ static mfn_t ept_get_entry(struct domain for ( i = EPT_DEFAULT_GAW; i > 0; i-- ) { ret = ept_next_level(d, 1, &table, &gfn_remainder, - i * EPT_TABLE_ORDER, 0); + i * EPT_TABLE_ORDER); if ( !ret ) goto out; else if ( ret == GUEST_TABLE_SUPER_PAGE ) @@ -365,13 +381,13 @@ static mfn_t ept_get_entry(struct domain return mfn; } -static uint64_t ept_get_entry_content(struct domain *d, unsigned long gfn) +static ept_entry_t ept_get_entry_content(struct domain *d, unsigned long gfn) { ept_entry_t *table = map_domain_page(mfn_x(pagetable_get_mfn(d->arch.phys_table))); unsigned long gfn_remainder = gfn; ept_entry_t *ept_entry; - uint64_t content = 0; + ept_entry_t content = { .epte = 0 }; u32 index; int i; int ret=0; @@ -383,7 +399,7 @@ static uint64_t ept_get_entry_content(st for ( i = EPT_DEFAULT_GAW; i > 0; i-- ) { ret = ept_next_level(d, 1, &table, &gfn_remainder, - i * EPT_TABLE_ORDER, 0); + i * EPT_TABLE_ORDER); if ( !ret ) goto out; else if ( ret == GUEST_TABLE_SUPER_PAGE ) @@ -392,7 +408,7 @@ static uint64_t ept_get_entry_content(st index = gfn_remainder >> (i * EPT_TABLE_ORDER); ept_entry = table + index; - content = ept_entry->epte; + content = *ept_entry; out: unmap_domain_page(table); @@ -429,29 +445,18 @@ void ept_change_entry_emt_with_range(str unsigned long end_gfn) { unsigned long gfn; - uint64_t epte; - p2m_type_t p2mt; + ept_entry_t e; int order = 0; - mfn_t mfn; - uint8_t o_igmt; - uint8_t o_emt; for ( gfn = start_gfn; gfn <= end_gfn; gfn++ ) { - epte = ept_get_entry_content(d, gfn); - if ( epte == 0 ) + e = ept_get_entry_content(d, gfn); + if ( e.epte == 0 || !mfn_valid(mfn_x(e.mfn)) ) continue; - mfn = _mfn((epte & EPTE_MFN_MASK) >> PAGE_SHIFT); - if ( !mfn_valid(mfn_x(mfn)) ) - continue; - - p2mt = (epte & EPTE_AVAIL1_MASK) >> EPTE_AVAIL1_SHIFT; - o_igmt = (epte & EPTE_IGMT_MASK) >> EPTE_IGMT_SHIFT; - o_emt = (epte & EPTE_EMT_MASK) >> EPTE_EMT_SHIFT; order = 0; - if ( epte & EPTE_SUPER_PAGE_MASK ) + if ( e.sp_avail ) { if ( !(gfn & ((1 << EPT_TABLE_ORDER) - 1)) && ((gfn + 0x1FF) <= end_gfn) ) @@ -461,22 +466,22 @@ void ept_change_entry_emt_with_range(str * Set emt for super page. */ order = EPT_TABLE_ORDER; - if ( need_modify_ept_entry(d, gfn, mfn, o_igmt, o_emt, p2mt) ) - ept_set_entry(d, gfn, mfn, order, p2mt); + if ( need_modify_ept_entry(d, gfn, e.mfn, e.igmt, e.emt, e.avail1) ) + ept_set_entry(d, gfn, e.mfn, order, e.avail1); gfn += 0x1FF; } else { /* Change emt for partial entries of the 2m area. */ - if ( need_modify_ept_entry(d, gfn, mfn, o_igmt, o_emt, p2mt) ) - ept_set_entry(d, gfn, mfn, order, p2mt); + if ( need_modify_ept_entry(d, gfn, e.mfn, e.igmt, e.emt, e.avail1) ) + ept_set_entry(d, gfn, e.mfn, order, e.avail1); gfn = ((gfn >> EPT_TABLE_ORDER) << EPT_TABLE_ORDER) + 0x1FF; } } else /* gfn assigned with 4k */ { - if ( need_modify_ept_entry(d, gfn, mfn, o_igmt, o_emt, p2mt) ) - ept_set_entry(d, gfn, mfn, order, p2mt); + if ( need_modify_ept_entry(d, gfn, e.mfn, e.igmt, e.emt, e.avail1) ) + ept_set_entry(d, gfn, e.mfn, order, e.avail1); } } } diff -r 582970a2d2dc -r dbe555567793 xen/include/asm-x86/hvm/vmx/vmx.h --- a/xen/include/asm-x86/hvm/vmx/vmx.h Tue Sep 15 09:13:38 2009 +0100 +++ b/xen/include/asm-x86/hvm/vmx/vmx.h Tue Sep 15 09:14:36 2009 +0100 @@ -33,9 +33,9 @@ typedef union { u64 r : 1, w : 1, x : 1, - emt : 3, - igmt : 1, - sp_avail : 1, + emt : 3, /* EPT Memory type */ + igmt : 1, /* Ignore MTRR memory type */ + sp_avail : 1, /* Is this a superpage? */ avail1 : 4, mfn : 45, rsvd : 5, _______________________________________________ Xen-changelog mailing list Xen-changelog@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-changelog
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |