[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v3 3/3] xen: rework paging_log_dirty_op to work with hvm guests
When the caller of paging_log_dirty_op is a hvm guest Xen would choke when trying to copy the dirty bitmap to the guest because the paging lock is already held. Fix this by independently mapping each page of the guest bitmap as needed without the paging lock held. Signed-off-by: Roger Pau Monnà <roger.pau@xxxxxxxxxx> Cc: Tim Deegan <tim@xxxxxxx> Cc: Jan Beulich <jbeulich@xxxxxxxx> Cc: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> --- Changes since v2: - Add checks for p2m_is_ram and p2m_is_discard_write when mapping a guest page. - Remove error checking from memset/memcpy, they unconditionally return dst. --- xen/arch/x86/mm/paging.c | 103 +++++++++++++++++++++++++++++++++++++++++------ 1 file changed, 91 insertions(+), 12 deletions(-) diff --git a/xen/arch/x86/mm/paging.c b/xen/arch/x86/mm/paging.c index b54d76a..cb5feb8 100644 --- a/xen/arch/x86/mm/paging.c +++ b/xen/arch/x86/mm/paging.c @@ -397,6 +397,53 @@ int paging_mfn_is_dirty(struct domain *d, mfn_t gmfn) return rv; } +static inline void *map_dirty_bitmap(XEN_GUEST_HANDLE_64(uint8) dirty_bitmap, + unsigned long pages, + struct page_info **page, + unsigned long *mapped_page) +{ + p2m_type_t p2mt; + uint32_t pfec; + unsigned long gfn; + + gfn = paging_gva_to_gfn(current, + (paddr_t)(((char *)dirty_bitmap.p) + (pages >> 3)), + &pfec); + if ( gfn == INVALID_GFN ) + return NULL; + + *page = get_page_from_gfn(current->domain, gfn, &p2mt, P2M_UNSHARE); + + if ( !p2m_is_ram(p2mt) ) + { + put_page(*page); + return NULL; + } + if ( p2m_is_paging(p2mt) ) + { + put_page(*page); + p2m_mem_paging_populate(current->domain, gfn); + return NULL; + } + if ( p2m_is_shared(p2mt) || p2m_is_discard_write(p2mt) ) + { + put_page(*page); + return NULL; + } + + *mapped_page = pages; + return __map_domain_page(*page); +} + +static inline void unmap_dirty_bitmap(void *addr, struct page_info *page) +{ + if ( addr != NULL ) + { + unmap_domain_page(addr); + put_page(page); + } +} + /* Read a domain's log-dirty bitmap and stats. If the operation is a CLEAN, * clear the bitmap and stats as well. */ @@ -409,9 +456,23 @@ static int paging_log_dirty_op(struct domain *d, mfn_t *l4 = NULL, *l3 = NULL, *l2 = NULL; unsigned long *l1 = NULL; int i4, i3, i2; + uint8_t *dirty_bitmap = NULL; + struct page_info *page; + unsigned long index_mapped = 0; if ( !resuming ) domain_pause(d); + + dirty_bitmap = map_dirty_bitmap(sc->dirty_bitmap, + resuming ? + d->arch.paging.preempt.log_dirty.done : + 0, + &page, &index_mapped); + if ( dirty_bitmap == NULL ) + { + domain_unpause(d); + return -EFAULT; + } paging_lock(d); if ( !d->arch.paging.preempt.dom ) @@ -448,21 +509,23 @@ static int paging_log_dirty_op(struct domain *d, goto out; } - l4 = paging_map_log_dirty_bitmap(d); i4 = d->arch.paging.preempt.log_dirty.i4; i3 = d->arch.paging.preempt.log_dirty.i3; + i2 = 0; pages = d->arch.paging.preempt.log_dirty.done; + again: + l4 = paging_map_log_dirty_bitmap(d); + for ( ; (pages < sc->pages) && (i4 < LOGDIRTY_NODE_ENTRIES); i4++, i3 = 0 ) { l3 = (l4 && mfn_valid(l4[i4])) ? map_domain_page(mfn_x(l4[i4])) : NULL; - for ( ; (pages < sc->pages) && (i3 < LOGDIRTY_NODE_ENTRIES); i3++ ) + for ( ; (pages < sc->pages) && (i3 < LOGDIRTY_NODE_ENTRIES); + i3++, i2 = 0 ) { l2 = ((l3 && mfn_valid(l3[i3])) ? map_domain_page(mfn_x(l3[i3])) : NULL); - for ( i2 = 0; - (pages < sc->pages) && (i2 < LOGDIRTY_NODE_ENTRIES); - i2++ ) + for ( ; (pages < sc->pages) && (i2 < LOGDIRTY_NODE_ENTRIES); i2++ ) { unsigned int bytes = PAGE_SIZE; l1 = ((l2 && mfn_valid(l2[i2])) ? @@ -471,15 +534,29 @@ static int paging_log_dirty_op(struct domain *d, bytes = (unsigned int)((sc->pages - pages + 7) >> 3); if ( likely(peek) ) { - if ( (l1 ? copy_to_guest_offset(sc->dirty_bitmap, - pages >> 3, (uint8_t *)l1, - bytes) - : clear_guest_offset(sc->dirty_bitmap, - pages >> 3, bytes)) != 0 ) + if ( pages >> (3 + PAGE_SHIFT) != + index_mapped >> (3 + PAGE_SHIFT) ) { - rv = -EFAULT; - goto out; + /* We need to map next page */ + paging_unlock(d); + unmap_dirty_bitmap(dirty_bitmap, page); + dirty_bitmap = map_dirty_bitmap(sc->dirty_bitmap, pages, + &page, &index_mapped); + paging_lock(d); + if ( dirty_bitmap == NULL ) + { + rv = -EFAULT; + goto out; + } + goto again; } + BUG_ON(((pages >> 3) % PAGE_SIZE) + bytes > PAGE_SIZE); + if ( l1 ) + memcpy(dirty_bitmap + ((pages >> 3) % PAGE_SIZE), + (uint8_t *)l1, bytes); + else + memset(dirty_bitmap + ((pages >> 3) % PAGE_SIZE), 0, + bytes); } pages += bytes << 3; if ( l1 ) @@ -549,12 +626,14 @@ static int paging_log_dirty_op(struct domain *d, * paging modes (shadow or hap). Safe because the domain is paused. */ d->arch.paging.log_dirty.clean_dirty_bitmap(d); } + unmap_dirty_bitmap(dirty_bitmap, page); domain_unpause(d); return rv; out: d->arch.paging.preempt.dom = NULL; paging_unlock(d); + unmap_dirty_bitmap(dirty_bitmap, page); domain_unpause(d); if ( l1 ) -- 1.9.5 (Apple Git-50.3) _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |