[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-devel] [PATCH] move dirty_vram in struct hvm_domain



Hi all,
this patch moves dirty_vram from struct domain to the more appropriate
struct hvm_domain (d->arch.hvm_domain), since it is an hvm only feature.

Signed-off-by: Stefano Stabellini <stefano.stabellini@xxxxxxxxxxxxx>

---

diff -r 6eff3fe96aff xen/arch/x86/mm/hap/hap.c
--- a/xen/arch/x86/mm/hap/hap.c Fri Jun 05 09:32:03 2009 +0100
+++ b/xen/arch/x86/mm/hap/hap.c Fri Jun 05 12:25:34 2009 +0100
@@ -58,8 +58,9 @@
 int hap_enable_vram_tracking(struct domain *d)
 {
     int i;
+    struct sh_dirty_vram *dirty_vram = d->arch.hvm_domain.dirty_vram;
 
-    if ( !d->dirty_vram )
+    if ( !dirty_vram )
         return -EINVAL;
 
     /* turn on PG_log_dirty bit in paging mode */
@@ -68,7 +69,7 @@
     hap_unlock(d);
 
     /* set l1e entries of P2M table to be read-only. */
-    for (i = d->dirty_vram->begin_pfn; i < d->dirty_vram->end_pfn; i++)
+    for (i = dirty_vram->begin_pfn; i < dirty_vram->end_pfn; i++)
         p2m_change_type(d, i, p2m_ram_rw, p2m_ram_logdirty);
 
     flush_tlb_mask(&d->domain_dirty_cpumask);
@@ -78,8 +79,9 @@
 int hap_disable_vram_tracking(struct domain *d)
 {
     int i;
+    struct sh_dirty_vram *dirty_vram = d->arch.hvm_domain.dirty_vram;
 
-    if ( !d->dirty_vram )
+    if ( !dirty_vram )
         return -EINVAL;
 
     hap_lock(d);
@@ -87,7 +89,7 @@
     hap_unlock(d);
 
     /* set l1e entries of P2M table with normal mode */
-    for (i = d->dirty_vram->begin_pfn; i < d->dirty_vram->end_pfn; i++)
+    for (i = dirty_vram->begin_pfn; i < dirty_vram->end_pfn; i++)
         p2m_change_type(d, i, p2m_ram_rw, p2m_ram_logdirty);
 
     flush_tlb_mask(&d->domain_dirty_cpumask);
@@ -97,12 +99,13 @@
 void hap_clean_vram_tracking(struct domain *d)
 {
     int i;
+    struct sh_dirty_vram *dirty_vram = d->arch.hvm_domain.dirty_vram;
 
-    if ( !d->dirty_vram )
+    if ( !dirty_vram )
         return;
 
     /* set l1e entries of P2M table to be read-only. */
-    for (i = d->dirty_vram->begin_pfn; i < d->dirty_vram->end_pfn; i++)
+    for (i = dirty_vram->begin_pfn; i < dirty_vram->end_pfn; i++)
         p2m_change_type(d, i, p2m_ram_rw, p2m_ram_logdirty);
 
     flush_tlb_mask(&d->domain_dirty_cpumask);
@@ -121,30 +124,32 @@
                          XEN_GUEST_HANDLE_64(uint8) dirty_bitmap)
 {
     long rc = 0;
+    struct sh_dirty_vram *dirty_vram = d->arch.hvm_domain.dirty_vram;
 
     if ( nr )
     {
-        if ( paging_mode_log_dirty(d) && d->dirty_vram )
+        if ( paging_mode_log_dirty(d) && dirty_vram )
         {
-            if ( begin_pfn != d->dirty_vram->begin_pfn ||
-                 begin_pfn + nr != d->dirty_vram->end_pfn )
+            if ( begin_pfn != dirty_vram->begin_pfn ||
+                 begin_pfn + nr != dirty_vram->end_pfn )
             {
                 paging_log_dirty_disable(d);
-                d->dirty_vram->begin_pfn = begin_pfn;
-                d->dirty_vram->end_pfn = begin_pfn + nr;
+                dirty_vram->begin_pfn = begin_pfn;
+                dirty_vram->end_pfn = begin_pfn + nr;
                 rc = paging_log_dirty_enable(d);
                 if (rc != 0)
                     goto param_fail;
             }
         }
-        else if ( !paging_mode_log_dirty(d) && !d->dirty_vram )
+        else if ( !paging_mode_log_dirty(d) && !dirty_vram )
         {
             rc -ENOMEM;
-            if ( (d->dirty_vram = xmalloc(struct sh_dirty_vram)) == NULL )
+            if ( (dirty_vram = xmalloc(struct sh_dirty_vram)) == NULL )
                 goto param_fail;
 
-            d->dirty_vram->begin_pfn = begin_pfn;
-            d->dirty_vram->end_pfn = begin_pfn + nr;
+            dirty_vram->begin_pfn = begin_pfn;
+            dirty_vram->end_pfn = begin_pfn + nr;
+            d->arch.hvm_domain.dirty_vram = dirty_vram;
             hap_vram_tracking_init(d);
             rc = paging_log_dirty_enable(d);
             if (rc != 0)
@@ -152,7 +157,7 @@
         }
         else
         {
-            if ( !paging_mode_log_dirty(d) && d->dirty_vram )
+            if ( !paging_mode_log_dirty(d) && dirty_vram )
                 rc = -EINVAL;
             else
                 rc = -ENODATA;
@@ -163,10 +168,10 @@
     }
     else
     {
-        if ( paging_mode_log_dirty(d) && d->dirty_vram ) {
+        if ( paging_mode_log_dirty(d) && dirty_vram ) {
             rc = paging_log_dirty_disable(d);
-            xfree(d->dirty_vram);
-            d->dirty_vram = NULL;
+            xfree(dirty_vram);
+            dirty_vram = d->arch.hvm_domain.dirty_vram = NULL;
         } else
             rc = 0;
     }
@@ -174,10 +179,10 @@
     return rc;
 
 param_fail:
-    if ( d->dirty_vram )
+    if ( dirty_vram )
     {
-        xfree(d->dirty_vram);
-        d->dirty_vram = NULL;
+        xfree(dirty_vram);
+        dirty_vram = d->arch.hvm_domain.dirty_vram = NULL;
     }
     return rc;
 }
@@ -220,11 +225,12 @@
 
 void hap_logdirty_init(struct domain *d)
 {
-    if ( paging_mode_log_dirty(d) && d->dirty_vram )
+    struct sh_dirty_vram *dirty_vram = d->arch.hvm_domain.dirty_vram;
+    if ( paging_mode_log_dirty(d) && dirty_vram )
     {
         paging_log_dirty_disable(d);
-        xfree(d->dirty_vram);
-        d->dirty_vram = NULL;
+        xfree(dirty_vram);
+        dirty_vram = d->arch.hvm_domain.dirty_vram = NULL;
     }
 
     /* Reinitialize logdirty mechanism */
diff -r 6eff3fe96aff xen/arch/x86/mm/shadow/common.c
--- a/xen/arch/x86/mm/shadow/common.c   Fri Jun 05 09:32:03 2009 +0100
+++ b/xen/arch/x86/mm/shadow/common.c   Fri Jun 05 12:25:34 2009 +0100
@@ -3259,11 +3259,11 @@
      * calls now that we've torn down the bitmap */
     d->arch.paging.mode &= ~PG_log_dirty;
 
-    if (d->dirty_vram) {
-        xfree(d->dirty_vram->sl1ma);
-        xfree(d->dirty_vram->dirty_bitmap);
-        xfree(d->dirty_vram);
-        d->dirty_vram = NULL;
+    if (d->arch.hvm_domain.dirty_vram) {
+        xfree(d->arch.hvm_domain.dirty_vram->sl1ma);
+        xfree(d->arch.hvm_domain.dirty_vram->dirty_bitmap);
+        xfree(d->arch.hvm_domain.dirty_vram);
+        d->arch.hvm_domain.dirty_vram = NULL;
     }
 
     shadow_unlock(d);
@@ -3583,6 +3583,7 @@
     int flush_tlb = 0;
     unsigned long i;
     p2m_type_t t;
+    struct sh_dirty_vram *dirty_vram = d->arch.hvm_domain.dirty_vram;
 
     if (end_pfn < begin_pfn
             || begin_pfn > d->arch.p2m->max_mapped_pfn
@@ -3591,16 +3592,16 @@
 
     shadow_lock(d);
 
-    if ( d->dirty_vram && (!nr ||
-             ( begin_pfn != d->dirty_vram->begin_pfn
-            || end_pfn   != d->dirty_vram->end_pfn )) )
+    if ( dirty_vram && (!nr ||
+             ( begin_pfn != dirty_vram->begin_pfn
+            || end_pfn   != dirty_vram->end_pfn )) )
     {
         /* Different tracking, tear the previous down. */
-        gdprintk(XENLOG_INFO, "stopping tracking VRAM %lx - %lx\n", 
d->dirty_vram->begin_pfn, d->dirty_vram->end_pfn);
-        xfree(d->dirty_vram->sl1ma);
-        xfree(d->dirty_vram->dirty_bitmap);
-        xfree(d->dirty_vram);
-        d->dirty_vram = NULL;
+        gdprintk(XENLOG_INFO, "stopping tracking VRAM %lx - %lx\n", 
dirty_vram->begin_pfn, dirty_vram->end_pfn);
+        xfree(dirty_vram->sl1ma);
+        xfree(dirty_vram->dirty_bitmap);
+        xfree(dirty_vram);
+        dirty_vram = d->arch.hvm_domain.dirty_vram = NULL;
     }
 
     if ( !nr )
@@ -3611,7 +3612,7 @@
 
     /* This should happen seldomly (Video mode change),
      * no need to be careful. */
-    if ( !d->dirty_vram )
+    if ( !dirty_vram )
     {
         /* Just recount from start. */
         for ( i = begin_pfn; i < end_pfn; i++ ) {
@@ -3623,29 +3624,30 @@
         gdprintk(XENLOG_INFO, "tracking VRAM %lx - %lx\n", begin_pfn, end_pfn);
 
         rc = -ENOMEM;
-        if ( (d->dirty_vram = xmalloc(struct sh_dirty_vram)) == NULL )
+        if ( (dirty_vram = xmalloc(struct sh_dirty_vram)) == NULL )
             goto out;
-        d->dirty_vram->begin_pfn = begin_pfn;
-        d->dirty_vram->end_pfn = end_pfn;
+        dirty_vram->begin_pfn = begin_pfn;
+        dirty_vram->end_pfn = end_pfn;
+        d->arch.hvm_domain.dirty_vram = dirty_vram;
 
-        if ( (d->dirty_vram->sl1ma = xmalloc_array(paddr_t, nr)) == NULL )
+        if ( (dirty_vram->sl1ma = xmalloc_array(paddr_t, nr)) == NULL )
             goto out_dirty_vram;
-        memset(d->dirty_vram->sl1ma, ~0, sizeof(paddr_t) * nr);
+        memset(dirty_vram->sl1ma, ~0, sizeof(paddr_t) * nr);
 
-        if ( (d->dirty_vram->dirty_bitmap = xmalloc_array(uint8_t, 
dirty_size)) == NULL )
+        if ( (dirty_vram->dirty_bitmap = xmalloc_array(uint8_t, dirty_size)) 
== NULL )
             goto out_sl1ma;
-        memset(d->dirty_vram->dirty_bitmap, 0, dirty_size);
+        memset(dirty_vram->dirty_bitmap, 0, dirty_size);
 
-        d->dirty_vram->last_dirty = NOW();
+        dirty_vram->last_dirty = NOW();
 
         /* Tell the caller that this time we could not track dirty bits. */
         rc = -ENODATA;
     }
-    else if (d->dirty_vram->last_dirty == -1)
+    else if (dirty_vram->last_dirty == -1)
     {
         /* still completely clean, just copy our empty bitmap */
         rc = -EFAULT;
-        if ( copy_to_guest(dirty_bitmap, d->dirty_vram->dirty_bitmap, 
dirty_size) == 0 )
+        if ( copy_to_guest(dirty_bitmap, dirty_vram->dirty_bitmap, dirty_size) 
== 0 )
             rc = 0;
     }
     else
@@ -3660,7 +3662,7 @@
             mfn_t mfn = gfn_to_mfn(d, begin_pfn + i, &t);
             struct page_info *page;
             int dirty = 0;
-            paddr_t sl1ma = d->dirty_vram->sl1ma[i];
+            paddr_t sl1ma = dirty_vram->sl1ma[i];
 
             if (mfn_x(mfn) == INVALID_MFN)
             {
@@ -3724,8 +3726,8 @@
 
             if ( dirty )
             {
-                d->dirty_vram->dirty_bitmap[i / 8] |= 1 << (i % 8);
-                d->dirty_vram->last_dirty = NOW();
+                dirty_vram->dirty_bitmap[i / 8] |= 1 << (i % 8);
+                dirty_vram->last_dirty = NOW();
             }
         }
 
@@ -3735,9 +3737,9 @@
 #endif
 
         rc = -EFAULT;
-        if ( copy_to_guest(dirty_bitmap, d->dirty_vram->dirty_bitmap, 
dirty_size) == 0 ) {
-            memset(d->dirty_vram->dirty_bitmap, 0, dirty_size);
-            if (d->dirty_vram->last_dirty + SECONDS(2) < NOW())
+        if ( copy_to_guest(dirty_bitmap, dirty_vram->dirty_bitmap, dirty_size) 
== 0 ) {
+            memset(dirty_vram->dirty_bitmap, 0, dirty_size);
+            if (dirty_vram->last_dirty + SECONDS(2) < NOW())
             {
                 /* was clean for more than two seconds, try to disable guest
                  * write access */
@@ -3746,7 +3748,7 @@
                     if (mfn_x(mfn) != INVALID_MFN)
                         flush_tlb |= sh_remove_write_access(d->vcpu[0], mfn, 
1, 0);
                 }
-                d->dirty_vram->last_dirty = -1;
+                dirty_vram->last_dirty = -1;
             }
             rc = 0;
         }
@@ -3756,10 +3758,10 @@
     goto out;
 
 out_sl1ma:
-    xfree(d->dirty_vram->sl1ma);
+    xfree(dirty_vram->sl1ma);
 out_dirty_vram:
-    xfree(d->dirty_vram);
-    d->dirty_vram = NULL;
+    xfree(dirty_vram);
+    dirty_vram = d->arch.hvm_domain.dirty_vram = NULL;
 
 out:
     shadow_unlock(d);
diff -r 6eff3fe96aff xen/arch/x86/mm/shadow/multi.c
--- a/xen/arch/x86/mm/shadow/multi.c    Fri Jun 05 09:32:03 2009 +0100
+++ b/xen/arch/x86/mm/shadow/multi.c    Fri Jun 05 12:25:34 2009 +0100
@@ -475,6 +475,7 @@
     guest_l1e_t guest_entry = { guest_intpte };
     shadow_l1e_t *sp = shadow_entry_ptr;
     struct domain *d = v->domain;
+    struct sh_dirty_vram *dirty_vram = d->arch.hvm_domain.dirty_vram;
     gfn_t target_gfn = guest_l1e_get_gfn(guest_entry);
     u32 pass_thru_flags;
     u32 gflags, sflags;
@@ -615,13 +616,13 @@
         }
     }
 
-    if ( unlikely((level == 1) && d->dirty_vram
-            && d->dirty_vram->last_dirty == -1
-            && gfn_x(target_gfn) >= d->dirty_vram->begin_pfn
-            && gfn_x(target_gfn) < d->dirty_vram->end_pfn) )
+    if ( unlikely((level == 1) && dirty_vram
+            && dirty_vram->last_dirty == -1
+            && gfn_x(target_gfn) >= dirty_vram->begin_pfn
+            && gfn_x(target_gfn) < dirty_vram->end_pfn) )
     {
         if ( ft & FETCH_TYPE_WRITE )
-            d->dirty_vram->last_dirty = NOW();
+            dirty_vram->last_dirty = NOW();
         else
             sflags &= ~_PAGE_RW;
     }
@@ -1042,22 +1043,23 @@
     mfn_t mfn = shadow_l1e_get_mfn(new_sl1e);
     int flags = shadow_l1e_get_flags(new_sl1e);
     unsigned long gfn;
-
-    if ( !d->dirty_vram         /* tracking disabled? */
+    struct sh_dirty_vram *dirty_vram = d->arch.hvm_domain.dirty_vram;
+
+    if ( !dirty_vram         /* tracking disabled? */
          || !(flags & _PAGE_RW) /* read-only mapping? */
          || !mfn_valid(mfn) )   /* mfn can be invalid in mmio_direct */
         return;
 
     gfn = mfn_to_gfn(d, mfn);
 
-    if ( (gfn >= d->dirty_vram->begin_pfn) && (gfn < d->dirty_vram->end_pfn) )
-    {
-        unsigned long i = gfn - d->dirty_vram->begin_pfn;
+    if ( (gfn >= dirty_vram->begin_pfn) && (gfn < dirty_vram->end_pfn) )
+    {
+        unsigned long i = gfn - dirty_vram->begin_pfn;
         struct page_info *page = mfn_to_page(mfn);
         
         if ( (page->u.inuse.type_info & PGT_count_mask) == 1 )
             /* Initial guest reference, record it */
-            d->dirty_vram->sl1ma[i] = pfn_to_paddr(mfn_x(sl1mfn))
+            dirty_vram->sl1ma[i] = pfn_to_paddr(mfn_x(sl1mfn))
                 | ((unsigned long)sl1e & ~PAGE_MASK);
     }
 }
@@ -1070,17 +1072,18 @@
     mfn_t mfn = shadow_l1e_get_mfn(old_sl1e);
     int flags = shadow_l1e_get_flags(old_sl1e);
     unsigned long gfn;
-
-    if ( !d->dirty_vram         /* tracking disabled? */
+    struct sh_dirty_vram *dirty_vram = d->arch.hvm_domain.dirty_vram;
+
+    if ( !dirty_vram         /* tracking disabled? */
          || !(flags & _PAGE_RW) /* read-only mapping? */
          || !mfn_valid(mfn) )   /* mfn can be invalid in mmio_direct */
         return;
 
     gfn = mfn_to_gfn(d, mfn);
 
-    if ( (gfn >= d->dirty_vram->begin_pfn) && (gfn < d->dirty_vram->end_pfn) )
-    {
-        unsigned long i = gfn - d->dirty_vram->begin_pfn;
+    if ( (gfn >= dirty_vram->begin_pfn) && (gfn < dirty_vram->end_pfn) )
+    {
+        unsigned long i = gfn - dirty_vram->begin_pfn;
         struct page_info *page = mfn_to_page(mfn);
         int dirty = 0;
         paddr_t sl1ma = pfn_to_paddr(mfn_x(sl1mfn))
@@ -1089,14 +1092,14 @@
         if ( (page->u.inuse.type_info & PGT_count_mask) == 1 )
         {
             /* Last reference */
-            if ( d->dirty_vram->sl1ma[i] == INVALID_PADDR ) {
+            if ( dirty_vram->sl1ma[i] == INVALID_PADDR ) {
                 /* We didn't know it was that one, let's say it is dirty */
                 dirty = 1;
             }
             else
             {
-                ASSERT(d->dirty_vram->sl1ma[i] == sl1ma);
-                d->dirty_vram->sl1ma[i] = INVALID_PADDR;
+                ASSERT(dirty_vram->sl1ma[i] == sl1ma);
+                dirty_vram->sl1ma[i] = INVALID_PADDR;
                 if ( flags & _PAGE_DIRTY )
                     dirty = 1;
             }
@@ -1106,10 +1109,10 @@
             /* We had more than one reference, just consider the page dirty. */
             dirty = 1;
             /* Check that it's not the one we recorded. */
-            if ( d->dirty_vram->sl1ma[i] == sl1ma )
+            if ( dirty_vram->sl1ma[i] == sl1ma )
             {
                 /* Too bad, we remembered the wrong one... */
-                d->dirty_vram->sl1ma[i] = INVALID_PADDR;
+                dirty_vram->sl1ma[i] = INVALID_PADDR;
             }
             else
             {
@@ -1119,8 +1122,8 @@
         }
         if ( dirty )
         {
-            d->dirty_vram->dirty_bitmap[i / 8] |= 1 << (i % 8);
-            d->dirty_vram->last_dirty = NOW();
+            dirty_vram->dirty_bitmap[i / 8] |= 1 << (i % 8);
+            dirty_vram->last_dirty = NOW();
         }
     }
 }
diff -r 6eff3fe96aff xen/include/asm-x86/hvm/domain.h
--- a/xen/include/asm-x86/hvm/domain.h  Fri Jun 05 09:32:03 2009 +0100
+++ b/xen/include/asm-x86/hvm/domain.h  Fri Jun 05 12:25:34 2009 +0100
@@ -66,6 +66,9 @@
     /* Memory ranges with pinned cache attributes. */
     struct list_head       pinned_cacheattr_ranges;
 
+    /* VRAM dirty support. */
+    struct sh_dirty_vram *dirty_vram;
+
     /* If one of vcpus of this domain is in no_fill_mode or
      * mtrr/pat between vcpus is not the same, set is_in_uc_mode
      */
diff -r 6eff3fe96aff xen/include/xen/sched.h
--- a/xen/include/xen/sched.h   Fri Jun 05 09:32:03 2009 +0100
+++ b/xen/include/xen/sched.h   Fri Jun 05 12:25:34 2009 +0100
@@ -267,9 +267,6 @@
      * cause a deadlock. Acquirers don't spin waiting; they preempt.
      */
     spinlock_t hypercall_deadlock_mutex;
-
-    /* VRAM dirty support. */
-    struct sh_dirty_vram *dirty_vram;
 
     /* transcendent memory, auto-allocated on first tmem op by each domain */
     void *tmem;

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.