[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-changelog] [xen-unstable] [IA64] Cleanup: remove unused declarations, add static and reindent



# HG changeset patch
# User Alex Williamson <alex.williamson@xxxxxx>
# Date 1190049981 21600
# Node ID b91d16ab68be801382c4b7ef99fa6b6929856bae
# Parent  0902e4aae8100c8130617ec7170cf26a6692bacb
[IA64] Cleanup: remove unused declarations, add static and reindent

Signed-off-by: Tristan Gingold <tgingold@xxxxxxx>
---
 xen/arch/ia64/vmx/vmmu.c                     |   67 ------------
 xen/arch/ia64/vmx/vtlb.c                     |  145 ++++++++++++++-------------
 xen/include/asm-ia64/linux-xen/asm/pgtable.h |    1 
 xen/include/asm-ia64/vmmu.h                  |   93 -----------------
 4 files changed, 82 insertions(+), 224 deletions(-)

diff -r 0902e4aae810 -r b91d16ab68be xen/arch/ia64/vmx/vmmu.c
--- a/xen/arch/ia64/vmx/vmmu.c  Mon Sep 17 11:08:46 2007 -0600
+++ b/xen/arch/ia64/vmx/vmmu.c  Mon Sep 17 11:26:21 2007 -0600
@@ -66,7 +66,7 @@ custom_param("vti_vhpt_size", parse_vhpt
  * Input:
  *  d: 
  */
-u64 get_mfn(struct domain *d, u64 gpfn)
+static u64 get_mfn(struct domain *d, u64 gpfn)
 {
 //    struct domain *d;
     u64    xen_gppn, xen_mppn, mpfn;
@@ -91,68 +91,6 @@ u64 get_mfn(struct domain *d, u64 gpfn)
     mpfn = mpfn | (((1UL <<(PAGE_SHIFT-ARCH_PAGE_SHIFT))-1)&gpfn);
     return mpfn;
     
-}
-
-/*
- * The VRN bits of va stand for which rr to get.
- */
-//ia64_rr vmmu_get_rr(struct vcpu *vcpu, u64 va)
-//{
-//    ia64_rr   vrr;
-//    vcpu_get_rr(vcpu, va, &vrr.rrval);
-//    return vrr;
-//}
-
-/*
-void recycle_message(thash_cb_t *hcb, u64 para)
-{
-    if(hcb->ht == THASH_VHPT)
-    {
-        printk("ERROR : vhpt recycle happenning!!!\n");
-    }
-    printk("hcb=%p recycled with %lx\n",hcb,para);
-}
- */
-
-/*
- * Purge all guest TCs in logical processor.
- * Instead of purging all LP TCs, we should only purge   
- * TCs that belong to this guest.
- */
-void
-purge_machine_tc_by_domid(domid_t domid)
-{
-#ifndef PURGE_GUEST_TC_ONLY
-    // purge all TCs
-    struct ia64_pal_retval  result;
-    u64 addr;
-    u32 count1,count2;
-    u32 stride1,stride2;
-    u32 i,j;
-    u64 psr;
-
-    result = ia64_pal_call_static(PAL_PTCE_INFO,0,0,0, 0);
-    if ( result.status != 0 ) {
-        panic ("PAL_PTCE_INFO failed\n");
-    }
-    addr = result.v0;
-    count1 = HIGH_32BITS(result.v1);
-    count2 = LOW_32BITS (result.v1);
-    stride1 = HIGH_32BITS(result.v2);
-    stride2 = LOW_32BITS (result.v2);
-
-    local_irq_save(psr);
-    for (i=0; i<count1; i++) {
-        for (j=0; j<count2; j++) {
-            ia64_ptce(addr);
-            addr += stride2;
-        }
-        addr += stride1;
-    }
-    local_irq_restore(psr);
-#else
-    // purge all TCs belong to this guest.
-#endif
 }
 
 static int init_domain_vhpt(struct vcpu *v)
@@ -313,7 +251,8 @@ fetch_code(VCPU *vcpu, u64 gip, IA64_BUN
     }
     if( gpip){
         mfn = gmfn_to_mfn(vcpu->domain, gpip >>PAGE_SHIFT);
-        if( mfn == INVALID_MFN )  panic_domain(vcpu_regs(vcpu),"fetch_code: 
invalid memory\n");
+        if (mfn == INVALID_MFN)
+            panic_domain(vcpu_regs(vcpu), "fetch_code: invalid memory\n");
         maddr = (mfn << PAGE_SHIFT) | (gpip & (PAGE_SIZE - 1));
     }else{
         tlb = vhpt_lookup(gip);
diff -r 0902e4aae810 -r b91d16ab68be xen/arch/ia64/vmx/vtlb.c
--- a/xen/arch/ia64/vmx/vtlb.c  Mon Sep 17 11:08:46 2007 -0600
+++ b/xen/arch/ia64/vmx/vtlb.c  Mon Sep 17 11:26:21 2007 -0600
@@ -69,9 +69,9 @@ __is_tr_overlap(thash_data_t *trp, u64 r
         return 0;
     }
     sa1 = trp->vadr;
-    ea1 = sa1 + PSIZE(trp->ps) -1;
+    ea1 = sa1 + PSIZE(trp->ps) - 1;
     eva -= 1;
-    if ( (sva>ea1) || (sa1>eva) )
+    if (sva > ea1 || sa1 > eva)
         return 0;
     else
         return 1;
@@ -85,10 +85,11 @@ static thash_data_t *__vtr_lookup(VCPU *
     int  i;
     u64 rid;
     vcpu_get_rr(vcpu, va, &rid);
-    rid = rid & RR_RID_MASK;;
+    rid &= RR_RID_MASK;
     if (is_data) {
         if (vcpu_quick_region_check(vcpu->arch.dtr_regions,va)) {
-            for (trp =(thash_data_t *) vcpu->arch.dtrs,i=0; i<NDTRS; i++, 
trp++) {
+            trp = (thash_data_t *)vcpu->arch.dtrs;
+            for (i = 0;  i < NDTRS; i++, trp++) {
                 if (__is_tr_translated(trp, rid, va)) {
                     return trp;
                 }
@@ -97,7 +98,8 @@ static thash_data_t *__vtr_lookup(VCPU *
     }
     else {
         if (vcpu_quick_region_check(vcpu->arch.itr_regions,va)) {
-            for (trp =(thash_data_t *) vcpu->arch.itrs,i=0; i<NITRS; i++, 
trp++) {
+            trp = (thash_data_t *)vcpu->arch.itrs;
+            for (i = 0; i < NITRS; i++, trp++) {
                 if (__is_tr_translated(trp, rid, va)) {
                     return trp;
                 }
@@ -107,35 +109,34 @@ static thash_data_t *__vtr_lookup(VCPU *
     return NULL;
 }
 
-
 static void thash_recycle_cch(thash_cb_t *hcb, thash_data_t *hash)
 {
     thash_data_t *p, *q;
-    int i=0;
-    
-    p=hash;
-    for(i=0; i < MAX_CCN_DEPTH; i++){
-        p=p->next;
-    }
-    q=hash->next;
-    hash->len=0;
-    hash->next=0;
-    p->next=hcb->cch_freelist;
-    hcb->cch_freelist=q;
-}
-
-
-
+    int i = 0;
+     
+    p = hash;
+    for (i = 0; i < MAX_CCN_DEPTH; i++) {
+        p = p->next;
+    }
+    q = hash->next;
+    hash->len = 0;
+    hash->next = 0;
+    p->next = hcb->cch_freelist;
+    hcb->cch_freelist = q;
+}
 
 static void vmx_vhpt_insert(thash_cb_t *hcb, u64 pte, u64 itir, u64 ifa)
 {
     u64 tag;
     ia64_rr rr;
     thash_data_t *head, *cch;
+
     pte = pte & ~PAGE_FLAGS_RV_MASK;
     rr.rrval = ia64_get_rr(ifa);
     head = (thash_data_t *)ia64_thash(ifa);
     tag = ia64_ttag(ifa);
+
+    /* Find a free (ie invalid) entry.  */
     cch = head;
     while (cch) {    
         if (INVALID_VHPT(cch))
@@ -143,16 +144,16 @@ static void vmx_vhpt_insert(thash_cb_t *
         cch = cch->next;
     }
     if (cch) {
+        /* As we insert in head, copy head.  */
         if (cch != head) {
             local_irq_disable();
             cch->page_flags = head->page_flags;
             cch->itir = head->itir;
-            cch->etag  = head->etag;
+            cch->etag = head->etag;
             head->ti = 1;
             local_irq_enable();
         }
-    }
-    else{
+    } else {
         if (head->len >= MAX_CCN_DEPTH) {
             thash_recycle_cch(hcb, head);
             cch = cch_alloc(hcb);
@@ -181,7 +182,7 @@ void thash_vhpt_insert(VCPU *v, u64 pte,
     ia64_rr mrr;
 
     mrr.rrval = ia64_get_rr(va);
-    phy_pte=translate_phy_pte(v, &pte, itir, va);
+    phy_pte = translate_phy_pte(v, &pte, itir, va);
 
     if (itir_ps(itir) >= mrr.ps) {
         vmx_vhpt_insert(vcpu_get_vhpt(v), phy_pte, itir, va);
@@ -232,26 +233,31 @@ thash_data_t * vhpt_lookup(u64 va)
 {
     thash_data_t *hash, *head;
     u64 tag, pte, itir;
+
     head = (thash_data_t *)ia64_thash(va);
-    hash=head;
+    hash = head;
     tag = ia64_ttag(va);
-    do{
-        if(hash->etag == tag)
+    do {
+        if (hash->etag == tag)
             break;
-        hash=hash->next;
-    }while(hash);
-    if(hash && hash!=head){
+        hash = hash->next;
+    } while(hash);
+    if (hash && hash != head) {
+        /* Put the entry on the front of the list (ie swap hash and head).  */
         pte = hash->page_flags;
         hash->page_flags = head->page_flags;
         head->page_flags = pte;
+
         tag = hash->etag;
         hash->etag = head->etag;
         head->etag = tag;
+
         itir = hash->itir;
         hash->itir = head->itir;
         head->itir = itir;
+
         head->len = hash->len;
-        hash->len=0;
+        hash->len = 0;
         return head;
     }
     return hash;
@@ -368,14 +374,15 @@ void thash_recycle_cch_all(thash_cb_t *h
 {
     int num;
     thash_data_t *head;
-    head=hcb->hash;
+
+    head = hcb->hash;
     num = (hcb->hash_sz/sizeof(thash_data_t));
-    do{
+    do {
         head->len = 0;
         head->next = 0;
         head++;
         num--;
-    }while(num);
+    } while(num);
     cch_mem_init(hcb);
 }
 
@@ -409,6 +416,7 @@ void vtlb_insert(VCPU *v, u64 pte, u64 i
     /* u64 gppn, ppns, ppne; */
     u64 tag, len;
     thash_cb_t *hcb = &v->arch.vtlb;
+
     vcpu_get_rr(v, va, &vrr.rrval);
     vrr.ps = itir_ps(itir);
     VMX(v, psbits[va >> 61]) |= (1UL << vrr.ps);
@@ -419,13 +427,13 @@ void vtlb_insert(VCPU *v, u64 pte, u64 i
             len = cch->len;
             cch->page_flags = pte;
             cch->len = len;
-            cch->itir=itir;
-            cch->etag=tag;
+            cch->itir = itir;
+            cch->etag = tag;
             return;
         }
         cch = cch->next;
     }
-    if (hash_table->len>=MAX_CCN_DEPTH){
+    if (hash_table->len >= MAX_CCN_DEPTH) {
         thash_recycle_cch(hcb, hash_table);
         cch = cch_alloc(hcb);
     }
@@ -450,11 +458,12 @@ int vtr_find_overlap(VCPU *vcpu, u64 va,
     u64 end, rid;
 
     vcpu_get_rr(vcpu, va, &rid);
-    rid = rid & RR_RID_MASK;;
+    rid &= RR_RID_MASK;
     end = va + PSIZE(ps);
     if (is_data) {
-        if (vcpu_quick_region_check(vcpu->arch.dtr_regions,va)) {
-            for (trp =(thash_data_t *) vcpu->arch.dtrs,i=0; i<NDTRS; i++, 
trp++) {
+        if (vcpu_quick_region_check(vcpu->arch.dtr_regions, va)) {
+            trp = (thash_data_t *)vcpu->arch.dtrs;
+            for (i = 0; i < NDTRS; i++, trp++) {
                 if (__is_tr_overlap(trp, rid, va, end )) {
                     return i;
                 }
@@ -463,7 +472,8 @@ int vtr_find_overlap(VCPU *vcpu, u64 va,
     }
     else {
         if (vcpu_quick_region_check(vcpu->arch.itr_regions,va)) {
-            for (trp =(thash_data_t *) vcpu->arch.itrs,i=0; i<NITRS; i++, 
trp++) {
+            trp = (thash_data_t *)vcpu->arch.itrs;
+            for (i = 0; i < NITRS; i++, trp++) {
                 if (__is_tr_overlap(trp, rid, va, end )) {
                     return i;
                 }
@@ -478,7 +488,7 @@ int vtr_find_overlap(VCPU *vcpu, u64 va,
  */
 void thash_purge_entries(VCPU *v, u64 va, u64 ps)
 {
-    if(vcpu_quick_region_check(v->arch.tc_regions,va))
+    if (vcpu_quick_region_check(v->arch.tc_regions, va))
         vtlb_purge(v, va, ps);
     vhpt_purge(v, va, ps);
 }
@@ -497,6 +507,7 @@ u64 translate_phy_pte(VCPU *v, u64 *pte,
     u64 ps, ps_mask, paddr, maddr;
 //    ia64_rr rr;
     union pte_flags phy_pte;
+
     ps = itir_ps(itir);
     ps_mask = ~((1UL << ps) - 1);
     phy_pte.val = *pte;
@@ -536,7 +547,7 @@ int thash_purge_and_insert(VCPU *v, u64 
 
     ps = itir_ps(itir);
     mrr.rrval = ia64_get_rr(ifa);
-    if(VMX_DOMAIN(v)){
+    if (VMX_DOMAIN(v)) {
         phy_pte = translate_phy_pte(v, &pte, itir, ifa);
 
         if (pte & VTLB_PTE_IO)
@@ -544,18 +555,18 @@ int thash_purge_and_insert(VCPU *v, u64 
         vtlb_purge(v, ifa, ps);
         vhpt_purge(v, ifa, ps);
         if (ps == mrr.ps) {
-            if(!(pte&VTLB_PTE_IO)){
+            if (!(pte & VTLB_PTE_IO)) {
                 vmx_vhpt_insert(&v->arch.vhpt, phy_pte, itir, ifa);
             }
             else{
                 vtlb_insert(v, pte, itir, ifa);
-                vcpu_quick_region_set(PSCBX(v,tc_regions),ifa);
+                vcpu_quick_region_set(PSCBX(v, tc_regions), ifa);
             }
         }
         else if (ps > mrr.ps) {
             vtlb_insert(v, pte, itir, ifa);
-            vcpu_quick_region_set(PSCBX(v,tc_regions),ifa);
-            if(!(pte&VTLB_PTE_IO)){
+            vcpu_quick_region_set(PSCBX(v, tc_regions), ifa);
+            if (!(pte & VTLB_PTE_IO)) {
                 vmx_vhpt_insert(&v->arch.vhpt, phy_pte, itir, ifa);
             }
         }
@@ -572,9 +583,9 @@ int thash_purge_and_insert(VCPU *v, u64 
     }
     else{
         phy_pte = translate_phy_pte(v, &pte, itir, ifa);
-        if(ps!=PAGE_SHIFT){
+        if (ps != PAGE_SHIFT) {
             vtlb_insert(v, pte, itir, ifa);
-            vcpu_quick_region_set(PSCBX(v,tc_regions),ifa);
+            vcpu_quick_region_set(PSCBX(v, tc_regions), ifa);
         }
         machine_tlb_purge(ifa, ps);
         vmx_vhpt_insert(&v->arch.vhpt, phy_pte, itir, ifa);
@@ -593,13 +604,13 @@ void thash_purge_all(VCPU *v)
     int num;
     thash_data_t *head;
     thash_cb_t  *vtlb,*vhpt;
-    vtlb =&v->arch.vtlb;
-    vhpt =&v->arch.vhpt;
+    vtlb = &v->arch.vtlb;
+    vhpt = &v->arch.vhpt;
 
     for (num = 0; num < 8; num++)
         VMX(v, psbits[num]) = 0;
     
-    head=vtlb->hash;
+    head = vtlb->hash;
     num = (vtlb->hash_sz/sizeof(thash_data_t));
     do{
         head->page_flags = 0;
@@ -608,10 +619,10 @@ void thash_purge_all(VCPU *v)
         head->next = 0;
         head++;
         num--;
-    }while(num);
+    } while(num);
     cch_mem_init(vtlb);
     
-    head=vhpt->hash;
+    head = vhpt->hash;
     num = (vhpt->hash_sz/sizeof(thash_data_t));
     do{
         head->page_flags = 0;
@@ -619,7 +630,7 @@ void thash_purge_all(VCPU *v)
         head->next = 0;
         head++;
         num--;
-    }while(num);
+    } while(num);
     cch_mem_init(vhpt);
     local_flush_tlb_all();
 }
@@ -635,19 +646,19 @@ void thash_purge_all(VCPU *v)
 
 thash_data_t *vtlb_lookup(VCPU *v, u64 va,int is_data)
 {
-    thash_data_t  *cch;
-    u64     psbits, ps, tag;
+    thash_data_t *cch;
+    u64 psbits, ps, tag;
     ia64_rr vrr;
-    thash_cb_t * hcb= &v->arch.vtlb;
-
-    cch = __vtr_lookup(v, va, is_data);;
+    thash_cb_t *hcb = &v->arch.vtlb;
+
+    cch = __vtr_lookup(v, va, is_data);
     if (cch)
        return cch;
 
-    if (vcpu_quick_region_check(v->arch.tc_regions,va) == 0)
+    if (vcpu_quick_region_check(v->arch.tc_regions, va) == 0)
         return NULL;
     psbits = VMX(v, psbits[(va >> 61)]);
-    vcpu_get_rr(v,va,&vrr.rrval);
+    vcpu_get_rr(v, va, &vrr.rrval);
     while (psbits) {
         ps = __ffs(psbits);
         psbits &= ~(1UL << ps);
@@ -676,16 +687,16 @@ static void thash_init(thash_cb_t *hcb, 
     hcb->pta.ve = 1;
     hcb->pta.size = sz;
     
-    head=hcb->hash;
+    head = hcb->hash;
     num = (hcb->hash_sz/sizeof(thash_data_t));
-    do{
+    do {
         head->page_flags = 0;
         head->itir = 0;
-        head->etag = 1UL<<63;
+        head->etag = 1UL << 63;
         head->next = 0;
         head++;
         num--;
-    }while(num);
+    } while(num);
 
     hcb->cch_free_idx = 0;
     hcb->cch_freelist = NULL;
diff -r 0902e4aae810 -r b91d16ab68be 
xen/include/asm-ia64/linux-xen/asm/pgtable.h
--- a/xen/include/asm-ia64/linux-xen/asm/pgtable.h      Mon Sep 17 11:08:46 
2007 -0600
+++ b/xen/include/asm-ia64/linux-xen/asm/pgtable.h      Mon Sep 17 11:26:21 
2007 -0600
@@ -73,6 +73,7 @@
 #ifdef XEN
 #define _PAGE_VIRT_D           (__IA64_UL(1) << 53)    /* Virtual dirty bit */
 #define _PAGE_PROTNONE         0
+#define _PAGE_PL_PRIV          (CONFIG_CPL0_EMUL << 7)
 
 #ifdef CONFIG_XEN_IA64_TLB_TRACK
 #define _PAGE_TLB_TRACKING_BIT          54
diff -r 0902e4aae810 -r b91d16ab68be xen/include/asm-ia64/vmmu.h
--- a/xen/include/asm-ia64/vmmu.h       Mon Sep 17 11:08:46 2007 -0600
+++ b/xen/include/asm-ia64/vmmu.h       Mon Sep 17 11:26:21 2007 -0600
@@ -28,7 +28,6 @@
 #define     DEFAULT_VHPT_SZ     (23) // 8M hash + 8M c-chain for VHPT
 #define     VTLB(v,_x)          (v->arch.vtlb._x)
 #define     VHPT(v,_x)          (v->arch.vhpt._x)
-#define     _PAGE_PL_PRIV       (CONFIG_CPL0_EMUL << 7)
 
 #ifndef __ASSEMBLY__
 
@@ -39,34 +38,6 @@
 #include <asm/regionreg.h>
 #include <asm/vmx_mm_def.h>
 #include <asm/bundle.h>
-//#define         THASH_TLB_TR            0
-//#define         THASH_TLB_TC            1
-
-
-// bit definition of TR, TC search cmobination
-//#define         THASH_SECTION_TR        (1<<0)
-//#define         THASH_SECTION_TC        (1<<1)
-
-/*
- * Next bit definition must be same with THASH_TLB_XX
-#define         PTA_BASE_SHIFT          (15)
- */
-
-
-
-
-#define HIGH_32BITS(x)  bits(x,32,63)
-#define LOW_32BITS(x)   bits(x,0,31)
-
-typedef union search_section {
-    struct {
-        u32 tr : 1;
-        u32 tc : 1;
-        u32 rsv: 30;
-    };
-    u32     v;
-} search_section_t;
-
 
 enum {
     ISIDE_TLB=0,
@@ -169,28 +140,6 @@ static inline u64 xen_to_arch_ppn(u64 xp
     return (xppn <<(PAGE_SHIFT- ARCH_PAGE_SHIFT));
 }
 
-typedef enum {
-    THASH_TLB=0,
-    THASH_VHPT
-} THASH_TYPE;
-
-struct thash_cb;
-/*
- * Use to calculate the HASH index of thash_data_t.
- */
-typedef u64 *(THASH_FN)(PTA pta, u64 va);
-typedef u64 *(TTAG_FN)(PTA pta, u64 va);
-typedef u64 *(GET_MFN_FN)(domid_t d, u64 gpfn, u64 pages);
-typedef void *(REM_NOTIFIER_FN)(struct thash_cb *hcb, thash_data_t *entry);
-typedef void (RECYCLE_FN)(struct thash_cb *hc, u64 para);
-typedef ia64_rr (GET_RR_FN)(struct vcpu *vcpu, u64 reg);
-typedef thash_data_t *(FIND_OVERLAP_FN)(struct thash_cb *hcb, 
-        u64 va, u64 ps, int rid, char cl, search_section_t s_sect);
-typedef thash_data_t *(FIND_NEXT_OVL_FN)(struct thash_cb *hcb);
-typedef void (REM_THASH_FN)(struct thash_cb *hcb, thash_data_t *entry);
-typedef void (INS_THASH_FN)(struct thash_cb *hcb, thash_data_t *entry, u64 va);
-
-
 typedef struct thash_cb {
     /* THASH base information */
     thash_data_t    *hash; // hash table pointer, aligned at thash_sz.
@@ -224,45 +173,6 @@ extern void thash_free(thash_cb_t *hcb);
 //extern void thash_insert(thash_cb_t *hcb, thash_data_t *entry, u64 va);
 //extern void thash_tr_insert(thash_cb_t *hcb, thash_data_t *entry, u64 va, 
int idx);
 extern int vtr_find_overlap(struct vcpu *vcpu, u64 va, u64 ps, int is_data);
-extern u64 get_mfn(struct domain *d, u64 gpfn);
-/*
- * Force to delete a found entry no matter TR or foreign map for TLB.
- *    NOTES:
- *      1: TLB entry may be TR, TC or Foreign Map. For TR entry,
- *         itr[]/dtr[] need to be updated too.
- *      2: This API must be called after thash_find_overlap() or
- *         thash_find_next_overlap().
- *      3: Return TRUE or FALSE
- *
- */
-extern void thash_remove(thash_cb_t *hcb, thash_data_t *entry);
-extern void thash_tr_remove(thash_cb_t *hcb, thash_data_t *entry/*, int idx*/);
-
-/*
- * Find an overlap entry in hash table and its collision chain.
- * Refer to SDM2 4.1.1.4 for overlap definition.
- *    PARAS:
- *      1: in: TLB format entry, rid:ps must be same with vrr[].
- *             va & ps identify the address space for overlap lookup
- *      2: section can be combination of TR, TC and FM. (THASH_SECTION_XX)
- *      3: cl means I side or D side.
- *    RETURNS:
- *      NULL to indicate the end of findings.
- *    NOTES:
- *
- */
-extern thash_data_t *thash_find_overlap(thash_cb_t *hcb, 
-                        thash_data_t *in, search_section_t s_sect);
-extern thash_data_t *thash_find_overlap_ex(thash_cb_t *hcb, 
-                u64 va, u64 ps, int rid, char cl, search_section_t s_sect);
-
-
-/*
- * Similar with thash_find_overlap but find next entry.
- *    NOTES:
- *      Intermediate position information is stored in hcb->priv.
- */
-extern thash_data_t *thash_find_next_overlap(thash_cb_t *hcb);
 
 /*
  * Find and purge overlap entries in hash table and its collision chain.
@@ -290,7 +200,6 @@ extern void thash_purge_all(struct vcpu 
  *
  */
 extern thash_data_t *vtlb_lookup(struct vcpu *v,u64 va,int is_data);
-extern int thash_lock_tc(thash_cb_t *hcb, u64 va, u64 size, int rid, char cl, 
int lock);
 
 
 #define   ITIR_RV_MASK      (((1UL<<32)-1)<<32 | 0x3)
@@ -298,12 +207,10 @@ extern int thash_lock_tc(thash_cb_t *hcb
 #define   PAGE_FLAGS_AR_PL_MASK ((0x7UL<<9)|(0x3UL<<7))
 extern u64 machine_ttag(PTA pta, u64 va);
 extern u64 machine_thash(PTA pta, u64 va);
-extern void purge_machine_tc_by_domid(domid_t domid);
 extern void machine_tlb_insert(struct vcpu *v, thash_data_t *tlb);
 extern ia64_rr vmmu_get_rr(struct vcpu *vcpu, u64 va);
 extern int init_domain_tlb(struct vcpu *v);
 extern void free_domain_tlb(struct vcpu *v);
-extern thash_data_t * vsa_thash(PTA vpta, u64 va, u64 vrr, u64 *tag);
 extern thash_data_t * vhpt_lookup(u64 va);
 extern void machine_tlb_purge(u64 va, u64 ps);
 extern unsigned long fetch_code(struct vcpu *vcpu, u64 gip, IA64_BUNDLE 
*pbundle);

_______________________________________________
Xen-changelog mailing list
Xen-changelog@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-changelog


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.