[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH] minios: Fix width of shifted 1
minios: Fix width of shifted 1 Shifting 1 more that 32 bits requires 1 to be long-typed. Signed-off-by: Samuel Thibault <samuel.thibault@xxxxxxxxxxxxx> diff -r 3a5c0f873c74 extras/mini-os/hypervisor.c --- a/extras/mini-os/hypervisor.c Fri May 23 12:23:11 2008 +0100 +++ b/extras/mini-os/hypervisor.c Fri May 23 15:40:36 2008 +0100 @@ -55,12 +55,12 @@ while ( l1 != 0 ) { l1i = __ffs(l1); - l1 &= ~(1 << l1i); + l1 &= ~(1UL << l1i); while ( (l2 = active_evtchns(cpu, s, l1i)) != 0 ) { l2i = __ffs(l2); - l2 &= ~(1 << l2i); + l2 &= ~(1UL << l2i); port = (l1i * (sizeof(unsigned long) * 8)) + l2i; do_event(port, regs); diff -r 3a5c0f873c74 extras/mini-os/lib/xmalloc.c --- a/extras/mini-os/lib/xmalloc.c Fri May 23 12:23:11 2008 +0100 +++ b/extras/mini-os/lib/xmalloc.c Fri May 23 15:40:36 2008 +0100 @@ -127,7 +127,7 @@ if ( hdr == NULL ) return NULL; - hdr->size = (1 << (pageorder + PAGE_SHIFT)); + hdr->size = (1UL << (pageorder + PAGE_SHIFT)); /* Debugging aid. */ hdr->freelist.next = hdr->freelist.prev = NULL; diff -r 3a5c0f873c74 extras/mini-os/mm.c --- a/extras/mini-os/mm.c Fri May 23 12:23:11 2008 +0100 +++ b/extras/mini-os/mm.c Fri May 23 15:40:36 2008 +0100 @@ -58,7 +58,7 @@ #define PAGES_PER_MAPWORD (sizeof(unsigned long) * 8) #define allocated_in_map(_pn) \ -(alloc_bitmap[(_pn)/PAGES_PER_MAPWORD] & (1<<((_pn)&(PAGES_PER_MAPWORD-1)))) +(alloc_bitmap[(_pn)/PAGES_PER_MAPWORD] & (1UL<<((_pn)&(PAGES_PER_MAPWORD-1)))) /* * Hint regarding bitwise arithmetic in map_{alloc,free}: @@ -80,13 +80,13 @@ if ( curr_idx == end_idx ) { - alloc_bitmap[curr_idx] |= ((1<<end_off)-1) & -(1<<start_off); + alloc_bitmap[curr_idx] |= ((1UL<<end_off)-1) & -(1UL<<start_off); } else { - alloc_bitmap[curr_idx] |= -(1<<start_off); - while ( ++curr_idx < end_idx ) alloc_bitmap[curr_idx] = ~0L; - alloc_bitmap[curr_idx] |= (1<<end_off)-1; + alloc_bitmap[curr_idx] |= -(1UL<<start_off); + while ( ++curr_idx < end_idx ) alloc_bitmap[curr_idx] = ~0UL; + alloc_bitmap[curr_idx] |= (1UL<<end_off)-1; } } @@ -102,13 +102,13 @@ if ( curr_idx == end_idx ) { - alloc_bitmap[curr_idx] &= -(1<<end_off) | ((1<<start_off)-1); + alloc_bitmap[curr_idx] &= -(1UL<<end_off) | ((1UL<<start_off)-1); } else { - alloc_bitmap[curr_idx] &= (1<<start_off)-1; + alloc_bitmap[curr_idx] &= (1UL<<start_off)-1; while ( ++curr_idx != end_idx ) alloc_bitmap[curr_idx] = 0; - alloc_bitmap[curr_idx] &= -(1<<end_off); + alloc_bitmap[curr_idx] &= -(1UL<<end_off); } } @@ -178,7 +178,7 @@ head = free_head[order]; while(!FREELIST_EMPTY(head)) { - for(count = 0; count < 1<< head->level; count++) + for(count = 0; count < 1UL<< head->level; count++) { if(count + virt_to_pfn(head) - pfn_start < 1000) chunks[count + virt_to_pfn(head) - pfn_start] = current; @@ -235,13 +235,13 @@ * Next chunk is limited by alignment of min, but also * must not be bigger than remaining range. */ - for ( i = PAGE_SHIFT; (1<<(i+1)) <= range; i++ ) - if ( min & (1<<i) ) break; + for ( i = PAGE_SHIFT; (1UL<<(i+1)) <= range; i++ ) + if ( min & (1UL<<i) ) break; ch = (chunk_head_t *)min; - min += (1<<i); - range -= (1<<i); + min += (1UL<<i); + range -= (1UL<<i); ct = (chunk_tail_t *)min-1; i -= PAGE_SHIFT; ch->level = i; @@ -280,8 +280,8 @@ { /* Split into two equal parts. */ i--; - spare_ch = (chunk_head_t *)((char *)alloc_ch + (1<<(i+PAGE_SHIFT))); - spare_ct = (chunk_tail_t *)((char *)spare_ch + (1<<(i+PAGE_SHIFT)))-1; + spare_ch = (chunk_head_t *)((char *)alloc_ch + (1UL<<(i+PAGE_SHIFT))); + spare_ct = (chunk_tail_t *)((char *)spare_ch + (1UL<<(i+PAGE_SHIFT)))-1; /* Create new header for spare chunk. */ spare_ch->level = i; @@ -294,7 +294,7 @@ free_head[i] = spare_ch; } - map_alloc(PHYS_PFN(to_phys(alloc_ch)), 1<<order); + map_alloc(PHYS_PFN(to_phys(alloc_ch)), 1UL<<order); return((unsigned long)alloc_ch); @@ -312,16 +312,16 @@ unsigned long mask; /* First free the chunk */ - map_free(virt_to_pfn(pointer), 1 << order); + map_free(virt_to_pfn(pointer), 1UL << order); /* Create free chunk */ freed_ch = (chunk_head_t *)pointer; - freed_ct = (chunk_tail_t *)((char *)pointer + (1<<(order + PAGE_SHIFT)))-1; + freed_ct = (chunk_tail_t *)((char *)pointer + (1UL<<(order + PAGE_SHIFT)))-1; /* Now, possibly we can conseal chunks together */ while(order < FREELIST_SIZE) { - mask = 1 << (order + PAGE_SHIFT); + mask = 1UL << (order + PAGE_SHIFT); if((unsigned long)freed_ch & mask) { to_merge_ch = (chunk_head_t *)((char *)freed_ch - mask); _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |