[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen-unstable] introduce and use nr_cpu_ids and nr_cpumask_bits
# HG changeset patch # User Jan Beulich <jbeulich@xxxxxxxx> # Date 1319181462 -7200 # Node ID 511d5e65a30231fab56b47ffe7b2f1483b023700 # Parent 6c583d35d76dda2236c81d9437ff9d57ab02c006 introduce and use nr_cpu_ids and nr_cpumask_bits The former is the runtime equivalent of NR_CPUS (and users of NR_CPUS, where necessary, get adjusted accordingly), while the latter is for the sole use of determining the allocation size when dynamically allocating CPU masks (done later in this series). Adjust accessors to use either of the two to bound their bitmap operations - which one gets used depends on whether accessing the bits in the gap between nr_cpu_ids and nr_cpumask_bits is benign but more efficient. Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx> Acked-by: Keir Fraser <keir@xxxxxxx> --- diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/acpi/cpu_idle.c --- a/xen/arch/x86/acpi/cpu_idle.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/acpi/cpu_idle.c Fri Oct 21 09:17:42 2011 +0200 @@ -900,7 +900,7 @@ if ( apic_id == BAD_APICID ) return -1; - for ( i = 0; i < NR_CPUS; i++ ) + for ( i = 0; i < nr_cpu_ids; i++ ) { if ( apic_id == x86_cpu_to_apicid[i] ) return i; diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/acpi/cpufreq/cpufreq.c --- a/xen/arch/x86/acpi/cpufreq/cpufreq.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/acpi/cpufreq/cpufreq.c Fri Oct 21 09:17:42 2011 +0200 @@ -209,7 +209,7 @@ if (!cpumask_test_cpu(cpu, mask)) cpu = cpumask_first(mask); - if (cpu >= NR_CPUS || !cpu_online(cpu)) + if (cpu >= nr_cpu_ids || !cpu_online(cpu)) return 0; policy = per_cpu(cpufreq_cpu_policy, cpu); diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/acpi/lib.c --- a/xen/arch/x86/acpi/lib.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/acpi/lib.c Fri Oct 21 09:17:42 2011 +0200 @@ -95,7 +95,7 @@ if (!(acpi_id + 1)) c = &boot_cpu_data; - else if (cpu >= NR_CPUS || !cpu_online(cpu)) + else if (cpu >= nr_cpu_ids || !cpu_online(cpu)) return -EINVAL; else c = cpu_data + cpu; diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/cpu/common.c --- a/xen/arch/x86/cpu/common.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/cpu/common.c Fri Oct 21 09:17:42 2011 +0200 @@ -532,7 +532,7 @@ printk(KERN_INFO "CPU: Hyper-Threading is disabled\n"); } else if (c->x86_num_siblings > 1 ) { - if (c->x86_num_siblings > NR_CPUS) { + if (c->x86_num_siblings > nr_cpu_ids) { printk(KERN_WARNING "CPU: Unsupported number of the siblings %d", c->x86_num_siblings); c->x86_num_siblings = 1; return; diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/cpu/mcheck/mce.c --- a/xen/arch/x86/cpu/mcheck/mce.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/cpu/mcheck/mce.c Fri Oct 21 09:17:42 2011 +0200 @@ -1487,7 +1487,7 @@ mc_msrinject = &op->u.mc_msrinject; target = mc_msrinject->mcinj_cpunr; - if (target >= NR_CPUS) + if (target >= nr_cpu_ids) return x86_mcerr("do_mca inject: bad target", -EINVAL); if (!cpu_online(target)) @@ -1514,7 +1514,7 @@ mc_mceinject = &op->u.mc_mceinject; target = mc_mceinject->mceinj_cpunr; - if (target >= NR_CPUS) + if (target >= nr_cpu_ids) return x86_mcerr("do_mca #MC: bad target", -EINVAL); if (!cpu_online(target)) diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/microcode.c --- a/xen/arch/x86/microcode.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/microcode.c Fri Oct 21 09:17:42 2011 +0200 @@ -126,7 +126,7 @@ info->error = error; info->cpu = next_cpu(info->cpu, cpu_online_map); - if ( info->cpu < NR_CPUS ) + if ( info->cpu < nr_cpu_ids ) return continue_hypercall_on_cpu(info->cpu, do_microcode_update, info); error = info->error; diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/mpparse.c --- a/xen/arch/x86/mpparse.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/mpparse.c Fri Oct 21 09:17:42 2011 +0200 @@ -28,6 +28,7 @@ #include <asm/mtrr.h> #include <asm/mpspec.h> #include <asm/io_apic.h> +#include <asm/setup.h> #include <mach_apic.h> #include <mach_mpparse.h> @@ -61,10 +62,31 @@ /* Internal processor count */ static unsigned int __devinitdata num_processors; +static unsigned int __initdata disabled_cpus; /* Bitmask of physically existing CPUs */ physid_mask_t phys_cpu_present_map; +void __init set_nr_cpu_ids(unsigned int max_cpus) +{ + if (!max_cpus) + max_cpus = num_processors + disabled_cpus; + if (max_cpus > NR_CPUS) + max_cpus = NR_CPUS; + else if (!max_cpus) + max_cpus = 1; + printk(XENLOG_INFO "SMP: Allowing %u CPUs (%d hotplug CPUs)\n", + max_cpus, max_t(int, max_cpus - num_processors, 0)); + nr_cpu_ids = max_cpus; + +#ifndef nr_cpumask_bits + nr_cpumask_bits = (max_cpus + (BITS_PER_LONG - 1)) & + ~(BITS_PER_LONG - 1); + printk(XENLOG_DEBUG "NR_CPUS:%u nr_cpumask_bits:%u\n", + NR_CPUS, nr_cpumask_bits); +#endif +} + /* * Intel MP BIOS table parsing routines: */ @@ -90,8 +112,11 @@ { int ver, apicid, cpu = 0; - if (!(m->mpc_cpuflag & CPU_ENABLED)) + if (!(m->mpc_cpuflag & CPU_ENABLED)) { + if (!hotplug) + ++disabled_cpus; return -EINVAL; + } apicid = mpc_apic_id(m, apicidx); @@ -115,9 +140,9 @@ set_apicid(apicid, &phys_cpu_present_map); - if (num_processors >= NR_CPUS) { - printk(KERN_WARNING "WARNING: NR_CPUS limit of %i reached." - " Processor ignored.\n", NR_CPUS); + if (num_processors >= nr_cpu_ids) { + printk(KERN_WARNING "WARNING: NR_CPUS limit of %u reached." + " Processor ignored.\n", nr_cpu_ids); return -ENOSPC; } diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/numa.c --- a/xen/arch/x86/numa.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/numa.c Fri Oct 21 09:17:42 2011 +0200 @@ -198,7 +198,7 @@ CPUs, as the number of CPUs is not known yet. We round robin the existing nodes. */ rr = first_node(node_online_map); - for (i = 0; i < NR_CPUS; i++) { + for (i = 0; i < nr_cpu_ids; i++) { if (cpu_to_node[i] != NUMA_NO_NODE) continue; numa_set_node(i, rr); @@ -280,7 +280,7 @@ memnodemap = _memnodemap; nodes_clear(node_online_map); node_set_online(0); - for (i = 0; i < NR_CPUS; i++) + for (i = 0; i < nr_cpu_ids; i++) numa_set_node(i, 0); node_to_cpumask[0] = cpumask_of_cpu(0); setup_node_bootmem(0, (u64)start_pfn << PAGE_SHIFT, (u64)end_pfn << PAGE_SHIFT); @@ -335,7 +335,7 @@ void __init init_cpu_to_node(void) { int i, node; - for (i = 0; i < NR_CPUS; i++) { + for (i = 0; i < nr_cpu_ids; i++) { u32 apicid = x86_cpu_to_apicid[i]; if (apicid == BAD_APICID) continue; diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/oprofile/nmi_int.c --- a/xen/arch/x86/oprofile/nmi_int.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/oprofile/nmi_int.c Fri Oct 21 09:17:42 2011 +0200 @@ -127,7 +127,7 @@ static void free_msrs(void) { int i; - for (i = 0; i < NR_CPUS; ++i) { + for (i = 0; i < nr_cpu_ids; ++i) { xfree(cpu_msrs[i].counters); cpu_msrs[i].counters = NULL; xfree(cpu_msrs[i].controls); diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/platform_hypercall.c --- a/xen/arch/x86/platform_hypercall.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/platform_hypercall.c Fri Oct 21 09:17:42 2011 +0200 @@ -442,7 +442,7 @@ break; } - if ( (g_info->xen_cpuid >= NR_CPUS) || + if ( (g_info->xen_cpuid >= nr_cpu_ids) || !cpu_present(g_info->xen_cpuid) ) { g_info->flags |= XEN_PCPU_FLAGS_INVALID; diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/setup.c --- a/xen/arch/x86/setup.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/setup.c Fri Oct 21 09:17:42 2011 +0200 @@ -51,7 +51,7 @@ boolean_param("nosmp", opt_nosmp); /* maxcpus: maximum number of CPUs to activate. */ -static unsigned int __initdata max_cpus = NR_CPUS; +static unsigned int __initdata max_cpus; integer_param("maxcpus", max_cpus); /* opt_watchdog: If true, run a watchdog NMI on each processor. */ @@ -230,7 +230,7 @@ * Among identical longest-prefix matches, pick the smallest APIC ID. */ for ( j = next_cpu(i, cpu_present_map); - j < NR_CPUS; + j < nr_cpu_ids; j = next_cpu(j, cpu_present_map) ) { diff = x86_cpu_to_apicid[j] ^ apicid; @@ -246,9 +246,9 @@ } /* If no match then there must be no CPUs remaining to consider. */ - if ( min_cpu >= NR_CPUS ) + if ( min_cpu >= nr_cpu_ids ) { - BUG_ON(next_cpu(i, cpu_present_map) < NR_CPUS); + BUG_ON(next_cpu(i, cpu_present_map) < nr_cpu_ids); break; } @@ -1203,6 +1203,17 @@ if ( smp_found_config ) get_smp_config(); + if ( opt_nosmp ) + { + max_cpus = 0; + set_nr_cpu_ids(1); + } + else + { + set_nr_cpu_ids(max_cpus); + max_cpus = nr_cpu_ids; + } + #ifdef CONFIG_X86_64 /* Low mappings were only needed for some BIOS table parsing. */ zap_low_mappings(); @@ -1254,9 +1265,6 @@ acpi_mmcfg_init(); #endif - if ( opt_nosmp ) - max_cpus = 0; - iommu_setup(); /* setup iommu if available */ smp_prepare_cpus(max_cpus); diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/smpboot.c --- a/xen/arch/x86/smpboot.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/smpboot.c Fri Oct 21 09:17:42 2011 +0200 @@ -535,7 +535,7 @@ int cpu; cpus_complement(tmp_map, cpu_present_map); cpu = first_cpu(tmp_map); - return (cpu < NR_CPUS) ? cpu : -ENODEV; + return (cpu < nr_cpu_ids) ? cpu : -ENODEV; } static int do_boot_cpu(int apicid, int cpu) diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/srat.c --- a/xen/arch/x86/srat.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/srat.c Fri Oct 21 09:17:42 2011 +0200 @@ -447,7 +447,7 @@ continue; setup_node_bootmem(i, nodes[i].start, nodes[i].end); } - for (i = 0; i < NR_CPUS; i++) { + for (i = 0; i < nr_cpu_ids; i++) { if (cpu_to_node[i] == NUMA_NO_NODE) continue; if (!node_isset(cpu_to_node[i], nodes_parsed)) diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/sysctl.c --- a/xen/arch/x86/sysctl.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/sysctl.c Fri Oct 21 09:17:42 2011 +0200 @@ -82,7 +82,7 @@ pi->nr_cpus = num_online_cpus(); pi->nr_nodes = num_online_nodes(); pi->max_node_id = MAX_NUMNODES-1; - pi->max_cpu_id = NR_CPUS-1; + pi->max_cpu_id = nr_cpu_ids - 1; pi->total_pages = total_pages; pi->free_pages = avail_domheap_pages(); pi->scrub_pages = 0; diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/tboot.c --- a/xen/arch/x86/tboot.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/tboot.c Fri Oct 21 09:17:42 2011 +0200 @@ -248,7 +248,7 @@ void *p; int i; - for ( i = 0; i < NR_CPUS; i++ ) + for ( i = 0; i < nr_cpu_ids; i++ ) { if ( !stack_base[i] ) continue; diff -r 6c583d35d76d -r 511d5e65a302 xen/arch/x86/traps.c --- a/xen/arch/x86/traps.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/arch/x86/traps.c Fri Oct 21 09:17:42 2011 +0200 @@ -3412,7 +3412,7 @@ { int i; /* Keep secondary tables in sync with IRQ updates. */ - for ( i = 1; i < NR_CPUS; i++ ) + for ( i = 1; i < nr_cpu_ids; i++ ) if ( idt_tables[i] != NULL ) _set_gate(&idt_tables[i][n], 14, dpl, addr); _set_gate(&idt_table[n], 14, dpl, addr); diff -r 6c583d35d76d -r 511d5e65a302 xen/common/cpu.c --- a/xen/common/cpu.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/common/cpu.c Fri Oct 21 09:17:42 2011 +0200 @@ -6,6 +6,12 @@ #include <xen/sched.h> #include <xen/stop_machine.h> +unsigned int __read_mostly nr_cpu_ids = NR_CPUS; +#ifndef nr_cpumask_bits +unsigned int __read_mostly nr_cpumask_bits + = BITS_TO_LONGS(NR_CPUS) * BITS_PER_LONG; +#endif + /* * cpu_bit_bitmap[] is a special, "compressed" data structure that * represents all NR_CPUS bits binary values of 1<<nr. @@ -80,7 +86,7 @@ if ( !cpu_hotplug_begin() ) return -EBUSY; - if ( (cpu >= NR_CPUS) || (cpu == 0) || !cpu_online(cpu) ) + if ( (cpu >= nr_cpu_ids) || (cpu == 0) || !cpu_online(cpu) ) { cpu_hotplug_done(); return -EINVAL; @@ -122,7 +128,7 @@ if ( !cpu_hotplug_begin() ) return -EBUSY; - if ( (cpu >= NR_CPUS) || cpu_online(cpu) || !cpu_present(cpu) ) + if ( (cpu >= nr_cpu_ids) || cpu_online(cpu) || !cpu_present(cpu) ) { cpu_hotplug_done(); return -EINVAL; diff -r 6c583d35d76d -r 511d5e65a302 xen/common/cpupool.c --- a/xen/common/cpupool.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/common/cpupool.c Fri Oct 21 09:17:42 2011 +0200 @@ -489,7 +489,7 @@ if ( cpu == XEN_SYSCTL_CPUPOOL_PAR_ANY ) cpu = first_cpu(cpupool_free_cpus); ret = -EINVAL; - if ( cpu >= NR_CPUS ) + if ( cpu >= nr_cpu_ids ) goto addcpu_out; ret = -EBUSY; if ( !cpu_isset(cpu, cpupool_free_cpus) ) @@ -517,7 +517,7 @@ cpu = op->cpu; if ( cpu == XEN_SYSCTL_CPUPOOL_PAR_ANY ) cpu = last_cpu(c->cpu_valid); - ret = (cpu < NR_CPUS) ? cpupool_unassign_cpu(c, cpu) : -EINVAL; + ret = (cpu < nr_cpu_ids) ? cpupool_unassign_cpu(c, cpu) : -EINVAL; cpupool_put(c); } break; diff -r 6c583d35d76d -r 511d5e65a302 xen/common/domain.c --- a/xen/common/domain.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/common/domain.c Fri Oct 21 09:17:42 2011 +0200 @@ -1023,7 +1023,7 @@ { struct migrate_info *info; - if ( (cpu >= NR_CPUS) || !cpu_online(cpu) ) + if ( (cpu >= nr_cpu_ids) || !cpu_online(cpu) ) return -EINVAL; info = this_cpu(continue_info); diff -r 6c583d35d76d -r 511d5e65a302 xen/common/kexec.c --- a/xen/common/kexec.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/common/kexec.c Fri Oct 21 09:17:42 2011 +0200 @@ -296,7 +296,7 @@ int nr = range->nr; int nr_bytes = 0; - if ( nr < 0 || nr >= NR_CPUS || !cpu_online(nr) ) + if ( nr < 0 || nr >= nr_cpu_ids || !cpu_online(nr) ) return -EINVAL; nr_bytes += sizeof_note("CORE", sizeof(ELF_Prstatus)); diff -r 6c583d35d76d -r 511d5e65a302 xen/common/keyhandler.c --- a/xen/common/keyhandler.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/common/keyhandler.c Fri Oct 21 09:17:42 2011 +0200 @@ -98,7 +98,7 @@ return; cpu = cycle_cpu(cpu, dump_execstate_mask); - if ( cpu < NR_CPUS ) + if ( cpu < nr_cpu_ids ) { smp_send_state_dump(cpu); return; diff -r 6c583d35d76d -r 511d5e65a302 xen/common/sched_credit.c --- a/xen/common/sched_credit.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/common/sched_credit.c Fri Oct 21 09:17:42 2011 +0200 @@ -374,7 +374,7 @@ INIT_LIST_HEAD(&spc->runq); spc->runq_sort_last = prv->runq_sort; - spc->idle_bias = NR_CPUS - 1; + spc->idle_bias = nr_cpu_ids - 1; if ( per_cpu(schedule_data, cpu).sched_priv == NULL ) per_cpu(schedule_data, cpu).sched_priv = spc; diff -r 6c583d35d76d -r 511d5e65a302 xen/common/sched_credit2.c --- a/xen/common/sched_credit2.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/common/sched_credit2.c Fri Oct 21 09:17:42 2011 +0200 @@ -2071,7 +2071,7 @@ INIT_LIST_HEAD(&prv->sdom); /* But un-initialize all runqueues */ - for ( i=0; i<NR_CPUS; i++) + for ( i = 0; i < nr_cpu_ids; i++ ) { prv->runq_map[i] = -1; prv->rqd[i].id = -1; diff -r 6c583d35d76d -r 511d5e65a302 xen/common/sched_sedf.c --- a/xen/common/sched_sedf.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/common/sched_sedf.c Fri Oct 21 09:17:42 2011 +0200 @@ -1201,7 +1201,7 @@ routine. Try to avoid unnecessary runs but: Save approximation: Always switch to scheduler!*/ ASSERT(d->processor >= 0); - ASSERT(d->processor < NR_CPUS); + ASSERT(d->processor < nr_cpu_ids); ASSERT(per_cpu(schedule_data, d->processor).curr); if ( should_switch(per_cpu(schedule_data, d->processor).curr, d, now) ) diff -r 6c583d35d76d -r 511d5e65a302 xen/common/schedule.c --- a/xen/common/schedule.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/common/schedule.c Fri Oct 21 09:17:42 2011 +0200 @@ -1353,7 +1353,7 @@ idle_domain = domain_create(DOMID_IDLE, 0, 0); BUG_ON(idle_domain == NULL); idle_domain->vcpu = idle_vcpu; - idle_domain->max_vcpus = NR_CPUS; + idle_domain->max_vcpus = nr_cpu_ids; if ( alloc_vcpu(idle_domain, 0, 0) == NULL ) BUG(); if ( ops.alloc_pdata && diff -r 6c583d35d76d -r 511d5e65a302 xen/common/sysctl.c --- a/xen/common/sysctl.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/common/sysctl.c Fri Oct 21 09:17:42 2011 +0200 @@ -183,7 +183,7 @@ uint32_t i, nr_cpus; struct xen_sysctl_cpuinfo cpuinfo; - nr_cpus = min_t(uint32_t, op->u.getcpuinfo.max_cpus, NR_CPUS); + nr_cpus = min(op->u.getcpuinfo.max_cpus, nr_cpu_ids); ret = xsm_getcpuinfo(); if ( ret ) diff -r 6c583d35d76d -r 511d5e65a302 xen/drivers/acpi/pmstat.c --- a/xen/drivers/acpi/pmstat.c Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/drivers/acpi/pmstat.c Fri Oct 21 09:17:42 2011 +0200 @@ -53,7 +53,7 @@ int ret = 0; const struct processor_pminfo *pmpt; - if ( !op || (op->cpuid >= NR_CPUS) || !cpu_online(op->cpuid) ) + if ( !op || (op->cpuid >= nr_cpu_ids) || !cpu_online(op->cpuid) ) return -EINVAL; pmpt = processor_pminfo[op->cpuid]; diff -r 6c583d35d76d -r 511d5e65a302 xen/include/asm-x86/setup.h --- a/xen/include/asm-x86/setup.h Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/include/asm-x86/setup.h Fri Oct 21 09:17:42 2011 +0200 @@ -20,6 +20,8 @@ int centaur_init_cpu(void); int transmeta_init_cpu(void); +void set_nr_cpu_ids(unsigned int max_cpus); + void numa_initmem_init(unsigned long start_pfn, unsigned long end_pfn); void arch_init_memory(void); void subarch_init_memory(void); diff -r 6c583d35d76d -r 511d5e65a302 xen/include/xen/cpumask.h --- a/xen/include/xen/cpumask.h Thu Oct 20 15:36:01 2011 +0100 +++ b/xen/include/xen/cpumask.h Fri Oct 21 09:17:42 2011 +0200 @@ -81,26 +81,43 @@ typedef struct cpumask{ DECLARE_BITMAP(bits, NR_CPUS); } cpumask_t; +extern unsigned int nr_cpu_ids; + +#if NR_CPUS > 4 * BITS_PER_LONG && !defined(__ia64__) +/* Assuming NR_CPUS is huge, a runtime limit is more efficient. Also, + * not all bits may be allocated. */ +extern unsigned int nr_cpumask_bits; +#else +# define nr_cpumask_bits (BITS_TO_LONGS(NR_CPUS) * BITS_PER_LONG) +#endif + +/* verify cpu argument to cpumask_* operators */ +static inline unsigned int cpumask_check(unsigned int cpu) +{ + ASSERT(cpu < nr_cpu_ids); + return cpu; +} + #define cpu_set(cpu, dst) cpumask_set_cpu(cpu, &(dst)) static inline void cpumask_set_cpu(int cpu, volatile cpumask_t *dstp) { - set_bit(cpu, dstp->bits); + set_bit(cpumask_check(cpu), dstp->bits); } #define cpu_clear(cpu, dst) cpumask_clear_cpu(cpu, &(dst)) static inline void cpumask_clear_cpu(int cpu, volatile cpumask_t *dstp) { - clear_bit(cpu, dstp->bits); + clear_bit(cpumask_check(cpu), dstp->bits); } -#define cpumask_setall(dst) __cpus_setall(dst, NR_CPUS) +#define cpumask_setall(dst) __cpus_setall(dst, nr_cpumask_bits) #define cpus_setall(dst) __cpus_setall(&(dst), NR_CPUS) static inline void __cpus_setall(cpumask_t *dstp, int nbits) { bitmap_fill(dstp->bits, nbits); } -#define cpumask_clear(dst) __cpus_clear(dst, NR_CPUS) +#define cpumask_clear(dst) __cpus_clear(dst, nr_cpumask_bits) #define cpus_clear(dst) __cpus_clear(&(dst), NR_CPUS) static inline void __cpus_clear(cpumask_t *dstp, int nbits) { @@ -108,24 +125,26 @@ } /* No static inline type checking - see Subtlety (1) above. */ -#define cpumask_test_cpu(cpu, cpumask) test_bit(cpu, (cpumask)->bits) +#define cpumask_test_cpu(cpu, cpumask) \ + test_bit(cpumask_check(cpu), (cpumask)->bits) #define cpu_isset(cpu, cpumask) test_bit((cpu), (cpumask).bits) #define cpu_test_and_set(cpu, cpumask) \ cpumask_test_and_set_cpu(cpu, &(cpumask)) static inline int cpumask_test_and_set_cpu(int cpu, cpumask_t *addr) { - return test_and_set_bit(cpu, addr->bits); + return test_and_set_bit(cpumask_check(cpu), addr->bits); } #define cpu_test_and_clear(cpu, cpumask) \ cpumask_test_and_clear_cpu(cpu, &(cpumask)) static inline int cpumask_test_and_clear_cpu(int cpu, cpumask_t *addr) { - return test_and_clear_bit(cpu, addr->bits); + return test_and_clear_bit(cpumask_check(cpu), addr->bits); } -#define cpumask_and(dst, src1, src2) __cpus_and(dst, src1, src2, NR_CPUS) +#define cpumask_and(dst, src1, src2) \ + __cpus_and(dst, src1, src2, nr_cpumask_bits) #define cpus_and(dst, src1, src2) __cpus_and(&(dst), &(src1), &(src2), NR_CPUS) static inline void __cpus_and(cpumask_t *dstp, const cpumask_t *src1p, const cpumask_t *src2p, int nbits) @@ -133,7 +152,8 @@ bitmap_and(dstp->bits, src1p->bits, src2p->bits, nbits); } -#define cpumask_or(dst, src1, src2) __cpus_or(dst, src1, src2, NR_CPUS) +#define cpumask_or(dst, src1, src2) \ + __cpus_or(dst, src1, src2, nr_cpumask_bits) #define cpus_or(dst, src1, src2) __cpus_or(&(dst), &(src1), &(src2), NR_CPUS) static inline void __cpus_or(cpumask_t *dstp, const cpumask_t *src1p, const cpumask_t *src2p, int nbits) @@ -141,7 +161,8 @@ bitmap_or(dstp->bits, src1p->bits, src2p->bits, nbits); } -#define cpumask_xor(dst, src1, src2) __cpus_xor(dst, src1, src2, NR_CPUS) +#define cpumask_xor(dst, src1, src2) \ + __cpus_xor(dst, src1, src2, nr_cpumask_bits) #define cpus_xor(dst, src1, src2) __cpus_xor(&(dst), &(src1), &(src2), NR_CPUS) static inline void __cpus_xor(cpumask_t *dstp, const cpumask_t *src1p, const cpumask_t *src2p, int nbits) @@ -149,7 +170,8 @@ bitmap_xor(dstp->bits, src1p->bits, src2p->bits, nbits); } -#define cpumask_andnot(dst, src1, src2) __cpus_andnot(dst, src1, src2, NR_CPUS) +#define cpumask_andnot(dst, src1, src2) \ + __cpus_andnot(dst, src1, src2, nr_cpumask_bits) #define cpus_andnot(dst, src1, src2) \ __cpus_andnot(&(dst), &(src1), &(src2), NR_CPUS) static inline void __cpus_andnot(cpumask_t *dstp, const cpumask_t *src1p, @@ -158,7 +180,8 @@ bitmap_andnot(dstp->bits, src1p->bits, src2p->bits, nbits); } -#define cpumask_complement(dst, src) __cpus_complement(dst, src, NR_CPUS) +#define cpumask_complement(dst, src) \ + __cpus_complement(dst, src, nr_cpumask_bits) #define cpus_complement(dst, src) __cpus_complement(&(dst), &(src), NR_CPUS) static inline void __cpus_complement(cpumask_t *dstp, const cpumask_t *srcp, int nbits) @@ -166,55 +189,62 @@ bitmap_complement(dstp->bits, srcp->bits, nbits); } -#define cpumask_equal(src1, src2) __cpus_equal(src1, src2, NR_CPUS) -#define cpus_equal(src1, src2) __cpus_equal(&(src1), &(src2), NR_CPUS) +#define cpumask_equal(src1, src2) __cpus_equal(src1, src2, nr_cpu_ids) +#define cpus_equal(src1, src2) __cpus_equal(&(src1), &(src2), nr_cpu_ids) static inline int __cpus_equal(const cpumask_t *src1p, const cpumask_t *src2p, int nbits) { return bitmap_equal(src1p->bits, src2p->bits, nbits); } -#define cpus_intersects(src1, src2) __cpus_intersects(&(src1), &(src2), NR_CPUS) +#define cpumask_intersects(src1, src2) \ + __cpus_intersects(src1, src2, nr_cpu_ids) +#define cpus_intersects(src1, src2) \ + __cpus_intersects(&(src1), &(src2), nr_cpu_ids) static inline int __cpus_intersects(const cpumask_t *src1p, const cpumask_t *src2p, int nbits) { return bitmap_intersects(src1p->bits, src2p->bits, nbits); } -#define cpus_subset(src1, src2) __cpus_subset(&(src1), &(src2), NR_CPUS) +#define cpumask_subset(src1, src2) __cpus_subset(src1, src2, nr_cpu_ids) +#define cpus_subset(src1, src2) __cpus_subset(&(src1), &(src2), nr_cpu_ids) static inline int __cpus_subset(const cpumask_t *src1p, const cpumask_t *src2p, int nbits) { return bitmap_subset(src1p->bits, src2p->bits, nbits); } -#define cpumask_empty(src) __cpus_empty(src, NR_CPUS) -#define cpus_empty(src) __cpus_empty(&(src), NR_CPUS) +#define cpumask_empty(src) __cpus_empty(src, nr_cpu_ids) +#define cpus_empty(src) __cpus_empty(&(src), nr_cpu_ids) static inline int __cpus_empty(const cpumask_t *srcp, int nbits) { return bitmap_empty(srcp->bits, nbits); } -#define cpumask_full(cpumask) __cpus_full(cpumask, NR_CPUS) -#define cpus_full(cpumask) __cpus_full(&(cpumask), NR_CPUS) +#define cpumask_full(cpumask) __cpus_full(cpumask, nr_cpu_ids) +#define cpus_full(cpumask) __cpus_full(&(cpumask), nr_cpu_ids) static inline int __cpus_full(const cpumask_t *srcp, int nbits) { return bitmap_full(srcp->bits, nbits); } -#define cpumask_weight(cpumask) __cpus_weight(cpumask, NR_CPUS) -#define cpus_weight(cpumask) __cpus_weight(&(cpumask), NR_CPUS) +#define cpumask_weight(cpumask) __cpus_weight(cpumask, nr_cpu_ids) +#define cpus_weight(cpumask) __cpus_weight(&(cpumask), nr_cpu_ids) static inline int __cpus_weight(const cpumask_t *srcp, int nbits) { return bitmap_weight(srcp->bits, nbits); } -#define cpus_copy(dest, src) cpumask_copy(&(dest), &(src)) -static inline void cpumask_copy(cpumask_t *dstp, const cpumask_t *srcp) +#define cpumask_copy(dest, src) __cpus_copy(dest, src, nr_cpumask_bits) +#define cpus_copy(dest, src) __cpus_copy(&(dest), &(src), NR_CPUS) +static inline void __cpus_copy(cpumask_t *dstp, const cpumask_t *srcp, int nbits) { - bitmap_copy(dstp->bits, srcp->bits, NR_CPUS); + bitmap_copy(dstp->bits, srcp->bits, nbits); } +#define cpumask_shift_right(dst, src, n) \ + __cpus_shift_right(dst, src, n, nr_cpumask_bits) #define cpus_shift_right(dst, src, n) \ __cpus_shift_right(&(dst), &(src), (n), NR_CPUS) static inline void __cpus_shift_right(cpumask_t *dstp, @@ -223,6 +253,8 @@ bitmap_shift_right(dstp->bits, srcp->bits, n, nbits); } +#define cpumask_shift_left(dst, src, n) \ + __cpus_shift_left(dst, src, n, nr_cpumask_bits) #define cpus_shift_left(dst, src, n) \ __cpus_shift_left(&(dst), &(src), (n), NR_CPUS) static inline void __cpus_shift_left(cpumask_t *dstp, @@ -231,22 +263,22 @@ bitmap_shift_left(dstp->bits, srcp->bits, n, nbits); } -#define cpumask_first(src) __first_cpu(src, NR_CPUS) -#define first_cpu(src) __first_cpu(&(src), NR_CPUS) +#define cpumask_first(src) __first_cpu(src, nr_cpu_ids) +#define first_cpu(src) __first_cpu(&(src), nr_cpu_ids) static inline int __first_cpu(const cpumask_t *srcp, int nbits) { return min_t(int, nbits, find_first_bit(srcp->bits, nbits)); } -#define cpumask_next(n, src) __next_cpu(n, src, NR_CPUS) -#define next_cpu(n, src) __next_cpu((n), &(src), NR_CPUS) +#define cpumask_next(n, src) __next_cpu(n, src, nr_cpu_ids) +#define next_cpu(n, src) __next_cpu((n), &(src), nr_cpu_ids) static inline int __next_cpu(int n, const cpumask_t *srcp, int nbits) { return min_t(int, nbits, find_next_bit(srcp->bits, nbits, n+1)); } -#define cpumask_last(src) __last_cpu(src, NR_CPUS) -#define last_cpu(src) __last_cpu(&(src), NR_CPUS) +#define cpumask_last(src) __last_cpu(src, nr_cpu_ids) +#define last_cpu(src) __last_cpu(&(src), nr_cpu_ids) static inline int __last_cpu(const cpumask_t *srcp, int nbits) { int cpu, pcpu = nbits; @@ -257,8 +289,8 @@ return pcpu; } -#define cpumask_cycle(n, src) __cycle_cpu(n, src, NR_CPUS) -#define cycle_cpu(n, src) __cycle_cpu((n), &(src), NR_CPUS) +#define cpumask_cycle(n, src) __cycle_cpu(n, src, nr_cpu_ids) +#define cycle_cpu(n, src) __cycle_cpu((n), &(src), nr_cpu_ids) static inline int __cycle_cpu(int n, const cpumask_t *srcp, int nbits) { int nxt = __next_cpu(n, srcp, nbits); @@ -317,7 +349,7 @@ #define cpus_addr(src) ((src).bits) #define cpumask_scnprintf(buf, len, src) \ - __cpumask_scnprintf((buf), (len), &(src), NR_CPUS) + __cpumask_scnprintf((buf), (len), &(src), nr_cpu_ids) static inline int __cpumask_scnprintf(char *buf, int len, const cpumask_t *srcp, int nbits) { @@ -325,7 +357,7 @@ } #define cpulist_scnprintf(buf, len, src) \ - __cpulist_scnprintf((buf), (len), &(src), NR_CPUS) + __cpulist_scnprintf((buf), (len), &(src), nr_cpu_ids) static inline int __cpulist_scnprintf(char *buf, int len, const cpumask_t *srcp, int nbits) { @@ -355,6 +387,11 @@ static inline bool_t alloc_cpumask_var(cpumask_var_t *mask) { + /* + * Once all direct cpumask assignments and all cpus_*() accessors + * still referencing NR_CPUS are gone, we could use nr_cpumask_bits + * to determine the allocation size here. + */ return (*mask = xmalloc(cpumask_t)) != NULL; } @@ -386,7 +423,7 @@ #if NR_CPUS > 1 #define for_each_cpu_mask(cpu, mask) \ for ((cpu) = first_cpu(mask); \ - (cpu) < NR_CPUS; \ + (cpu) < nr_cpu_ids; \ (cpu) = next_cpu((cpu), (mask))) #else /* NR_CPUS == 1 */ #define for_each_cpu_mask(cpu, mask) for ((cpu) = 0; (cpu) < 1; (cpu)++) _______________________________________________ Xen-changelog mailing list Xen-changelog@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-changelog
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |