This includes the conversion from for_each_cpu_mask() to for_each-cpu(). Signed-off-by: Jan Beulich <jbeulich@suse.com> --- a/xen/arch/ia64/linux-xen/iosapic.c +++ b/xen/arch/ia64/linux-xen/iosapic.c @@ -704,7 +704,7 @@ get_target_cpu (unsigned int gsi, int ve cpu_mask = node_to_cpumask(iosapic_lists[iosapic_index].node); - for_each_cpu_mask(numa_cpu, cpu_mask) { + for_each_cpu(numa_cpu, &cpu_mask) { if (!cpu_online(numa_cpu)) cpumask_clear_cpu(numa_cpu, &cpu_mask); } @@ -717,8 +717,8 @@ get_target_cpu (unsigned int gsi, int ve /* Use vector assigment to distribute across cpus in node */ cpu_index = vector % num_cpus; - for (numa_cpu = first_cpu(cpu_mask) ; i < cpu_index ; i++) - numa_cpu = next_cpu(numa_cpu, cpu_mask); + for (numa_cpu = cpumask_first(&cpu_mask) ; i < cpu_index ; i++) + numa_cpu = cpumask_next(numa_cpu, &cpu_mask); if (numa_cpu != NR_CPUS) return cpu_physical_id(numa_cpu); --- a/xen/arch/ia64/linux-xen/mca.c +++ b/xen/arch/ia64/linux-xen/mca.c @@ -1415,7 +1415,7 @@ ia64_mca_cmc_poll (void *dummy) #endif { /* Trigger a CMC interrupt cascade */ - platform_send_ipi(first_cpu(cpu_online_map), IA64_CMCP_VECTOR, IA64_IPI_DM_INT, 0); + platform_send_ipi(cpumask_first(&cpu_online_map), IA64_CMCP_VECTOR, IA64_IPI_DM_INT, 0); } /* @@ -1505,7 +1505,7 @@ ia64_mca_cpe_poll (void *dummy) #endif { /* Trigger a CPE interrupt cascade */ - platform_send_ipi(first_cpu(cpu_online_map), IA64_CPEP_VECTOR, IA64_IPI_DM_INT, 0); + platform_send_ipi(cpumask_first(&cpu_online_map), IA64_CPEP_VECTOR, IA64_IPI_DM_INT, 0); } #endif /* CONFIG_ACPI */ --- a/xen/arch/ia64/linux-xen/smp.c +++ b/xen/arch/ia64/linux-xen/smp.c @@ -462,7 +462,7 @@ on_selected_cpus(const cpumask_t *select call_data = &data; wmb(); - for_each_cpu_mask(cpu, *selected) + for_each_cpu(cpu, selected) send_IPI_single(cpu, IPI_CALL_FUNC); while (atomic_read(wait ? &data.finished : &data.started) != nr_cpus) --- a/xen/arch/ia64/linux-xen/smpboot.c +++ b/xen/arch/ia64/linux-xen/smpboot.c @@ -687,9 +687,9 @@ clear_cpu_sibling_map(int cpu) { int i; - for_each_cpu_mask(i, *per_cpu(cpu_sibling_mask, cpu)) + for_each_cpu(i, per_cpu(cpu_sibling_mask, cpu)) cpumask_clear_cpu(cpu, per_cpu(cpu_sibling_mask, i)); - for_each_cpu_mask(i, *per_cpu(cpu_core_mask, cpu)) + for_each_cpu(i, per_cpu(cpu_core_mask, cpu)) cpumask_clear_cpu(cpu, per_cpu(cpu_core_mask, i)); cpumask_clear(per_cpu(cpu_sibling_mask, cpu)); --- a/xen/arch/ia64/vmx/vacpi.c +++ b/xen/arch/ia64/vmx/vacpi.c @@ -191,7 +191,7 @@ void vacpi_init(struct domain *d) s->last_gtime = NOW(); /* Set up callback to fire SCIs when the MSB of TMR_VAL changes */ - init_timer(&s->timer, pmt_timer_callback, d, first_cpu(cpu_online_map)); + init_timer(&s->timer, pmt_timer_callback, d, cpumask_first(&cpu_online_map)); pmt_timer_callback(d); } --- a/xen/arch/ia64/xen/dom0_ops.c +++ b/xen/arch/ia64/xen/dom0_ops.c @@ -618,7 +618,7 @@ long arch_do_sysctl(xen_sysctl_t *op, XE XEN_GUEST_HANDLE_64(uint32) arr; uint32_t i, val, max_array_ent = ti->max_cpu_index; - ti->max_cpu_index = last_cpu(cpu_online_map); + ti->max_cpu_index = cpumask_last(&cpu_online_map); max_array_ent = min(max_array_ent, ti->max_cpu_index); arr = ti->cpu_to_core; --- a/xen/arch/ia64/xen/domain.c +++ b/xen/arch/ia64/xen/domain.c @@ -501,7 +501,7 @@ int vcpu_initialise(struct vcpu *v) if (!VMX_DOMAIN(v)) init_timer(&v->arch.hlt_timer, hlt_timer_fn, v, - first_cpu(cpu_online_map)); + cpumask_any(&cpu_online_map)); return 0; } --- a/xen/arch/ia64/xen/vhpt.c +++ b/xen/arch/ia64/xen/vhpt.c @@ -463,7 +463,7 @@ __domain_flush_vtlb_track_entry(struct d local_purge = 0; } } else { - for_each_cpu_mask(cpu, entry->pcpu_dirty_mask) { + for_each_cpu(cpu, &entry->pcpu_dirty_mask) { /* Invalidate VHPT entries. */ cpu_flush_vhpt_range(cpu, vaddr, 1L << ps); @@ -559,7 +559,7 @@ void flush_tlb_mask(const cpumask_t *mas if (cpumask_subset(mask, cpumask_of(cpu))) return; - for_each_cpu_mask (cpu, *mask) + for_each_cpu (cpu, mask) if (cpu != smp_processor_id()) smp_call_function_single (cpu, (void (*)(void *))flush_tlb_vhpt_all, NULL, 1); --- a/xen/arch/x86/acpi/cpu_idle.c +++ b/xen/arch/x86/acpi/cpu_idle.c @@ -251,7 +251,7 @@ void cpuidle_wakeup_mwait(cpumask_t *mas cpumask_and(&target, mask, &cpuidle_mwait_flags); /* CPU is MWAITing on the cpuidle_mwait_wakeup flag. */ - for_each_cpu_mask(cpu, target) + for_each_cpu(cpu, &target) mwait_wakeup(cpu) = 0; cpumask_andnot(mask, mask, &target); --- a/xen/arch/x86/acpi/cpufreq/cpufreq.c +++ b/xen/arch/x86/acpi/cpufreq/cpufreq.c @@ -487,7 +487,7 @@ static int acpi_cpufreq_target(struct cp return -EAGAIN; } - for_each_cpu_mask(j, online_policy_cpus) + for_each_cpu(j, &online_policy_cpus) cpufreq_statistic_update(j, perf->state, next_perf_state); perf->state = next_perf_state; --- a/xen/arch/x86/acpi/cpufreq/powernow.c +++ b/xen/arch/x86/acpi/cpufreq/powernow.c @@ -130,7 +130,7 @@ static int powernow_cpufreq_target(struc on_selected_cpus(cmd.mask, transition_pstate, &cmd, 1); - for_each_cpu_mask(j, online_policy_cpus) + for_each_cpu(j, &online_policy_cpus) cpufreq_statistic_update(j, perf->state, next_perf_state); perf->state = next_perf_state; --- a/xen/arch/x86/genapic/x2apic.c +++ b/xen/arch/x86/genapic/x2apic.c @@ -72,7 +72,7 @@ static void __send_IPI_mask_x2apic( local_irq_save(flags); - for_each_cpu_mask ( cpu, *cpumask ) + for_each_cpu ( cpu, cpumask ) { if ( !cpu_online(cpu) || (cpu == smp_processor_id()) ) continue; --- a/xen/arch/x86/hpet.c +++ b/xen/arch/x86/hpet.c @@ -182,7 +182,7 @@ again: now = NOW(); /* find all expired events */ - for_each_cpu_mask(cpu, *ch->cpumask) + for_each_cpu(cpu, ch->cpumask) { s_time_t deadline; --- a/xen/arch/x86/irq.c +++ b/xen/arch/x86/irq.c @@ -125,7 +125,7 @@ static int __init __bind_irq_vector(int if ( desc->arch.vector != IRQ_VECTOR_UNASSIGNED ) return -EBUSY; trace_irq_mask(TRC_HW_IRQ_BIND_VECTOR, irq, vector, &online_mask); - for_each_cpu_mask(cpu, online_mask) + for_each_cpu(cpu, &online_mask) per_cpu(vector_irq, cpu)[vector] = irq; desc->arch.vector = vector; cpumask_copy(desc->arch.cpu_mask, &online_mask); @@ -223,7 +223,7 @@ static void __clear_irq_vector(int irq) vector = desc->arch.vector; cpumask_and(&tmp_mask, desc->arch.cpu_mask, &cpu_online_map); - for_each_cpu_mask(cpu, tmp_mask) { + for_each_cpu(cpu, &tmp_mask) { ASSERT( per_cpu(vector_irq, cpu)[vector] == irq ); per_cpu(vector_irq, cpu)[vector] = -1; } @@ -248,7 +248,7 @@ static void __clear_irq_vector(int irq) old_vector = desc->arch.old_vector; cpumask_and(&tmp_mask, desc->arch.old_cpu_mask, &cpu_online_map); - for_each_cpu_mask(cpu, tmp_mask) { + for_each_cpu(cpu, &tmp_mask) { ASSERT( per_cpu(vector_irq, cpu)[old_vector] == irq ); TRACE_3D(TRC_HW_IRQ_MOVE_FINISH, irq, old_vector, cpu); per_cpu(vector_irq, cpu)[old_vector] = -1; @@ -455,7 +455,7 @@ static int __assign_irq_vector( else irq_used_vectors = irq_get_used_vector_mask(irq); - for_each_cpu_mask(cpu, *mask) { + for_each_cpu(cpu, mask) { int new_cpu; int vector, offset; @@ -485,7 +485,7 @@ next: && test_bit(vector, irq_used_vectors) ) goto next; - for_each_cpu_mask(new_cpu, tmp_mask) + for_each_cpu(new_cpu, &tmp_mask) if (per_cpu(vector_irq, new_cpu)[vector] != -1) goto next; /* Found one! */ @@ -497,7 +497,7 @@ next: desc->arch.old_vector = desc->arch.vector; } trace_irq_mask(TRC_HW_IRQ_ASSIGN_VECTOR, irq, vector, &tmp_mask); - for_each_cpu_mask(new_cpu, tmp_mask) + for_each_cpu(new_cpu, &tmp_mask) per_cpu(vector_irq, new_cpu)[vector] = irq; desc->arch.vector = vector; cpumask_copy(desc->arch.cpu_mask, &tmp_mask); --- a/xen/arch/x86/microcode.c +++ b/xen/arch/x86/microcode.c @@ -125,7 +125,7 @@ static long do_microcode_update(void *_i if ( error ) info->error = error; - info->cpu = next_cpu(info->cpu, cpu_online_map); + info->cpu = cpumask_next(info->cpu, &cpu_online_map); if ( info->cpu < nr_cpu_ids ) return continue_hypercall_on_cpu(info->cpu, do_microcode_update, info); @@ -158,7 +158,7 @@ int microcode_update(XEN_GUEST_HANDLE(co info->buffer_size = len; info->error = 0; - info->cpu = first_cpu(cpu_online_map); + info->cpu = cpumask_first(&cpu_online_map); return continue_hypercall_on_cpu(info->cpu, do_microcode_update, info); } --- a/xen/arch/x86/platform_hypercall.c +++ b/xen/arch/x86/platform_hypercall.c @@ -366,7 +366,7 @@ ret_t do_platform_op(XEN_GUEST_HANDLE(xe goto out; guest_from_compat_handle(idletimes, op->u.getidletime.idletime); - for_each_cpu_mask ( cpu, *cpumap ) + for_each_cpu ( cpu, cpumap ) { if ( idle_vcpu[cpu] == NULL ) cpumask_clear_cpu(cpu, cpumap); @@ -460,7 +460,7 @@ ret_t do_platform_op(XEN_GUEST_HANDLE(xe g_info->flags |= XEN_PCPU_FLAGS_ONLINE; } - g_info->max_present = last_cpu(cpu_present_map); + g_info->max_present = cpumask_last(&cpu_present_map); put_cpu_maps(); --- a/xen/arch/x86/setup.c +++ b/xen/arch/x86/setup.c @@ -229,9 +229,9 @@ static void __init normalise_cpu_order(v * Find remaining CPU with longest-prefix match on APIC ID. * Among identical longest-prefix matches, pick the smallest APIC ID. */ - for ( j = next_cpu(i, cpu_present_map); + for ( j = cpumask_next(i, &cpu_present_map); j < nr_cpu_ids; - j = next_cpu(j, cpu_present_map) ) + j = cpumask_next(j, &cpu_present_map) ) { diff = x86_cpu_to_apicid[j] ^ apicid; while ( diff & (diff-1) ) @@ -248,12 +248,12 @@ static void __init normalise_cpu_order(v /* If no match then there must be no CPUs remaining to consider. */ if ( min_cpu >= nr_cpu_ids ) { - BUG_ON(next_cpu(i, cpu_present_map) < nr_cpu_ids); + BUG_ON(cpumask_next(i, &cpu_present_map) < nr_cpu_ids); break; } /* Switch the best-matching CPU with the next CPU in logical order. */ - j = next_cpu(i, cpu_present_map); + j = cpumask_next(i, &cpu_present_map); apicid = x86_cpu_to_apicid[min_cpu]; x86_cpu_to_apicid[min_cpu] = x86_cpu_to_apicid[j]; x86_cpu_to_apicid[j] = apicid; --- a/xen/arch/x86/smp.c +++ b/xen/arch/x86/smp.c @@ -182,7 +182,7 @@ void send_IPI_mask_phys(const cpumask_t local_irq_save(flags); - for_each_cpu_mask ( query_cpu, *mask ) + for_each_cpu ( query_cpu, mask ) { if ( !cpu_online(query_cpu) || (query_cpu == smp_processor_id()) ) continue; --- a/xen/arch/x86/smpboot.c +++ b/xen/arch/x86/smpboot.c @@ -248,7 +248,7 @@ static void set_cpu_sibling_map(int cpu) if ( c[cpu].x86_num_siblings > 1 ) { - for_each_cpu_mask ( i, cpu_sibling_setup_map ) + for_each_cpu ( i, &cpu_sibling_setup_map ) { if ( cpu_has(c, X86_FEATURE_TOPOEXT) ) { if ( (c[cpu].phys_proc_id == c[i].phys_proc_id) && @@ -273,7 +273,7 @@ static void set_cpu_sibling_map(int cpu) return; } - for_each_cpu_mask ( i, cpu_sibling_setup_map ) + for_each_cpu ( i, &cpu_sibling_setup_map ) { if ( c[cpu].phys_proc_id == c[i].phys_proc_id ) { @@ -814,7 +814,7 @@ remove_siblinginfo(int cpu) int sibling; struct cpuinfo_x86 *c = cpu_data; - for_each_cpu_mask ( sibling, *per_cpu(cpu_core_mask, cpu) ) + for_each_cpu ( sibling, per_cpu(cpu_core_mask, cpu) ) { cpumask_clear_cpu(cpu, per_cpu(cpu_core_mask, sibling)); /* Last thread sibling in this cpu core going down. */ @@ -822,7 +822,7 @@ remove_siblinginfo(int cpu) c[sibling].booted_cores--; } - for_each_cpu_mask(sibling, *per_cpu(cpu_sibling_mask, cpu)) + for_each_cpu(sibling, per_cpu(cpu_sibling_mask, cpu)) cpumask_clear_cpu(cpu, per_cpu(cpu_sibling_mask, sibling)); cpumask_clear(per_cpu(cpu_sibling_mask, cpu)); cpumask_clear(per_cpu(cpu_core_mask, cpu)); --- a/xen/arch/x86/sysctl.c +++ b/xen/arch/x86/sysctl.c @@ -103,7 +103,7 @@ long arch_do_sysctl( uint32_t i, max_cpu_index, last_online_cpu; xen_sysctl_topologyinfo_t *ti = &sysctl->u.topologyinfo; - last_online_cpu = last_cpu(cpu_online_map); + last_online_cpu = cpumask_last(&cpu_online_map); max_cpu_index = min_t(uint32_t, ti->max_cpu_index, last_online_cpu); ti->max_cpu_index = last_online_cpu; --- a/xen/common/cpu.c +++ b/xen/common/cpu.c @@ -205,7 +205,7 @@ void enable_nonboot_cpus(void) printk("Enabling non-boot CPUs ...\n"); - for_each_cpu_mask ( cpu, frozen_cpus ) + for_each_cpu ( cpu, &frozen_cpus ) { if ( (error = cpu_up(cpu)) ) { --- a/xen/common/cpupool.c +++ b/xen/common/cpupool.c @@ -494,7 +494,7 @@ int cpupool_do_sysctl(struct xen_sysctl_ op->cpupool_id, cpu); spin_lock(&cpupool_lock); if ( cpu == XEN_SYSCTL_CPUPOOL_PAR_ANY ) - cpu = first_cpu(cpupool_free_cpus); + cpu = cpumask_first(&cpupool_free_cpus); ret = -EINVAL; if ( cpu >= nr_cpu_ids ) goto addcpu_out; --- a/xen/common/domctl.c +++ b/xen/common/domctl.c @@ -190,7 +190,7 @@ static unsigned int default_vcpu0_locati cpu = cpumask_first(&cpu_exclude_map); if ( cpumask_weight(&cpu_exclude_map) > 1 ) cpu = cpumask_next(cpu, &cpu_exclude_map); - for_each_cpu_mask(i, *online) + for_each_cpu(i, online) { if ( cpumask_test_cpu(i, &cpu_exclude_map) ) continue; @@ -541,7 +541,7 @@ long do_domctl(XEN_GUEST_HANDLE(xen_domc cpu = (i == 0) ? default_vcpu0_location(online) : - cycle_cpu(d->vcpu[i-1]->processor, *online); + cpumask_cycle(d->vcpu[i-1]->processor, online); if ( alloc_vcpu(d, i, cpu) == NULL ) goto maxvcpu_out; --- a/xen/common/keyhandler.c +++ b/xen/common/keyhandler.c @@ -128,7 +128,7 @@ static void dump_registers(unsigned char return; /* Normal handling: synchronously dump the remaining CPUs'' states. */ - for_each_cpu_mask ( cpu, dump_execstate_mask ) + for_each_cpu ( cpu, &dump_execstate_mask ) { smp_send_state_dump(cpu); while ( cpumask_test_cpu(cpu, &dump_execstate_mask) ) --- a/xen/common/perfc.c +++ b/xen/common/perfc.c @@ -211,14 +211,14 @@ static int perfc_copy_info(XEN_GUEST_HAN { case TYPE_SINGLE: case TYPE_S_SINGLE: - for_each_cpu_mask ( cpu, perfc_cpumap ) + for_each_cpu ( cpu, &perfc_cpumap ) perfc_vals[v++] = per_cpu(perfcounters, cpu)[j]; ++j; break; case TYPE_ARRAY: case TYPE_S_ARRAY: memset(perfc_vals + v, 0, perfc_d[i].nr_vals * sizeof(*perfc_vals)); - for_each_cpu_mask ( cpu, perfc_cpumap ) + for_each_cpu ( cpu, &perfc_cpumap ) { perfc_t *counters = per_cpu(perfcounters, cpu) + j; unsigned int k; --- a/xen/common/sched_credit2.c +++ b/xen/common/sched_credit2.c @@ -521,7 +521,7 @@ runq_tickle(const struct scheduler *ops, cpumask_andnot(&mask, &rqd->active, &rqd->idle); cpumask_andnot(&mask, &mask, &rqd->tickled); - for_each_cpu_mask(i, mask) + for_each_cpu(i, &mask) { struct csched_vcpu * cur; @@ -1051,7 +1051,7 @@ choose_cpu(const struct scheduler *ops, else { d2printk("d%dv%d +\n", svc->vcpu->domain->domain_id, svc->vcpu->vcpu_id); - new_cpu = first_cpu(svc->migrate_rqd->active); + new_cpu = cpumask_first(&svc->migrate_rqd->active); goto out_up; } } @@ -1061,7 +1061,7 @@ choose_cpu(const struct scheduler *ops, min_avgload = MAX_LOAD; /* Find the runqueue with the lowest instantaneous load */ - for_each_cpu_mask(i, prv->active_queues) + for_each_cpu(i, &prv->active_queues) { struct csched_runqueue_data *rqd; s_time_t rqd_avgload; @@ -1099,7 +1099,7 @@ choose_cpu(const struct scheduler *ops, else { BUG_ON(cpumask_empty(&prv->rqd[min_rqi].active)); - new_cpu = first_cpu(prv->rqd[min_rqi].active); + new_cpu = cpumask_first(&prv->rqd[min_rqi].active); } out_up: @@ -1179,7 +1179,7 @@ void migrate(const struct scheduler *ops on_runq=1; } __runq_deassign(svc); - svc->vcpu->processor = first_cpu(trqd->active); + svc->vcpu->processor = cpumask_first(&trqd->active); __runq_assign(svc, trqd); if ( on_runq ) { @@ -1219,7 +1219,7 @@ retry: st.load_delta = 0; - for_each_cpu_mask(i, prv->active_queues) + for_each_cpu(i, &prv->active_queues) { s_time_t delta; @@ -1618,7 +1618,7 @@ csched_schedule( { int rq; other_rqi = -2; - for_each_cpu_mask ( rq, CSCHED_PRIV(ops)->active_queues ) + for_each_cpu ( rq, &CSCHED_PRIV(ops)->active_queues ) { if ( scurr->rqd == &CSCHED_PRIV(ops)->rqd[rq] ) { @@ -1803,7 +1803,7 @@ csched_dump(const struct scheduler *ops) "\tdefault-weight = %d\n", cpumask_weight(&prv->active_queues), CSCHED_DEFAULT_WEIGHT); - for_each_cpu_mask(i, prv->active_queues) + for_each_cpu(i, &prv->active_queues) { s_time_t fraction; --- a/xen/common/sched_sedf.c +++ b/xen/common/sched_sedf.c @@ -442,7 +442,7 @@ static int sedf_pick_cpu(const struct sc online = SEDF_CPUONLINE(v->domain->cpupool); cpumask_and(&online_affinity, v->cpu_affinity, online); - return first_cpu(online_affinity); + return cpumask_first(&online_affinity); } /* @@ -1322,7 +1322,7 @@ static int sedf_adjust_weights(struct cp { struct vcpu *p; struct domain *d; - unsigned int cpu, nr_cpus = last_cpu(cpu_online_map) + 1; + unsigned int cpu, nr_cpus = cpumask_last(&cpu_online_map) + 1; int *sumw = xzalloc_array(int, nr_cpus); s_time_t *sumt = xzalloc_array(s_time_t, nr_cpus); --- a/xen/common/schedule.c +++ b/xen/common/schedule.c @@ -1450,7 +1450,7 @@ void schedule_dump(struct cpupool *c) printk("Scheduler: %s (%s)\n", sched->name, sched->opt_name); SCHED_OP(sched, dump_settings); - for_each_cpu_mask (i, *cpus) + for_each_cpu (i, cpus) { pcpu_schedule_lock(i); printk("CPU[%02d] ", i); --- a/xen/common/softirq.c +++ b/xen/common/softirq.c @@ -74,7 +74,7 @@ void cpumask_raise_softirq(const cpumask cpumask_t send_mask; cpumask_clear(&send_mask); - for_each_cpu_mask(cpu, *mask) + for_each_cpu(cpu, mask) if ( !test_and_set_bit(nr, &softirq_pending(cpu)) ) cpumask_set_cpu(cpu, &send_mask); --- a/xen/common/stop_machine.c +++ b/xen/common/stop_machine.c @@ -101,7 +101,7 @@ int stop_machine_run(int (*fn)(void *), smp_wmb(); - for_each_cpu_mask ( i, allbutself ) + for_each_cpu ( i, &allbutself ) tasklet_schedule_on_cpu(&per_cpu(stopmachine_tasklet, i), i); stopmachine_set_state(STOPMACHINE_PREPARE); --- a/xen/common/timer.c +++ b/xen/common/timer.c @@ -548,7 +548,7 @@ static struct keyhandler dump_timerq_key static void migrate_timers_from_cpu(unsigned int old_cpu) { - unsigned int new_cpu = first_cpu(cpu_online_map); + unsigned int new_cpu = cpumask_any(&cpu_online_map); struct timers *old_ts, *new_ts; struct timer *t; bool_t notify = 0; --- a/xen/drivers/acpi/pmstat.c +++ b/xen/drivers/acpi/pmstat.c @@ -223,7 +223,7 @@ static int get_cpufreq_para(struct xen_s if ( !(affected_cpus = xzalloc_array(uint32_t, op->u.get_para.cpu_num)) ) return -ENOMEM; - for_each_cpu_mask(cpu, *policy->cpus) + for_each_cpu(cpu, policy->cpus) affected_cpus[j++] = cpu; ret = copy_to_guest(op->u.get_para.affected_cpus, affected_cpus, op->u.get_para.cpu_num); --- a/xen/drivers/cpufreq/cpufreq_ondemand.c +++ b/xen/drivers/cpufreq/cpufreq_ondemand.c @@ -122,7 +122,7 @@ static void dbs_check_cpu(struct cpu_dbs return; /* Get Idle Time */ - for_each_cpu_mask(j, *policy->cpus) { + for_each_cpu(j, policy->cpus) { uint64_t idle_ns, total_idle_ns; uint64_t load, load_freq, freq_avg; struct cpu_dbs_info_s *j_dbs_info; @@ -233,7 +233,7 @@ int cpufreq_governor_dbs(struct cpufreq_ dbs_enable++; - for_each_cpu_mask(j, *policy->cpus) { + for_each_cpu(j, policy->cpus) { struct cpu_dbs_info_s *j_dbs_info; j_dbs_info = &per_cpu(cpu_dbs_info, j); j_dbs_info->cur_policy = policy; --- a/xen/drivers/passthrough/vtd/iommu.c +++ b/xen/drivers/passthrough/vtd/iommu.c @@ -1033,7 +1033,7 @@ static void dma_msi_set_affinity(struct msg.address_lo = (MSI_ADDRESS_HEADER << (MSI_ADDRESS_HEADER_SHIFT + 8)); msg.address_lo |= MSI_PHYSICAL_MODE << 2; msg.address_lo |= MSI_REDIRECTION_HINT_MODE << 3; - dest = cpu_physical_id(first_cpu(mask)); + dest = cpu_physical_id(cpumask_first(mask)); msg.address_lo |= dest << MSI_TARGET_CPU_SHIFT; #endif --- a/xen/include/asm-ia64/linux-xen/asm/acpi.h +++ b/xen/include/asm-ia64/linux-xen/asm/acpi.h @@ -139,7 +139,7 @@ extern int __initdata nid_to_pxm_map[MAX #ifdef CONFIG_ACPI_NUMA extern cpumask_t early_cpu_possible_map; #define for_each_possible_early_cpu(cpu) \ - for_each_cpu_mask((cpu), early_cpu_possible_map) + for_each_cpu(cpu, &early_cpu_possible_map) static inline void per_cpu_scan_finalize(int min_cpus, int reserve_cpus) { --- a/xen/include/asm-x86/flushtlb.h +++ b/xen/include/asm-x86/flushtlb.h @@ -52,7 +52,7 @@ static inline int NEED_FLUSH(u32 cpu_sta #define tlbflush_filter(mask, page_timestamp) \ do { \ unsigned int cpu; \ - for_each_cpu_mask ( cpu, mask ) \ + for_each_cpu ( cpu, &(mask) ) \ if ( !NEED_FLUSH(per_cpu(tlbflush_time, cpu), page_timestamp) ) \ cpumask_clear_cpu(cpu, &(mask)); \ } while ( 0 ) --- a/xen/include/xen/cpumask.h +++ b/xen/include/xen/cpumask.h @@ -37,18 +37,19 @@ * void cpumask_shift_right(dst, src, n) Shift right * void cpumask_shift_left(dst, src, n) Shift left * - * int first_cpu(mask) Number lowest set bit, or NR_CPUS - * int next_cpu(cpu, mask) Next cpu past ''cpu'', or NR_CPUS - * int last_cpu(mask) Number highest set bit, or NR_CPUS - * int cycle_cpu(cpu, mask) Next cpu cycling from ''cpu'', or NR_CPUS + * int cpumask_first(mask) Number lowest set bit, or NR_CPUS + * int cpumask_next(cpu, mask) Next cpu past ''cpu'', or NR_CPUS + * int cpumask_last(mask) Number highest set bit, or NR_CPUS + * int cpumask_any(mask) Any cpu in mask, or NR_CPUS + * int cpumask_cycle(cpu, mask) Next cpu cycling from ''cpu'', or NR_CPUS * - * cpumask_t cpumask_of_cpu(cpu) Return cpumask with bit ''cpu'' set + * const cpumask_t *cpumask_of(cpu) Return cpumask with bit ''cpu'' set * unsigned long *cpumask_bits(mask) Array of unsigned long''s in mask * * int cpumask_scnprintf(buf, len, mask) Format cpumask for printing * int cpulist_scnprintf(buf, len, mask) Format cpumask as list for printing * - * for_each_cpu_mask(cpu, mask) for-loop cpu over mask + * for_each_cpu(cpu, mask) for-loop cpu over mask * * int num_online_cpus() Number of online CPUs * int num_possible_cpus() Number of all possible CPUs @@ -210,42 +211,43 @@ static inline void cpumask_shift_left(cp bitmap_shift_left(dstp->bits, srcp->bits, n, nr_cpumask_bits); } -#define cpumask_first(src) __first_cpu(src, nr_cpu_ids) -#define first_cpu(src) __first_cpu(&(src), nr_cpu_ids) -static inline int __first_cpu(const cpumask_t *srcp, int nbits) +static inline int cpumask_first(const cpumask_t *srcp) { - return min_t(int, nbits, find_first_bit(srcp->bits, nbits)); + return min_t(int, nr_cpu_ids, find_first_bit(srcp->bits, nr_cpu_ids)); } -#define cpumask_next(n, src) __next_cpu(n, src, nr_cpu_ids) -#define next_cpu(n, src) __next_cpu((n), &(src), nr_cpu_ids) -static inline int __next_cpu(int n, const cpumask_t *srcp, int nbits) +static inline int cpumask_next(int n, const cpumask_t *srcp) { - return min_t(int, nbits, find_next_bit(srcp->bits, nbits, n+1)); + /* -1 is a legal arg here. */ + if (n != -1) + cpumask_check(n); + + return min_t(int, nr_cpu_ids, + find_next_bit(srcp->bits, nr_cpu_ids, n + 1)); } -#define cpumask_last(src) __last_cpu(src, nr_cpu_ids) -#define last_cpu(src) __last_cpu(&(src), nr_cpu_ids) -static inline int __last_cpu(const cpumask_t *srcp, int nbits) +static inline int cpumask_last(const cpumask_t *srcp) { - int cpu, pcpu = nbits; - for (cpu = __first_cpu(srcp, nbits); - cpu < nbits; - cpu = __next_cpu(cpu, srcp, nbits)) + int cpu, pcpu = nr_cpu_ids; + + for (cpu = cpumask_first(srcp); + cpu < nr_cpu_ids; + cpu = cpumask_next(cpu, srcp)) pcpu = cpu; return pcpu; } -#define cpumask_cycle(n, src) __cycle_cpu(n, src, nr_cpu_ids) -#define cycle_cpu(n, src) __cycle_cpu((n), &(src), nr_cpu_ids) -static inline int __cycle_cpu(int n, const cpumask_t *srcp, int nbits) -{ - int nxt = __next_cpu(n, srcp, nbits); - if (nxt == nbits) - nxt = __first_cpu(srcp, nbits); +static inline int cpumask_cycle(int n, const cpumask_t *srcp) +{ + int nxt = cpumask_next(n, srcp); + + if (nxt == nr_cpu_ids) + nxt = cpumask_first(srcp); return nxt; } +#define cpumask_any(srcp) cpumask_first(srcp) + /* * Special-case data structure for "single bit set only" constant CPU masks. * @@ -262,8 +264,6 @@ static inline const cpumask_t *cpumask_o return (const cpumask_t *)(p - cpu / BITS_PER_LONG); } -#define cpumask_of_cpu(cpu) (*cpumask_of(cpu)) - #if defined(__ia64__) /* XXX needs cleanup */ #define CPU_MASK_LAST_WORD BITMAP_LAST_WORD_MASK(NR_CPUS) @@ -366,12 +366,13 @@ static inline void free_cpumask_var(cpum #endif #if NR_CPUS > 1 -#define for_each_cpu_mask(cpu, mask) \ - for ((cpu) = first_cpu(mask); \ - (cpu) < nr_cpu_ids; \ - (cpu) = next_cpu((cpu), (mask))) +#define for_each_cpu(cpu, mask) \ + for ((cpu) = cpumask_first(mask); \ + (cpu) < nr_cpu_ids; \ + (cpu) = cpumask_next(cpu, mask)) #else /* NR_CPUS == 1 */ -#define for_each_cpu_mask(cpu, mask) for ((cpu) = 0; (cpu) < 1; (cpu)++) +#define for_each_cpu(cpu, mask) \ + for ((cpu) = 0; (cpu) < 1; (cpu)++, (void)(mask)) #endif /* NR_CPUS */ /* @@ -458,18 +459,9 @@ extern cpumask_t cpu_present_map; ((present) ? cpumask_set_cpu(cpu, &cpu_present_map) \ : cpumask_clear_cpu(cpu, &cpu_present_map)) -#define any_online_cpu(mask) \ -({ \ - int cpu; \ - for_each_cpu_mask(cpu, (mask)) \ - if (cpu_online(cpu)) \ - break; \ - cpu; \ -}) - -#define for_each_possible_cpu(cpu) for_each_cpu_mask((cpu), cpu_possible_map) -#define for_each_online_cpu(cpu) for_each_cpu_mask((cpu), cpu_online_map) -#define for_each_present_cpu(cpu) for_each_cpu_mask((cpu), cpu_present_map) +#define for_each_possible_cpu(cpu) for_each_cpu(cpu, &cpu_possible_map) +#define for_each_online_cpu(cpu) for_each_cpu(cpu, &cpu_online_map) +#define for_each_present_cpu(cpu) for_each_cpu(cpu, &cpu_present_map) /* Copy to/from cpumap provided by control tools. */ struct xenctl_cpumap; _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com http://lists.xensource.com/xen-devel