/* * Uniprocessor-only support functions. The counterpart to kernel/smp.c */ #include #include #include #include #include int smp_call_function_single(int cpu, void (*func) (void *info), void *info, int wait) { unsigned long flags; WARN_ON(cpu != 0); local_irq_save(flags); func(info); local_irq_restore(flags); return 0; } EXPORT_SYMBOL(smp_call_function_single); int smp_call_function_single_async(int cpu, struct call_single_data *csd) { unsigned long flags; local_irq_save(flags); csd->func(csd->info); local_irq_restore(flags); return 0; } EXPORT_SYMBOL(smp_call_function_single_async); int on_each_cpu(smp_call_func_t func, void *info, int wait) { unsigned long flags; local_irq_save(flags); func(info); local_irq_restore(flags); return 0; } EXPORT_SYMBOL(on_each_cpu); /* * Note we still need to test the mask even for UP * because we actually can get an empty mask from * code that on SMP might call us without the local * CPU in the mask. */ void on_each_cpu_mask(const struct cpumask *mask, smp_call_func_t func, void *info, bool wait) { unsigned long flags; if (cpumask_test_cpu(0, mask)) { local_irq_save(flags); func(info); local_irq_restore(flags); } } EXPORT_SYMBOL(on_each_cpu_mask); /* * Preemption is disabled here to make sure the cond_func is called under the * same condtions in UP and SMP. */ void on_each_cpu_cond(bool (*cond_func)(int cpu, void *info), smp_call_func_t func, void *info, bool wait, gfp_t gfp_flags) { unsigned long flags; preempt_disable(); if (cond_func(0, info)) { local_irq_save(flags); func(info); local_irq_restore(flags); } preempt_enable(); } EXPORT_SYMBOL(on_each_cpu_cond); int smp_call_on_cpu(unsigned int cpu, int (*func)(void *), void *par, bool phys) { int ret; if (cpu != 0) return -ENXIO; if (phys) hypervisor_pin_vcpu(0); ret = func(par); if (phys) hypervisor_pin_vcpu(-1); return ret; } EXPORT_SYMBOL_GPL(smp_call_on_cpu); 8b79c321e4a1f31f98194e4f6b6cae7'>commitdiff
diff options
context:
space:
mode:
Diffstat (limited to 'include/trace/events/migrate.h')