ASSERT(cpus_weight(dirty_mask) <= 1);
if ( unlikely(!cpu_isset(cpu, dirty_mask) && !cpus_empty(dirty_mask)) )
{
- /* Other cpus call __sync_lazy_execstate from flush ipi handler. */
+ /* Other cpus call __sync_local_execstate from flush ipi handler. */
flush_tlb_mask(&dirty_mask);
}
BUG();
}
-int __sync_lazy_execstate(void)
+int __sync_local_execstate(void)
{
unsigned long flags;
int switch_required;
void sync_local_execstate(void)
{
- (void)__sync_lazy_execstate();
+ (void)__sync_local_execstate();
}
void sync_vcpu_execstate(struct vcpu *v)
{
if ( cpu_isset(smp_processor_id(), v->vcpu_dirty_cpumask) )
- (void)__sync_lazy_execstate();
+ sync_local_execstate();
- /* Other cpus call __sync_lazy_execstate from flush ipi handler. */
+ /* Other cpus call __sync_local_execstate from flush ipi handler. */
flush_tlb_mask(&v->vcpu_dirty_cpumask);
}
ack_APIC_irq();
perfc_incr(ipis);
irq_enter();
- if ( !__sync_lazy_execstate() ||
+ if ( !__sync_local_execstate() ||
(flush_flags & (FLUSH_TLB_GLOBAL | FLUSH_CACHE)) )
flush_area_local(flush_va, flush_flags);
cpu_clear(smp_processor_id(), flush_cpumask);
int cpu = raw_smp_processor_id();
/* Previous non-idle state should be synchronised already. */
- if (__sync_lazy_execstate())
+ if (__sync_local_execstate())
BUG();
cpucount --;
* before changing cpu_online_map. If we are running non-idle vcpu,
* we will synchronously sync the state in context_switch() later.
*/
- __sync_lazy_execstate();
+ sync_local_execstate();
/* It's now safe to remove this processor from the online map */
cpu_clear(cpu, cpu_online_map);
{
/* If we really are idling, perform lazy context switch now. */
if ( (v = idle_vcpu[smp_processor_id()]) == current )
- __sync_lazy_execstate();
+ sync_local_execstate();
/* We must now be running on the idle page table. */
ASSERT(read_cr3() == __pa(idle_pg_table));
}
void propagate_page_fault(unsigned long addr, u16 error_code);
void *do_page_walk(struct vcpu *v, unsigned long addr);
-int __sync_lazy_execstate(void);
+int __sync_local_execstate(void);
/* Arch-specific portion of memory_op hypercall. */
long arch_memory_op(int op, XEN_GUEST_HANDLE(void) arg);