return apic_tmict || !timeout;
}
-static void cf_check apic_timer_interrupt(struct cpu_user_regs *regs)
+static void cf_check apic_timer_interrupt(void)
{
ack_APIC_irq();
perfc_incr(apic_timer);
/*
* Spurious interrupts should _never_ happen with our APIC/SMP architecture.
*/
-static void cf_check spurious_interrupt(struct cpu_user_regs *regs)
+static void cf_check spurious_interrupt(void)
{
/*
* Check if this is a vectored interrupt (most likely, as this is probably
is_spurious = !nmi_check_continuation();
if (this_cpu(state_dump_pending)) {
this_cpu(state_dump_pending) = false;
- dump_execstate(regs);
+ dump_execstate(get_irq_regs());
is_spurious = false;
}
* This interrupt should never happen with our APIC/SMP architecture
*/
-static void cf_check error_interrupt(struct cpu_user_regs *regs)
+static void cf_check error_interrupt(void)
{
static const char *const esr_fields[] = {
", Send CS error",
* This interrupt handles performance counters interrupt
*/
-static void cf_check pmu_interrupt(struct cpu_user_regs *regs)
+static void cf_check pmu_interrupt(void)
{
ack_APIC_irq();
vpmu_do_interrupt();
#define MCE_RING 0x1
static DEFINE_PER_CPU(int, last_state);
-static void cf_check intel_thermal_interrupt(struct cpu_user_regs *regs)
+static void cf_check intel_thermal_interrupt(void)
{
uint64_t msr_content;
unsigned int cpu = smp_processor_id();
clear_cmci();
}
-static void cf_check cmci_interrupt(struct cpu_user_regs *regs)
+static void cf_check cmci_interrupt(void)
{
mctelem_cookie_t mctc;
struct mca_summary bs;
}
}
-static void cf_check xen_evtchn_upcall(struct cpu_user_regs *regs)
+static void cf_check xen_evtchn_upcall(void)
{
struct vcpu_info *vcpu_info = this_cpu(vcpu_info);
unsigned long pending;
};
/* Handle VT-d posted-interrupt when VCPU is blocked. */
-static void cf_check pi_wakeup_interrupt(struct cpu_user_regs *regs)
+static void cf_check pi_wakeup_interrupt(void)
{
struct vmx_vcpu *vmx, *tmp;
spinlock_t *lock = &per_cpu(vmx_pi_blocking, smp_processor_id()).lock;
}
/* Handle VT-d posted-interrupt when VCPU is running. */
-static void cf_check pi_notification_interrupt(struct cpu_user_regs *regs)
+static void cf_check pi_notification_interrupt(void)
{
ack_APIC_irq();
this_cpu(irq_count)++;
#define platform_legacy_irq(irq) ((irq) < 16)
-void cf_check event_check_interrupt(struct cpu_user_regs *regs);
-void cf_check invalidate_interrupt(struct cpu_user_regs *regs);
-void cf_check call_function_interrupt(struct cpu_user_regs *regs);
-void cf_check irq_move_cleanup_interrupt(struct cpu_user_regs *regs);
+void cf_check event_check_interrupt(void);
+void cf_check invalidate_interrupt(void);
+void cf_check call_function_interrupt(void);
+void cf_check irq_move_cleanup_interrupt(void);
uint8_t alloc_hipriority_vector(void);
-void set_direct_apic_vector(
- uint8_t vector, void (*handler)(struct cpu_user_regs *regs));
-void alloc_direct_apic_vector(
- uint8_t *vector, void (*handler)(struct cpu_user_regs *regs));
+void set_direct_apic_vector(uint8_t vector, void (*handler)(void));
+void alloc_direct_apic_vector(uint8_t *vector, void (*handler)(void));
void do_IRQ(struct cpu_user_regs *regs);
desc->handler->enable(desc);
}
-void cf_check irq_move_cleanup_interrupt(struct cpu_user_regs *regs)
+void cf_check irq_move_cleanup_interrupt(void)
{
unsigned vector, me;
return next++;
}
-static void (*direct_apic_vector[X86_NR_VECTORS])(struct cpu_user_regs *regs);
-void set_direct_apic_vector(
- uint8_t vector, void (*handler)(struct cpu_user_regs *regs))
+static void (*direct_apic_vector[X86_NR_VECTORS])(void);
+void set_direct_apic_vector(uint8_t vector, void (*handler)(void))
{
BUG_ON(direct_apic_vector[vector] != NULL);
direct_apic_vector[vector] = handler;
}
-void alloc_direct_apic_vector(
- uint8_t *vector, void (*handler)(struct cpu_user_regs *regs))
+void alloc_direct_apic_vector(uint8_t *vector, void (*handler)(void))
{
static DEFINE_SPINLOCK(lock);
if ( irq < 0 )
{
if ( direct_apic_vector[vector] )
- direct_apic_vector[vector](regs);
+ direct_apic_vector[vector]();
else
{
const char *kind = ", LAPIC";
static const void *flush_va;
static unsigned int flush_flags;
-void cf_check invalidate_interrupt(struct cpu_user_regs *regs)
+void cf_check invalidate_interrupt(void)
{
unsigned int flags = flush_flags;
ack_APIC_irq();
send_IPI_mask(&cpu_online_map, APIC_DM_NMI);
}
-void cf_check event_check_interrupt(struct cpu_user_regs *regs)
+void cf_check event_check_interrupt(void)
{
ack_APIC_irq();
perfc_incr(ipis);
this_cpu(irq_count)++;
}
-void cf_check call_function_interrupt(struct cpu_user_regs *regs)
+void cf_check call_function_interrupt(void)
{
ack_APIC_irq();
perfc_incr(ipis);