#define PREFIX "ACPI: "
-bool_t __initdata acpi_noirq; /* skip ACPI IRQ initialization */
-bool_t __initdata acpi_ht = 1; /* enable HT */
+bool __initdata acpi_noirq; /* skip ACPI IRQ initialization */
+bool __initdata acpi_ht = true; /* enable HT */
-bool_t __initdata acpi_lapic;
-bool_t __initdata acpi_ioapic;
+bool __initdata acpi_lapic;
+bool __initdata acpi_ioapic;
/* acpi_skip_timer_override: Skip IRQ0 overrides. */
-static bool_t acpi_skip_timer_override __initdata;
+static bool __initdata acpi_skip_timer_override;
boolean_param("acpi_skip_timer_override", acpi_skip_timer_override);
static u64 acpi_lapic_addr __initdata = APIC_DEFAULT_PHYS_BASE;
{
struct acpi_madt_local_x2apic *processor =
container_of(header, struct acpi_madt_local_x2apic, header);
- bool_t enabled = 0;
+ bool enabled = false;
if (BAD_MADT_ENTRY(processor, end))
return -EINVAL;
if (processor->lapic_flags & ACPI_MADT_ENABLED) {
x86_acpiid_to_apicid[processor->uid] =
processor->local_apic_id;
- enabled = 1;
+ enabled = true;
}
/*
{
struct acpi_madt_local_apic *processor =
container_of(header, struct acpi_madt_local_apic, header);
- bool_t enabled = 0;
+ bool enabled = false;
if (BAD_MADT_ENTRY(processor, end))
return -EINVAL;
/* Record local apic id only when enabled */
if (processor->lapic_flags & ACPI_MADT_ENABLED) {
x86_acpiid_to_apicid[processor->processor_id] = processor->id;
- enabled = 1;
+ enabled = true;
}
/*
*/
error = acpi_parse_madt_lapic_entries();
if (!error) {
- acpi_lapic = 1;
+ acpi_lapic = true;
generic_bigsmp_probe();
/*
*/
error = acpi_parse_madt_ioapic_entries();
if (!error) {
- acpi_ioapic = 1;
+ acpi_ioapic = true;
smp_found_config = 1;
clustered_apic_check();
* other side effects.
*
* side effects of acpi_boot_init:
- * acpi_lapic = 1 if LAPIC found
- * acpi_ioapic = 1 if IOAPIC found
+ * acpi_lapic = true if LAPIC found
+ * acpi_ioapic = true if IOAPIC found
* if (acpi_lapic && acpi_ioapic) smp_found_config = 1;
* ...
*
void (*__read_mostly lapic_timer_off)(void);
void (*__read_mostly lapic_timer_on)(void);
-bool_t lapic_timer_init(void)
+bool lapic_timer_init(void)
{
if ( boot_cpu_has(X86_FEATURE_ARAT) )
{
lapic_timer_on = pit_broadcast_exit;
}
else
- return 0;
+ return false;
- return 1;
+ return true;
}
static uint64_t (*__read_mostly tick_to_ns)(uint64_t) = acpi_pm_tick_to_ns;
void (*__read_mostly pm_idle_save)(void);
unsigned int max_cstate __read_mostly = ACPI_PROCESSOR_MAX_POWER - 1;
integer_param("max_cstate", max_cstate);
-static bool_t __read_mostly local_apic_timer_c2_ok;
+static bool __read_mostly local_apic_timer_c2_ok;
boolean_param("lapic_timer_c2_ok", local_apic_timer_c2_ok);
struct acpi_processor_power *__read_mostly processor_powers[NR_CPUS];
cpumask_andnot(mask, mask, &target);
}
-bool_t arch_skip_send_event_check(unsigned int cpu)
+bool arch_skip_send_event_check(unsigned int cpu)
{
/*
* This relies on softirq_pending() and mwait_wakeup() to access data
* may not be sent if software enters core C6 during an interrupt service
* routine. So we don't enter deep Cx state if there is an EOI pending.
*/
-bool_t errata_c6_eoi_workaround(void)
+static bool errata_c6_eoi_workaround(void)
{
static int8_t fix_needed = -1;
cpu_id = get_cpu_id(cpu);
if ( cpu_id == -1 )
{
- static bool_t warn_once = 1;
+ static bool warn_once = true;
+
if ( warn_once || opt_cpu_info )
printk(XENLOG_WARNING "No CPU ID for APIC ID %#x\n", cpu);
- warn_once = 0;
+ warn_once = false;
return -EINVAL;
}
hpet_disable_legacy_broadcast();
}
-bool_t cpuidle_using_deep_cstate(void)
+bool cpuidle_using_deep_cstate(void)
{
return xen_cpuidle && max_cstate > (local_apic_timer_c2_ok ? 2 : 1);
}
static struct cpufreq_driver acpi_cpufreq_driver;
-static bool_t __read_mostly acpi_pstate_strict;
+static bool __read_mostly acpi_pstate_strict;
boolean_param("acpi_pstate_strict", acpi_pstate_strict);
static int check_est_cpu(unsigned int cpuid)
:"=r"(n_hi), "=r"(n_lo) \
:"0"(n_hi), "1"(n_lo))
-extern bool_t acpi_lapic, acpi_ioapic, acpi_noirq;
-extern bool_t acpi_ht;
-extern bool acpi_force, acpi_disabled;
+extern bool acpi_lapic, acpi_ioapic, acpi_noirq;
+extern bool acpi_force, acpi_ht, acpi_disabled;
extern u32 acpi_smi_cmd;
extern u8 acpi_enable_value, acpi_disable_value;
void acpi_pic_sci_set_trigger(unsigned int, u16);
extern void (*pm_idle_save)(void);
-bool_t lapic_timer_init(void);
+bool lapic_timer_init(void);
extern void (*lapic_timer_off)(void);
extern void (*lapic_timer_on)(void);
#define HVM_DPCI_SOFTIRQ (NR_COMMON_SOFTIRQS + 5)
#define NR_ARCH_SOFTIRQS 6
-bool_t arch_skip_send_event_check(unsigned int cpu);
+bool arch_skip_send_event_check(unsigned int cpu);
#endif /* __ASM_SOFTIRQ_H__ */
extern s8 xen_cpuidle;
extern struct cpuidle_governor *cpuidle_current_governor;
-bool_t cpuidle_using_deep_cstate(void);
+bool cpuidle_using_deep_cstate(void);
void cpuidle_disable_deep_cstate(void);
extern void cpuidle_wakeup_mwait(cpumask_t *mask);