#include <xen/sched.h>
/* Alternate p2m on/off per domain */
-static inline bool_t altp2m_active(const struct domain *d)
+static inline bool altp2m_active(const struct domain *d)
{
/* Not implemented on ARM. */
- return 0;
+ return false;
}
/* Alternate p2m VCPU */
};
#define __AARCH64_INSN_FUNCS(abbr, mask, val) \
-static always_inline bool_t aarch64_insn_is_##abbr(u32 code) \
+static always_inline bool aarch64_insn_is_##abbr(u32 code) \
{ return (code & (mask)) == (val); } \
static always_inline u32 aarch64_insn_get_##abbr##_value(void) \
{ return (val); }
struct vtimer phys_timer;
struct vtimer virt_timer;
- bool_t vtimer_initialized;
+ bool vtimer_initialized;
} __cacheline_aligned;
void vcpu_show_execution_state(struct vcpu *);
unsigned long raw_clear_guest(void *to, unsigned len);
int access_guest_memory_by_ipa(struct domain *d, paddr_t ipa, void *buf,
- uint32_t size, bool_t is_write);
+ uint32_t size, bool is_write);
#define __raw_copy_to_guest raw_copy_to_guest
#define __raw_copy_from_guest raw_copy_from_guest
* If true, and an access fault comes in and there is no vm_event listener,
* pause domain. Otherwise, remove access restrictions.
*/
- bool_t access_required;
+ bool access_required;
/* Defines if mem_access is in use for the domain. */
- bool_t mem_access_enabled;
+ bool mem_access_enabled;
/*
* Default P2M access type for each page in the the domain: new pages,
/* get host p2m table */
#define p2m_get_hostp2m(d) (&(d)->arch.p2m)
-static inline bool_t p2m_vm_event_sanity_check(struct domain *d)
+static inline bool p2m_vm_event_sanity_check(struct domain *d)
{
- return 1;
+ return true;
}
/*