}
else
{
- unsigned long first_mfn = alloc_boot_pages(1, 1);
+ mfn_t first_mfn = alloc_boot_pages(1, 1);
- clear_page(mfn_to_virt(first_mfn));
- pte = mfn_to_xen_entry(_mfn(first_mfn), WRITEALLOC);
+ clear_page(mfn_to_virt(mfn_x(first_mfn)));
+ pte = mfn_to_xen_entry(first_mfn, WRITEALLOC);
pte.pt.table = 1;
write_pte(p, pte);
- first = mfn_to_virt(first_mfn);
+ first = mfn_to_virt(mfn_x(first_mfn));
}
pte = mfn_to_xen_entry(_mfn(mfn), WRITEALLOC);
unsigned long nr_pages = (pe - ps) >> PAGE_SHIFT;
unsigned long nr_pdxs = pfn_to_pdx(nr_pages);
unsigned long frametable_size = nr_pdxs * sizeof(struct page_info);
- unsigned long base_mfn;
+ mfn_t base_mfn;
const unsigned long mapping_size = frametable_size < MB(32) ? MB(2) : MB(32);
#ifdef CONFIG_ARM_64
lpae_t *second, pte;
- unsigned long nr_second, second_base;
+ unsigned long nr_second;
+ mfn_t second_base;
int i;
#endif
/* Compute the number of second level pages. */
nr_second = ROUNDUP(frametable_size, FIRST_SIZE) >> FIRST_SHIFT;
second_base = alloc_boot_pages(nr_second, 1);
- second = mfn_to_virt(second_base);
+ second = mfn_to_virt(mfn_x(second_base));
for ( i = 0; i < nr_second; i++ )
{
- clear_page(mfn_to_virt(second_base + i));
- pte = mfn_to_xen_entry(_mfn(second_base + i), WRITEALLOC);
+ clear_page(mfn_to_virt(mfn_x(mfn_add(second_base, i))));
+ pte = mfn_to_xen_entry(mfn_add(second_base, i), WRITEALLOC);
pte.pt.table = 1;
write_pte(&xen_first[first_table_offset(FRAMETABLE_VIRT_START)+i], pte);
}
- create_mappings(second, 0, base_mfn, frametable_size >> PAGE_SHIFT, mapping_size);
+ create_mappings(second, 0, mfn_x(base_mfn), frametable_size >> PAGE_SHIFT,
+ mapping_size);
#else
- create_mappings(xen_second, FRAMETABLE_VIRT_START,
- base_mfn, frametable_size >> PAGE_SHIFT, mapping_size);
+ create_mappings(xen_second, FRAMETABLE_VIRT_START, mfn_x(base_mfn),
+ frametable_size >> PAGE_SHIFT, mapping_size);
#endif
memset(&frame_table[0], 0, nr_pdxs * sizeof(struct page_info));
init_boot_pages(pfn_to_paddr(boot_mfn_start), pfn_to_paddr(boot_mfn_end));
/* Copy the DTB. */
- fdt = mfn_to_virt(alloc_boot_pages(dtb_pages, 1));
+ fdt = mfn_to_virt(mfn_x(alloc_boot_pages(dtb_pages, 1)));
copy_from_paddr(fdt, dtb_paddr, dtb_size);
device_tree_flattened = fdt;
dtb_pages = (dtb_size + PAGE_SIZE-1) >> PAGE_SHIFT;
/* Copy the DTB. */
- fdt = mfn_to_virt(alloc_boot_pages(dtb_pages, 1));
+ fdt = mfn_to_virt(mfn_x(alloc_boot_pages(dtb_pages, 1)));
copy_from_paddr(fdt, dtb_paddr, dtb_size);
device_tree_flattened = fdt;
{
unsigned long s = (unsigned long)start;
unsigned long e = (unsigned long)end;
- unsigned long step, mfn;
+ unsigned long step;
+ mfn_t mfn;
ASSERT(!(s & ((1 << L2_PAGETABLE_SHIFT) - 1)));
for ( ; s < e; s += step << PAGE_SHIFT )
while ( step && s + (step << PAGE_SHIFT) > e + (4 << PAGE_SHIFT) )
step >>= PAGETABLE_ORDER;
mfn = alloc_boot_pages(step, step);
- map_pages_to_xen(s, mfn, step, PAGE_HYPERVISOR);
+ map_pages_to_xen(s, mfn_x(mfn), step, PAGE_HYPERVISOR);
}
memset(start, 0, end - start);
return ptr;
}
- return mfn_to_virt(alloc_boot_pages(1, 1));
+ return mfn_to_virt(mfn_x(alloc_boot_pages(1, 1)));
}
void free_xen_pagetable(void *v)
static int __init allocate_cachealigned_memnodemap(void)
{
unsigned long size = PFN_UP(memnodemapsize * sizeof(*memnodemap));
- unsigned long mfn = alloc_boot_pages(size, 1);
+ unsigned long mfn = mfn_x(alloc_boot_pages(size, 1));
memnodemap = mfn_to_virt(mfn);
mfn <<= PAGE_SHIFT;
/* Callback for SLIT parsing */
void __init acpi_numa_slit_init(struct acpi_table_slit *slit)
{
- unsigned long mfn;
+ mfn_t mfn;
+
if (!slit_valid(slit)) {
printk(KERN_INFO "ACPI: SLIT table looks invalid. "
"Not used.\n");
return;
}
mfn = alloc_boot_pages(PFN_UP(slit->header.length), 1);
- acpi_slit = mfn_to_virt(mfn);
+ acpi_slit = mfn_to_virt(mfn_x(mfn));
memcpy(acpi_slit, slit, slit->header.length);
}
}
}
-unsigned long __init alloc_boot_pages(
- unsigned long nr_pfns, unsigned long pfn_align)
+mfn_t __init alloc_boot_pages(unsigned long nr_pfns, unsigned long pfn_align)
{
unsigned long pg, _e;
unsigned int i = nr_bootmem_regions;
if ( pg + nr_pfns > PFN_DOWN(highmem_start) )
continue;
r->s = pg + nr_pfns;
- return pg;
+ return _mfn(pg);
}
#endif
_e = r->e;
r->e = pg;
bootmem_region_add(pg + nr_pfns, _e);
- return pg;
+ return _mfn(pg);
}
BUG();
void *ptr;
if (system_state == SYS_STATE_early_boot)
- return mfn_to_virt(alloc_boot_pages(PFN_UP(sz), 1));
+ return mfn_to_virt(mfn_x(alloc_boot_pages(PFN_UP(sz), 1)));
ptr = xmalloc_bytes(sz);
ASSERT(!ptr || is_xmalloc_memory(ptr));
/* Boot-time allocator. Turns into generic allocator after bootstrap. */
void init_boot_pages(paddr_t ps, paddr_t pe);
-unsigned long alloc_boot_pages(
- unsigned long nr_pfns, unsigned long pfn_align);
+mfn_t alloc_boot_pages(unsigned long nr_pfns, unsigned long pfn_align);
void end_boot_allocator(void);
/* Xen suballocator. These functions are interrupt-safe. */