struct page_info *new_page;
void *sp, *dp;
- new_page = alloc_domheap_page(NULL, MEMF_bits(max_bitsize));
+ new_page = alloc_domheap_page(e, MEMF_no_owner |
+ MEMF_bits(max_bitsize));
if ( new_page == NULL )
{
gop.status = GNTST_address_too_big;
/* Allocate a chunk's worth of anonymous output pages. */
for ( j = 0; j < (1UL << out_chunk_order); j++ )
{
- page = alloc_domheap_pages(NULL, exch.out.extent_order, memflags);
+ page = alloc_domheap_pages(d, exch.out.extent_order,
+ MEMF_no_owner | memflags);
if ( unlikely(page == NULL) )
{
rc = -ENOMEM;
ASSERT(!in_irq());
- bits = domain_clamp_alloc_bitsize(d, bits ? : (BITS_PER_LONG+PAGE_SHIFT));
+ bits = domain_clamp_alloc_bitsize(memflags & MEMF_no_owner ? NULL : d,
+ bits ? : (BITS_PER_LONG+PAGE_SHIFT));
if ( (zone_hi = min_t(unsigned int, bits_to_zone(bits), zone_hi)) == 0 )
return NULL;
+ if ( memflags & MEMF_no_owner )
+ memflags |= MEMF_no_refcount;
+
if ( dma_bitsize && ((dma_zone = bits_to_zone(dma_bitsize)) < zone_hi) )
pg = alloc_heap_pages(dma_zone + 1, zone_hi, order, memflags, d);
memflags, d)) == NULL)) )
return NULL;
- if ( (d != NULL) && assign_pages(d, pg, order, memflags) )
+ if ( d && !(memflags & MEMF_no_owner) &&
+ assign_pages(d, pg, order, memflags) )
{
free_heap_pages(pg, order);
return NULL;
#define MEMF_no_dma (1U<<_MEMF_no_dma)
#define _MEMF_exact_node 4
#define MEMF_exact_node (1U<<_MEMF_exact_node)
+#define _MEMF_no_owner 5
+#define MEMF_no_owner (1U<<_MEMF_no_owner)
#define _MEMF_node 8
#define MEMF_node_mask ((1U << (8 * sizeof(nodeid_t))) - 1)
#define MEMF_node(n) ((((n) + 1) & MEMF_node_mask) << _MEMF_node)