unsigned long flags;
cpumask_t dest_mask;
- if (!cpumask_intersects(mask, &cpu_online_map))
+ if ( mask && !cpumask_intersects(mask, &cpu_online_map) )
return BAD_APICID;
spin_lock_irqsave(&vector_lock, flags);
- ret = _assign_irq_vector(desc, mask);
+ ret = _assign_irq_vector(desc, mask ?: TARGET_CPUS);
spin_unlock_irqrestore(&vector_lock, flags);
- if (ret < 0)
+ if ( ret < 0 )
return BAD_APICID;
- cpumask_copy(desc->affinity, mask);
- cpumask_and(&dest_mask, mask, desc->arch.cpu_mask);
+ if ( mask )
+ {
+ cpumask_copy(desc->affinity, mask);
+ cpumask_and(&dest_mask, mask, desc->arch.cpu_mask);
+ }
+ else
+ {
+ cpumask_setall(desc->affinity);
+ cpumask_copy(&dest_mask, desc->arch.cpu_mask);
+ }
cpumask_and(&dest_mask, &dest_mask, &cpu_online_map);
return cpu_mask_to_apicid(&dest_mask);
desc = irq_to_desc(iommu->msi.irq);
spin_lock(&desc->lock);
- set_msi_affinity(desc, &cpu_online_map);
+ set_msi_affinity(desc, NULL);
spin_unlock(&desc->lock);
amd_iommu_msi_enable(iommu, IOMMU_CONTROL_ENABLED);
const struct acpi_rhsa_unit *rhsa = drhd_to_rhsa(drhd);
unsigned int node = rhsa ? pxm_to_node(rhsa->proximity_domain)
: NUMA_NO_NODE;
- const cpumask_t *cpumask = &cpu_online_map;
+ const cpumask_t *cpumask = NULL;
struct irq_desc *desc;
if ( node < MAX_NUMNODES && node_online(node) &&
- cpumask_intersects(&node_to_cpumask(node), cpumask) )
+ cpumask_intersects(&node_to_cpumask(node), &cpu_online_map) )
cpumask = &node_to_cpumask(node);
desc = irq_to_desc(drhd->iommu->msi.irq);