spin_lock(&s->lock);
- rc = alloc_unbound_xen_event_channel(v, s->domid, NULL);
+ rc = alloc_unbound_xen_event_channel(v->domain, v->vcpu_id, s->domid,
+ NULL);
if ( rc < 0 )
goto fail2;
{
struct domain *d = s->domain;
- rc = alloc_unbound_xen_event_channel(v, s->domid, NULL);
+ rc = alloc_unbound_xen_event_channel(v->domain, 0, s->domid, NULL);
if ( rc < 0 )
goto fail3;
return 0;
fail3:
- free_xen_event_channel(v, sv->ioreq_evtchn);
+ free_xen_event_channel(v->domain, sv->ioreq_evtchn);
fail2:
spin_unlock(&s->lock);
list_del(&sv->list_entry);
if ( v->vcpu_id == 0 && s->bufioreq.va != NULL )
- free_xen_event_channel(v, s->bufioreq_evtchn);
+ free_xen_event_channel(v->domain, s->bufioreq_evtchn);
- free_xen_event_channel(v, sv->ioreq_evtchn);
+ free_xen_event_channel(v->domain, sv->ioreq_evtchn);
xfree(sv);
break;
list_del(&sv->list_entry);
if ( v->vcpu_id == 0 && s->bufioreq.va != NULL )
- free_xen_event_channel(v, s->bufioreq_evtchn);
+ free_xen_event_channel(v->domain, s->bufioreq_evtchn);
- free_xen_event_channel(v, sv->ioreq_evtchn);
+ free_xen_event_channel(v->domain, sv->ioreq_evtchn);
xfree(sv);
}
{
int old_port, new_port;
- new_port = alloc_unbound_xen_event_channel(v, remote_domid, NULL);
+ new_port = alloc_unbound_xen_event_channel(v->domain, v->vcpu_id,
+ remote_domid, NULL);
if ( new_port < 0 )
return new_port;
/* xchg() ensures that only we call free_xen_event_channel(). */
old_port = xchg(p_port, new_port);
- free_xen_event_channel(v, old_port);
+ free_xen_event_channel(v->domain, old_port);
return 0;
}
int alloc_unbound_xen_event_channel(
- struct vcpu *local_vcpu, domid_t remote_domid,
+ struct domain *ld, unsigned int lvcpu, domid_t remote_domid,
xen_event_channel_notification_t notification_fn)
{
struct evtchn *chn;
- struct domain *d = local_vcpu->domain;
int port, rc;
- spin_lock(&d->event_lock);
+ spin_lock(&ld->event_lock);
- rc = get_free_port(d);
+ rc = get_free_port(ld);
if ( rc < 0 )
goto out;
port = rc;
- chn = evtchn_from_port(d, port);
+ chn = evtchn_from_port(ld, port);
- rc = xsm_evtchn_unbound(XSM_TARGET, d, chn, remote_domid);
+ rc = xsm_evtchn_unbound(XSM_TARGET, ld, chn, remote_domid);
if ( rc )
goto out;
chn->state = ECS_UNBOUND;
chn->xen_consumer = get_xen_consumer(notification_fn);
- chn->notify_vcpu_id = local_vcpu->vcpu_id;
+ chn->notify_vcpu_id = lvcpu;
chn->u.unbound.remote_domid = remote_domid;
out:
- spin_unlock(&d->event_lock);
+ spin_unlock(&ld->event_lock);
return rc < 0 ? rc : port;
}
-
-void free_xen_event_channel(
- struct vcpu *local_vcpu, int port)
+void free_xen_event_channel(struct domain *d, int port)
{
struct evtchn *chn;
- struct domain *d = local_vcpu->domain;
spin_lock(&d->event_lock);
med->blocked = 0;
/* Allocate event channel */
- rc = alloc_unbound_xen_event_channel(d->vcpu[0],
- current->domain->domain_id,
+ rc = alloc_unbound_xen_event_channel(d, 0, current->domain->domain_id,
notification_fn);
if ( rc < 0 )
goto err;
}
/* Free domU's event channel and leave the other one unbound */
- free_xen_event_channel(d->vcpu[0], med->xen_port);
+ free_xen_event_channel(d, med->xen_port);
/* Unblock all vCPUs */
for_each_vcpu ( d, v )
typedef void (*xen_event_channel_notification_t)(
struct vcpu *v, unsigned int port);
int alloc_unbound_xen_event_channel(
- struct vcpu *local_vcpu, domid_t remote_domid,
+ struct domain *ld, unsigned int lvcpu, domid_t remote_domid,
xen_event_channel_notification_t notification_fn);
-void free_xen_event_channel(
- struct vcpu *local_vcpu, int port);
+void free_xen_event_channel(struct domain *d, int port);
/* Query if event channel is in use by the guest */
int guest_enabled_event(struct vcpu *v, uint32_t virq);