Re: [PATCH] xen-evtchn: Bind dyn evtchn:qemu-dm interrupt to next online VCPU

From: Juergen Gross
Date: Fri Jun 02 2017 - 11:10:16 EST


On 02/06/17 16:49, Anoob Soman wrote:
> A HVM domian booting generates around 200K (evtchn:qemu-dm xen-dyn)
> interrupts,in a short period of time. All these evtchn:qemu-dm are bound
> to VCPU 0, until irqbalance sees these IRQ and moves it to a different VCPU.
> In one configuration, irqbalance runs every 10 seconds, which means
> irqbalance doesn't get to see these burst of interrupts and doesn't
> re-balance interrupts most of the time, making all evtchn:qemu-dm to be
> processed by VCPU0. This cause VCPU0 to spend most of time processing
> hardirq and very little time on softirq. Moreover, if dom0 kernel PREEMPTION
> is disabled, VCPU0 never runs watchdog (process context), triggering a
> softlockup detection code to panic.
>
> Binding evtchn:qemu-dm to next online VCPU, will spread hardirq
> processing evenly across different CPU. Later, irqbalance will try to balance
> evtchn:qemu-dm, if required.
>
> Signed-off-by: Anoob Soman <anoob.soman@xxxxxxxxxx>
> ---
> drivers/xen/events/events_base.c | 9 +++++++--
> drivers/xen/evtchn.c | 36 +++++++++++++++++++++++++++++++++++-
> include/xen/events.h | 1 +
> 3 files changed, 43 insertions(+), 3 deletions(-)
>
> diff --git a/drivers/xen/events/events_base.c b/drivers/xen/events/events_base.c
> index b52852f..8224ec1 100644
> --- a/drivers/xen/events/events_base.c
> +++ b/drivers/xen/events/events_base.c
> @@ -1303,10 +1303,9 @@ void rebind_evtchn_irq(int evtchn, int irq)
> }
>
> /* Rebind an evtchn so that it gets delivered to a specific cpu */
> -static int rebind_irq_to_cpu(unsigned irq, unsigned tcpu)
> +int xen_rebind_evtchn_to_cpu(int evtchn, unsigned tcpu)
> {
> struct evtchn_bind_vcpu bind_vcpu;
> - int evtchn = evtchn_from_irq(irq);
> int masked;
>
> if (!VALID_EVTCHN(evtchn))
> @@ -1338,6 +1337,12 @@ static int rebind_irq_to_cpu(unsigned irq, unsigned tcpu)
>
> return 0;
> }
> +EXPORT_SYMBOL_GPL(xen_rebind_evtchn_to_cpu);
> +
> +static int rebind_irq_to_cpu(unsigned irq, unsigned tcpu)
> +{
> + return xen_rebind_evtchn_to_cpu(evtchn_from_irq(irq), tcpu);
> +}
>
> static int set_affinity_irq(struct irq_data *data, const struct cpumask *dest,
> bool force)
> diff --git a/drivers/xen/evtchn.c b/drivers/xen/evtchn.c
> index 10f1ef5..1192f24 100644
> --- a/drivers/xen/evtchn.c
> +++ b/drivers/xen/evtchn.c
> @@ -58,6 +58,8 @@
> #include <xen/xen-ops.h>
> #include <asm/xen/hypervisor.h>
>
> +static DEFINE_PER_CPU(int, bind_last_selected_cpu);
> +
> struct per_user_data {
> struct mutex bind_mutex; /* serialize bind/unbind operations */
> struct rb_root evtchns;
> @@ -421,6 +423,36 @@ static void evtchn_unbind_from_user(struct per_user_data *u,
> del_evtchn(u, evtchn);
> }
>
> +static void evtchn_bind_interdom_next_vcpu(int evtchn)
> +{
> + unsigned int selected_cpu, irq;
> + struct irq_desc *desc = NULL;
> + unsigned long flags;
> +
> + irq = irq_from_evtchn(evtchn);
> + desc = irq_to_desc(irq);
> +
> + if (!desc)
> + return;
> +
> + raw_spin_lock_irqsave(&desc->lock, flags);
> + selected_cpu = this_cpu_read(bind_last_selected_cpu);
> + selected_cpu = cpumask_next_and(selected_cpu,
> + desc->irq_common_data.affinity, cpu_online_mask);
> +
> + if (unlikely(selected_cpu >= nr_cpu_ids))
> + selected_cpu = cpumask_first_and(desc->irq_common_data.affinity,
> + cpu_online_mask);
> +
> + raw_spin_unlock_irqrestore(&desc->lock, flags);
> + this_cpu_write(bind_last_selected_cpu, selected_cpu);
> +
> + local_irq_disable();
> + /* unmask expects irqs to be disabled */
> + xen_rebind_evtchn_to_cpu(evtchn, selected_cpu);
> + local_irq_enable();

I'd prefer the to have irq disabled from taking the lock until here.
This will avoid problems due to preemption and will be faster as it
avoids one irq on/off cycle. So:

local_irq_disable();
raw_spin_lock();
...
raw_spin_unlock();
this_cpu_write();
xen_rebind_evtchn_to_cpu();
local_irq_enable();


Juergen