Re: [PATCH] IB/hfi1: Fix potential deadlock on &sde->flushlist_lock

From: Chengfeng Ye
Date: Tue Jul 04 2023 - 13:42:50 EST


> Plus, we already in context where interrupts are stopped.

Indeed they can be called from .ndo_start_xmit callback and
the document said it is with bh disabled.

But I found some call chain from the user process that seems could
be called from irq disabled context. For sdma_send_txlist(),
there is a call chain.

-> hfi1_write_iter() (.write_iter callback)
-> hfi1_user_sdma_process_request()
-> user_sdma_send_pkts()
-> sdma_send_txlist()

The .write_iter seems not to disable irq by default, as mentioned by
https://www.kernel.org/doc/Documentation/filesystems/vfs.txt
And I didn't find any explicit disabling or bh or irq along the call path,
and also see several copy_from_usr() which cannot be invoked under
irq context.


For sdma_send_txreq(), there is a call chain.

-> qp_priv_alloc()
-> iowait_init() (register _hfi1_do_tid_send() as a work queue)
-> _hfi1_do_tid_send() (workqueue)
-> hfi1_do_tid_send()
-> hfi1_verbs_send()
-> sr(qp, ps, 0) (sr could points to hfi1_verbs_send_dm())
-> hfi1_verbs_send_dma()
-> sdma_send_txreq()

_hfi1_do_tid_send() is a work queue without irq disabled by default,
I also check the remaining call path and also found that there is no explicit
irq disable, instead the call site of hfi1_verbs_send() is exactly after
spin_lock_irq_restore(), seems like a hint that it is probably called withirq
enable.

Another hint is that the lock acquisition of
spin_lock_irqsave(&sde->tail_lock, flags);
just before my patch in the same function also disable irq, seems like another
hint that this function could be called with interrupt disable,
otherwise the lock/unlock
for sde->tail_lock does not need to disable irq?

Would be appreciated if you could further check with this.


> It can't work as exactly after "ret = -ECOMM;" line, there is "goto unlock"
> and there hfi1 calls to spin_unlock_irqrestore(..) with same "flags".

Yeah, that's my negligence, and sorry for this. Once you confirm that
there should
be some fixes, I would like to provide with v2 patch with the correct fix.

Best Regards,
Chengfeng


Leon Romanovsky <leon@xxxxxxxxxx> 于2023年7月4日周二 19:48写道:
>
> On Wed, Jun 28, 2023 at 04:59:25AM +0000, Chengfeng Ye wrote:
> > As &sde->flushlist_lock is acquired by timer sdma_err_progress_check()
> > through layer of calls under softirq context, other process
> > context code acquiring the lock should disable irq.
> >
> > Possible deadlock scenario
> > sdma_send_txreq()
> > -> spin_lock(&sde->flushlist_lock)
> > <timer interrupt>
> > -> sdma_err_progress_check()
> > -> __sdma_process_event()
> > -> sdma_set_state()
> > -> sdma_flush()
> > -> spin_lock_irqsave(&sde->flushlist_lock, flags) (deadlock here)
> >
> > This flaw was found using an experimental static analysis tool we are
> > developing for irq-related deadlock.
> >
> > The tentative patch fix the potential deadlock by spin_lock_irqsave().
> >
> > Signed-off-by: Chengfeng Ye <dg573847474@xxxxxxxxx>
> > ---
> > drivers/infiniband/hw/hfi1/sdma.c | 8 ++++----
> > 1 file changed, 4 insertions(+), 4 deletions(-)
> >
> > diff --git a/drivers/infiniband/hw/hfi1/sdma.c b/drivers/infiniband/hw/hfi1/sdma.c
> > index bb2552dd29c1..0431f575c861 100644
> > --- a/drivers/infiniband/hw/hfi1/sdma.c
> > +++ b/drivers/infiniband/hw/hfi1/sdma.c
> > @@ -2371,9 +2371,9 @@ int sdma_send_txreq(struct sdma_engine *sde,
> > tx->sn = sde->tail_sn++;
> > trace_hfi1_sdma_in_sn(sde, tx->sn);
> > #endif
> > - spin_lock(&sde->flushlist_lock);
> > + spin_lock_irqsave(&sde->flushlist_lock, flags);
> > list_add_tail(&tx->list, &sde->flushlist);
> > - spin_unlock(&sde->flushlist_lock);
> > + spin_unlock_irqrestore(&sde->flushlist_lock, flags);
> > iowait_inc_wait_count(wait, tx->num_desc);
> > queue_work_on(sde->cpu, system_highpri_wq, &sde->flush_worker);
> > ret = -ECOMM;
>
> It can't work as exactly after "ret = -ECOMM;" line, there is "goto unlock"
> and there hfi1 calls to spin_unlock_irqrestore(..) with same "flags".
>
> Plus, we already in context where interrupts are stopped.
>
> Thanks
>
> > @@ -2459,7 +2459,7 @@ int sdma_send_txlist(struct sdma_engine *sde, struct iowait_work *wait,
> > *count_out = total_count;
> > return ret;
> > unlock_noconn:
> > - spin_lock(&sde->flushlist_lock);
> > + spin_lock_irqsave(&sde->flushlist_lock, flags);
> > list_for_each_entry_safe(tx, tx_next, tx_list, list) {
> > tx->wait = iowait_ioww_to_iow(wait);
> > list_del_init(&tx->list);
> > @@ -2472,7 +2472,7 @@ int sdma_send_txlist(struct sdma_engine *sde, struct iowait_work *wait,
> > flush_count++;
> > iowait_inc_wait_count(wait, tx->num_desc);
> > }
> > - spin_unlock(&sde->flushlist_lock);
> > + spin_unlock_irqrestore(&sde->flushlist_lock, flags);
> > queue_work_on(sde->cpu, system_highpri_wq, &sde->flush_worker);
> > ret = -ECOMM;
> > goto update_tail;
> > --
> > 2.17.1
> >