Re: [PATCH V2] PCI/DOE: Detect on stack work items automatically

From: Ira Weiny
Date: Sat Nov 19 2022 - 12:27:58 EST


On Fri, Nov 18, 2022 at 09:11:58PM -0800, Dan Williams wrote:
> Li, Ming wrote:
> > On 11/19/2022 3:46 AM, Dan Williams wrote:
> > > Ira Weiny wrote:
> > >> On Fri, Nov 18, 2022 at 09:20:38AM +0000, David Laight wrote:
> > >>> From: ira.weiny@xxxxxxxxx
> > >>>> Sent: 18 November 2022 00:05
> > >>>>
> > >>>> Work item initialization needs to be done with either
> > >>>> INIT_WORK_ONSTACK() or INIT_WORK() depending on how the work item is
> > >>>> allocated.
> > >>>>
> > >>>> The callers of pci_doe_submit_task() allocate struct pci_doe_task on the
> > >>>> stack and pci_doe_submit_task() incorrectly used INIT_WORK().
> > >>>>
> > >>>> Jonathan suggested creating doe task allocation macros such as
> > >>>> DECLARE_CDAT_DOE_TASK_ONSTACK().[1] The issue with this is the work
> > >>>> function is not known to the callers and must be initialized correctly.
> > >>>>
> > >>>> A follow up suggestion was to have an internal 'pci_doe_work' item
> > >>>> allocated by pci_doe_submit_task().[2] This requires an allocation which
> > >>>> could restrict the context where tasks are used.
> > >>>>
> > >>>> Another idea was to have an intermediate step to initialize the task
> > >>>> struct with a new call.[3] This added a lot of complexity.
> > >>>>
> > >>>> Lukas pointed out that object_is_on_stack() is available to detect this
> > >>>> automatically.
> > >>>>
> > >>>> Use object_is_on_stack() to determine the correct init work function to
> > >>>> call.
> > >>>
> > >>> This is all a bit strange.
> > >>> The 'onstack' flag is needed for the diagnostic check:
> > >>> is_on_stack = object_is_on_stack(addr);
> > >>> if (is_on_stack == onstack)
> > >>> return;
> > >>> pr_warn(...);
> > >>> WARN_ON(1);
> > >>>
> > >>
> > >> :-(
> > >>
> > >>> So setting the flag to the location of the buffer just subverts the check.
> > >>> It that is sane there ought to be a proper way to do it.
> > >>
> > >> Ok this brings me back to my previous point and suggested patch.[*] The
> > >> fundamental bug is that the work item is allocated in different code from
> > >> the code which uses it. Separating the work item from the task.
> > >>
> > >> [*] https://lore.kernel.org/linux-cxl/20221014151045.24781-1-Jonathan.Cameron@xxxxxxxxxx/T/#m63c636c5135f304480370924f4d03c00357be667
> > >>
> > >> Bjorn would this solution be acceptable and just use GFP_KERNEL and mark the
> > >> required context for pci_doe_submit_task()?
> > >
> > > It is a waste to have an allocation when one is not needed. The value of
> > > having INIT_WORK_ONSTACK and DECLARE_COMPLETION_ONSTACK is to be clear
> > > at the call site that an async context cares about this stack frame not
> > > going out of scope.
> > >
> > > However, coming full circle, we have zero async users today, and having
> > > the completion and work struct in the task are causing a maintenance
> > > burden. So let's just rip it out for now with something like the
> > > following and circle back to add async support later when it becomes
> > > necessary: (only compile tested)
> > >
> > >
> > > diff --git a/drivers/cxl/core/pci.c b/drivers/cxl/core/pci.c
> > > index 0dbbe8d39b07..69873cdcc911 100644
> > > --- a/drivers/cxl/core/pci.c
> > > +++ b/drivers/cxl/core/pci.c
> > > @@ -488,21 +488,14 @@ static struct pci_doe_mb *find_cdat_doe(struct device *uport)
> > > CXL_DOE_TABLE_ACCESS_TABLE_TYPE_CDATA) | \
> > > FIELD_PREP(CXL_DOE_TABLE_ACCESS_ENTRY_HANDLE, (entry_handle)))
> > >
> > > -static void cxl_doe_task_complete(struct pci_doe_task *task)
> > > -{
> > > - complete(task->private);
> > > -}
> > > -
> > > struct cdat_doe_task {
> > > u32 request_pl;
> > > u32 response_pl[32];
> > > - struct completion c;
> > > struct pci_doe_task task;
> > > };
> > >
> > > #define DECLARE_CDAT_DOE_TASK(req, cdt) \
> > > struct cdat_doe_task cdt = { \
> > > - .c = COMPLETION_INITIALIZER_ONSTACK(cdt.c), \
> > > .request_pl = req, \
> > > .task = { \
> > > .prot.vid = PCI_DVSEC_VENDOR_ID_CXL, \
> > > @@ -511,8 +504,6 @@ struct cdat_doe_task cdt = { \
> > > .request_pl_sz = sizeof(cdt.request_pl), \
> > > .response_pl = cdt.response_pl, \
> > > .response_pl_sz = sizeof(cdt.response_pl), \
> > > - .complete = cxl_doe_task_complete, \
> > > - .private = &cdt.c, \
> > > } \
> > > }
> > >
> > > @@ -523,12 +514,12 @@ static int cxl_cdat_get_length(struct device *dev,
> > > DECLARE_CDAT_DOE_TASK(CDAT_DOE_REQ(0), t);
> > > int rc;
> > >
> > > - rc = pci_doe_submit_task(cdat_doe, &t.task);
> > > + rc = pci_doe_submit_task_wait(cdat_doe, &t.task);
> > > if (rc < 0) {
> > > dev_err(dev, "DOE submit failed: %d", rc);
> > > return rc;
> > > }
> > > - wait_for_completion(&t.c);
> > > +
> > > if (t.task.rv < sizeof(u32))
> > > return -EIO;
> > >
> > > @@ -552,12 +543,11 @@ static int cxl_cdat_read_table(struct device *dev,
> > > u32 *entry;
> > > int rc;
> > >
> > > - rc = pci_doe_submit_task(cdat_doe, &t.task);
> > > + rc = pci_doe_submit_task_wait(cdat_doe, &t.task);
> > > if (rc < 0) {
> > > dev_err(dev, "DOE submit failed: %d", rc);
> > > return rc;
> > > }
> > > - wait_for_completion(&t.c);
> > > /* 1 DW header + 1 DW data min */
> > > if (t.task.rv < (2 * sizeof(u32)))
> > > return -EIO;
> > > diff --git a/drivers/pci/doe.c b/drivers/pci/doe.c
> > > index e402f05068a5..115a8ff14afc 100644
> > > --- a/drivers/pci/doe.c
> > > +++ b/drivers/pci/doe.c
> > > @@ -18,7 +18,6 @@
> > > #include <linux/mutex.h>
> > > #include <linux/pci.h>
> > > #include <linux/pci-doe.h>
> > > -#include <linux/workqueue.h>
> > >
> > > #define PCI_DOE_PROTOCOL_DISCOVERY 0
> > >
> > > @@ -40,7 +39,6 @@
> > > * @cap_offset: Capability offset
> > > * @prots: Array of protocols supported (encoded as long values)
> > > * @wq: Wait queue for work item
> > > - * @work_queue: Queue of pci_doe_work items
> > > * @flags: Bit array of PCI_DOE_FLAG_* flags
> > > */
> > > struct pci_doe_mb {
> > > @@ -49,7 +47,6 @@ struct pci_doe_mb {
> > > struct xarray prots;
> > >
> > > wait_queue_head_t wq;
> > > - struct workqueue_struct *work_queue;
> > > unsigned long flags;
> > > };
> > >
> > > @@ -211,7 +208,6 @@ static int pci_doe_recv_resp(struct pci_doe_mb *doe_mb, struct pci_doe_task *tas
> > > static void signal_task_complete(struct pci_doe_task *task, int rv)
> > > {
> > > task->rv = rv;
> > > - task->complete(task);
> > > }
> > >
> > > static void signal_task_abort(struct pci_doe_task *task, int rv)
> > > @@ -231,10 +227,9 @@ static void signal_task_abort(struct pci_doe_task *task, int rv)
> > > signal_task_complete(task, rv);
> > > }
> > >
> > > -static void doe_statemachine_work(struct work_struct *work)
> > > +
> > > +static void exec_task(struct pci_doe_task *task)
> > > {
> > > - struct pci_doe_task *task = container_of(work, struct pci_doe_task,
> > > - work);
> > > struct pci_doe_mb *doe_mb = task->doe_mb;
> > > struct pci_dev *pdev = doe_mb->pdev;
> > > int offset = doe_mb->cap_offset;
> > > @@ -295,18 +290,12 @@ static void doe_statemachine_work(struct work_struct *work)
> > > signal_task_complete(task, rc);
> > > }
> > >
> > > -static void pci_doe_task_complete(struct pci_doe_task *task)
> > > -{
> > > - complete(task->private);
> > > -}
> > > -
> > > static int pci_doe_discovery(struct pci_doe_mb *doe_mb, u8 *index, u16 *vid,
> > > u8 *protocol)
> > > {
> > > u32 request_pl = FIELD_PREP(PCI_DOE_DATA_OBJECT_DISC_REQ_3_INDEX,
> > > *index);
> > > u32 response_pl;
> > > - DECLARE_COMPLETION_ONSTACK(c);
> > > struct pci_doe_task task = {
> > > .prot.vid = PCI_VENDOR_ID_PCI_SIG,
> > > .prot.type = PCI_DOE_PROTOCOL_DISCOVERY,
> > > @@ -314,17 +303,13 @@ static int pci_doe_discovery(struct pci_doe_mb *doe_mb, u8 *index, u16 *vid,
> > > .request_pl_sz = sizeof(request_pl),
> > > .response_pl = &response_pl,
> > > .response_pl_sz = sizeof(response_pl),
> > > - .complete = pci_doe_task_complete,
> > > - .private = &c,
> > > };
> > > int rc;
> > >
> > > - rc = pci_doe_submit_task(doe_mb, &task);
> > > + rc = pci_doe_submit_task_wait(doe_mb, &task);
> > > if (rc < 0)
> > > return rc;
> > >
> > > - wait_for_completion(&c);
> > > -
> > > if (task.rv != sizeof(response_pl))
> > > return -EIO;
> > >
> > > @@ -376,13 +361,6 @@ static void pci_doe_xa_destroy(void *mb)
> > > xa_destroy(&doe_mb->prots);
> > > }
> > >
> > > -static void pci_doe_destroy_workqueue(void *mb)
> > > -{
> > > - struct pci_doe_mb *doe_mb = mb;
> > > -
> > > - destroy_workqueue(doe_mb->work_queue);
> > > -}
> > > -
> > > static void pci_doe_flush_mb(void *mb)
> > > {
> > > struct pci_doe_mb *doe_mb = mb;
> > > @@ -393,9 +371,6 @@ static void pci_doe_flush_mb(void *mb)
> > > /* Cancel an in progress work item, if necessary */
> > > set_bit(PCI_DOE_FLAG_CANCEL, &doe_mb->flags);
> > > wake_up(&doe_mb->wq);
> > > -
> > > - /* Flush all work items */
> > > - flush_workqueue(doe_mb->work_queue);
> > > }
> > >
> > > /**
> > > @@ -429,19 +404,6 @@ struct pci_doe_mb *pcim_doe_create_mb(struct pci_dev *pdev, u16 cap_offset)
> > > if (rc)
> > > return ERR_PTR(rc);
> > >
> > > - doe_mb->work_queue = alloc_ordered_workqueue("%s %s DOE [%x]", 0,
> > > - dev_driver_string(&pdev->dev),
> > > - pci_name(pdev),
> > > - doe_mb->cap_offset);
> > > - if (!doe_mb->work_queue) {
> > > - pci_err(pdev, "[%x] failed to allocate work queue\n",
> > > - doe_mb->cap_offset);
> > > - return ERR_PTR(-ENOMEM);
> > > - }
> > > - rc = devm_add_action_or_reset(dev, pci_doe_destroy_workqueue, doe_mb);
> > > - if (rc)
> > > - return ERR_PTR(rc);
> > > -
> > > /* Reset the mailbox by issuing an abort */
> > > rc = pci_doe_abort(doe_mb);
> > > if (rc) {
> > > @@ -496,23 +458,20 @@ bool pci_doe_supports_prot(struct pci_doe_mb *doe_mb, u16 vid, u8 type)
> > > EXPORT_SYMBOL_GPL(pci_doe_supports_prot);
> > >
> > > /**
> > > - * pci_doe_submit_task() - Submit a task to be processed by the state machine
> > > + * pci_doe_submit_task_wait() - Submit and execute a task
> > > *
> > > * @doe_mb: DOE mailbox capability to submit to
> > > - * @task: task to be queued
> > > + * @task: task to be run
> > > *
> > > - * Submit a DOE task (request/response) to the DOE mailbox to be processed.
> > > - * Returns upon queueing the task object. If the queue is full this function
> > > - * will sleep until there is room in the queue.
> > > - *
> > > - * task->complete will be called when the state machine is done processing this
> > > - * task.
> > > + * Submit and run DOE task (request/response) to the DOE mailbox to be
> > > + * processed.
> > > *
> > > * Excess data will be discarded.
> > > *
> > > - * RETURNS: 0 when task has been successfully queued, -ERRNO on error
> > > + * RETURNS: 0 when task was executed, the @task->rv holds the status
> > > + * result of the executed opertion, -ERRNO on failure to submit.
> > > */
> > > -int pci_doe_submit_task(struct pci_doe_mb *doe_mb, struct pci_doe_task *task)
> > > +int pci_doe_submit_task_wait(struct pci_doe_mb *doe_mb, struct pci_doe_task *task)
> > > {
> > > if (!pci_doe_supports_prot(doe_mb, task->prot.vid, task->prot.type))
> > > return -EINVAL;
> > > @@ -529,8 +488,8 @@ int pci_doe_submit_task(struct pci_doe_mb *doe_mb, struct pci_doe_task *task)
> > > return -EIO;
> > >
> > > task->doe_mb = doe_mb;
> > > - INIT_WORK(&task->work, doe_statemachine_work);
> > > - queue_work(doe_mb->work_queue, &task->work);
> > > + exec_task(task);
> > > +
> > > return 0;
> > > }
> > > -EXPORT_SYMBOL_GPL(pci_doe_submit_task);
> > > +EXPORT_SYMBOL_GPL(pci_doe_submit_task_wait);
> > > diff --git a/include/linux/pci-doe.h b/include/linux/pci-doe.h
> > > index ed9b4df792b8..c94122a66221 100644
> > > --- a/include/linux/pci-doe.h
> > > +++ b/include/linux/pci-doe.h
> > > @@ -30,8 +30,6 @@ struct pci_doe_mb;
> > > * @response_pl_sz: Size of the response payload (bytes)
> > > * @rv: Return value. Length of received response or error (bytes)
> > > * @complete: Called when task is complete
> > > - * @private: Private data for the consumer
> > > - * @work: Used internally by the mailbox
> > > * @doe_mb: Used internally by the mailbox
> > > *
> > > * The payload sizes and rv are specified in bytes with the following
> > > @@ -50,11 +48,6 @@ struct pci_doe_task {
> > > u32 *response_pl;
> > > size_t response_pl_sz;
> > > int rv;
> > > - void (*complete)(struct pci_doe_task *task);
> > > - void *private;
> > > -
> > > - /* No need for the user to initialize these fields */
> > > - struct work_struct work;
> > > struct pci_doe_mb *doe_mb;
> > > };
> > >
> > > @@ -72,6 +65,5 @@ struct pci_doe_task {
> > >
> > > struct pci_doe_mb *pcim_doe_create_mb(struct pci_dev *pdev, u16 cap_offset);
> > > bool pci_doe_supports_prot(struct pci_doe_mb *doe_mb, u16 vid, u8 type);
> > > -int pci_doe_submit_task(struct pci_doe_mb *doe_mb, struct pci_doe_task *task);
> > > -
> > > +int pci_doe_submit_task_wait(struct pci_doe_mb *doe_mb, struct pci_doe_task *task);
> > > #endif
> > >
> > >
> >
> > good to see that we can have a sync interface.
> > I think that we still need some methods to garantee doe_task can be
> > handled one by one in doe_mb? When more than one kernel thread are
> > going to transfer data over a same doe_mb, only one kernel thread can
> > use it and others will failed in exec_task().
> >
>
> Oh, good catch, yes, this likely needs a mutex_lock_interruptible() over
> exec_task(), or similar.

Indeed, good catch. I'll get to this today; coded up and tested.

Thanks,
Ira