Re: [RFC] vdpa: clear the device when opening/releasing it

From: Jason Wang
Date: Mon Nov 14 2022 - 21:24:23 EST


On Mon, Nov 14, 2022 at 12:58 PM Longpeng (Mike, Cloud Infrastructure
Service Product Dept.) <longpeng2@xxxxxxxxxx> wrote:
>
>
>
> 在 2022/11/14 12:02, Jason Wang 写道:
> > On Fri, Nov 11, 2022 at 11:36 PM Longpeng(Mike) <longpeng2@xxxxxxxxxx> wrote:
> >>
> >> From: Longpeng <longpeng2@xxxxxxxxxx>
> >>
> >> We should do some deeply cleanup when opening or releasing the device,
> >> e.g trigger FLR if it is PCIe device.
> >
> > Why is this needed? We're resetting at the virtio level instead of the
> > transport level.
> >
> Some existing virtio hardware needs to trigger FLR to clean some
> internal private states.

If those internal private states are virtio specific, it needs to be
cleaned upon virtio/vdpa reset instead of FLR. Otherwise it's a layer
violation.

> Otherwise, the internal logic would be confused
> when we reassign the function to another VM.
>
> The vendor driver can decide whether the device should do deep cleanup
> if we provide more information about the context. The driver can ignore
> the parameter if the device does not need to reset at the transport level.

Driver won't need to care about the transport specific stuff as it's
hidden below the vDPA bus.

Thanks

>
> > Thanks
> >
> >>
> >> Signed-off-by: Longpeng <longpeng2@xxxxxxxxxx>
> >> ---
> >> drivers/vdpa/alibaba/eni_vdpa.c | 2 +-
> >> drivers/vdpa/ifcvf/ifcvf_main.c | 2 +-
> >> drivers/vdpa/mlx5/net/mlx5_vnet.c | 2 +-
> >> drivers/vdpa/vdpa_sim/vdpa_sim.c | 2 +-
> >> drivers/vdpa/vdpa_user/vduse_dev.c | 2 +-
> >> drivers/vdpa/virtio_pci/vp_vdpa.c | 2 +-
> >> drivers/vhost/vdpa.c | 4 ++--
> >> drivers/virtio/virtio_vdpa.c | 2 +-
> >> include/linux/vdpa.h | 7 ++++---
> >> 9 files changed, 13 insertions(+), 12 deletions(-)
> >>
> >> diff --git a/drivers/vdpa/alibaba/eni_vdpa.c b/drivers/vdpa/alibaba/eni_vdpa.c
> >> index 5a09a09cca70..07215b174dd6 100644
> >> --- a/drivers/vdpa/alibaba/eni_vdpa.c
> >> +++ b/drivers/vdpa/alibaba/eni_vdpa.c
> >> @@ -226,7 +226,7 @@ static void eni_vdpa_set_status(struct vdpa_device *vdpa, u8 status)
> >> eni_vdpa_free_irq(eni_vdpa);
> >> }
> >>
> >> -static int eni_vdpa_reset(struct vdpa_device *vdpa)
> >> +static int eni_vdpa_reset(struct vdpa_device *vdpa, bool clear)
> >> {
> >> struct eni_vdpa *eni_vdpa = vdpa_to_eni(vdpa);
> >> struct virtio_pci_legacy_device *ldev = &eni_vdpa->ldev;
> >> diff --git a/drivers/vdpa/ifcvf/ifcvf_main.c b/drivers/vdpa/ifcvf/ifcvf_main.c
> >> index f9c0044c6442..b9a6ac18f358 100644
> >> --- a/drivers/vdpa/ifcvf/ifcvf_main.c
> >> +++ b/drivers/vdpa/ifcvf/ifcvf_main.c
> >> @@ -496,7 +496,7 @@ static void ifcvf_vdpa_set_status(struct vdpa_device *vdpa_dev, u8 status)
> >> ifcvf_set_status(vf, status);
> >> }
> >>
> >> -static int ifcvf_vdpa_reset(struct vdpa_device *vdpa_dev)
> >> +static int ifcvf_vdpa_reset(struct vdpa_device *vdpa_dev, bool clear)
> >> {
> >> struct ifcvf_adapter *adapter;
> >> struct ifcvf_hw *vf;
> >> diff --git a/drivers/vdpa/mlx5/net/mlx5_vnet.c b/drivers/vdpa/mlx5/net/mlx5_vnet.c
> >> index 90913365def4..6f06f9c464a3 100644
> >> --- a/drivers/vdpa/mlx5/net/mlx5_vnet.c
> >> +++ b/drivers/vdpa/mlx5/net/mlx5_vnet.c
> >> @@ -2560,7 +2560,7 @@ static void init_group_to_asid_map(struct mlx5_vdpa_dev *mvdev)
> >> mvdev->group2asid[i] = 0;
> >> }
> >>
> >> -static int mlx5_vdpa_reset(struct vdpa_device *vdev)
> >> +static int mlx5_vdpa_reset(struct vdpa_device *vdev, bool clear)
> >> {
> >> struct mlx5_vdpa_dev *mvdev = to_mvdev(vdev);
> >> struct mlx5_vdpa_net *ndev = to_mlx5_vdpa_ndev(mvdev);
> >> diff --git a/drivers/vdpa/vdpa_sim/vdpa_sim.c b/drivers/vdpa/vdpa_sim/vdpa_sim.c
> >> index b071f0d842fb..7438a89ce939 100644
> >> --- a/drivers/vdpa/vdpa_sim/vdpa_sim.c
> >> +++ b/drivers/vdpa/vdpa_sim/vdpa_sim.c
> >> @@ -504,7 +504,7 @@ static void vdpasim_set_status(struct vdpa_device *vdpa, u8 status)
> >> spin_unlock(&vdpasim->lock);
> >> }
> >>
> >> -static int vdpasim_reset(struct vdpa_device *vdpa)
> >> +static int vdpasim_reset(struct vdpa_device *vdpa, bool clear)
> >> {
> >> struct vdpasim *vdpasim = vdpa_to_sim(vdpa);
> >>
> >> diff --git a/drivers/vdpa/vdpa_user/vduse_dev.c b/drivers/vdpa/vdpa_user/vduse_dev.c
> >> index 35dceee3ed56..e5fee28233c0 100644
> >> --- a/drivers/vdpa/vdpa_user/vduse_dev.c
> >> +++ b/drivers/vdpa/vdpa_user/vduse_dev.c
> >> @@ -691,7 +691,7 @@ static void vduse_vdpa_set_config(struct vdpa_device *vdpa, unsigned int offset,
> >> /* Now we only support read-only configuration space */
> >> }
> >>
> >> -static int vduse_vdpa_reset(struct vdpa_device *vdpa)
> >> +static int vduse_vdpa_reset(struct vdpa_device *vdpa, bool clear)
> >> {
> >> struct vduse_dev *dev = vdpa_to_vduse(vdpa);
> >> int ret = vduse_dev_set_status(dev, 0);
> >> diff --git a/drivers/vdpa/virtio_pci/vp_vdpa.c b/drivers/vdpa/virtio_pci/vp_vdpa.c
> >> index d35fac5cde11..3db25b622a57 100644
> >> --- a/drivers/vdpa/virtio_pci/vp_vdpa.c
> >> +++ b/drivers/vdpa/virtio_pci/vp_vdpa.c
> >> @@ -226,7 +226,7 @@ static void vp_vdpa_set_status(struct vdpa_device *vdpa, u8 status)
> >> vp_modern_set_status(mdev, status);
> >> }
> >>
> >> -static int vp_vdpa_reset(struct vdpa_device *vdpa)
> >> +static int vp_vdpa_reset(struct vdpa_device *vdpa, bool clear)
> >> {
> >> struct vp_vdpa *vp_vdpa = vdpa_to_vp(vdpa);
> >> struct virtio_pci_modern_device *mdev = vp_vdpa_to_mdev(vp_vdpa);
> >> diff --git a/drivers/vhost/vdpa.c b/drivers/vhost/vdpa.c
> >> index 166044642fd5..fdda08cd7e7a 100644
> >> --- a/drivers/vhost/vdpa.c
> >> +++ b/drivers/vhost/vdpa.c
> >> @@ -212,7 +212,7 @@ static int vhost_vdpa_reset(struct vhost_vdpa *v)
> >>
> >> v->in_batch = 0;
> >>
> >> - return vdpa_reset(vdpa);
> >> + return vdpa_reset(vdpa, true);
> >> }
> >>
> >> static long vhost_vdpa_get_device_id(struct vhost_vdpa *v, u8 __user *argp)
> >> @@ -269,7 +269,7 @@ static long vhost_vdpa_set_status(struct vhost_vdpa *v, u8 __user *statusp)
> >> vhost_vdpa_unsetup_vq_irq(v, i);
> >>
> >> if (status == 0) {
> >> - ret = vdpa_reset(vdpa);
> >> + ret = vdpa_reset(vdpa, false);
> >> if (ret)
> >> return ret;
> >> } else
> >> diff --git a/drivers/virtio/virtio_vdpa.c b/drivers/virtio/virtio_vdpa.c
> >> index 9670cc79371d..8f6ae689547e 100644
> >> --- a/drivers/virtio/virtio_vdpa.c
> >> +++ b/drivers/virtio/virtio_vdpa.c
> >> @@ -99,7 +99,7 @@ static void virtio_vdpa_reset(struct virtio_device *vdev)
> >> {
> >> struct vdpa_device *vdpa = vd_get_vdpa(vdev);
> >>
> >> - vdpa_reset(vdpa);
> >> + vdpa_reset(vdpa, false);
> >> }
> >>
> >> static bool virtio_vdpa_notify(struct virtqueue *vq)
> >> diff --git a/include/linux/vdpa.h b/include/linux/vdpa.h
> >> index 6d0f5e4e82c2..a0b917743b66 100644
> >> --- a/include/linux/vdpa.h
> >> +++ b/include/linux/vdpa.h
> >> @@ -218,6 +218,7 @@ struct vdpa_map_file {
> >> * @status: virtio device status
> >> * @reset: Reset device
> >> * @vdev: vdpa device
> >> + * @clear: need device/function level clear or not, e.g pcie_flr.
> >> * Returns integer: success (0) or error (< 0)
> >> * @suspend: Suspend or resume the device (optional)
> >> * @vdev: vdpa device
> >> @@ -322,7 +323,7 @@ struct vdpa_config_ops {
> >> u32 (*get_vendor_id)(struct vdpa_device *vdev);
> >> u8 (*get_status)(struct vdpa_device *vdev);
> >> void (*set_status)(struct vdpa_device *vdev, u8 status);
> >> - int (*reset)(struct vdpa_device *vdev);
> >> + int (*reset)(struct vdpa_device *vdev, bool clear);
> >> int (*suspend)(struct vdpa_device *vdev);
> >> size_t (*get_config_size)(struct vdpa_device *vdev);
> >> void (*get_config)(struct vdpa_device *vdev, unsigned int offset,
> >> @@ -427,14 +428,14 @@ static inline struct device *vdpa_get_dma_dev(struct vdpa_device *vdev)
> >> return vdev->dma_dev;
> >> }
> >>
> >> -static inline int vdpa_reset(struct vdpa_device *vdev)
> >> +static inline int vdpa_reset(struct vdpa_device *vdev, bool clear)
> >> {
> >> const struct vdpa_config_ops *ops = vdev->config;
> >> int ret;
> >>
> >> down_write(&vdev->cf_lock);
> >> vdev->features_valid = false;
> >> - ret = ops->reset(vdev);
> >> + ret = ops->reset(vdev, clear);
> >> up_write(&vdev->cf_lock);
> >> return ret;
> >> }
> >> --
> >> 2.23.0
> >>
> >
> > .
>