Re: [RFC 2/2] firmware: arm_scmi: Add SCMI QTI Memlat vendor protocol

From: Matthias Kaehlcke
Date: Thu Nov 03 2022 - 16:02:36 EST


Hi Sibi,

On Thu, Nov 03, 2022 at 10:28:32AM +0530, Sibi Sankar wrote:
> Add support for the SCMI QTI memlat (memory latency) vendor protocol.
> The QTI memlat vendor protocol takes in several tuneables including the
> IPM ratio (Instructions Per Miss), bus bandwidth requirements and PMU
> maps to enable frequency scaling of various buses (L3/LLCC/DDR) performed
> by the memory latency governor running on the CPUSS Control Processor.
>
> Signed-off-by: Sibi Sankar <quic_sibis@xxxxxxxxxxx>
> ---
> drivers/firmware/arm_scmi/Kconfig | 10 +
> drivers/firmware/arm_scmi/Makefile | 1 +
> drivers/firmware/arm_scmi/qcom_memlat_vendor.c | 269 +++++++++++++++++++++++++
> include/linux/scmi_protocol.h | 36 ++++
> 4 files changed, 316 insertions(+)
> create mode 100644 drivers/firmware/arm_scmi/qcom_memlat_vendor.c
>
> diff --git a/drivers/firmware/arm_scmi/Kconfig b/drivers/firmware/arm_scmi/Kconfig
> index a14f65444b35..814a3fc37dc1 100644
> --- a/drivers/firmware/arm_scmi/Kconfig
> +++ b/drivers/firmware/arm_scmi/Kconfig
> @@ -136,6 +136,16 @@ config ARM_SCMI_TRANSPORT_VIRTIO_ATOMIC_ENABLE
>
> endif #ARM_SCMI_PROTOCOL
>
> +config QTI_SCMI_MEMLAT_PROTOCOL
> + tristate "Qualcomm Technologies, Inc. SCMI MEMLAT vendor Protocol"
> + depends on ARM_SCMI_PROTOCOL && QCOM_CPUCP_MBOX
> + help
> + The SCMI QTI memlat vendor protocol adds support for the frequency
> + scaling of buses (L3/LLCC/DDR) by the QTI HW memlat governor running
> + on the CPUSS Control Processor (CPUCP).
> +
> + Say Y here if you want to build this driver.
> +
> config ARM_SCMI_POWER_DOMAIN
> tristate "SCMI power domain driver"
> depends on ARM_SCMI_PROTOCOL || (COMPILE_TEST && OF)
> diff --git a/drivers/firmware/arm_scmi/Makefile b/drivers/firmware/arm_scmi/Makefile
> index 9ea86f8cc8f7..78e6d72fb9bb 100644
> --- a/drivers/firmware/arm_scmi/Makefile
> +++ b/drivers/firmware/arm_scmi/Makefile
> @@ -11,6 +11,7 @@ scmi-protocols-y = base.o clock.o perf.o power.o reset.o sensors.o system.o volt
> scmi-module-objs := $(scmi-bus-y) $(scmi-driver-y) $(scmi-protocols-y) \
> $(scmi-transport-y)
> obj-$(CONFIG_ARM_SCMI_PROTOCOL) += scmi-module.o
> +obj-$(CONFIG_QTI_SCMI_MEMLAT_PROTOCOL) += qcom_memlat_vendor.o
> obj-$(CONFIG_ARM_SCMI_POWER_DOMAIN) += scmi_pm_domain.o
> obj-$(CONFIG_ARM_SCMI_POWER_CONTROL) += scmi_power_control.o
>
> diff --git a/drivers/firmware/arm_scmi/qcom_memlat_vendor.c b/drivers/firmware/arm_scmi/qcom_memlat_vendor.c
> new file mode 100644
> index 000000000000..4b7db309e633
> --- /dev/null
> +++ b/drivers/firmware/arm_scmi/qcom_memlat_vendor.c
> @@ -0,0 +1,269 @@
> +// SPDX-License-Identifier: GPL-2.0-only
> +/*
> + * Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
> + */
> +
> +#include <linux/module.h>
> +#include <linux/scmi_protocol.h>
> +
> +#include "protocols.h"
> +
> +#define MAX_MAP_ENTRIES 14
> +#define MAX_PMU_ENTRIES 24
> +
> +enum scmi_memlat_protocol_cmd {
> + MEMLAT_SET_CPU_GROUP = 0x10,
> + MEMLAT_SET_MONITOR = 0x11,
> + MEMLAT_COMMON_PMU_MAP = 0x12,
> + MEMLAT_MON_PMU_MAP = 0x13,
> + MEMLAT_IPM_RATIO = 0x14,
> + MEMLAT_STALL_RATIO = 0x15,
> + MEMLAT_SAMPLE_MS = 0x18,
> + MEMLAT_MON_FREQ_MAP = 0x19,
> + MEMLAT_START_MONITOR = 0x1c,
> + MEMLAT_STOP_MONITOR = 0x1d,
> +};
> +
> +struct node_msg {
> + u32 cpumask;
> + u32 mon_type;
> +};
> +
> +struct scalar_param_msg {
> + u32 cpumask;
> + u32 mon_type;
> + u32 val;
> +};
> +
> +struct map_table {
> + u32 v1;
> + u32 v2;
> +};
> +
> +struct map_param_msg {
> + u32 cpumask;
> + u32 mon_type;
> + u32 nr_rows;
> + struct map_table tbl[MAX_MAP_ENTRIES];
> +};
> +
> +struct pmu_map_msg {
> + u32 cpumask;
> + u32 mon_type;
> + u32 nr_entries;
> + u32 pmu[MAX_PMU_ENTRIES];
> +};
> +
> +static int scmi_set_cpugrp_mon(const struct scmi_protocol_handle *ph,
> + u32 cpus_mpidr, u32 mon_type, u32 msg_id)
> +{
> + int ret = 0;

no need to initialize

> + struct scmi_xfer *t;
> + struct node_msg *msg;
> +
> + ret = ph->xops->xfer_get_init(ph, msg_id, sizeof(*msg), sizeof(*msg), &t);
> + if (ret)
> + return ret;
> +
> + msg = t->tx.buf;
> + msg->cpumask = cpu_to_le32(cpus_mpidr);
> + msg->mon_type = cpu_to_le32(mon_type);
> + ret = ph->xops->do_xfer(ph, t);
> + ph->xops->xfer_put(ph, t);
> +
> + return ret;
> +}
> +
> +static int scmi_set_mon(const struct scmi_protocol_handle *ph, u32 cpus_mpidr, u32 mon_type)
> +{
> + return scmi_set_cpugrp_mon(ph, cpus_mpidr, mon_type, MEMLAT_SET_MONITOR);
> +}
> +
> +static int scmi_set_cpu_grp(const struct scmi_protocol_handle *ph, u32 cpus_mpidr, u32 mon_type)
> +{
> + return scmi_set_cpugrp_mon(ph, cpus_mpidr, mon_type, MEMLAT_SET_CPU_GROUP);
> +}
> +
> +static int scmi_send_pmu_map_command(const struct scmi_protocol_handle *ph, u32 cpus_mpidr,
> + u32 mon_type, u32 nr_entries, void *buf, u32 msg_id)
> +{
> + u32 *dst;
> + int ret, i = 0;

initialization is not needed

> + struct scmi_xfer *t;
> + struct pmu_map_msg *msg;
> + struct map_table *src = buf;
> +
> + if (nr_entries > MAX_PMU_ENTRIES)
> + return -EINVAL;
> +
> + ret = ph->xops->xfer_get_init(ph, msg_id, sizeof(*msg), sizeof(*msg), &t);
> + if (ret)
> + return ret;
> +
> + msg = t->tx.buf;
> + msg->cpumask = cpu_to_le32(cpus_mpidr);
> + msg->mon_type = cpu_to_le32(mon_type);

The above 7 lines are a recurring pattern. Might be worth to have a wrapper for
it. The datatype of 'msg' varies though, so it would have to be a macro :(

> + msg->nr_entries = cpu_to_le32(nr_entries);
> + dst = msg->pmu;
> +
> + for (i = 0; i < nr_entries; i++)
> + dst[i] = cpu_to_le32(src[i].v2);
> +
> + ret = ph->xops->do_xfer(ph, t);
> + ph->xops->xfer_put(ph, t);
> + return ret;

This above 3 lines also recurring, consider a wrapper. With that the above
would become:

return scmi_do_xfer(ph, t);
> +}
> +
> +static int scmi_common_pmu_map(const struct scmi_protocol_handle *ph, u32 cpus_mpidr,
> + u32 mon_type, u32 nr_entries, void *buf)
> +{
> + return scmi_send_pmu_map_command(ph, cpus_mpidr, mon_type, nr_entries,
> + buf, MEMLAT_COMMON_PMU_MAP);
> +}
> +
> +static int scmi_mon_pmu_map(const struct scmi_protocol_handle *ph, u32 cpus_mpidr,
> + u32 mon_type, u32 nr_entries, void *buf)
> +{
> + return scmi_send_pmu_map_command(ph, cpus_mpidr, mon_type, nr_entries,
> + buf, MEMLAT_MON_PMU_MAP);
> +}
> +
> +static int scmi_freq_map(const struct scmi_protocol_handle *ph, u32 cpus_mpidr,
> + u32 mon_type, u32 nr_rows, void *buf)
> +{
> + int ret, i = 0;

initialization is unnecessary

> + struct scmi_xfer *t;
> + struct map_param_msg *msg;
> + struct map_table *tbl, *src = buf;
> +
> + if (nr_rows > MAX_MAP_ENTRIES)
> + return -EINVAL;
> +
> + ret = ph->xops->xfer_get_init(ph, MEMLAT_MON_FREQ_MAP, sizeof(*msg),
> + sizeof(*msg), &t);
> + if (ret)
> + return ret;
> +
> + msg = t->tx.buf;
> + msg->cpumask = cpu_to_le32(cpus_mpidr);
> + msg->mon_type = cpu_to_le32(mon_type);
> + msg->nr_rows = cpu_to_le32(nr_rows);
> + tbl = msg->tbl;
> +
> + for (i = 0; i < nr_rows; i++) {
> + tbl[i].v1 = cpu_to_le32(src[i].v1);
> + tbl[i].v2 = cpu_to_le32(src[i].v2);
> + }
> +
> + ret = ph->xops->do_xfer(ph, t);
> + ph->xops->xfer_put(ph, t);
> + return ret;
> +}
> +
> +static int scmi_set_tunable(const struct scmi_protocol_handle *ph, u32 cpus_mpidr,
> + u32 msg_id, u32 mon_type, u32 val)
> +{
> + int ret = 0;

drop initialization

> + struct scmi_xfer *t;
> + struct scalar_param_msg *msg;
> +
> + ret = ph->xops->xfer_get_init(ph, msg_id, sizeof(*msg), sizeof(*msg), &t);
> + if (ret)
> + return ret;
> +
> + msg = t->tx.buf;
> + msg->cpumask = cpu_to_le32(cpus_mpidr);
> + msg->mon_type = cpu_to_le32(mon_type);
> + msg->val = cpu_to_le32(val);
> + ret = ph->xops->do_xfer(ph, t);
> + ph->xops->xfer_put(ph, t);
> +
> + return ret;
> +}
> +
> +static int scmi_ipm_ratio(const struct scmi_protocol_handle *ph, u32 cpus_mpidr,
> + u32 mon_type, u32 val)
> +{
> + return scmi_set_tunable(ph, cpus_mpidr, MEMLAT_IPM_RATIO, mon_type, val);
> +}
> +
> +static int scmi_stall_ratio(const struct scmi_protocol_handle *ph, u32 cpus_mpidr,
> + u32 mon_type, u32 val)
> +{
> + return scmi_set_tunable(ph, cpus_mpidr, MEMLAT_STALL_RATIO, mon_type, val);
> +}
> +
> +static int scmi_sample_ms(const struct scmi_protocol_handle *ph, u32 cpus_mpidr,
> + u32 mon_type, u32 val)
> +{
> + return scmi_set_tunable(ph, cpus_mpidr, MEMLAT_SAMPLE_MS, mon_type, val);
> +}
> +
> +static int scmi_send_start_stop(const struct scmi_protocol_handle *ph,
> + u32 cpus_mpidr, u32 mon_type, u32 msg_id)
> +{
> + int ret = 0;

drop init

> + struct scmi_xfer *t;
> + struct scalar_param_msg *msg;
> +
> + ret = ph->xops->xfer_get_init(ph, msg_id, sizeof(*msg), sizeof(*msg), &t);
> + if (ret)
> + return ret;
> +
> + msg = t->tx.buf;
> + msg->cpumask = cpu_to_le32(cpus_mpidr);
> + msg->mon_type = cpu_to_le32(mon_type);
> + ret = ph->xops->do_xfer(ph, t);
> + ph->xops->xfer_put(ph, t);
> +
> + return ret;
> +}
> +
> +static int scmi_stop_mon(const struct scmi_protocol_handle *ph, u32 cpus_mpidr, u32 mon_type)
> +{
> + return scmi_send_start_stop(ph, cpus_mpidr, mon_type, MEMLAT_STOP_MONITOR);
> +}
> +
> +static int scmi_start_mon(const struct scmi_protocol_handle *ph, u32 cpus_mpidr, u32 mon_type)
> +{
> + return scmi_send_start_stop(ph, cpus_mpidr, mon_type, MEMLAT_START_MONITOR);
> +}
> +
> +static struct scmi_vendor_memlat_ops memlat_ops = {
> + .set_cpu_grp = scmi_set_cpu_grp,
> + .freq_map = scmi_freq_map,
> + .set_mon = scmi_set_mon,
> + .common_pmu_map = scmi_common_pmu_map,
> + .mon_pmu_map = scmi_mon_pmu_map,
> + .ipm_ratio = scmi_ipm_ratio,
> + .stall_ratio = scmi_stall_ratio,
> + .sample_ms = scmi_sample_ms,
> + .start_monitor = scmi_start_mon,
> + .stop_monitor = scmi_stop_mon,
> +};
> +
> +static int scmi_vendor_memlat_protocol_init(const struct scmi_protocol_handle *ph)
> +{
> + int ret;
> + u32 version;
> +
> + ret = ph->xops->version_get(ph, &version);
> + if (ret)
> + return ret;
> +
> + dev_dbg(ph->dev, "Memlat Version %d.%d\n",
> + PROTOCOL_REV_MAJOR(version), PROTOCOL_REV_MINOR(version));
> +
> + return 0;
> +}
> +
> +static const struct scmi_protocol scmi_vendor_memlat = {
> + .id = SCMI_VENDOR_PROTOCOL_MEMLAT,
> + .owner = THIS_MODULE,
> + .instance_init = &scmi_vendor_memlat_protocol_init,
> + .ops = &memlat_ops,
> +};
> +module_scmi_protocol(scmi_vendor_memlat);
> +
> +MODULE_DESCRIPTION("Qualcomm Technologies, Inc. SCMI Memlat Protocol");
> +MODULE_LICENSE("GPL");
> diff --git a/include/linux/scmi_protocol.h b/include/linux/scmi_protocol.h
> index 4f765bc788ff..57abb5be45c9 100644
> --- a/include/linux/scmi_protocol.h
> +++ b/include/linux/scmi_protocol.h
> @@ -677,6 +677,40 @@ struct scmi_powercap_proto_ops {
> };
>
> /**
> + * struct scmi_vendor_memlat_ops - represents the various operations provided
> + * by SCMI QTI HW Memlat Vendor Protocol
> + *
> + * @cpu_grp: set the cpugrp
> + * @set_mon: set the supported monitors
> + * @common_pmu_map: sets the common PMU map supported by governor
> + * @mon_pmu_map: sets the additional PMU map supported by governor
> + * @ipm_ratio: sets the ratio_ceil needed for hw memlat governor
> + * @stall_ratio: sets the stall_floor needed for hw memlat governor
> + * @sample_ms: sets the poll iterval of the governor
> + * @freq_map: sets the freq_map of the governor
> + * @start_mon: starts the monitor in firmware
> + * @stop_mon: stops the monitor in firmware
> + */
> +struct scmi_vendor_memlat_ops {
> + int (*set_cpu_grp)(const struct scmi_protocol_handle *ph, u32 cpus_mpidr, u32 mon_type);
> + int (*set_mon)(const struct scmi_protocol_handle *ph, u32 cpus_mpidr, u32 mon_type);
> + int (*common_pmu_map)(const struct scmi_protocol_handle *ph, u32 cpus_mpidr, u32 mon_type,
> + u32 nr_rows, void *buf);
> + int (*mon_pmu_map)(const struct scmi_protocol_handle *ph, u32 cpus_mpidr, u32 mon_type,
> + u32 nr_rows, void *buf);
> + int (*ipm_ratio)(const struct scmi_protocol_handle *ph, u32 cpus_mpidr,
> + u32 mon_type, u32 val);
> + int (*stall_ratio)(const struct scmi_protocol_handle *ph, u32 cpus_mpidr,
> + u32 mon_type, u32 val);
> + int (*sample_ms)(const struct scmi_protocol_handle *ph, u32 cpus_mpidr,
> + u32 mon_type, u32 val);
> + int (*freq_map)(const struct scmi_protocol_handle *ph, u32 cpus_mpidr, u32 mon_type,
> + u32 nr_rows, void *buf);
> + int (*start_monitor)(const struct scmi_protocol_handle *ph, u32 cpus_mpidr, u32 mon_type);
> + int (*stop_monitor)(const struct scmi_protocol_handle *ph, u32 cpus_mpidr, u32 mon_type);
> +};
> +
> +/**
> * struct scmi_notify_ops - represents notifications' operations provided by
> * SCMI core
> * @devm_event_notifier_register: Managed registration of a notifier_block for
> @@ -785,6 +819,8 @@ enum scmi_std_protocol {
> SCMI_PROTOCOL_POWERCAP = 0x18,
> };
>
> +#define SCMI_VENDOR_PROTOCOL_MEMLAT 0x80
> +
> enum scmi_system_events {
> SCMI_SYSTEM_SHUTDOWN,
> SCMI_SYSTEM_COLDRESET,
> --
> 2.7.4
>