[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAA8EJpoN3QSi72UUDV+=YkqySgXkiODUFJvwpBYfSmnkSotuZA@mail.gmail.com>
Date: Thu, 5 Dec 2024 13:30:02 +0200
From: Dmitry Baryshkov <dmitry.baryshkov@...aro.org>
To: Sibi Sankar <quic_sibis@...cinc.com>
Cc: myungjoo.ham@...sung.com, Kyungmin.park@...sung.com, cw00.choi@...sung.com,
Viresh Kumar <viresh.kumar@...aro.org>, sudeep.holla@....com, cristian.marussi@....com,
andersson@...nel.org, konrad.dybcio@...aro.org, robh+dt@...nel.org,
krzysztof.kozlowski+dt@...aro.org, linux-kernel@...r.kernel.org,
linux-arm-msm@...r.kernel.org, devicetree@...r.kernel.org,
linux-arm-kernel@...ts.infradead.org, quic_rgottimu@...cinc.com,
quic_kshivnan@...cinc.com, conor+dt@...nel.org, arm-scmi@...r.kernel.org,
Amir Vajid <avajid@...cinc.com>
Subject: Re: [PATCH V4 4/5] soc: qcom: Introduce SCMI based Memlat (Memory
Latency) governor
On Thu, 5 Dec 2024 at 12:53, Sibi Sankar <quic_sibis@...cinc.com> wrote:
>
>
>
> On 11/14/24 18:02, Dmitry Baryshkov wrote:
> > On Thu, Nov 14, 2024 at 09:43:53AM +0530, Sibi Sankar wrote:
> >>
> >>
> >> On 10/26/24 23:46, Dmitry Baryshkov wrote:
> >>> On Tue, Oct 22, 2024 at 01:48:25PM +0530, Sibi Sankar wrote:
> >>>>
> >>>>
> >>>> On 10/7/24 23:27, Dmitry Baryshkov wrote:
> >>>>> On Mon, Oct 07, 2024 at 11:40:22AM GMT, Sibi Sankar wrote:
> >
> >>>>>
> >>>>>> +};
> >>>>>> +
> >>>>>> +struct map_param_msg {
> >>>>>> + u32 hw_type;
> >>>>>> + u32 mon_idx;
> >>>>>> + u32 nr_rows;
> >>>>>> + struct map_table tbl[MAX_MAP_ENTRIES];
> >>>>>> +} __packed;
> >>>>>> +
> >>>>>> +struct node_msg {
> >>>>>> + u32 cpumask;
> >>>>>> + u32 hw_type;
> >>>>>> + u32 mon_type;
> >>>>>> + u32 mon_idx;
> >>>>>> + char mon_name[MAX_NAME_LEN];
> >>>>>> +};
> >>>>>> +
> >>>>>> +struct scalar_param_msg {
> >>>>>> + u32 hw_type;
> >>>>>> + u32 mon_idx;
> >>>>>> + u32 val;
> >>>>>> +};
> >>>>>> +
> >>>>>> +enum common_ev_idx {
> >>>>>> + INST_IDX,
> >>>>>> + CYC_IDX,
> >>>>>> + CONST_CYC_IDX,
> >>>>>> + FE_STALL_IDX,
> >>>>>> + BE_STALL_IDX,
> >>>>>> + NUM_COMMON_EVS
> >>>>>> +};
> >>>>>> +
> >>>>>> +enum grp_ev_idx {
> >>>>>> + MISS_IDX,
> >>>>>> + WB_IDX,
> >>>>>> + ACC_IDX,
> >>>>>> + NUM_GRP_EVS
> >>>>>> +};
> >>>>>> +
> >>>>>> +#define EV_CPU_CYCLES 0
> >>>>>> +#define EV_INST_RETIRED 2
> >>>>>> +#define EV_L2_D_RFILL 5
> >>>>>> +
> >>>>>> +struct ev_map_msg {
> >>>>>> + u32 num_evs;
> >>>>>> + u32 hw_type;
> >>>>>> + u32 cid[NUM_COMMON_EVS];
> >>>>>> +};
> >>>>>> +
> >>>>>> +struct cpufreq_memfreq_map {
> >>>>>> + unsigned int cpufreq_mhz;
> >>>>>> + unsigned int memfreq_khz;
> >>>>>> +};
> >>>>>> +
> >>>>>> +struct scmi_monitor_info {
> >>>>>> + struct cpufreq_memfreq_map *freq_map;
> >>>>>> + char mon_name[MAX_NAME_LEN];
> >>>>>> + u32 mon_idx;
> >>>>>> + u32 mon_type;
> >>>>>> + u32 ipm_ceil;
> >>>>>> + u32 mask;
> >>>>>> + u32 freq_map_len;
> >>>>>> +};
> >>>>>> +
> >>>>>> +struct scmi_memory_info {
> >>>>>> + struct scmi_monitor_info *monitor[MAX_MONITOR_CNT];
> >>>>>> + u32 hw_type;
> >>>>>> + int monitor_cnt;
> >>>>>> + u32 min_freq;
> >>>>>> + u32 max_freq;
> >>>>>> +};
> >>>>>> +
> >>>>>> +struct scmi_memlat_info {
> >>>>>> + struct scmi_protocol_handle *ph;
> >>>>>> + const struct qcom_generic_ext_ops *ops;
> >>>>>> + struct scmi_memory_info *memory[MAX_MEMORY_TYPES];
> >>>>>> + u32 cluster_info[NR_CPUS];
> >>>>>> + int memory_cnt;
> >>>>>> +};
> >>>>>> +
> >>>>>> +static int populate_cluster_info(u32 *cluster_info)
> >>>>>> +{
> >>>>>> + char name[MAX_NAME_LEN];
> >>>>>> + int i = 0;
> >>>>>> +
> >>>>>> + struct device_node *cn __free(device_node) = of_find_node_by_path("/cpus");
> >>>>>> + if (!cn)
> >>>>>> + return -ENODEV;
> >>>>>> +
> >>>>>> + struct device_node *map __free(device_node) = of_get_child_by_name(cn, "cpu-map");
> >>>>>> + if (!map)
> >>>>>> + return -ENODEV;
> >>>>>> +
> >>>>>> + do {
> >>>>>> + snprintf(name, sizeof(name), "cluster%d", i);
> >>>>>> + struct device_node *c __free(device_node) = of_get_child_by_name(map, name);
> >>>>>> + if (!c)
> >>>>>> + break;
> >>>>>> +
> >>>>>> + *(cluster_info + i) = of_get_child_count(c);
> >>>>>> + i++;
> >>>>>> + } while (1);
> >>>>>
> >>>>> Can you use existing API from drivers/base/arch_topology.c? If not, can
> >>>>> it be extended to support your usecase?
> >>>>
> >>>> ack. But I'm pretty sure it's going to take a while for reaching such
> >>>> an agreement so I'll drop this feature during the next re-spin.
> >>>
> >>> Why? What kind of API do you actually need? The arch_topology.c simply
> >>> exports a table of struct cpu_topology. Is it somehow different from
> >>> what you are parsing manually?
> >>
> >> yup, we had to figure out the physical id of the cpu
> >> since cpus can be disabled by the bootloader using
> >> status = "failed" property and we have to pass this
> >> onto the cpucp memlat algorithm.
> >
> > Isn't it equal to the index in the cpu_topology table?
>
> from what I see cpu_topology indexes remain unpopulated
> for cpus that are disabled since get_cpu_for_node
> ignores those?
Why do you need cpu_topology for disabled aka non-existing CPU devices?
>
> >
> >>
> >>>
> >>>>
> >>>>>
> >>>>>> +
> >>>>>> + return 0;
> >>>>>> +}
> >>>>>> +
> >>>>>> +static void populate_physical_mask(struct device_node *np, u32 *mask, u32 *cluster_info)
> >>>>>> +{
> >>>>>> + struct device_node *dev_phandle __free(device_node);
> >>>>>> + int cpu, i = 0, physical_id;
> >>>>>> +
> >>>>>> + do {
> >>>>>> + dev_phandle = of_parse_phandle(np, "cpus", i++);
> >>>>>> + cpu = of_cpu_node_to_id(dev_phandle);
> >>>>>> + if (cpu != -ENODEV) {
> >>>>>> + physical_id = topology_core_id(cpu);
> >>>>>> + for (int j = 0; j < topology_cluster_id(cpu); j++)
> >>>>>> + physical_id += *(cluster_info + j);
> >>>>>> + *mask |= BIT(physical_id);
> >>>>>> + }
> >>>>>> + } while (dev_phandle);
> >>>>>> +}
> >>>>>> +
> >>>>>> +static struct cpufreq_memfreq_map *init_cpufreq_memfreq_map(struct device *dev,
> >>>>>> + struct scmi_memory_info *memory,
> >>>>>> + struct device_node *of_node,
> >>>>>> + u32 *cnt)
> >>>>>> +{
> >>>>>> + struct device_node *tbl_np __free(device_node), *opp_np __free(device_node);
> >>>>>> + struct cpufreq_memfreq_map *tbl;
> >>>>>> + int ret, i = 0;
> >>>>>> + u32 level, len;
> >>>>>> + u64 rate;
> >>>>>> +
> >>>>>> + tbl_np = of_parse_phandle(of_node, "operating-points-v2", 0);
> >>>>>
> >>>>> Please use existing API to parse OPP tables or document a reason why it
> >>>>> can't be used.
> >>>>
> >>>> Thanks, I had them documented as opens in the coverletter. Dropped them
> >>>> since no one had any comments on it during V3. Will add them as comments
> >>>> to this driver instead.
> >>>>
> >>>> https://lore.kernel.org/lkml/20240702191440.2161623-1-quic_sibis@quicinc.com/
> >>>>
> >>>> re-copying things again:
> >>>> opp-tables are used but they don't get to be added to the scmi device
> >>>> (thus we rely on a lot of manual parsing) because the memlat client driver
> >>>> doesn't vote on these resources clocks/interconnects/power-domain
> >>>> from the kernel and some of the resources aren't modeled in the first
> >>>> place like DDR_QOS.
> >>>
> >>> As discussed offline, please consider extending the OPP to be able to
> >>> get the struct opp_table for the particular phandle. Another option
> >>> might be to change the memlat driver by having a separate device for
> >>> each monitor. This way you can use existing API to parse OPP tables and
> >>> to get necessary data from those tables.
> >>
> >> + Viresh
> >>
> >> Spoke with Viresh offline and he had stricter requirements
> >> than what you proposed. He definitely wanted the opp-tables
> >> to be assoiciated with devices at the very least and have
> >> all opp parsing logic within the opp-framework. Given that
> >> we have to model all these dummy devices just to add the
> >> tables I'll re-check the feasibility of movign the tables
> >> into the driver itself. Will move the patch series back
> >> into RFC and re-post just the vendor protocol since that's
> >> close to merge
> >
> > I don't think it's sensible to move the tables to the driver. Instead
> > adding a device per monitor sounds like a better idea.
>
> yeah, I would like to keep this in dt as well. But in order
> to be able to do that through the opp core we would need
> to put in a clock and interconnect paths so that the framework
> is able to add the table to the device.
> I should be able
> to list the scmi perf domain as the clock phandle but inorder
> to be able to convert the kbps values for the interconnect I
> would need to store the bus width and so on which is currently
> abstracted by the interconnect framework. Also in the future
> we may have to model dummy devices just to get the table parsed
> if the devices aren't modelled and controlled in the kernel.
> All of these seems to indicate that having the tables in the
> driver is a better alternative, lol.
Or fix the OPP API. I don't think that having monitor tables for the
SCMI-based platforms in the driver code is going to scale.
>
> -Sibi
>
> >
> >>>>>> + if (!tbl_np)
> >>>>>> + return ERR_PTR(-ENODEV);
> >>>>>> +
> >>>>>> + len = min(of_get_available_child_count(tbl_np), MAX_MAP_ENTRIES);
> >>>>>> + if (len == 0)
> >>>>>> + return ERR_PTR(-ENODEV);
> >>>>>> +
> >>>>>> + tbl = devm_kzalloc(dev, (len + 1) * sizeof(struct cpufreq_memfreq_map),
> >>>>>> + GFP_KERNEL);
> >>>>>> + if (!tbl)
> >>>>>> + return ERR_PTR(-ENOMEM);
> >>>>>> +
> >>>>>> + for_each_available_child_of_node(tbl_np, opp_np) {
> >>>>>> + ret = of_property_read_u64_index(opp_np, "opp-hz", 0, &rate);
> >>>>>> + if (ret < 0)
> >>>>>> + return ERR_PTR(ret);
> >>>>>> +
> >>>>>> + tbl[i].cpufreq_mhz = rate / HZ_PER_MHZ;
> >>>>>> +
> >>>>>> + if (memory->hw_type != QCOM_MEM_TYPE_DDR_QOS) {
> >>>>>> + ret = of_property_read_u64_index(opp_np, "opp-hz", 1, &rate);
> >>>>>> + if (ret < 0)
> >>>>>> + return ERR_PTR(ret);
> >>>>>> +
> >>>>>> + tbl[i].memfreq_khz = rate / HZ_PER_KHZ;
> >>>>>> + } else {
> >>>>>> + ret = of_property_read_u32(opp_np, "opp-level", &level);
> >>>>>> + if (ret < 0)
> >>>>>> + return ERR_PTR(ret);
> >>>>>> +
> >>>>>> + tbl[i].memfreq_khz = level;
> >>>>>> + }
> >>>>>> +
> >>>>>> + dev_dbg(dev, "Entry%d CPU:%u, Mem:%u\n", i, tbl[i].cpufreq_mhz, tbl[i].memfreq_khz);
> >>>>>> + i++;
> >>>>>> + }
> >>>>>> + *cnt = len;
> >>>>>> +
> >>>>>> + return tbl;
> >>>>>> +}
> >>>>>> +
> >>>>>> +static int process_scmi_memlat_of_node(struct scmi_device *sdev, struct scmi_memlat_info *info)
> >>>>>> +{
> >>>>>> + struct scmi_monitor_info *monitor;
> >>>>>> + struct scmi_memory_info *memory;
> >>>>>> + char name[MAX_NAME_LEN];
> >>>>>> + u64 memfreq[2];
> >>>>>> + int ret;
> >>>>>> +
> >>>>>> + ret = populate_cluster_info(info->cluster_info);
> >>>>>> + if (ret < 0) {
> >>>>>> + dev_err_probe(&sdev->dev, ret, "failed to populate cluster info\n");
> >>>>>> + goto err;
> >>>>>> + }
> >>>>>> +
> >>>>>> + of_node_get(sdev->dev.of_node);
> >>>>>> + do {
> >>>>>> + snprintf(name, sizeof(name), "memory-%d", info->memory_cnt);
> >>>>>> + struct device_node *memory_np __free(device_node) =
> >>>>>> + of_find_node_by_name(sdev->dev.of_node, name);
> >>>>>> +
> >>>>>> + if (!memory_np)
> >>>>>> + break;
> >>>>>> +
> >>>>>> + if (info->memory_cnt >= MAX_MEMORY_TYPES)
> >>>>>> + return dev_err_probe(&sdev->dev, -EINVAL,
> >>>>>> + "failed to parse unsupported memory type\n");
> >>>>>> +
> >>>>>> + memory = devm_kzalloc(&sdev->dev, sizeof(*memory), GFP_KERNEL);
> >>>>>> + if (!memory) {
> >>>>>> + ret = -ENOMEM;
> >>>>>> + goto err;
> >>>>>> + }
> >>>>>> +
> >>>>>> + ret = of_property_read_u32(memory_np, "qcom,memory-type", &memory->hw_type);
> >>>>>> + if (ret) {
> >>>>>> + dev_err_probe(&sdev->dev, ret, "failed to read memory type\n");
> >>>>>> + goto err;
> >>>>>> + }
> >>>>>> +
> >>>>>> + ret = of_property_read_u64_array(memory_np, "freq-table-hz", memfreq, 2);
> >>>>>> + if (ret && (ret != -EINVAL)) {
> >>>>>> + dev_err_probe(&sdev->dev, ret, "failed to read min/max freq\n");
> >>>>>> + goto err;
> >>>>>> + }
> >>>>>
> >>>>> Can we get this information from the OPP table instead?
> >>>>
> >>>> we don't list all the available ddr/llcc freqs in the opp-table
> >>>> so that we can keep the table constant across platforms.
> >>>
> >>> NO. Use opp-supported-hw to limit data to a particular platform. There
> >>> is no reason to keep min/max out of the OPP table.
> >>
> >> if we are movign the opp-tables into driver data for the reasons
> >> described above, this can probably stay?
> >
> > No. They duplicate the information that can be a part of the tables. It
> > doesn't matter if the tables are in the driver or in DT.
> >
> >
--
With best wishes
Dmitry
Powered by blists - more mailing lists