[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <87v8prtgcj.fsf@linux.ibm.com>
Date: Tue, 13 Sep 2022 10:59:56 -0500
From: Nathan Lynch <nathanl@...ux.ibm.com>
To: Michal Suchánek <msuchanek@...e.de>
Cc: Laurent Dufour <ldufour@...ux.ibm.com>,
Tyrel Datwyler <tyreld@...ux.ibm.com>,
linuxppc-dev@...ts.ozlabs.org, linux-kernel@...r.kernel.org,
Michal Hocko <mhocko@...e.com>, Lee@...sune.suse.cz,
Chun-Yi <jlee@...e.com>
Subject: Re: [PATCH] powerpc/pseries: add lparctl driver for
platform-specific functions
Michal Suchánek <msuchanek@...e.de> writes:
> On Fri, Aug 12, 2022 at 02:14:21PM -0500, Nathan Lynch wrote:
>> Laurent Dufour <ldufour@...ux.ibm.com> writes:
>> > Le 30/07/2022 à 02:04, Nathan Lynch a écrit :
>> >> +static long lparctl_get_sysparm(struct lparctl_get_system_parameter __user *argp)
>> >> +{
>> >> + struct lparctl_get_system_parameter *gsp;
>> >> + long ret;
>> >> + int fwrc;
>> >> +
>> >> + /*
>> >> + * Special case to allow user space to probe the command.
>> >> + */
>> >> + if (argp == NULL)
>> >> + return 0;
>> >> +
>> >> + gsp = memdup_user(argp, sizeof(*gsp));
>> >> + if (IS_ERR(gsp)) {
>> >> + ret = PTR_ERR(gsp);
>> >> + goto err_return;
>> >> + }
>> >> +
>> >> + ret = -EINVAL;
>> >> + if (gsp->rtas_status != 0)
>> >> + goto err_free;
>> >> +
>> >> + do {
>> >> + static_assert(sizeof(gsp->data) <= sizeof(rtas_data_buf));
>> >> +
>> >> + spin_lock(&rtas_data_buf_lock);
>> >> + memset(rtas_data_buf, 0, sizeof(rtas_data_buf));
>> >> + memcpy(rtas_data_buf, gsp->data, sizeof(gsp->data));
>> >> + fwrc = rtas_call(rtas_token("ibm,get-system-parameter"), 3, 1,
>> >> + NULL, gsp->token, __pa(rtas_data_buf),
>> >> + sizeof(gsp->data));
>> >> + if (fwrc == 0)
>> >> + memcpy(gsp->data, rtas_data_buf, sizeof(gsp->data));
>> >
>> > May be the amount of data copied out to the user space could be
>> > gsp->length. This would prevent copying 4K bytes all the time.
>> >
>> > In a more general way, the size of the RTAS buffer is quite big, and I'm
>> > wondering if all the data need to be copied back and forth to the kernel.
>> >
>> > Unless there are a high frequency of calls this doesn't make sense, and
>> > keeping the code simple might be the best way. Otherwise limiting the bytes
>> > copied could help a bit.
>>
>> This is not intended to be a high-bandwidth interface and I don't think
>> there's much of a performance concern here, so I'd rather just keep the
>> copy sizes involved constant.
>
> But that's absolutely horrible!
?
> The user wants the VPD data, all of it. And you only give one page with
> this interface.
The code here is for system parameters, which have a known maximum size,
unlike VPD. There's no code for VPD retrieval in this patch.
But I'm happy to constructively discuss how a VPD ioctl interface should
work.
> Worse, the call is not reentrant so you need to lock against other users
> calling the call while the current caller is retrieving the inidividual
> pagaes.
>
> You could do that per process, but then processes with userspace
> threading would want the data as well so you would have to save the
> arguments of the last call, and compare to arguments of any subsequent
> call to determine if you can let it pass or block.
>
> And when you do all that there will be a process that retrieves a couple
> of pages and goes out for lunch or loses interest completely, blocking
> out everyone from accessing the interface at all.
Right, the ibm,get-vpd RTAS function is tricky to expose to user space.
It needs to be called repeatedly until all data has been returned, 4KB
at a time.
Only one ibm,get-vpd sequence can be in progress at any time. If an
ibm,get-vpd sequence is begun while another sequence is already
outstanding, the first one is invalidated -- I would guess -1 or some
other error is returned on its next call.
So a new system-call level interface for VPD retrieval probably should
not expose the repeating sequence-based nature of the RTAS function to
user space, to prevent concurrent clients from interfering with each
other. That implies that the kernel should buffer the VPD results
internally; at least that's the only idea I've had so far. Open to
other suggestions.
Powered by blists - more mailing lists