lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <a9814644-96e3-456f-90b7-8705a102c938@amd.com>
Date: Tue, 26 Aug 2025 12:18:37 -0500
From: Mario Limonciello <mario.limonciello@....com>
To: Lizhi Hou <lizhi.hou@....com>, ogabbay@...nel.org,
 quic_jhugo@...cinc.com, jacek.lawrynowicz@...ux.intel.com,
 dri-devel@...ts.freedesktop.org
Cc: linux-kernel@...r.kernel.org, max.zhen@....com, sonal.santan@....com
Subject: Re: [PATCH V1] accel/amdxdna: Add ioctl DRM_IOCTL_AMDXDNA_GET_ARRAY

On 8/25/2025 11:48 PM, Lizhi Hou wrote:
> 
> On 8/25/25 14:28, Mario Limonciello wrote:
>> On 8/22/2025 12:23 PM, Lizhi Hou wrote:
>>> Add interface for applications to get information array. The application
>>> provides a buffer pointer along with information type, maximum number of
>>> entries and maximum size of each entry. The buffer may also contain 
>>> match
>>> conditions based on the information type. After the ioctl completes, the
>>> actual number of entries and entry size are returned.
>>>
>>> Signed-off-by: Lizhi Hou <lizhi.hou@....com>
>>
>> How does userspace discover whether or not the new IOCTL call is 
>> supported?  Just a test call?
> The kernel header version will be used to determine whether the 
> application which uses new IOCTL will be compiled or not.
> 

But it's not actually an application compile time decision, it's a 
runtime decision.  IE I can compile an application with the headers on 
kernel 6.18 that has this, but if I try to run it on 6.15 it's going to 
barf.

To some extent that comes with the territory, but I'm wondering if a 
better solution going forward would be for there to be a dedicated 
version command that you bump.

That way applications can call the version command and if it fails you 
know it's an old version pre-version command (and likely pre-this 
command).  If it responds with a valid number you can know what commands 
went with what number.

> Thanks,
> 
> Lizhi
> 
>>
>>> ---
>>>   drivers/accel/amdxdna/aie2_pci.c        | 114 ++++++++++++++++++------
>>>   drivers/accel/amdxdna/amdxdna_pci_drv.c |  21 +++++
>>>   drivers/accel/amdxdna/amdxdna_pci_drv.h |   1 +
>>>   include/uapi/drm/amdxdna_accel.h        | 109 ++++++++++++++++++++++
>>>   4 files changed, 220 insertions(+), 25 deletions(-)
>>>
>>> diff --git a/drivers/accel/amdxdna/aie2_pci.c b/drivers/accel/ 
>>> amdxdna/aie2_pci.c
>>> index 16ac0cab4f44..b8bfc0700798 100644
>>> --- a/drivers/accel/amdxdna/aie2_pci.c
>>> +++ b/drivers/accel/amdxdna/aie2_pci.c
>>> @@ -785,10 +785,11 @@ static int aie2_get_clock_metadata(struct 
>>> amdxdna_client *client,
>>>     static int aie2_hwctx_status_cb(struct amdxdna_hwctx *hwctx, void 
>>> *arg)
>>>   {
>>> -    struct amdxdna_drm_query_hwctx __user *buf, *tmp __free(kfree) = 
>>> NULL;
>>> -    struct amdxdna_drm_get_info *get_info_args = arg;
>>> +    struct amdxdna_drm_hwctx_entry __user *buf, *tmp __free(kfree) = 
>>> NULL;
>>> +    struct amdxdna_drm_get_array *array_args = arg;
>>> +    u32 size;
>>>   -    if (get_info_args->buffer_size < sizeof(*tmp))
>>> +    if (!array_args->num_element)
>>>           return -EINVAL;
>>>         tmp = kzalloc(sizeof(*tmp), GFP_KERNEL);
>>> @@ -801,14 +802,23 @@ static int aie2_hwctx_status_cb(struct 
>>> amdxdna_hwctx *hwctx, void *arg)
>>>       tmp->num_col = hwctx->num_col;
>>>       tmp->command_submissions = hwctx->priv->seq;
>>>       tmp->command_completions = hwctx->priv->completed;
>>> -
>>> -    buf = u64_to_user_ptr(get_info_args->buffer);
>>> -
>>> -    if (copy_to_user(buf, tmp, sizeof(*tmp)))
>>> +    tmp->pasid = hwctx->client->pasid;
>>> +    tmp->priority = hwctx->qos.priority;
>>> +    tmp->gops = hwctx->qos.gops;
>>> +    tmp->fps = hwctx->qos.fps;
>>> +    tmp->dma_bandwidth = hwctx->qos.dma_bandwidth;
>>> +    tmp->latency = hwctx->qos.latency;
>>> +    tmp->frame_exec_time = hwctx->qos.frame_exec_time;
>>> +    tmp->state = AMDXDNA_HWCTX_STATE_ACTIVE;
>>> +
>>> +    buf = u64_to_user_ptr(array_args->buffer);
>>> +    size = min(sizeof(*tmp), array_args->element_size);
>>> +
>>> +    if (copy_to_user(buf, tmp, size))
>>>           return -EFAULT;
>>>   -    get_info_args->buffer += sizeof(*tmp);
>>> -    get_info_args->buffer_size -= sizeof(*tmp);
>>> +    array_args->buffer += size;
>>> +    array_args->num_element--;
>>>         return 0;
>>>   }
>>> @@ -816,23 +826,24 @@ static int aie2_hwctx_status_cb(struct 
>>> amdxdna_hwctx *hwctx, void *arg)
>>>   static int aie2_get_hwctx_status(struct amdxdna_client *client,
>>>                    struct amdxdna_drm_get_info *args)
>>>   {
>>> +    struct amdxdna_drm_get_array array_args;
>>>       struct amdxdna_dev *xdna = client->xdna;
>>> -    struct amdxdna_drm_get_info info_args;
>>>       struct amdxdna_client *tmp_client;
>>>       int ret;
>>>         drm_WARN_ON(&xdna->ddev, !mutex_is_locked(&xdna->dev_lock));
>>>   -    info_args.buffer = args->buffer;
>>> -    info_args.buffer_size = args->buffer_size;
>>> -
>>> +    array_args.element_size = sizeof(struct amdxdna_drm_query_hwctx);
>>> +    array_args.buffer = args->buffer;
>>> +    array_args.num_element = args->buffer_size / 
>>> array_args.element_size;
>>>       list_for_each_entry(tmp_client, &xdna->client_list, node) {
>>> -        ret = amdxdna_hwctx_walk(tmp_client, &info_args, 
>>> aie2_hwctx_status_cb);
>>> +        ret = amdxdna_hwctx_walk(tmp_client, &array_args,
>>> +                     aie2_hwctx_status_cb);
>>>           if (ret)
>>>               break;
>>>       }
>>>   -    args->buffer_size = (u32)(info_args.buffer - args->buffer);
>>> +    args->buffer_size -= (u32)(array_args.buffer - args->buffer);
>>>       return ret;
>>>   }
>>>   @@ -876,6 +887,58 @@ static int aie2_get_info(struct amdxdna_client 
>>> *client, struct amdxdna_drm_get_i
>>>       return ret;
>>>   }
>>>   +static int aie2_query_ctx_status_array(struct amdxdna_client *client,
>>> +                       struct amdxdna_drm_get_array *args)
>>> +{
>>> +    struct amdxdna_drm_get_array array_args;
>>> +    struct amdxdna_dev *xdna = client->xdna;
>>> +    struct amdxdna_client *tmp_client;
>>> +    int ret;
>>> +
>>> +    drm_WARN_ON(&xdna->ddev, !mutex_is_locked(&xdna->dev_lock));
>>> +
>>> +    array_args.element_size = min(args->element_size,
>>> +                      sizeof(struct amdxdna_drm_hwctx_entry));
>>> +    array_args.buffer = args->buffer;
>>> +    array_args.num_element = args->num_element * args->element_size /
>>> +                array_args.element_size;
>>> +    list_for_each_entry(tmp_client, &xdna->client_list, node) {
>>> +        ret = amdxdna_hwctx_walk(tmp_client, &array_args,
>>> +                     aie2_hwctx_status_cb);
>>> +        if (ret)
>>> +            break;
>>> +    }
>>> +
>>> +    args->element_size = array_args.element_size;
>>> +    args->num_element = (u32)((array_args.buffer - args->buffer) /
>>> +                  args->element_size);
>>> +
>>> +    return ret;
>>> +}
>>> +
>>> +static int aie2_get_array(struct amdxdna_client *client,
>>> +              struct amdxdna_drm_get_array *args)
>>> +{
>>> +    struct amdxdna_dev *xdna = client->xdna;
>>> +    int ret, idx;
>>> +
>>> +    if (!drm_dev_enter(&xdna->ddev, &idx))
>>> +        return -ENODEV;
>>> +
>>> +    switch (args->param) {
>>> +    case DRM_AMDXDNA_HW_CONTEXT_ALL:
>>> +        ret = aie2_query_ctx_status_array(client, args);
>>> +        break;
>>> +    default:
>>> +        XDNA_ERR(xdna, "Not supported request parameter %u", args- 
>>> >param);
>>> +        ret = -EOPNOTSUPP;
>>> +    }
>>> +    XDNA_DBG(xdna, "Got param %d", args->param);
>>> +
>>> +    drm_dev_exit(idx);
>>> +    return ret;
>>> +}
>>> +
>>>   static int aie2_set_power_mode(struct amdxdna_client *client,
>>>                      struct amdxdna_drm_set_state *args)
>>>   {
>>> @@ -925,15 +988,16 @@ static int aie2_set_state(struct amdxdna_client 
>>> *client,
>>>   }
>>>     const struct amdxdna_dev_ops aie2_ops = {
>>> -    .init           = aie2_init,
>>> -    .fini           = aie2_fini,
>>> -    .resume         = aie2_hw_resume,
>>> -    .suspend        = aie2_hw_suspend,
>>> -    .get_aie_info   = aie2_get_info,
>>> -    .set_aie_state    = aie2_set_state,
>>> -    .hwctx_init     = aie2_hwctx_init,
>>> -    .hwctx_fini     = aie2_hwctx_fini,
>>> -    .hwctx_config   = aie2_hwctx_config,
>>> -    .cmd_submit     = aie2_cmd_submit,
>>> +    .init = aie2_init,
>>> +    .fini = aie2_fini,
>>> +    .resume = aie2_hw_resume,
>>> +    .suspend = aie2_hw_suspend,
>>> +    .get_aie_info = aie2_get_info,
>>> +    .set_aie_state = aie2_set_state,
>>> +    .hwctx_init = aie2_hwctx_init,
>>> +    .hwctx_fini = aie2_hwctx_fini,
>>> +    .hwctx_config = aie2_hwctx_config,
>>> +    .cmd_submit = aie2_cmd_submit,
>>>       .hmm_invalidate = aie2_hmm_invalidate,
>>> +    .get_array = aie2_get_array,
>>>   };
>>> diff --git a/drivers/accel/amdxdna/amdxdna_pci_drv.c b/drivers/accel/ 
>>> amdxdna/amdxdna_pci_drv.c
>>> index 8ef5e4f27f5e..ee89485299bc 100644
>>> --- a/drivers/accel/amdxdna/amdxdna_pci_drv.c
>>> +++ b/drivers/accel/amdxdna/amdxdna_pci_drv.c
>>> @@ -164,6 +164,26 @@ static int amdxdna_drm_get_info_ioctl(struct 
>>> drm_device *dev, void *data, struct
>>>       return ret;
>>>   }
>>>   +static int amdxdna_drm_get_array_ioctl(struct drm_device *dev, 
>>> void *data,
>>> +                       struct drm_file *filp)
>>> +{
>>> +    struct amdxdna_client *client = filp->driver_priv;
>>> +    struct amdxdna_dev *xdna = to_xdna_dev(dev);
>>> +    struct amdxdna_drm_get_array *args = data;
>>> +    int ret;
>>> +
>>> +    if (!xdna->dev_info->ops->get_array)
>>> +        return -EOPNOTSUPP;
>>> +
>>> +    if (args->pad || !args->num_element)
>>> +        return -EINVAL;
>>> +
>>> +    mutex_lock(&xdna->dev_lock);
>>> +    ret = xdna->dev_info->ops->get_array(client, args);
>>> +    mutex_unlock(&xdna->dev_lock);
>>> +    return ret;
>>> +}
>>> +
>>>   static int amdxdna_drm_set_state_ioctl(struct drm_device *dev, void 
>>> *data, struct drm_file *filp)
>>>   {
>>>       struct amdxdna_client *client = filp->driver_priv;
>>> @@ -195,6 +215,7 @@ static const struct drm_ioctl_desc 
>>> amdxdna_drm_ioctls[] = {
>>>       DRM_IOCTL_DEF_DRV(AMDXDNA_EXEC_CMD, 
>>> amdxdna_drm_submit_cmd_ioctl, 0),
>>>       /* AIE hardware */
>>>       DRM_IOCTL_DEF_DRV(AMDXDNA_GET_INFO, amdxdna_drm_get_info_ioctl, 
>>> 0),
>>> +    DRM_IOCTL_DEF_DRV(AMDXDNA_GET_ARRAY, 
>>> amdxdna_drm_get_array_ioctl, 0),
>>>       DRM_IOCTL_DEF_DRV(AMDXDNA_SET_STATE, 
>>> amdxdna_drm_set_state_ioctl, DRM_ROOT_ONLY),
>>>   };
>>>   diff --git a/drivers/accel/amdxdna/amdxdna_pci_drv.h b/drivers/ 
>>> accel/amdxdna/amdxdna_pci_drv.h
>>> index b6b3b424d1d5..72d6696d49da 100644
>>> --- a/drivers/accel/amdxdna/amdxdna_pci_drv.h
>>> +++ b/drivers/accel/amdxdna/amdxdna_pci_drv.h
>>> @@ -58,6 +58,7 @@ struct amdxdna_dev_ops {
>>>       int (*cmd_submit)(struct amdxdna_hwctx *hwctx, struct 
>>> amdxdna_sched_job *job, u64 *seq);
>>>       int (*get_aie_info)(struct amdxdna_client *client, struct 
>>> amdxdna_drm_get_info *args);
>>>       int (*set_aie_state)(struct amdxdna_client *client, struct 
>>> amdxdna_drm_set_state *args);
>>> +    int (*get_array)(struct amdxdna_client *client, struct 
>>> amdxdna_drm_get_array *args);
>>>   };
>>>     /*
>>> diff --git a/include/uapi/drm/amdxdna_accel.h b/include/uapi/drm/ 
>>> amdxdna_accel.h
>>> index ce523e9ccc52..e19e4cd04ffa 100644
>>> --- a/include/uapi/drm/amdxdna_accel.h
>>> +++ b/include/uapi/drm/amdxdna_accel.h
>>> @@ -34,6 +34,7 @@ enum amdxdna_drm_ioctl_id {
>>>       DRM_AMDXDNA_EXEC_CMD,
>>>       DRM_AMDXDNA_GET_INFO,
>>>       DRM_AMDXDNA_SET_STATE,
>>> +    DRM_AMDXDNA_GET_ARRAY = 10,
>>>   };
>>>     /**
>>> @@ -455,6 +456,110 @@ struct amdxdna_drm_get_info {
>>>       __u64 buffer; /* in/out */
>>>   };
>>>   +#define AMDXDNA_HWCTX_STATE_IDLE    0
>>> +#define AMDXDNA_HWCTX_STATE_ACTIVE    1
>>> +
>>> +/**
>>> + * struct amdxdna_drm_hwctx_entry - The hardware context array entry
>>> + */
>>> +struct amdxdna_drm_hwctx_entry {
>>> +    /** @context_id: Context ID. */
>>> +    __u32 context_id;
>>> +    /** @start_col: Start AIE array column assigned to context. */
>>> +    __u32 start_col;
>>> +    /** @num_col: Number of AIE array columns assigned to context. */
>>> +    __u32 num_col;
>>> +    /** @hwctx_id: The real hardware context id. */
>>> +    __u32 hwctx_id;
>>> +    /** @pid: ID of process which created this context. */
>>> +    __s64 pid;
>>> +    /** @command_submissions: Number of commands submitted. */
>>> +    __u64 command_submissions;
>>> +    /** @command_completions: Number of commands completed. */
>>> +    __u64 command_completions;
>>> +    /** @migrations: Number of times been migrated. */
>>> +    __u64 migrations;
>>> +    /** @preemptions: Number of times been preempted. */
>>> +    __u64 preemptions;
>>> +    /** @errors: Number of errors happened. */
>>> +    __u64 errors;
>>> +    /** @priority: Context priority. */
>>> +    __u64 priority;
>>> +    /** @heap_usage: Usage of device heap buffer. */
>>> +    __u64 heap_usage;
>>> +    /** @suspensions: Number of times been suspended. */
>>> +    __u64 suspensions;
>>> +    /**
>>> +     * @state: Context state.
>>> +     * %AMDXDNA_HWCTX_STATE_IDLE
>>> +     * %AMDXDNA_HWCTX_STATE_ACTIVE
>>> +     */
>>> +    __u32 state;
>>> +    /** @pasid: PASID been bound. */
>>> +    __u32 pasid;
>>> +    /** @gops: Giga operations per second. */
>>> +    __u32 gops;
>>> +    /** @fps: Frames per second. */
>>> +    __u32 fps;
>>> +    /** @dma_bandwidth: DMA bandwidth. */
>>> +    __u32 dma_bandwidth;
>>> +    /** @latency: Frame response latency. */
>>> +    __u32 latency;
>>> +    /** @frame_exec_time: Frame execution time. */
>>> +    __u32 frame_exec_time;
>>> +    /** @txn_op_idx: Index of last control code executed. */
>>> +    __u32 txn_op_idx;
>>> +    /** @ctx_pc: Program counter. */
>>> +    __u32 ctx_pc;
>>> +    /** @fatal_error_type: Fatal error type if context crashes. */
>>> +    __u32 fatal_error_type;
>>> +    /** @fatal_error_exception_type: Firmware exception type. */
>>> +    __u32 fatal_error_exception_type;
>>> +    /** @fatal_error_exception_pc: Firmware exception program 
>>> counter. */
>>> +    __u32 fatal_error_exception_pc;
>>> +    /** @fatal_error_app_module: Exception module name. */
>>> +    __u32 fatal_error_app_module;
>>> +};
>>> +
>>> +#define DRM_AMDXDNA_HW_CONTEXT_ALL    0
>>> +
>>> +/**
>>> + * struct amdxdna_drm_get_array - Get information array.
>>> + */
>>> +struct amdxdna_drm_get_array {
>>> +    /**
>>> +     * @param:
>>> +     *
>>> +     * Supported params:
>>> +     *
>>> +     * %DRM_AMDXDNA_HW_CONTEXT_ALL:
>>> +     * Returns all created hardware contexts.
>>> +     */
>>> +    __u32 param;
>>> +    /**
>>> +     * @element_size:
>>> +     *
>>> +     * Specifies maximum element size and returns the actual element 
>>> size.
>>> +     */
>>> +    __u32 element_size;
>>> +    /**
>>> +     * @num_element:
>>> +     *
>>> +     * Specifies maximum number of elements and returns the actual 
>>> number
>>> +     * of elements.
>>> +     */
>>> +    __u32 num_element; /* in/out */
>>> +    /** @pad: MBZ */
>>> +    __u32 pad;
>>> +    /**
>>> +     * @buffer:
>>> +     *
>>> +     * Specifies the match conditions and returns the matched 
>>> information
>>> +     * array.
>>> +     */
>>> +    __u64 buffer;
>>> +};
>>> +
>>>   enum amdxdna_drm_set_param {
>>>       DRM_AMDXDNA_SET_POWER_MODE,
>>>       DRM_AMDXDNA_WRITE_AIE_MEM,
>>> @@ -519,6 +624,10 @@ struct amdxdna_drm_set_power_mode {
>>>       DRM_IOWR(DRM_COMMAND_BASE + DRM_AMDXDNA_SET_STATE, \
>>>            struct amdxdna_drm_set_state)
>>>   +#define DRM_IOCTL_AMDXDNA_GET_ARRAY \
>>> +    DRM_IOWR(DRM_COMMAND_BASE + DRM_AMDXDNA_GET_ARRAY, \
>>> +         struct amdxdna_drm_get_array)
>>> +
>>>   #if defined(__cplusplus)
>>>   } /* extern c end */
>>>   #endif
>>


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ