[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <CF2PMAWK64QJ.110OLIYYWWRIF@skynet-linux>
Date: Mon, 18 Oct 2021 23:00:18 +0530
From: "Sireesh Kodali" <sireeshkodali1@...il.com>
To: "Alex Elder" <elder@...e.org>, <phone-devel@...r.kernel.org>,
<~postmarketos/upstreaming@...ts.sr.ht>, <netdev@...r.kernel.org>,
<linux-kernel@...r.kernel.org>, <linux-arm-msm@...r.kernel.org>,
<elder@...nel.org>
Cc: "David S. Miller" <davem@...emloft.net>,
"Jakub Kicinski" <kuba@...nel.org>
Subject: Re: [RFC PATCH 09/17] net: ipa: Add support for using BAM as a DMA
transport
On Thu Oct 14, 2021 at 4:00 AM IST, Alex Elder wrote:
> On 9/19/21 10:08 PM, Sireesh Kodali wrote:
> > BAM is used on IPA v2.x. Since BAM already has a nice dmaengine driver,
> > the IPA driver only makes calls the dmaengine API.
> > Also add BAM transaction support to IPA's trasaction abstraction layer.
> >
> > BAM transactions should use NAPI just like GSI transactions, but just
> > use callbacks on each transaction for now.
>
> This is where things get a little more complicated. I'm not really
> familiar with the BAM interface and would really like to give this
> a much deeper review, and I won't be doing that now.
>
> At first glance, it looks reasonably clean to me, and it surprises
> me a little that this different system can be used with a relatively
> small amount of change. Much looks duplicated, so it could be that
> a little more work abstracting might avoid that (but I haven't looked
> that closely).
>
BAM is handled by the bam_dma driver, which supports the dmaengine API, so
all the functions are like so:
bam_function()
{
bookkeeping();
dmaengine_api_call();
bookkeeping();
}
gsi_function()
{
bookkeeping();
gsi_register_rws();
gsi_misc_ops();
bookkeeping();
}
Some of the bookkeeping code is common between BAM and GSI, but the
current abstraction doesn't allow sharing that code. As I stated
previously, we might be able to share more code (or possibly all code)
between BAM and GSI if GSI was implemented as a dmaengine driver. But
reimplementing GSI like this might be rather time consuming, and there
might be easier solutions to improve code sharing between BAM and GSI.
Regards,
Sireesh
> -Alex
>
> >
> > Signed-off-by: Sireesh Kodali <sireeshkodali1@...il.com>
> > ---
> > drivers/net/ipa/Makefile | 2 +-
> > drivers/net/ipa/bam.c | 525 ++++++++++++++++++++++++++++++
> > drivers/net/ipa/gsi.c | 1 +
> > drivers/net/ipa/ipa_data.h | 1 +
> > drivers/net/ipa/ipa_dma.h | 18 +-
> > drivers/net/ipa/ipa_dma_private.h | 2 +
> > drivers/net/ipa/ipa_main.c | 20 +-
> > drivers/net/ipa/ipa_trans.c | 14 +-
> > drivers/net/ipa/ipa_trans.h | 4 +
> > 9 files changed, 569 insertions(+), 18 deletions(-)
> > create mode 100644 drivers/net/ipa/bam.c
> >
> > diff --git a/drivers/net/ipa/Makefile b/drivers/net/ipa/Makefile
> > index 3cd021fb992e..4abebc667f77 100644
> > --- a/drivers/net/ipa/Makefile
> > +++ b/drivers/net/ipa/Makefile
> > @@ -2,7 +2,7 @@ obj-$(CONFIG_QCOM_IPA) += ipa.o
> >
> > ipa-y := ipa_main.o ipa_power.o ipa_reg.o ipa_mem.o \
> > ipa_table.o ipa_interrupt.o gsi.o ipa_trans.o \
> > - ipa_gsi.o ipa_smp2p.o ipa_uc.o \
> > + ipa_gsi.o ipa_smp2p.o ipa_uc.o bam.o \
> > ipa_endpoint.o ipa_cmd.o ipa_modem.o \
> > ipa_resource.o ipa_qmi.o ipa_qmi_msg.o \
> > ipa_sysfs.o
> > diff --git a/drivers/net/ipa/bam.c b/drivers/net/ipa/bam.c
> > new file mode 100644
> > index 000000000000..0726e385fee5
> > --- /dev/null
> > +++ b/drivers/net/ipa/bam.c
> > @@ -0,0 +1,525 @@
> > +// SPDX-License-Identifier: GPL-2.0
> > +
> > +/* Copyright (c) 2020, The Linux Foundation. All rights reserved.
> > + */
> > +
> > +#include <linux/completion.h>
> > +#include <linux/dma-mapping.h>
> > +#include <linux/dmaengine.h>
> > +#include <linux/interrupt.h>
> > +#include <linux/io.h>
> > +#include <linux/kernel.h>
> > +#include <linux/mutex.h>
> > +#include <linux/netdevice.h>
> > +#include <linux/platform_device.h>
> > +
> > +#include "ipa_gsi.h"
> > +#include "ipa.h"
> > +#include "ipa_dma.h"
> > +#include "ipa_dma_private.h"
> > +#include "ipa_gsi.h"
> > +#include "ipa_trans.h"
> > +#include "ipa_data.h"
> > +
> > +/**
> > + * DOC: The IPA Smart Peripheral System Interface
> > + *
> > + * The Smart Peripheral System is a means to communicate over BAM pipes to
> > + * the IPA block. The Modem also uses BAM pipes to communicate with the IPA
> > + * core.
> > + *
> > + * Refer the GSI documentation, because BAM is a precursor to GSI and more or less
> > + * the same, conceptually (maybe, IDK, I have no docs to go through).
> > + *
> > + * Each channel here corresponds to 1 BAM pipe configured in BAM2BAM mode
> > + *
> > + * IPA cmds are transferred one at a time, each in one BAM transfer.
> > + */
> > +
> > +/* Get and configure the BAM DMA channel */
> > +int bam_channel_init_one(struct ipa_dma *bam,
> > + const struct ipa_gsi_endpoint_data *data, bool command)
> > +{
> > + struct dma_slave_config bam_config;
> > + u32 channel_id = data->channel_id;
> > + struct ipa_channel *channel = &bam->channel[channel_id];
> > + int ret;
> > +
> > + /*TODO: if (!bam_channel_data_valid(bam, data))
> > + return -EINVAL;*/
> > +
> > + channel->dma_subsys = bam;
> > + channel->dma_chan = dma_request_chan(bam->dev, data->channel_name);
> > + channel->toward_ipa = data->toward_ipa;
> > + channel->tlv_count = data->channel.tlv_count;
> > + channel->tre_count = data->channel.tre_count;
> > + if (IS_ERR(channel->dma_chan)) {
> > + dev_err(bam->dev, "failed to request BAM channel %s: %d\n",
> > + data->channel_name,
> > + (int) PTR_ERR(channel->dma_chan));
> > + return PTR_ERR(channel->dma_chan);
> > + }
> > +
> > + ret = ipa_channel_trans_init(bam, data->channel_id);
> > + if (ret)
> > + goto err_dma_chan_free;
> > +
> > + if (data->toward_ipa) {
> > + bam_config.direction = DMA_MEM_TO_DEV;
> > + bam_config.dst_maxburst = channel->tlv_count;
> > + } else {
> > + bam_config.direction = DMA_DEV_TO_MEM;
> > + bam_config.src_maxburst = channel->tlv_count;
> > + }
> > +
> > + dmaengine_slave_config(channel->dma_chan, &bam_config);
> > +
> > + if (command)
> > + ret = ipa_cmd_pool_init(channel, 256);
> > +
> > + if (!ret)
> > + return 0;
> > +
> > +err_dma_chan_free:
> > + dma_release_channel(channel->dma_chan);
> > + return ret;
> > +}
> > +
> > +static void bam_channel_exit_one(struct ipa_channel *channel)
> > +{
> > + if (channel->dma_chan) {
> > + dmaengine_terminate_sync(channel->dma_chan);
> > + dma_release_channel(channel->dma_chan);
> > + }
> > +}
> > +
> > +/* Get channels from BAM_DMA */
> > +int bam_channel_init(struct ipa_dma *bam, u32 count,
> > + const struct ipa_gsi_endpoint_data *data)
> > +{
> > + int ret = 0;
> > + u32 i;
> > +
> > + for (i = 0; i < count; ++i) {
> > + bool command = i == IPA_ENDPOINT_AP_COMMAND_TX;
> > +
> > + if (!data[i].channel_name || data[i].ee_id == GSI_EE_MODEM)
> > + continue;
> > +
> > + ret = bam_channel_init_one(bam, &data[i], command);
> > + if (ret)
> > + goto err_unwind;
> > + }
> > +
> > + return ret;
> > +
> > +err_unwind:
> > + while (i--) {
> > + if (ipa_gsi_endpoint_data_empty(&data[i]))
> > + continue;
> > +
> > + bam_channel_exit_one(&bam->channel[i]);
> > + }
> > + return ret;
> > +}
> > +
> > +/* Inverse of bam_channel_init() */
> > +void bam_channel_exit(struct ipa_dma *bam)
> > +{
> > + u32 channel_id = BAM_CHANNEL_COUNT_MAX - 1;
> > +
> > + do
> > + bam_channel_exit_one(&bam->channel[channel_id]);
> > + while (channel_id--);
> > +}
> > +
> > +/* Inverse of bam_init() */
> > +static void bam_exit(struct ipa_dma *bam)
> > +{
> > + mutex_destroy(&bam->mutex);
> > + bam_channel_exit(bam);
> > +}
> > +
> > +/* Return the channel id associated with a given channel */
> > +static u32 bam_channel_id(struct ipa_channel *channel)
> > +{
> > + return channel - &channel->dma_subsys->channel[0];
> > +}
> > +
> > +static void
> > +bam_channel_tx_update(struct ipa_channel *channel, struct ipa_trans *trans)
> > +{
> > + u64 byte_count = trans->byte_count + trans->len;
> > + u64 trans_count = trans->trans_count + 1;
> > +
> > + byte_count -= channel->compl_byte_count;
> > + channel->compl_byte_count += byte_count;
> > + trans_count -= channel->compl_trans_count;
> > + channel->compl_trans_count += trans_count;
> > +
> > + ipa_gsi_channel_tx_completed(channel->dma_subsys, bam_channel_id(channel),
> > + trans_count, byte_count);
> > +}
> > +
> > +static void
> > +bam_channel_rx_update(struct ipa_channel *channel, struct ipa_trans *trans)
> > +{
> > + /* FIXME */
> > + u64 byte_count = trans->byte_count + trans->len;
> > +
> > + channel->byte_count += byte_count;
> > + channel->trans_count++;
> > +}
> > +
> > +/* Consult hardware, move any newly completed transactions to completed list */
> > +static void bam_channel_update(struct ipa_channel *channel)
> > +{
> > + struct ipa_trans *trans;
> > +
> > + list_for_each_entry(trans, &channel->trans_info.pending, links) {
> > + enum dma_status trans_status =
> > + dma_async_is_tx_complete(channel->dma_chan,
> > + trans->cookie, NULL, NULL);
> > + if (trans_status == DMA_COMPLETE)
> > + break;
> > + }
> > + /* Get the transaction for the latest completed event. Take a
> > + * reference to keep it from completing before we give the events
> > + * for this and previous transactions back to the hardware.
> > + */
> > + refcount_inc(&trans->refcount);
> > +
> > + /* For RX channels, update each completed transaction with the number
> > + * of bytes that were actually received. For TX channels, report
> > + * the number of transactions and bytes this completion represents
> > + * up the network stack.
> > + */
> > + if (channel->toward_ipa)
> > + bam_channel_tx_update(channel, trans);
> > + else
> > + bam_channel_rx_update(channel, trans);
> > +
> > + ipa_trans_move_complete(trans);
> > +
> > + ipa_trans_free(trans);
> > +}
> > +
> > +/**
> > + * bam_channel_poll_one() - Return a single completed transaction on a channel
> > + * @channel: Channel to be polled
> > + *
> > + * Return: Transaction pointer, or null if none are available
> > + *
> > + * This function returns the first entry on a channel's completed transaction
> > + * list. If that list is empty, the hardware is consulted to determine
> > + * whether any new transactions have completed. If so, they're moved to the
> > + * completed list and the new first entry is returned. If there are no more
> > + * completed transactions, a null pointer is returned.
> > + */
> > +static struct ipa_trans *bam_channel_poll_one(struct ipa_channel *channel)
> > +{
> > + struct ipa_trans *trans;
> > +
> > + /* Get the first transaction from the completed list */
> > + trans = ipa_channel_trans_complete(channel);
> > + if (!trans) {
> > + bam_channel_update(channel);
> > + trans = ipa_channel_trans_complete(channel);
> > + }
> > +
> > + if (trans)
> > + ipa_trans_move_polled(trans);
> > +
> > + return trans;
> > +}
> > +
> > +/**
> > + * bam_channel_poll() - NAPI poll function for a channel
> > + * @napi: NAPI structure for the channel
> > + * @budget: Budget supplied by NAPI core
> > + *
> > + * Return: Number of items polled (<= budget)
> > + *
> > + * Single transactions completed by hardware are polled until either
> > + * the budget is exhausted, or there are no more. Each transaction
> > + * polled is passed to ipa_trans_complete(), to perform remaining
> > + * completion processing and retire/free the transaction.
> > + */
> > +static int bam_channel_poll(struct napi_struct *napi, int budget)
> > +{
> > + struct ipa_channel *channel;
> > + int count = 0;
> > +
> > + channel = container_of(napi, struct ipa_channel, napi);
> > + while (count < budget) {
> > + struct ipa_trans *trans;
> > +
> > + count++;
> > + trans = bam_channel_poll_one(channel);
> > + if (!trans)
> > + break;
> > + ipa_trans_complete(trans);
> > + }
> > +
> > + if (count < budget)
> > + napi_complete(&channel->napi);
> > +
> > + return count;
> > +}
> > +
> > +/* Setup function for a single channel */
> > +static void bam_channel_setup_one(struct ipa_dma *bam, u32 channel_id)
> > +{
> > + struct ipa_channel *channel = &bam->channel[channel_id];
> > +
> > + if (!channel->dma_subsys)
> > + return; /* Ignore uninitialized channels */
> > +
> > + if (channel->toward_ipa) {
> > + netif_tx_napi_add(&bam->dummy_dev, &channel->napi,
> > + bam_channel_poll, NAPI_POLL_WEIGHT);
> > + } else {
> > + netif_napi_add(&bam->dummy_dev, &channel->napi,
> > + bam_channel_poll, NAPI_POLL_WEIGHT);
> > + }
> > + napi_enable(&channel->napi);
> > +}
> > +
> > +static void bam_channel_teardown_one(struct ipa_dma *bam, u32 channel_id)
> > +{
> > + struct ipa_channel *channel = &bam->channel[channel_id];
> > +
> > + if (!channel->dma_subsys)
> > + return; /* Ignore uninitialized channels */
> > +
> > + netif_napi_del(&channel->napi);
> > +}
> > +
> > +/* Setup function for channels */
> > +static int bam_channel_setup(struct ipa_dma *bam)
> > +{
> > + u32 channel_id = 0;
> > + int ret;
> > +
> > + mutex_lock(&bam->mutex);
> > +
> > + do
> > + bam_channel_setup_one(bam, channel_id);
> > + while (++channel_id < BAM_CHANNEL_COUNT_MAX);
> > +
> > + /* Make sure no channels were defined that hardware does not support */
> > + while (channel_id < BAM_CHANNEL_COUNT_MAX) {
> > + struct ipa_channel *channel = &bam->channel[channel_id++];
> > +
> > + if (!channel->dma_subsys)
> > + continue; /* Ignore uninitialized channels */
> > +
> > + dev_err(bam->dev, "channel %u not supported by hardware\n",
> > + channel_id - 1);
> > + channel_id = BAM_CHANNEL_COUNT_MAX;
> > + goto err_unwind;
> > + }
> > +
> > + mutex_unlock(&bam->mutex);
> > +
> > + return 0;
> > +
> > +err_unwind:
> > + while (channel_id--)
> > + bam_channel_teardown_one(bam, channel_id);
> > +
> > + mutex_unlock(&bam->mutex);
> > +
> > + return ret;
> > +}
> > +
> > +/* Inverse of bam_channel_setup() */
> > +static void bam_channel_teardown(struct ipa_dma *bam)
> > +{
> > + u32 channel_id;
> > +
> > + mutex_lock(&bam->mutex);
> > +
> > + channel_id = BAM_CHANNEL_COUNT_MAX;
> > + do
> > + bam_channel_teardown_one(bam, channel_id);
> > + while (channel_id--);
> > +
> > + mutex_unlock(&bam->mutex);
> > +}
> > +
> > +static int bam_setup(struct ipa_dma *bam)
> > +{
> > + return bam_channel_setup(bam);
> > +}
> > +
> > +static void bam_teardown(struct ipa_dma *bam)
> > +{
> > + bam_channel_teardown(bam);
> > +}
> > +
> > +static u32 bam_channel_tre_max(struct ipa_dma *bam, u32 channel_id)
> > +{
> > + struct ipa_channel *channel = &bam->channel[channel_id];
> > +
> > + /* Hardware limit is channel->tre_count - 1 */
> > + return channel->tre_count - (channel->tlv_count - 1);
> > +}
> > +
> > +static u32 bam_channel_trans_tre_max(struct ipa_dma *bam, u32 channel_id)
> > +{
> > + struct ipa_channel *channel = &bam->channel[channel_id];
> > +
> > + return channel->tlv_count;
> > +}
> > +
> > +static int bam_channel_start(struct ipa_dma *bam, u32 channel_id)
> > +{
> > + return 0;
> > +}
> > +
> > +static int bam_channel_stop(struct ipa_dma *bam, u32 channel_id)
> > +{
> > + struct ipa_channel *channel = &bam->channel[channel_id];
> > +
> > + return dmaengine_terminate_sync(channel->dma_chan);
> > +}
> > +
> > +static void bam_channel_reset(struct ipa_dma *bam, u32 channel_id, bool doorbell)
> > +{
> > + bam_channel_stop(bam, channel_id);
> > +}
> > +
> > +static int bam_channel_suspend(struct ipa_dma *bam, u32 channel_id)
> > +{
> > + struct ipa_channel *channel = &bam->channel[channel_id];
> > +
> > + return dmaengine_pause(channel->dma_chan);
> > +}
> > +
> > +static int bam_channel_resume(struct ipa_dma *bam, u32 channel_id)
> > +{
> > + struct ipa_channel *channel = &bam->channel[channel_id];
> > +
> > + return dmaengine_resume(channel->dma_chan);
> > +}
> > +
> > +static void bam_suspend(struct ipa_dma *bam)
> > +{
> > + /* No-op for now */
> > +}
> > +
> > +static void bam_resume(struct ipa_dma *bam)
> > +{
> > + /* No-op for now */
> > +}
> > +
> > +static void bam_trans_callback(void *arg)
> > +{
> > + ipa_trans_complete(arg);
> > +}
> > +
> > +static void bam_trans_commit(struct ipa_trans *trans, bool unused)
> > +{
> > + struct ipa_channel *channel = &trans->dma_subsys->channel[trans->channel_id];
> > + enum ipa_cmd_opcode opcode = IPA_CMD_NONE;
> > + struct ipa_cmd_info *info;
> > + struct scatterlist *sg;
> > + u32 byte_count = 0;
> > + u32 i;
> > + enum dma_transfer_direction direction;
> > +
> > + if (channel->toward_ipa)
> > + direction = DMA_MEM_TO_DEV;
> > + else
> > + direction = DMA_DEV_TO_MEM;
> > +
> > + /* assert(trans->used > 0); */
> > +
> > + info = trans->info ? &trans->info[0] : NULL;
> > + for_each_sg(trans->sgl, sg, trans->used, i) {
> > + bool last_tre = i == trans->used - 1;
> > + dma_addr_t addr = sg_dma_address(sg);
> > + u32 len = sg_dma_len(sg);
> > + u32 dma_flags = 0;
> > + struct dma_async_tx_descriptor *desc;
> > +
> > + byte_count += len;
> > + if (info)
> > + opcode = info++->opcode;
> > +
> > + if (opcode != IPA_CMD_NONE) {
> > + len = opcode;
> > + dma_flags |= DMA_PREP_IMM_CMD;
> > + }
> > +
> > + if (last_tre)
> > + dma_flags |= DMA_PREP_INTERRUPT;
> > +
> > + desc = dmaengine_prep_slave_single(channel->dma_chan, addr, len,
> > + direction, dma_flags);
> > +
> > + if (last_tre) {
> > + desc->callback = bam_trans_callback;
> > + desc->callback_param = trans;
> > + }
> > +
> > + desc->cookie = dmaengine_submit(desc);
> > +
> > + if (last_tre)
> > + trans->cookie = desc->cookie;
> > +
> > + if (direction == DMA_DEV_TO_MEM)
> > + dmaengine_desc_attach_metadata(desc, &trans->len, sizeof(trans->len));
> > + }
> > +
> > + if (channel->toward_ipa) {
> > + /* We record TX bytes when they are sent */
> > + trans->len = byte_count;
> > + trans->trans_count = channel->trans_count;
> > + trans->byte_count = channel->byte_count;
> > + channel->trans_count++;
> > + channel->byte_count += byte_count;
> > + }
> > +
> > + ipa_trans_move_pending(trans);
> > +
> > + dma_async_issue_pending(channel->dma_chan);
> > +}
> > +
> > +/* Initialize the BAM DMA channels
> > + * Actual hw init is handled by the BAM_DMA driver
> > + */
> > +int bam_init(struct ipa_dma *bam, struct platform_device *pdev,
> > + enum ipa_version version, u32 count,
> > + const struct ipa_gsi_endpoint_data *data)
> > +{
> > + struct device *dev = &pdev->dev;
> > + int ret;
> > +
> > + bam->dev = dev;
> > + bam->version = version;
> > + bam->setup = bam_setup;
> > + bam->teardown = bam_teardown;
> > + bam->exit = bam_exit;
> > + bam->suspend = bam_suspend;
> > + bam->resume = bam_resume;
> > + bam->channel_tre_max = bam_channel_tre_max;
> > + bam->channel_trans_tre_max = bam_channel_trans_tre_max;
> > + bam->channel_start = bam_channel_start;
> > + bam->channel_stop = bam_channel_stop;
> > + bam->channel_reset = bam_channel_reset;
> > + bam->channel_suspend = bam_channel_suspend;
> > + bam->channel_resume = bam_channel_resume;
> > + bam->trans_commit = bam_trans_commit;
> > +
> > + init_dummy_netdev(&bam->dummy_dev);
> > +
> > + ret = bam_channel_init(bam, count, data);
> > + if (ret)
> > + return ret;
> > +
> > + mutex_init(&bam->mutex);
> > +
> > + return 0;
> > +}
> > diff --git a/drivers/net/ipa/gsi.c b/drivers/net/ipa/gsi.c
> > index 39d9ca620a9f..ac0b9e748fa1 100644
> > --- a/drivers/net/ipa/gsi.c
> > +++ b/drivers/net/ipa/gsi.c
> > @@ -2210,6 +2210,7 @@ int gsi_init(struct ipa_dma *gsi, struct platform_device *pdev,
> > gsi->channel_reset = gsi_channel_reset;
> > gsi->channel_suspend = gsi_channel_suspend;
> > gsi->channel_resume = gsi_channel_resume;
> > + gsi->trans_commit = gsi_trans_commit;
> >
> > /* GSI uses NAPI on all channels. Create a dummy network device
> > * for the channel NAPI contexts to be associated with.
> > diff --git a/drivers/net/ipa/ipa_data.h b/drivers/net/ipa/ipa_data.h
> > index 6d329e9ce5d2..7d62d49f414f 100644
> > --- a/drivers/net/ipa/ipa_data.h
> > +++ b/drivers/net/ipa/ipa_data.h
> > @@ -188,6 +188,7 @@ struct ipa_gsi_endpoint_data {
> > u8 channel_id;
> > u8 endpoint_id;
> > bool toward_ipa;
> > + const char *channel_name; /* used only for BAM DMA channels */
> >
> > struct gsi_channel_data channel;
> > struct ipa_endpoint_data endpoint;
> > diff --git a/drivers/net/ipa/ipa_dma.h b/drivers/net/ipa/ipa_dma.h
> > index 1a23e6ac5785..3000182ae689 100644
> > --- a/drivers/net/ipa/ipa_dma.h
> > +++ b/drivers/net/ipa/ipa_dma.h
> > @@ -17,7 +17,11 @@
> >
> > /* Maximum number of channels and event rings supported by the driver */
> > #define GSI_CHANNEL_COUNT_MAX 23
> > +#define BAM_CHANNEL_COUNT_MAX 20
> > #define GSI_EVT_RING_COUNT_MAX 24
> > +#define IPA_CHANNEL_COUNT_MAX MAX(GSI_CHANNEL_COUNT_MAX, \
> > + BAM_CHANNEL_COUNT_MAX)
> > +#define MAX(a, b) ((a > b) ? a : b)
> >
> > /* Maximum TLV FIFO size for a channel; 64 here is arbitrary (and high) */
> > #define GSI_TLV_MAX 64
> > @@ -119,6 +123,8 @@ struct ipa_channel {
> > struct gsi_ring tre_ring;
> > u32 evt_ring_id;
> >
> > + struct dma_chan *dma_chan;
> > +
> > u64 byte_count; /* total # bytes transferred */
> > u64 trans_count; /* total # transactions */
> > /* The following counts are used only for TX endpoints */
> > @@ -154,7 +160,7 @@ struct ipa_dma {
> > u32 irq;
> > u32 channel_count;
> > u32 evt_ring_count;
> > - struct ipa_channel channel[GSI_CHANNEL_COUNT_MAX];
> > + struct ipa_channel channel[IPA_CHANNEL_COUNT_MAX];
> > struct gsi_evt_ring evt_ring[GSI_EVT_RING_COUNT_MAX];
> > u32 event_bitmap; /* allocated event rings */
> > u32 modem_channel_bitmap; /* modem channels to allocate */
> > @@ -303,7 +309,7 @@ static inline void ipa_dma_resume(struct ipa_dma *dma_subsys)
> > }
> >
> > /**
> > - * ipa_dma_init() - Initialize the GSI subsystem
> > + * ipa_init/bam_init() - Initialize the GSI/BAM subsystem
> > * @dma_subsys: Address of ipa_dma structure embedded in an IPA structure
> > * @pdev: IPA platform device
> > * @version: IPA hardware version (implies GSI version)
> > @@ -312,14 +318,18 @@ static inline void ipa_dma_resume(struct ipa_dma *dma_subsys)
> > *
> > * Return: 0 if successful, or a negative error code
> > *
> > - * Early stage initialization of the GSI subsystem, performing tasks
> > - * that can be done before the GSI hardware is ready to use.
> > + * Early stage initialization of the GSI/BAM subsystem, performing tasks
> > + * that can be done before the GSI/BAM hardware is ready to use.
> > */
> >
> > int gsi_init(struct ipa_dma *dma_subsys, struct platform_device *pdev,
> > enum ipa_version version, u32 count,
> > const struct ipa_gsi_endpoint_data *data);
> >
> > +int bam_init(struct ipa_dma *dma_subsys, struct platform_device *pdev,
> > + enum ipa_version version, u32 count,
> > + const struct ipa_gsi_endpoint_data *data);
> > +
> > /**
> > * ipa_dma_exit() - Exit the DMA subsystem
> > * @dma_subsys: ipa_dma address previously passed to a successful gsi_init() call
> > diff --git a/drivers/net/ipa/ipa_dma_private.h b/drivers/net/ipa/ipa_dma_private.h
> > index 40148a551b47..1db53e597a61 100644
> > --- a/drivers/net/ipa/ipa_dma_private.h
> > +++ b/drivers/net/ipa/ipa_dma_private.h
> > @@ -16,6 +16,8 @@ struct ipa_channel;
> >
> > #define GSI_RING_ELEMENT_SIZE 16 /* bytes; must be a power of 2 */
> >
> > +void gsi_trans_commit(struct ipa_trans *trans, bool ring_db);
> > +
> > /* Return the entry that follows one provided in a transaction pool */
> > void *ipa_trans_pool_next(struct ipa_trans_pool *pool, void *element);
> >
> > diff --git a/drivers/net/ipa/ipa_main.c b/drivers/net/ipa/ipa_main.c
> > index ba06e3ad554c..ea6c4347f2c6 100644
> > --- a/drivers/net/ipa/ipa_main.c
> > +++ b/drivers/net/ipa/ipa_main.c
> > @@ -60,12 +60,15 @@
> > * core. The GSI implements a set of "channels" used for communication
> > * between the AP and the IPA.
> > *
> > - * The IPA layer uses GSI channels to implement its "endpoints". And while
> > - * a GSI channel carries data between the AP and the IPA, a pair of IPA
> > - * endpoints is used to carry traffic between two EEs. Specifically, the main
> > - * modem network interface is implemented by two pairs of endpoints: a TX
> > + * The IPA layer uses GSI channels or BAM pipes to implement its "endpoints".
> > + * And while a GSI channel carries data between the AP and the IPA, a pair of
> > + * IPA endpoints is used to carry traffic between two EEs. Specifically, the
> > + * main modem network interface is implemented by two pairs of endpoints: a TX
> > * endpoint on the AP coupled with an RX endpoint on the modem; and another
> > * RX endpoint on the AP receiving data from a TX endpoint on the modem.
> > + *
> > + * For BAM based transport, a pair of BAM pipes are used for TX and RX between
> > + * the AP and IPA, and between IPA and other EEs.
> > */
> >
> > /* The name of the GSI firmware file relative to /lib/firmware */
> > @@ -716,8 +719,13 @@ static int ipa_probe(struct platform_device *pdev)
> > if (ret)
> > goto err_reg_exit;
> >
> > - ret = gsi_init(&ipa->dma_subsys, pdev, ipa->version, data->endpoint_count,
> > - data->endpoint_data);
> > + if (IPA_HAS_GSI(ipa->version))
> > + ret = gsi_init(&ipa->dma_subsys, pdev, ipa->version, data->endpoint_count,
> > + data->endpoint_data);
> > + else
> > + ret = bam_init(&ipa->dma_subsys, pdev, ipa->version, data->endpoint_count,
> > + data->endpoint_data);
> > +
> > if (ret)
> > goto err_mem_exit;
> >
> > diff --git a/drivers/net/ipa/ipa_trans.c b/drivers/net/ipa/ipa_trans.c
> > index 22755f3ce3da..444f44846da8 100644
> > --- a/drivers/net/ipa/ipa_trans.c
> > +++ b/drivers/net/ipa/ipa_trans.c
> > @@ -254,7 +254,7 @@ struct ipa_trans *ipa_channel_trans_complete(struct ipa_channel *channel)
> > }
> >
> > /* Move a transaction from the allocated list to the pending list */
> > -static void ipa_trans_move_pending(struct ipa_trans *trans)
> > +void ipa_trans_move_pending(struct ipa_trans *trans)
> > {
> > struct ipa_channel *channel = &trans->dma_subsys->channel[trans->channel_id];
> > struct ipa_trans_info *trans_info = &channel->trans_info;
> > @@ -539,7 +539,7 @@ static void gsi_trans_tre_fill(struct gsi_tre *dest_tre, dma_addr_t addr,
> > * pending list. Finally, updates the channel ring pointer and optionally
> > * rings the doorbell.
> > */
> > -static void __gsi_trans_commit(struct ipa_trans *trans, bool ring_db)
> > +void gsi_trans_commit(struct ipa_trans *trans, bool ring_db)
> > {
> > struct ipa_channel *channel = &trans->dma_subsys->channel[trans->channel_id];
> > struct gsi_ring *ring = &channel->tre_ring;
> > @@ -604,9 +604,9 @@ static void __gsi_trans_commit(struct ipa_trans *trans, bool ring_db)
> > /* Commit a GSI transaction */
> > void ipa_trans_commit(struct ipa_trans *trans, bool ring_db)
> > {
> > - if (trans->used)
> > - __gsi_trans_commit(trans, ring_db);
> > - else
> > + if (trans->used) {
> > + trans->dma_subsys->trans_commit(trans, ring_db);
> > + } else
> > ipa_trans_free(trans);
> > }
> >
> > @@ -618,7 +618,7 @@ void ipa_trans_commit_wait(struct ipa_trans *trans)
> >
> > refcount_inc(&trans->refcount);
> >
> > - __gsi_trans_commit(trans, true);
> > + trans->dma_subsys->trans_commit(trans, true);
> >
> > wait_for_completion(&trans->completion);
> >
> > @@ -638,7 +638,7 @@ int ipa_trans_commit_wait_timeout(struct ipa_trans *trans,
> >
> > refcount_inc(&trans->refcount);
> >
> > - __gsi_trans_commit(trans, true);
> > + trans->dma_subsys->trans_commit(trans, true);
> >
> > remaining = wait_for_completion_timeout(&trans->completion,
> > timeout_jiffies);
> > diff --git a/drivers/net/ipa/ipa_trans.h b/drivers/net/ipa/ipa_trans.h
> > index b93342414360..5f41e3e6f92a 100644
> > --- a/drivers/net/ipa/ipa_trans.h
> > +++ b/drivers/net/ipa/ipa_trans.h
> > @@ -10,6 +10,7 @@
> > #include <linux/refcount.h>
> > #include <linux/completion.h>
> > #include <linux/dma-direction.h>
> > +#include <linux/dmaengine.h>
> >
> > #include "ipa_cmd.h"
> >
> > @@ -61,6 +62,7 @@ struct ipa_trans {
> > struct scatterlist *sgl;
> > struct ipa_cmd_info *info; /* array of entries, or null */
> > enum dma_data_direction direction;
> > + dma_cookie_t cookie;
> >
> > refcount_t refcount;
> > struct completion completion;
> > @@ -149,6 +151,8 @@ struct ipa_trans *ipa_channel_trans_alloc(struct ipa_dma *dma_subsys, u32 channe
> > */
> > void ipa_trans_free(struct ipa_trans *trans);
> >
> > +void ipa_trans_move_pending(struct ipa_trans *trans);
> > +
> > /**
> > * ipa_trans_cmd_add() - Add an immediate command to a transaction
> > * @trans: Transaction
> >
Powered by blists - more mailing lists