[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <CAC=U0a2MskDVgB0rVBoQ2vt+TRUw1royBm++LturL02V_j1Z-w@mail.gmail.com>
Date: Thu, 22 Dec 2022 14:00:33 -0500
From: Kamal Dasu <kdasu.kdev@...il.com>
To: Christian Löhle <CLoehle@...erstone.com>
Cc: "keescook@...omium.org" <keescook@...omium.org>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
"gmpy.liaowx@...il.com" <gmpy.liaowx@...il.com>,
"linux-mmc@...r.kernel.org" <linux-mmc@...r.kernel.org>,
"f.fainelli@...il.com" <f.fainelli@...il.com>,
"tony.luck@...el.com" <tony.luck@...el.com>,
"anton@...msg.org" <anton@...msg.org>
Subject: Re: [RFC 1/1] mmc: Add mmc pstore backend support
On Sat, Dec 17, 2022 at 9:24 AM Christian Löhle <CLoehle@...erstone.com> wrote:
"Please check your patch, I think you're accidentally reverting all
recent mmc changes."
I will send a v2 RFC patch by fixing this.
>
> Two things right of the bat:
> MMC and SD are managed flash, you don't need to erase before writing somewhere.
> (although you can)
Thanks for your input. It is my understanding that the pstore fs is
using a zoned filesystem to populate sys/fs/pstore when reading and
writing.
I just modelled it like the mtdpstore and seems like panic_write is
registered only with the fs/pstore/zone api.
> There is a lot of things that need to be considered to get writes workable on all cards and hosts, it should be easier to make use of the core/block.c code somehow instead of doing this yourself.
I am not sure what needs to be considered for mmc storage. Can you
please elaborate.
>
>
> -----Original Message-----
> From: Kamal Dasu <kdasu.kdev@...il.com>
> Sent: Freitag, 16. Dezember 2022 22:54
> To: keescook@...omium.org; linux-kernel@...r.kernel.org; gmpy.liaowx@...il.com; linux-mmc@...r.kernel.org
> Cc: f.fainelli@...il.com; tony.luck@...el.com; anton@...msg.org; Kamal Dasu <kdasu.kdev@...il.com>
> Subject: [RFC 1/1] mmc: Add mmc pstore backend support
>
> Adding mmc pstore backend support to be able to write to pstore device. Code is based on mtdpstore.
>
> Signed-off-by: Kamal Dasu <kdasu.kdev@...il.com>
> ---
> drivers/mmc/core/Kconfig | 13 +
> drivers/mmc/core/Makefile | 1 +
> drivers/mmc/core/block.c | 61 ++--
> drivers/mmc/core/block.h | 9 +
> drivers/mmc/core/core.c | 28 +-
> drivers/mmc/core/mmcpstore.c | 594 +++++++++++++++++++++++++++++++++++
> drivers/mmc/host/sdhci.c | 166 +++++-----
> include/linux/mmc/host.h | 3 +
> 8 files changed, 746 insertions(+), 129 deletions(-) create mode 100644 drivers/mmc/core/mmcpstore.c
>
> diff --git a/drivers/mmc/core/Kconfig b/drivers/mmc/core/Kconfig index 6f25c34e4fec..fd46753ed746 100644
> --- a/drivers/mmc/core/Kconfig
> +++ b/drivers/mmc/core/Kconfig
> @@ -34,9 +34,22 @@ config PWRSEQ_SIMPLE
> This driver can also be built as a module. If so, the module
> will be called pwrseq_simple.
>
> +config MMC_PSTORE_BACKEND
> + bool "Log panic/oops to a MMC buffer"
> + depends on MMC_BLOCK
> + help
> + This option will let you create platform backend to store kmsg
> + crash dumps to a user specified MMC device. This is primarily
> + based on pstore/blk.
> +
> +config MMC_PSTORE
> + tristate
> + select PSTORE_BLK
> +
> config MMC_BLOCK
> tristate "MMC block device driver"
> depends on BLOCK
> + select MMC_PSTORE if MMC_PSTORE_BACKEND=y
> default y
> help
> Say Y here to enable the MMC block device driver support.
> diff --git a/drivers/mmc/core/Makefile b/drivers/mmc/core/Makefile index 6a907736cd7a..c3f1199deb71 100644
> --- a/drivers/mmc/core/Makefile
> +++ b/drivers/mmc/core/Makefile
> @@ -16,6 +16,7 @@ obj-$(CONFIG_PWRSEQ_EMMC) += pwrseq_emmc.o
> mmc_core-$(CONFIG_DEBUG_FS) += debugfs.o
> obj-$(CONFIG_MMC_BLOCK) += mmc_block.o
> mmc_block-objs := block.o queue.o
> +mmc_block-$(CONFIG_MMC_PSTORE) += mmcpstore.o
> obj-$(CONFIG_MMC_TEST) += mmc_test.o
> obj-$(CONFIG_SDIO_UART) += sdio_uart.o
> mmc_core-$(CONFIG_MMC_CRYPTO) += crypto.o
> diff --git a/drivers/mmc/core/block.c b/drivers/mmc/core/block.c index 20da7ed43e6d..95da015c65b8 100644
> --- a/drivers/mmc/core/block.c
> +++ b/drivers/mmc/core/block.c
> @@ -134,7 +134,6 @@ struct mmc_blk_data {
> * track of the current selected device partition.
> */
> unsigned int part_curr;
> -#define MMC_BLK_PART_INVALID UINT_MAX /* Unknown partition active */
> int area_type;
>
> /* debugfs files (only in main mmc_blk_data) */ @@ -514,6 +513,19 @@ static int __mmc_blk_ioctl_cmd(struct mmc_card *card, struct mmc_blk_data *md,
> if (idata->ic.data_timeout_ns)
> data.timeout_ns = idata->ic.data_timeout_ns;
>
> + if ((cmd.flags & MMC_RSP_R1B) == MMC_RSP_R1B) {
> + /*
> + * Pretend this is a data transfer and rely on the
> + * host driver to compute timeout. When all host
> + * drivers support cmd.cmd_timeout for R1B, this
> + * can be changed to:
> + *
> + * mrq.data = NULL;
> + * cmd.cmd_timeout = idata->ic.cmd_timeout_ms;
> + */
> + data.timeout_ns = idata->ic.cmd_timeout_ms * 1000000;
> + }
> +
> mrq.data = &data;
> }
>
> @@ -975,39 +987,33 @@ static unsigned int mmc_blk_data_timeout_ms(struct mmc_host *host,
> return ms;
> }
>
> -/*
> - * Attempts to reset the card and get back to the requested partition.
> - * Therefore any error here must result in cancelling the block layer
> - * request, it must not be reattempted without going through the mmc_blk
> - * partition sanity checks.
> - */
> static int mmc_blk_reset(struct mmc_blk_data *md, struct mmc_host *host,
> int type)
> {
> int err;
> - struct mmc_blk_data *main_md = dev_get_drvdata(&host->card->dev);
>
> if (md->reset_done & type)
> return -EEXIST;
>
> md->reset_done |= type;
> err = mmc_hw_reset(host->card);
> - /*
> - * A successful reset will leave the card in the main partition, but
> - * upon failure it might not be, so set it to MMC_BLK_PART_INVALID
> - * in that case.
> - */
> - main_md->part_curr = err ? MMC_BLK_PART_INVALID : main_md->part_type;
> - if (err)
> - return err;
> /* Ensure we switch back to the correct partition */
> - if (mmc_blk_part_switch(host->card, md->part_type))
> - /*
> - * We have failed to get back into the correct
> - * partition, so we need to abort the whole request.
> - */
> - return -ENODEV;
> - return 0;
> + if (err) {
> + struct mmc_blk_data *main_md =
> + dev_get_drvdata(&host->card->dev);
> + int part_err;
> +
> + main_md->part_curr = main_md->part_type;
> + part_err = mmc_blk_part_switch(host->card, md->part_type);
> + if (part_err) {
> + /*
> + * We have failed to get back into the correct
> + * partition, so we need to abort the whole request.
> + */
> + return -ENODEV;
> + }
> + }
> + return err;
> }
>
> static inline void mmc_blk_reset_success(struct mmc_blk_data *md, int type) @@ -1865,9 +1871,8 @@ static void mmc_blk_mq_rw_recovery(struct mmc_queue *mq, struct request *req)
> return;
>
> /* Reset before last retry */
> - if (mqrq->retries + 1 == MMC_MAX_RETRIES &&
> - mmc_blk_reset(md, card->host, type))
> - return;
> + if (mqrq->retries + 1 == MMC_MAX_RETRIES)
> + mmc_blk_reset(md, card->host, type);
>
> /* Command errors fail fast, so use all MMC_MAX_RETRIES */
> if (brq->sbc.error || brq->cmd.error)
> @@ -2977,6 +2982,9 @@ static int mmc_blk_probe(struct mmc_card *card)
> if (ret)
> goto out;
>
> + if (mmc_card_mmc(card) || mmc_card_sd(card))
> + mmcpstore_register(card, md->disk->disk_name);
> +
> /* Add two debugfs entries */
> mmc_blk_add_debugfs(card, md);
>
> @@ -3126,6 +3134,7 @@ static void __exit mmc_blk_exit(void)
> unregister_blkdev(MMC_BLOCK_MAJOR, "mmc");
> unregister_chrdev_region(mmc_rpmb_devt, MAX_DEVICES);
> bus_unregister(&mmc_rpmb_bus_type);
> + unregister_mmcpstore();
> }
>
> module_init(mmc_blk_init);
> diff --git a/drivers/mmc/core/block.h b/drivers/mmc/core/block.h index 31153f656f41..974081d0ab4e 100644
> --- a/drivers/mmc/core/block.h
> +++ b/drivers/mmc/core/block.h
> @@ -16,5 +16,14 @@ void mmc_blk_mq_recovery(struct mmc_queue *mq); struct work_struct;
>
> void mmc_blk_mq_complete_work(struct work_struct *work);
> +#if IS_ENABLED(CONFIG_MMC_PSTORE)
> +sector_t mmc_blk_get_part(struct mmc_card *card, int part_num, sector_t
> +*size); void mmcpstore_register(struct mmc_card *card, const char
> +*disk_name); void unregister_mmcpstore(void); #else static inline void
> +mmcpstore_register(struct mmc_card *card,
> + const char *disk_name) {}
> +static inline void unregister_mmcpstore(void) {} #endif
>
> #endif
> diff --git a/drivers/mmc/core/core.c b/drivers/mmc/core/core.c index 426c7f66b349..ef53a2578824 100644
> --- a/drivers/mmc/core/core.c
> +++ b/drivers/mmc/core/core.c
> @@ -56,7 +56,7 @@ static const unsigned freqs[] = { 400000, 300000, 200000, 100000 };
> /*
> * Enabling software CRCs on the data blocks can be a significant (30%)
> * performance cost, and for other reasons may not always be desired.
> - * So we allow it to be disabled.
> + * So we allow it it to be disabled.
> */
> bool use_spi_crc = 1;
> module_param(use_spi_crc, bool, 0);
> @@ -97,8 +97,8 @@ static void mmc_should_fail_request(struct mmc_host *host,
> !should_fail(&host->fail_mmc_request, data->blksz * data->blocks))
> return;
>
> - data->error = data_errors[get_random_u32_below(ARRAY_SIZE(data_errors))];
> - data->bytes_xfered = get_random_u32_below(data->bytes_xfered >> 9) << 9;
> + data->error = data_errors[prandom_u32() % ARRAY_SIZE(data_errors)];
> + data->bytes_xfered = (prandom_u32() % (data->bytes_xfered >> 9)) << 9;
> }
>
> #else /* CONFIG_FAIL_MMC_REQUEST */
> @@ -142,7 +142,8 @@ void mmc_request_done(struct mmc_host *host, struct mmc_request *mrq)
> int err = cmd->error;
>
> /* Flag re-tuning needed on CRC errors */
> - if (!mmc_op_tuning(cmd->opcode) &&
> + if (cmd->opcode != MMC_SEND_TUNING_BLOCK &&
> + cmd->opcode != MMC_SEND_TUNING_BLOCK_HS200 &&
> !host->retune_crc_disable &&
> (err == -EILSEQ || (mrq->sbc && mrq->sbc->error == -EILSEQ) ||
> (mrq->data && mrq->data->error == -EILSEQ) || @@ -526,7 +527,7 @@ EXPORT_SYMBOL(mmc_cqe_post_req);
> * mmc_cqe_recovery - Recover from CQE errors.
> * @host: MMC host to recover
> *
> - * Recovery consists of stopping CQE, stopping eMMC, discarding the queue
> + * Recovery consists of stopping CQE, stopping eMMC, discarding the
> + queue in
> * in eMMC, and discarding the queue in CQE. CQE must call
> * mmc_cqe_request_done() on all requests. An error is returned if the eMMC
> * fails to discard its queue.
> @@ -1133,13 +1134,7 @@ u32 mmc_select_voltage(struct mmc_host *host, u32 ocr)
> mmc_power_cycle(host, ocr);
> } else {
> bit = fls(ocr) - 1;
> - /*
> - * The bit variable represents the highest voltage bit set in
> - * the OCR register.
> - * To keep a range of 2 values (e.g. 3.2V/3.3V and 3.3V/3.4V),
> - * we must shift the mask '3' with (bit - 1).
> - */
> - ocr &= 3 << (bit - 1);
> + ocr &= 3 << bit;
> if (bit != host->ios.vdd)
> dev_warn(mmc_dev(host), "exceeding card's volts\n");
> }
> @@ -1483,11 +1478,6 @@ void mmc_init_erase(struct mmc_card *card)
> card->pref_erase = 0;
> }
>
> -static bool is_trim_arg(unsigned int arg) -{
> - return (arg & MMC_TRIM_OR_DISCARD_ARGS) && arg != MMC_DISCARD_ARG;
> -}
> -
> static unsigned int mmc_mmc_erase_timeout(struct mmc_card *card,
> unsigned int arg, unsigned int qty) { @@ -1770,7 +1760,7 @@ int mmc_erase(struct mmc_card *card, unsigned int from, unsigned int nr,
> !(card->ext_csd.sec_feature_support & EXT_CSD_SEC_ER_EN))
> return -EOPNOTSUPP;
>
> - if (mmc_card_mmc(card) && is_trim_arg(arg) &&
> + if (mmc_card_mmc(card) && (arg & MMC_TRIM_ARGS) &&
> !(card->ext_csd.sec_feature_support & EXT_CSD_SEC_GB_CL_EN))
> return -EOPNOTSUPP;
>
> @@ -1800,7 +1790,7 @@ int mmc_erase(struct mmc_card *card, unsigned int from, unsigned int nr,
> * identified by the card->eg_boundary flag.
> */
> rem = card->erase_size - (from % card->erase_size);
> - if ((arg & MMC_TRIM_OR_DISCARD_ARGS) && card->eg_boundary && nr > rem) {
> + if ((arg & MMC_TRIM_ARGS) && (card->eg_boundary) && (nr > rem)) {
> err = mmc_do_erase(card, from, from + rem - 1, arg);
> from += rem;
> if ((err) || (to <= from))
> diff --git a/drivers/mmc/core/mmcpstore.c b/drivers/mmc/core/mmcpstore.c new file mode 100644 index 000000000000..f881be42aa41
> --- /dev/null
> +++ b/drivers/mmc/core/mmcpstore.c
> @@ -0,0 +1,594 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * MMC pstore support based on mtdpstore, pstore/blk
> + *
> + */
> +
> +#define pr_fmt(fmt) "mmcpstore: " fmt
> +
> +#include <linux/kernel.h>
> +#include <linux/module.h>
> +#include <linux/pstore_blk.h>
> +#include <linux/blkdev.h>
> +#include <linux/mount.h>
> +#include <linux/slab.h>
> +#include <linux/mmc/mmc.h>
> +#include <linux/mmc/host.h>
> +#include <linux/mmc/card.h>
> +#include <linux/scatterlist.h>
> +#include "block.h"
> +#include "card.h"
> +#include "core.h"
> +
> +#define MMC_READ_REQ 0
> +#define MMC_WRITE_REQ 1
> +#define MMC_PANIC_WRITE_REQ 2
> +
> +static struct mmcpstore_context {
> + struct pstore_blk_config conf;
> + struct pstore_device_info dev;
> + struct mmc_card *card;
> + struct mmc_request *mrq;
> + sector_t start_sect;
> + sector_t size;
> + unsigned long *rmmap; /* removed bit map */
> + unsigned long *usedmap; /* used bit map */
> +
> +} oops_cxt;
> +
> +static void mmc_wait_for_pstore_req(struct mmc_host *host,
> + struct mmc_request *mrq,
> + int poll)
> +{
> + unsigned int timeout;
> +
> + mmc_start_request(host, mrq);
> +
> + if (poll && mrq->data) {
> + timeout = mrq->data->timeout_ns / NSEC_PER_MSEC;
> + host->ops->req_completion_poll(host, timeout);
> + }
> +}
> +
> +static void mmc_prep_req(struct mmc_request *mrq,
> + unsigned int sect_offset, unsigned int nsects,
> + struct scatterlist *sg, u32 opcode, unsigned int flags) {
> + mrq->cmd->opcode = opcode;
> + mrq->cmd->arg = sect_offset;
> + mrq->cmd->flags = MMC_RSP_R1 | MMC_CMD_ADTC;
> +
> + if (nsects == 1) {
> + mrq->stop = NULL;
> + } else {
> + mrq->stop->opcode = MMC_STOP_TRANSMISSION;
> + mrq->stop->arg = 0;
> + mrq->stop->flags = MMC_RSP_R1B | MMC_CMD_AC;
> + }
> +
> + mrq->data->blksz = SECTOR_SIZE;
> + mrq->data->blocks = nsects;
> + mrq->data->flags = flags;
> + mrq->data->sg = sg;
> + mrq->data->sg_len = 1;
> +}
> +
> +static int mmcpstore_req(const char *buf,
> + size_t size, loff_t sect_offset, int write) {
> + struct mmcpstore_context *cxt = &oops_cxt;
> + struct mmc_request *mrq = cxt->mrq;
> + struct mmc_card *card = cxt->card;
> + struct mmc_host *host = card->host;
> + struct scatterlist sg;
> + u32 opcode;
> + u32 nsects = size >> SECTOR_SHIFT;
> +
> +
> + if (MMC_READ_REQ)
> + opcode = (nsects > 1) ? MMC_READ_MULTIPLE_BLOCK :
> + MMC_READ_SINGLE_BLOCK;
> + else
> + opcode = (nsects > 1) ? MMC_WRITE_MULTIPLE_BLOCK :
> + MMC_WRITE_BLOCK;
> +
> +
> + sg_init_one(&sg, buf, size);
> + mmc_prep_req(mrq, sect_offset, nsects, &sg, opcode, MMC_DATA_WRITE);
> + mmc_set_data_timeout(mrq->data, cxt->card);
> +
> + mmc_claim_host(host);
> + mmc_wait_for_pstore_req(host, mrq,
> + (write == MMC_PANIC_WRITE_REQ) ? 1 : 0);
> + mmc_release_host(host);
> +
> + return 0;
> +}
> +
> +static inline void mmcpstore_mark_used(struct mmcpstore_context *cxt,
> + loff_t off)
> +{
> + u64 zonenum = div_u64(off, cxt->conf.kmsg_size);
> +
> + dev_dbg(&cxt->card->dev, "mark zone %llu used\n", zonenum);
> + set_bit(zonenum, cxt->usedmap);
> +}
> +
> +static inline void mmcpstore_mark_unused(struct mmcpstore_context *cxt,
> + loff_t off)
> +{
> + u64 zonenum = div_u64(off, cxt->conf.kmsg_size);
> +
> + dev_dbg(&cxt->card->dev, "mark zone %llu unused\n", zonenum);
> + clear_bit(zonenum, cxt->usedmap);
> +}
> +
> +static inline void mmcpstore_block_mark_unused(struct mmcpstore_context *cxt,
> + loff_t off)
> +{
> + u32 zonecnt = cxt->card->erase_size / cxt->conf.kmsg_size;
> + u64 zonenum;
> +
> + off = ALIGN_DOWN(off, cxt->card->erase_size);
> + zonenum = div_u64(off, cxt->conf.kmsg_size);
> + while (zonecnt > 0) {
> + dev_dbg(&cxt->card->dev, "mark zone %llu unused\n", zonenum);
> + clear_bit(zonenum, cxt->usedmap);
> + zonenum++;
> + zonecnt--;
> + }
> +}
> +
> +static inline int mmcpstore_is_used(struct mmcpstore_context *cxt,
> +loff_t off) {
> + u64 zonenum = div_u64(off, cxt->conf.kmsg_size);
> +
> + return test_bit(zonenum, cxt->usedmap); }
> +
> +static int mmcpstore_block_is_used(struct mmcpstore_context *cxt,
> + loff_t off)
> +{
> + u32 zonecnt = cxt->card->erase_size / cxt->conf.kmsg_size;
> + u64 zonenum;
> +
> + off = ALIGN_DOWN(off, cxt->card->erase_size);
> + zonenum = div_u64(off, cxt->conf.kmsg_size);
> + while (zonecnt > 0) {
> + if (test_bit(zonenum, cxt->usedmap))
> + return true;
> + zonenum++;
> + zonecnt--;
> + }
> + return false;
> +}
> +
> +static int mmcpstore_is_empty(struct mmcpstore_context *cxt, char *buf,
> + size_t size)
> +{
> + size_t sz;
> + int i;
> +
> + sz = min_t(uint32_t, size, SECTOR_SIZE / 4);
> + for (i = 0; i < sz; i++) {
> + if (buf[i] != (char)0x00)
> + return false;
> + }
> +
> + return true;
> +}
> +
> +static void mmcpstore_mark_removed(struct mmcpstore_context *cxt,
> +loff_t off) {
> + u64 zonenum = div_u64(off, cxt->conf.kmsg_size);
> +
> + dev_dbg(&cxt->card->dev, "mark zone %llu removed\n", zonenum);
> + set_bit(zonenum, cxt->rmmap);
> +}
> +
> +static void mmcpstore_block_clear_removed(struct mmcpstore_context *cxt,
> + loff_t off)
> +{
> + u32 zonecnt = cxt->card->erase_size / cxt->conf.kmsg_size;
> + u64 zonenum;
> +
> + off = ALIGN_DOWN(off, cxt->card->erase_size);
> + zonenum = div_u64(off, cxt->conf.kmsg_size);
> + while (zonecnt > 0) {
> + clear_bit(zonenum, cxt->rmmap);
> + zonenum++;
> + zonecnt--;
> + }
> +}
> +
> +static int mmcpstore_block_is_removed(struct mmcpstore_context *cxt,
> + loff_t off)
> +{
> + u32 zonecnt = cxt->card->erase_size / cxt->conf.kmsg_size;
> + u64 zonenum;
> +
> + off = ALIGN_DOWN(off, cxt->card->erase_size);
> + zonenum = div_u64(off, cxt->conf.kmsg_size);
> + while (zonecnt > 0) {
> + if (test_bit(zonenum, cxt->rmmap))
> + return true;
> + zonenum++;
> + zonecnt--;
> + }
> + return false;
> +}
> +
> +static int mmcpstore_erase_do(struct mmcpstore_context *cxt, size_t
> +size, loff_t off) {
> + int ret;
> +
> + off = ALIGN_DOWN(off, cxt->card->erase_size);
> + dev_dbg(&cxt->card->dev, "try to erase off 0x%llx\n", off);
> +
> + ret = mmc_erase(cxt->card, cxt->start_sect, size >> 9,
> + MMC_ERASE_ARG);
> + if (!ret)
> + mmcpstore_block_clear_removed(cxt, off);
> + else
> + dev_err(&cxt->card->dev,
> + "erase of region [0x%llx, 0x%llx] on \"%s\" failed\n",
> + off, (long long unsigned) cxt->card->erase_size,
> + cxt->conf.device);
> +
> + return ret;
> +}
> +
> +/*
> + * called while removing file
> + *
> + * Avoiding over erasing, do erase block only when the whole block is unused.
> + * If the block contains valid log, do erase lazily on flush_removed()
> +when
> + * unregister.
> + */
> +static ssize_t mmcpstore_erase(size_t size, loff_t off) {
> + struct mmcpstore_context *cxt = &oops_cxt;
> +
> + mmcpstore_mark_unused(cxt, off);
> +
> + /* If the block still has valid data, mmcpstore do erase lazily */
> + if (likely(mmcpstore_block_is_used(cxt, off))) {
> + mmcpstore_mark_removed(cxt, off);
> + return 0;
> + }
> +
> + /* all zones are unused, erase it */
> + return mmcpstore_erase_do(cxt, size, off); }
> +
> +/*
> + * What is security for mmcpstore?
> + * As there is no erase for panic case, we should ensure at least one
> +zone
> + * is writable. Otherwise, panic write will fail.
> + * If zone is used, write operation will return -ENOMSG, which means
> +that
> + * pstore/blk will try one by one until gets an empty zone. So, it is
> +not
> + * needed to ensure the next zone is empty, but at least one.
> + */
> +static int mmcpstore_security(struct mmcpstore_context *cxt, loff_t
> +off) {
> + int ret = 0, i;
> + u32 zonenum = (u32)div_u64(off, cxt->conf.kmsg_size);
> + u32 zonecnt = (u32)div_u64(cxt->size, cxt->conf.kmsg_size);
> + u32 blkcnt = (u32)div_u64(cxt->size, cxt->card->erase_size);
> + u32 erasesize = cxt->card->erase_size;
> +
> + for (i = 0; i < zonecnt; i++) {
> + u32 num = (zonenum + i) % zonecnt;
> +
> + /* found empty zone */
> + if (!test_bit(num, cxt->usedmap))
> + return 0;
> + }
> +
> + /* If there is no any empty zone, we have no way but to do erase */
> + while (blkcnt--) {
> + div64_u64_rem(off + erasesize, cxt->size, (u64 *)&off);
> +
> + ret = mmcpstore_erase_do(cxt, cxt->size, off);
> + if (!ret) {
> + mmcpstore_block_mark_unused(cxt, off);
> + break;
> + }
> + }
> +
> + if (ret)
> + dev_err(&cxt->card->dev, "all blocks bad!\n");
> +
> + dev_dbg(&cxt->card->dev, "end security\n");
> +
> + return ret;
> +}
> +
> +static ssize_t mmcpstore_panic_write(const char *buf, size_t size,
> +loff_t off) {
> + struct mmcpstore_context *cxt = &oops_cxt;
> + int ret;
> +
> + /* zone is used, please try next one */
> + if (mmcpstore_is_used(cxt, off))
> + return -ENOMSG;
> +
> + ret = mmcpstore_req(buf, size, cxt->start_sect + off,
> + MMC_PANIC_WRITE_REQ);
> + if (ret)
> + return 0;
> +
> + mmcpstore_mark_used(cxt, off);
> + dev_dbg(&cxt->card->dev, "mmcpstore panic write %llu \n", size);
> +
> + return size;
> +}
> +
> +static ssize_t mmcpstore_write(const char *buf, size_t size, loff_t
> +off) {
> + struct mmcpstore_context *cxt = &oops_cxt;
> + int ret;
> +
> + /* zone is used, please try next one */
> + if (mmcpstore_is_used(cxt, off))
> + return -ENOMSG;
> +
> + ret = mmcpstore_req(buf, size, cxt->start_sect + off, MMC_WRITE_REQ);
> + if (ret)
> + return 0;
> +
> + mmcpstore_mark_used(cxt, off);
> +
> + mmcpstore_security(cxt, off);
> +
> + return size;
> +}
> +
> +static ssize_t mmcpstore_read(char *buf, size_t size, loff_t off) {
> + struct mmcpstore_context *cxt = &oops_cxt;
> + int ret;
> +
> + ret = mmcpstore_req(buf, size, cxt->start_sect + off, MMC_READ_REQ);
> + if (ret)
> + return 0;
> +
> + if (mmcpstore_is_empty(cxt, buf, size))
> + mmcpstore_mark_unused(cxt, off);
> + else
> + mmcpstore_mark_used(cxt, off);
> +
> + mmcpstore_security(cxt, off);
> +
> + return size;
> +}
> +
> +static struct block_device *mmcpstore_open_backend(const char *device)
> +{
> + struct block_device *bdev;
> + dev_t devt;
> +
> + bdev = blkdev_get_by_path(device, FMODE_READ, NULL);
> +
> + if (IS_ERR(bdev)) {
> + devt = name_to_dev_t(device);
> + if (devt == 0)
> + return ERR_PTR(-ENODEV);
> +
> + bdev = blkdev_get_by_dev(devt, FMODE_READ, NULL);
> + if (IS_ERR(bdev))
> + return bdev;
> +}
> +
> + return bdev;
> +}
> +
> +static void mmcpstore_close_backend(struct block_device *bdev) {
> + if (!bdev)
> + return;
> + blkdev_put(bdev, FMODE_READ);
> +}
> +
> +void mmcpstore_register(struct mmc_card *card, const char *disk_name) {
> + struct mmcpstore_context *cxt = &oops_cxt;
> + struct pstore_blk_config *conf = &cxt->conf;
> + struct block_device *bdev;
> + struct mmc_command *stop;
> + struct mmc_command *cmd;
> + struct mmc_request *mrq;
> + struct mmc_data *data;
> + int ret;
> + unsigned long longcnt;
> +
> + ret = pstore_blk_get_config(conf);
> + if (!conf->device[0]) {
> + pr_debug("psblk backend is empty\n");
> + return;
> + }
> +
> + bdev = mmcpstore_open_backend(conf->device);
> + if (IS_ERR(bdev)) {
> + pr_err("%s failed to open with %ld\n",
> + conf->device, PTR_ERR(bdev));
> + return;
> + }
> +
> + mmcpstore_close_backend(bdev);
> +
> + dev_info(&card->dev, "found pstore mmc device %s\n", disk_name);
> +
> + cxt->start_sect = get_start_sect(bdev);
> + cxt->size = bdev_nr_bytes(bdev);
> + cxt->card = card;
> +
> + if (cxt->size < conf->kmsg_size * 2) {
> + dev_err(&card->dev, "pstore partition %s not big enough\n",
> + disk_name);
> + return;
> + }
> +
> + mrq = kzalloc(sizeof(struct mmc_request), GFP_KERNEL);
> + if (!mrq)
> + goto out;
> +
> + cmd = kzalloc(sizeof(struct mmc_command), GFP_KERNEL);
> + if (!cmd)
> + goto free_mrq;
> +
> + stop = kzalloc(sizeof(struct mmc_command), GFP_KERNEL);
> + if (!stop)
> + goto free_cmd;
> +
> + data = kzalloc(sizeof(struct mmc_data), GFP_KERNEL);
> + if (!data)
> + goto free_stop;
> +
> +
> + if (unlikely(conf->kmsg_size % SECTOR_SIZE)) {
> + dev_err(&cxt->card->dev,
> + "record size %lu KB must align to %d Bytes\n",
> + conf->kmsg_size / 1024,
> + SECTOR_SIZE);
> + goto free_data;
> + }
> +
> + longcnt = BITS_TO_LONGS(div_u64(cxt->size, conf->kmsg_size));
> + cxt->rmmap = kcalloc(longcnt, sizeof(long), GFP_KERNEL);
> + if (!cxt->rmmap)
> + goto free_data;
> +
> + cxt->usedmap = kcalloc(longcnt, sizeof(long), GFP_KERNEL);
> + if (!cxt->usedmap) {
> + kfree(cxt->rmmap);
> + goto free_data;
> + }
> +
> +
> + mrq->cmd = cmd;
> + mrq->data = data;
> + mrq->stop = stop;
> + cxt->mrq = mrq;
> +
> + cxt->dev.flags = PSTORE_FLAGS_DMESG;
> + cxt->dev.zone.read = mmcpstore_read;
> + cxt->dev.zone.write = mmcpstore_write;
> + cxt->dev.zone.erase = mmcpstore_erase;
> + cxt->dev.zone.panic_write = mmcpstore_panic_write;
> + cxt->dev.zone.total_size = cxt->size;
> +
> + ret = register_pstore_device(&cxt->dev);
> + if (ret) {
> + pr_err("%s registering with psblk failed (%d)\n",
> + conf->device, ret);
> + goto free_data;
> + }
> +
> + pr_info("%s start sector %lld size %lld registered as psblk backend\n",
> + conf->device, cxt->start_sect, cxt->size);
> + pr_info("%s erase_size %d sector size %d kmesg size %lu \n", conf->device,
> + cxt->card->erase_size, SECTOR_SIZE, conf->kmsg_size);
> +
> + return;
> +
> +free_data:
> + kfree(data);
> +free_stop:
> + kfree(stop);
> +free_cmd:
> + kfree(cmd);
> +free_mrq:
> + kfree(mrq);
> +out:
> + pr_info("%s register pstoreblk backend device failed\n",
> +conf->device);
> +
> + return;
> +}
> +
> +static int mmcpstore_flush_removed_do(struct mmcpstore_context *cxt,
> + loff_t off, size_t size)
> +{
> + u_char *buf;
> + int ret;
> +
> + buf = kmalloc(cxt->card->erase_size, GFP_KERNEL);
> + if (!buf)
> + return -ENOMEM;
> +
> + /* 1st. read to cache */
> + ret = mmcpstore_read(buf, cxt->card->erase_size, off);
> + if (ret < 0)
> + goto free;
> +
> + /* 2nd. erase block */
> + ret = mmcpstore_erase_do(cxt, cxt->size, off);
> + if (ret)
> + goto free;
> +
> + /* 3rd. write back */
> + while (size) {
> + unsigned int zonesize = cxt->conf.kmsg_size;
> +
> + /* there is valid data on block, write back */
> + if (mmcpstore_is_used(cxt, off))
> + {
> + ret = mmcpstore_req(buf, zonesize, cxt->start_sect + off,
> + MMC_WRITE_REQ);
> +
> + if (ret)
> + dev_err(&cxt->card->dev,
> + "write failure at %lld (%u bytes), err %d\n",
> + off, zonesize, ret);
> + }
> +
> + off += zonesize;
> + size -= min_t(unsigned int, zonesize, size);
> + }
> +
> +free:
> + kfree(buf);
> + return ret;
> +}
> +
> +
> +/*
> + * What does mmcpstore_flush_removed() do?
> + * When user remove any log file on pstore filesystem, mmcpstore should
> +do
> + * something to ensure log file removed. If the whole block is no
> +longer used,
> + * it's nice to erase the block. However if the block still contains
> +valid log,
> + * what mmcpstore can do is to erase and write the valid log back.
> + */
> +static int mmcpstore_flush_removed(struct mmcpstore_context *cxt) {
> + int ret;
> + loff_t off;
> + u32 blkcnt = (u32)div_u64(cxt->size, cxt->card->erase_size);
> +
> + for (off = 0; blkcnt > 0; blkcnt--, off += cxt->card->erase_size) {
> +
> + ret = mmcpstore_block_is_removed(cxt, off);
> + if (!ret)
> + continue;
> +
> + ret = mmcpstore_flush_removed_do(cxt, off, cxt->card->erase_size);
> + if (ret)
> + return ret;
> + }
> + return 0;
> +}
> +
> +void unregister_mmcpstore(void)
> +{
> + struct mmcpstore_context *cxt = &oops_cxt;
> +
> + mmcpstore_flush_removed(cxt);
> + unregister_pstore_device(&cxt->dev);
> + kfree(cxt->mrq->data);
> + kfree(cxt->mrq->stop);
> + kfree(cxt->mrq->cmd);
> + kfree(cxt->mrq);
> + kfree(cxt->usedmap);
> + kfree(cxt->rmmap);
> + cxt->card = NULL;
> +}
> diff --git a/drivers/mmc/host/sdhci.c b/drivers/mmc/host/sdhci.c index f3af1bd0f7b9..08e4ecd13693 100644
> --- a/drivers/mmc/host/sdhci.c
> +++ b/drivers/mmc/host/sdhci.c
> @@ -270,11 +270,6 @@ enum sdhci_reset_reason {
>
> static void sdhci_reset_for_reason(struct sdhci_host *host, enum sdhci_reset_reason reason) {
> - if (host->quirks2 & SDHCI_QUIRK2_ISSUE_CMD_DAT_RESET_TOGETHER) {
> - sdhci_do_reset(host, SDHCI_RESET_CMD | SDHCI_RESET_DATA);
> - return;
> - }
> -
> switch (reason) {
> case SDHCI_RESET_FOR_INIT:
> sdhci_do_reset(host, SDHCI_RESET_CMD | SDHCI_RESET_DATA); @@ -378,7 +373,6 @@ static void sdhci_init(struct sdhci_host *host, int soft)
> if (soft) {
> /* force clock reconfiguration */
> host->clock = 0;
> - host->reinit_uhs = true;
> mmc->ops->set_ios(mmc, &mmc->ios);
> }
> }
> @@ -531,6 +525,7 @@ static inline bool sdhci_has_requests(struct sdhci_host *host)
>
> static void sdhci_read_block_pio(struct sdhci_host *host) {
> + unsigned long flags;
> size_t blksize, len, chunk;
> u32 scratch;
> u8 *buf;
> @@ -540,6 +535,8 @@ static void sdhci_read_block_pio(struct sdhci_host *host)
> blksize = host->data->blksz;
> chunk = 0;
>
> + local_irq_save(flags);
> +
> while (blksize) {
> BUG_ON(!sg_miter_next(&host->sg_miter));
>
> @@ -566,10 +563,13 @@ static void sdhci_read_block_pio(struct sdhci_host *host)
> }
>
> sg_miter_stop(&host->sg_miter);
> +
> + local_irq_restore(flags);
> }
>
> static void sdhci_write_block_pio(struct sdhci_host *host) {
> + unsigned long flags;
> size_t blksize, len, chunk;
> u32 scratch;
> u8 *buf;
> @@ -580,6 +580,8 @@ static void sdhci_write_block_pio(struct sdhci_host *host)
> chunk = 0;
> scratch = 0;
>
> + local_irq_save(flags);
> +
> while (blksize) {
> BUG_ON(!sg_miter_next(&host->sg_miter));
>
> @@ -606,6 +608,8 @@ static void sdhci_write_block_pio(struct sdhci_host *host)
> }
>
> sg_miter_stop(&host->sg_miter);
> +
> + local_irq_restore(flags);
> }
>
> static void sdhci_transfer_pio(struct sdhci_host *host) @@ -701,14 +705,16 @@ static int sdhci_pre_dma_transfer(struct sdhci_host *host,
> return sg_count;
> }
>
> -static char *sdhci_kmap_atomic(struct scatterlist *sg)
> +static char *sdhci_kmap_atomic(struct scatterlist *sg, unsigned long
> +*flags)
> {
> - return kmap_local_page(sg_page(sg)) + sg->offset;
> + local_irq_save(*flags);
> + return kmap_atomic(sg_page(sg)) + sg->offset;
> }
>
> -static void sdhci_kunmap_atomic(void *buffer)
> +static void sdhci_kunmap_atomic(void *buffer, unsigned long *flags)
> {
> - kunmap_local(buffer);
> + kunmap_atomic(buffer);
> + local_irq_restore(*flags);
> }
>
> void sdhci_adma_write_desc(struct sdhci_host *host, void **desc, @@ -750,6 +756,7 @@ static void sdhci_adma_table_pre(struct sdhci_host *host,
> struct mmc_data *data, int sg_count)
> {
> struct scatterlist *sg;
> + unsigned long flags;
> dma_addr_t addr, align_addr;
> void *desc, *align;
> char *buffer;
> @@ -781,9 +788,9 @@ static void sdhci_adma_table_pre(struct sdhci_host *host,
> SDHCI_ADMA2_MASK;
> if (offset) {
> if (data->flags & MMC_DATA_WRITE) {
> - buffer = sdhci_kmap_atomic(sg);
> + buffer = sdhci_kmap_atomic(sg, &flags);
> memcpy(align, buffer, offset);
> - sdhci_kunmap_atomic(buffer);
> + sdhci_kunmap_atomic(buffer, &flags);
> }
>
> /* tran, valid */
> @@ -844,6 +851,7 @@ static void sdhci_adma_table_post(struct sdhci_host *host,
> int i, size;
> void *align;
> char *buffer;
> + unsigned long flags;
>
> if (data->flags & MMC_DATA_READ) {
> bool has_unaligned = false;
> @@ -866,9 +874,9 @@ static void sdhci_adma_table_post(struct sdhci_host *host,
> size = SDHCI_ADMA2_ALIGN -
> (sg_dma_address(sg) & SDHCI_ADMA2_MASK);
>
> - buffer = sdhci_kmap_atomic(sg);
> + buffer = sdhci_kmap_atomic(sg, &flags);
> memcpy(buffer, align, size);
> - sdhci_kunmap_atomic(buffer);
> + sdhci_kunmap_atomic(buffer, &flags);
>
> align += SDHCI_ADMA2_ALIGN;
> }
> @@ -1457,7 +1465,7 @@ static void sdhci_set_transfer_mode(struct sdhci_host *host,
> if (host->quirks2 &
> SDHCI_QUIRK2_CLEAR_TRANSFERMODE_REG_BEFORE_CMD) {
> /* must not clear SDHCI_TRANSFER_MODE when tuning */
> - if (!mmc_op_tuning(cmd->opcode))
> + if (cmd->opcode != MMC_SEND_TUNING_BLOCK_HS200)
> sdhci_writew(host, 0x0, SDHCI_TRANSFER_MODE);
> } else {
> /* clear Auto CMD settings for no data CMDs */ @@ -1698,7 +1706,8 @@ static bool sdhci_send_command(struct sdhci_host *host, struct mmc_command *cmd)
> flags |= SDHCI_CMD_INDEX;
>
> /* CMD19 is special in that the Data Present Select should be set */
> - if (cmd->data || mmc_op_tuning(cmd->opcode))
> + if (cmd->data || cmd->opcode == MMC_SEND_TUNING_BLOCK ||
> + cmd->opcode == MMC_SEND_TUNING_BLOCK_HS200)
> flags |= SDHCI_CMD_DATA;
>
> timeout = jiffies;
> @@ -2284,46 +2293,11 @@ void sdhci_set_uhs_signaling(struct sdhci_host *host, unsigned timing) } EXPORT_SYMBOL_GPL(sdhci_set_uhs_signaling);
>
> -static bool sdhci_timing_has_preset(unsigned char timing) -{
> - switch (timing) {
> - case MMC_TIMING_UHS_SDR12:
> - case MMC_TIMING_UHS_SDR25:
> - case MMC_TIMING_UHS_SDR50:
> - case MMC_TIMING_UHS_SDR104:
> - case MMC_TIMING_UHS_DDR50:
> - case MMC_TIMING_MMC_DDR52:
> - return true;
> - }
> - return false;
> -}
> -
> -static bool sdhci_preset_needed(struct sdhci_host *host, unsigned char timing) -{
> - return !(host->quirks2 & SDHCI_QUIRK2_PRESET_VALUE_BROKEN) &&
> - sdhci_timing_has_preset(timing);
> -}
> -
> -static bool sdhci_presetable_values_change(struct sdhci_host *host, struct mmc_ios *ios) -{
> - /*
> - * Preset Values are: Driver Strength, Clock Generator and SDCLK/RCLK
> - * Frequency. Check if preset values need to be enabled, or the Driver
> - * Strength needs updating. Note, clock changes are handled separately.
> - */
> - return !host->preset_enabled &&
> - (sdhci_preset_needed(host, ios->timing) || host->drv_type != ios->drv_type);
> -}
> -
> void sdhci_set_ios(struct mmc_host *mmc, struct mmc_ios *ios) {
> struct sdhci_host *host = mmc_priv(mmc);
> - bool reinit_uhs = host->reinit_uhs;
> - bool turning_on_clk = false;
> u8 ctrl;
>
> - host->reinit_uhs = false;
> -
> if (ios->power_mode == MMC_POWER_UNDEFINED)
> return;
>
> @@ -2349,8 +2323,6 @@ void sdhci_set_ios(struct mmc_host *mmc, struct mmc_ios *ios)
> sdhci_enable_preset_value(host, false);
>
> if (!ios->clock || ios->clock != host->clock) {
> - turning_on_clk = ios->clock && !host->clock;
> -
> host->ops->set_clock(host, ios->clock);
> host->clock = ios->clock;
>
> @@ -2377,17 +2349,6 @@ void sdhci_set_ios(struct mmc_host *mmc, struct mmc_ios *ios)
>
> host->ops->set_bus_width(host, ios->bus_width);
>
> - /*
> - * Special case to avoid multiple clock changes during voltage
> - * switching.
> - */
> - if (!reinit_uhs &&
> - turning_on_clk &&
> - host->timing == ios->timing &&
> - host->version >= SDHCI_SPEC_300 &&
> - !sdhci_presetable_values_change(host, ios))
> - return;
> -
> ctrl = sdhci_readb(host, SDHCI_HOST_CONTROL);
>
> if (!(host->quirks & SDHCI_QUIRK_NO_HISPD_BIT)) { @@ -2408,21 +2369,8 @@ void sdhci_set_ios(struct mmc_host *mmc, struct mmc_ios *ios)
> if (host->version >= SDHCI_SPEC_300) {
> u16 clk, ctrl_2;
>
> - /*
> - * According to SDHCI Spec v3.00, if the Preset Value
> - * Enable in the Host Control 2 register is set, we
> - * need to reset SD Clock Enable before changing High
> - * Speed Enable to avoid generating clock glitches.
> - */
> - clk = sdhci_readw(host, SDHCI_CLOCK_CONTROL);
> - if (clk & SDHCI_CLOCK_CARD_EN) {
> - clk &= ~SDHCI_CLOCK_CARD_EN;
> - sdhci_writew(host, clk, SDHCI_CLOCK_CONTROL);
> - }
> -
> - sdhci_writeb(host, ctrl, SDHCI_HOST_CONTROL);
> -
> if (!host->preset_enabled) {
> + sdhci_writeb(host, ctrl, SDHCI_HOST_CONTROL);
> /*
> * We only need to set Driver Strength if the
> * preset value enable is not set.
> @@ -2444,20 +2392,46 @@ void sdhci_set_ios(struct mmc_host *mmc, struct mmc_ios *ios)
> }
>
> sdhci_writew(host, ctrl_2, SDHCI_HOST_CONTROL2);
> - host->drv_type = ios->drv_type;
> + } else {
> + /*
> + * According to SDHC Spec v3.00, if the Preset Value
> + * Enable in the Host Control 2 register is set, we
> + * need to reset SD Clock Enable before changing High
> + * Speed Enable to avoid generating clock gliches.
> + */
> +
> + /* Reset SD Clock Enable */
> + clk = sdhci_readw(host, SDHCI_CLOCK_CONTROL);
> + clk &= ~SDHCI_CLOCK_CARD_EN;
> + sdhci_writew(host, clk, SDHCI_CLOCK_CONTROL);
> +
> + sdhci_writeb(host, ctrl, SDHCI_HOST_CONTROL);
> +
> + /* Re-enable SD Clock */
> + host->ops->set_clock(host, host->clock);
> }
>
> + /* Reset SD Clock Enable */
> + clk = sdhci_readw(host, SDHCI_CLOCK_CONTROL);
> + clk &= ~SDHCI_CLOCK_CARD_EN;
> + sdhci_writew(host, clk, SDHCI_CLOCK_CONTROL);
> +
> host->ops->set_uhs_signaling(host, ios->timing);
> host->timing = ios->timing;
>
> - if (sdhci_preset_needed(host, ios->timing)) {
> + if (!(host->quirks2 & SDHCI_QUIRK2_PRESET_VALUE_BROKEN) &&
> + ((ios->timing == MMC_TIMING_UHS_SDR12) ||
> + (ios->timing == MMC_TIMING_UHS_SDR25) ||
> + (ios->timing == MMC_TIMING_UHS_SDR50) ||
> + (ios->timing == MMC_TIMING_UHS_SDR104) ||
> + (ios->timing == MMC_TIMING_UHS_DDR50) ||
> + (ios->timing == MMC_TIMING_MMC_DDR52))) {
> u16 preset;
>
> sdhci_enable_preset_value(host, true);
> preset = sdhci_get_preset_value(host);
> ios->drv_type = FIELD_GET(SDHCI_PRESET_DRV_MASK,
> preset);
> - host->drv_type = ios->drv_type;
> }
>
> /* Re-enable SD Clock */
> @@ -3369,6 +3343,8 @@ static void sdhci_adma_show_error(struct sdhci_host *host)
>
> static void sdhci_data_irq(struct sdhci_host *host, u32 intmask) {
> + u32 command;
> +
> /*
> * CMD19 generates _only_ Buffer Read Ready interrupt if
> * use sdhci_send_tuning.
> @@ -3377,7 +3353,9 @@ static void sdhci_data_irq(struct sdhci_host *host, u32 intmask)
> * SDHCI_INT_DATA_AVAIL always there, stuck in irq storm.
> */
> if (intmask & SDHCI_INT_DATA_AVAIL && !host->data) {
> - if (mmc_op_tuning(SDHCI_GET_CMD(sdhci_readw(host, SDHCI_COMMAND)))) {
> + command = SDHCI_GET_CMD(sdhci_readw(host, SDHCI_COMMAND));
> + if (command == MMC_SEND_TUNING_BLOCK ||
> + command == MMC_SEND_TUNING_BLOCK_HS200) {
> host->tuning_done = 1;
> wake_up(&host->buf_ready_int);
> return;
> @@ -3653,6 +3631,24 @@ static irqreturn_t sdhci_irq(int irq, void *dev_id)
> return result;
> }
>
> +static int sdhci_completion_poll(struct mmc_host *mmc, unsigned long
> +msecs) {
> + int result;
> + struct sdhci_host *host = mmc_priv(mmc);
> +
> + DBG("IRQ Poll\n");
> + while(1) {
> + result = sdhci_irq(host->irq, host);
> + mdelay(1);
> + msecs--;
> +
> + if (!msecs)
> + break;
> + }
> +
> + return result;
> +}
> +
> static irqreturn_t sdhci_thread_irq(int irq, void *dev_id) {
> struct sdhci_host *host = dev_id;
> @@ -3790,7 +3786,6 @@ int sdhci_resume_host(struct sdhci_host *host)
> sdhci_init(host, 0);
> host->pwr = 0;
> host->clock = 0;
> - host->reinit_uhs = true;
> mmc->ops->set_ios(mmc, &mmc->ios);
> } else {
> sdhci_init(host, (mmc->pm_flags & MMC_PM_KEEP_POWER)); @@ -3853,7 +3848,6 @@ int sdhci_runtime_resume_host(struct sdhci_host *host, int soft_reset)
> /* Force clock and power re-program */
> host->pwr = 0;
> host->clock = 0;
> - host->reinit_uhs = true;
> mmc->ops->start_signal_voltage_switch(mmc, &mmc->ios);
> mmc->ops->set_ios(mmc, &mmc->ios);
>
> @@ -4055,6 +4049,8 @@ struct sdhci_host *sdhci_alloc_host(struct device *dev,
>
> host->max_timeout_count = 0xE;
>
> + host->mmc_host_ops.req_completion_poll = sdhci_completion_poll;
> +
> return host;
> }
>
> @@ -4754,6 +4750,8 @@ int sdhci_setup_host(struct sdhci_host *host)
> /* This may alter mmc->*_blk_* parameters */
> sdhci_allocate_bounce_buffer(host);
>
> + pr_info("block size, assuming max_blk_sz %u bytes max_req_sz %u \n",
> +mmc->max_blk_size, mmc->max_req_size);
> +
> return 0;
>
> unreg:
> diff --git a/include/linux/mmc/host.h b/include/linux/mmc/host.h index 8fdd3cf971a3..bb908d778843 100644
> --- a/include/linux/mmc/host.h
> +++ b/include/linux/mmc/host.h
> @@ -212,6 +212,9 @@ struct mmc_host_ops {
>
> /* Initialize an SD express card, mandatory for MMC_CAP2_SD_EXP. */
> int (*init_sd_express)(struct mmc_host *host, struct mmc_ios *ios);
> +
> + int (*req_completion_poll)(struct mmc_host *host,
> + unsigned long timeout);
> };
>
> struct mmc_cqe_ops {
> --
> 2.17.1
>
>
>
> Hyperstone GmbH | Reichenaustr. 39a | 78467 Konstanz
> Managing Director: Dr. Jan Peter Berns.
> Commercial register of local courts: Freiburg HRB381782
>
Powered by blists - more mailing lists