[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <163415517425.1358430.17611745343501937698.b4-ty@kernel.org>
Date: Wed, 13 Oct 2021 20:59:46 +0100
From: Mark Brown <broonie@...nel.org>
To: Arnd Bergmann <arnd@...nel.org>,
Thierry Reding <thierry.reding@...il.com>,
Laxman Dewangan <ldewangan@...dia.com>,
Jonathan Hunter <jonathanh@...dia.com>
Cc: Mark Brown <broonie@...nel.org>, linux-spi@...r.kernel.org,
Zhang Qilong <zhangqilong3@...wei.com>,
Geert Uytterhoeven <geert+renesas@...der.be>,
Arnd Bergmann <arnd@...db.de>,
Guenter Roeck <linux@...ck-us.net>,
Dmitry Osipenko <digetx@...il.com>,
linux-kernel@...r.kernel.org,
Yang Yingliang <yangyingliang@...wei.com>,
linux-tegra@...r.kernel.org
Subject: Re: [PATCH] spi: tegra20: fix build with CONFIG_PM_SLEEP=n
On Wed, 13 Oct 2021 16:45:23 +0200, Arnd Bergmann wrote:
> From: Arnd Bergmann <arnd@...db.de>
>
> There is another one of these warnings:
>
> drivers/spi/spi-tegra20-slink.c:1197:12: error: 'tegra_slink_runtime_resume' defined but not used [-Werror=unused-function]
> 1197 | static int tegra_slink_runtime_resume(struct device *dev)
> | ^~~~~~~~~~~~~~~~~~~~~~~~~~
>
> [...]
Applied to
https://git.kernel.org/pub/scm/linux/kernel/git/broonie/spi.git for-next
Thanks!
[1/1] spi: tegra20: fix build with CONFIG_PM_SLEEP=n
commit: 7dc9b9562740d858332894447c9779b146559239
All being well this means that it will be integrated into the linux-next
tree (usually sometime in the next 24 hours) and sent to Linus during
the next merge window (or sooner if it is a bug fix), however if
problems are discovered then the patch may be dropped or reverted.
You may get further e-mails resulting from automated or manual testing
and review of the tree, please engage with people reporting problems and
send followup patches addressing any issues that are reported if needed.
If any updates are required or you are submitting further changes they
should be sent as incremental updates against current git, existing
patches will not be replaced.
Please add any relevant lists and maintainers to the CCs when replying
to this mail.
Thanks,
Mark
Powered by blists - more mailing lists