lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Fri, 22 Mar 2019 12:17:05 +0100 From: Greg Kroah-Hartman <gregkh@...uxfoundation.org> To: linux-kernel@...r.kernel.org Cc: Greg Kroah-Hartman <gregkh@...uxfoundation.org>, stable@...r.kernel.org, Guoqing Jiang <gqjiang@...e.com>, Aditya Pakki <pakki001@....edu>, Song Liu <songliubraving@...com> Subject: [PATCH 5.0 206/238] md: Fix failed allocation of md_register_thread 5.0-stable review patch. If anyone has any objections, please let me know. ------------------ From: Aditya Pakki <pakki001@....edu> commit e406f12dde1a8375d77ea02d91f313fb1a9c6aec upstream. mddev->sync_thread can be set to NULL on kzalloc failure downstream. The patch checks for such a scenario and frees allocated resources. Committer node: Added similar fix to raid5.c, as suggested by Guoqing. Cc: stable@...r.kernel.org # v3.16+ Acked-by: Guoqing Jiang <gqjiang@...e.com> Signed-off-by: Aditya Pakki <pakki001@....edu> Signed-off-by: Song Liu <songliubraving@...com> Signed-off-by: Greg Kroah-Hartman <gregkh@...uxfoundation.org> --- drivers/md/raid10.c | 2 ++ drivers/md/raid5.c | 2 ++ 2 files changed, 4 insertions(+) --- a/drivers/md/raid10.c +++ b/drivers/md/raid10.c @@ -3939,6 +3939,8 @@ static int raid10_run(struct mddev *mdde set_bit(MD_RECOVERY_RUNNING, &mddev->recovery); mddev->sync_thread = md_register_thread(md_do_sync, mddev, "reshape"); + if (!mddev->sync_thread) + goto out_free_conf; } return 0; --- a/drivers/md/raid5.c +++ b/drivers/md/raid5.c @@ -7402,6 +7402,8 @@ static int raid5_run(struct mddev *mddev set_bit(MD_RECOVERY_RUNNING, &mddev->recovery); mddev->sync_thread = md_register_thread(md_do_sync, mddev, "reshape"); + if (!mddev->sync_thread) + goto abort; } /* Ok, everything is just fine now */
Powered by blists - more mailing lists