lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20240501072449.GH2575892@kernel.org>
Date: Wed, 1 May 2024 08:24:49 +0100
From: Simon Horman <horms@...nel.org>
To: Alexandra Winter <wintera@...ux.ibm.com>
Cc: David Miller <davem@...emloft.net>, Jakub Kicinski <kuba@...nel.org>,
	Paolo Abeni <pabeni@...hat.com>, Eric Dumazet <edumazet@...gle.com>,
	netdev@...r.kernel.org, linux-s390@...r.kernel.org,
	Heiko Carstens <hca@...ux.ibm.com>,
	Vasily Gorbik <gor@...ux.ibm.com>,
	Alexander Gordeev <agordeev@...ux.ibm.com>,
	Thorsten Winkler <twinkler@...ux.ibm.com>,
	Wenjia Zhang <wenjia@...ux.ibm.com>,
	Christian Borntraeger <borntraeger@...ux.ibm.com>,
	Sven Schnelle <svens@...ux.ibm.com>
Subject: Re: [PATCH net] s390/qeth: Fix kernel panic after setting hsuid

+ Wenjia, Christian, Sven

[ I am not trimming this email in order to leave the full
  context for those added to the CC list.
  All I have added below is my own Reviewed-by tag. ]

On Tue, Apr 30, 2024 at 11:10:04AM +0200, Alexandra Winter wrote:
> Symptom:
> When the hsuid attribute is set for the first time on an IQD Layer3
> device while the corresponding network interface is already UP,
> the kernel will try to execute a napi function pointer that is NULL.
> 
> Example:
> ---------------------------------------------------------------------------
> [ 2057.572696] illegal operation: 0001 ilc:1 [#1] SMP
> [ 2057.572702] Modules linked in: af_iucv qeth_l3 zfcp scsi_transport_fc sunrpc nft_fib_inet nft_fib_ipv4 nft_fib_ipv6 nft_fib nft_reject_inet nf_reject_ipv4 nf_reject_ipv6
> nft_reject nft_ct nf_tables_set nft_chain_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 ip_set nf_tables libcrc32c nfnetlink ghash_s390 prng xts aes_s390 des_s390 de
> s_generic sha3_512_s390 sha3_256_s390 sha512_s390 vfio_ccw vfio_mdev mdev vfio_iommu_type1 eadm_sch vfio ext4 mbcache jbd2 qeth_l2 bridge stp llc dasd_eckd_mod qeth dasd_mod
>  qdio ccwgroup pkey zcrypt
> [ 2057.572739] CPU: 6 PID: 60182 Comm: stress_client Kdump: loaded Not tainted 4.18.0-541.el8.s390x #1
> [ 2057.572742] Hardware name: IBM 3931 A01 704 (LPAR)
> [ 2057.572744] Krnl PSW : 0704f00180000000 0000000000000002 (0x2)
> [ 2057.572748]            R:0 T:1 IO:1 EX:1 Key:0 M:1 W:0 P:0 AS:3 CC:3 PM:0 RI:0 EA:3
> [ 2057.572751] Krnl GPRS: 0000000000000004 0000000000000000 00000000a3b008d8 0000000000000000
> [ 2057.572754]            00000000a3b008d8 cb923a29c779abc5 0000000000000000 00000000814cfd80
> [ 2057.572756]            000000000000012c 0000000000000000 00000000a3b008d8 00000000a3b008d8
> [ 2057.572758]            00000000bab6d500 00000000814cfd80 0000000091317e46 00000000814cfc68
> [ 2057.572762] Krnl Code:#0000000000000000: 0000                illegal
>                          >0000000000000002: 0000                illegal
>                           0000000000000004: 0000                illegal
>                           0000000000000006: 0000                illegal
>                           0000000000000008: 0000                illegal
>                           000000000000000a: 0000                illegal
>                           000000000000000c: 0000                illegal
>                           000000000000000e: 0000                illegal
> [ 2057.572800] Call Trace:
> [ 2057.572801] ([<00000000ec639700>] 0xec639700)
> [ 2057.572803]  [<00000000913183e2>] net_rx_action+0x2ba/0x398
> [ 2057.572809]  [<0000000091515f76>] __do_softirq+0x11e/0x3a0
> [ 2057.572813]  [<0000000090ce160c>] do_softirq_own_stack+0x3c/0x58
> [ 2057.572817] ([<0000000090d2cbd6>] do_softirq.part.1+0x56/0x60)
> [ 2057.572822]  [<0000000090d2cc60>] __local_bh_enable_ip+0x80/0x98
> [ 2057.572825]  [<0000000091314706>] __dev_queue_xmit+0x2be/0xd70
> [ 2057.572827]  [<000003ff803dd6d6>] afiucv_hs_send+0x24e/0x300 [af_iucv]
> [ 2057.572830]  [<000003ff803dd88a>] iucv_send_ctrl+0x102/0x138 [af_iucv]
> [ 2057.572833]  [<000003ff803de72a>] iucv_sock_connect+0x37a/0x468 [af_iucv]
> [ 2057.572835]  [<00000000912e7e90>] __sys_connect+0xa0/0xd8
> [ 2057.572839]  [<00000000912e9580>] sys_socketcall+0x228/0x348
> [ 2057.572841]  [<0000000091514e1a>] system_call+0x2a6/0x2c8
> [ 2057.572843] Last Breaking-Event-Address:
> [ 2057.572844]  [<0000000091317e44>] __napi_poll+0x4c/0x1d8
> [ 2057.572846]
> [ 2057.572847] Kernel panic - not syncing: Fatal exception in interrupt
> -------------------------------------------------------------------------------------------
> 
> Analysis:
> There is one napi structure per out_q: card->qdio.out_qs[i].napi
> The napi.poll functions are set during qeth_open().
> 
> Since
> commit 1cfef80d4c2b ("s390/qeth: Don't call dev_close/dev_open (DOWN/UP)")
> qeth_set_offline()/qeth_set_online() no longer call dev_close()/
> dev_open(). So if qeth_free_qdio_queues() cleared
> card->qdio.out_qs[i].napi.poll while the network interface was UP and the
> card was offline, they are not set again.
> 
> Reproduction:
> chzdev -e $devno layer2=0
> ip link set dev $network_interface up
> echo 0 > /sys/bus/ccwgroup/devices/0.0.$devno/online
> echo foo > /sys/bus/ccwgroup/devices/0.0.$devno/hsuid
> echo 1 > /sys/bus/ccwgroup/devices/0.0.$devno/online
> -> Crash (can be enforced e.g. by af_iucv connect(), ip link down/up, ...)
> 
> Note that a Completion Queue (CQ) is only enabled or disabled, when hsuid
> is set for the first time or when it is removed.
> 
> Workarounds:
> - Set hsuid before setting the device online for the first time
> or
> - Use chzdev -d $devno; chzdev $devno hsuid=xxx; chzdev -e $devno;
> to set hsuid on an existing device. (this will remove and recreate the
> network interface)
> 
> Fix:
> There is no need to free the output queues when a completion queue is
> added or removed.
> card->qdio.state now indicates whether the inbound buffer pool and the
> outbound queues are allocated.
> card->qdio.c_q indicates whether a CQ is allocated.
> 
> Fixes: 1cfef80d4c2b ("s390/qeth: Don't call dev_close/dev_open (DOWN/UP)")
> Signed-off-by: Alexandra Winter <wintera@...ux.ibm.com>

Reviewed-by: Simon Horman <horms@...nel.org>

> ---
>  drivers/s390/net/qeth_core_main.c | 61 ++++++++++++++-----------------
>  1 file changed, 27 insertions(+), 34 deletions(-)
> 
> diff --git a/drivers/s390/net/qeth_core_main.c b/drivers/s390/net/qeth_core_main.c
> index f0b8b709649f..a3adaec5504e 100644
> --- a/drivers/s390/net/qeth_core_main.c
> +++ b/drivers/s390/net/qeth_core_main.c
> @@ -364,30 +364,33 @@ static int qeth_cq_init(struct qeth_card *card)
>  	return rc;
>  }
>  
> +static void qeth_free_cq(struct qeth_card *card)
> +{
> +	if (card->qdio.c_q) {
> +		qeth_free_qdio_queue(card->qdio.c_q);
> +		card->qdio.c_q = NULL;
> +	}
> +}
> +
>  static int qeth_alloc_cq(struct qeth_card *card)
>  {
>  	if (card->options.cq == QETH_CQ_ENABLED) {
>  		QETH_CARD_TEXT(card, 2, "cqon");
> -		card->qdio.c_q = qeth_alloc_qdio_queue();
>  		if (!card->qdio.c_q) {
> -			dev_err(&card->gdev->dev, "Failed to create completion queue\n");
> -			return -ENOMEM;
> +			card->qdio.c_q = qeth_alloc_qdio_queue();
> +			if (!card->qdio.c_q) {
> +				dev_err(&card->gdev->dev,
> +					"Failed to create completion queue\n");
> +				return -ENOMEM;
> +			}
>  		}
>  	} else {
>  		QETH_CARD_TEXT(card, 2, "nocq");
> -		card->qdio.c_q = NULL;
> +		qeth_free_cq(card);
>  	}
>  	return 0;
>  }
>  
> -static void qeth_free_cq(struct qeth_card *card)
> -{
> -	if (card->qdio.c_q) {
> -		qeth_free_qdio_queue(card->qdio.c_q);
> -		card->qdio.c_q = NULL;
> -	}
> -}
> -
>  static enum iucv_tx_notify qeth_compute_cq_notification(int sbalf15,
>  							int delayed)
>  {
> @@ -2628,6 +2631,10 @@ static int qeth_alloc_qdio_queues(struct qeth_card *card)
>  
>  	QETH_CARD_TEXT(card, 2, "allcqdbf");
>  
> +	/* completion */
> +	if (qeth_alloc_cq(card))
> +		goto out_err;
> +
>  	if (atomic_cmpxchg(&card->qdio.state, QETH_QDIO_UNINITIALIZED,
>  		QETH_QDIO_ALLOCATED) != QETH_QDIO_UNINITIALIZED)
>  		return 0;
> @@ -2663,10 +2670,6 @@ static int qeth_alloc_qdio_queues(struct qeth_card *card)
>  		queue->priority = QETH_QIB_PQUE_PRIO_DEFAULT;
>  	}
>  
> -	/* completion */
> -	if (qeth_alloc_cq(card))
> -		goto out_freeoutq;
> -
>  	return 0;
>  
>  out_freeoutq:
> @@ -2677,6 +2680,8 @@ static int qeth_alloc_qdio_queues(struct qeth_card *card)
>  	qeth_free_buffer_pool(card);
>  out_buffer_pool:
>  	atomic_set(&card->qdio.state, QETH_QDIO_UNINITIALIZED);
> +	qeth_free_cq(card);
> +out_err:
>  	return -ENOMEM;
>  }
>  
> @@ -2684,11 +2689,12 @@ static void qeth_free_qdio_queues(struct qeth_card *card)
>  {
>  	int i, j;
>  
> +	qeth_free_cq(card);
> +
>  	if (atomic_xchg(&card->qdio.state, QETH_QDIO_UNINITIALIZED) ==
>  		QETH_QDIO_UNINITIALIZED)
>  		return;
>  
> -	qeth_free_cq(card);
>  	for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j) {
>  		if (card->qdio.in_q->bufs[j].rx_skb) {
>  			consume_skb(card->qdio.in_q->bufs[j].rx_skb);
> @@ -3742,24 +3748,11 @@ static void qeth_qdio_poll(struct ccw_device *cdev, unsigned long card_ptr)
>  
>  int qeth_configure_cq(struct qeth_card *card, enum qeth_cq cq)
>  {
> -	int rc;
> -
> -	if (card->options.cq ==  QETH_CQ_NOTAVAILABLE) {
> -		rc = -1;
> -		goto out;
> -	} else {
> -		if (card->options.cq == cq) {
> -			rc = 0;
> -			goto out;
> -		}
> -
> -		qeth_free_qdio_queues(card);
> -		card->options.cq = cq;
> -		rc = 0;
> -	}
> -out:
> -	return rc;
> +	if (card->options.cq == QETH_CQ_NOTAVAILABLE)
> +		return -1;
>  
> +	card->options.cq = cq;
> +	return 0;
>  }
>  EXPORT_SYMBOL_GPL(qeth_configure_cq);
>  
> -- 
> 2.40.1
> 
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ