lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Wed, 16 May 2018 15:04:08 +0100
From:   Roman Gushchin <guro@...com>
To:     Mike Rapoport <rppt@...ux.vnet.ibm.com>
CC:     Shuah Khan <shuah@...nel.org>, Tejun Heo <tj@...nel.org>,
        <linux-kselftest@...r.kernel.org>, <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] selftests: cgroup: memcontrol: add basic test for socket
 accounting

Hi Mike!

On Wed, May 16, 2018 at 02:28:09PM +0300, Mike Rapoport wrote:
> The test verifies that with active TCP traffic memory.current and
> memory.stat.sock have similar values.
> 
> Signed-off-by: Mike Rapoport <rppt@...ux.vnet.ibm.com>
> ---
>  tools/testing/selftests/cgroup/test_memcontrol.c | 184 +++++++++++++++++++++++
>  1 file changed, 184 insertions(+)
> 
> diff --git a/tools/testing/selftests/cgroup/test_memcontrol.c b/tools/testing/selftests/cgroup/test_memcontrol.c
> index beae06c9c899..0efdb1009175 100644
> --- a/tools/testing/selftests/cgroup/test_memcontrol.c
> +++ b/tools/testing/selftests/cgroup/test_memcontrol.c
> @@ -9,6 +9,12 @@
>  #include <sys/stat.h>
>  #include <sys/types.h>
>  #include <unistd.h>
> +#include <sys/socket.h>
> +#include <sys/wait.h>
> +#include <arpa/inet.h>
> +#include <netinet/in.h>
> +#include <netdb.h>
> +#include <errno.h>
>  
>  #include "../kselftest.h"
>  #include "cgroup_util.h"
> @@ -772,6 +778,183 @@ static int test_memcg_oom_events(const char *root)
>  	return ret;
>  }
>  
> +struct tcp_server_args {
> +	unsigned short port;
> +	int ctl[2];
> +};
> +
> +static int tcp_server(const char *cgroup, void *arg)
> +{
> +	struct tcp_server_args *srv_args = arg;
> +	struct sockaddr_in6 saddr = { 0 };
> +	socklen_t slen = sizeof(saddr);
> +	int sk, client_sk, ctl_fd, yes = 1, ret = -1;
> +
> +	close(srv_args->ctl[0]);
> +	ctl_fd = srv_args->ctl[1];
> +
> +	saddr.sin6_family = AF_INET6;
> +	saddr.sin6_addr = in6addr_any;
> +	saddr.sin6_port = htons(srv_args->port);
> +
> +	sk = socket(AF_INET6, SOCK_STREAM, 0);
> +	if (sk < 0)
> +		return ret;
> +
> +	if (setsockopt(sk, SOL_SOCKET, SO_REUSEADDR, &yes, sizeof(yes)) < 0)
> +		goto cleanup;
> +
> +	if (bind(sk, (struct sockaddr *)&saddr, slen)) {
> +		write(ctl_fd, &errno, sizeof(errno));
> +		goto cleanup;
> +	}
> +
> +	if (listen(sk, 1))
> +		goto cleanup;
> +
> +	ret = 0;
> +	if (write(ctl_fd, &ret, sizeof(ret)) != sizeof(ret)) {
> +		ret = -1;
> +		goto cleanup;
> +	}
> +
> +	client_sk = accept(sk, NULL, NULL);
> +	if (client_sk < 0)
> +		goto cleanup;
> +
> +	ret = -1;
> +	for (;;) {
> +		uint8_t buf[0x100000];
> +
> +		if (write(client_sk, buf, sizeof(buf)) <= 0) {
> +			if (errno == ECONNRESET)
> +				ret = 0;
> +			break;
> +		}
> +	}
> +
> +	close(client_sk);
> +
> +cleanup:
> +	close(sk);
> +	return ret;
> +}
> +
> +static int tcp_client(const char *cgroup, unsigned short port)
> +{
> +	const char server[] = "localhost";
> +	struct addrinfo *ai;
> +	char servport[6];
> +	int retries = 0x10; /* nice round number */
> +	int sk, ret;
> +
> +	snprintf(servport, sizeof(servport), "%hd", port);
> +	ret = getaddrinfo(server, servport, NULL, &ai);
> +	if (ret)
> +		return ret;
> +
> +	sk = socket(ai->ai_family, ai->ai_socktype, ai->ai_protocol);
> +	if (sk < 0)
> +		goto free_ainfo;
> +
> +	ret = connect(sk, ai->ai_addr, ai->ai_addrlen);
> +	if (ret < 0)
> +		goto close_sk;
> +
> +	ret = KSFT_FAIL;
> +	while (retries--) {
> +		uint8_t buf[0x100000];
> +		long current, sock;
> +
> +		if (read(sk, buf, sizeof(buf)) <= 0)
> +			goto close_sk;
> +
> +		current = cg_read_long(cgroup, "memory.current");
> +		if (current < 0)
> +			goto close_sk;
> +
> +		sock = cg_read_key_long(cgroup, "memory.stat", "sock ");
> +		if (sock < 0)
> +			goto close_sk;
> +
> +		if (values_close(current, sock, 3)) {
> +			ret = KSFT_PASS;
> +			break;
> +		}

The test is flapping (at least on my dev machine) because of this condition.

I believe it's because of the batching we're using on the page charge path.
So, in theory, it should be possible to calculate the maximum difference
like num_cpus * PAGE_SIZE * batch_size.

Alternatively, just bump allowed error percentage :)

> +	}
> +
> +close_sk:
> +	close(sk);

It would be great to check that sock and current are getting 0 values
after we're closing the socket.


Thanks!

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ