lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <A2F3C0B9-FFA2-4EB3-8A20-A0D5D89A8C63@gmail.com>
Date:   Sat, 01 Jun 2019 21:18:42 -0700
From:   "Jonathan Lemon" <jonathan.lemon@...il.com>
To:     "Song Liu" <songliubraving@...com>
Cc:     Networking <netdev@...r.kernel.org>,
        "Kernel Team" <Kernel-team@...com>, bjorn.topel@...el.com,
        magnus.karlsson@...el.com
Subject: Re: [PATCH v3 bpf-next 2/2] libbpf: remove qidconf and better support
 external bpf programs.



On 1 Jun 2019, at 16:05, Song Liu wrote:

>> On May 31, 2019, at 11:57 AM, Jonathan Lemon 
>> <jonathan.lemon@...il.com> wrote:
>>
>> Use the recent change to XSKMAP bpf_map_lookup_elem() to test if
>> there is a xsk present in the map instead of duplicating the work
>> with qidconf.
>>
>> Fix things so callers using XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD
>> bypass any internal bpf maps, so xsk_socket__{create|delete} works
>> properly.
>>
>> Signed-off-by: Jonathan Lemon <jonathan.lemon@...il.com>
>> ---
>> tools/lib/bpf/xsk.c | 79 
>> +++++++++------------------------------------
>> 1 file changed, 16 insertions(+), 63 deletions(-)
>>
>> diff --git a/tools/lib/bpf/xsk.c b/tools/lib/bpf/xsk.c
>> index 38667b62f1fe..7ce7494b5b50 100644
>> --- a/tools/lib/bpf/xsk.c
>> +++ b/tools/lib/bpf/xsk.c
>> @@ -60,10 +60,8 @@ struct xsk_socket {
>> 	struct xsk_umem *umem;
>> 	struct xsk_socket_config config;
>> 	int fd;
>> -	int xsks_map;
>> 	int ifindex;
>> 	int prog_fd;
>> -	int qidconf_map_fd;
>> 	int xsks_map_fd;
>> 	__u32 queue_id;
>> 	char ifname[IFNAMSIZ];
>> @@ -265,15 +263,11 @@ static int xsk_load_xdp_prog(struct xsk_socket 
>> *xsk)
>> 	/* This is the C-program:
>> 	 * SEC("xdp_sock") int xdp_sock_prog(struct xdp_md *ctx)
>> 	 * {
>> -	 *     int *qidconf, index = ctx->rx_queue_index;
>> +	 *     int index = ctx->rx_queue_index;
>> 	 *
>> 	 *     // A set entry here means that the correspnding queue_id
>> 	 *     // has an active AF_XDP socket bound to it.
>> -	 *     qidconf = bpf_map_lookup_elem(&qidconf_map, &index);
>> -	 *     if (!qidconf)
>> -	 *         return XDP_ABORTED;
>> -	 *
>> -	 *     if (*qidconf)
>> +	 *     if (bpf_map_lookup_elem(&xsks_map, &index))
>> 	 *         return bpf_redirect_map(&xsks_map, index, 0);
>> 	 *
>> 	 *     return XDP_PASS;
>> @@ -286,15 +280,10 @@ static int xsk_load_xdp_prog(struct xsk_socket 
>> *xsk)
>> 		BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_1, -4),
>> 		BPF_MOV64_REG(BPF_REG_2, BPF_REG_10),
>> 		BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4),
>> -		BPF_LD_MAP_FD(BPF_REG_1, xsk->qidconf_map_fd),
>> +		BPF_LD_MAP_FD(BPF_REG_1, xsk->xsks_map_fd),
>> 		BPF_EMIT_CALL(BPF_FUNC_map_lookup_elem),
>> 		BPF_MOV64_REG(BPF_REG_1, BPF_REG_0),
>> -		BPF_MOV32_IMM(BPF_REG_0, 0),
>> -		/* if r1 == 0 goto +8 */
>> -		BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 8),
>> 		BPF_MOV32_IMM(BPF_REG_0, 2),
>> -		/* r1 = *(u32 *)(r1 + 0) */
>> -		BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_1, 0),
>> 		/* if r1 == 0 goto +5 */
>> 		BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 5),
>> 		/* r2 = *(u32 *)(r10 - 4) */
>> @@ -366,18 +355,11 @@ static int xsk_create_bpf_maps(struct 
>> xsk_socket *xsk)
>> 	if (max_queues < 0)
>> 		return max_queues;
>>
>> -	fd = bpf_create_map_name(BPF_MAP_TYPE_ARRAY, "qidconf_map",
>> +	fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
>> 				 sizeof(int), sizeof(int), max_queues, 0);
>> 	if (fd < 0)
>> 		return fd;
>> -	xsk->qidconf_map_fd = fd;
>>
>> -	fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
>> -				 sizeof(int), sizeof(int), max_queues, 0);
>> -	if (fd < 0) {
>> -		close(xsk->qidconf_map_fd);
>> -		return fd;
>> -	}
>> 	xsk->xsks_map_fd = fd;
>>
>> 	return 0;
>> @@ -385,10 +367,8 @@ static int xsk_create_bpf_maps(struct xsk_socket 
>> *xsk)
>>
>> static void xsk_delete_bpf_maps(struct xsk_socket *xsk)
>> {
>> -	close(xsk->qidconf_map_fd);
>> +	bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
>> 	close(xsk->xsks_map_fd);
>> -	xsk->qidconf_map_fd = -1;
>> -	xsk->xsks_map_fd = -1;
>> }
>>
>> static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
>> @@ -417,10 +397,9 @@ static int xsk_lookup_bpf_maps(struct xsk_socket 
>> *xsk)
>> 	if (err)
>> 		goto out_map_ids;
>>
>> -	for (i = 0; i < prog_info.nr_map_ids; i++) {
>> -		if (xsk->qidconf_map_fd != -1 && xsk->xsks_map_fd != -1)
>> -			break;
>> +	xsk->xsks_map_fd = -1;
>>
>> +	for (i = 0; i < prog_info.nr_map_ids; i++) {
>> 		fd = bpf_map_get_fd_by_id(map_ids[i]);
>> 		if (fd < 0)
>> 			continue;
>> @@ -431,11 +410,6 @@ static int xsk_lookup_bpf_maps(struct xsk_socket 
>> *xsk)
>> 			continue;
>> 		}
>>
>> -		if (!strcmp(map_info.name, "qidconf_map")) {
>> -			xsk->qidconf_map_fd = fd;
>> -			continue;
>> -		}
>> -
>> 		if (!strcmp(map_info.name, "xsks_map")) {
>> 			xsk->xsks_map_fd = fd;
>> 			continue;
>> @@ -445,40 +419,18 @@ static int xsk_lookup_bpf_maps(struct 
>> xsk_socket *xsk)
>> 	}
>>
>> 	err = 0;
>> -	if (xsk->qidconf_map_fd < 0 || xsk->xsks_map_fd < 0) {
>> +	if (xsk->xsks_map_fd == -1)
>> 		err = -ENOENT;
>> -		xsk_delete_bpf_maps(xsk);
>> -	}
>>
>> out_map_ids:
>> 	free(map_ids);
>> 	return err;
>> }
>>
>> -static void xsk_clear_bpf_maps(struct xsk_socket *xsk)
>> -{
>> -	int qid = false;
>> -
>> -	bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0);
>> -	bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
>> -}
>> -
>> static int xsk_set_bpf_maps(struct xsk_socket *xsk)
>> {
>> -	int qid = true, fd = xsk->fd, err;
>> -
>> -	err = bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, 
>> &qid, 0);
>> -	if (err)
>> -		goto out;
>> -
>> -	err = bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, &fd, 
>> 0);
>> -	if (err)
>> -		goto out;
>> -
>> -	return 0;
>> -out:
>> -	xsk_clear_bpf_maps(xsk);
>> -	return err;
>> +	return bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id,
>> +				   &xsk->fd, 0);
>> }
>>
>> static int xsk_setup_xdp_prog(struct xsk_socket *xsk)
>> @@ -514,6 +466,7 @@ static int xsk_setup_xdp_prog(struct xsk_socket 
>> *xsk)
>>
>> out_load:
>> 	close(xsk->prog_fd);
>> +	xsk->prog_fd = -1;
>
> I found xsk->prog_fd confusing. Why do we need to set it here?

I suppose this one isn't strictly required - I set it as a guard out of 
habit.
xsk is (currently) immediately freed by the caller, so it can be 
removed.


The main logic is:

         xsk->prog_fd = -1;
         if (!(xsk->config.libbpf_flags & 
XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) {
                 err = xsk_setup_xdp_prog(xsk);

The user may pass INHIBIT_PROG_LOAD, which bypasses setting up the xdp 
program
(and any maps associated with the program), allowing installation of a 
custom
program.  The cleanup behavior is then gated on prog_fd being -1,

>
> I think we don't need to call xsk_delete_bpf_maps() in out_load path?

Hmm, there's two out_load paths, but only one needs the delete maps 
call.  Let
me redo the error handling so it's a bit more explicit.


>
>> out_maps:
>> 	xsk_delete_bpf_maps(xsk);
>> 	return err;
>> @@ -643,9 +596,7 @@ int xsk_socket__create(struct xsk_socket 
>> **xsk_ptr, const char *ifname,
>> 		goto out_mmap_tx;
>> 	}
>>
>> -	xsk->qidconf_map_fd = -1;
>> -	xsk->xsks_map_fd = -1;
>> -
>> +	xsk->prog_fd = -1;
>> 	if (!(xsk->config.libbpf_flags & 
>> XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) {
>> 		err = xsk_setup_xdp_prog(xsk);
>> 		if (err)
>> @@ -708,8 +659,10 @@ void xsk_socket__delete(struct xsk_socket *xsk)
>> 	if (!xsk)
>> 		return;
>>
>> -	xsk_clear_bpf_maps(xsk);
>> -	xsk_delete_bpf_maps(xsk);
>> +	if (xsk->prog_fd != -1) {
>> +		xsk_delete_bpf_maps(xsk);
>> +		close(xsk->prog_fd);
>
> Here, we use prog_fd != -1 to gate xsk_delete_bpf_maps(), which is
> confusing. I looked at the code for quite sometime, but still cannot
> confirm it is correct.

See above reasoning - with INHIBIT_PROG_LOAD, there is no 
library-provided
program or maps, so cleanup actions are skipped.
-- 
Jonathan

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ