lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Date:   Tue, 21 Nov 2017 11:44:05 +0100 (CET)
From:   Jiri Kosina <jikos@...nel.org>
To:     luca abeni <luca.abeni@...tannapisa.it>,
        Peter Zijlstra <peterz@...radead.org>,
        Daniel Bristot de Oliveira <bristot@...hat.com>
cc:     Juri Lelli <juri.lelli@....com>,
        Linus Torvalds <torvalds@...ux-foundation.org>,
        Mathieu Poirier <mathieu.poirier@...aro.org>,
        Mike Galbraith <efault@....de>,
        Steven Rostedt <rostedt@...dmis.org>,
        Thomas Gleixner <tglx@...utronix.de>,
        linux-kernel@...r.kernel.org
Subject: [PATCH] sched/deadline: Use bools for the state flags

From: Jiri Kosina <jkosina@...e.cz>

Commit

	799ba82de01e ("sched/deadline: Use C bitfields for the state flags")

converted state flags into one-bit signed int. Signed one-bit type can be 
either 0 or -1, which is going to cause a problem once 1 is assigned to it 
and then the value later tested against 1.

The current code is okay, as all the checks are (non-)zero check, but I 
believe that we'd rather be safe than sorry here; remove the fragility by 
converting the state flags to bool.

This also silences annoying sparse complaints about this very issue when 
compiling any code that includes sched.h.

Fixes: 799ba82de01e ("sched/deadline: Use C bitfields for the state flags")
Cc: luca abeni <luca.abeni@...tannapisa.it>
Cc: Peter Zijlstra (Intel) <peterz@...radead.org>
Cc: Daniel Bristot de Oliveira <bristot@...hat.com>
Cc: Juri Lelli <juri.lelli@....com>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>
Cc: Mathieu Poirier <mathieu.poirier@...aro.org>
Cc: Mike Galbraith <efault@....de>
Cc: Peter Zijlstra <peterz@...radead.org>
Cc: Steven Rostedt <rostedt@...dmis.org>
Cc: Thomas Gleixner <tglx@...utronix.de>
Signed-off-by: Jiri Kosina <jkosina@...e.cz>
---
 include/linux/sched.h   |  8 ++++----
 kernel/sched/core.c     |  8 ++++----
 kernel/sched/deadline.c | 32 ++++++++++++++++----------------
 3 files changed, 24 insertions(+), 24 deletions(-)

diff --git a/include/linux/sched.h b/include/linux/sched.h
index a5dc7c98b0a2..b19fa1b96726 100644
--- a/include/linux/sched.h
+++ b/include/linux/sched.h
@@ -473,10 +473,10 @@ struct sched_dl_entity {
 	 * conditions between the inactive timer handler and the wakeup
 	 * code.
 	 */
-	int				dl_throttled      : 1;
-	int				dl_boosted        : 1;
-	int				dl_yielded        : 1;
-	int				dl_non_contending : 1;
+	bool				dl_throttled;
+	bool				dl_boosted;
+	bool				dl_yielded;
+	bool				dl_non_contending;
 
 	/*
 	 * Bandwidth enforcement timer. Each -deadline task has its
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index 75554f366fd3..f9bbf0f55e0e 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -3738,20 +3738,20 @@ void rt_mutex_setprio(struct task_struct *p, struct task_struct *pi_task)
 	if (dl_prio(prio)) {
 		if (!dl_prio(p->normal_prio) ||
 		    (pi_task && dl_entity_preempt(&pi_task->dl, &p->dl))) {
-			p->dl.dl_boosted = 1;
+			p->dl.dl_boosted = true;
 			queue_flag |= ENQUEUE_REPLENISH;
 		} else
-			p->dl.dl_boosted = 0;
+			p->dl.dl_boosted = false;
 		p->sched_class = &dl_sched_class;
 	} else if (rt_prio(prio)) {
 		if (dl_prio(oldprio))
-			p->dl.dl_boosted = 0;
+			p->dl.dl_boosted = false;
 		if (oldprio < prio)
 			queue_flag |= ENQUEUE_HEAD;
 		p->sched_class = &rt_sched_class;
 	} else {
 		if (dl_prio(oldprio))
-			p->dl.dl_boosted = 0;
+			p->dl.dl_boosted = false;
 		if (rt_prio(oldprio))
 			p->rt.timeout = 0;
 		p->sched_class = &fair_sched_class;
diff --git a/kernel/sched/deadline.c b/kernel/sched/deadline.c
index 2473736c7616..78cb00ae8b2f 100644
--- a/kernel/sched/deadline.c
+++ b/kernel/sched/deadline.c
@@ -133,7 +133,7 @@ void dl_change_utilization(struct task_struct *p, u64 new_bw)
 	rq = task_rq(p);
 	if (p->dl.dl_non_contending) {
 		sub_running_bw(p->dl.dl_bw, &rq->dl);
-		p->dl.dl_non_contending = 0;
+		p->dl.dl_non_contending = false;
 		/*
 		 * If the timer handler is currently running and the
 		 * timer cannot be cancelled, inactive_task_timer()
@@ -251,7 +251,7 @@ static void task_non_contending(struct task_struct *p)
 		return;
 	}
 
-	dl_se->dl_non_contending = 1;
+	dl_se->dl_non_contending = true;
 	get_task_struct(p);
 	hrtimer_start(timer, ns_to_ktime(zerolag_time), HRTIMER_MODE_REL);
 }
@@ -271,7 +271,7 @@ static void task_contending(struct sched_dl_entity *dl_se, int flags)
 		add_rq_bw(dl_se->dl_bw, dl_rq);
 
 	if (dl_se->dl_non_contending) {
-		dl_se->dl_non_contending = 0;
+		dl_se->dl_non_contending = false;
 		/*
 		 * If the timer handler is currently running and the
 		 * timer cannot be cancelled, inactive_task_timer()
@@ -676,9 +676,9 @@ static void replenish_dl_entity(struct sched_dl_entity *dl_se,
 	}
 
 	if (dl_se->dl_yielded)
-		dl_se->dl_yielded = 0;
+		dl_se->dl_yielded = false;
 	if (dl_se->dl_throttled)
-		dl_se->dl_throttled = 0;
+		dl_se->dl_throttled = false;
 }
 
 /*
@@ -1051,7 +1051,7 @@ static inline void dl_check_constrained_dl(struct sched_dl_entity *dl_se)
 	    dl_time_before(rq_clock(rq), dl_next_period(dl_se))) {
 		if (unlikely(dl_se->dl_boosted || !start_dl_timer(p)))
 			return;
-		dl_se->dl_throttled = 1;
+		dl_se->dl_throttled = true;
 		if (dl_se->runtime > 0)
 			dl_se->runtime = 0;
 	}
@@ -1154,7 +1154,7 @@ static void update_curr_dl(struct rq *rq)
 
 throttle:
 	if (dl_runtime_exceeded(dl_se) || dl_se->dl_yielded) {
-		dl_se->dl_throttled = 1;
+		dl_se->dl_throttled = true;
 		__dequeue_task_dl(rq, curr, 0);
 		if (unlikely(dl_se->dl_boosted || !start_dl_timer(curr)))
 			enqueue_task_dl(rq, curr, ENQUEUE_REPLENISH);
@@ -1206,7 +1206,7 @@ static enum hrtimer_restart inactive_task_timer(struct hrtimer *timer)
 		if (p->state == TASK_DEAD && dl_se->dl_non_contending) {
 			sub_running_bw(p->dl.dl_bw, dl_rq_of_se(&p->dl));
 			sub_rq_bw(p->dl.dl_bw, dl_rq_of_se(&p->dl));
-			dl_se->dl_non_contending = 0;
+			dl_se->dl_non_contending = false;
 		}
 
 		raw_spin_lock(&dl_b->lock);
@@ -1216,14 +1216,14 @@ static enum hrtimer_restart inactive_task_timer(struct hrtimer *timer)
 
 		goto unlock;
 	}
-	if (dl_se->dl_non_contending == 0)
+	if (!dl_se->dl_non_contending)
 		goto unlock;
 
 	sched_clock_tick();
 	update_rq_clock(rq);
 
 	sub_running_bw(dl_se->dl_bw, &rq->dl);
-	dl_se->dl_non_contending = 0;
+	dl_se->dl_non_contending = false;
 unlock:
 	task_rq_unlock(rq, p, &rf);
 	put_task_struct(p);
@@ -1492,7 +1492,7 @@ static void yield_task_dl(struct rq *rq)
 	 * it and the bandwidth timer will wake it up and will give it
 	 * new scheduling parameters (thanks to dl_yielded=1).
 	 */
-	rq->curr->dl.dl_yielded = 1;
+	rq->curr->dl.dl_yielded = true;
 
 	update_rq_clock(rq);
 	update_curr_dl(rq);
@@ -1565,7 +1565,7 @@ static void migrate_task_rq_dl(struct task_struct *p)
 	raw_spin_lock(&rq->lock);
 	if (p->dl.dl_non_contending) {
 		sub_running_bw(p->dl.dl_bw, &rq->dl);
-		p->dl.dl_non_contending = 0;
+		p->dl.dl_non_contending = false;
 		/*
 		 * If the timer handler is currently running and the
 		 * timer cannot be cancelled, inactive_task_timer()
@@ -2232,7 +2232,7 @@ static void switched_from_dl(struct rq *rq, struct task_struct *p)
 	 * while SCHED_OTHER in the meanwhile.
 	 */
 	if (p->dl.dl_non_contending)
-		p->dl.dl_non_contending = 0;
+		p->dl.dl_non_contending = false;
 
 	/*
 	 * Since this might be the only -deadline task on the rq,
@@ -2563,9 +2563,9 @@ void __dl_clear_params(struct task_struct *p)
 	dl_se->dl_bw = 0;
 	dl_se->dl_density = 0;
 
-	dl_se->dl_throttled = 0;
-	dl_se->dl_yielded = 0;
-	dl_se->dl_non_contending = 0;
+	dl_se->dl_throttled = false;
+	dl_se->dl_yielded = false;
+	dl_se->dl_non_contending = false;
 }
 
 bool dl_param_changed(struct task_struct *p, const struct sched_attr *attr)

-- 
Jiri Kosina
SUSE Labs

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ