lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <b637ec0b0711201238i746ca9bala3812ca93479c6fa@mail.gmail.com>
Date:	Tue, 20 Nov 2007 21:38:29 +0100
From:	"Fabio Comolli" <fabio.comolli@...il.com>
To:	"Ingo Molnar" <mingo@...e.hu>
Cc:	linux-kernel@...r.kernel.org
Subject: Re: [patch/backport] CFS scheduler, -v24, for v2.6.24-rc3, v2.6.23.8, v2.6.22.13, v2.6.21.7

Hi.
First, thanks for this patch. I don't have numbers but working with my
laptop feels much better.
Just a question: does the patch include the fix (divide by zero....)
you just posted in the stable review for 2.6.23.9?

Thanks and regards,
Fabio



On Nov 19, 2007 4:17 PM, Ingo Molnar <mingo@...e.hu> wrote:
>
> By popular demand, here is release -v24 of the CFS scheduler patch.
>
> It is a full backport of the latest & greatest scheduler code to
> v2.6.24-rc3, v2.6.23.8, v2.6.22.13, v2.6.21.7. The patches can be
> downloaded from the usual place:
>
>     http://people.redhat.com/mingo/cfs-scheduler/
>
> There's tons of changes since v22 was released:
>
>     36 files changed, 2359 insertions(+), 1082 deletions(-)
>
> that's 187 individual commits from 32 authors.
>
> So even if CFS v22 worked well for you, please try this release too and
> report regressions (if any).
>
> There are countless improvements in -v24 (see the shortlog further below
> for details), but here are a few highlights:
>
>  - improved interactivity via Peter Ziljstra's "virtual slices" feature.
>    As load increases, the scheduler shortens the virtual timeslices that
>    tasks get, so that applications observe the same constant latency for
>    getting on the CPU. (This goes on until the slices reach a minimum
>    granularity value)
>
>  - CONFIG_FAIR_USER_SCHED is now available across all backported
>    kernels and the per user weights are configurable via
>    /sys/kernel/uids/. Group scheduling got refined all around.
>
>  - performance improvements
>
>  - bugfixes
>
> 99% of these changes are already upstream in Linus's git tree and they
> will be released as part of v2.6.24. (there are 4 pending commits that
> are in the small 2.6.24-rc3-v24 patch.)
>
> As usual, any sort of feedback, bugreport, fix and suggestion is more
> than welcome!
>
>         Ingo
>
> ------------------>
> Adrian Bunk (3):
>       sched: make kernel/sched.c:account_guest_time() static
>       sched: proper prototype for kernel/sched.c:migration_init()
>       sched: make sched_nr_latency static
>
> Alexey Dobriyan (1):
>       sched: uninline scheduler
>
> Andi Kleen (5):
>       sched: cleanup: remove unnecessary gotos
>       sched: cleanup: refactor common code of sleep_on / wait_for_completion
>       sched: cleanup: refactor normalize_rt_tasks
>       sched: remove stale comment from sched_group_set_shares()
>       sched: fix return value of wait_for_completion_interruptible()
>
> Arjan van de Ven (1):
>       Make scheduler debug file operations const
>
> Balbir Singh (1):
>       sched: fix delay accounting regression
>
> Christian Borntraeger (1):
>       sched: fix accounting of interrupts during guest execution on s390
>
> Cliff Wickman (1):
>       hotplug cpu: migrate a task within its cpuset
>
> Dhaval Giani (1):
>       sched: group scheduling, sysfs tunables
>
> Dmitry Adamushko (16):
>       sched: clean up struct load_stat
>       sched: clean up schedstat block in dequeue_entity()
>       sched: sched_setscheduler() fix
>       sched: add set_curr_task() calls
>       sched: do not keep current in the tree and get rid of sched_entity::fair_key
>       sched: optimize task_new_fair()
>       sched: simplify sched_class::yield_task()
>       sched: rework enqueue/dequeue_entity() to get rid of set_curr_task()
>       sched: yield fix
>       sched: fix __pick_next_entity()
>       sched: tidy up SCHED_RR
>       sched: cleanup, remove calc_weighted()
>       sched: cleanup, make dequeue_entity() and update_stats_wait_end() similar
>       sched: fix group scheduling for SCHED_BATCH
>       sched: fix __set_task_cpu() SMP race
>       sched: remove activate_idle_task()
>
> Eric Dumazet (1):
>       sched: cleanup, use NSEC_PER_MSEC and NSEC_PER_SEC
>
> Eugene Teo (1):
>       Fix tsk->exit_state usage
>
> Gautham R Shenoy (1):
>       sched: fix rt ptracer monopolizing CPU
>
> Hiroshi Shimamoto (1):
>       sched: clean up sched_fork()
>
> Ingo Molnar (80):
>       sched: fix sysctl_sched_child_runs_first flag
>       sched: resched task in task_new_fair()
>       sched: small sched_debug cleanup
>       sched: debug: track maximum 'slice'
>       sched: uniform tunings
>       sched: use constants if !CONFIG_SCHED_DEBUG
>       sched: remove stat_gran
>       sched: remove precise CPU load
>       sched: remove precise CPU load calculations #2
>       sched: track cfs_rq->curr on !group-scheduling too
>       sched: cleanup: simplify cfs_rq_curr() methods
>       sched: uninline __enqueue_entity()/__dequeue_entity()
>       sched: speed up update_load_add/_sub()
>       sched: clean up calc_weighted()
>       sched: introduce se->vruntime
>       sched: move sched_feat() definitions
>       sched: optimize vruntime based scheduling
>       sched: simplify check_preempt() methods
>       sched: wakeup granularity increase
>       sched: add se->vruntime debugging
>       sched: remove SCHED_FEAT_SKIP_INITIAL
>       sched: add more vruntime statistics
>       sched: debug: update exec_clock only when SCHED_DEBUG
>       sched: remove wait_runtime limit
>       sched: remove wait_runtime fields and features
>       sched: fix delay accounting performance regression
>       sched: prettify /proc/sched_debug output
>       sched: enhance debug output
>       sched: kernel/sched_fair.c whitespace cleanups
>       sched debug: BKL usage statistics
>       sched: remove unneeded tunables
>       sched debug: print settings
>       sched debug: more width for parameter printouts
>       sched: entity_key() fix
>       sched: remove condition from set_task_cpu()
>       sched: remove last_min_vruntime effect
>       sched: undo some of the recent changes
>       sched: fix sign check error in place_entity()
>       sched: fix sched_fork()
>       sched: remove set_leftmost()
>       sched: clean up schedstats, cnt -> count
>       sched: cleanup, remove stale comment
>       sched: mark scheduling classes as const
>       sched: whitespace cleanups
>       sched: vslice fixups for non-0 nice levels
>       sched: optimize schedule() a bit on SMP
>       sched: tweak wakeup granularity
>       sched: run sched_domain_debug() if CONFIG_SCHED_DEBUG=y
>       sched: break out if printing a warning in sched_domain_debug()
>       sched: style cleanup
>       sched: kfree(NULL) is valid
>       sched: cleanup: rename SCHED_FEAT_USE_TREE_AVG to SCHED_FEAT_TREE_AVG
>       sched: cleanup: rename task_grp to task_group
>       sched: cleanup: function prototype cleanups
>       sched: fix: move the CPU check into ->task_new_fair()
>       sched: update comment
>       sched: clean up is_migration_thread()
>       sched: do not normalize kernel threads via SysRq-N
>       sched: do not wakeup-preempt with SCHED_BATCH tasks
>       sched: speed up context-switches a bit
>       sched: reintroduce cache-hot affinity
>       sched: debug: increase width of debug line
>       sched: debug, improve migration statistics
>       sched: allow the immediate migration of cache-cold tasks
>       sched: affine sync wakeups
>       sched: sync wakeups preempt too
>       sched: cleanup, fix spacing
>       sched: cleanup, make struct rq comments more consistent
>       sched: add KERN_CONT annotation
>       sched: fix fastcall mismatch in completion APIs
>       sched: clean up sched_domain_debug()
>       sched: fix style of swap() macro in kernel/sched_fair.c
>       sched: fix style in kernel/sched.c
>       sched: reintroduce SMP tunings again
>       sched: turn off PREEMPT_RESTRICT
>       sched: remove PREEMPT_RESTRICT
>       sched: wakeup preemption fix
>       sched: clean up the wakeup preempt check
>       sched: clean up the wakeup preempt check, #2
>       sched: reorder SCHED_FEAT_ bits
>
> James Bottomley (1):
>       sched: fix incorrect assumption that cpu 0 exists
>
> Ken Chen (2):
>       sched: fix improper load balance across sched domain
>       sched: reduce schedstat variable overhead a bit
>
> Laurent Vivier (2):
>       sched: guest CPU accounting: maintain stats in account_system_time()
>       sched: don't clear PF_VCPU in scheduler
>
> Matthias Kaehlcke (1):
>       sched: use list_for_each_entry_safe() in __wake_up_common()
>
> Michael Neuling (2):
>       Add scaled time to taskstats based process accounting
>       kernel/sched.c: remove bogus comment from account_user_time
>
> Mike Galbraith (3):
>       sched: fix SMP migration latencies
>       sched: fix formatting of /proc/sched_debug
>       sched: prevent wakeup over-scheduling
>
> Milton Miller (7):
>       sched: domain sysctl fixes: use kcalloc()
>       sched: domain sysctl fixes: use for_each_online_cpu()
>       sched: domain sysctl fixes: unregister the sysctl table before domains
>       sched: domain sysctl fixes: do not crash on allocation failure
>       sched: domain sysctl fixes: add terminator comment
>       sched: more robust sd-sysctl entry freeing
>       sched: fix sched_domain sysctl registration again
>
> Oleg Nesterov (3):
>       do CPU_DEAD migrating under read_lock(tasklist) instead of write_lock_irq(tasklist)
>       migration_call(CPU_DEAD): use spin_lock_irq() instead of task_rq_lock()
>       sched: fix SCHED_FIFO tasks & FAIR_GROUP_SCHED
>
> Paul E. McKenney (1):
>       sched: export cpu_clock()
>
> Paul Jackson (2):
>       cpuset: remove sched domain hooks from cpusets
>       cpuset sched_load_balance flag
>
> Paul Menage (4):
>       Task Control Groups: example CPU accounting subsystem
>       Fix cpusets update_cpumask
>       sched: clean up some control group code
>       sched: report CPU usage in CFS cgroup directories
>
> Pavel Emelyanov (3):
>       pid namespaces: changes to show virtual ids to user
>       Uninline find_task_by_xxx set of functions
>       Use helpers to obtain task pid in printks
>
> Peter Williams (2):
>       sched: reduce balance-tasks overhead
>       sched: isolate SMP balancing code a bit more
>
> Peter Zijlstra (21):
>       sched: simplify SCHED_FEAT_* code
>       sched: new task placement for vruntime
>       sched: simplify adaptive latency
>       sched: clean up new task placement
>       sched: add tree based averages
>       sched: handle vruntime 64-bit overflow
>       sched: better min_vruntime tracking
>       sched: add vslice
>       sched debug: check spread
>       sched: max_vruntime() simplification
>       sched: clean up min_vruntime use
>       sched: speed up and simplify vslice calculations
>       sched: another wakeup_granularity fix
>       sched: disable sleeper_fairness on SCHED_BATCH
>       sched: disable forced preemption by default
>       sched: activate task_hot() only on fair-scheduled tasks
>       sched: fix unconditional irq lock
>       sched: fix vslice
>       sched: documentation: place_entity() comments
>       sched: reintroduce the sched_min_granularity tunable
>       sched: avoid large irq-latencies in smp-balancing
>
> S.Caglar Onur (1):
>       sched debug: BKL usage statistics, fix
>
> Satyam Sharma (1):
>       sched: use show_regs() to improve __schedule_bug() output
>
> Srivatsa Vaddagiri (16):
>       sched: group-scheduler core
>       sched: revert recent removal of set_curr_task()
>       sched: fix minor bug in yield
>       sched: print nr_running and load in /proc/sched_debug
>       sched: print &rq->cfs stats
>       sched: clean up code under CONFIG_FAIR_GROUP_SCHED
>       sched: add fair-user scheduler
>       sched: group scheduler wakeup latency fix
>       sched: group scheduler SMP migration fix
>       sched: group scheduler, fix coding style issues
>       sched: group scheduler, fix bloat
>       sched: group scheduler, fix latency
>       sched: fix new task startup crash
>       Hook up group scheduler with control groups
>       sched: move rcu_head to task_group struct
>       sched: fix copy_namespace() <-> sched_fork() dependency in do_fork
>
> Zou Nan hai (1):
>       sched: some proc entries are missed in sched_domain sys_ctl debug code
>
> -
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to majordomo@...r.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at  http://www.tux.org/lkml/
>
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ