lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <200706121852.58161.Martin@lichtvoll.de>
Date:	Tue, 12 Jun 2007 18:52:51 +0200
From:	Martin Steigerwald <Martin@...htvoll.de>
To:	ck@....kolivas.org
Cc:	linux kernel mailing list <linux-kernel@...r.kernel.org>
Subject: Re: call for more SD versus CFS comparisons (was: Re: [ck] Mainline plans)

Am Dienstag 12 Juni 2007 schrieb Miguel Figueiredo:
> Hi all,
>
> some results based on massing_intr.c by Satoru, can be found on
> http://people.redhat.com/mingo/cfs-scheduler/tools/massive_intr.c

Hi Miquel, Ingo, Con!

I have been a week without internet access. I have been testing 2.6.21.3 + 
sws2 2.2.10 and cfs-v15 and ck2 on both my T42 and my Amarok machine T23 
[1].

From subjective feedback I can't tell a difference anymore. But I usually 
do not play 3D games or use Beryl / Compiz (except for showing it off).

I did not yet test with disabled NEED_RESCHED CFS workaround as Ingo asked 
my in private mail. This is a work-around to fix the audio skipping 
problems on my Amarok machine. I will compile a CFS v16 and test with 
this work-around disabled.

I had one issue with CFS v15 at least, but its not easy to reproduce. I 
usually test Amarok playback while kernel package make (kernel compile), 
opening a lot of KDE apps (Konqueror mainly) and moving the Amarok window 
like mad. After it I close the windows of that KDE apps again. For 
Konqueror I can use the menu entry "Close all instances" since I have 20+ 
of them open. When I do this, all windows will be closed and I 
occassionally get some audio skips there. AFAIR it didn't happen with SD 
there on my tests with it.

One issue on SD I had was some audio skips *directly* after resume (with 
suspend2). I did not see these with CFS-v15.

Both are issues that are not easy to trackdown, test reproducably for me.

I am willing to run other tests on my Amarok machine. Since I can't tell a 
subjective difference anymore this likely involved running some 
benchmarks. Test would be whether sound playback is 100% OK during 
running that benchmark.

To make it comparative it would be nice if a certain test producedure is 
done on several machines by several people. What would be good test 
scenarios? 

I am interested in desktop loads, and would like stuff like this:

1) Generating high load on the X server.

2) Starting loads of applications and stopping them again quickly. Maybe 
its possible to simulate my manually starting KDE apps by clicking around 
in the startbar wildly and stopping them again (to track this CFS issue).

3) Starting one or more computational expensive tasks like kernel 
compiles.

4) Putting load onto the VM layer / block layer. I am not sure about this. 
We are testing CPU schedulers, not IO schedulers, but different VM 
scenarious are still important IMHO.

These are specially interesting IMHO when being combined. I tried to 
simulate this with my manual point and click testings. (Kernel compile + 
starting apps + moving Amarok window like mad.)

My main concerns during those tests are:

1) Is audio playback 100% okay? Instead of only hearing whether there are 
glitches there possible is a way to *measure* them? Could be extended to 
video playback tough I would expect glitches at a certain load... (well 
at some point eventually audio play back will stop too)

2) Is the mouse pointer always respsonsive? (When there is something I do 
not like it is a frozen mouse pointer as I had a lot with earlier CFS 
versions previous to above workaround). Maybe this can be measured too? 

Is that the massive_intr testing?

3) How is interactivity? How long does it take till an applications react 
to a mouse click... but then how on earth do you measure this?

I guess there are other important criterias to have a look - especially 
also the behavior on a server for example - how does the scheduler behave 
on a webserver with many clients issuing differently sized requests?

Any suggestions?

Now what would be a testing procedure that is affordable for testers and 
as relevant to scheduler testing as it could be? My problem with using 
benchmarks is that I first have to spent lots of time to figure out how 
they work and how to produce a reasonable setup. If someone - preferably 
Linux kernel scheduler experts - invent a testing procedure I could 
follow step by step and then post the results it would be easier for me 
to collect some useful informations for you Kernel developers.

Without any more extensive testing aside from above mentioned difficult to 
reproduce artifacts in my subjective perception SD (as in 2.6.21-ck2) and 
CFS-v15 are on par now. So based on that other criteria - code size, 
design criteria, maintainership, responding to bug reports, further 
benchmarks and real world usage tests - would need to be used to decide 
which scheduler should go into the kernel. 

As I read several times SD is smaller regarding to code size. OTOH to me 
it seems that Ingo Molnar has more time and ability to support 
maintainership, but still to what I have seen Con did his best here too 
and fixed all known issues with SD.

[1] http://martin-steigerwald.de/amarok-machine/

Regards,
-- 
Martin 'Helios' Steigerwald - http://www.Lichtvoll.de
GPG: 03B0 0D6C 0040 0710 4AFA  B82F 991B EAAC A599 84C7

Download attachment "signature.asc " of type "application/pgp-signature" (190 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ