lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <952949.96475.qm@web52503.mail.re2.yahoo.com>
Date:	Sun, 19 Aug 2007 06:22:37 -0700 (PDT)
From:	Marc Perkel <mperkel@...oo.com>
To:	Willy Tarreau <w@....eu>, Jiri Slaby <jirislaby@...il.com>
Cc:	Marc Perkel <mperkel@...oo.com>, linux-kernel@...r.kernel.org
Subject: Re: The vi editor causes brain damage


--- Willy Tarreau <w@....eu> wrote:

> On Sun, Aug 19, 2007 at 09:15:22AM +0200, Jiri Slaby
> wrote:
> > Marc Perkel napsal(a):
> > > Let me give you and example of the difference
> between
> > > Linux open source world brain damaged thinking
> and
> > > what it's like out here in the real world.
> > > 
> > > Go to a directory with 10k files and type:
> > > 
> > > rm *
> > > 
> > > What do you get?
> > > 
> > > /bin/rm: Argument list too long
> > 
> > What does this have to do with rm command?
> 
> Nothing, and no more with linux development. Marc
> confuses shell and rm.
> Under DOS, when he types "del *", the shell calls
> the builtin function
> "del" and passes it only one argument "*". The del
> function is then
> responsible for iterating through the files using
> getfirst/getnext.
> 
> This is also why mostly only builtin shell commands
> support "*", while
> most external commands do not support it, since they
> have to re-implement
> the same code to iterate through the files (try
> "debug c*.com", it will
> not work).
> 
> Under unix, the shell resolves "*" and passes the
> 10000 file names to
> the "rm" command. Now, execve() may fail because
> 10000 names in arguments
> can require too much memory. That's why find and
> xargs were invented!
> 
> The solution is easy : find . -maxdepth 1 | xargs rm
> 
> So this has nothing to do with rm, nor with rm being
> open-source, and
> even less with rm being written with vi, and Marc's
> rant is totally
> wrong and off-topic. Maybe he was drunk when
> posting, or maybe someone
> used his keyboard to make him look like a complete
> fool. Or maybe he
> really is.
> 
> Willy
> (please do not follow up on this OT thread,
> responses to /dev/null)
> 

The important point that you are missing here is that
the Linux world is willing to live with an rm command
that is broken and the Windows and DOS world isn't.
This isn't about the rm command it's about programming
standards. It's about that the Linux community isn't
committed to getting it right.

Just like my thinking outside the box thread when I
try to say "this is broken" people don't go fix it.
Instead I get an explanation why Linux isn't capable
of having an rm command that will delete an unlimited
number of files.

I bet there are Microsoft people out there laughing at
this.

THINK ABOUT IT PEOPLE !!!

20 years, a million programmers, tens of millions of
users and RM is BROKEN. Am I the only one who has a
problem with this? If so - I'm normal - and Linux is a
cult.


Marc Perkel
Junk Email Filter dot com
http://www.junkemailfilter.com


       
____________________________________________________________________________________Ready for the edge of your seat? 
Check out tonight's top picks on Yahoo! TV. 
http://tv.yahoo.com/
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ