[<prev] [next>] [day] [month] [year] [list]
Message-Id: <1212694580.2194.63.camel@haakon2.linux-iscsi.org>
Date: Thu, 05 Jun 2008 12:36:20 -0700
From: "Nicholas A. Bellinger" <nab@...ux-iscsi.org>
To: "Ross S. W. Walker" <rwalker@...allion.com>
Cc: Jerome Martin <tramjoe.merin@...il.com>,
"Linux-iSCSI.org Target Dev"
<linux-iscsi-target-dev@...glegroups.com>,
LKML <linux-kernel@...r.kernel.org>,
Christoph Hellwig <hch@....de>,
"H. Peter Anvin" <hpa@...or.com>, Andrew Morton <akpm@...l.org>
Subject: RE: Vpools and v3.0-UPSTREAM LIO
On Thu, 2008-06-05 at 14:28 -0400, Ross S. W. Walker wrote:
> Jerome Martin wrote:
>
> > On Thu, Jun 5, 2008 at 5:39 PM, Ross S. W. Walker
> > <rwalker@...allion.com> wrote:
> >
> >
> > How about something like "ipool"?
> >
> > Maybe if I had a better picture of what it is your trying to
> > achieve I can throw some better suggestions out too?
> >
> >
> >
> > Well, basically the idea is to have a cluster of machines
> > driven by heartbeatv2 (pacemaker) implementing "storage" and
> > "vhost" roles for arbitrary nodes. The storage part is based
> > on lvm + drbd + lio-target + open/iSCSI + ext3 (ocfs2
> > planned for v2). The vhost part is based on linux-vservers,
> > with plans to add vmware server in v2 and then maybe
> > xen/openVZ depending on contribs. The term "pool" relates to
> > the fact that the vservers are "pooled" by storage chunks,
> > and the "v" stands for vservers. The effect is "Vserver POOLS".
>
> How about a "stateless" storage pool where each node knows
> nothing about any other node in the storage pool.
>
> You have central servers that then organize each node in
> the pool by available storage, performance of that storage,
> etc.
Having central servers (and having to worry about the redundancy of said
servers) would increase the complexity a great deal I would think.
> The storage is then mapped out based on these
> characteristics, redundancy and striping is handled in
> the mapping, also based on performance characteristics
> and served up to servers that need it.
>
Hmmm, so what are the advantages what we have working so far..?
> > An other property of that cluster scheme is that it is
> > designed to work at N+M redundancy (N active, M spares per
> > resource). That could be used for devising a name too.
> >
> > I'm currently trying to find some acronym based on the
> > software stack used :
> > - DRBD
> > - iSCSI (LIO-target, no other target being able to fullfill
> > the project requirements, and either open/iSCSI or core iscsi)
> > - virtual servers (to stay implementation-agnostic)
> > - linux-ha / pacemaker (the actual heartbeat part is
> > non-essential, will be replaced by openAIS in v2 I think)
> >
> > So maybe something like funny combinations like "DIVHA,
> > software that sings" or "PAID, free software" or anything
> > along those lines would work :-) To be honest, I liked
> > "vpools", and don't really have a good idea yet.
>
> How about DIVA?
>
> This product can be added on to already existing iSCSI
> infrastructures and provides integrated high-level
> storage management services that would otherwise
> require separate products.
>
Hmm, not bad. I usually have to go for a walk or two to come up with
good project names.. :-)
--nab
> -Ross
>
> ______________________________________________________________________
> This e-mail, and any attachments thereto, is intended only for use by
> the addressee(s) named herein and may contain legally privileged
> and/or confidential information. If you are not the intended recipient
> of this e-mail, you are hereby notified that any dissemination,
> distribution or copying of this e-mail, and any attachments thereto,
> is strictly prohibited. If you have received this e-mail in error,
> please immediately notify the sender and permanently delete the
> original and any copy or printout thereof.
>
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists