Re: [Clusters_sig] Re: [ANNOUNCE] Linux Cluster Summit 2005

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> Supporting this vision -- a box is in N clusters from M separate
> cluster system vendors, at the same time, and these N clusters know nothing
> about each other -- is in my opinion a reasonable plan of attack

In memory of the Amadeus consortia of companies that built and shipped a commerical single system image cluster Unix called SVR4/MK (based on the Chorus microkernel), I wonder today what the impact of web services based distributed systems management plus virtualized hosting environments might offer this analysis. Perhaps the classic SSI approach to Unix will become something confined to the boundaries of an embedded multi-core cluster that manifests a single computer to the external web services based distributed network. Is distributed fork/exec really the correct scale out primitive or something confined to these new multi-core super-servers. Iow, considering web services, where should the line be drawn between a single (internally clustered) multi-core distributed memory system, versus the larger external network - how many cores per single system image OS instance is expected? 64, 128? Compared to SVR4/MK, these days, what's on the outside looking in, is web services and
  grid.  Returning to the reality of many vendor's enterprise* business, the suitespot for h/a clusters still seems to be somewhere around ~8 dual-CPU nodes with many customers deploying multiple similar clusters. Nodes are never in multiple clusters at once, rather, individual nodes are members of a cluster and that cluster might be a member of a cluster of clusters.

(* excluding HPC)

> in the "galaxy cluster" vision, a cluster can represent itself as a node, to
> other nodes participating in the same cluster, without revealing

Intergalactic scale is interesting but so too is a microscopic view of each node. 
Perhaps it's a cluster too; running a single system image Linux over multiple cores. I suppose this is what you are thinking, presuming your vantage point is a single CPU core.

Robert


>>> David Nicol <davidnicol@xxxxxxxxx> 05/20/05 10:44 PM >>>
On 5/18/05, Daniel Phillips <phillips@xxxxxxxxxx> wrote:
> Linux Cluster Summit 2005
> 
> June 20 and 21, Walldorf, Germany (Near Heidelberg)
> 
> Sponsors: Red Hat, SAP AG, Oracle, HP, and Fujitsu-Siemens.
> 
> The goal of the two-day Linux Cluster Summit workshop is to bring
> together the key individuals who can realize a general purpose
> clustering API for Linux, including, kernel components, userspace
> libraries and internal and external interfaces.  Results of this
> workshop will be presented to the Kernel Summit the following month.


target vision for cluster infrastructure (thoughts on reading
interview with Andrew Morton in
 Ziff-Davis eweek)  
		
April 21, 2005, edited May 20
I was surprised to see that cluster infrastructure is still missing,
yet pleased that the need for it is more widely perceived today
http://www.uwsg.iu.edu/hypermail/linux/kernel/0409.0/0238.html

than it was four years ago when the linux-cluster mailing list was formed.
http://mail.nl.linux.org/linux-cluster/2001-02/msg00000.html

although there is nothing but spam in its archive since july 2002.

A quick review of more recent developments indicates that little has changed.

There is no need for standardization accross cluster infrastructures at any one
installation, and the sense that a discussion is over "whose version
gets included" rather than "what can we add to make things easier for
everyone, even if doing so will actually hurt the prospects for the clustering
infrastructure I am promoting" still leads to benchmark wars whenever the
subject comes up.  So I gather from glancing at discussion on LKML from
last september that there has been some progress but not much.

Four years ago, I proposed a target vision for linux cluster kernel development
which I believe still makes sense. (And now I know to call it a
"target vision!")

At the time,  I had no good answer for why we would bother to implement support
for something that nobody would immediately use, and sky-pie descriptions of
wide area computing grids seemed silly. (they may still.)  

The vision was, that a Linux box could be in multiple clusters at
once, or could, if so configured, be a "cluster router" similar to the
file system shareres/retransmitters one can set up to run interference
between disparate network file systems.

Supporting this vision -- a box is in N clusters from M separate
cluster system vendors, at the same time, and these N clusters know nothing
about each other -- is in my opinion a reasonable plan of attack for selecting
features to include, or interfaces to mandate conformity to in cluster
system vendors, rather than getting into detailed fights about whose
implementation of feature F belongs in the core distribution.

In the automatic process migration model, it is easy to imagine a wide
cluster where each node might represent a cluster rather than a unit, and would
want to hide the details of the cluster it is representing.  Four
years ago, Mosix allowed pretty wide clusters containing nodes not directly
reachable from each other, but node id numbers needed to be unique across
the whole universe.

in the "galaxy cluster" vision, a cluster can represent itself as a node, to
other nodes participating in the same cluster, without revealing
internal details of the galaxy (because from far enough away, a galaxy
looks, at first, like a single star).

The closest thing to implementing this vision that was available when
I last reviewed what was available was implementing Condor to link separate
Mosix clusters.

I remember a few near-consensuses being reached on the linux-cluster
mailing list.

These included:

   Defining a standard interface for  requesting and obtaining
cluster services and enforcing compliance to it makes sense.

   Arguing about whose implementation of any particular clustering
feature is best does not make sense.  (Given free exchange of techniques
and a standard interface, the in fact better techniques will gradually nudge
out the in fact inferior ones with no shoving required.)

    A standard cluster configuration interface (CCI) defined as a fs
of its own makes sense, rather that squatting within /proc

   the CCI can be mounted anywhere, (possibly back
 within /proc) so multiple clusters on the same box will not
collide with each other -- each gets its own CCI, and all syscalls to
cluster parts include a pointer to a cluster configuration object, of which
there can be more than one defined 

   The first order of business therefore was to take a survey of
services provided by clustering infrastructures and work out standardizable
interfaces to these services

That's what I remember.  The survey of services may or may not have
been performed
formally, I know that a survey of cluster services provided can be
done easily -- is done
often, whenever anyone tries to select what kind of cluster they want to set up.

The role of the linux kernel maintainer, in the question of supporting
multiple disparate
cluster platforms, is NOT to choose one, but is to set up ground rules
under which
they can all play nice together.  Just like the file systems all play
nice together currently.

The thought of having two different spindles each with their own file
system is not
something that anyone blinks at anymore, but it was once revolutionary.

Having one computer participating in N clusters, at the same time, may in the
future be a similar non-issue.

Pertaining to the list of cluster services, here's a quick and small list of the
ones that spring to my mind as being valiud for inclusion into the CCI, without
doing too much additional research:

   services (including statistics) that cluster membership provides
to processes on the node should be defined and offered through the CCI

   node identification in each cluster, abstracted into that cluster

   information about other nodes

   extended PID must include cluster-ID as well as node-ID when discussing
   PID extension mechanisms:  if I am process 00F3 on my own node, I might
   have an extended pid of 000400F3 on a cluster in which I am node 4 and an
   extended pid of 001000F3 on a cluster in which I am node sixteen.

   the publish/subscribe model (just read that one today) is very good

   standardize a publish/subscribe RPC interface in terms of operations on
   filesystem entities within the CCI

Based on discussion on the cap-talk mailing list, i'd like to suggest
that publish/subscribe get implemented in terms of one-off capability
tickets, but that's a perfect example of the kind of implementatin
detail I'm saying we
do not need to define.  How a particular clustering system
implements remote procedure call is not relevant to mandating a standard for
how clustering systems, should their engineers choose to have their
product comply with a standard, may present available RPCs in the CCI, and how
processes on nodes subscribed to their clusters may call an available
RPC through
the CCI.

The big insight that I hope you to take away from this e-mail, if you
haven't had it already (I have been out of touch with the insight level of
LKML for a long time) is that clustering integration into the kernel makes
sense as a standards establishment and enforcement effort, rather than a
technology selection and inclusion effort (at least at first -- once
the CCI exists,
cluster providers might all rush to provide their "CCI driver modules" and then
we're back to selection and inclusion) and that is a different kind of effort.

Although not a new kind.  While writing that paragraph I realized that
the file system interface and the entire module interface, and any other
kind of plug-it-in-and-it-works-the-same interface linux supports, sound,
video, et cetera -- are all standards enforcement problems rather than
technology
selection problems.

Not recognizing that clustering is such a problem is what I believe is
holding back cluster infrastructure from appearing in the kernel.

So last septembers thread about message passing services, in my
vision, is improper. The question is not, how do we pass messages,
but, we have nodes that we know by node-id, and we have messages
that we wish to pass to them, how do we provide a mechanism so that
knowing only those things, node Id and message, an entity that
wishes to pass a message can ask the cluster system to pass the message?

Given modularization, it will then be possible to drop in and replace systems as
needed or as appropriate.

--
David L Nicol
Director of Research and Development
Mindtrust LLC, Kansas City, Missouri

_______________________________________________
Clusters_sig mailing list
Clusters_sig@xxxxxxxxxxxxxx
http://lists.osdl.org/mailman/listinfo/clusters_sig


--

Linux-cluster@xxxxxxxxxx
http://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux