collberg@dna.lth.se (Christian S. Collberg) (03/14/91)
A while back I sent out a request for information about distributed applications running on networks of workstations. I received quite a few replies and also requests for copies. I have therefore summarized the responses. Thomas Fahringer of Universitaet Wien reminded me about a similar survey he had made a short while back (which I saved and then forgot about...). His summary, which in turn contains the summary of a survey conducted by Karsten Morisse, is included at the end of this summary. Thanks to everyone who replied! Christian Collberg Christian.Collberg@dna.lth.se ----------------------------------------------------------------------- "The Benevolent Bandit Laboratory: A Testbed for Distributed Algorithms" R.E. Felderman, E.M. Schooler, L. Kleinrock IEEE Journal on Selected Areas in Communications Feb 1989 Vol 7, No 2 pp 303-311 Describes a system we built at UCLA to run distributed applications on a network of PC-ATs running DOS. Bob Felderman feldy@cs.ucla.edu UCLA Computer Science ...!{rutgers,ucbvax}!cs.ucla.edu!feldy ----------------------------------------------------------------------- Check the v-kernel and the x-kernel developed at our CS department. RAUL IZAHI izahi@nova.stanford.edu ----------------------------------------------------------------------- >From sven@tde.lth.se Tue Mar 5 21:49:08 1991 Check out Cosmic Environment/REctive Kernel Sven ----------------------------------------------------------------------- >From jonathan%maligne-lk.cs.UAlberta.CA@scapa.cs.UAlberta.CA I have a chess program that runs on a network of workstations, if that is the type of application you are looking for. The details were published in the Journal of Parallel and Distributed Computing, 1988 I think. ----------------------------------------------------------------------- >From ht@cogsci.edinburgh.ac.uk Wed Mar 6 10:50:04 1991 For a parser, see Thompson, Henry S. 1991, "Chart Parsing for Loosely Coupled Parallel Systems", in Tomita ed. Current Issues in Parsing Technology, Kluwer, ISBN 0-7923-9131-4 (from Proceedings of the Internationl Workshop on Parsing Technologies, 1989). ----------------------------------------------------------------------- >From stratton@computing-maths.cardiff.ac.uk Have you got in touch with Parasoft - they provide Express - that allows running of programs across (say) networked SUN's, or transputers, or Ncube 1/2, or Cray (soon) etc. They guarantee portability! They can probably give you details fo applications available (such as their own circuit router {i think}, and their debugger, profiler etc). Hope this is of some use, Andy.S. ----------------------------------------------------------------------- You might check with: Scientific Computing Associates, Inc. 246 Church St. Suite 307 New Haven, CT 06510 David Gelernter and James Philbin wrote an article in BYTE, May 1990 pp 213.. "Spending Your Free Time" which describes a PC-network implementation of C-Linda. Some of SCA's literature describes applications ported to this system. Jack Merner Jack N. Merner | Jack.Merner@tortuga.SanDiego.NCR.COM NCR Corporation, MS4440 | An otherwise reliable source ! 16550 W. Bernardo Dr. | Ph: (619)485-2184 San Diego, CA 92127 | Fax: (619)485-2020 ----------------------------------------------------------------------- There is a distributed program that can run on networks of NeXT computers. This program is called Zilla. It can be used to distribute the computation of anything you might want (if you can figure out a way to split up the work). NeXT has used it to factor large numbers (that have only a few factors). Zilla is unique in that it can be set up only to use otherwise idle cycles. I think that NeXT is currently testing an improved version of Zilla. You might also want to check into a new language called PCN. PCN can run on networks of NeXT machines (and possibly Suns if you are using the cosmic environment). John Garnett University of Texas at Austin garnett@cs.utexas.edu Department of Computer Science Austin, Texas ----------------------------------------------------------------------- >From zhou@ifi.unizh.ch Thu Mar 7 09:41:30 1991 I am working on a distributed implementation(simulation) of a spreadsheet program on our local network of Sparcstations. Honbo Zhou. ----------------------------------------------------------------------- We have modified a calibration program to run in a distributed fashion, using ISIS on a network of Sun Sparc machines. Try ftp'ing to water.ca.gov, and look for pub/pstscrpt.out. The postscript output is ~ 250K. Ralph Finch 916-445-0088 rfinch@water.ca.gov ...ucbvax!ucdavis!caldwr!rfinch Any opinions expressed are my own; they do not represent the DWR ----------------------------------------------------------------------- We (the distributed problem solving group at EDRC, CMU) are working on distributed AI applications over a network of UNIX workstations. The applications span various domains --(3 active at present)are: 1) Assistance to human decision-makers (operators) in real-time control applications like electric power networks. 2) Design of High-Rise buildings. 3) Solution of a set of complex nonLinear algebriac equations. A startingPoint reference: 1) S.N.Talukdar, V.C.Ramesh, J.C.Nixon, "A distributed system of control specialists for real-time operations", proceedings of the 3rd international symposium on expert systems applications to power systems, Tokyo, Japan, April 1991. Ramesh vcr@cs.cmu.edu ----------------------------------------------------------------------- I recently read "Modules, Objects and Distributed Programming: Issues in RPC and Remote Object Invocation" H.M.Levy, E.D.Tempero in Software- Practice and Experience, Vol21(1), jan. 91. Its an interesting comparison between two programming styles. -- Bruno BARON -- CRI, Ecole des Mines de Paris, 35 rue Saint Honore', 77305 Fontainebleau, -- France -- baron@ensmp.fr -- tel: (1)64.69.48.38 -- fax: (1)64.69.47.09 ----------------------------------------------------------------------- From: segall@caip.rutgers.edu (Ed Segall) Organization: Rutgers Univ., New Brunswick, N.J. Check out recent issues of IEEE Transactions on Parallel and Distributed Systems, and of the Journal of Parallel and Distributed Processing (Academic Press). ----------------------------------------------------------------------- I know of a program ISIS, which can be used for distributed applications, and it is available from ananomous ftp, you better contact : ken@cs.cornell.edu or reen@cs.cornell.edu (administrative Aide) Geert v.d. Heijden heijden@prles6.prl.philips.nl Philips Research Lab. Eindhoven the Netherlands I can give you a information summary on the public version of ISIS, which I received from Ken Birman: ISIS is available in two forms. The description below is for the public copy, which is provided in source form and at no fee (except postage and our manual, but you can also obtain the system and manual by FTP). There is also a commercially enhanced, supported release of the system by a company I founded 3 years ago. I will have a copy of the product announcement sent to you. This version is priced inexpensively and supports everything the public version does, but has a number of useful enhancements. And, needless to say, the possibility of support is important to some potential users. We have several contacts in the Netherlands and eventually hope to have a local support organzation close to you. We also have some contacts who could provide short courses in distributed systems and specifically on programmign with ISIS, if you need this at some stage of your effort. Please let me know if you have other questions. I hope you find the system useful. And, if you run into any problems at all, don't hesitate to ask us for help. We stand behind our software -- both the public release and the commercial one. If you have problems setting things up, we will be happy to help you work around them. Ken Birman --- ISIS V2.1 blurb --- This is to announce the availability of a public distribution of the ISIS System, a toolkit for distributed and fault-tolerant programming. The initial version of ISIS runs on UNIX on SUN, DEC, GOULD, AUX and HP systems; ports to other UNIX-like systems are planned for the future. No kernel changes are needed to support ISIS; you just roll it in and should be able to use it immediately. The current implementation of ISIS performs well in networks of up to about 100-200 sites. Most users, however, run on a smaller number of sites (16-32 is typical) and other sites connect as "remote clients" that don't actually run ISIS directly. In this mode many hundreds of ISIS users can be clustered around a smaller set of ISIS "mother sites"; many users with large networks favor such an architecture. --- Who might find ISIS useful? --- You will find ISIS useful if you are interested in developing relatively sophisticated distributed programs under UNIX (eventu- ally, other systems too). These include programs that distribute computations over multiple processes, need fault-tolerance, coor- dinate activities underway at several places in a network, recover automatically from software and hardware crashes, and/or dynamically reconfigure while maintaining some sort of distri- buted correctness constraint at all times. ISIS is also useful in building certain types of distributed real time systems. Here are examples of problems to which ISIS has been applied: o On the factory floor, we are working with an industrial research group that is using ISIS to program decentralized cell controllers. They need to arrive at a modular, expand- able, fault-tolerant distributed system. ISIS makes it pos- sible for them to build such a system without a huge invest- ment of effort. (The ISIS group also working closely with an automation standards consortium called ANSA, headed by Andrew Herbert in Cambridge). o As part of a network file system, we built an interface to the UNIX NFS (we call ours "DECEIT") that supports tran- sparent file replication and fault-tolerance. DECEIT speaks NFS protocols but employs ISIS internally to maintain a consistent distributed state. For most operations, DECEIT performance is at worst 50-75% of that of a normal NFS -- despite supporting file replication and fault-tolerance. Interestingly, for many common operations, DECEIT substantially outperforms NFS (!) and it is actually fairly hard to come up with workloads that demonstate replication-related degradation. o A parallel "make" program. Here, ISIS was used within a control program that splits up large software recompilation tasks and runs them on idle workstations, tolerating failures and dynamically adapting if a workstation is reclaimed by its owner. o A system for monitoring and reacting to sensors scattered around the network, in software or in hardware. This system, Meta, is actually included as part of our ISIS V2.1 release. We are adding a high level language to it now, Lomita, in which you can specify reactive control rules or embed such rules into your C or Fortran code, or whatever. o In a hospital, we have looked at using ISIS to manage repli- cated data and to coordinate activities that may span multi- ple machines. The problem here is the need for absolute correctness: if a doctor is to trust a network to carry out orders that might impact on patient health, there is no room for errors due to race conditions or failures. At the same time, cost considerations argue for distributed systems that can be expanded slowly in a fully decentralized manner. ISIS addresses both of these issues: it makes it far easier to build a reliable, correct, distributed system that will manage replicated data and provide complex distributed behaviors. And, ISIS is designed to scale well. o For programming numerical algorithms. One group at Cornell used ISIS to distribute matrix computations over large numbers of workstations. They did this because the worksta- tions were available, mostly idle, and added up to a tremen- dous computational engine. Another group, at LANL, uses ISIS in a parallel plasma physics application. o In a graphics rendering application. Over an extended period, a Cornell graphics group (not even in our department) has used ISIS to build distributed rendering software for image generation. They basically use a set of machines as a parallel processor, with a server that farms out rendering tasks and a variable set of slave computing units that join up when their host machine is fairly idle and drop out if the owner comes back to use the machine again. This is a nice load sharing paradigm and makes for sexy demos too. o In a wide-area seismic monitoring system (i.e. a system that has both local-area networks and wide-area connections between them), developed by a company called SAIC on a DARPA contract. The system gathers seismic data remotely, preprocesses it, and ships event descriptions to a free-standing analysis "hub", which must run completely automatically (their people in San Diego don't like to be phoned in the middle of the night to debug problems in Norway). The hub may request data transfers and other complex computations, raising a number of wide-area programming problems. In addition, the hub system itself has a lot of programs in various languages and just keeping it running can be a challenge. o On brokerage and banking trading floors. Here, ISIS tends to be an adjunct to a technology for distributing quotes, because the special solutions for solving that specific problem are so fast that it is hard for us to compete with them (we normally don't have the freedom of specifying the hardware... many "ticker plant vendors" wire the whole floor for you). However, to the extent that these systems have problems requiring fault-tolerance, simple database integration mechanisms, dynamic restart of services, and in general need "reactive monitoring and control" mechanisms, ISIS works well. And, with our newer versions of the ISIS protocols, performance is actually good enough to handle distribution of stock quotes or other information directly in ISIS, although one has to be a bit careful in super performance intensive settings. (The commercial ISIS release should compete well with the sorts of commercial alternatives listed above on a performance basis, but more than 10 trading groups are using ISIS V2.1 despite the fact that it is definitely slower!). The problems above are characterized by several features. First, they would all be very difficult to solve using remote procedure calls or transactions against some shared database. They have complex, distributed correctness constraints on them: what hap- pens at site "a" often requires a coordinated action at site "b" to be correct. And, they do a lot of work in the application program itself, so that the ISIS communication mechanism is not the bottleneck. If you have an application like this, or are interested in taking on this kind of application, ISIS may be a big win for you. Instead of investing resources in building an environment within which to solve your application, using ISIS means that you can tackle the application immediately, and get something working much faster than if you start with RPC (remote procedure calls). On the other hand, don't think of ISIS as competing with RPC or database transactions. We are oriented towards online control and coordination problems, fault-tolerance of main-memory databases, etc. ISIS normally co-exists with other mechanisms, such as conventional streams and RPC, databases, or whatever. The system is highly portable and not very intrusive, and many of our users employ it to control some form of old code running a computation they don't want to touch at any price. --- What ISIS does --- The ISIS system has been under development for several years at Cornell University. After an initial focus on transactional "resilient objects", the emphasis shifted in 1986 to a toolkit style of programming. This approach stresses distributed con- sistency in applications that manage replicated data or that require distributed actions to be taken in response to events occurring in the system. An "event" could be a user request on a distributed service, a change to the system configuration result- ing from a process or site failure or recovery, a timeout, etc. The ISIS toolkit uses a subroutine call style interface similar to the interface to any conventional operating system. The pri- mary difference, however, is that ISIS functions as a meta- operating system. ISIS system calls result in actions that may span multiple processes and machines in the network. Moreover, ISIS provides a novel "virtual consistency" property to its users. This property makes it easy to build software in which currently executing processes behave in a coordinated way, main- tain replicated data, or otherwise satisfy a system-wide correct- ness property. Moreover, virtual synchrony makes even complex operations look atomic, which generally implies that toolkit functions will not interfere with one another. One can take advantage of this to develop distributed ISIS software in a sim- ple step-by-step style, starting with a non-distributed program, then adding replicated data or backup processes for fault- tolerance or higher availability, then extending the distributed solution to support dynamic reconfiguration, etc. ISIS provides a really unique style of distributed programming -- at least if your distributed computing problems run up against the issues we address. For such applications, the ISIS programming style is both easy and intuitive. ISIS is really intended for, and is good at, problems that draw heavily on replication of data and coordination of actions by a set of processes that know about one another's existence. For example, in a factory, one might need to coordinate the actions of a set of machine-controlled drills at a manufacturing cell. Each drill would do its part of the overall work to be done, using a coordinated scheduling policy that avoids collisions between the drill heads, and with fault-tolerance mechanisms to deal with bits breaking. ISIS is ideally suited to solving prob- lems like this one. Similar problems arise in any distributed setting, be it local-area network software for the office or a CAD problem, or the automation of a critical care system in a hospital. ISIS is not intended for transactional database applications. If this is what you need, you should obtain one of the many such systems that are now available. On the other hand, ISIS would be useful if your goal is to build a front-end in a setting that needs databases. The point is that most database systems are designed to avoid interference between simultaneously executing processes. If your application also needs cooperation between processes doing things concurrently at several places, you may find this aspect hard to solve using just a database because databases force the interactions to be done indirectly through the shared data. ISIS is good for solving this kind of problem, because it provides a direct way to replicate control informa- tion, coordinate the actions of the front-end processes, and to detect and react to failures. ISIS itself runs as a user-domain program on UNIX systems sup- porting the TCP/IP protocol suite. It currently is operational on SUN, DEC, GOULD and HP versions of UNIX. Language interfaces for C, C++, FORTRAN, and Common LISP (both Lucid and Allegro) are included, and a new C-Prolog interface is being tested now. Recent ports available in V2.1 include AUX for the Apple Mac. II, AIX on the IBM RS/6000 and also the older PC/RT. A Cray UNICOS port is (still) under development at LANL, and a DEC VMS port is being done by ISIS Distributed Systems, Inc. ISIS runs over Mach on anything that supports Mach but will probably look a little unnatural to you if you use the Mach primitives. We are planning a version of ISIS that would be more transparent in a Mach context, but it will be some time before this becomes available. Meanwhile, you can use ISIS but may find some aspects of the interface inconsistent with the way that Mach does things. The actual set of tools includes the following: o High performance mechanisms supporting lightweight tasks in UNIX, a simple message-passing facility, and a very simple and uniform addressing mechanism. Users do not work directly with things like ports, sockets, binding, connect- ing, etc. ISIS handles all of this. o A process "grouping" facility, which permits processes to dynamically form and leave symbolically-named associations. The system serializes changes to the membership of each group: all members see the same sequence of changes. Groups names can be used as a location-transparent address. o A suite of broadcast protocols integrated with a group addressing mechanism. This suite operates in a way that makes it look as if all broadcasts are received "simultane- ously" by all the members of a group, and are received in the same "view" of group membership. o Ways of obtaining distributed executions. When a request arrives in a group, or a distributed event takes place, ISIS supports any of a variety of execution styles, ranging from a redundant computation to a coordinator-cohort computation in which one process takes the requested actions while oth- ers back it up, taking over if the coordinator fails. o Replicated data with 1-copy consistency guarantees. o Synchronization facilities, based on token passing or read/write locks. o Facilities for watching a for a process or site (computer) to fail or recover, triggering execution of subroutines pro- vided by the user when the watched-for event occurs. If several members of a group watch for the same event, all will see it at the same "time" with respect to arriving mes- sages to the group and other events, such as group member- ship changes. o A facility for joining a group and atomically obtaining copies of any variables or data structures that comprise its "state" at the instant before the join takes place. The programmer who designs a group can specify state information in addition to the state automatically maintained by ISIS. o Automatic restart of applications when a computer recovers from a crash, including log-based recovery (if desired) for cases when all representatives of a service fail simultane- ously. o Ways to build transactions or to deal with transactional files and database systems external to ISIS. ISIS itself doesn't know about files or transactions. However, as noted above, this tool is pretty unsophisticated as transactional tools go... o Spooler/long-haul mechanism, for saving data to be sent to a group next time it recovers, or for sending from one ISIS LAN to another, physically remote one (e.g. from your Norway site to your San Diego installation). Note: ISIS will not normally run over communication links subject to frequent failures, al- though this long-haul interface has no such restrictions. Everything in ISIS is fault-tolerant. Our programming manual has been written in a tutorial style, and gives details on each of these mechanisms. It includes examples of typical small ISIS applications and how they can be solved. The distribution of the system includes demos, such as the parallel make facility men- tioned above; this large ISIS application program illustrates many system features. To summarize, ISIS provides a broad range of tools, including some that require algorithms that would be very hard to support in other systems or to implement by hand. Performance is quite good: most tools require between 1/20 and 1/5 second to execute on a SUN 3/60, although the actual numbers depend on how big processes groups get, the speed of the network, the locations of processes involved, etc. Overall, however, the system is really quite fast when compared with, say, file access over the network. For certain common operations a five to ten-fold performance improvement is expected within two years, as we implement a col- lection of optimizations. The system scales well with the size of the network, and system overhead is largely independent of network size. On a machine that is not participating in any ISIS application, the overhead of having ISIS running is negligible. In certain communication scenarios, ISIS performance can be quite good. These involve streaming data within a single group or certain client-server interaction patterns, and make use of a new BYPASS communication protocol suite. Future ISIS development is likely to stress extensions and optimizations at this level of the system. In addition, a lot of effort is going into scaling the system to larger environments. --- You can get a copy of ISIS now --- Version V2.1 of ISIS is now fully operational and is being made available to the public. This version consists of a C implementations for UNIX, and has been ported to AIX, SUN, UNIX, MACH, ULTRIX, Gould UNIX, HP-UX, AUX and APOLLO UNIX (release 10.1). Performance is uniformly good. A 400 page tutorial and sys- tem manual containing numerous programming examples is also available. Online manual pages are also provided. The remainder of this posting focuses on how to get ISIS, and how to get the manual. Everything is free except bound copies of the manual. Source is included, but the system is in the public domain, and is released on condition that any ports to other sys- tems or minor modifications remain in the public domain. The manual is copyrighted by the project and is available in hard- copy form or as a DVI file, with figures available for free on request. We have placed a compressed TAR images in the following places: * cu-arpa.cs.cornell.edu (anonymous login, binary mode pub/ISISV21.TAR.Z) * Doc: cu-arpa.cs.cornell.edu (pub/ISISV21-DOC.TAR.Z) * uunet.uu.net (anonymous login, binary mode networks/ISIS/ISISV21.TAR.Z) * mcsun.eu.net (anonymous login, binary mode networks/ISIS/ISISV21.TAR.Z) Also available are DVI and PS versions of our manual. Bound copies will be available at $25 each. A package of figures to glue into the DVI version will be provided free of charge. A tape containing ISIS will be provided upon payment of a charge to cover our costs in making the tape. Our resources are limited and we do not wish to do much of this. --- Copyright, restrictions --- V2.1 of ISIS is subject to a restrictive copyright; basically, you can use it without changing it in any way you like, but are not permitted to develop "derivative versions" without discussing this with us. V2.1 differs substantially from V1.3.1, which was released in the public domain and remains available without any restrictions whatsoever. On the other hand, whereas previous versions of ISIS required export licenses to be sent to certain eastern-block countries, the present version seems not to be subject to this restriction. Contact the US Dept. of Commerce for details if you plan to export ISIS to a country that might be subject to restrictions. Any place in Europe, Japan, etc. should be fine and no license is required. --- Commercial support --- We are working with a local company, ISIS Distributed Systems Inc., to provide support services for ISIS. This company will prepare distributions and work to fix bugs. Support contracts are available for an annual fee; without a contract, we will do our best to be helpful but make no promises. Other services that IDS plans to provide will include consulting on fault-tolerant distributed systems design, instruction on how to work with ISIS, bug identification and fixes, and contractual joint software development projects. The company is also prepared to port ISIS to other systems or other programming languages. Contact "birman@gvax.cs.cornell.edu" for more information. --- If you want ISIS, but have questions, let us know --- Send mail to isis@cs.cornell.edu, subject "I want ISIS", with electronic and physical mailing details. We will send you a form for acknowledging agreement with the conditions for release of the software and will later contact you with details on how to actually copy the system off our machine to yours. --- You can read more about ISIS if you like --- The following papers and documents are available from Cornell. We don't distribute papers by e-mail. Requests for papers should be transmitted to "isis@cs.cornell.edu". 1. Exploiting replication. K. Birman and T. Joseph. This is a preprint of a chapter that will appear in: Arctic 88, An advanced course on operating systems, Tromso, Norway (July 1988). 50pp. 2. Reliable broadcast protocols. T. Joseph and K. Birman. This is a preprint of a chapter that will appear in: Arctic 88, An advanced course on operating systems, Tromso, Norway (July 1988). 30pp. 3. ISIS: A distributed programming environment. User's guide and reference manual. K. Birman, T. Joseph, F. Schmuck. Cornell University, March 1988. 275pp. 4. Exploiting virtual synchrony in distributed systems. K. Birman and T. Joseph. Proc. 11th ACM Symposium on Operating Systems Principles (SOSP), Nov. 1987. 12pp. 5. Reliable communication in an unreliable environment. K. Birman and T. Joseph. ACM Transactions on Computer Systems, Feb. 1987. 29pp. 6. Low cost management of replicated data in fault-tolerant distributed systems. T. Joseph and K. Birman. ACM Transac- tions on Computer Systems, Feb. 1986. 15pp. 7. Fast causal multicast. K. Birman, A. Schiper, P. Stephenson. Dept. of Computer Science TR, May 1990. 8. Distributed application management. K. Marzullo, M. Wood, R. Cooper, K. Birman. Dept. of Computer Science TR, June 1990. We will be happy to provide reprints of these papers. Unless we get an overwhelming number of requests, we plan no fees except for the manual. We also maintain a mailing list for individuals who would like to receive publications generated by the project on an ongoing basis. The last two papers can be copied using FTP from cu-arpa.cs.cornell.edu. If you want to learn about the virtual synchrony as an approach to distributed computing, the best place to start is with refer- ence [1]. If you want to learn more about the ISIS system, how- ever, start with the manual. It has been written in a tutorial style and should be easily accessible to anyone familiar with the C programming language. References [7] and [8] are typical of our recent publications (there are others -- contact Maureen Robinson for details). ---------------------------------------------------------------- ---------------------------------------------------------------- /* THANKS TO ALL WHO CONTRIBUTED TO THIS SUMMARY */ HI, Before you read the summary of my request for references on "Parallelizing applications on a multi-workstation network" (I posted about a week ago) two more questions: I got a very interesting hint of two projects dealing exactly with the topic of my netnews request: 1. "The Multi-Satellite Star", by Michael Stumm at Stanford University 2. "Marionette: Support for Highly Parallel Distributed Programs in Unix" by Mark Sullivan, at Univ. of Berkeley, CA Unfortunately I couldn't get any more information about this projects. No e-mail addresses, technical reports nor published paper. Is anyone out there who knows anything about the two projects mentioned above (e-mail addresses, published papers, etc.)? It should be easy for you guys at Stanford and Berkeley University to find out about, right? Please let me know about. Thanks in advance. ------------------------------------------------------------------ Now to the SUMMARY. I got more than 30 answers to my network request. I am still receiving at least 3 responses a day. Anyway I think it is time to post my summary as promised. Some entries are undoubtably incomplete. Corrections and additions are appreciated. I also included the name of the contributor for most of the contributors. I hope that this does not violate some aspect of netiquette that I am unaware of. Please forgive the faux pas otherwise. Thanks to all those who contributed! Thomas Fahringer Universitaet Wien Institut fuer Statistik und Informatik Rathausstrasse 19/II/3 1010 Wien Austria tf@eacpc1.tuwien.ac.at thomasf@globe.edrc.cmu.edu tf%eacpc1.tuwien.ac.at@VMXA.TUWIEN.AC.AT -------------------------------------------------------------- 1. People here at Yale and elsewhere are looking at using C-LINDA to effectively use a network of workstations (homogeneous or heterogeneous). The tuple space is assumed to be distributed throughout the local memories of the nodes and the compiler performs some optimisations to reduce data movement during a process's communication with tuple space. i myself am using C-Linda on a Sun-Sparc network for various linear algebra algorithms and have gotten good speedups. contact Doug Gilmore at "gilmore@sca.com " for further details: >From: David Kaminsky <kaminsky-david@CS.YALE.EDU> I am working on a similar problem using TSnet. TSnet is a network version of Linda. >From: <bremner@cs.sfu.ca> Carriero, N., Gelernter, D. [1986], "The S/Net's Linda Kernel", ACM Transactions on Computer Systems,4,2, May 1986, pp. 110-129. Carriero, N., Gelernter, D. [1988,1], "Linda in Context", Research Report YALEU/DCS/RR-622, Yale University, Department of Computer Science, April 1988. Carriero, N., Gelernter, D. [1988,2], "How to Write Parallel Programs: A Guide to the Perplexed", Research Report YALEU/DCS/RR-628, Yale University, Department of Computer Science, April 1988. Carriero, N., Gelernter, D. [1989] Technical Correspondence, Communications of the ACM, 32,19, pp. 1256-1258 Davidson, C., [1989], ibid, pp. 1249-1251 Gelernter, D. [1984], "Dynamic global name spaces on network computers", Proceedings of International Conference on Parallel Processing, August, 1984, pp. 25-31. Gelernter, D. [1985], "Generative Communication ,in Linda", ACM Transactions on Programming Languages and Systems, 7, 1, pp. 80-112. Leler, Wm [1990], "Linda Meets Unix", IEEE Computer, 23, 2, pp. 43-54. ------------------------------------------------------------- 2. >From: anand@top.cis.syr.edu We just had a discussion of this topic on the net a while back. I have used ISIS for parallel processing of the type that you are interested in. >From: "Chang L. Lee" <clee@polyslo.CalPoly.EDU> You might want to look at ISIS from Cornell. It's a distributed system toolkit. The idea is to build applications served by process groups, with the virtual synchrony model helping to make the actual implementation less painful. --------------------------------------------------------------- 3. TCGMSG Send/receive subroutines .. version 3.0 (9/27/90) -------------------------------------------------------- Robert J. Harrison tel: (708) 972-7197 E-mail: harrison@tcg.anl.gov, harrison@anlchm.bitnet letter: Bldg. 200, Theoretical Chemistry Group, Argonne National Laboratory, 9700 S. Cass Avenue, Argonne, IL 60439. These routines have been written with the objective of providing a robust, portable, high performance message passing system for distributed memory FORTRAN applications. The C interface is also portable. The syntax is nearly identical to that of the iPSC subroutines, but the functionality is restricted to improve efficiency and speed implementation. On machines with vector hardware sustained interprocess communication rates of 6.0Mb/s have been observed. This toolkit (referred to as TCGMSG) only strives to provide the minimal functionallity needed for our applications. It is only a stop gap until some better model becomes widely (and cheaply) available. However, I believe that many (not all) chemistry and physics problems are readily and efficiently coded with this simple functionality, and that such effort will not be wasted when better tools are found. ---------------------------------------------------------------- 4. >From: Rochelle Grober <argosy!rocky@decwrl.dec.com> Have you ever heard of what has been done on Apollo workstations (now owned by HP)? Their animated movies are created by a "master" node grabbing any other computer on the net and shipping off a frame and appropriate code to process the frame using ray tracing algorithms. The master locates available computers, coordinates shippping off a frame and appropriate code to process the frame using ray tracing algorithms. The master locates available computers, coordinates shipping out and retrieving the frames and code, etc. It can grow with the number of computers available. The way most of the movies have been made is that the job is run during off hours and weekends. I believe one of their claims was something like 25,000 hours of compute time to generate one of their shors was accomplished over two weekends using the company's at that time ~1000 node network. I, mayself participated in a project that took our large data files and broke them up into blocks, shipping blocks and copy of the necessary code to up to five other computers on the net. It worked very well, and this was in 1986. The apollo operating system and system support is designed to facilitate this kind of network distribution of work, so the code to do this sort of subtasking took a knowledgeable person a day to write, and perhaps a couple to ensure it worked properly. ------------------------------------------------------------------------- 5. >From: Colin Brough <cmb@castle.edinburgh.ac.uk> Have you heard of CSTools, from Meiko Scientific? It is a message passing environment, originally designed for their Computing Surface reconfigurable processor arrays. There is now a version for a Sun cluster, as well as the Transputer and i860 versions. ----------------------------------------------------------------------- 6. >From: "V.S.Sunderam" <vss@mathcs.emory.edu> We have a system called PVM that does what you describe. A paper on this has appeared in "Concurrency: Practice & Experience" December 1990. ---------------------------------------------------------------------- 7. >From: "Veikko M. Keha" <keha@hydra.Helsinki.FI> I am working on a methodology that tries to automate the process of converting a serial program into objects that are executed parallel in a local area network. This "Implicit Parallelism Model" is targeted to programmers who are used to write serial programs and don't wont to know much about parallelism and its complexity. The model is based on remote procedure calls. The source code is modified by a precompiler to produce such code that remote procedure calls are executed parallel. The model is suitable to be used with any 3rd generation language. I have built some prototypes to demonstrate the model. The language has been C++. ---------------------------------------------------------------------- 8. >From: rgb@mcc.com (Roger Bolick) You posted a request for information concerning multi-workstation programming. If I understand your search correctly then our project may be of interest. It is a programming environment for C++ with a runtime kernel running on either a Unix box or on the native hardware which supports remote objects. This is used on both multi-computer hardware as well as multi-workstations. This means that you program runs on the available hardware as described in that applications configuration file, assigning objects to as many nodes as you have. Of course its not that simple and there are limitations, but that is why its still in research. ------------------------------------------------------------------- 9. >From: Karsten Morisse <kamo@uni-paderborn.de> I collected quite a few responses to my query on how to connect an ensemble of suns into a multiprocessing team. Here is a summary of what I received. (It took some time for the mail to percolate overseas and back, that is the reason for the delay in my replying.) I heard about 9 different projects: 1. ISIS (Cornell) If you want ISIS, send mail to "croft@gvax.cs.cornell.edu," subject "I want ISIS". 2. Cosmic Environment (Caltech) You can obtain a programming guide by sending e-mail to chuck@vlsi.caltech.edu, or postal mail to: 3. DOMINO (U. Maryland, College Park) DOMINO is a message passing environment for parallel computation. See the Computer Science Dept. (U. Maryland) tech report # TR-1648 (April, 1986) by D. P. O'Leary, G. W. Stewart, and R. A. van de Geijn. 4. DPUP (U. of Colorado) DPUP stands for Distributed Processing Utilities Package. What follows is an abstract from a technical report written at the Computer Science Dept. at the University of Colorado by T. J. Garner, et. al "DPUP is a library of utilities that support distributed concurrent computing on a local area network of computers. The library is built upon the interprocess communication facilities in Berkeley Unix 4.2BSD." 5. TORiS (Toronto) TORis implements a shared memory communication model. Contact Orran Krieger at the University of Toronto for more information: UUCP: {decvax,ihnp4,linus,utzoo,uw-beaver}!utcsri!eecg!okrieg ARPA: okrieg%eecg.toronto.edu@relay.cs.net CSNET: okrieg@eecg.toronto.edu CDNNET: okrieg@eecg.toronto.cdn 6. LINDA (Yale, Scientific Computing Associates) Linda is a parallel programming language for shared memory implementations. It is simple and has only six operators. C-linda has been implemented for a network of SUNs in the internet domain. With LAN-LINDA (also called TSnet) you can write parallel or distributed programs in C and run them on a network of workstations. TSnet has been tested on Sun and IBM RT workstations. Contact David Gelernter (project head) or Mauricio Arango at: gelernter@cs.yale.edu arango@cs.yale.edu TSnet and other Linda systems are being distributed through Scientific Computing Associates. Contact Dennis Philbin Scientific Computing Associates 246 Church St., Suite 307 New Haven, CT 06510 203-777-7442 7. SR (U. Arizona) "SR (Synchronizing Resources) is designed for writing distributed programs. The main language constructs are resources and operations. Resources encapsulate processes and variables they share; operations provide the primary mechanism for process interaction. SR provides a novel integratiotion of the mechanisms for invoking and servicing operations. Consequently, all of local and remote procedure call, rendezvous, message passing, dynamic process creation, multicast, and semaphores are supported. An overview of the language and implementation appeared in the January, 1988, issue of TOPLAS (ACM Transactions on Programming Languages and Systems 10,1, 51-86). "SR is available by anonymous FTP from Arizona.EDU (128.196.128.118 or 192.12.69.1). [Copy over the README file for an explanation.] You may reach the members of the SR project electronically at: uunet!Arizona!sr-project or by surface mail at: SR Project Department of Computer Science University of Arizona Tucson, AZ 85721 (602) 621-2018 8. MAITRD (U.C. Berkeley/U. Wash) "The maitr'd software is remote process server that is designed to farm out cpu expensive jobs to less loaded machines. It has a small amount of built-in intelligence, in that it attempts to send jobs to the least loaded machine of the set which is accepting off-site jobs." `Maitrd' is available via anonymous ftp from june.cs.washington.edu (128.95.1.4) as ~ftp/pub/Maitrd.tar.Z. There is also a heterogeneous systems rpc package `hrpc.tar.Z'. Contact Brian Bershad at U. Washington (brian@june.cs.washington.edu.) for more information. 9. PARMACS (Argonne) David Levine at Argonne National Laboratory tells us about a "generic package to do send/recv message passing" with "different versions (c, c++, fortran) [that] work on different machines." For more information, send email to netlib@mcs.anl.gov, with subject (or body) ``send index from parmacs.'' For more information send email to levine@mcs.anl.gov or by uucp: {alliant,sequent,rogue}!anlams!levine. ------------------------------------------------------------------ 10. >From: Karen Tracey <kmt@pclsys48.pcl.nd.edu> My implementation platform for this work is the ARCADE distributed system. If you are not familiar with ARCADE I can also give you references on it. The goal of the ARCADE project is to develop an environment which allows programmers to easily build distributed applications that consist of cooperating tasks which may run on heterogeneous machines. ARCADE contains a number of facilities (data transfer & sharing, inter-task control & synchronization) that simplify development of cooperative distributed applications. "ARCADE: A Platform for Heterogeneous Distributed Operating Systems", by David L. Cohn, William P. Delaney, and Karen M. Tracey, appeared in Proceedings of 1989 USENIX Workship on Experiences with Distributed and Multiprocessor Systems, Fort Lauderdale, FL, October 1989. ----------------------------------------------------------------- 11. >From: David Taylor <ddt@ccwf.cc.utexas.edu> I'm not sure about this, but I believe you may be interested in a system called Amoeba (sp?). I can't remember whether it's a true parallel-processing environment or simply an OS that distributes processes on a network, but it's probably worth looking into. ------------------------------------------------------------------- 12. >From : roman@CCSF.Caltech.EDU (Roman Salvador) We (ParaSoft Corporation) sell a system (parallel libraries, profilers, debugger, semi-automatic data-distributer, automatic parallelizer, ...) to do parallel processing on networks of workstations and most other parallel computers. ------------------------------------------------------------------- 13. >From: Joe Hummel <jhummel@ICS.UCI.EDU> The Univ. of Washington has such a system, Amber. It runs a single appl in parallel on a network of shared-memory workstations (e.g. Dec Firefly). See 12th ACM Syp on Operating Systems, Dec. '89, pp 1147-158. Also, take a look at Munin, at system from Rice Univ. They have it running on a network of Suns. See '90 ACM Sigplan PPoPP conference. ------------------------------------------------------------------ 14. We did some work at UCLA using PCs on an ethernet Felderman, R.E., Schooler, E.M., Kleinrock L., "The Benevolent Bandit Laboratory: A Testbed for Distributed Algorithms", IEEE Journal on Selected Areas in Communications, Vol. 7, No. 2, February 1989. ---------------------------------------------------------------------- 15. "J. Eric Townsend" <JET@uh.edu> I would suggest you look at CalTech's "Cosmic Environment", which lets you write iPSC/2 code and run it on a network of workstations. chuck@vlsi.caltech.edu is who you need to talk to. --------------------------------------------------------------------- 16. >From: Andy Newman <andy@research.canon.oz.au> Get in touch with AT&T and get some information on their Concurrent C product. They have a version that runs on multiple machines in the network and others that run on actual multi-processors. Its basically C with Hoare's (sp?) CSP constructs added to it (a sort of super-Occam). ----------------------------------------------------------------------- 17. >From: paddy@daimi.aau.dk About 2 years ago I was involved in a project which given an Ada program with (logical) site annotations converted it into a set of Ada programs which could be compiled using existing compilers and the code run on separate machines connected by an ethernet. There was a set of utilities which invoked our convertor, followed by the Ada compiler/linker etc. ----------------------------------------------------------------------- 18. >From: Paulo V Rocha <P.Rocha@cs.ucl.ac.uk> The PYGMALION Programming Environment, a ESPRIT II project, uses a multi-workstation environment (up to 3 workstations if I am not wrong) to run neural network applications. It uses remote procedure calls to communicate. ---------------------------------------------------------------------- 19. >From: adamb@cs.utk.edu We're currently working on a system called DagNet which will allow the programmer to specify subroutines and their dependencies and then have the subroutines scheduled around the internet. In DagNet the data distribution is precisely associated with the dependencies. We currently have a prototype working but the the system is still going through changes. I also developed a system called Phred which allows the visual specification and analysis of parallel programs. Currently I'm working on designing an execution system which will actually execute Phred programs over a network of machines. Unfortunately the executioner is still on paper at this point. Phred does have interesting properties for sharing data among parallel processes which might interest you. ------------------------------------------------------------------ 20. >From: eric@castle.edinburgh.ac.uk Meiko (from Bristol UK -- don't have actual address but somebody on the net must) sell a product called CS-Tools which will run jobs over a network of Suns (SPARCstations and SLC etc) and their own boxes. Don't know how well it works. (I use it every day but I only run on Meiko's Computing Surface so can't verify its behaviour on the Suns.) -------------------------------------------------------------------- 21. >From: Ozalp Babaoglu <ozalp@dm.unibo.it> Paralex: An Environment for Parallel Programming in Distributed Systems One of the many advantages of distributed systems is their ability to execute several computations on behalf of a single application in parallel, thus improving performance. In fact, at a certain level of aabstraction, there is little difference between a distributed system and a losely-coupled multiprocessor computer. We cannot, however, treat distributed systems as if they were uniform multiprocessor parallel machines due to the following characteristics: o High latency, low bandwidth communication o Presence of heterogeneous processor architectures o Communication link and processor failures o Multiple independent administrative domains. Thus, if we can address these issues, a distributed computing resource such as a collection of workstations could be viewed and used as if it were a poor man's ``super computer.'' To make a distributed system suitable for long-running parallel computations, support must be provided for fault tolerance. Many hours of computation can be wasted not only if there are hardware failures, but also if one of the processors is turned off, rebooted or disconnected from the network. Given that the components of the system (workstations) may be under the control of several independent administrative domais (typically a single individual who ``owns'' the workstation), these events are much more plausible and frequent than real hardware failures. ---------------------------------------------------------------------- 22. >From: eisen@cc.gatech.edu (Greg Eisenhauer) I'm not sure if it's exactly what you're looking for, but you might look at my old work on Distributed Ada. The idea was that you took a single application (program) and gave it to a compiler along with a specification of how you wanted parts of it to be distributed across a distributed memory architecture. We did most of the development here on a network of Suns and eventually moved to an Intel iPSC/2 Hypercube. ----------------------------------------------------------------------- 23. >From: Ralph Noack <mdaeng!rwn@utacfd.uta.edu> We've bought a package called Express by ParaSoft. It runs on sun workstation networks and pc or macs with transputer cards. It has two different modes of programming for parallelizing a task. 1) cubix: a single executable with appropriate calls to library routines to: find id number, exchange data between nodes, etc. I've written a small program with solves a simple pde on multiple nodes. So it works. I could not get the main application running do to problems with their current preprocessor(it translates fortran write statements to subr calls so all io plays through node 0). They say a new version of the preprocessor will be released soon. 2) host + node executables. A single host task communicated/controls multiple tasks running a different executable. ----------------------------------------------------------------------- 24. from rwolski@lll-crg.llnl.gov (Richard Wolski) About a week ago I posted a request for reference information on partitioning and scheduling for heterogeneous systems. I am pleased to say the the response has been almost overwhelming. While I haven't completely digested everything I have received, here is a summary of what I think is pertinent. Dr. Shahid Bokhari at icase suggests: ------------------------------------- Author = "Shahid H. Bokhari", Year = "July 1979", Journal = "IEEE Transactions on Software Engineering", Number = "5", Pages = "341-349", Title = "Dual processor scheduling with dynamic reassignment", Volume = "SE-5", Author = "Shahid H. Bokhari", Year = "November 1981", Journal = "IEEE Transactions on Software Engineering", Number = "6", Pages = "583-589", Title = "A shortest tree algorithm for optimal assignments across space and time in a distributed processor system", Volume = "SE-7", Author = "Shahid H. Bokhari", *** RECOMMENDED *** Title = "Partitioning problems in parallel, pipelined and distributed computing", Journal = "IEEE Transactions on Computers", Year = "January, 1988", Number ="1", Volume="C-37", Pages="48-57" Author = "Shahid H. Bokhari", Title = "Assignment problems in parallel and distributed computing", Publisher = "Kluwer", Address = "Boston", Year = "1987" Author = "Patricia J. Carstensen", Title = "The Complexity of Some Problems in Parametric Linear and Combinatorial Programming", Year = "1983", Institution = "Department of Mathematics, University of Michigan", Author = "K. W. Doty", Author = "P. L. McEntire", Author = "J. G. O'Reilly", Title = "Task allocation in a distributed computer system", Journal = "Proceedings of the IEEE Infocom 82", Pages = "33-38", Year = "1982", Author = "Dan Gusfield", Title = "Parametric combinatorial computing and a problem of program module distribution", Journal = "Journal of the ACM", Volume = "30", Number = "3", Pages = "551-563", Year = "July 1983", Author = "Robert E. Larson", Author = "Paul E. McIntyre", Author = "John G. O'Reilly", Title = "Tutorial: Distributed Control", Publisher = "IEEE Computer Society Press", Address = "Silver Spring, MD", Year = "1982", Author = "Virginia M. Lo", **** RECOMMENDED **** Title = "Heuristic algorithms for task assignments in distributed systems", Journal = "Proceedings of the 4th International Conference on Distributed Processing Systems", Pages = "30-39", Year = "May 1984", Author = "Janet Michel", Author = "Andries van Dam", Title = "Experience with distributed processing on a host/satellite system", Journal = "Computer Graphics (SIGGRAPH Newsletter)", Volume = "10", Number = "2", Year = "1976", Author = "Camille C. Price", Author = "Udo W. Pooch", Title = "Search Techniques for a nonlinear multiprocessor scheduling problem", Journal = "Naval Research Logistics Quarterly", Volume = "29", Number = "2", Pages = "213-233", Year = "June 1982", Author = "Gururaj S. Rao", Author = "Harold S. Stone", Author = "T. C. Hu", Title = "Assignment of tasks in a distributed processor system with limited memory", Journal = "IEEE TC", Volume = "C-28", Number = "4", Pages = "291-299", Year = "April 1979", Author = "Harold S. Stone", **** RECOMMENDED **** Title = "Multiprocessor scheduling with the aid of network flow algorithms", Journal = "IEEE Transactions on Software Engineering", Volume = "SE-3", Number = "1", Pages = "85-93", Year = "January 1977", Author = "Harold S. Stone", Year = "1977", Number = "ECE-CS-77-7", Institution = "Department of Electrical & Computer Engineering, University of Massachusetts, Amherst", Title = "Program assignment in three-processor systems and tricutset partitioning of graphs" Author = "Harold S. Stone", Title = "Critical load factors in two-processor distributed systems", Journal = "IEEE Transactions on Software Engineering", Volume = "SE-4", Number = "3", Pages = "254-258", Year = "May 1978", Author = "Donald F. Towsley", Title = "Allocating programs containing branches and loops within a multiple processor system", Journal = "IEEE Transactions on Software Engineering", Volume = "SE-12", Pages = "1018-1024", Year = "October 1986", Author = "Andries van Dam", Author = "George M. Stabler", Author = "Richard J. Harrington", Title = "Intelligent satellites for interactive graphics", Journal = "Proceedings of the IEEE", Volume = "62", Number = "4", Pages = "483-492", Year = "April 1974", >From Alessandro Forin at CMU: @article ( IEEECOMP, key = "Agora" , author = "Bisiani, R. and Forin, A." , title = "Multilanguage Parallel Programming on Heterogeneous Systems" , journal = "IEEE Transactions on Computers", publisher= "IEEE-CS" , month = "August" , year = "1988" , ) inproceedings ( BISI87G, key = "bisi87g" , author = "Bisiani,R. and Lecouat,F." , title = "A Planner for the Automatization of Programming Environment Tasks" , booktitle= "21st Hawaii International Conference on System Sciences" , publisher= "IEEE" , month = "January" , year = "1988" , bibdate = "Fri Aug 28 09:44:54 1987" , ) @inproceedings ( DBGWKSHP, key = "Agora" , author = "Forin, Alessandro" , title = "Debugging of Heterogeneous Parallel Systems" , booktitle= "Intl. Workshop on Parallel and Distributed Debugging", publisher= "SIGPLAN Notices, V24-1 Jan. 1989", address = "Madison, WI", month = "May" , year = "1988" , pages = "130-141", ) @techreport ( ASMREPORT, key = "Agora" , author = "R. Bisiani, F. Alleva, F. Correrini, A. Forin, F. Lecouat, R. L erner", title = "Heterogeneous Parallel Processing, The Agora Shared Memory" , institution= "Carnegie-Mellon University" , address = "Comp. Science Dept." , type = "Tech. Report" , number = "CMU-CS-87-112" , month = "March" , year = "1987" , ) Dr Michael Coffin at Unoiversity of Waterloo suggests: ------------------------------------------------------ AUTHOR = "Michael H. Coffin", TITLE = "Par: {A}n Approach to Architecture-Independent Parallel Programming", SCHOOL = "Department of Computer Science, The University of Arizona", MONTH = aug, YEAR = "1990", ADDRESS = "Tucson, Arizona" } Dr. David Skillicorn at Queens University suggests: --------------------------------------------------- TITLE = {The Purdue Dual {MACE} Operating System}, INSTITUTION = {Purdue University}, KEYWORDS = {Abell1}, YEAR = {1978}, MONTH = {NOV}, } @ARTICLE{bib:002, AUTHOR = {Guy T. Almes and Andrew P. Black and Edward D. Lazowska and Jerre D. Noe}, TITLE = {The Eden System: A Technical Review}, JOURNAL = {IEEE Transactions on Software Engineering}, PAGES = {43--59}, KEYWORDS = {Almes1}, YEAR = {1985}, MONTH = {JAN}, } @INPROCEEDINGS{bib:003, AUTHOR = {D.E. Bailey and J.E. Cuny}, TITLE = {An Approach to Programming Process Interconnection Structures: Aggregate Rewriting Graph Grammars}, BOOKTITLE = {Proceedings of PARLE '87 Parallel Architectures and Languages Europe, Volume II}, PAGES = {112--123}, ORGANIZATION = {Springer-Verlag, Lecture Notes in Computer Science}, ADDRESS = {Eindhoven, The Netherlands}, YEAR = {1987}, MONTH = {June}, } @ARTICLE{bib:004, AUTHOR = {A. Barak and A. Litman}, TITLE = {{MOS}: a Multicomputer Distributed Operating System}, JOURNAL = {Software: Practice and Experience}, KEYWORDS = {Barak1}, LENGTH = {725}, YEAR = {1985}, MONTH = {AUG}, } @ARTICLE{bib:005, AUTHOR = {A. Barak and A. Shiloh}, TITLE = {A Distributed Load Balancing Policy for a Multicomputer}, JOURNAL = {Software: Practice and Experience}, KEYWORDS = {Barak2}, LENGTH = {901}, YEAR = {1985}, MONTH = {SEP}, } @ARTICLE{bib:006, AUTHOR = {? Bartlett and et al}, TITLE = {A NonStop Kernel}, JOURNAL = {PROC of the 8th SOSP}, KEYWORDS = {Bartle1}, YEAR = {1981}, MONTH = {OCT}, } @ARTICLE{bib:007, AUTHOR = {M.J. Berger and S.H. Bokhari}, TITLE = {A Partitioning Strategy for Nonuniform Problems on Multiprocessors}, JOURNAL = {IEEE Transactions on Computers}, VOLUME = {C-36, No.5}, PAGES = {570--580}, KEYWORDS = {rectangular partition with uniform workload}, YEAR = {1987}, MONTH = {May}, } @INPROCEEDINGS{bib:008, AUTHOR = {Andrew P. Black}, TITLE = {Supporting Distributed Applications: Experience with Eden}, JOURNAL = {PROC of the 10th SOSP}, KEYWORDS = {Black1}, YEAR = {1985}, MONTH = {DEC}, } @ARTICLE{bib:011, ***** RECOMMENDED ***** AUTHOR = {Shahid H. Bokhari}, TITLE = {On the Mapping Problem}, JOURNAL = {IEEE Transactions on Computers}, VOLUME = {C-30}, NUMBER = {3}, PAGES = {207--214}, KEYWORDS = {grecommended,}, YEAR = {1981}, MONTH = {March}, ABSTRACT = {This paper is important because it points out that the mapping problem is akin to graph traversal and is at least P-complete. Also see ICPP79. Reproduced in the 1984 tutorial: Interconnection Networks for parallel and distributed processing by Wu and Feng.}, } @ARTICLE{bib:015, AUTHOR = {W.W. Chu and L.J. Holloway and M.T. Lan and K. Efe}, TITLE = {Task Allocation in Distributed Data Processing}, JOURNAL = {Computer}, PAGES = {57--69}, YEAR = {1980}, MONTH = {November}, } @INPROCEEDINGS{bib:018, AUTHOR = {J.G. Donnett and M. Starkey and D.B. Skillicorn}, TITLE = {Effective Algorithms for Partitioning Distributed Programs}, BOOKTITLE = {Proceedings of the Seventh Annual International Phoenix Conference on Computers and Communications}, PAGES = {363--369}, YEAR = {1988}, MONTH = {March 16--18}, } @MISC{bib:025, **** RECOMMENDED **** AUTHOR = {D.A. Hornig}, TITLE = {Automatic Partitioning and Scheduling on a Network of Personal Computers}, INSTITUTION = {Carnegie Mellon University, Department of Computer Science,}, YEAR = {1984}, MONTH = {November}, ABSTRACT = {This Ph.D thesis describes the development of a language Stardust in which indications are given of the running time of each function. The run-time evironment then schedules the functions based on the costs of message passing and load balancing. There is some discussion of granularity. The language contains no explicit partitioning.}, } @ARTICLE{bib:027, AUTHOR = {P. Hudak and B. Goldberg}, TITLE = {Distributed Execution of Functional Programs Using Serial Combinators}, JOURNAL = {IEEE Transactions on Computers}, VOLUME = {C34, No.10}, PAGES = {881--891}, YEAR = {1985}, MONTH = {October}, } @ARTICLE{bib:031, **** RECOMMENDED **** AUTHOR = {F.C.H. Lin and R.M. Keller}, TITLE = {The Gradient Model Load Balancing Method}, JOURNAL = {IEEE Transactions on Software Engineering}, VOLUME = {SE-13, No.1}, PAGES = {32--38}, YEAR = {1987}, MONTH = {January}, } @INPROCEEDINGS{bib:037, AUTHOR = {L.J. Miller}, TITLE = {A Heterogeneous Multiprocessor Design and the Distributed Scheduling of its Task Group Workload}, BOOKTITLE = {Proceedings of 9th Annual Symposium on Computer Architecture}, PAGES = {283--290}, YEAR = {1982}, MONTH = {April}, } @ARTICLE{bib:042, AUTHOR = {D.A. Padua and M.J. Wolfe}, TITLE = {Advanced Compiler Optimizations for Supercomputers}, JOURNAL = {Communications of the ACM}, VOLUME = {29, No.12}, PAGES = {1184--1201}, YEAR = {1986}, MONTH = {December}, } @ARTICLE{bib:043, AUTHOR = {Michael L. Powell and Barton P. Miller}, TITLE = {Process Migration in DEMOS/MP}, JOURNAL = {PROC of the 9th SOSP}, KEYWORDS = {Powell1}, LENGTH = {110}, YEAR = {1983}, MONTH = {DEC}, } @ARTICLE{bib:044, AUTHOR = {G.S. Rao and H.S. Stone and T.C. Hu}, TITLE = {Assignment of Tasks in a Distributed Processor System with Limited Memory}, JOURNAL = {IEEE Transactions on Computers}, VOLUME = {C-28, No.4}, PAGES = {291--299}, YEAR = {1979}, MONTH = {April}, } @ARTICLE{bib:046, **** RECOMMENDED **** AUTHOR = {C.-C Shen and W.-H. Tsai}, TITLE = {A Graph Matching Approach to Optimal Task Assignment in Distributed Computing Systems Using a Minimax Criterion}, JOURNAL = {IEEE Transactions on Computers}, VOLUME = {C-34, No.3}, PAGES = {197--203}, YEAR = {1985}, MONTH = {March}, } @ARTICLE{bib:054, AUTHOR = {H. Widjaja}, TITLE = {An Effective Structured Approach to Finding Optimal Partitions}, JOURNAL = {Computing}, VOLUME = {29, No.3}, PAGES = {241--262}, YEAR = {1982}, } @INPROCEEDINGS{bib:055, AUTHOR = {E. Williams}, TITLE = {Assigning Processes to Processors in Distributed Systems}, BOOKTITLE = {Proceedings of 1983 International Conference on Parallel}, PAGES = {404--406}, YEAR = {1983}, MONTH = {August}, } @INPROCEEDINGS{bib:056, AUTHOR = {F. Ercal and J. Ramanujam and P. Sadayappan}, TITLE = {Task Allocation onto a Hypercube by Recursive Mincut}, BOOKTITLE = {Hypercube Conference}, YEAR = {1988}, } @article{, author = {J.-L. Gaudiot and J.I. Pi and M.L. Campbell}, title = {Program Graph Allocation in Distributed Multicomputers}, journal = {Parallel Computing}, volume = {7}, year = {1988}, pages = {227 -- 247}, } David Hudak at the University of Michigan writes: ------------------------------------------------- "Performance Evaluation and Prediction for Parallel Algorithms on the BBN GP1000", F. Bodin, D. Windheiser, W. Jalby, etc., ACM International Conference on Supercomputing, 1990, pp. 401 - 413. "The Impact of Synchronization and Granularity on Parallel Systems", Ding-Kai Chen, Hong-Men Su, and Pen-Chung Yew, International Symposium on Computer Architecture, 1990, p. 239 - 248 Also: for interesting work on dynamic partitioning, check Polychronopou los' article, (IEEE Computer, '86 I think) on Guided Self-Scheduling Really, the guys you want to read about are: Jalby, Polychronopoulos, Dennis Gannon, Sameh, Windheiser, and, of course, me. (Oh, Reed had an IEEE paper '87 on stencils and program partitioning, and Vrsalovic had a good tech report from CMU.) Bill Schilit at Columbia suggests: ---------------------------------- Parallel Processing: the Cm* experience, Edward F. Gehringer, et. al. Digital Press Dr. David Finkel at Worcester Polythecnic Institute writes: ----------------------------------------------------------- "Evaluating Dynamic Load Sharing in Distributed Computer Systems", Computer Systems: Science and Engineering 5 (1990), 89 - 94. "Load Indices for Load Sharing in Heterogeneous Distributed Computing Systems", with David Hatch,Proceedings of the 1990 UKSC Conference on Computer Simulation, Brighton, 1990, 202 - 206. Zbigniew Chamski (Zbigniew.Chamski@irisa.fr) suggests: ------------------------------------------------------ @string{IEEES = "IEEE Software"} **** RECOMMENDED **** @string{ECEDOSU = "Electrical and Computer Engineering Department, Oregon State University"} @article{ KrLe88, author = "Kruatrachue, B. and Lewis, T.", title = "Grain Size Determination for Parallel Processing", journal = IEEES, year = 1988, volume = 5, number = 1, pages = "23--32", month = jan} @phdthesis{ **** RECOMMENDED **** Krua87, author = "Kruatrachue, B.", title = "Static Task Scheduling and Grain Packing in Parallel Processing Systems", school = ECEDOSU, year = 1987, address = "{Corvallis, OR, USA}"} @PhdThesis{ElRe89, author = "El-Rewini, H.", title = "Architecture-Independent Task Partitioning and Scheduling on Arbitrary Parallel Processing Systems", school = "Department of Computer Science, Oregon State Universi ty", year = "1989", address = "{Corvallis, OR, USA}", month = nov} I would also add the following recommendations: McCreary, C., and Gill, H., "Automatic Determination of Grain Size for Efficient Parallel Processing", CACM, September 1989, pp. 1073-1078. Van Tilborg, A., Wittie, L., "Wave Scheduling -- Decentralized Scheduling of Task Forces in Multicomputers", IEEE Transactions on Computers, 33:835-844, September 1984. Berman, F., "Why is Mapping Hard for Parallel Computers?", Proceedings of the IEEE Parallel/Distributed Computing Networks Seminar, Jan. 31, 1990. Sarkar, V., "Partitioning and Scheduling for Execution on Multiprocessors", Ph.D. dissertation, Stanford Tech. Report No. CSL-TR-87-328, April 1987. -- -------------------------------------------------------------------- Christian.Collberg@dna.lu.se -- =========================== MODERATOR ============================== Steve Stevenson {steve,fpst}@hubcap.clemson.edu Department of Computer Science, comp.parallel Clemson University, Clemson, SC 29634-1906 (803)656-5880.mabell