Advertisement

The Open Run-Time Environment (OpenRTE): A Transparent Multi-cluster Environment for High-Performance Computing

  • R. H. Castain
  • T. S. Woodall
  • D. J. Daniel
  • J. M. Squyres
  • B. Barrett
  • G. E. Fagg
Part of the Lecture Notes in Computer Science book series (LNCS, volume 3666)

Abstract

The Open Run-Time Environment (OpenRTE)—a spin-off from the Open MPI project—was developed to support distributed high-performance computing applications operating in a heterogeneous environment. The system transparently provides support for interprocess communication, resource discovery and allocation, and process launch across a variety of platforms. In addition, users can launch their applications remotely from their desktop, disconnect from them, and reconnect at a later time to monitor progress. This paper will describe the capabilities of the OpenRTE system, describe its architecture, and discuss future directions for the project.

Keywords

Message Passing Interface Resource Discovery Error Management Interprocess Communication Remote Process 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Geist, A., et al.: MPI-2: Extending the Message-Passing Interface. In: Euro-Par 1996 Parallel Processing, pp. 128–135. Springer, Heidelberg (1996)Google Scholar
  2. 2.
    Bernholdt, D.E., et al.: A component architecture for high-performance scientific computing. (to appear) in Intl. J. High-Performance Computing ApplicationsGoogle Scholar
  3. 3.
    Gabriel, E., et al.: Open MPI: Goals, concept, and design of a next generation mpi implementation. In: 11th European PVM/MPI Users’ Group Meeting (2004)Google Scholar
  4. 4.
    Aulwes, R.T., et al.: Architecture of LA-MPI, a network-fault-tolerant mpi. In: 18th Intl Parallel and Distributed Processing Symposiun (2004)Google Scholar
  5. 5.
    Fagg, G., Dongarra, J.: HARNESS Fault Tolerant MPI Design, Usage and Performance Issues. Future Generation Computer Systems 18(8), 1127–1142 (2002)zbMATHCrossRefGoogle Scholar
  6. 6.
    Foster, I., Kesselman, C.: Globus: A metacomputing infrastructure toolkit. Intl J. Supercomputer Applications 11(2), 115–128 (1997)CrossRefGoogle Scholar
  7. 7.
    Kronstadt, E.P.: Petascale computing. In: 19th IEEE Intl Parallel and Distributed Processing Symposium, Denver, CO, USA (April 2005)Google Scholar
  8. 8.
    Message Passing Interface Forum. MPI: A Message Passing Interface. In: Proc. of Supercomputing 1993, November 1993, pp. 878–883. IEEE Computer Society Press, Los Alamitos (1993)Google Scholar
  9. 9.
    Squyres, J.M., Lumsdaine, A.: A Component Architecture for LAM/MPI. In: 10th European PVM/MPI Users’ Group Meeting (2003)Google Scholar
  10. 10.
    Sunderam, V., Kurzyniec, D.: Lightweight self-organizing frameworks for metacomputing. In: 11th International Symposium on High Performance Distributed Computing, Edinburgh, UK (July 2002)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2005

Authors and Affiliations

  • R. H. Castain
    • 1
  • T. S. Woodall
    • 1
  • D. J. Daniel
    • 1
  • J. M. Squyres
    • 2
  • B. Barrett
    • 2
  • G. E. Fagg
    • 3
  1. 1.Los Alamos National Lab 
  2. 2.Indiana University 
  3. 3.University of TennesseeKnoxville

Personalised recommendations