Component Specification for Parallel Coupling Infrastructure

  • J. Walter Larson
  • Boyana Norris
Part of the Lecture Notes in Computer Science book series (LNCS, volume 4707)


Coupled systems comprise multiple mutually interacting subsystems, and are an increasingly common computational science application, most notably as multiscale and multiphysics models. Parallel computing, and in particular message-passing programming have spurred the development of these models, but also present a parallel coupling problem (PCP) in the form of intermodel data dependencies. The PCP complicates model coupling through requirements for the description, transfer, and transformation of the distributed data that models in a parallel coupled system exchange. Component-based software engineering has been proposed as one means of conquering software complexity in scientific applications, and given the compound nature of coupled models, it is a natural approach to addressing the parallel coupling problem. We define a software component specification for solving the parallel coupling problem. This design draws from the already successful Common Component Architecture (CCA). We abstract the parallel coupling problem’s elements and map them onto a set of CCA components, defining a parallel coupling infrastructure toolkit. We discuss a reference implementation based on the Model Coupling Toolkit. We demonstrate how these components might be deployed to solve a relevant coupling problems in climate modeling.


Parallel Composition Community Climate System Model Reference Implementation Parallel Climate Model Communication Schedule 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Manabe, S., Bryan, K.: Climate calculations with a combined ocean-atmosphere model. Journal of the Atmospheric Sciences 26(4), 786–789 (1969)CrossRefGoogle Scholar
  2. 2.
    Larson, J., Jacob, R., Ong, E.: The Model Coupling Toolkit: A new Fortran90 toolkit for building multi-physics parallel coupled models. Int. J. High Perf. Comp. App. 19(3), 277–292 (2005)CrossRefGoogle Scholar
  3. 3.
    Larson, J.W.: Some organising principles for coupling in multiphysics and multiscale models. Preprint ANL/MCS-P1414-0207, Mathematics and Computer Science Division, Argonne National Laboratory (2006)Google Scholar
  4. 4.
    Bryan, F.O., Kauffman, B.G., Large, W.G., Gent, P.R.: The NCAR CSM flux coupler. NCAR Tech. Note 424, NCAR, Boulder, CO (1996)Google Scholar
  5. 5.
    Jacob, R., Schafer, C., Foster, I., Tobis, M., Anderson, J.: Computational design and performance of the Fast Ocean Atmosphere Model. In: Alexandrov, V.N., Dongarra, J.J., Juliano, B.A., Renner, R.S., Tan, C.J.K. (eds.) ICCS 2001. LNCS, vol. 2073, pp. 175–184. Springer, Heidelberg (2001)Google Scholar
  6. 6.
    Bettge, T., Craig, A., James, R., Wayland, V., Strand, G.: The DOE Parallel Climate Model (PCM): The Computational Highway and Backroads. In: Alexandrov, V.N., Dongarra, J.J., Juliano, B.A., Renner, R.S., Tan, C.J.K. (eds.) ICCS 2001. LNCS, vol. 2073, pp. 148–156. Springer, Heidelberg (2001)Google Scholar
  7. 7.
    Drummond, L.A., Demmel, J., Mechose, C.R., Robinson, H., Sklower, K., Spahr, J.A.: A data broker for distirbuted computing environments. In: Alexandrov, V.N., Dongarra, J.J., Juliano, B.A., Renner, R.S., Tan, C.J.K. (eds.) ICCS 2001. LNCS, vol. 2073, pp. 31–40. Springer, Heidelberg (2001)Google Scholar
  8. 8.
    Valcke, S., Redler, R., Vogelsang, R., Declat, D., Ritzdorf, H., Schoenemeyer, T.: OASIS4 user’s guide. PRISM Report Series 3, CERFACS, Toulouse, France (2004)Google Scholar
  9. 9.
    Hill, C., DeLuca, C., Balaji, V., Suarez, M., da Silva, A.: The ESMF Joint Specification Team: The architecture of the earth system modeling framework. Computing in Science and Engineering 6, 18–28 (2004)Google Scholar
  10. 10.
    Toth, G., Sokolov, I.V., Gombosi, T.I., Chesney, D.R., Clauer, C.R., Zeeuw, D.D., Hansen, K.C., Kane, K.J., Manchester, W.B., Oehmke, R.C., Powell, K.G., Ridley, A.J., Roussev, I.I., Stout, Q.F., Volberg, O., Wolf, R.A., Sazykin, S., Chan, A., Yu, B., Kota, J.: Space weather modeling framework: A new tool for the space science community. Journal of Geophysical Research 110, A12226 (2005)CrossRefGoogle Scholar
  11. 11.
    Bertrand, F., Bramley, R., Bernholdt, D.E., Kohl, J.A., Sussman, A., Larson, J.W., Damevski, K.B.: Data redistribution and remote method invocation for coupled components. Journal of Parallel and Distributed Computing 66(7), 931–946 (2006)zbMATHCrossRefGoogle Scholar
  12. 12.
    Joppich, W., Kurschner, M.: MpCCI - a tool for the simulation of coupled applications. Concurrency and Computation: Practice and Experience 18(2), 183–192 (2006)CrossRefGoogle Scholar
  13. 13.
    Jacob, R., Larson, J., Ong, E.: M x N communication and parallel interpolation in ccsm3 using the Model Coupling Tookit. Int. J. High Perf. Comp. App. 19(3), 293–308 (2005)CrossRefGoogle Scholar
  14. 14.
    The MCT Development Team: Model Coupling Toolkit (MCT) web site (2007),
  15. 15.
    Craig, A.P., Kaufmann, B., Jacob, R., Bettge, T., Larson, J., Ong, E., Ding, C., He, H.: cpl6: The new extensible high-performance parallel coupler for the community climate system model. Int. J. High Perf. Comp. App. 19(3), 309–327 (2005)CrossRefGoogle Scholar
  16. 16.
    Foster, I.: Designing and Building Parallel Programs: Concepts and Tools for Parallel Software Engineering. Addison Wesley, Reading, Massachusetts (1995)zbMATHGoogle Scholar
  17. 17.
    Szyperski, C.: Component Software: Beyond Object-Oriented Programming. ACM Press, New York (1999)Google Scholar
  18. 18.
    Heineman, G.T., Council, W.T.: Component Based Software Engineering: Putting the Pieces Together. Addison-Wesley, New York (1999)Google Scholar
  19. 19.
    Bernholdt, D.E., Allan, B.A., Armstrong, R., Bertrand, F., Chiu, K., Dahlgren, T.L., Damevski, K., Elwasif, W.R., Epperly, T.G.W., Govindaraju, M., Katz, D.S., Kohl, J.A., Krishnan, M., Kumfert, G., Larson, J.W., Lefantzi, S., Lewis, M.J., Malony, A.D., Mclnnes, L.C., Nieplocha, J., Norris, B., Parker, S.G., Ray, J., Shende, S., Windus, T.L., Zhou, S.: A component architecture for high-performance scientific computing. Int. J. High Perf. Comp. App. 20(2), 163–202 (2006)CrossRefGoogle Scholar
  20. 20.
    CCA Forum: CCA Forum web site (2007),
  21. 21.
    Dahlgren, T., Epperly, T., Kumfert, G.: Babel User’s Guide. CASC, Lawrence Livermore National Laboratory. version 0.9.0 edn. (January 2004)Google Scholar
  22. 22.
    Ong, E.T., Larson, J.W., Norris, B., Jacob, R.L., Tobis, M., Steder, M.: Multilingual interfaces for parallel coupling in multiphysics and multiscale systems. In: Shi, Y. (ed.) ICCS 2007. LNCS, vol. 4487, pp. 924–931. Springer, Heidelberg (2007)CrossRefGoogle Scholar
  23. 23.
    Lee, J.-Y., Sussman, A.: High performance communication between parallel programs (Appears with the Proceedings of IPDPS 2005). In: Proceedings of 2005 Joint Workshop on High-Performance Grid Computing and High-Level Parallel Programming Models (HIPS-HPGC 2005), Apr 2005, IEEE Computer Society Press, Los Alamitos (2005)Google Scholar
  24. 24.
    Sussman, A.: Building complex coupled physical simulations on the grid with InterComm. Engineering with Computers 22(3–4), 311–323 (2006)CrossRefGoogle Scholar
  25. 25.
    Jones, P.W.: A user’s guide for SCRIP: A spherical coordinate remapping and interpolation package. Technical report, Los Alamos National Laboratory, Los Alamos, NM (1998)Google Scholar
  26. 26.
    Jones, P.W.: First and second-order conservative remapping schemes for grids in spherical coordinates. Mon. Wea. Rev. 127, 2204–2210 (1999)CrossRefGoogle Scholar
  27. 27.
    Interoperable Technologies for Advanced Petascale Simulation Team: ITAPS web site (2007),

Copyright information

© Springer-Verlag Berlin Heidelberg 2007

Authors and Affiliations

  • J. Walter Larson
    • 1
    • 2
  • Boyana Norris
    • 1
  1. 1.Mathematics and Computer Science Division, Argonne National Laboratory, Argonne, IL 60439USA
  2. 2.ANU Supercomputer Facility, The Australian National University, Canberra ACT 0200Australia

Personalised recommendations