Parallel Computing Strategy Design Based on COC

  • Jing-Jing Zhou
Part of the Communications in Computer and Information Science book series (CCIS, volume 236)


Sparse and unstructured computations are widely involved in engineering and scientific applications. It means that data arrays could be indexed indirectly through the values of other arrays or non-affine subscripts. Data access pattern would not be known until runtime. So far all the parallel computing strategies for this kind of irregular problem are single network topology oriented, which cannot fully exploit the advantages of modern hierarchical computing architecture, like grid. We proposed a hybrid parallel computing strategy RP, shorted for “Replicated and Partially-shared”, to improve the performance of irregular applications in the COC (Cluster of Clusters) environment.


Irregular Applications Replicated and Partially-shared Cluster of Clusters 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    High Performance Fortran Forum, High Performance Fortran language specification, version 1.0, Technical Report CRPC-TR92225, Houston, Tex (1993)Google Scholar
  2. 2.
    Frumkin, M., Jin, H., Yan, J.: Implementation of NAS Parallel Benchmarks in High Performance Fortran, Technical Report NAS-98-009Google Scholar
  3. 3.
    Das, R., Uysal, M., Saltz, J., Hwang, Y.-S.S.: Communication Optimizations for Irregular Scientific Computations on Distributed Memory ArchitecturesGoogle Scholar
  4. 4.
    Koelbel, C.H., Loveman, D.B., Schreiber, R.S., Steel Jr., G.L., Zosel, M.E.: The High Performance Fortran Handbook. MIT Press, Cambridge (1994)Google Scholar
  5. 5.
    Brezany, P., Bubak, M., Malawski, M., Zajaac, K.: Large-Scale Scientific Irregular Computing on Clusters and Grids. In: Proceedings of the International Conference on Computational Science (2002)Google Scholar
  6. 6.
    Minyi, G.: Automatic Parallelization and Optimization for Irregular Scientific Applications. In: Proceedings of the 18th International Parallel and Distributed Processing Symposium (2004)Google Scholar
  7. 7.
    Han, H., Tseng, C.-W.: Exploiting Locality for Irregular Scientific Codes. IEEE Transactions on Parallel and Distributed Systems (2006)Google Scholar
  8. 8.
    Hu, Y.C., Lu, H., Cox, A.L., Zwaenepoel, W.: OpenMP for Networks of SMPs. Journal of Parallel and Distributed Computing 60(12), 1512–1530 (2000)CrossRefzbMATHGoogle Scholar
  9. 9.
    El-Ghazawi, T., Carlson, W., Draper, J.: UPC Language Specifications V1.0 (February 2001)Google Scholar
  10. 10.
    Min, S.-J., Basumallik, A.y., Eigenmann, R.: Optimizing OpenMP programs on Software Distributed Shared Memory Systems. International Journal of Parallel Programming 31(3), 225–249 (2003)CrossRefzbMATHGoogle Scholar
  11. 11.
    Basumallik, A., Eigenmann, R.: Optimizing Irregular Shared-Memory Applications for Distributed-Memory Systems. In: Proceedings of the 11th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming (March 2006)Google Scholar
  12. 12.
    Feng, X.-B., Chen, L., Wang, Y.-R., An, X.-M., Ma, L., Sang, C.-L., Zhang, Z.-Q.: Integrating Parallelizing Compilation Technologies for SMP Clusters. Journal of Computer Science and Technology 20(1), 125–133 (2005)CrossRefGoogle Scholar
  13. 13.
    Catalyurek, U.V., Aykanat, C.: A Fine-Grain Hypergraph Model for 2D Decomposition of Sparse Matrices, ipdps. In: 15th International Parallel and Distributed Processing Symposium (IPDPS 2001) Workshops, vol. 3, p. 30118b (2001)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2011

Authors and Affiliations

  • Jing-Jing Zhou
    • 1
  1. 1.College of Information & Electronic EngineeringZhejiang Gongshang UniversityChina

Personalised recommendations