Abstract
To exploit the potential of many core processors, a serial code is generally optimized for a particular compiler called a target compiler, so that the compiler can understand the code structure for automatic parallelization. However, the performance of such a serial code is always not portable to a new system that uses a different compiler. To improve the performance portability, this paper proposes an OpenMP parallelization method by using compiler messages of the target compiler. Since the compiler messages from the target compiler are also useful to identify key loop nests even for the different system, a programmer can use the message to easily parallelize a serial code with low programming effort. Furthermore, programmer’s intention of the optimization can be migrated to other systems through the OpenMP parallelization, which results in high performance portability. The experimental results indicate that the OpenMP codes parallelized by the proposed method can achieve a comparable or even better performance than the automatically parallelized codes by various compilers.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Top 500 supercomputers sites. http://www.top500.org/
Mustafa, D., Aurangzeb, A., Eigenmann, R.: Performance analysis and tuning of automatically parallelized openmp applications. In: Proceedings of the 7th International Conference on OpenMP in the Petascale era, IWOMP’11, pp. 151–164. Springer, Berlin, Heidelberg. http://dl.acm.org/citation.cfm?id=2023025.2023041 (2011)
Himeno benchmark. http://accc.riken.jp/2444.htm
Soga, T., Musa, A., Shimomura, Y., Egawa, R., Itakura, K., Takizawa, H., Okabe, K., Kobayashi, H.: Performance evaluation of nec sx-9 using real science and engineering applications. In: Proceedings of the Conference on High Performance Computing Networking, Storage and Analysis, pp. 1–12 (2009). doi:10.1145/1654059.1654088
Nakahashi, K.: High-density mesh flow computations with pre-/post-data compressions. AIAA Paper 2005-4876 (2005)
Acknowledgements
This research was partially supported by Grant-in- Aid for Scientific Research (S) #21226018 and Core Research of Evolutional Science and Technology of Japan Science and Technology Agency (JST CREST) “An Evolutionary Approach to Construction of a Software Development Environment for Massively-Parallel Heterogeneous Systems”.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Komatsu, K., Egawa, R., Takizawa, H., Kobayashi, H. (2015). Performance Evaluation of an OpenMP Parallelization by Using Automatic Parallelization Information. In: Resch, M., Bez, W., Focht, E., Kobayashi, H., Patel, N. (eds) Sustained Simulation Performance 2014. Springer, Cham. https://doi.org/10.1007/978-3-319-10626-7_10
Download citation
DOI: https://doi.org/10.1007/978-3-319-10626-7_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-10625-0
Online ISBN: 978-3-319-10626-7
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)