Abstract
Software performance can be assessed in terms of stability, usability, and reliability which in turn can be obtained by testing and put emphasis on testing strategies. Testing and controlled monitoring increase the life period of software by allowing it to be transformed to meet the customers’ desires. Software with modest design is more reliable and stable than the software with composite design. Simple design is easier to understand, which minimizes time that will be spent understanding the system, and it is easier to revise and helps to realize the software performance. This work emphasizes on performance measuring criteria and its release decisions based on quantitative approach. From business perspective a strong relationship between business requirements and software performance exists, so it is necessary to measure the performance of software before it gets released. As business requirements changes, software needs to be able to adapt to these changes without decreasing reliability. Software which is adaptable and is more stable will deliver maximum performance in its operational environment and will be easier for the development team to take proper release decisions. The purpose of this work is to envisage and propose a new paradigm of judging software reliability in its operational environment and address issues related to releasing of software. The results show that our proposed model after incorporating the concept of window gives optimal output in terms of operational reliability by eliminating maximum number of faults.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
N.F. Schneidewind, Measuring testing efficiency & effectiveness for software up gradation and its impact on CBP. R & M Eng. J 23(2), 1–6 (2003)
R Majumdar, P.K. Kapur, S.K. Khatri, Measuring testing efficiency & effectiveness for software up gradation and its impact on CBP. in 2016 International Conference on Innovation and Challenges in Cyber Security (ICICCS-INBUSH), pp. 123–128, 2016
R. Majumdar, P.K. Kapur, S.K. Khatri, Measuring testing efficiency: An alternative approach, in 2015 4th International Conference on Reliability, Infocom Technologies and Optimization (ICRITO) (Trends and Future Directions), pp. 1–5, 2015
P.K. Kapur, V.S.S. Yadavali, A.K. Shrivastava, A comparative study of vulnerability discovery modeling and software reliability growth modeling, in 2015 International Conference on Futuristic Trends on Computational Analysis and Knowledge Management (ABLAZE), pp. 246–251, 2015
P.K. Kapur, H. Pham, A.G. Aggarwal, G. Kaur, Two dimensional multi-release software reliability modeling and optimal release planning. IEEE Trans. Reliab., 57–63 (2012)
S.K. Khatri, A. Sonik, R. Majumdar, Modelling the successive software release time of S-shaped model with imperfect debugging, in International Conference on Information Technology, Amity University, 6–7 October 2016
J.D. Musa, Software Reliability Engineering: More Reliable Software, Faster and Cheaper (TataMcGraw-HillEducation, Springer, 2004.) http:johnmusa.com/book.htm
P.K. Kapur, R.B. Garg, S. Kumar, Contributions to Hardware and Software Reliability (World Scientific, Singapore, 1999)
M.R. Lyu, Handbook of Software Reliability Engineering (McGraw Hill, Springer, 1996)
P.K. Kapur, S. Agarwal, R.B. Garg, Bi-criterion release policy for exponential software reliability growth models. Oper. Res. (Recherche Operationanelle) 28, 165–180 (1994)
C. Stringfellow et al., An empirical method for selecting software reliability growth models. Empir. Softw. Eng., 319–343 (2002)
https://bitdepth.thomasrutter.com/2010/04/02/stable-vs-stable-what-stable-means-in-software/
S.Z. Ke, C.Y. Huang, Software reliability analysis considering the variation of testing-effort and change-point, in Proceedings of the International Workshop on Innovative Software Development Methodologies and Practices, pp 30–39, ISBN: 978-1-4503-3226-2, https://doi.org/10.1145/2666581.2666588
S.K. Khatri, S. John, R. Majumdar, Quantifying software reliability using testing effort, in International Conference on Information Technology, Amity University, 6–7 October 2016
P.K. Kapur, H. Pham, A. Gupta, P.C. Jha, Software Reliability Assessment with OR Application (Springer, Berlin, 2011)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
Majumdar, R., Gupta, R., Singh, A. (2018). Software Performance Measuring Benchmarks. In: Woungang, I., Dhurandher, S. (eds) International Conference on Wireless, Intelligent, and Distributed Environment for Communication. WIDECOM 2018. Lecture Notes on Data Engineering and Communications Technologies, vol 18. Springer, Cham. https://doi.org/10.1007/978-3-319-75626-4_8
Download citation
DOI: https://doi.org/10.1007/978-3-319-75626-4_8
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-75625-7
Online ISBN: 978-3-319-75626-4
eBook Packages: EngineeringEngineering (R0)