Software Performance Measuring Benchmarks

  • Rana MajumdarEmail author
  • Ritu Gupta
  • Abhilasha Singh
Conference paper
Part of the Lecture Notes on Data Engineering and Communications Technologies book series (LNDECT, volume 18)


Software performance can be assessed in terms of stability, usability, and reliability which in turn can be obtained by testing and put emphasis on testing strategies. Testing and controlled monitoring increase the life period of software by allowing it to be transformed to meet the customers’ desires. Software with modest design is more reliable and stable than the software with composite design. Simple design is easier to understand, which minimizes time that will be spent understanding the system, and it is easier to revise and helps to realize the software performance. This work emphasizes on performance measuring criteria and its release decisions based on quantitative approach. From business perspective a strong relationship between business requirements and software performance exists, so it is necessary to measure the performance of software before it gets released. As business requirements changes, software needs to be able to adapt to these changes without decreasing reliability. Software which is adaptable and is more stable will deliver maximum performance in its operational environment and will be easier for the development team to take proper release decisions. The purpose of this work is to envisage and propose a new paradigm of judging software reliability in its operational environment and address issues related to releasing of software. The results show that our proposed model after incorporating the concept of window gives optimal output in terms of operational reliability by eliminating maximum number of faults.


Software reliability Release decisions Performance measurement window 


  1. 1.
    N.F. Schneidewind, Measuring testing efficiency & effectiveness for software up gradation and its impact on CBP. R & M Eng. J 23(2), 1–6 (2003)Google Scholar
  2. 2.
    R Majumdar, P.K. Kapur, S.K. Khatri, Measuring testing efficiency & effectiveness for software up gradation and its impact on CBP. in 2016 International Conference on Innovation and Challenges in Cyber Security (ICICCS-INBUSH), pp. 123–128, 2016Google Scholar
  3. 3.
    R. Majumdar, P.K. Kapur, S.K. Khatri, Measuring testing efficiency: An alternative approach, in 2015 4th International Conference on Reliability, Infocom Technologies and Optimization (ICRITO) (Trends and Future Directions), pp. 1–5, 2015Google Scholar
  4. 4.
    P.K. Kapur, V.S.S. Yadavali, A.K. Shrivastava, A comparative study of vulnerability discovery modeling and software reliability growth modeling, in 2015 International Conference on Futuristic Trends on Computational Analysis and Knowledge Management (ABLAZE), pp. 246–251, 2015Google Scholar
  5. 5.
    P.K. Kapur, H. Pham, A.G. Aggarwal, G. Kaur, Two dimensional multi-release software reliability modeling and optimal release planning. IEEE Trans. Reliab., 57–63 (2012)Google Scholar
  6. 6.
    S.K. Khatri, A. Sonik, R. Majumdar, Modelling the successive software release time of S-shaped model with imperfect debugging, in International Conference on Information Technology, Amity University, 6–7 October 2016Google Scholar
  7. 7.
    J.D. Musa, Software Reliability Engineering: More Reliable Software, Faster and Cheaper (TataMcGraw-HillEducation, Springer, 2004.) Scholar
  8. 8.
    P.K. Kapur, R.B. Garg, S. Kumar, Contributions to Hardware and Software Reliability (World Scientific, Singapore, 1999)CrossRefGoogle Scholar
  9. 9.
    M.R. Lyu, Handbook of Software Reliability Engineering (McGraw Hill, Springer, 1996)Google Scholar
  10. 10.
    P.K. Kapur, S. Agarwal, R.B. Garg, Bi-criterion release policy for exponential software reliability growth models. Oper. Res. (Recherche Operationanelle) 28, 165–180 (1994)zbMATHGoogle Scholar
  11. 11.
    C. Stringfellow et al., An empirical method for selecting software reliability growth models. Empir. Softw. Eng., 319–343 (2002)Google Scholar
  12. 12.
  13. 13.
    S.Z. Ke, C.Y. Huang, Software reliability analysis considering the variation of testing-effort and change-point, in Proceedings of the International Workshop on Innovative Software Development Methodologies and Practices, pp 30–39, ISBN: 978-1-4503-3226-2,
  14. 14.
    S.K. Khatri, S. John, R. Majumdar, Quantifying software reliability using testing effort, in International Conference on Information Technology, Amity University, 6–7 October 2016Google Scholar
  15. 15.
    P.K. Kapur, H. Pham, A. Gupta, P.C. Jha, Software Reliability Assessment with OR Application (Springer, Berlin, 2011)CrossRefGoogle Scholar

Copyright information

© Springer International Publishing AG, part of Springer Nature 2018

Authors and Affiliations

  1. 1.Amity School of Engineering and Technology, Amity UniversityNoidaIndia

Personalised recommendations