Software testing is an activity which is aimed for evaluating quality of a program and also for improving it, by identifying defects and problems. Software testing strives for achieving its goals (both implicit and explicit) but it does have certain limitations, still testing can be done more effectively if certain established principles are be followed. In spite of having limitations, software testing continues to dominate other verification techniques like static analysis, model checking, and proofs. So it is indispensable to understand the goals, principles and limitations of software testing so that the effectiveness of software testing could be maximized.
General TermsSoftware Engineering, Software Testing.
Artificial Intelligence (AI) is making a significant impact in multiple areas like medical, military, industrial, domestic, law, arts as AI is capable to perform several roles such as managing smart factories, driving autonomous vehicles, creating accurate weather forecasts, detecting cancer and personal assistants, etc. Software testing is the process of putting the software to test for some abnormal behaviour of the software. Software testing is a tedious, laborious and most time-consuming process. Automation tools have been developed that help to automate some activities of the testing process to enhance quality and timely delivery. Over time with the inclusion of continuous integration and continuous delivery (CI/CD) pipeline, automation tools are becoming less effective. The testing community is turning to AI to fill the gap as AI is able to check the code for bugs and errors without any human intervention and in a much faster way than humans. In this study, we aim to recognize the impact of AI technologies on various software testing activities or facets in the STLC. Further, the study aims to recognize and explain some of the biggest challenges software testers face while applying AI to testing. The paper also proposes some key contributions of AI in the future to the domain of software testing.
Software reliability growth models (SRGM) are employed to aid us in predicting and estimating reliability in the software development process. Many SRGM proposed in the past claim to be effective over previous models. While some earlier research had raised concern regarding use of delayed S-shaped SRGM, researchers later indicated that the model performs well when appropriate testing-effort function (TEF) is used. This paper proposes and evaluates an approach to incorporate the log-logistic (LL) testing-effort function into delayed S-shaped SRGMs with imperfect debugging based on non-homogeneous Poisson process (NHPP). The model parameters are estimated by weighted least square estimation (WLSE) and maximum likelihood estimation (MLE) methods. The experimental results obtained after applying the model on real data sets and statistical methods for analysis are presented. The results obtained suggest that performance of the proposed model is better than the other existing models. The authors can conclude that the log-logistic TEF is appropriate for incorporating into delayed S-shaped software reliability growth models.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.