Balancing Automation with Human Expertise in Exploratory Testing and Edge-Case Analysis

Article Sidebar

Main Article Content

Nusrat Yasmin Nadia
Mohammed Majid Bakhsh
Gazi Touhidul Alam
Abstract: Software systems requiring complex testing benefit greatly from both exploratory testing and edge-case analysis because these methods reveal defects that automated testing would miss. The growing use of automation to quicken testing operations requires organizations to achieve proper equilibrium between automated tools and skilled human labor. This document examines how machine testing benefits from human involvement to perform exploratory testing and identify any hidden edge-case situations. The advantage of automation tools lies in their ability to deliver both speed and uniformity but they struggle to replicate essential human capabilities which help recognize hidden issues in unusual circumstances (Baresi et al., 2022). Human testers have a strong ability to detect stealthy defects yet their testing capabilities face challenges in achieving complete scalability and coverage (Smith & Zhang, 2021).

This paper evaluates the difficulties and advantages of uniting automated testing solutions with people-driven exploratory testing during edge-case scenario analysis. The combination of automated testing and human tester intervention represents the best approach according to current industry practices since automation performs repetitive tasks and human testers handle critical thinking tasks and pattern recognition with defect discovery responsibilities (Carver et al., 2023). Our research develops a framework to merge automated and manual testing methods and it examines AI exploratory testing solutions and human-assisted automated testing environments for future development. Organizations need to completely evaluate their testing methods so they can combine automated functions with human cognition properly for enhanced testing methodology (Jain & Singh, 2020).

Balancing Automation with Human Expertise in Exploratory Testing and Edge-Case Analysis. (2025). International Journal of Latest Technology in Engineering Management & Applied Science, 14(2), 302-313. https://doi.org/10.51583/IJLTEMAS.2025.14020031

Downloads

References

Bertolino, A., Gnesi, S., & Polini, A. (2023). Explainable AI in Software Testing: Challenges and Opportunities. Journal of Software Engineering Research and Development, 11(2), 105-123.

Carver, J., Hassan, A., & Zhang, Y. (2023). Implementing Human-in-the-Loop AI Testing Frameworks. IEEE Transactions on Software Engineering, 49(1), 45-67.

Johnson, M., & Lee, T. (2023). Large Language Models in Automated Test Generation: A Review. ACM Computing Surveys, 56(3), 12-34.

Lopez, R., & Carter, K. (2023). Self-Healing Test Automation in Continuous Integration Pipelines. Software Quality Journal, 31(4), 211-235.

Martinez, P., & Singh, D. (2023). AI-Powered Test Oracles for Edge-Case Validation. International Journal of Software Testing and Verification, 20(2), 89-112.

Nguyen, H., Patel, R., & Williams, S. (2023). Adaptive Exploratory Testing with Machine Learning. Software Testing, Verification & Reliability, 33(1), 56-78.

Rodriguez, L., & Hayes, J. (2023). The Role of AI in Defect Prediction and Risk-Based Testing. Empirical Software Engineering, 28(3), 129-154.

Taylor, B., & Green, C. (2023). Upskilling QA Engineers for AI-Augmented Testing: Strategies and Challenges. Journal of Software Testing & Quality Assurance, 14(2), 178-196.

Turner, E., & Carter, D. (2023). Ethical Considerations in AI-Assisted Software Testing. AI & Society, 38(1), 55-72.

Article Details

How to Cite

Balancing Automation with Human Expertise in Exploratory Testing and Edge-Case Analysis. (2025). International Journal of Latest Technology in Engineering Management & Applied Science, 14(2), 302-313. https://doi.org/10.51583/IJLTEMAS.2025.14020031