Trends in SDLC Document Review using Generative AI

Authors

  • Prashant Gulave Department of Computer Science, JSPM’s Rajarshi Shahu College of Engineering, Pune, Maharashtra, India Author
  • Dr. Kavita Moholkar Department of Computer Science, JSPM’s Rajarshi Shahu College of Engineering, Pune, Maharashtra, India Author

DOI:

https://doi.org/10.32628/CSEIT25112777

Keywords:

Generative AI, SDLC, Document Review, Large Language Models, NLP

Abstract

This research paper explores the evolving role of Generative AI in Software Development Life Cycle (SDLC) document review. With AI-driven advancements in Natural Language Processing (NLP), models such as GPT, BERT, and domain-specific LLMs have been adapted to evaluate requirement specifications, test plans, and design documents. We present an analysis of how these models are being fine-tuned for document validation, compliance checking, and contextual feedback generation in the software industry. The paper also examines the integration of rule-based methods with AI, providing structured feedback for engineering domain documentation. Furthermore, we discuss emerging trends, challenges, and future research directions for enhancing AI-based document review in SDLC.

Downloads

Download data is not yet available.

References

Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. (2019). "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding." In Proceedings of NAACL-HLT.

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... & Polosukhin, I. (2017). "Attention is All You Need." NeurIPS.

Brown, T., Mann, B., Ryder, N., et al. (2020). "Language Models are Few-Shot Learners." NeurIPS.

Radford, A., Wu, J., Child, R., et al. (2019). "Language Models are Unsupervised Multitask Learners." OpenAI.

Zhang, Y., Roller, S., Goyal, N., et al. (2020). "Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension." ACL.

Howard, J., & Ruder, S. (2018). "Universal Language Model Fine-tuning for Text Classification." ACL.

Raffel, C., Shazeer, N., Roberts, A., et al. (2020). "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer." JMLR.

OpenAI. (2023). "GPT-4 Technical Report." arXiv preprint arXiv:2303.08774.

McCandlish, S., Radford, A., et al. (2018). "An Empirical Model of Large-Batch Training." arXiv.

Vaswani, A., Parmar, N., et al. (2018). "Scaling Neural Machine Translation." NeurIPS.

Downloads

Published

02-04-2025

Issue

Section

Research Articles