Sequential Requirements Prediction in Synthetic Fintech-Like Backlogs Using an Interpretable Hybrid of Transition Rules and Transformer Models

Authors

  • Diana Laily Fithri Muria Kudus University, Indonesia
  • Soni Adiyono Muria Kudus University, Indonesia
  • Muhammad Arifin Muria Kudus University, Indonesia
Pages Icon

DOI:

https://doi.org/10.63158/journalisi.v8i2.1469

Keywords:

Requirements Engineering, Sequential Requirements Prediction, Synthetic Fintech-Like Backlogs, Semantic Normalization, Transformer

Abstract

Fintech software development is characterized by rapid product iteration and stringent regulatory requirements, resulting in changing requirements as an interrelated set rather than individual elements. In this research, Sequential Requirements Prediction is proposed as a decision support task in Requirements Engineering, where the time-ordered prefix of completed backlog items is used to predict the next likely canonical requirement type as Top-k ranked output. To mitigate noise and inconsistency in backlog data, an LLM-aided semantic normalization step maps diverse requirement descriptions to a closed set of fintech requirement types. The research compares an interpretable rule-based Markov-1 predictor with Transformer-based sequential predictors under a case-level time-aware split. The proposed method is evaluated on a synthetic fintech-like backlog dataset consisting of 900 cases, 5,252 events, and 18 canonical requirement types. The best-performing model, Transformer + normalization + augmentation (M4), achieved Recall@5 = 0.638889, MRR@5 = 0.536806, and NDCG@5 = 0.566667. These results surpassed the rule-based predictor and non-normalized Transformer model. In addition, augmentation further improved Recall@5 from 0.493056 to 0.527778 in the rare-type subset. These findings suggest the methodological promise of the proposed framework for sequence-aware and compliance-conscious backlog analytics in synthetic fintech-like settings.

Downloads

Download data is not yet available.

References

[1] S. V. A. Ammu, S. S. Sehra, S. K. Sehra, and J. Singh, “Amalgamation of Classical and Large Language Models for Duplicate Bug Detection: A Comparative Study,” Computers, Materials and Continua, vol. 83, no. 1, pp. 435–453, 2025, doi: https://doi.org/10.32604/cmc.2025.057792.

[2] T. F. Boka, Z. Niu, and R. B. Neupane, “A survey of sequential recommendation systems: Techniques, evaluation, and future directions,” Inf. Syst., vol. 125, p. 102427, 2024, doi: https://doi.org/10.1016/j.is.2024.102427.

[3] L.-W. Pan, W.-K. Pan, M.-Y. Wei, H.-Z. Yin, and Z. Ming, “A survey on sequential recommendation,” Front. Comput. Sci., vol. 20, no. 3, p. 2003606, 2025, doi: 10.1007/s11704-025-41329-w.

[4] Z. Zheng et al., “Towards an understanding of large language models in software engineering tasks,” Empir. Softw. Eng., vol. 30, no. 2, p. 50, 2024, doi: 10.1007/s10664-024-10602-0.

[5] S. Saleem, M. N. Asim, L. Van Elst, and A. Dengel, “Generative language models potential for requirement engineering applications: insights into current strengths and limitations,” Complex & Intelligent Systems, vol. 11, no. 6, p. 278, 2025, doi: 10.1007/s40747-024-01707-6.

[6] J. Zhang, J. Zhou, N. Niu, J. Hua, and C. Liu, “Synergistic enhancement of requirement-to-code traceability: A framework combining large language model based data augmentation and an advanced encoder,” Inf. Softw. Technol., vol. 193, p. 108045, 2026, doi: 10.1016/j.infsof.2026.108045.

[7] A. Majidzadeh, M. Ashtiani, and M. Zakeri-Nasrabadi, “Multi-type requirements traceability prediction by code data augmentation and fine-tuning MS-CodeBERT,” Comput. Stand. Interfaces, vol. 90, p. 103850, 2024, doi: 10.1016/j.csi.2024.103850.

[8] A. Ferrari and P. Spoletini, “Formal requirements engineering and large language models: A two-way roadmap,” Inf. Softw. Technol., vol. 181, p. 107697, 2025, doi: 10.1016/j.infsof.2025.107697.

[9] F. R. Arnob, R. H. Mollik, P. Goyal, and R. Bryce, “Bug Triaging Based on Transformer Models Utilizing Commit Messages,” in The 22nd International Conference on Information Technology-New Generations (ITNG 2025), S. Latifi, Ed., Cham: Springer Nature Switzerland, 2025, pp. 354–366.

[10] K. Ronanki, B. Cabrero-Daniel, and C. Berger, “ChatGPT as a Tool for User Story Quality Evaluation: Trustworthy Out of the Box?,” in Agile Processes in Software Engineering and Extreme Programming – Workshops, P. Kruchten and P. Gregory, Eds., Cham: Springer Nature Switzerland, 2024, pp. 173–181.

[11] J. Zhang, J. Zhou, N. Niu, J. Hua, and C. Liu, “Synergistic Enhancement of Requirement-to-Code Traceability: A Framework Combining Large Language Model based Data Augmentation and an Advanced Encoder,” Oct. 2025, [Online]. Available: http://arxiv.org/abs/2509.20149

[12] B. Wang, Z. Zou, X. Liang, H. Jin, and P. Liang, “HGNNLink: recovering requirements-code traceability links with text and dependency-aware heterogeneous graph neural networks,” Automated Software Engineering, vol. 32, no. 2, p. 55, 2025, doi: 10.1007/s10515-025-00528-2.

[13] S. Rezig, Z. Achour, and N. Rezg, “Retraction:Using data mining methods for predicting sequential maintenance activities,” Nov. 07, 2018, Multidisciplinary Digital Publishing Institute (MDPI). doi: 10.3390/app8112184.

[14] A. Majidzadeh, M. Ashtiani, and M. Zakeri-Nasrabadi, “Multi-type requirements traceability prediction by code data augmentation and fine-tuning MS-CodeBERT,” Comput. Stand. Interfaces, vol. 90, p. 103850, 2024, doi: 10.1016/j.csi.2024.103850.

[15] A. Vogelsang and J. Fischbach, “Using Large Language Models for Natural Language Processing Tasks in Requirements Engineering: A Systematic Guideline,” May 2024, [Online]. Available: http://arxiv.org/abs/2402.13823

[16] Z. Zheng et al., “Towards an understanding of large language models in software engineering tasks,” Empir. Softw. Eng., vol. 30, no. 2, p. 50, 2024, doi: 10.1007/s10664-024-10602-0.

[17] N. Marques, R. R. Silva, and J. Bernardino, “Using ChatGPT in Software Requirements Engineering: A Comprehensive Review,” Future Internet, vol. 16, no. 6, 2024, doi: 10.3390/fi16060180.

[18] N. N. Tabari, “Automated Agile User Story Quality Assessment with Natural Language Processing,” 2024.

[19] A. Sharma and A. Kumar Tripathi, “Evaluating user story quality with LLMs: a comparative study,” J. Intell. Inf. Syst., vol. 63, no. 4, pp. 1423–1451, 2025, doi: 10.1007/s10844-025-00939-3.

[20] T. F. Boka, Z. Niu, and R. B. Neupane, “A survey of sequential recommendation systems: Techniques, evaluation, and future directions,” Inf. Syst., vol. 125, p. 102427, 2024, doi: 10.1016/j.is.2024.102427.

[21] C. Ellsel and R. Stark, “Advancing Requirements Engineering with Large Language Models,” Procedia CIRP, vol. 136, pp. 701–706, 2025, doi: 10.1016/j.procir.2025.08.120.

[22] A. Hemmat, M. Sharbaf, S. Kolahdouz-Rahimi, K. Lano, and S. Y. Tehrani, “Research directions for using LLM in software requirement engineering: a systematic review,” Front. Comput. Sci., vol. 7, 2025, doi: 10.3389/fcomp.2025.1519437.

[23] X. Du, Z. Liu, C. Li, X. Ma, Y. Li, and X. Wang, “LLM-BRC: A large language model-based bug report classification framework,” Software Quality Journal, vol. 32, no. 3, pp. 985–1005, 2024, doi: 10.1007/s11219-024-09675-3.

Downloads

Published

2026-04-12

Issue

Section

Articles

How to Cite

[1]
D. L. Fithri, S. Adiyono, and M. Arifin, “Sequential Requirements Prediction in Synthetic Fintech-Like Backlogs Using an Interpretable Hybrid of Transition Rules and Transformer Models”, journalisi, vol. 8, no. 2, pp. 1871–1890, Apr. 2026, doi: 10.63158/journalisi.v8i2.1469.

Most read articles by the same author(s)