dc.identifier.citation |
[1] H. Naveed, A. U. Khan, S. Qiu, et al., “A comprehensive overview of large language mod- els,” arXiv preprint arXiv:2307.06435, 2023. [2] H. W. Chung, L. Hou, S. Longpre, et al., “Scaling instruction-finetuned language models,” arXiv preprint arXiv:2210.11416, 2022. [3] N. Fatima, A. S. Imran, Z. Kastrati, S. M. Daudpota, and A. Soomro, “A systematic liter- ature review on text generation using deep neural network models,” IEEE Access, vol. 10, pp. 53 490–53 503, 2022. [4] J. Li, T. Tang, W. X. Zhao, J.-Y. Nie, and J.-R. Wen, “Pretrained language models for text generation: A survey,” arXiv preprint arXiv:2201.05273, 2022. [5] K. Pandya and M. Holia, “Automating customer service using langchain: Building custom open-source gpt chatbot for organizations,” arXiv preprint arXiv:2310.05421, 2023. [6] M. A. K. Raiaan, M. S. H. Mukta, K. Fatema, et al., “A review on large language models: Architectures, applications, taxonomies, open issues and challenges,” IEEE Access, 2024. [7] C. Raffel, N. Shazeer, A. Roberts, et al., Exploring the limits of transfer learning with a unified text-to-text transformer, 2023. arXiv: 1910.10683[cs.LG]. [8] M. Khorshed, “2022 report on investment mechanism of islami bank bangladesh limited i research report investment mechanism of islami bank bangladesh limited course title: Research matodology submitted to submitted by mohammed khorshed bba in finance report on investment mechanism of islami subject: Submission of research report titled on”report on investment mechanism of islami bank bangladesh limited.”,” Ph.D. dissertation, Jun. 2022. 41 Bibliography 42 [9] S. Liu, Y. Chen, X. Xie, J. Siow, and Y. Liu, Retrieval-augmented generation for code summarization via hybrid gnn, 2021. arXiv: 2006.05405[cs.LG]. [10] Y. Gao, Y. Xiong, X. Gao, et al., Retrieval-augmented generation forlarge language mod- els: A survey, 2024. arXiv: 2312.10997[cs.CL]. [11] Z. Sun and Z. Wu, Handbook of Research on Foundations and Applications of Intelligent Business Analytics. IGI Global, 2022. [12] M. Zhang and J. Li, “A commentary of gpt-3 in mit technology review 2021,” Fundamental Research, vol. 1, no. 6, pp. 831–833, 2021. [13] L. Floridi and M. Chiriatti, “Gpt-3: Its nature, scope, limits, and consequences,” Minds and Machines, vol. 30, pp. 681–694, 2020. [14] H. Touvron, L. Martin, K. Stone, et al., “Llama 2: Open foundation and fine-tuned chat models,” arXiv preprint arXiv:2307.09288, 2023 |
en_US |