Published July 30, 2024 | Version CC-BY-NC-ND 4.0
Journal article Open

Promoting Project Outcomes: A Development Approach to Generative AI and LLM-Based Software Applications' Deployment

  • 1. Ph.D.; Director, Master Program in Industrial Engineering, Mehr Astan University, Gilan, Iran.

Contributors

  • 1. Generative AI Researcher and Developer, Tehran, Iran.
  • 2. Ph.D.; Director, Master Program in Industrial Engineering, Mehr Astan University, Gilan, Iran.

Description

Abstract: In the dynamic realm of artificial intelligence, the emergence of Generative Artificial Intelligence (GAI) has marked a revolutionary stride, particularly in the context of project execution models. This paper delves deep into the sophisticated architectures of GAI, mainly focusing on Large Language Models (LLMs) such as GPT-3 and BERT and their practical applications across varied scenarios. The intricacies of deploying these models have been effectively unraveled to ensure resonating with the specific demands of distinct cases, falling within departmental integration, medical diagnostics, or tailored training modules. Central to the proposed exposition is the innovative "Forward and Back Systematic Approach" designed for executing GAI projects. This approach is meticulously structured to enhance efficiency and ensure a harmonious alignment with the nuanced requirements of diverse applications. We dissect some strategies, including leveraging Private Generalized LLM APIs, in-context learning (ICL), and fine-tuning methodologies, to empower these models to adapt and excel. Furthermore, the proposed platform underscores the pivotal role of evaluation criteria in refining GAI project outcomes, ensuring each model's prowess. It is not strictly theoretical but yields tangible benefits in real-world applications. Under the aegis of this comprehensive exploration, the result of the study would serve as a beacon for enthusiasts and professionals navigating the GAI landscape by offering insights into optimizing robust models for specific and case-driven utilities. Standing on the brink of a modern era in AI, this paper contributes a substantial framework and critical analysis, steering the course for future innovations and applications of GAI.

Files

D363614040924.pdf

Files (603.7 kB)

Name Size Download all
md5:d4ecb9243e3bbc08e5e6c957aaf86b7d
603.7 kB Preview Download

Additional details

Identifiers

Dates

Accepted
2024-07-15
Manuscript received on 04 July 2024 | Revised Manuscript received on 09 July 2024 | Manuscript Accepted on 15 July 2024 | Manuscript published on 30 July 2024.

References

  • Ng, A. and Jordan, M. (2001).On discriminative vs. generative classifiers: A comparison of logistic regression and naive Bayes.In Advances in Neural Information Processing Systems, volume 14, pages 841––848.
  • Hu, L. (2023). Generative AI and Future. Retrieved on January 23 from https://pub.towardsai.net/generative-ai-and-future-c3b1695876f2.
  • Jovanović, M. (2023). Generative Artificial Intelligence: Trends and Prospects. https://www.computer.org/csdl/magazine/co/2022/10/09903869/1H0G6xvtREk.0.1109/MC.2022.3192720.
  • Abukmeil, M., Ferrari, S., Genovese, A., Piuri, V., & Scotti, F. (2021). A survey of unsupervised generativemodels for exploratory data analysis and representation learning. Acm computing surveys (csur), 54(5), https://doi.org/10.1145/3450963
  • Gui, J., Sun, Z., Wen, Y., Tao, D., & Ye, J. (2021). A review on generative adversarial networks: Algorithms, theory, and applications. IEEE Transactions on Knowledge and Data Engineering. doi: 10.1109/TKDE.2021.3130191. https://doi.org/10.1109/TKDE.2021.3130191
  • Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., et al. (2020). Language models are few-shot learners. Advances in Neural Information Processing Systems, 33: 1877-1901.
  • Aydın, Ö., & Karaarslan, E. (2023). Is ChatGPT leading generative AI? What is beyond expectations? Academic Platform Journal of Engineering and Smart Systems. https://doi.org/10.2139/ssrn.4341500
  • Pavlik, J. V. (2023). Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education. Journalism & Mass Communication Educator, 0(0). https://doi.org/10.1177/10776958221149577
  • Kulkarni, A., Shivananda, A., Kulkarni, A., & Gudivada, D. (2023). Applied Generative AI for Beginners: Practical Knowledge on Diffusion Models, ChatGPT, and Other LLMs. https://doi.org/10.1007/978-1-4842-9994-4
  • Spitale, G., Biller-Andorno, N., and Germani, F. (2023).AI model GPT-3 (dis) informs us better than humans.arXiv:2301.11924. https://doi.org/10.1126/sciadv.adh1850
  • Ji, Z., Lee, N., Frieske, R., Yu, T., Su, D., Xu, Y., Ishii, E., Bang, Y. J., Madotto, A., and Fung, P. (2023).Survey of hallucination in natural language generation.ACM Computing Surveys, 55(12):1–38. https://doi.org/10.1145/3571730
  • B. A. y Arcas, (2022). "Do large language models understand us?" Daedalus, vol. 151, no. 2, pp. 183–197. https://doi.org/10.1162/daed_a_01909
  • Zellers, R., Holtzman, A., Peters, M. E., Mottaghi, R., Choi, Y., Kuehlkamp, A., ... & Farhadi, A. (2023). A Survey for In-context Learning. ar5iv. Retrieved from https://ar5iv.labs.arxiv.org/html/2301.00234
  • Bai, Y., Jones, A., Ndousse, K., Askell, A., Chen, A., DasSarma, N., Drain, D., Fort, S., Ganguli, D., Henighan, T., Joseph, N., Kadavath, S., Kernion, J., Conerly, T., El Showk, S., Elhage, N., Hatfield-Dodds, Z., Hernandez, D., Hume, T., Johnston, S., Kravec, S., Lovitt, L., Nanda, N., Olsson, C., Amodei, D., Brown, T. B., Clark, J., McCandlish, S., Olah, C., Mann, B., Kaplan, J. (2022). Training a helpful and harmless assistant with reinforcement learning from human feedback. CoRR, abs/2204.05862.
  • Chung, H. W., Hou, L., Longpre, S., Zoph, B., Tay, Y., Fedus, W., Li, E., Wang, X., Dehghani, M., Brahma, S., Webson, A., Gu, S. S., Dai, Z., Suzgun, M., Chen, X., Chowdhery, A., Narang, S., Mishra, G., Yu, A., Zhao, V. Y., Huang, Y., Dai, A. M., Yu, H., Petrov, S., Chi, E. H., Dean, J., Devlin, J., Roberts, A., Zhou, D., Le, Q. V., & Wei, J. (2022). Scaling instruction-finetuned language models. CoRR, abs/2210.11416.
  • Guo, Z., Jin, R., Liu, C., Huang, Y., Shi, D., Supryadi, Linhao, Y., Liu, Y., Li, J., Xiong, B., & Xiong, D. (2023). Evaluating Large Language Models: A Comprehensive Survey. arXiv. https://arxiv.org/abs/2310.19736
  • Pichai, S. (2023).An important next step on our AIjourney.Accessed on 03/16/23.
  • Mehdi, Y. (2023). Reinventing search with a new AI-powered Microsoft Bing and Edge, your copilot for the web. Official Microsoft Blog, 7.
  • Blodgett, S. L., Barocas, S., Daum'e, H., and Wallach, H. M. (2020).Language (technology) is power: A critical survey of "bias" in nlp.In Annual Meeting of the Association for Computational Linguistics. https://doi.org/10.18653/v1/2020.acl-main.485
  • Jones, E. and Steinhardt, J. (2022).Capturing failures of large language models via human cognitive biases.In Oh, A. H., Agarwal, A., Belgrave, D., and Cho, K., editors, Advances in Neural Information Processing Systems.
  • Adik, K. (2023, July 13). Generative AI Project Life Cycle. Medium. https://medium.com/@kanikaadik07/generative-ai-project-life-cycle-55ce9092e24a
  • Alenezi, M., & Almuairfi, S. (2019). Security Risks in the Software Development Lifecycle. In International Journal of Recent Technology and Engineering (IJRTE) (Vol. 8, Issue 3, pp. 7048–7055). https://doi.org/10.35940/ijrte.c5374.098319