Generating Minutes of Meeting Using LLM
Authors/Creators
Description
Meetings represent the predominant communication and decision-making medium within organizational frameworks. However, the manual extraction of meeting transcripts from audio recordings is both time-consuming and labor-intensive. This project aims to automate this process by utilizing advanced Large Language Models (LLM) and Speech Recognition techniques through various Python libraries and APIs. The workflow encompasses uploading video or audio content, extracting audio from meeting videos, and transcribing it into textual transcripts, summary and generates minutes via various machine learning models. Sophisticated speaker diarization and identification techniques are applied for segmentation, providing an individual understanding of each speaker's contributions during the meeting, thereby enhancing contextual richness. Furthermore, the project generates concise summaries of the meeting, distilling key insights, action items, and decision points. By automating transcription, speaker segmentation, and summarization, this project streamlines the extraction of valuable insights from meeting audio and video recordings. The resulting transcripts, summaries, and keynotes facilitate better information sharing, knowledge retention, and informed decision-making within organizations, bridging the gap between unstructured audio data and actionable insights.
Files
Generating Minutes Of Meeting Using LLM -Formatted Paper.pdf
Files
(436.9 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:a797be002dc926caeb0bad267aea6387
|
436.9 kB | Preview Download |
Additional details
References
- 1. Xue, B., Hu, S., Xu, J., Geng, M., Liu, X., & Meng, H. (2022). Bayesian neural network language modeling for speech recognition. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 30, 2900-2917.
- 2. Horiguchi, S., Fujita, Y., Watanabe, S., Xue, Y., & Garcia, P. (2022). Encoder-decoder based attractors for end-to-end neural diarization. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 30, 1493-1507.
- 3. Kanda, N., Xiao, X., Gaur, Y., Wang, X., Meng, Z., Chen, Z., & Yoshioka, T. (2022, May). Transcribe-to-diarize: Neural speaker diarization for unlimited number of speakers using end-to-end speaker-attributed ASR. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 8082-8086). IEEE.
- 4. See, A., Liu, P. J., & Manning, C. D. (2017). Get to the point: Summarization with pointer-generator networks. arXiv preprint arXiv:1704.04368.
- 5. Banerjee, S., Mitra, P., & Sugiyama, K. (2015, September). Generating abstractive summaries from meeting transcripts. In Proceedings of the 2015 ACM Symposium on Document Engineering (pp. 51-60).
- 6. Kryściński, W., McCann, B., Xiong, C., & Socher, R. (2019). Evaluating the factual consistency of abstractive text summarization. arXiv preprint arXiv:1910.12840.