A Cascaded Architecture for Extractive Summarization of Multimedia Content via Audio-to-Text Alignment
By: Tanzir Hossain , Ar-Rafi Islam , Md. Sabbir Hossain and more
Potential Business Impact:
Summarizes long videos into short, easy text.
This study presents a cascaded architecture for extractive summarization of multimedia content via audio-to-text alignment. The proposed framework addresses the challenge of extracting key insights from multimedia sources like YouTube videos. It integrates audio-to-text conversion using Microsoft Azure Speech with advanced extractive summarization models, including Whisper, Pegasus, and Facebook BART XSum. The system employs tools such as Pytube, Pydub, and SpeechRecognition for content retrieval, audio extraction, and transcription. Linguistic analysis is enhanced through named entity recognition and semantic role labeling. Evaluation using ROUGE and F1 scores demonstrates that the cascaded architecture outperforms conventional summarization methods, despite challenges like transcription errors. Future improvements may include model fine-tuning and real-time processing. This study contributes to multimedia summarization by improving information retrieval, accessibility, and user experience.
Similar Papers
Breaking the Barriers of Text-Hungry and Audio-Deficient AI
Sound
Lets computers understand and speak any language.
Rethinking Multimodal Sentiment Analysis: A High-Accuracy, Simplified Fusion Architecture
Computation and Language
Helps computers understand feelings from talking, seeing, and hearing.
CFSum: A Transformer-Based Multi-Modal Video Summarization Framework With Coarse-Fine Fusion
CV and Pattern Recognition
Makes video summaries using sound, text, and pictures.