TY - GEN
T1 - AMP-BiLSTM
T2 - 18th IEEE International Conference on Semantic Computing, ICSC 2024
AU - Lin, Sheng Jie
AU - Chen, Chien Chin
AU - Chang, Yung Chun
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - With the rise of conversation-oriented streaming videos, the platforms that host them like Twitch have rapidly become prominent information hubs. However, the lengthy nature of such streams often deters viewers from consuming the full content. To mitigate this, we propose AMP-BiLSTM, a novel highlight extraction method which focuses on the textual information in streamer discourses and viewer responses rather than visual features. This approach addresses the limitations of previous methods, which primarily centered on analyzing visual features, and were thus insufficient for conversation-oriented streams where highlights emerge from dialogues and viewer interactions. AMP-BiLSTM is built on techniques of Attention, Multi-channel, and Position enrichment integrated into a Bidirectional Long Short-Term Memory (BiLSTM) network. Through experiments on a real-world dataset, we found that streamer discourses and viewer messages provide significant utility for highlight extraction in conversation-oriented streaming videos. Furthermore, our proposed Multi-channel and self-attention techniques effectively distill text streams into semantically-rich embeddings. The experiment results demonstrate that AMP-BiLSTM outperforms several state-of-the-art methods for deep learning-based highlight extraction, thus showing promise for improved conversation-oriented streaming video content digestion.
AB - With the rise of conversation-oriented streaming videos, the platforms that host them like Twitch have rapidly become prominent information hubs. However, the lengthy nature of such streams often deters viewers from consuming the full content. To mitigate this, we propose AMP-BiLSTM, a novel highlight extraction method which focuses on the textual information in streamer discourses and viewer responses rather than visual features. This approach addresses the limitations of previous methods, which primarily centered on analyzing visual features, and were thus insufficient for conversation-oriented streams where highlights emerge from dialogues and viewer interactions. AMP-BiLSTM is built on techniques of Attention, Multi-channel, and Position enrichment integrated into a Bidirectional Long Short-Term Memory (BiLSTM) network. Through experiments on a real-world dataset, we found that streamer discourses and viewer messages provide significant utility for highlight extraction in conversation-oriented streaming videos. Furthermore, our proposed Multi-channel and self-attention techniques effectively distill text streams into semantically-rich embeddings. The experiment results demonstrate that AMP-BiLSTM outperforms several state-of-the-art methods for deep learning-based highlight extraction, thus showing promise for improved conversation-oriented streaming video content digestion.
KW - Attention Mechanism
KW - BiLSTM
KW - Conversation-Oriented Streaming Video
KW - Deep Learning
KW - Highlight Extraction
KW - Multi-channel Analysis
KW - Position Enrichment
UR - http://www.scopus.com/inward/record.url?scp=85192259821&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85192259821&partnerID=8YFLogxK
U2 - 10.1109/ICSC59802.2024.00009
DO - 10.1109/ICSC59802.2024.00009
M3 - Conference contribution
AN - SCOPUS:85192259821
T3 - Proceedings - IEEE International Conference on Semantic Computing, ICSC
SP - 9
EP - 16
BT - Proceedings - 18th IEEE International Conference on Semantic Computing, ICSC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 5 February 2024 through 7 February 2024
ER -