*Result*: Hybrid Deep Learning Techniques for Video Processing using attention-based LSTM and 3D CNN.

Title:
Hybrid Deep Learning Techniques for Video Processing using attention-based LSTM and 3D CNN.
Authors:
Nagaraju, P.1 nagarajupampati1234@gmail.com, Manchala, Sadanandam1 msadanandam@kakatiya.ac.in
Source:
Journal of Computational Analysis & Applications. 2025, Vol. 34 Issue 8, p406-433. 28p.
Database:
Academic Search Index

*Further Information*

*Large-scale video classification is an essential task in computer vision, with applications spanning autonomous systems, surveillance, and content retrieval. Videos are complex as they involve both spatial and temporal dynamics, requiring models that can process this vast amount of information efficiently. Traditional approaches struggle with the high dimensionality and temporal continuity of video data. By combining a Long Short-Term Memory network, a tweaked and optimized 3D Convolutional Neural Network, and attention processes, we provide a unique method for video action identification. When it comes to handling the complexities of real-world situations, this synergy improves overall performance and gives an edge over current approaches. The ability of our method to extract both temporal and spatial information from video sequences, together with the addition of an attention mechanism that highlights certain regions in the sequences to improve identification accuracy, makes it unique. Complex situations, including those involving several actors or objects or occlusion, are especially wellsuited for the model. It successfully deals with the subjectivity and unpredictability present in action annotations in datasets. Additionally, we use a variety of preprocessing methods to enhance model performance. Based on thorough experimental tests on the UCF101 and HMDB51 benchmark datasets, we show that our suggested strategy performs noticeably better than the current state-of-the-art techniques in action recognition. These results highlight our approach's potential for future developments in the field of video action detection research. [ABSTRACT FROM AUTHOR]*