This research paper discusses a method for human action recognition utilizing spatio-temporal features through transfer learning via convolutional neural networks (CNNs) like ResNet50 and VGG16. The approach achieves a remarkable accuracy of 89.71% on the UCF-101 dataset, analyzing video data to categorize activities such as basketball playing and cycling. Additionally, the study emphasizes the challenges in accurately recognizing human actions due to environmental factors while proposing enhancements in personal security and elder care applications.
Related topics: