Presentation + Paper
7 June 2024 Enhancing human action recognition with GAN-based data augmentation
Author Affiliations +
Abstract
Deep Neural Networks (DNNs) have emerged as a powerful tool for human action recognition, yet their reliance on vast amounts of high-quality labeled data poses significant challenges. A promising alternative is to train the network on generated synthetic data. However, existing synthetic data generation pipelines require complex simulation environments. Our novel solution bypasses this requirement by employing Generative Adversarial Networks (GANs) to generate synthetic data from only a small existing real-world dataset. Our training pipeline extracts the motion from each training video and augments it across various subject appearances within the training set. This approach increases the diversity in both motion and subject representations, thus significantly enhancing the model's performance. A rigorous evaluation of the model's performance is presented under diverse scenarios, including ground and aerial views. Moreover, an insightful analysis of critical factors influencing human action recognition performance, such as gesture motion diversity and subject appearance, is presented.
Conference Presentation
(2024) Published by SPIE. Downloading of the abstract is permitted for personal use only.
Prasanna Reddy Pulakurthi, Celso M. De Melo, Raghuveer Rao, and Majid Rabbani "Enhancing human action recognition with GAN-based data augmentation", Proc. SPIE 13035, Synthetic Data for Artificial Intelligence and Machine Learning: Tools, Techniques, and Applications II, 130350O (7 June 2024); https://doi.org/10.1117/12.3021572
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Education and training

Video

Action recognition

Motion models

Neural networks

Deep learning

Gesture recognition

Back to Top