I recently had the opportunity to present my research at the Fifth Doctoral Consortium in Computer Science (JIPII 2025), hosted by the Escuela Superior de Ingeniería at the Universidad de Cádiz. Although I joined online, it was a great experience to share ideas and connect with other PhD students and researchers working on diverse areas of computer science.
The work I presented is titled: “Towards Recognizing Human Action in Videos using Spiking Neural Networks”. Co-authored with Elisa Guerrero Vázquez, María de la Paz Guerrero Lebrero, and Hayat Yedjour.
Recognizing human actions in video is a fascinating yet tricky task. Videos come with all sorts of challenges — different lighting conditions, moving cameras, cluttered backgrounds, and variation in how people move. While deep learning has made great progress in this area, it's still very resource-hungry, especially when it comes to processing video data in real-time or on low-power devices.
That’s where Spiking Neural Networks (SNNs) come in. These are bio-inspired models that mimic how neurons in the brain fire sparsely and only when needed — making them potentially much more energy-efficient than traditional deep networks.
In our approach, we tried to combine the best of both worlds:
- We first use a lightweight deep model to convert RGB video frames into event-based streams (more compact and time-aware representations).
- Then, we feed these into a spiking neural network trained using surrogate gradients — a training trick that helps SNNs learn more effectively.
The goal is to reduce redundancy, focus on meaningful temporal features, and ultimately recognize actions with much less energy consumption than typical methods.
This kind of work opens doors for bringing human activity recognition into real-world settings where resources are limited — like wearable devices, mobile robots, or smart home systems. It’s also part of a broader movement in AI research to explore neuromorphic and energy-aware computing — something I’m personally very interested in.