AIResearch AIResearch
Back to articles
Robotics

AI Predicts Human Actions from First-Person Video

A new method uses egocentric video to anticipate future activities, improving human-robot interaction and setting records in AI challenges.

AI Research
November 14, 2025
4 min read
AI Predicts Human Actions from First-Person Video

Imagine a robot that can predict your next move in the kitchen before you even make it, allowing for seamless collaboration in tasks like cooking or elder care. This vision is closer to reality thanks to research from the University of Maryland, which introduces a novel AI system called Egocentric Object Manipulation Graphs (Ego-OMG). Designed to model and anticipate human actions from first-person video, this approach achieved top performance in the EPIC-Kitchens Action Anticipation Challenge, outperforming previous methods by large margins. Its success lies in combining multiple levels of activity understanding—appearance, dynamics, and semantic structure—into a single framework that could transform how machines interact with people in everyday settings.

The key finding of this research is that Ego-OMG can accurately predict future actions in manipulation activities, such as those in kitchen environments, by analyzing video segments up to 60 seconds long. Specifically, it anticipates the next most likely action after the observation period ends, using a graph-based representation that captures how hands interact with objects over time. In tests, this method ranked first on the Unseen test set and second on the Seen test set of the EPIC-Kitchens challenge, with Top-1 accuracy scores of 16.02% and 11.80%, respectively, showing significant improvements over earlier approaches like RULSTM and TSN.

To achieve this, the researchers developed a methodology that integrates two main streams of data processing. The first stream uses a Channel-Separated Convolutional Network (CSN) to analyze short-term dynamics from 32 consecutive video frames, focusing on appearance and motion features similar to how one might track quick movements in a sports replay. This network was pre-trained on a large dataset and processes inputs with techniques like horizontal flipping and color jittering to enhance robustness. The second stream constructs a graph representation of the activity, where nodes represent objects and hands, and edges denote interactions, such as when a hand makes or breaks contact with an object. This graph is embedded using a Graph Convolutional Network (GCN) to create vector representations that capture long-term semantic relationships, akin to mapping out a story's plot points to guess what happens next. These representations are then fed into an LSTM (a type of recurrent neural network) to model the sequence of states and predict future actions, with the outputs from both streams combined via a late fusion approach for final classification.

The results analysis reveals that Ego-OMG's performance degrades gracefully as the anticipation time increases, maintaining an advantage over methods relying solely on short-term features. For instance, in ablation studies, the joint use of GCN and CSN streams consistently outperformed individual components, with the GCN stream excelling in longer-term predictions—achieving higher accuracy than CSN alone when anticipating actions five seconds ahead. The researchers also found that using an LSTM to aggregate state sequences provided better results than alternatives like mean pooling, emphasizing the importance of temporal order. Additionally, initializing node features with pre-trained word embeddings (GloVe) boosted accuracy, as shown in comparisons where identity matrix initialization led to significantly lower scores, such as dropping from 12.81% to 6.67% in one test.

In terms of real-world context, this research matters because it addresses the growing need for AI systems that can anticipate human behavior in practical scenarios. For example, in human-robot interaction, better action prediction could enable robots to assist in cooperative tasks—like helping an elderly person cook—by providing earlier feedback and reducing cognitive load. The use of egocentric video, which offers a consistent perspective with hands clearly visible, makes this approach particularly suited for applications in surveillance, navigation, and automated assistance, where understanding relational activities is crucial. By focusing on manipulation tasks, the method taps into a domain where human agents drive meaningful change, potentially leading to more intuitive and responsive robotic partners.

However, the study acknowledges limitations, including biases in the EPIC-Kitchens dataset, such as its focus on subjects from specific socioeconomic backgrounds and regions in North America and Europe, which may not generalize to diverse global contexts. The method also struggles with scenarios involving multiple objects held simultaneously, as it currently predicts only up to two objects per hand, limiting its applicability in highly dexterous tasks. Furthermore, the object classifier tends to misclassify held items due to occlusion by hands, especially with small objects like utensils, though the researchers mitigated this by incorporating temporal constraints from past predictions.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn