Artificial intelligence systems often struggle with understanding why things happen, not just what happens. This limitation becomes critical in applications like medical diagnosis or autonomous vehicles, where knowing the cause of an outcome matters as much as predicting the outcome itself. A new approach developed by researchers at Mexico's National Institute of Astrophysics, Optics and Electronics gives AI the ability to learn causal relationships through direct interaction with its environment, much like humans do.
The key finding shows that AI can simultaneously learn how the world works while making optimal decisions. The researchers demonstrated that their method successfully identifies which variables cause which effects across different scenarios, from medical treatment decisions to controlling complex lighting systems. Unlike traditional machine learning that only finds statistical patterns, this approach captures the actual causal structure underlying observed phenomena.
The methodology uses Bayesian random graphs to represent uncertainty about causal relationships. Think of it as the AI maintaining a mental map of possible connections between different elements in its environment, where each potential cause-effect link has a probability score representing how likely it exists. As the AI interacts with its environment—like a doctor trying different treatments or a robot flipping switches—it updates these probability scores based on what actually happens. The system doesn't assume it knows anything about the underlying causal structure beforehand; instead, it learns this structure through trial and observation.
In the disease treatment experiment, the AI played the role of a doctor deciding between treatment A, treatment B, or both for a patient with possible diseases. The results showed that after multiple interactions, the AI correctly learned that treatment directly affects patient survival and causes adverse reactions, while diseases influence both survival and reaction likelihood. The system achieved performance similar to established Q-learning algorithms but with the added benefit of understanding why specific treatments worked or failed.
For real-world applications, this means AI systems could explain their decisions in critical domains. A self-driving car could understand not just that braking prevented an accident, but why braking was necessary. Medical AI could explain which treatment factors actually cause recovery versus those that merely correlate with it. The approach also allows knowledge transfer—once an AI learns causal relationships in one domain, that understanding can apply to similar problems elsewhere.
The main limitation is that the method requires the AI to actively intervene and experiment, which isn't always possible or safe in real-world scenarios. The researchers also note that learning becomes more challenging when multiple causes produce the same effect, creating ambiguity about what actually drives outcomes. Future work needs to address how quickly these causal relationships can be learned with minimal experimentation and how the approach scales to extremely complex environments with hundreds of interacting variables.
Original Source
Read the complete research paper
About the Author
Guilherme A.
Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.
Connect on LinkedIn