AIResearch AIResearch
Back to articles
Science

AI Learns Cause and Effect from Human Knowledge

AI learns cause and effect from human knowledge, making scientific discoveries more accurate by incorporating established facts automatically.

AI Research
November 14, 2025
3 min read
AI Learns Cause and Effect from Human Knowledge

Understanding cause and effect is fundamental to science, from medicine to economics, but traditional methods often rely on costly experiments or struggle with noisy data. A new study introduces 'interventional constraints,' a way for artificial intelligence to incorporate high-level human knowledge about causal relationships, improving accuracy and reliability in discovering how variables influence each other. This approach ensures AI models align with established facts, such as the known positive effect of PIP3 on Akt in cell signaling, without needing extensive experimental data.

Researchers developed a method that enforces inequality constraints on the total causal effect between variables, ensuring the direction and sign of influences match prior knowledge. For example, if biological evidence indicates PIP3 activates Akt, the AI model is constrained to reflect a positive total effect, preventing incorrect conclusions like inhibition. This differs from earlier techniques that only shaped the model's structure without regulating effect strengths, bridging a gap in causal discovery.

The methodology uses a two-stage optimization process. First, a gradient-based algorithm learns an initial causal model that satisfies acyclicity—ensuring no cycles in the causal graph. Then, a sequential quadratic programming method refines the model to incorporate interventional constraints, adjusting weights to meet inequality conditions on total effects. This approach handles the non-convex nature of the problem, where traditional optimizers fall short, by iteratively solving subproblems to converge on a solution that respects both data and knowledge.

Experiments on synthetic and real-world datasets, including the Sachs dataset on immune cell signaling, demonstrate significant improvements. With interventional constraints, models showed higher accuracy in recovering true causal structures, with metrics like false discovery rate and structural intervention distance improving by up to 50% in some cases. For instance, in tests with 20 variables and small sample sizes, the method correctly identified additional interactions, such as PKA's inhibition of P38, which were missed by constraint-free approaches. The sign consistency of effects also increased, ensuring models not only find connections but also their correct directional influence.

This advancement matters because it makes AI-driven causal discovery more trustworthy and applicable in real-world scenarios. In fields like healthcare, where ethical and cost limitations restrict experiments, integrating expert knowledge can lead to better treatment insights without compromising privacy. Economists could use it to model tax impacts on spending more reliably, leveraging existing evidence without full datasets. By reducing reliance on large-scale data, the method offers a scalable path for domains with limited resources.

Limitations include scalability challenges due to computational intensity, as the optimization becomes demanding with many constraints. The paper notes that the approach is currently tailored to linear models, and extending it to nonlinear settings or systems with hidden confounders remains for future work. Additionally, the method's performance can vary with the choice of constraint thresholds, requiring careful tuning to avoid underfitting or overcorrection in complex networks.

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn