Tech

How causal models fix offline reinforcement learning’s generalization problem

Share
Share
Breaking the spurious link: How causal models fix offline RL's generalization problem
The heat map of the three offline data sets in the car driving model. Credit: Frontiers of Computer Science (2024). DOI: 10.1007/s11704-024-3946-y

Researchers from Nanjing University and Carnegie Mellon University have introduced an AI approach that improves how machines learn from past data—a process known as offline reinforcement learning. This type of machine learning is essential for allowing systems to make decisions using only historical information without needing real-time interaction with the world.

By focusing on the authentic cause-and-effect relationships within the data, the new method enables autonomous systems—like driverless cars and medical decision-support systems—to make safer and more reliable choices. The work is published in the journal Frontiers of Computer Science.

From misleading signals to true causality: A new learning paradigm

Traditionally, offline reinforcement learning has struggled because it sometimes picks up misleading patterns from biased historical data. To illustrate, imagine learning how to drive by only watching videos of someone else behind the wheel.

If that driver always turns on the windshield wipers when slowing down in the rain, you might incorrectly think that turning on the wipers causes the car to slow down. In reality, it is the act of braking that slows the vehicle.

The new AI method corrects this misunderstanding by teaching the system to recognize that the braking action, not the activation of the windshield wipers, is responsible for slowing the car.

Enhancing safety in autonomous systems

With the ability to identify genuine cause-and-effect relationships, the new approach makes autonomous systems much safer, smarter, and more dependable. Industries such as autonomous vehicles, health care, and robotics benefit significantly because these systems are often used when precise and trustworthy decision-making is critical.

Lead researcher Prof. Yang Yu stated, “Our study harnesses the power of causal reasoning to cut through the noise in historical data, enabling systems to make decisions that are both more accurate and safer—an advancement that could improve how autonomous technology is deployed across industries.”

For policymakers and industry leaders, these findings could support improved regulatory standards, safer deployment practices, and increased public trust in automated systems. Additionally, from a scientific perspective, the research paves the way for more robust studies on AI awareness of causality.

A causal approach that outperforms traditional models

The researchers found that traditional AI models sometimes mistake unrelated actions as causally linked, which can result in dangerous outcomes. They demonstrated that many of these errors are significantly reduced by incorporating causal structure into these models. Moreover, the new method—referred to as a new causal AI approach—has been shown to perform consistently better than existing techniques (i.e., MOPO, MOReL, COMBO, LNCM) when tested in practical scenarios.

To achieve these promising results, the research team developed a method that identifies genuine causal relationships from historical data using specialized statistical tests designed for sequential and continuous data. This approach helps accurately discern the true causes behind observed actions and reduces the computational complexity that often hampers traditional methods, making the system more efficient and practical.

This research enhances our understanding of AI capabilities by embedding causal reasoning into offline reinforcement learning. It offers practical improvements in the safety and effectiveness of autonomous systems in everyday life.

More information:
Zhengmao Zhu et al, Offline model-based reinforcement learning with causal structured world models, Frontiers of Computer Science (2024). DOI: 10.1007/s11704-024-3946-y

Provided by
Higher Education Press

Citation:
Breaking the spurious link: How causal models fix offline reinforcement learning’s generalization problem (2025, April 28)
retrieved 28 April 2025
from

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.

Share

Leave a comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Related Articles
Could the ‘Angry Magpie’ save your business from insider threat and data-related attacks?
Tech

Could the ‘Angry Magpie’ save your business from insider threat and data-related attacks?

Browsers are the new frontline, but today’s DLP can’t see the real...

Smart surfaces could represent a powerless solution to multipath signal interference
Tech

Smart surfaces could represent a powerless solution to multipath signal interference

This study demonstrates a passive metasurface technology that uses a time-varying mechanism...

Dual scalable annealing processors overcome capacity and precision limits
Tech

Dual scalable annealing processors overcome capacity and precision limits

The proposed system enables simultaneous expansion of the number of spins and...

All-organic solar cells achieve record efficiency by doubling previous performance
Tech

All-organic solar cells achieve record efficiency by doubling previous performance

Example of damage to the lower layer of a solar cell disposal...