Skip to content

Implementation of Predictive Coding Enhances Meta-RL To Achieve Interpretable Bayes-Optimal Belief Representation Under Partial Observability. Kuo et al. (NeurIPS 2025)

License

Notifications You must be signed in to change notification settings

walkerlab/metaRL-predictive-representation

Repository files navigation

metaRL-predictive-representation

Code for the paper: "Predictive Coding Enhances Meta-RL To Achieve Interpretable Bayes-Optimal Belief Representation Under Partial Observability" Po-Chen Kuo, Han Hou, Will Dabney, Edgar Y. Walker. Published in NeurIPS 2025. (arXiv: https://arxiv.org/abs/2510.22039)

@inproceedings{kuo2025predictive,
  title={Predictive Coding Enhances Meta-RL To Achieve Interpretable Bayes-Optimal Belief Representation Under Partial Observability},
  author={Kuo, Po-Chen and Hou, Han and Dabney, Will and Walker, Edgar Y.},
  booktitle={Neural Information Processing Systems (NeurIPS)},
  year={2025}

Overview

The entry point for model training is in train.py, where args and configs are parsed to initialize and train the models. The meta-RL models, including both black-box meta-RL and meta-RL with predictive modeuls, are defined in metalearner.py. State machine simulation analysis, including comments on how to use the script, is demonstrated in sms.py. The main requirements are listed in requirements.txt.

Model training

To train meta-RL with predictive modules on a task (e.g. OracleBandit), use:

python train.py --exp_label=mpc --env_name=OracleBanditDeterministic-v0

To train black-box meta-RL (e.g. RL2) on a task (e.g. OracleBandit), use:

python train.py --exp_label=rl2 --env_name=OracleBanditDeterministic-v0

You can find other tasks defined in the environments/ folder. The trained models and checkpoints will be saved at ./logs by default. The default configs are in the config/ folder. Hyperparameters can be overwritten command line arguments.

State machine simulation

To evalute representational equivalence on trained models, modify the sms.py script to add the paths to the trained models and adjust parameters to the tasks you're evaluating on accordingly.

About

Implementation of Predictive Coding Enhances Meta-RL To Achieve Interpretable Bayes-Optimal Belief Representation Under Partial Observability. Kuo et al. (NeurIPS 2025)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages