BlendRL: A Framework for Merging Symbolic and Neural Policy Learning

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

Abstract

Humans can leverage both abstract reasoning and intuitive reactions. In contrast, reinforcement learning policies are typically encoded in either opaque systems like neural networks or symbolic systems that rely on predefined symbols and rules. This disjointed approach severely limits the agents' capabilities, as they often lack either the flexible low-level reaction characteristic of neural agents or the interpretable reasoning of symbolic agents. To overcome this challenge, we introduce BlendRL, a neuro-symbolic RL framework that harmoniously integrates both paradigms within RL agents that use mixtures of both logic and neural policies. We empirically demonstrate that BlendRL agents outperform both neural and symbolic baselines in standard Atari environments, and showcase their robustness to environmental changes. Additionally, we analyze the interaction between neural and symbolic policies, illustrating how their hybrid use helps agents overcome each other's limitations.

Original languageEnglish
Title of host publication13th International Conference on Learning Representations, ICLR 2025
PublisherCurran Associates
Pages10518-10549
Number of pages32
ISBN (Electronic)9798331320850
Publication statusPublished - 2025
Event13th international Conference on Learning Representations, ICLR 2025 - Singapore, Singapore
Duration: 24 Apr 202528 Apr 2025

Conference

Conference13th international Conference on Learning Representations, ICLR 2025
Abbreviated titleICLR 2025
Country/TerritorySingapore
CitySingapore
Period24/04/2528/04/25

Bibliographical note

Publisher Copyright:
© 2025 13th International Conference on Learning Representations, ICLR 2025. All rights reserved.

Fingerprint

Dive into the research topics of 'BlendRL: A Framework for Merging Symbolic and Neural Policy Learning'. Together they form a unique fingerprint.

Cite this