Kyran Flynn
← Back to Portfolio

Causal Structure-Informed Reinforcement Learning

Key Idea — Built graph neural network-based reinforcement learning models leveraging causal structure for improved data-efficiency in multi-intervention environments.

ProgramColumbia IEOR, advised by Prof. Lily Xu.

Tech StackPyTorch, PyG, GurobiPy, Gurobi-ML, PyWhy, NetworkX.

Overview

We experimented with graph neural network (GNN)–based reinforcement learning (RL) frameworks to improve model data-efficiency in data-limited environments with multiple intervention options. Standard RL algorithms often overlook the causal relationships between interventions and system states, so we focused on learning them and leveraging the resulting graph structure using GNNs. Our goal was to explore whether this approach could reduce data needs, and scale effectively to high-dimensional settings.

Implementation

Developed model architectures, synthetic problems, and testing framework from scratch

Each intervention (sub-action) and state variable is represented as a node in a bipartite graph

A graph neural network (GNN) propagates information between state and sub-action nodes to estimate Q-values jointly

The combinatorial optimization step for selecting the best intervention set is solved via Gurobi using the learned Q-value function

This approach enables scalable training and decision-making in environments with large binary sub-action spaces

Highlights

Integrated GNN message passing into a custom Q-learning loop for multi-action reinforcement learning.

Designed a modular PyTorch Geometric training pipeline with Gurobi-based optimization for policy selection.

Demonstrated improved learning stability and faster convergence on synthetic multi-intervention testbeds.

Applied causal inference tools from the PyWhy framework to improve interpretability of learned dependencies.