Skip to content
/ SHIRE Public

This repository contains code associated with SHIRE: Enhancing Sample Efficiency using Human Intuition in Reinforcement Learning (ICRA 2025)

License

Notifications You must be signed in to change notification settings

amoghj98/SHIRE

Repository files navigation

SHIRE: Enhancing Sample Efficiency using Human Intuition in Reinforcement Learning

This repository contains code associated with SHIRE: Enhancing Sample Efficiency using Human Intuition in Reinforcement Learning This code has most recently been tested with Python 3.10 and PyTorch 2.5.1

Introduction

The ability of neural networks to perform robotic perception and control tasks such as depth and optical flow estimation, simultaneous localization and mapping (SLAM), and automatic control has led to their widespread adoption in recent years. Deep Reinforcement Learning has been used extensively in these settings, as it does not have the unsustainable training costs associated with supervised learning. However, DeepRL suffers from poor sample efficiency, i.e., it requires a large number of environmental interactions to converge to an acceptable solution. Modern RL algorithms such as Deep Q Learning and Soft Actor-Critic attempt to remedy this shortcoming but can not provide the explainability required in applications such as autonomous robotics. Humans intuitively understand the long-time-horizon sequential tasks common in robotics. Properly using such intuition can make RL policies more explainable while enhancing their sample efficiency. In this work, we propose SHIRE, a novel framework for encoding human intuition using Probabilistic Graphical Models (PGMs) and using it in the Deep RL training pipeline to enhance sample efficiency. Our framework achieves 25-78% sample efficiency gains across the environments we evaluate at negligible overhead cost. Additionally, by teaching RL agents the encoded elementary behavior, SHIRE enhances policy explainability.

Installation

Clone this repository using:
git clone [email protected]:amoghj98/SHIRE.git

Create a conda environment using the provided yaml file as follows:

conda create -n intuitiveRL -f intuitiveRL.yml

Activate the environment using:

conda activate $ENV_NAME

Training RL Agents

RL agents can be trained using the run.bash file as follows, providing a convenient encapsulation of the otherwise cumbersome task of using the large number of available cmd-line args.:
bash run.bash

This script internally calls the script intuitiveTrain.py, which supports a large number of command-line arguments. These args are documented in the file "argConfig.py"

Citations

If you find this work useful in your research, please consider citing:

A. Joshi, A. Kosta and K. Roy, "SHIRE: Enhancing Sample Efficiency using Human Intuition in Reinforcement Learning", In Proceedings of the 2025 IEEE International Conference on Robotics and Automation (ICRA), Atlanta, USA, 2025

@inproceedings{joshi2025shire,
  author={Joshi, Amogh and Kosta, Adarsh and Roy, Kaushik},
  booktitle={2025 IEEE International Conference on Robotics and Automation (ICRA)},
  title={SHIRE: Enhancing Sample Efficiency using Human Intuition in REinforcement Learning},
  pages={},
  year={2025},
  organization={IEEE}
}

An arXiv version is also available: Amogh Joshi, Adarsh Kosta and Kaushik Roy, "SHIRE: Enhancing Sample Efficiency using Human Intuition in REinforcement Learning", arXiv preprint, 2024

@misc{joshi2025shireenhancingsampleefficiency,
    title={SHIRE: Enhancing Sample Efficiency using Human Intuition in REinforcement Learning}, 
    author={Amogh Joshi and Adarsh Kumar Kosta and Kaushik Roy},
    year={2025},
    eprint={2409.09990},
    archivePrefix={arXiv},
    primaryClass={cs.LG},
    url={https://arxiv.org/abs/2409.09990}, 
}

Authors

Amogh Joshi, Adarsh Kosta and Kaushik Roy

About

This repository contains code associated with SHIRE: Enhancing Sample Efficiency using Human Intuition in Reinforcement Learning (ICRA 2025)

Topics

Resources

License

Stars

Watchers

Forks

Packages

No packages published