Skip to content

Monishver11/Gaze-RL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

57 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Gaze-RL

Implementation for "Deep Decision Making with RL" course project titled "Gaze-Guided Reinforcement Learning for Visual Search"

Visual search is a fundamental task that humans perform effortlessly, but it remains challenging for autonomous agents. This project explores how human visual attention patterns can be leveraged to improve reinforcement learning for object search tasks in the AI2-THOR environment.

Our approach combines:

  • A pretrained gaze prediction model
  • Three integration methods for incorporating gaze information:
    • Channel integration
    • Bottleneck integration
    • Weighted integration
  • A dual-level integration framework that utilises gaze for both perception and reward shaping

Setup and Installation

#Clone the repository
git clone https://github.com/Monishver11/Gaze-RL.git
cd Gaze-RL

#Create a conda environment
conda env create -f environment.yml
conda activate gaze-rl-stable

#Install dependencies
pip install -r requirements.txt

Usage

To train an agent with gaze integration:

cd Gaze-RL/
python src/train_gaze_guided_rl_final.py --target Microwave --integration channel --exp-name gaze_channel_exp

For a detailed explanation of this project, including motivation, methodology, and comprehensive results, visit our blog post.

About

Discover how gaze prediction from human eye-tracking enhances AI agents in object search tasks. By integrating visual attention into reinforcement learning through three novel methods, our approach enables faster, more effective navigation in simulated environments.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors