Skip to content

Pytorch Implementation of Knowing When to Look: Adaptive Attention via A Visual Sentinel for Image Captioning

Notifications You must be signed in to change notification settings

yufengm/Adaptive

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

AdaptiveAttention

Pytorch Implementation of Adaptive Attention Model for Image Captioning

Knowing When to Look: Adaptive Attention via A Visual Sentinel for Image Captioning [Paper] [Review]

Dataset Preparation

First we will need to download the MS-COCO dataset. So create a data folder and run the download bash script

mkdir data && ./download.sh

Afterwards, we should create the Karpathy split for training, validation and test.

python KarpathySplit.py

Then we can build the vocabulary by running

python build_vocab.py

The vocab.pkl should be saved in the data folder.

Now we will need to resize all the images in both train and val folder. Here I create a new folder under data, i.e., 'resized'. Then we may run resize.py to resize all images into 256 x 256. You may specify different locations inside resize.py

mkdir data/resized && python resize.py

After all images are resized. Now we can train our Adaptive Attention model with

python train.py

About

Pytorch Implementation of Knowing When to Look: Adaptive Attention via A Visual Sentinel for Image Captioning

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published