Skip to content

Jhdo/basic_vqa

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

75 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

basic_vqa

Pytorch implementation of the paper - VQA: Visual Question Answering (https://arxiv.org/pdf/1505.00468.pdf).

Usage

1. Clone the repositories

$ git clone https://github.com/tbmoon/basic_vqa.git

2. Download and unzip the dataset from official url of VQA: https://visualqa.org/download.html

$ cd basic_vqa/utils
$ chmod +x download_and_unzip_datasets.csh
$ ./download_and_unzip_datasets.csh

3. Preproccess input data for (images, questions and answers)

$ python resize_images.py --input_dir='../datasets/Images' --output_dir='../datasets/Resized_Images'  
$ python make_vacabs_for_questions_answers.py --input_dir='../datasets'
$ python build_vqa_inputs.py

4. Train model for VQA task

$ python train.py

Results

  • Loss and Accuracy on VQA datasets v2

    • Experiment 0: Prediction of '<UNK>' is accepted as Answer.

    • Experiment 1: Prediction of '<UNK>' is not accepted as Answer.

train1

References

About

Pytorch VQA : Visual Question Answering (https://arxiv.org/pdf/1505.00468.pdf)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 82.9%
  • Python 15.3%
  • Shell 1.8%