VQA Challenge
VQA Challenge is available at this link.
In MMF, we provide the starter code for various baseline models for this challenge. VQA2.0 dataset will also be automatically downloaded during first training.
In this tutorial, we provide steps for running training and evaluation with VisualBERT model on VQA2.0 dataset and generating submission file for the challenge. The same steps can be used for other models.
#
InstallationFollow the prerequisites for installation of mmf here.
#
Training and Evaluation#
TrainingFor running training on train set, run the following command:
This will train the visual_bert
model on the dataset and generate the checkpoints and best trained model (visual_bert_final.pth
) will be stored in an experiment folder under the ./save
directory by default.
#
EvaluationNext run evaluation on the validation set:
This will give you the performance of your model on the validation set. The metric will be VQA Accuracy.
#
Predictions for ChallengeAfter training the model and evaluated on the validation set, we will generate the predictions on the test-dev
and test-std
set. The prediction file should contain the following for each sample:
- Question ID,
question_id
- Answer,
answer
With MMF you can directly generate the predictions in the required submission format with the following command:
This command will output where the generated predictions JSON file is stored.
#
Submission for ChallengeNext you can upload the generated json file to EvalAI page for VQA here. To check your results, you can go in 'My submissions' section and check the phase where you submitted your results file.