You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Refer to #30 about the PR to add caption and vqa task.
Refer to #77 about the PR to merge the code for caption task and vqa task (all of the work was completed on add_vqa branch) to master
Now the merge has completed, we submit this issue to keep track of multiple issues that need to be investigated/resolved
Issue 1: the training loss and perplexity is not decreasing for the VQA task
For the vqa task, the training loss and perplexity is not decreasing with the tests we have done up to about 14K images and their associated VQA. The test was done on Colab and we could not scale up for more images due to Colab I/O error to handle too many files in a folder.
We need to investigate this issue from at least two avenues:
Scale up more training data on a hardware where I/O is not causing error for more images in a folder. This is for purpose to see whether the non-decreasing perplexity is due to training dataset is too small
Inspect the algorithm for VQA and see whether it should be revised
Issue 2: need to implement a evaluate for caption task and vqa task in eval.py
Any other issue will be entered here as they show up
The text was updated successfully, but these errors were encountered:
Refer to #30 about the PR to add caption and vqa task.
Refer to #77 about the PR to merge the code for caption task and vqa task (all of the work was completed on add_vqa branch) to master
Now the merge has completed, we submit this issue to keep track of multiple issues that need to be investigated/resolved
Issue 1: the training loss and perplexity is not decreasing for the VQA task
For the vqa task, the training loss and perplexity is not decreasing with the tests we have done up to about 14K images and their associated VQA. The test was done on Colab and we could not scale up for more images due to Colab I/O error to handle too many files in a folder.
We need to investigate this issue from at least two avenues:
Issue 2: need to implement a evaluate for caption task and vqa task in eval.py
Any other issue will be entered here as they show up
The text was updated successfully, but these errors were encountered: