-
Notifications
You must be signed in to change notification settings - Fork 1.6k
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Showing
8 changed files
with
35,840 additions
and
0 deletions.
There are no files selected for viewing
Binary file not shown.
22,864 changes: 22,864 additions & 0 deletions
22,864
deep-learning/Transformer-Tutorials/ViLT/Fine_tuning_ViLT_for_VQA.ipynb
Large diffs are not rendered by default.
Oops, something went wrong.
1,051 changes: 1,051 additions & 0 deletions
1,051
...learning/Transformer-Tutorials/ViLT/Inference_with_ViLT_(visual_question_answering).ipynb
Large diffs are not rendered by default.
Oops, something went wrong.
3,196 changes: 3,196 additions & 0 deletions
3,196
deep-learning/Transformer-Tutorials/ViLT/Masked_language_modeling_with_ViLT.ipynb
Large diffs are not rendered by default.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,10 @@ | ||
# ViLT notebooks | ||
In this directory, you can find several notebooks that illustrate how to use NAVER AI Lab's [ViLT](https://arxiv.org/abs/2102.03334) both for fine-tuning on custom data as well as inference. It currently includes the following notebooks: | ||
|
||
- fine-tuning ViLT for visual question answering (VQA) (based on the [VQAv2 dataset](https://visualqa.org/)) | ||
- performing inference with ViLT to illustrate visual question answering (VQA) | ||
- masked language modeling (MLM) with a pre-trained ViLT model | ||
- performing inference with ViLT for image-text retrieval | ||
- performing inference with ViLT to illustrate natural language for visual reasoning (based on the [NLVRv2 dataset](https://lil.nlp.cornell.edu/nlvr/)). | ||
|
||
All models can be found on the [hub](https://huggingface.co/models?search=vilt). |
Oops, something went wrong.