Skip to content

Commit

Permalink
Merge pull request #406 from sce-taid/master
Browse files Browse the repository at this point in the history
fixing typo
  • Loading branch information
takumiohym authored Oct 20, 2023
2 parents bf83faf + 073efda commit 7de35c3
Show file tree
Hide file tree
Showing 2 changed files with 6 additions and 6 deletions.
6 changes: 3 additions & 3 deletions notebooks/multi_modal/labs/image_captioning.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -571,9 +571,9 @@
"decoder_gru = GRU(...)\n",
"gru_output, gru_state = decoder_gru(embed_x)\n",
"\n",
"decoder_atention = Attention()\n",
"decoder_attention = Attention()\n",
"# TODO: Define the inputs to the Attention layer\n",
"context_vector = decoder_atention([...])\n",
"context_vector = decoder_attention([...])\n",
"\n",
"addition = Add()([gru_output, context_vector])\n",
"\n",
Expand Down Expand Up @@ -728,7 +728,7 @@
"gru_output, gru_state = decoder_gru(...)\n",
"\n",
"# Reuse other layers as well\n",
"context_vector = decoder_atention([...])\n",
"context_vector = decoder_attention([...])\n",
"addition_output = Add()([...])\n",
"layer_norm_output = layer_norm(...)\n",
"\n",
Expand Down
6 changes: 3 additions & 3 deletions notebooks/multi_modal/solutions/image_captioning.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -716,8 +716,8 @@
")\n",
"gru_output, gru_state = decoder_gru(embed_x)\n",
"\n",
"decoder_atention = Attention()\n",
"context_vector = decoder_atention([gru_output, encoder_output])\n",
"decoder_attention = Attention()\n",
"context_vector = decoder_attention([gru_output, encoder_output])\n",
"\n",
"addition = Add()([gru_output, context_vector])\n",
"\n",
Expand Down Expand Up @@ -927,7 +927,7 @@
"gru_output, gru_state = decoder_gru(embed_x, initial_state=gru_state_input)\n",
"\n",
"# Reuse other layers as well\n",
"context_vector = decoder_atention([gru_output, encoder_output])\n",
"context_vector = decoder_attention([gru_output, encoder_output])\n",
"addition_output = Add()([gru_output, context_vector])\n",
"layer_norm_output = layer_norm(addition_output)\n",
"\n",
Expand Down

0 comments on commit 7de35c3

Please sign in to comment.