diff --git a/languages/master-language/model.py b/languages/master-language/model.py deleted file mode 100644 index a2f649e..0000000 --- a/languages/master-language/model.py +++ /dev/null @@ -1,20 +0,0 @@ -import tensorflow as tf -import opennmt as onmt -class SoftcatalaModel(onmt.models.Transformer): - """Defines a base Transformer model using relative position representations as - described in https://arxiv.org/abs/1803.02155. - """ - - def __init__(self): - super().__init__(share_embeddings=onmt.models.sequence_to_sequence.EmbeddingsSharingLevel.AUTO) - -model = SoftcatalaModel - -# TransformerRelative: *** model cfg: num_units: 512, num_heads: 8, ffn_inner_dim: 2048, maximum_relative_position: 20 -# -# num_units: The number of hidden units. -# num_heads: The number of heads in each self-attention layers. -# ffn_inner_dim: The inner dimension of the feed forward layers. -# maximum_relative_position: Maximum relative position representation -# Where the input / output dimensions are much greater than the hidden input dimension. - diff --git a/languages/master-language/train.sh b/languages/master-language/train.sh index 657ba8a..63ef197 100755 --- a/languages/master-language/train.sh +++ b/languages/master-language/train.sh @@ -1,2 +1,2 @@ #--num_gpus 4 for -> multiple gpus -onmt-main --mixed_precision --model model.py --config data.yml --auto_config train --with_eval +onmt-main --mixed_precision --model_type TransformerBigRelative --config data.yml --auto_config train --with_eval