From 8aa20927498e1eca91257b2a1d85f08228adf98e Mon Sep 17 00:00:00 2001 From: Marco Date: Mon, 2 May 2022 08:00:29 -0700 Subject: [PATCH] fix two small typos --- the_annotated_transformer.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/the_annotated_transformer.py b/the_annotated_transformer.py index 4aa1d46..1abb7cb 100644 --- a/the_annotated_transformer.py +++ b/the_annotated_transformer.py @@ -199,7 +199,7 @@ def step(self): # comprehension, abstractive summarization, textual entailment and # learning task-independent sentence representations. End-to-end # memory networks are based on a recurrent attention mechanism instead -# of sequencealigned recurrence and have been shown to perform well on +# of sequence-aligned recurrence and have been shown to perform well on # simple-language question answering and language modeling tasks. # # To the best of our knowledge, however, the Transformer is the first @@ -1852,7 +1852,7 @@ def average(model, models): # # > With the addtional extensions in the last section, the OpenNMT-py # > replication gets to 26.9 on EN-DE WMT. Here I have loaded in those -# > parameters to our reimplemenation. +# > parameters to our reimplementation. # %% # Load data and model for output checks