Skip to content

From scratch implementation of a Neural Machine Translation Model (including attention and beam search) in PyTorch – for the Neural Network practical @ KIT.

Notifications You must be signed in to change notification settings

retkowski/NeuralMachineTranslation

Repository files navigation

Attentional Neural Machine Translation

Scope

Standard attentional-based NMT framework (Bahdanau et al. 2014)

  • Data Preparer
  • (Bidirectional) GRU/LSTM-based encoder
  • Global Attention Layers: MLP/Dot Product (Luong et al. 2015)
  • GRU/LSTM-based decoder
  • Beam Search
  • Teacher Forcing

Requirements

  • Data Preparation: Convert words to index prior to one-hot encoding for training, validation and test sets.
  • Training: Train the NMT architecture.
  • Decoding/Testing: Realize the beam search to translate sentences from German to English.

Evaluation

Perplexity and BLEU scores on validation/test sets.

Papers

BLEU calculation

We used for the BLEU calculation the multi_bleu.pl script from the moses project: https://github.com/moses-smt/mosesdecoder/blob/master/scripts/generic/multi-bleu.perl

About

From scratch implementation of a Neural Machine Translation Model (including attention and beam search) in PyTorch – for the Neural Network practical @ KIT.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •