Neural-Machine-Translation-with-Attention
PublicI implement encoder-decoder based seq2seq models with attention using Keras. The encoder can be a Bidirectional LSTM, a simple LSTM, or a GRU, and the decoder can be an LSTM or a GRU. I evaluate the models on an English-French dataset.