A PyTorch implementation of the Transformer model in "Attention is All Y...
A TensorFlow Implementation of the Transformer: Attention Is All You Need
Sequence-to-sequence framework with a focus on Neural Machine Translatio...
Pre-training of Deep Bidirectional Transformers for Language Understandi...
list of efficient attention modules
My implementation of the original transformer model (Vaswani et al.). I'...
A PyTorch implementation of Speech Transformer, an End-to-End ASR with T...
A Keras+TensorFlow Implementation of the Transformer: Attention Is All Y...
A Benchmark of Text Classification in PyTorch
Neural Machine Translation with Keras
A Pytorch Implementation of "Attention is All You Need" and "Weighted Tr...
Open-Source Toolkit for End-to-End Korean Automatic Speech Recognition l...
pytorch implementation of Attention is all you need
Implementation of Vaswani, Ashish, et al. "Attention is all you need." A...
A recurrent attention module consisting of an LSTM cell which can query ...