Speech Transformer Tf2.0 Save

transformer for ASR-systerm (via tensorflow2.0)

Project README

ASR-Transformer-plus-2DAttention

A tensorflow2.0 implementation of Speech Transformer [1], an end-to-end automatic speech recognition with Transformer [4] network, which directly converts acoustic features to character sequence using a single nueral network. ps: 2DAttention used for feature extraction ,more details can be found in paper[1]

Install

  • Python3 (recommend Anaconda)
  • Tensorflow 2.0 alpha

Dataset

  • Librispeech: LibriSpeech is a corpus of approximately 1000 hours of 16kHz read English speech, prepared by Vassil Panayotov with the assistance of Daniel Povey. The data is derived from read audiobooks from the LibriVox project, and has been carefully segmented and aligned.
  • download Librispeech dataset for free

Model

  • model_architecture

model_architecture

  • input: The input acoustic features were 80-dimensional filterbanks extracted with a hop size of 10ms and a window size of 25ms, extended with temporal first and second order differences and per-speaker mean subtraction and variance normalization.
  • output: The output alphabet of target text consisted of 31 classes, 26 lowercase letters, apostrophe, space, unknown tokens, start-of-sequence tokens and end-of-sequence tokens.
  • 2D-attention: described in [1]

2D-ATT

  • pre_net: described in [1]

pre_nET

Reference

  • [1] Linhao Dong, Shuang Xu,and Bo Xu. “Speech-transformer:A no-recurrence sequence-to-sequence model for speech recognition” in ICASSP 2018
  • [2] Transformer model for language understanding. See transformer tutorial
  • [3] A PyTorch implementation of Speech Transformer. See speech transformer pytorch
  • [4] Ashish Vaswani, Noam Shazeer, et al. “Attention is all you need” in NIPS 2017
Open Source Agenda is not affiliated with "Speech Transformer Tf2.0" Project. README Source: xingchensong/Speech-Transformer-tf2.0
Stars
112
Open Issues
0
Last Commit
5 years ago

Open Source Agenda Badge

Open Source Agenda Rating