Fall 2018 term, arxiv submission in progress.
Modified RNN based Seq2Seq Architecture (i.e Tacotron) to reduce model size and help speed up training, the following is implemented:
- Self attention (Similar to transformer paper)
- Guided Attention & Forced incremental Attention (From DCTTS paper)
- grapheme & Phoneme mixing input
Listen to audio samples here.