Skip to content

Toolkit for efficient experimentation with various sequence-to-sequence models

License

Notifications You must be signed in to change notification settings

borisgin/OpenSeq2Seq

This branch is 88 commits behind NVIDIA/OpenSeq2Seq:master.

Folders and files

NameName
Last commit message
Last commit date

Latest commit

f3624dc · Apr 17, 2019
Mar 25, 2019
Nov 29, 2018
Apr 13, 2018
Apr 10, 2019
Apr 11, 2019
Apr 17, 2019
Apr 15, 2019
Apr 9, 2019
Dec 10, 2018
Sep 23, 2018
Jul 3, 2018
Apr 18, 2018
Oct 2, 2018
Sep 23, 2018
Oct 2, 2018
Dec 11, 2018
Apr 9, 2019
Apr 1, 2019
Jan 30, 2019

Repository files navigation

License Documentation

OpenSeq2Seq

OpenSeq2Seq: toolkit for distributed and mixed precision training of sequence-to-sequence models

OpenSeq2Seq main goal is to allow researchers to most effectively explore various sequence-to-sequence models. The efficiency is achieved by fully supporting distributed and mixed-precision training. OpenSeq2Seq is built using TensorFlow and provides all the necessary building blocks for training encoder-decoder models for neural machine translation, automatic speech recognition, speech synthesis, and language modeling.

Documentation and installation instructions

https://nvidia.github.io/OpenSeq2Seq/

Features

  1. Models for:
    1. Neural Machine Translation
    2. Automatic Speech Recognition
    3. Speech Synthesis
    4. Language Modeling
    5. NLP tasks (sentiment analysis)
  2. Data-parallel distributed training
    1. Multi-GPU
    2. Multi-node
  3. Mixed precision training for NVIDIA Volta/Turing GPUs

Software Requirements

  1. Python >= 3.5
  2. TensorFlow >= 1.10
  3. CUDA >= 9.0, cuDNN >= 7.0
  4. Horovod >= 0.13 (using Horovod is not required, but is highly recommended for multi-GPU setup)

Acknowledgments

Speech-to-text workflow uses some parts of Mozilla DeepSpeech project.

Text-to-text workflow uses some functions from Tensor2Tensor and Neural Machine Translation (seq2seq) Tutorial.

Disclaimer

This is a research project, not an official NVIDIA product.

Related resources

Paper

If you use OpenSeq2Seq, please cite this paper

@misc{openseq2seq,
    title={Mixed-Precision Training for NLP and Speech Recognition with OpenSeq2Seq},
    author={Oleksii Kuchaiev and Boris Ginsburg and Igor Gitman and Vitaly Lavrukhin and Jason Li and Huyen Nguyen and Carl Case and Paulius Micikevicius},
    year={2018},
    eprint={1805.10387},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}

About

Toolkit for efficient experimentation with various sequence-to-sequence models

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 93.8%
  • C++ 3.2%
  • Shell 2.1%
  • Other 0.9%