torchaudio/models/__init__.py (51 lines of code) (raw):

from .conformer import Conformer from .conv_tasnet import ConvTasNet from .deepspeech import DeepSpeech from .emformer import Emformer from .rnnt import RNNT, emformer_rnnt_base, emformer_rnnt_model from .rnnt_decoder import Hypothesis, RNNTBeamSearch from .tacotron2 import Tacotron2 from .wav2letter import Wav2Letter from .wav2vec2 import ( Wav2Vec2Model, HuBERTPretrainModel, wav2vec2_model, wav2vec2_base, wav2vec2_large, wav2vec2_large_lv60k, hubert_base, hubert_large, hubert_xlarge, hubert_pretrain_model, hubert_pretrain_base, hubert_pretrain_large, hubert_pretrain_xlarge, ) from .wavernn import WaveRNN __all__ = [ "Wav2Letter", "WaveRNN", "ConvTasNet", "DeepSpeech", "Wav2Vec2Model", "HuBERTPretrainModel", "wav2vec2_model", "wav2vec2_base", "wav2vec2_large", "wav2vec2_large_lv60k", "hubert_base", "hubert_large", "hubert_xlarge", "hubert_pretrain_model", "hubert_pretrain_base", "hubert_pretrain_large", "hubert_pretrain_xlarge", "Tacotron2", "Conformer", "Emformer", "Hypothesis", "RNNT", "RNNTBeamSearch", "emformer_rnnt_base", "emformer_rnnt_model", ]