Project Name | Stars | Downloads | Repos Using This | Packages Using This | Most Recent Commit | Total Releases | Latest Release | Open Issues | License | Language |
---|---|---|---|---|---|---|---|---|---|---|
Transformers | 112,668 | 64 | 1,869 | 2 hours ago | 114 | July 18, 2023 | 865 | apache-2.0 | Python | |
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. | ||||||||||
Deepspeech | 23,017 | 29 | 14 | 16 days ago | 100 | December 19, 2020 | 135 | mpl-2.0 | C++ | |
DeepSpeech is an open source embedded (offline, on-device) speech-to-text engine which can run in real time on devices ranging from a Raspberry Pi 4 to high power GPU servers. | ||||||||||
Deeplearningexamples | 11,463 | 15 days ago | 279 | Jupyter Notebook | ||||||
State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterprise-grade infrastructure. | ||||||||||
Deep Learning Drizzle | 10,767 | 9 months ago | 6 | HTML | ||||||
Drench yourself in Deep Learning, Reinforcement Learning, Machine Learning, Computer Vision, and NLP by learning from these exciting lectures!! | ||||||||||
Nemo | 7,999 | 2 | 8 | 4 hours ago | 69 | August 04, 2023 | 83 | apache-2.0 | Python | |
NeMo: a toolkit for conversational AI | ||||||||||
Espnet | 7,046 | 5 | 7 hours ago | 32 | August 03, 2023 | 226 | apache-2.0 | Python | ||
End-to-End Speech Processing Toolkit | ||||||||||
Asrt_speechrecognition | 6,942 | 11 days ago | 1 | October 23, 2020 | 101 | gpl-3.0 | Python | |||
A Deep-Learning-Based Chinese Speech Recognition System 基于深度学习的中文语音识别系统 | ||||||||||
Speechbrain | 6,627 | 2 hours ago | 148 | apache-2.0 | Python | |||||
A PyTorch-based Speech Toolkit | ||||||||||
Wav2letter | 6,287 | a month ago | 106 | other | C++ | |||||
Facebook AI Research's Automatic Speech Recognition Toolkit | ||||||||||
Vosk Api | 6,043 | 30 | a month ago | 37 | December 14, 2022 | 408 | apache-2.0 | Jupyter Notebook | ||
Offline speech recognition API for Android, iOS, Raspberry Pi and servers with Python, Java, C# and Node |
SpeechBrain is an open-source and all-in-one conversational AI toolkit based on PyTorch.
The goal is to create a single, flexible, and user-friendly toolkit that can be used to easily develop state-of-the-art speech technologies, including systems for speech recognition, speaker recognition, speech enhancement, speech separation, language identification, multi-microphone signal processing, and many others.
Please, star our project on github (see top-right corner) if you appreciate our contribution to the community!
SpeechBrain is currently in beta.
| Tutorials | Website | Documentation | Contributing | HuggingFace | YouTube |
In March 2023, PyTorch introduced a new version, PyTorch 2.0, which offers numerous enhancements to the community. At present, the majority of SpeechBrain is compatible with PyTorch 2.0. However, certain sections of the code remain incompatible, and we are actively working towards full compatibility with PyTorch 2.0. For the time being, we recommend users continue utilizing PyTorch 1.13, as this is the version employed in our experiments.
If you wish to use SpeechBrain alongside PyTorch 2.0 and encounter any issues, kindly inform us by responding to this issue.
SpeechBrain provides various useful tools to speed up and facilitate research on speech and language technologies:
Brain
class is a fully-customizable tool for managing training and evaluation loops over data. The annoying details of training loops are handled for you while retaining complete flexibility to override any part of the process when needed.SpeechBrain supports state-of-the-art methods for end-to-end speech recognition:
dataset
to facilitate the training over a large text dataset.SpeechBrain provides efficient (GPU-friendly) speech augmentation and feature extraction pipelines:
SpeechBrain provides different models for speaker recognition, identification, and diarization on different datasets:
We have models for converting characters into a sequence of phonemes. In particular, we have Transformer- and RNN-based models operating at the sentence level (i.e, converting a full sentence into a corresponding sequence of phonemes). The models are trained with both data from Wikipedia and LibriSpeech.
SpeechBrain provides different models for language identification. In particular, our best model is based on an ECAPA-TDNN trained with the voxlingua107 dataset.
Combining multiple microphones is a powerful approach to achieving robustness in adverse acoustic environments:
The recipes released with speechbrain implement speech processing systems with competitive or state-of-the-art performance. In the following, we report the best performance achieved on some popular benchmarks:
Dataset | Task | System | Performance |
---|---|---|---|
LibriSpeech | Speech Recognition | wav2vec2 | WER=1.90% (test-clean) |
LibriSpeech | Speech Recognition | CNN + Conformer | WER=2.0% (test-clean) |
TIMIT | Speech Recognition | CRDNN + distillation | PER=13.1% (test) |
TIMIT | Speech Recognition | wav2vec2 + CTC/Att. | PER=8.04% (test) |
CommonVoice (English) | Speech Recognition | wav2vec2 + CTC | WER=15.69% (test) |
CommonVoice (French) | Speech Recognition | wav2vec2 + CTC | WER=9.96% (test) |
CommonVoice (Italian) | Speech Recognition | wav2vec2 + seq2seq | WER=9.86% (test) |
CommonVoice (Kinyarwanda) | Speech Recognition | wav2vec2 + seq2seq | WER=18.91% (test) |
AISHELL (Mandarin) | Speech Recognition | wav2vec2 + CTC | CER=5.06% (test) |
Fisher-callhome (spanish) | Speech translation | conformer (ST + ASR) | BLEU=48.04 (test) |
VoxCeleb2 | Speaker Verification | ECAPA-TDNN | EER=0.80% (vox1-test) |
AMI | Speaker Diarization | ECAPA-TDNN | DER=3.01% (eval) |
VoiceBank | Speech Enhancement | MetricGAN+ | PESQ=3.08 (test) |
WSJ2MIX | Speech Separation | SepFormer | SDRi=22.6 dB (test) |
WSJ3MIX | Speech Separation | SepFormer | SDRi=20.0 dB (test) |
WHAM! | Speech Separation | SepFormer | SDRi= 16.4 dB (test) |
WHAMR! | Speech Separation | SepFormer | SDRi= 14.0 dB (test) |
Libri2Mix | Speech Separation | SepFormer | SDRi= 20.6 dB (test-clean) |
Libri3Mix | Speech Separation | SepFormer | SDRi= 18.7 dB (test-clean) |
LibryParty | Voice Activity Detection | CRDNN | F-score=0.9477 (test) |
IEMOCAP | Emotion Recognition | wav2vec2 | Accuracy=79.8% (test) |
CommonLanguage | Language Recognition | ECAPA-TDNN | Accuracy=84.9% (test) |
Timers and Such | Spoken Language Understanding | CRDNN | Intent Accuracy=89.2% (test) |
SLURP | Spoken Language Understanding | HuBERT | Intent Accuracy=87.54% (test) |
VoxLingua 107 | Identification | ECAPA-TDNN | Sentence Accuracy=93.3% (test) |
For more details, take a look at the corresponding implementation in recipes/dataset/.
Beyond providing recipes for training the models from scratch, SpeechBrain shares several pre-trained models (coupled with easy-inference functions) on HuggingFace. In the following, we report some of them:
Task | Dataset | Model |
---|---|---|
Speech Recognition | LibriSpeech | CNN + Transformer |
Speech Recognition | LibriSpeech | CRDNN |
Speech Recognition | CommonVoice(English) | wav2vec + CTC |
Speech Recognition | CommonVoice(French) | wav2vec + CTC |
Speech Recognition | CommonVoice(Italian) | wav2vec + CTC |
Speech Recognition | CommonVoice(Kinyarwanda) | wav2vec + CTC |
Speech Recognition | AISHELL(Mandarin) | wav2vec + seq2seq |
Text-to-Speech | LJSpeech | Tacotron2 |
Speaker Recognition | Voxceleb | ECAPA-TDNN |
Speech Separation | WHAMR! | SepFormer |
Speech Enhancement | Voicebank | MetricGAN+ |
Speech Enhancement | WHAMR! | SepFormer |
Spoken Language Understanding | Timers and Such | CRDNN |
Language Identification | CommonLanguage | ECAPA-TDNN |
The full list of pre-trained models can be found on HuggingFace
SpeechBrain is designed to speed up the research and development of speech technologies. Hence, our code is backed-up with different levels of documentation:
We are currently implementing speech synthesis pipelines and real-time speech processing pipelines. An interface with the Finite State Transducers (FST) implemented by the Kaldi 2 team is under development.
(documentation) (tutorials)
.. ..
| readthedocs | > | Colab |
\/ \/
^ |
(release) | v
.. .. (landing) ..
| PyPI | > | github.io | (page) | templates | (reference)
\/ \/ > \/ (implementation)
| | |
v v v
.. .. .. .~~~~~~~~~~~~~.
| HyperPyYAML |~~~| speechbrain | > | recipes | > | HuggingFace |
\/ \/ \/ \~~~~~~~~~~~~~/
(usability) (source/modules) (use cases) (pretrained models)
| | |
v v v
.~~~~~~~~~~~~~. .~~~~~~~~. ..
| PyTorch | -> | GDrive | | Inference |
\~~~~~~~~~~~~~/ \~~~~~~~~/ \/
(checkpoints) (results) (code snippets)
SpeechBrain has been presented at Interspeech 2021 and 2022 as well as ASRU 2021. When possible, we will provide some ressources here:
SpeechBrain is constantly evolving. New features, tutorials, and documentation will appear over time. SpeechBrain can be installed via PyPI. Moreover, a local installation can be used by those users who want to run experiments and modify/customize the toolkit. SpeechBrain supports both CPU and GPU computations. For most all the recipes, however, a GPU is necessary during training. Please note that CUDA must be properly installed to use GPUs.
Once you have created your Python environment (Python 3.7+) you can simply type:
pip install speechbrain
Then you can access SpeechBrain with:
import speechbrain as sb
Once you have created your Python environment (Python 3.7+) you can simply type:
git clone https://github.com/speechbrain/speechbrain.git
cd speechbrain
pip install -r requirements.txt
pip install --editable .
Then you can access SpeechBrain with:
import speechbrain as sb
Any modification made to the speechbrain
package will be automatically interpreted as we installed it with the --editable
flag.
Please, run the following script to make sure your installation is working:
pytest tests
pytest --doctest-modules speechbrain
In SpeechBrain, you can run experiments in this way:
> cd recipes/<dataset>/<task>/
> python experiment.py params.yaml
The results will be saved in the output_folder
specified in the yaml file. The folder is created by calling sb.core.create_experiment_directory()
in experiment.py
. Both detailed logs and experiment outputs are saved there. Furthermore, less verbose logs are output to stdout.
As a community-based and open-source project, SpeechBrain needs the help of its community to grow in the right direction. Opening the roadmap to our users enables the toolkit to benefit from new ideas, new research axes, or even new technologies. The roadmap will be available in our GitHub Discussions and will list all the changes and updates that need to be done in the current version of SpeechBrain. Users are more than welcome to propose new items via new Discussions topics!
We provide users with different resources to learn how to use SpeechBrain:
SpeechBrain is released under the Apache License, version 2.0. The Apache license is a popular BSD-like license. SpeechBrain can be redistributed for free, even for commercial purposes, although you can not take off the license headers (and under some circumstances, you may have to distribute a license document). Apache is not a viral license like the GPL, which forces you to release your modifications to the source code. Note that this project has no connection to the Apache Foundation, other than that we use the same license terms.
We constantly update the community using Twitter. Feel free to follow us
Please, cite SpeechBrain if you use it for your research or business.
@misc{speechbrain,
title={{SpeechBrain}: A General-Purpose Speech Toolkit},
author={Mirco Ravanelli and Titouan Parcollet and Peter Plantinga and Aku Rouhe and Samuele Cornell and Loren Lugosch and Cem Subakan and Nauman Dawalatabad and Abdelwahab Heba and Jianyuan Zhong and Ju-Chieh Chou and Sung-Lin Yeh and Szu-Wei Fu and Chien-Feng Liao and Elena Rastorgueva and Franois Grondin and William Aris and Hwidong Na and Yan Gao and Renato De Mori and Yoshua Bengio},
year={2021},
eprint={2106.04624},
archivePrefix={arXiv},
primaryClass={eess.AS},
note={arXiv:2106.04624}
}