README.md 5.18 KB
Newer Older
Soumith Chintala's avatar
Soumith Chintala committed
1
torchaudio: an audio library for PyTorch
Vincent QB's avatar
Vincent QB committed
2
========================================
Soumith Chintala's avatar
Soumith Chintala committed
3

moto's avatar
moto committed
4
[![Build Status](https://circleci.com/gh/pytorch/audio.svg?style=svg)](https://app.circleci.com/pipelines/github/pytorch/audio)
moto's avatar
moto committed
5
[![Documentation](https://img.shields.io/badge/dynamic/json.svg?label=docs&url=https%3A%2F%2Fpypi.org%2Fpypi%2Ftorchaudio%2Fjson&query=%24.info.version&colorB=brightgreen&prefix=v)](https://pytorch.org/audio/main/)
moto's avatar
moto committed
6
7
[![Anaconda Badge](https://anaconda.org/pytorch/torchaudio/badges/downloads.svg)](https://anaconda.org/pytorch/torchaudio)
[![Anaconda-Server Badge](https://anaconda.org/pytorch/torchaudio/badges/platforms.svg)](https://anaconda.org/pytorch/torchaudio)
8

moto's avatar
moto committed
9
10
![TorchAudio Logo](docs/source/_static/img/logo.png)

jamarshon's avatar
jamarshon committed
11
The aim of torchaudio is to apply [PyTorch](https://github.com/pytorch/pytorch) to
12
the audio domain. By supporting PyTorch, torchaudio follows the same philosophy
jamarshon's avatar
jamarshon committed
13
14
of providing strong GPU acceleration, having a focus on trainable features through
the autograd system, and having consistent style (tensor names and dimension names).
15
Therefore, it is primarily a machine learning library and not a general signal
16
processing library. The benefits of PyTorch can be seen in torchaudio through
17
having all the computations be through PyTorch operations which makes it easy
jamarshon's avatar
jamarshon committed
18
19
to use and feel like a natural extension.

moto's avatar
moto committed
20
- [Support audio I/O (Load files, Save files)](http://pytorch.org/audio/main/)
moto's avatar
moto committed
21
  - Load a variety of audio formats, such as `wav`, `mp3`, `ogg`, `flac`, `opus`, `sphere`, into a torch Tensor using SoX
moto's avatar
moto committed
22
23
  - [Kaldi (ark/scp)](http://pytorch.org/audio/main/kaldi_io.html)
- [Dataloaders for common audio datasets](http://pytorch.org/audio/main/datasets.html)
24
- Common audio transforms
moto's avatar
moto committed
25
    - [Spectrogram, AmplitudeToDB, MelScale, MelSpectrogram, MFCC, MuLawEncoding, MuLawDecoding, Resample](http://pytorch.org/audio/main/transforms.html)
26
- Compliance interfaces: Run code using PyTorch that align with other libraries
moto's avatar
moto committed
27
    - [Kaldi: spectrogram, fbank, mfcc](https://pytorch.org/audio/main/compliance.kaldi.html)
Soumith Chintala's avatar
Soumith Chintala committed
28
29
30
31

Installation
------------

32
Please refer to https://pytorch.org/audio/main/installation.html for installation and build process of TorchAudio.
33

Soumith Chintala's avatar
Soumith Chintala committed
34
35
36
37
38
Quick Usage
-----------

```python
import torchaudio
39
40
41

waveform, sample_rate = torchaudio.load('foo.wav')  # load tensor from file
torchaudio.save('foo_save.wav', waveform, sample_rate)  # save tensor to file
Soumith Chintala's avatar
Soumith Chintala committed
42
43
```

44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
Backend Dispatch
----------------

By default in OSX and Linux, torchaudio uses SoX as a backend to load and save files.
The backend can be changed to [SoundFile](https://pysoundfile.readthedocs.io/en/latest/)
using the following. See [SoundFile](https://pysoundfile.readthedocs.io/en/latest/)
for installation instructions.

```python
import torchaudio
torchaudio.set_audio_backend("soundfile")  # switch backend

waveform, sample_rate = torchaudio.load('foo.wav')  # load tensor from file, as usual
torchaudio.save('foo_save.wav', waveform, sample_rate)  # save tensor to file, as usual
```

moto's avatar
moto committed
60
61
62
**Note**
- SoundFile currently does not support mp3.
- "soundfile" backend is not supported by TorchScript.
63

Soumith Chintala's avatar
Soumith Chintala committed
64
API Reference
Vincent QB's avatar
Vincent QB committed
65
-------------
SeanNaren's avatar
SeanNaren committed
66

moto's avatar
moto committed
67
API Reference is located here: http://pytorch.org/audio/main/
Vincent QB's avatar
Vincent QB committed
68

jamarshon's avatar
jamarshon committed
69
70
71
Contributing Guidelines
-----------------------

Nicolas Hug's avatar
Nicolas Hug committed
72
Please refer to [CONTRIBUTING.md](./CONTRIBUTING.md)
Vincent QB's avatar
Vincent QB committed
73

yangarbiter's avatar
yangarbiter committed
74
75
76
77
78
79
80
81
82
83
84
85
86
87
Citation
--------

If you find this package useful, please cite as:

```bibtex
@article{yang2021torchaudio,
  title={TorchAudio: Building Blocks for Audio and Speech Processing},
  author={Yao-Yuan Yang and Moto Hira and Zhaoheng Ni and Anjali Chourdia and Artyom Astafurov and Caroline Chen and Ching-Feng Yeh and Christian Puhrsch and David Pollack and Dmitriy Genzel and Donny Greenberg and Edward Z. Yang and Jason Lian and Jay Mahadeokar and Jeff Hwang and Ji Chen and Peter Goldsborough and Prabhat Roy and Sean Narenthiran and Shinji Watanabe and Soumith Chintala and Vincent Quenneville-Bélair and Yangyang Shi},
  journal={arXiv preprint arXiv:2110.15018},
  year={2021}
}
```

Vincent QB's avatar
Vincent QB committed
88
89
90
91
92
93
Disclaimer on Datasets
----------------------

This is a utility library that downloads and prepares public datasets. We do not host or distribute these datasets, vouch for their quality or fairness, or claim that you have license to use the dataset. It is your responsibility to determine whether you have permission to use the dataset under the dataset's license.

If you're a dataset owner and wish to update any part of it (description, citation, etc.), or do not want your dataset to be included in this library, please get in touch through a GitHub issue. Thanks for your contribution to the ML community!
94
95
96
97
98
99
100
101
102

Pre-trained Model License
-------------------------

The pre-trained models provided in this library may have their own licenses or terms and conditions derived from the dataset used for training. It is your responsibility to determine whether you have permission to use the models for your use case.

For instance, SquimObjective model is released under the Creative Commons Attribution 4.0 International license. See [DNS 2020 license](https://github.com/microsoft/DNS-Challenge/blob/interspeech2020/master/LICENSE) for additional details.

Other pre-trained models that have different license are noted in documentation. Please checkout the [documentation page](https://pytorch.org/audio/main/).