Contributing

We encourage contributions to soundata, especially new dataset loaders. To contribute a new loader, follow the steps indicated below and create a Pull Request (PR) to the github repository. For any doubt or comment about your contribution, you can always submit an issue or open a discussion in the repository.

Installing soundata for development purposes

To install soundata for development purposes:

  • First run:

git clone https://github.com/soundata/soundata.git
  • Then, after opening source data library you have to install the dependencies for updating the documentation and running tests:

pip install .
pip install ."[tests]"
pip install ."[docs]"

We recommend using miniconda or pyenv to manage your Python versions and install all soundata requirements. You will want to install the latest supported Python versions (see README.md). Once conda or pyenv and the Python versions are configured, install pytest. Make sure you’ve installed all the necessary pytest plugins needed (e.g. pytest-cov) to automatically test your code successfully.

Before running the tests, make sure to have formatted soundata/ and tests/ with black.

black soundata/ tests/

Also, make sure that they pass flake8 and mypy tests specified in lint-python.yml github action workflow.

flake8 soundata --count --select=E9,F63,F7,F82 --show-source --statistics
python -m mypy soundata --ignore-missing-imports --allow-subclassing-any

Finally, run:

pytest -vv --cov-report term-missing --cov-report=xml --cov=soundata tests/ --local

All tests should pass!

Note

Soundata assumes that your system has the zip library installed for unzipping files.

Writing a new dataset loader

The steps to add a new dataset loader to soundata are:

  1. Create an index

  2. Create a module

  3. Add tests

  4. Submit your loader

Before starting, if your dataset is not fully downloadable you should:

  1. Contact the soundata team by opening an issue or PR so we can discuss how to proceed with the closed dataset.

  2. Show that the version used to create the checksum is the “canonical” one, either by getting the version from the dataset creator, or by verifying equivalence with several other copies of the dataset.

To reduce friction, we will make commits on top of contributors PRs by default unless the please-do-not-edit flag is used.

1. Create an index

Soundata’s structure relies on indexes. Indexes are dictionaries that contain information about the structure of the dataset which is necessary for the loading and validating functionalities of Soundata. In particular, indexes contain information about the files included in the dataset, their location and checksums, see some example indexes below. To create an index, the necessary steps are:

  1. Create a script in scripts/, called make_<datasetname>_index.py, which generates an index file.

  2. Then run the script on the canonical versions of the dataset and save the index in soundata/datasets/indexes/ as <datasetname>_index.json.

The function make_<datasetname>_index.py should automate the generation of an index by computing the MD5 checksums for given files in a dataset located at data_path. Users can adapt this function to create an index for their dataset by adding their file paths and using the md5 function to generate checksums for their files.

Here’s an example of an index to use as a guide:

More examples of scripts used to create dataset indexes can be found in the scripts folder.

Example index with clips

Most sound datasets are organized as a collection of clips and annotations. In such case, the index should make use of the clips top-level key. Under this clips top-level key, you should store a dictionary where the keys are the unique clip ids of the dataset, and the values are dictionaries of files associated with a clip id, along with their checksums. These files can be for instance audio files or annotations related to the clip id. File paths are relative to the top level directory of a dataset.

2. Create a module

Once the index is created you can create the loader. For that, we suggest you use the following template and adjust it for your dataset. To quickstart a new module:

  1. Copy the example below and save it to soundata/datasets/<your_dataset_name>.py

  2. Find & Replace Example with the <your_dataset_name>.

  3. Remove any lines beginning with # – which are there as guidelines.

You may find these examples useful as references:

For many more examples, see the datasets folder.

3. Add tests

To finish your contribution, please include tests that check the integrity of your loader. For this, follow these steps:

  1. Make a toy version of the dataset in the tests folder tests/resources/sound_datasets/my_dataset/, so you can test against little data. For example:

    • Include all audio and annotation files for one clip of the dataset

    • For each audio/annotation file, reduce the audio length to 1-2 seconds and remove all but a few of the annotations.

    • If the dataset has a metadata file, reduce the length to a few lines.

  2. Test all of the dataset specific code, e.g. the public attributes of the Clip class, the load functions and any other custom functions you wrote. See the tests folder for reference. If your loader has a custom download function, add tests similar to this mirdata loader.

  3. Locally run pytest -s tests/test_full_dataset.py --local --dataset my_dataset before submitting your loader to make sure everything is working.

Note

We have written automated tests for all loader’s cite, download, validate, load, clip_ids functions, as well as some basic edge cases of the Clip class, so you don’t need to write tests for these!

Running your tests locally

Before creating a PR you should run the tests. But before that, make sure to have formatted soundata/ and tests/ with black.

black soundata/ tests/

Also, make sure that they pass flake8 and mypy tests specified in lint-python.yml github action workflow.

flake8 soundata --count --select=E9,F63,F7,F82 --show-source --statistics
python -m mypy soundata --ignore-missing-imports --allow-subclassing-any

Finally, run all the tests locally like this:

pytest -vv --cov-report term-missing --cov-report=xml --cov=soundata tests/ --local

The --local flag skips tests that are built to run only on the remote testing environment.

To run one specific test file:

pytest tests/test_urbansed.py

Finally, there is one local test you should run, which we can’t easily run in our testing environment.

pytest -s tests/test_full_dataset.py --local --dataset dataset

Where dataset is the name of the module of the dataset you added. The -s tells pytest not to skip print statments, which is useful here for seeing the download progress bar when testing the download function.

This tests that your dataset downloads, validates, and loads properly for every clip. This test takes a long time for some datasets, but it’s important to ensure the integrity of the library.

The --skip-download flag can be added to pytest command to run the tests skipping the download. This will skip the downloading step. Note that this is just for convenience during debugging - the tests should eventually all pass without this flag.

Working with big datasets

In the development of large datasets, it is advisable to create an index as small as possible to optimize the implementation process of the dataset loader and pass the tests.

Working with remote indexes

For the end-user there is no difference between the remote and local indexes. However, indexes can get large when there are a lot of clips in the dataset. In these cases, storing and accessing an index remotely can be convenient. Large indexes can be added to REMOTES, and will be downloaded with the rest of the dataset. For example:

"index": download_utils.RemoteFileMetadata(
    filename="remote_index.json.zip",
    url="https://zenodo.org/record/.../remote_index.json.zip?download=1",
    checksum="810f1c003f53cbe58002ba96e6d4d138",
)

Unlike local indexes, the remote indexes will live in the data_home directory. When creating the Dataset object, specify the custom_index_path to where the index will be downloaded (as a relative path to data_home).

Reducing the testing space usage

We are trying to keep the test resources folder size as small as possible, because it can get really heavy as new loaders are added. We kindly ask the contributors to reduce the size of the testing data if possible (e.g. trimming the audio clips, keeping just two rows for csv files).

4. Submit your loader

Before you submit your loader make sure to:

  1. Add your module to docs/source/soundata.rst following an alphabetical order.

  2. Add your module to docs/source/table.rst following an alphabetical order as follows:

* - Dataset
  - Downloadable?
  - Annotations
  - Clips
  - Hours
  - Usecase
  - License

An example of this for the UrbanSound8k dataset:

* - UrbanSound8K
  - - audio: ✅
    - annotations: ✅
  - :ref:`tags`
  - 8732
  - 8.75
  - Urban sound classification
  - .. image:: https://licensebuttons.net/l/by-nc/4.0/80x15.png
       :target: https://creativecommons.org/licenses/by-nc/4.0

You can find license badges images and links here.

Pull Request template

When starting your PR please use the new_loader.md template, it will simplify the reviewing process and also help you make a complete PR. You can do that by adding &template=new_loader.md at the end of the url when you are creating the PR :

...soundata/soundata/compare?expand=1 will become ...soundata/soundata/compare?expand=1&template=new_loader.md.

Docs

Staged docs for every new PR are built and accessible at soundata.github.io/preview/PR-<#PR_ID> in which <#PR_ID> is the pull request ID. To quickly troubleshoot any issues, you can build the docs locally by nagivating to the docs folder, and running make html (note, you must have sphinx installed). Then open the generated _build/source/index.html file in your web browser to view.

Troubleshooting

If github shows a red X next to your latest commit, it means one of our checks is not passing. This could mean:

  1. running black has failed – this means that your code is not formatted according to black’s code-style. To fix this, simply run the following from inside the top level folder of the repository:

black soundata/ tests/
  1. Your code does not pass flake8 test.

flake8 soundata --count --select=E9,F63,F7,F82 --show-source --statistics
  1. Your code does not pass mypy test.

python -m mypy soundata --ignore-missing-imports --allow-subclassing-any
  1. the test coverage is too low – this means that there are too many new lines of code introduced that are not tested.

  2. the docs build has failed – this means that one of the changes you made to the documentation has caused the build to fail. Check the formatting in your changes and make sure they are consistent.

  3. the tests have failed – this means at least one of the tests is failing. Run the tests locally to make sure they are passing. If they are passing locally but failing in the check, open an issue and we can help debug.

Documentation

This documentation is in rst format. It is built using Sphinx and hosted on readthedocs. The API documentation is built using autodoc, which autogenerates documentation from the code’s docstrings. We use the napoleon plugin for building docs in Google docstring style. See the next section for docstring conventions.

soundata uses Google’s Docstring formatting style. Here are some common examples.

Note

The small formatting details in these examples are important. Differences in new lines, indentation, and spacing make a difference in how the documentation is rendered. For example writing Returns: will render correctly, but Returns or Returns : will not.

Functions:

def add_to_list(list_of_numbers, scalar):
    """Add a scalar to every element of a list.
    You can write a continuation of the function description here on the next line.

    You can optionally write more about the function here. If you want to add an example
    of how this function can be used, you can do it like below.

    Example:
        .. code-block:: python

        foo = add_to_list([1, 2, 3], 2)

    Args:
        list_of_numbers (list): A short description that fits on one line.
        scalar (float):
            Description of the second parameter. If there is a lot to say you can
            overflow to a second line.

    Returns:
        list: Description of the return. The type here is not in parentheses

    """
    return [x + scalar for x in list_of_numbers]

Functions with more than one return value:

def multiple_returns():
    """This function has no arguments, but more than one return value. Autodoc with napoleon doesn't handle this well,
    and we use this formatting as a workaround.

    Returns:
        * int - the first return value
        * bool - the second return value

    """
    return 42, True

One-line docstrings

def some_function():
    """
    One line docstrings must be on their own separate line, or autodoc does not build them properly
    """
    ...

Objects

"""Description of the class
overflowing to a second line if it's long

Some more details here

Args:
    foo (str): First argument to the __init__ method
    bar (int): Second argument to the __init__ method

Attributes:
    foobar (str): First clip attribute
    barfoo (bool): Second clip attribute

Cached Properties:
    foofoo (list): Cached properties are special soundata attributes
    barbar (None): They are lazy loaded properties.
    barf (bool): Document them with this special header.

"""

Conventions

Loading from files

We use the following libraries for loading data from files:

Format

library

audio (wav, mp3, …)

librosa

json

json

csv

csv

jams

jams

Clip Attributes

Custom clip attributes should be global, clip-level data. For some datasets, there is a separate, dataset-level metadata file with clip-level metadata, e.g. as a csv. When a single file is needed for more than one clip, we recommend using writing a _metadata cached property (which returns a dictionary, either keyed by clip_id or freeform) in the Dataset class (see the dataset module example code above). When this is specified, it will populate a clip’s hidden _clip_metadata field, which can be accessed from the clip class.

For example, if _metadata returns a dictionary of the form:

{
    'clip1': {
        'microphone-type': 'Awesome',
        'recording-date': '27.10.2021'
    },
    'clip2': {
        'microphone-type': 'Less_awesome',
        'recording-date': '27.10.2021'
    }
}

the _clip metadata for clip_id=clip2 will be:

{
    'microphone-type': 'Less_awesome',
    'recording-date': '27.10.2021'
}

Load methods vs Clip properties

Clip properties and cached properties should be simple, and directly call a load_* method. Like this example from urbansed:

@property
def split(self):
    """The data splits (e.g. train)

    Returns
        * str - split

    """
    return self._clip_metadata.get("split")

@core.cached_property
def events(self) -> Optional[annotations.Events]:
    """The audio events

    Returns
        * annotations.Events - audio event object

    """
    return load_events(self.txt_path)

There should be no additional logic in a clip property/cached property, and instead all logic should be done in the load method. We separate these because the clip properties are only usable when data is available locally - when data is remote, the load methods are used instead.

Missing Data

Clip properties that are available for some clips and not for others should be set to None when whey are not available. Like this example in the tau2019aus loader:

@property
def tags(self):
    scene_label = self._clip_metadata.get("scene_label")
    if scene_label is None:
        return None
    else:
        return annotations.Tags([scene_label], "open", np.array([1.0]))

The index should only contain key-values for files that exist.

Custom Decorators

cached_property

This is used primarily for Clip classes.

This decorator causes an Object’s function to behave like an attribute (aka, like the @property decorator), but caches the value in memory after it is first accessed. This is used for data which is relatively large and loaded from files.

docstring_inherit

This decorator is used for children of the Dataset class, and copies the Attributes from the parent class to the docstring of the child. This gives us clear and complete docs without a lot of copy-paste.

copy_docs

This decorator is used mainly for a dataset’s load_ functions, which are attached to a loader’s Dataset class. The attached function is identical, and this decorator simply copies the docstring from another function.

coerce_to_bytes_io/coerce_to_string_io

These are two decorators used to simplify the loading of various Clip members in addition to giving users the ability to use file streams instead of paths in case the data is in a remote location e.g. GCS. The decorators modify the function to:

  • Return None if None if passed in.

  • Open a file if a string path is passed in either ‘w’ mode for string_io or wb for bytes_io and pass the file handle to the decorated function.

  • Pass the file handle to the decorated function if a file-like object is passed.

This cannot be used if the function to be decorated takes multiple arguments. coerce_to_bytes_io should not be used if trying to load an mp3 with librosa as libsndfile does not support mp3 yet and audioread expects a path.