Source code for mirdata.datasets.gtzan_genre

"""GTZAN-Genre Dataset Loader

.. admonition:: Dataset Info
    :class: dropdown

    This dataset was used for the well known genre classification paper:

    .. code-block:: latex

        "Musical genre classification of audio signals " by G. Tzanetakis and
        P. Cook in IEEE Transactions on Audio and Speech Processing 2002.

    The dataset consists of 1000 audio tracks each 30 seconds long. It
    contains 10 genres, each represented by 100 tracks. The tracks are all
    22050 Hz mono 16-bit audio files in .wav format.

"""

from typing import BinaryIO, Optional, TextIO, Tuple

from deprecated.sphinx import deprecated
import librosa
import numpy as np

from mirdata import download_utils, jams_utils, core, io, annotations


BIBTEX = """@article{tzanetakis2002gtzan,
  title={GTZAN genre collection},
  author={Tzanetakis, George and Cook, P},
  journal={Music Analysis, Retrieval and Synthesis for Audio Signals},
  year={2002}
}"""

INDEXES = {
    "default": "1.0",
    "test": "1.0",
    "1.0": core.Index(
        filename="gtzan_genre_index_1.0.json",
        partial_download=["all", "tempo_beat_annotations"],
    ),
    "mini": core.Index(
        filename="gtzan_genre_1.0_mini_index.json",
        partial_download=["mini", "tempo_beat_annotations"],
    ),
}

REMOTES = {
    "all": download_utils.RemoteFileMetadata(
        filename="genres.tar.gz",
        url="http://opihi.cs.uvic.ca/sound/genres.tar.gz",
        checksum="5b3d6dddb579ab49814ab86dba69e7c7",
        destination_dir="gtzan_genre",
    ),
    "mini": download_utils.RemoteFileMetadata(
        filename="main.zip",
        url="https://github.com/TempoBeatDownbeat/gtzan_mini/archive/refs/heads/main.zip",
        checksum="44f7f23af8363d96c59663a987f29a4c",
    ),
    "tempo_beat_annotations": download_utils.RemoteFileMetadata(
        filename="annot.zip",
        url="https://github.com/TempoBeatDownbeat/gtzan_tempo_beat/archive/refs/heads/main.zip",
        checksum="4baa58112697a8087de04558d6e97442",
    ),
}

LICENSE_INFO = "Unfortunately we couldn't find the license information for the GTZAN_genre dataset."


[docs] class Track(core.Track): """gtzan_genre Track class Args: track_id (str): track id of the track Attributes: audio_path (str): path to the audio file genre (str): annotated genre track_id (str): track id Cached Properties: beats (BeatData): human-labeled beat annotations tempo (float): global tempo annotations """ def __init__(self, track_id, data_home, dataset_name, index, metadata): super().__init__(track_id, data_home, dataset_name, index, metadata) self.genre = track_id.split(".")[0] if self.genre == "hiphop": self.genre = "hip-hop" self.audio_path = self.get_path("audio") self.beats_path = self.get_path("beats") self.tempo_path = self.get_path("tempo") @core.cached_property def beats(self) -> Optional[annotations.BeatData]: return load_beats(self.beats_path) @core.cached_property def tempo(self) -> Optional[float]: return load_tempo(self.tempo_path) @property def audio(self) -> Optional[Tuple[np.ndarray, float]]: """The track's audio Returns: * np.ndarray - audio signal * float - sample rate """ return load_audio(self.audio_path)
[docs] def to_jams(self): """Get the track's data in jams format Returns: jams.JAMS: the track's data in jams format """ return jams_utils.jams_converter( tags_gtzan_data=[(self.genre, "gtzan-genre")], beat_data=[(self.beats, None)], tempo_data=[(self.tempo, None)], metadata={ "title": "Unknown track", "artist": "Unknown artist", "release": "Unknown album", "duration": 30.0, "curator": "George Tzanetakis", }, )
[docs] @io.coerce_to_string_io def load_beats(fhandle: TextIO) -> annotations.BeatData: """Load GTZAN format beat data from a file Args: fhandle (str or file-like): path or file-like object pointing to a beat annotation file Returns: BeatData: loaded beat data """ beats = np.loadtxt(fhandle, ndmin=2) times = beats[:, 0] try: positions = beats[:, 1] except IndexError: positions = None beat_data = annotations.BeatData( times=times, time_unit="s", positions=positions, position_unit="bar_index" ) return beat_data
[docs] @io.coerce_to_string_io def load_tempo(fhandle: TextIO) -> float: """Load GTZAN format tempo data from a file Args: fhandle (str or file-like): path or file-like object pointing to a beat annotation file Returns: tempo (float): loaded tempo data """ tempo = np.loadtxt(fhandle, ndmin=2) return float(tempo)
[docs] @io.coerce_to_bytes_io def load_audio(fhandle: BinaryIO) -> Tuple[np.ndarray, float]: """Load a GTZAN audio file. Args: fhandle (str or file-like): File-like object or path to audio file Returns: * np.ndarray - the mono audio signal * float - The sample rate of the audio file """ audio, sr = librosa.load(fhandle, sr=22050, mono=True) return audio, sr
[docs] @core.docstring_inherit(core.Dataset) class Dataset(core.Dataset): """ The gtzan_genre dataset """ def __init__(self, data_home=None, version="default"): super().__init__( data_home, version, name="gtzan_genre", track_class=Track, bibtex=BIBTEX, indexes=INDEXES, remotes=REMOTES, license_info=LICENSE_INFO, )
[docs] @deprecated(reason="Use mirdata.datasets.gtzan_genre.load_audio", version="0.3.4") def load_audio(self, *args, **kwargs): return load_audio(*args, **kwargs)