COMP346: Music Intelligence - Generative Audio Modelling & Responsible AI Music

0.0(0)
Studied by 0 people
call kaiCall Kai
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
GameKnowt Play
Card Sorting

1/26

flashcard set

Earn XP

Description and Tags

These vocabulary flashcards cover the key technical concepts of generative audio modelling, neural codecs, specific AI models (MusicLM, MusicGen, Jukebox), and the ethical frameworks for Responsible AI Music (RAIM).

Last updated 6:22 PM on 5/14/26
Name
Mastery
Learn
Test
Matching
Spaced
Call with Kai

No analytics yet

Send a link to your students to track their progress

27 Terms

1
New cards

Symbolic Music Generation

Generative music modelling using compact representations such as MIDI, MusicXML, ABC, or Kern, where music is treated as a language modelling problem.

2
New cards

Audio Music Generation

The process of generating sound based on spectrograms (2D approach) or waveforms (1D approach), capturing rich performance and production info at the cost of high dimensionality.

3
New cards

The Audio Bottleneck

The computational challenge of predicting audio sample-by-sample due to high density; standard CD-quality audio features 44,10044,100 individual data points per second.

4
New cards

Autoencoder (AE)

A neural network module that maps high-dimensional input to a smaller, compressed "Latent Space" (embedding) via an Encoder and reconstructs it via a Decoder.

5
New cards

Variational Autoencoder (VAE)

A generative AE that maps inputs to a continuous probability distribution (via mean and variance) rather than a fixed point, forcing the model to learn a smooth latent space.

6
New cards

Reconstruction Loss

A component of the VAE loss function that measures how accurately the decoder rebuilt the original input from the latent representation.

7
New cards

KL Divergence

A regularization term in the VAE loss function that forces learned distributions to stay close to a standard normal distribution to prevent gaps in the latent space.

8
New cards

Neural Audio Codecs (NACs)

Neural networks that learn the optimal way to convert raw waveforms into an "alphabet" of discrete acoustic tokens for efficient AI processing.

9
New cards

Vector Quantisation (VQ)

The process of turning continuous VAE embeddings into discrete tokens by "snapping" a latent representation to the nearest cluster center in a fixed codebook.

10
New cards

Codebook

A fixed set of cluster centers in a latent space used in Vector Quantisation to represent audio samples as discrete indices.

11
New cards

VQ-VAE

A Variational Autoencoder that utilizes a vector-quantized latent space to learn discrete representations of data.

12
New cards

Residual Vector Quantisation (RVQ)

A technique using multiple sequential codebooks where the first level captures coarse features and subsequent levels quantize the "residual" error to capture fine details like timbre and reverb.

13
New cards

OpenAI Jukebox

A 2020 generative model using three levels of hierarchical VQ-VAE to capture both global musical structure and fine acoustic details, such as singing voices.

14
New cards

Semantic Tokens

Tokens extracted using models like w2v-BERT that capture long-term coherence, including melody, rhythm, and speech content (the "what" of the audio).

15
New cards

Acoustic Tokens

Tokens extracted from Neural Audio Codecs that capture high-fidelity details required for reconstruction (the "how" of the audio).

16
New cards

MusicLM

A Google generative model that builds on the AudioLM pipeline by adding text conditioning through a joint text-audio embedding model called MuLan.

17
New cards

MusicGen

A Meta generative model that uses token interleaving (delay pattern) and the EnCodec tokenizer to maintain musical structure without separate semantic tokens.

18
New cards

Vocoder

A specialized neural network, such as Hi-Fi GAN, used to translate 2D spectrogram representations back into 1D audio waves.

19
New cards

Fréchet Audio Distance (FAD)

An evaluation metric for acoustic quality that compares the statistical distribution of AI-generated audio against a dataset of real studio-quality music; lower scores indicate higher fidelity.

20
New cards

CLAP Score

A metric for semantic alignment (Contrastive Language-Audio Pretraining) that measures how well generated audio matches the provided text prompt.

21
New cards

RAIM Framework

The Responsible AI Music framework, an interdisciplinary effort to identify features for the ethical and responsible development of generative music systems.

22
New cards

Human Agency and Oversight

A RAIM principle ensuring individuals can influence and monitor AI systems while maintaining control over the creative process.

23
New cards

Robustness and Safety

A RAIM requirement that generative systems remain reliable and secure, preventing the production of harmful content or exposure to adversarial attacks.

24
New cards

Privacy and Data Governance

A RAIM requirement ensuring that training data is legally acquired and that the system does not leak private information or violate copyrights.

25
New cards

Transparency

A principle encompassing explainability regarding the data, model architecture, and business models used in generative AI systems.

26
New cards

Diversity, Fairness, and Non-Discrimination

The implementation of mechanisms to avoid unfair bias and ensure accessibility and fair treatment for all users of generative systems.

27
New cards

Accountability

A requirement ensuring AI systems are responsible for their design, implementation, and impact throughout their entire lifecycle.