Papers
arxiv:2603.13685

Evaluating Compositional Structure in Audio Representations

Published on Mar 14
Authors:
,
,
,

Abstract

A benchmark for evaluating compositionality in audio representations is proposed through two tasks that test consistency under additive transformations and reconstructibility from attribute-level primitives, using synthetic datasets with controlled acoustic variations.

AI-generated summary

We propose a benchmark for evaluating compositionality in audio representations. Audio compositionality refers to representing sound scenes in terms of constituent sources and attributes, and combining them systematically. While central to auditory perception, this property is largely absent from current evaluation protocols. Our framework adapts ideas from vision and language to audio through two tasks: A-COAT, which tests consistency under additive transformations, and A-TRE, which probes reconstructibility from attribute-level primitives. Both tasks are supported by large synthetic datasets with controlled variation in acoustic attributes, providing the first benchmark of compositional structure in audio embeddings.

Community

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2603.13685
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2603.13685 in a model README.md to link it from this page.

Datasets citing this paper 2

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2603.13685 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.