Papers
arxiv:2601.14960

VCNAC: A Variable-Channel Neural Audio Codec for Mono, Stereo, and Surround Sound

Published on Jan 21
Authors:
,
,
,

Abstract

A unified neural audio codec architecture supports variable channel configurations through shared representations and channel compatibility objectives while maintaining high reconstruction quality across different audio formats.

AI-generated summary

We present VCNAC, a variable channel neural audio codec. Our approach features a single encoder and decoder parametrization that enables native inference for different channel setups, from mono speech to cinematic 5.1 channel surround audio. Channel compatibility objectives ensure that multi-channel content maintains perceptual quality when decoded to fewer channels. The shared representation enables training of generative language models on a single set of codebooks while supporting inference-time scalability across modalities and channel configurations. Evaluation using objective spatial audio metrics and subjective listening tests demonstrates that our unified approach maintains high reconstruction quality across mono, stereo, and surround audio configurations.

Community

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2601.14960 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2601.14960 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2601.14960 in a Space README.md to link it from this page.

Collections including this paper 1