Datasets:
File size: 8,242 Bytes
303c9c8 ca08d7d d12c88a ca08d7d d12c88a 518f728 f369161 d12c88a 518f728 d12c88a ca08d7d 518f728 ca08d7d 303c9c8 69aaf14 303c9c8 69aaf14 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 |
---
license: mit
task_categories:
- image-to-text
- text-to-image
- audio-classification
- image-classification
- tabular-classification
tags:
- audio
- image
- multimodal
- visualization
- audio-visualization
- 3d-visualization
- synthetic
- proof-of-concept
- frequency-estimation
- generative-audio
- music-visualization
---
[](https://webxos.netlify.app)
[](https://github.com/webxos/webxos)
[](https://huggingface.co/webxos)
[](https://x.com/webxos)
<div style="
background: #00FF00;
border-left: 3px solid #00FF00;
padding: 12px;
margin: 10px 0;
font-family: 'Fira Code', 'Courier New', monospace;
color: #00FF00;
border-radius: 0 4px 4px 0;
font-size: 8px;
line-height: 1.1;
max-width: 100%;
overflow: hidden;
">
<pre style="
font-size: 6px;
line-height: 1.1;
margin: 0;
padding: 0;
color: #00FF00;
letter-spacing: 0px;
word-spacing: 0px;
">
AAA UUUUUUUU UUUUUUUUDDDDDDDDDDDDD IIIIIIIIII OOOOOOOOO FFFFFFFFFFFFFFFFFFFFFF OOOOOOOOO RRRRRRRRRRRRRRRRR MMMMMMMM MMMMMMMM
A:::A U::::::U U::::::UD::::::::::::DDD I::::::::I OO:::::::::OO F::::::::::::::::::::F OO:::::::::OO R::::::::::::::::R M:::::::M M:::::::M
A:::::A U::::::U U::::::UD:::::::::::::::DD I::::::::I OO:::::::::::::OO F::::::::::::::::::::F OO:::::::::::::OO R::::::RRRRRR:::::R M::::::::M M::::::::M
A:::::::A UU:::::U U:::::UUDDD:::::DDDDD:::::DII::::::IIO:::::::OOO:::::::OFF::::::FFFFFFFFF::::FO:::::::OOO:::::::ORR:::::R R:::::RM:::::::::M M:::::::::M
A:::::::::A U:::::U U:::::U D:::::D D:::::D I::::I O::::::O O::::::O F:::::F FFFFFFO::::::O O::::::O R::::R R:::::RM::::::::::M M::::::::::M
A:::::A:::::A U:::::D D:::::U D:::::D D:::::DI::::I O:::::O O:::::O F:::::F O:::::O O:::::O R::::R R:::::RM:::::::::::M M:::::::::::M
A:::::A A:::::A U:::::D D:::::U D:::::D D:::::DI::::I O:::::O O:::::O F::::::FFFFFFFFFF O:::::O O:::::O R::::RRRRRR:::::R M:::::::M::::M M::::M:::::::M
A:::::A A:::::A U:::::D D:::::U D:::::D D:::::DI::::I O:::::O O:::::O F:::::::::::::::F O:::::O O:::::O R:::::::::::::RR M::::::M M::::M M::::M M::::::M
A:::::A A:::::A U:::::D D:::::U D:::::D D:::::DI::::I O:::::O O:::::O F:::::::::::::::F O:::::O O:::::O R::::RRRRRR:::::R M::::::M M::::M::::M M::::::M
A:::::AAAAAAAAA:::::A U:::::D D:::::U D:::::D D:::::DI::::I O:::::O O:::::O F::::::FFFFFFFFFF O:::::O O:::::O R::::R R:::::RM::::::M M:::::::M M::::::M
A:::::::::::::::::::::A U:::::D D:::::U D:::::D D:::::DI::::I O:::::O O:::::O F:::::F O:::::O O:::::O R::::R R:::::RM::::::M M:::::M M::::::M
A:::::AAAAAAAAAAAAA:::::A U::::::U U::::::U D:::::D D:::::D I::::I O::::::O O::::::O F:::::F O::::::O O::::::O R::::R R:::::RM::::::M MMMMM M::::::M
A:::::A A:::::AU:::::::UUU:::::::U DDD:::::DDDDD:::::DII::::::IIO:::::::OOO:::::::OFF:::::::FF O:::::::OOO:::::::ORR:::::R R:::::RM::::::M M::::::M
A:::::A A:::::AUU:::::::::::::UU D:::::::::::::::DD I::::::::I OO:::::::::::::OO F::::::::FF OO:::::::::::::OO R::::::R R:::::RM::::::M M::::::M
A:::::A A:::::A UU:::::::::UU D::::::::::::DDD I::::::::I OO:::::::::OO F::::::::FF OO:::::::::OO R::::::R R:::::RM::::::M M::::::M
AAAAAAA AAAAAAA UUUUUUUUU DDDDDDDDDDDDD IIIIIIIIII OOOOOOOOO FFFFFFFFFFF OOOOOOOOO RRRRRRRR RRRRRRRMMMMMMMM MMMMMMMM
</pre>
</div>
## Audioform_Dataset_v1
This dataset is the very first output from **AUDIOFORM** — a Three.js powered 3D audio visualization tool that turns audio files
into beautiful, timestamped visual frames with rich metadata. **AUDIOFORM** by webXOS is available for download in the /audioform/
folder of this repo so developers can create their own similar datasets. Audio for is a synthetic harmonic oscilator that runs in HTML,
think of it as the "Hello World" / MNIST-style dataset application for audio-to-visual multimodal machine learning.
This dataset contains **10 captured frames** from a short uploaded WAV file (played at 1× speed), together with per-frame
metadata including dominant frequency, timestamp, and capture info.
## Dataset Structure
```
audioform_dataset/
├── images/
│ ├── frame_0001.png
│ ├── frame_0002.png
│ └── ... (10 PNG frames total)
├── metadata.csv # Main metadata file (Hugging Face viewer uses this)
└── README.md
```
```
| Column | Type | Description | Example Value |
|---------------|---------|-----------------------------------------------------------------------------|-----------------------------------|
| `file_name` | string | Relative path to the visualization PNG (required by Hugging Face) | `images/frame_0001.png` |
| `frame_id` | int | Sequential frame number (0-based) | 0, 1, 2, …, 9 |
| `timestamp` | float | Time in seconds when the frame was captured from the audio | 5.365, 6.219, 9.504 |
| `frequency` | int | Dominant / main detected audio frequency at capture time (Hz) | 0 (in this tiny sample) |
| `time_scale` | int | Playback speed multiplier used during visualization | 1 |
| `capture_date`| string | UTC ISO timestamp when the frame was rendered | 2026-01-13T19:57:36.427Z |
```
See how fast a tiny diffusion model / GAN / LoRA can memorize & regenerate these exact 10 styles. Use the frames as
style references for ControlNet, IP-Adapter, or fine-tuning SD to adopt this neon 3D audio-viz aesthetic.
```
This dataset shows the **format** AUDIOFORM produces.
→ Feed it real music, voices, field recordings, synths
→ Generate 1k–100k+ frames
→ Add labels (genre, instrument, mood, multiple freq peaks…)
→ Unlock serious applications:
- Music video auto-generation
- Visual audio classifiers
- Audio-conditioned image/video generation
- Interactive music → 3D art installations
- Novel multimodal music understanding models
```
## Dataset Description
This dataset was generated using AUDIOFORM, a 3D audio visualization system.
- **Total Frames**: 10
- **Generation Date**: 2026-01-13
- **Audio Type**: Uploaded WAV File
- **Time Scaling**: 1x
## Dataset Structure
- `images/`: Contains all captured frames in PNG format
- `metadata.csv`: Contains classification data for each frame
## Metadata Columns
- `file_name`: Relative path to the image file (e.g., images/frame_0001.png) - **REQUIRED for Hugging Face**
- `frame_id`: Unique identifier for each frame
- `timestamp`: Time in seconds when frame was captured
- `frequency`: Audio frequency at capture time (Hz)
- `time_scale`: Playback speed multiplier
- `capture_date`: ISO date string of capture
## Intended Use
This dataset is intended for training machine learning models on audio visualization patterns, waveform classification, or generative AI tasks.
|