alohajazz commited on
Commit
f46fe95
Β·
verified Β·
1 Parent(s): b4c2d8f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +87 -3
README.md CHANGED
@@ -1,3 +1,87 @@
1
- ---
2
- license: cc-by-nc-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ tags:
4
+ - audio
5
+ - sound-event-detection
6
+ - underwater-acoustics
7
+ - self-supervised-learning
8
+ - nature-ai
9
+ - beats
10
+ datasets:
11
+ - world-dapt
12
+ language:
13
+ - en
14
+ metrics:
15
+ - f1
16
+ library_name: torch
17
+ pipeline_tag: audio-classification
18
+ ---
19
+
20
+ # OceanBEATs
21
+
22
+ **OceanBEATs** is a foundation model for underwater acoustic monitoring, adapted from [BEATs](https://github.com/microsoft/unilm/tree/master/beats) via Domain-Adaptive Pretraining (DAPT) on approximately 4,400 hours of global ocean soundscapes (World-DAPT corpus).
23
+
24
+ This model serves as the "ears" for embodied acoustic agents described in our paper: **"Embodied acoustic agents with self-supervised audio for unknown-aware underwater soundscapes under label and false-positive constraints"** (Under Review, *npj Artificial Intelligence*).
25
+
26
+ ## Model Details
27
+
28
+ - **Model Type:** Audio Transformer (BEATs architecture)
29
+ - **Pretraining:** Masked Audio Modeling + DAPT (SimCLR/InfoNCE) on underwater data
30
+ - **Input:** 16kHz mono audio waveform
31
+ - **Backbone:** BEATs AS-2M (iter3+)
32
+
33
+ ## Available Files
34
+
35
+ This repository hosts the pretrained weights required to reproduce the results in our paper.
36
+
37
+ 1. **`beats_dapt_topup_encoder.pt`**
38
+ * The core encoder (backbone) adapted to underwater acoustics.
39
+ * Use this for feature extraction, unknown detection (CCED2), or fine-tuning on new marine datasets.
40
+
41
+ 2. **`sed_head_56_topup_ep8.pt`**
42
+ * A 56-class Sound Event Detection (SED) head trained on coastal/lagoon data (Okinawa, Japan).
43
+ * Detects fish, mammals, vessels, and environmental sounds.
44
+
45
+ ## Usage
46
+
47
+ These weights are designed to be used with the official code repository:
48
+
49
+ **GitHub Repository:** [BiologgingSolutions/embodied-ocean-cced2-dgpu](https://github.com/BiologgingSolutions/embodied-ocean-cced2-dgpu)
50
+
51
+ Please download the `.pt` files and place them in the `weights/` directory of the cloned GitHub repository.
52
+
53
+ ```bash
54
+ # Example directory structure after download
55
+ embodied-ocean-cced2-dgpu/
56
+ └── weights/
57
+ β”œβ”€β”€ beats_dapt_topup_encoder.pt
58
+ β”œβ”€β”€ sed_head_56_topup_ep8.pt
59
+ └── cced2/ ...
60
+ ```
61
+
62
+ ## License & Data Availability
63
+
64
+ **License:** CC BY-NC 4.0 (Creative Commons Attribution-NonCommercial 4.0 International)
65
+
66
+ These weights are released for **non-commercial research purposes only**.
67
+ Commercial use is strictly prohibited without prior permission from the authors.
68
+
69
+ > **Note:** The source code for using these models is released under the **MIT License** at the GitHub repository linked above.
70
+
71
+ ## Citation
72
+
73
+ If you use this model in your research, please cite our paper:
74
+
75
+ ```bibtex
76
+ @article{noda2025embodied,
77
+ title={Embodied acoustic agents with self-supervised audio for unknown-aware underwater soundscapes under label and false-positive constraints},
78
+ author={Noda, Takuji and Koizumi, Takuya},
79
+ journal={npj Artificial Intelligence (Special Collection: Embodied AI)},
80
+ note={Under Review},
81
+ year={2025}
82
+ }
83
+ ```
84
+
85
+ ## Acknowledgements
86
+
87
+ The base model architecture is based on BEATs (Microsoft). We acknowledge the creators of the BEATs model and the various open-source ocean acoustic datasets (SanctSound, ONC, PALAOA, etc.) used for DAPT.