AlexanderGbd commited on
Commit
4e042c2
·
verified ·
1 Parent(s): a94109c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +68 -3
README.md CHANGED
@@ -1,3 +1,68 @@
1
- ---
2
- license: cc-by-nc-sa-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-sa-4.0
3
+ pipeline_tag: audio-classification
4
+ tags:
5
+ - autrainer
6
+ - audio
7
+ - orthoptera-tagging
8
+ - HearTheSpecies
9
+ ---
10
+
11
+ # InsectNet for the Biodiversity Exploratories
12
+ Model that tags audio files as belonging to one or more of 29 (t.b.d. below) prevalent Orthoptera species within the Biodiversity Exploratories.
13
+ We also have a Silence, Buzz, and Bird tag, but these predictions should be ignored and are only incorporated for the training.
14
+
15
+ # Installation
16
+
17
+ To use the model, you have to install autrainer, e.g. via pip:
18
+
19
+ ```
20
+ pip install autrainer
21
+ ```
22
+
23
+ This model has been trained and tested with autrainer version `0.6.0`.
24
+ For more information about autrainer, please refer to: https://autrainer.github.io/autrainer/index.html
25
+
26
+ # Usage
27
+
28
+ The model can be applied on all wav files present in a folder (`<data-root>`) and stored in another folder (`<output-root>`):
29
+
30
+ ```
31
+ autrainer inference hf:AlexanderGbd/InsectNetLocal -r <data-root> <output-root> -w 4 -s 4 -sr 96000
32
+ ```
33
+ , where `-w` is the window size in seconds, `-s` is the step size in seconds and `-sr` is the sampling rate.
34
+ For other possible inference settings and all usable parameters, please have a look at the autrainer documentation.
35
+ However, the above settings are recommended.
36
+
37
+ ## Training
38
+
39
+ ### Pretraining
40
+
41
+ TODO
42
+
43
+ ### Dataset
44
+
45
+ TODO
46
+
47
+
48
+ ### Features
49
+
50
+ The audio recordings were resampled to 96kHz, as we wanted to avoid losing too much frequency information from the species. Log-Mel spectrograms were then extracted using torchlibrosa.
51
+
52
+ ### Training process
53
+
54
+ The model has been trained for 30 epochs. At the end of each epoch, the model was evaluated on our validation set.
55
+ We release the state that achieved the best performance on this validation set.
56
+ All training hyperparameters can be found inside `conf/config.yaml` inside the model folder.
57
+
58
+
59
+ ## Evaluation
60
+
61
+ The performance on the test set reached a (macro) f1-score of 0.70.
62
+
63
+
64
+ ## Acknowledgments
65
+
66
+ TODO
67
+
68
+ Please acknowledge the work which produced the original model. We would appreciate an acknowledgment to autrainer.