Update README.md
Browse files
README.md
CHANGED
|
@@ -46,4 +46,94 @@ library_name: pyannote-audio
|
|
| 46 |
tags:
|
| 47 |
- music
|
| 48 |
---
|
| 49 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
tags:
|
| 47 |
- music
|
| 48 |
---
|
| 49 |
+
# SoundSlayerAI
|
| 50 |
+
|
| 51 |
+
SoundSlayerAI is an innovative project that focuses on music-related tasks and utilizes the power of the "pyannote-audio" library. This project aims to provide various functionalities for audio analysis and processing, making it easier to work with music datasets.
|
| 52 |
+
|
| 53 |
+
## Datasets
|
| 54 |
+
|
| 55 |
+
SoundSlayerAI makes use of the following datasets:
|
| 56 |
+
|
| 57 |
+
- Fhrozen/AudioSet2K22
|
| 58 |
+
- Chr0my/Epidemic_sounds
|
| 59 |
+
- ChristophSchuhmann/lyrics-index
|
| 60 |
+
- Cropinky/rap_lyrics_english
|
| 61 |
+
- tsterbak/eurovision-lyrics-1956-2023
|
| 62 |
+
- brunokreiner/genius-lyrics
|
| 63 |
+
- google/MusicCaps
|
| 64 |
+
- ccmusic-database/music_genre
|
| 65 |
+
- Hyeon2/riffusion-musiccaps-dataset
|
| 66 |
+
- SamAct/autotrain-data-musicprompt
|
| 67 |
+
- Chr0my/Epidemic_music
|
| 68 |
+
- juliensimon/autonlp-data-song-lyrics
|
| 69 |
+
- Datatang/North_American_English_Speech_Data_by_Mobile_Phone_and_PC
|
| 70 |
+
- Chr0my/freesound.org
|
| 71 |
+
- teticio/audio-diffusion-256
|
| 72 |
+
- KELONMYOSA/dusha_emotion_audio
|
| 73 |
+
- Ar4ikov/iemocap_audio_text_splitted
|
| 74 |
+
- flexthink/ljspeech
|
| 75 |
+
- mozilla-foundation/common_voice_13_0
|
| 76 |
+
- facebook/voxpopuli
|
| 77 |
+
- SocialGrep/one-million-reddit-jokes
|
| 78 |
+
- breadlicker45/human-midi-rlhf
|
| 79 |
+
- breadlicker45/midi-gpt-music-small
|
| 80 |
+
- projectlosangeles/Los-Angeles-MIDI-Dataset
|
| 81 |
+
- huggingartists/epic-rap-battles-of-history
|
| 82 |
+
- SocialGrep/one-million-reddit-confessions
|
| 83 |
+
- shahules786/prosocial-nsfw-reddit
|
| 84 |
+
- Thewillonline/reddit-sarcasm
|
| 85 |
+
- autoevaluate/autoeval-eval-futin__guess-vi-4200fb-2012366606
|
| 86 |
+
- lmsys/chatbot_arena_conversations
|
| 87 |
+
- mozilla-foundation/common_voice_11_0
|
| 88 |
+
- mozilla-foundation/common_voice_4_0
|
| 89 |
+
|
| 90 |
+
## Library
|
| 91 |
+
|
| 92 |
+
The core library used in this project is "pyannote-audio." This library provides a wide range of functionalities for audio analysis and processing, making it an excellent choice for working with music datasets. The "pyannote-audio" library offers a comprehensive set of tools and algorithms for tasks such as audio segmentation, speaker diarization, music transcription, and more.
|
| 93 |
+
|
| 94 |
+
## Metrics
|
| 95 |
+
|
| 96 |
+
To evaluate the performance of SoundSlayerAI, several metrics are employed, including:
|
| 97 |
+
|
| 98 |
+
- Accuracy
|
| 99 |
+
- Bertscore
|
| 100 |
+
- BLEU
|
| 101 |
+
- BLEURT
|
| 102 |
+
- Brier Score
|
| 103 |
+
- Character
|
| 104 |
+
|
| 105 |
+
These metrics help assess the effectiveness and accuracy of the implemented algorithms and models.
|
| 106 |
+
|
| 107 |
+
## Language
|
| 108 |
+
|
| 109 |
+
The SoundSlayerAI project primarily focuses on the English language. The datasets and models used in this project are optimized for English audio and text analysis tasks.
|
| 110 |
+
|
| 111 |
+
## Usage
|
| 112 |
+
|
| 113 |
+
To use SoundSlayerAI, follow these steps:
|
| 114 |
+
|
| 115 |
+
1. Install the required dependencies by running `pip install pyannote-audio`.
|
| 116 |
+
|
| 117 |
+
2. Import the necessary modules from the "pyannote.audio" package to access the desired functionalities.
|
| 118 |
+
|
| 119 |
+
3. Load the audio data or use the provided datasets to perform tasks such as audio segmentation, speaker diarization, music transcription, and more.
|
| 120 |
+
|
| 121 |
+
4. Apply the appropriate algorithms and models from the "pyannote.audio" library to process and analyze the audio data.
|
| 122 |
+
|
| 123 |
+
5. Evaluate the results using the specified metrics, such as accuracy, bertscore, BLEU, BLEURT, brier_score, and character.
|
| 124 |
+
|
| 125 |
+
6. Iterate and refine your approach to achieve the desired outcomes for your music-related tasks.
|
| 126 |
+
|
| 127 |
+
## License
|
| 128 |
+
|
| 129 |
+
SoundSlayerAI is released under the Openrail license. Please refer to the LICENSE file for more details.
|
| 130 |
+
|
| 131 |
+
## Contributions
|
| 132 |
+
|
| 133 |
+
Contributions to SoundSlayerAI are welcome! If you have any ideas, bug fixes, or enhancements, feel free to submit a pull request or open an issue on the GitHub repository.
|
| 134 |
+
|
| 135 |
+
## Contact
|
| 136 |
+
|
| 137 |
+
For any inquiries or questions regarding SoundSlayerAI, please reach out to the project maintainer at [insert email address].
|
| 138 |
+
|
| 139 |
+
Thank you for your interest in SoundSlayerAI!
|