Spaces:
Sleeping
Sleeping
| title: Frugal AI Challenge Submission | |
| emoji: π | |
| colorFrom: blue | |
| colorTo: green | |
| sdk: docker | |
| pinned: false | |
| ## π Audio classification | |
| ### Strategy for solving the problem | |
| To minimize energy consumption, we deliberately **chose not to use deep learning techniques** such as CNN-based spectrogram analysis, LSTM on raw audio signals or transformer models, which are generally **more computationally intensive**. | |
| Instead, a more **lightweight approach** was adopted: | |
| - Feature extraction from the audio signal (MFCCs and spectral contrast) | |
| - Training a simple machine learning model (decision tree) on these extracted features | |
| Potential Improvements (Not Yet Tested) | |
| - Hyperparameter tuning for better performance | |
| - Exploring alternative lightweight ML models, such as logistic regression or k-nearest neighbors | |
| - Feature extraction without Librosa, using NumPy directly to compute basic signal properties, further reducing dependencies and overhead. | |
| The model is exported from the notebook `notebooks\Audio_Challenge.ipynb` and saved as `model_audio.pkl` | |
| ## π Text classification | |
| ### Evaluate locally | |
| To evaluate the model locally, you can use the following command: | |
| ```bash | |
| python main.py --config config_evaluation_{model_name}.json | |
| ``` | |
| where `{model_name}` is either `distilBERT` or `embeddingML`. | |
| ### Models Description | |
| #### DistilBERT Model | |
| The model uses the `distilbert-base-uncased` model from the Hugging Face Transformers library, fine-tuned on the | |
| training dataset (see below). | |
| #### Embedding + ML Model | |
| The model uses a simple embedding layer followed by a classic ML model. Currently, the embedding layer is a simple | |
| TF-IDF vectorizer, and the ML model is a logistic regression. | |