kanishka089 commited on
Commit
95259e5
Β·
verified Β·
1 Parent(s): 154bac6

Upload 3 files

Browse files
Infant.ipynb ADDED
@@ -0,0 +1,724 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cells": [
3
+ {
4
+ "cell_type": "code",
5
+ "execution_count": 12,
6
+ "id": "9b5d89c1",
7
+ "metadata": {
8
+ "collapsed": true
9
+ },
10
+ "outputs": [
11
+ {
12
+ "name": "stdout",
13
+ "output_type": "stream",
14
+ "text": [
15
+ "Defaulting to user installation because normal site-packages is not writeable\n",
16
+ "Requirement already satisfied: tensorflow in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (2.17.0)\n",
17
+ "Requirement already satisfied: librosa in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (0.10.2.post1)\n",
18
+ "Requirement already satisfied: numpy in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (1.26.4)\n",
19
+ "Requirement already satisfied: pandas in c:\\programdata\\anaconda3\\lib\\site-packages (1.4.2)\n",
20
+ "Requirement already satisfied: matplotlib in c:\\programdata\\anaconda3\\lib\\site-packages (3.5.1)\n",
21
+ "Requirement already satisfied: scikit-learn in c:\\programdata\\anaconda3\\lib\\site-packages (1.0.2)\n",
22
+ "Requirement already satisfied: resampy in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (0.4.3)\n",
23
+ "Requirement already satisfied: xgboost in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (2.1.1)\n",
24
+ "Requirement already satisfied: tensorflow-intel==2.17.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow) (2.17.0)\n",
25
+ "Requirement already satisfied: tensorboard<2.18,>=2.17 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (2.17.0)\n",
26
+ "Requirement already satisfied: absl-py>=1.0.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (2.1.0)\n",
27
+ "Requirement already satisfied: flatbuffers>=24.3.25 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (24.3.25)\n",
28
+ "Requirement already satisfied: six>=1.12.0 in c:\\programdata\\anaconda3\\lib\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (1.16.0)\n",
29
+ "Requirement already satisfied: keras>=3.2.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (3.4.1)\n",
30
+ "Requirement already satisfied: packaging in c:\\programdata\\anaconda3\\lib\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (21.3)\n",
31
+ "Requirement already satisfied: ml-dtypes<0.5.0,>=0.3.1 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (0.4.0)\n",
32
+ "Requirement already satisfied: wrapt>=1.11.0 in c:\\programdata\\anaconda3\\lib\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (1.12.1)\n",
33
+ "Requirement already satisfied: termcolor>=1.1.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (2.4.0)\n",
34
+ "Requirement already satisfied: grpcio<2.0,>=1.24.3 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (1.65.4)\n",
35
+ "Requirement already satisfied: libclang>=13.0.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (18.1.1)\n",
36
+ "Requirement already satisfied: astunparse>=1.6.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (1.6.3)\n",
37
+ "Requirement already satisfied: gast!=0.5.0,!=0.5.1,!=0.5.2,>=0.2.1 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (0.6.0)\n",
38
+ "Requirement already satisfied: opt-einsum>=2.3.2 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (3.3.0)\n",
39
+ "Requirement already satisfied: requests<3,>=2.21.0 in c:\\programdata\\anaconda3\\lib\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (2.27.1)\n",
40
+ "Requirement already satisfied: typing-extensions>=3.6.6 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (4.12.2)\n",
41
+ "Requirement already satisfied: protobuf!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5,<5.0.0dev,>=3.20.3 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (4.25.4)\n",
42
+ "Requirement already satisfied: h5py>=3.10.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (3.11.0)\n",
43
+ "Requirement already satisfied: setuptools in c:\\programdata\\anaconda3\\lib\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (61.2.0)\n",
44
+ "Requirement already satisfied: google-pasta>=0.1.1 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (0.2.0)\n",
45
+ "Requirement already satisfied: tensorflow-io-gcs-filesystem>=0.23.1 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorflow-intel==2.17.0->tensorflow) (0.31.0)\n",
46
+ "Requirement already satisfied: pooch>=1.1 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from librosa) (1.8.2)\n",
47
+ "Requirement already satisfied: scipy>=1.2.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from librosa) (1.13.1)\n",
48
+ "Requirement already satisfied: lazy-loader>=0.1 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from librosa) (0.4)\n",
49
+ "Requirement already satisfied: numba>=0.51.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from librosa) (0.60.0)\n",
50
+ "Requirement already satisfied: soxr>=0.3.2 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from librosa) (0.4.0)\n",
51
+ "Requirement already satisfied: audioread>=2.1.9 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from librosa) (3.0.1)\n",
52
+ "Requirement already satisfied: decorator>=4.3.0 in c:\\programdata\\anaconda3\\lib\\site-packages (from librosa) (5.1.1)\n",
53
+ "Requirement already satisfied: soundfile>=0.12.1 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from librosa) (0.12.1)\n",
54
+ "Requirement already satisfied: msgpack>=1.0 in c:\\programdata\\anaconda3\\lib\\site-packages (from librosa) (1.0.2)\n",
55
+ "Requirement already satisfied: joblib>=0.14 in c:\\programdata\\anaconda3\\lib\\site-packages (from librosa) (1.1.0)\n",
56
+ "Requirement already satisfied: python-dateutil>=2.8.1 in c:\\programdata\\anaconda3\\lib\\site-packages (from pandas) (2.8.2)\n",
57
+ "Requirement already satisfied: pytz>=2020.1 in c:\\programdata\\anaconda3\\lib\\site-packages (from pandas) (2021.3)\n",
58
+ "Requirement already satisfied: cycler>=0.10 in c:\\programdata\\anaconda3\\lib\\site-packages (from matplotlib) (0.11.0)\n",
59
+ "Requirement already satisfied: pillow>=6.2.0 in c:\\programdata\\anaconda3\\lib\\site-packages (from matplotlib) (9.0.1)\n",
60
+ "Requirement already satisfied: kiwisolver>=1.0.1 in c:\\programdata\\anaconda3\\lib\\site-packages (from matplotlib) (1.3.2)\n",
61
+ "Requirement already satisfied: fonttools>=4.22.0 in c:\\programdata\\anaconda3\\lib\\site-packages (from matplotlib) (4.25.0)\n",
62
+ "Requirement already satisfied: pyparsing>=2.2.1 in c:\\programdata\\anaconda3\\lib\\site-packages (from matplotlib) (3.0.4)\n",
63
+ "Requirement already satisfied: threadpoolctl>=2.0.0 in c:\\programdata\\anaconda3\\lib\\site-packages (from scikit-learn) (2.2.0)\n",
64
+ "Requirement already satisfied: wheel<1.0,>=0.23.0 in c:\\programdata\\anaconda3\\lib\\site-packages (from astunparse>=1.6.0->tensorflow-intel==2.17.0->tensorflow) (0.37.1)\n",
65
+ "Requirement already satisfied: rich in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from keras>=3.2.0->tensorflow-intel==2.17.0->tensorflow) (13.7.1)\n",
66
+ "Requirement already satisfied: namex in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from keras>=3.2.0->tensorflow-intel==2.17.0->tensorflow) (0.0.8)\n",
67
+ "Requirement already satisfied: optree in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from keras>=3.2.0->tensorflow-intel==2.17.0->tensorflow) (0.12.1)\n",
68
+ "Requirement already satisfied: llvmlite<0.44,>=0.43.0dev0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from numba>=0.51.0->librosa) (0.43.0)\n",
69
+ "Requirement already satisfied: platformdirs>=2.5.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from pooch>=1.1->librosa) (4.2.2)\n",
70
+ "Requirement already satisfied: charset-normalizer~=2.0.0 in c:\\programdata\\anaconda3\\lib\\site-packages (from requests<3,>=2.21.0->tensorflow-intel==2.17.0->tensorflow) (2.0.4)\n",
71
+ "Requirement already satisfied: certifi>=2017.4.17 in c:\\programdata\\anaconda3\\lib\\site-packages (from requests<3,>=2.21.0->tensorflow-intel==2.17.0->tensorflow) (2021.10.8)\n",
72
+ "Requirement already satisfied: urllib3<1.27,>=1.21.1 in c:\\programdata\\anaconda3\\lib\\site-packages (from requests<3,>=2.21.0->tensorflow-intel==2.17.0->tensorflow) (1.26.9)\n",
73
+ "Requirement already satisfied: idna<4,>=2.5 in c:\\programdata\\anaconda3\\lib\\site-packages (from requests<3,>=2.21.0->tensorflow-intel==2.17.0->tensorflow) (3.3)\n",
74
+ "Requirement already satisfied: cffi>=1.0 in c:\\programdata\\anaconda3\\lib\\site-packages (from soundfile>=0.12.1->librosa) (1.15.0)\n",
75
+ "Requirement already satisfied: pycparser in c:\\programdata\\anaconda3\\lib\\site-packages (from cffi>=1.0->soundfile>=0.12.1->librosa) (2.21)\n",
76
+ "Requirement already satisfied: markdown>=2.6.8 in c:\\programdata\\anaconda3\\lib\\site-packages (from tensorboard<2.18,>=2.17->tensorflow-intel==2.17.0->tensorflow) (3.3.4)\n",
77
+ "Requirement already satisfied: tensorboard-data-server<0.8.0,>=0.7.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from tensorboard<2.18,>=2.17->tensorflow-intel==2.17.0->tensorflow) (0.7.2)\n",
78
+ "Requirement already satisfied: werkzeug>=1.0.1 in c:\\programdata\\anaconda3\\lib\\site-packages (from tensorboard<2.18,>=2.17->tensorflow-intel==2.17.0->tensorflow) (2.0.3)\n",
79
+ "Requirement already satisfied: markdown-it-py>=2.2.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from rich->keras>=3.2.0->tensorflow-intel==2.17.0->tensorflow) (3.0.0)\n",
80
+ "Requirement already satisfied: pygments<3.0.0,>=2.13.0 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from rich->keras>=3.2.0->tensorflow-intel==2.17.0->tensorflow) (2.18.0)\n",
81
+ "Requirement already satisfied: mdurl~=0.1 in c:\\users\\msi\\appdata\\roaming\\python\\python39\\site-packages (from markdown-it-py>=2.2.0->rich->keras>=3.2.0->tensorflow-intel==2.17.0->tensorflow) (0.1.2)\n",
82
+ "Note: you may need to restart the kernel to use updated packages.\n"
83
+ ]
84
+ }
85
+ ],
86
+ "source": [
87
+ "pip install tensorflow librosa numpy pandas matplotlib scikit-learn resampy xgboost"
88
+ ]
89
+ },
90
+ {
91
+ "cell_type": "code",
92
+ "execution_count": 8,
93
+ "id": "2317d7f3",
94
+ "metadata": {},
95
+ "outputs": [
96
+ {
97
+ "name": "stderr",
98
+ "output_type": "stream",
99
+ "text": [
100
+ "C:\\Users\\MSI\\AppData\\Local\\Temp\\ipykernel_2700\\1497878668.py:24: UserWarning: PySoundFile failed. Trying audioread instead.\n",
101
+ " audio, sample_rate = librosa.load(file_name, res_type='kaiser_fast')\n"
102
+ ]
103
+ },
104
+ {
105
+ "name": "stdout",
106
+ "output_type": "stream",
107
+ "text": [
108
+ "Error encountered while parsing file: soundclips\\discomfort\\Minta Gendong AUD-20150509-WA0000.wav, \n",
109
+ "Feature extraction failed for file: soundclips\\discomfort\\Minta Gendong AUD-20150509-WA0000.wav\n",
110
+ "Error encountered while parsing file: soundclips\\discomfort\\recordgntipopok.wav, \n",
111
+ "Feature extraction failed for file: soundclips\\discomfort\\recordgntipopok.wav\n",
112
+ "Error encountered while parsing file: soundclips\\hungry\\Lapar AUD-20150509-WA0001.wav, \n",
113
+ "Feature extraction failed for file: soundclips\\hungry\\Lapar AUD-20150509-WA0001.wav\n",
114
+ "Error encountered while parsing file: soundclips\\hungry\\record-baby-1 cari puting.wav, \n",
115
+ "Feature extraction failed for file: soundclips\\hungry\\record-baby-1 cari puting.wav\n",
116
+ "Error encountered while parsing file: soundclips\\hungry\\record-baby2 puting dilepas.wav, \n",
117
+ "Feature extraction failed for file: soundclips\\hungry\\record-baby2 puting dilepas.wav\n",
118
+ "Error encountered while parsing file: soundclips\\tired\\Bangun Tidur AUD-20150509-WA0002.wav, \n",
119
+ "Feature extraction failed for file: soundclips\\tired\\Bangun Tidur AUD-20150509-WA0002.wav\n"
120
+ ]
121
+ }
122
+ ],
123
+ "source": [
124
+ "import os\n",
125
+ "import numpy as np\n",
126
+ "import librosa\n",
127
+ "import pandas as pd\n",
128
+ "from sklearn.model_selection import train_test_split\n",
129
+ "from sklearn.preprocessing import LabelEncoder\n",
130
+ "from tensorflow.keras.utils import to_categorical\n",
131
+ "import xgboost as xgb\n",
132
+ "from sklearn.metrics import accuracy_score\n",
133
+ "from tensorflow.keras.models import Sequential\n",
134
+ "from tensorflow.keras.layers import Dense, Dropout, Activation\n",
135
+ "from tensorflow.keras.optimizers import Adam\n",
136
+ "from tensorflow.keras.callbacks import ModelCheckpoint\n",
137
+ "\n",
138
+ "# Path to the dataset\n",
139
+ "dataset_path = 'soundclips'\n",
140
+ "\n",
141
+ "# List of categories\n",
142
+ "categories = ['belly_pain', 'burping', 'discomfort', 'hungry', 'tired']\n",
143
+ "\n",
144
+ "# Function to extract features from audio files\n",
145
+ "def extract_features(file_name):\n",
146
+ " try:\n",
147
+ " audio, sample_rate = librosa.load(file_name, res_type='kaiser_fast')\n",
148
+ " mfccs = librosa.feature.mfcc(y=audio, sr=sample_rate, n_mfcc=40)\n",
149
+ " mfccs_scaled = np.mean(mfccs.T, axis=0)\n",
150
+ " \n",
151
+ " return mfccs_scaled\n",
152
+ " except Exception as e:\n",
153
+ " print(f\"Error encountered while parsing file: {file_name}, {e}\")\n",
154
+ " return None\n",
155
+ "\n",
156
+ "# Create DataFrame to hold features and labels\n",
157
+ "features = []\n",
158
+ "labels = []\n",
159
+ "\n",
160
+ "# Iterate through each category\n",
161
+ "for category in categories:\n",
162
+ " category_path = os.path.join(dataset_path, category)\n",
163
+ " if not os.path.exists(category_path):\n",
164
+ " print(f\"Directory does not exist: {category_path}\")\n",
165
+ " continue\n",
166
+ " \n",
167
+ " for file in os.listdir(category_path):\n",
168
+ " file_path = os.path.join(category_path, file)\n",
169
+ " data = extract_features(file_path)\n",
170
+ " if data is not None and len(data) > 0:\n",
171
+ " features.append(data)\n",
172
+ " labels.append(category)\n",
173
+ " else:\n",
174
+ " print(f\"Feature extraction failed for file: {file_path}\")\n",
175
+ "\n",
176
+ "# Convert to numpy arrays\n",
177
+ "features = np.array(features)\n",
178
+ "labels = np.array(labels)\n",
179
+ "\n",
180
+ "# Check if features array is empty\n",
181
+ "if features.size == 0:\n",
182
+ " raise ValueError(\"No features extracted. Please check the dataset and ensure audio files are present and readable.\")\n",
183
+ "\n",
184
+ "# Encode the labels\n",
185
+ "le = LabelEncoder()\n",
186
+ "labels_encoded = le.fit_transform(labels)\n",
187
+ "labels_categorical = to_categorical(labels_encoded)\n",
188
+ "\n",
189
+ "# Split the dataset\n",
190
+ "X_train, X_test, y_train, y_test = train_test_split(features, labels_categorical, test_size=0.2, random_state=42)\n"
191
+ ]
192
+ },
193
+ {
194
+ "cell_type": "code",
195
+ "execution_count": 9,
196
+ "id": "955d889d",
197
+ "metadata": {},
198
+ "outputs": [
199
+ {
200
+ "name": "stdout",
201
+ "output_type": "stream",
202
+ "text": [
203
+ "Epoch 1/100\n"
204
+ ]
205
+ },
206
+ {
207
+ "name": "stderr",
208
+ "output_type": "stream",
209
+ "text": [
210
+ "C:\\Users\\MSI\\AppData\\Roaming\\Python\\Python39\\site-packages\\keras\\src\\layers\\core\\dense.py:87: UserWarning: Do not pass an `input_shape`/`input_dim` argument to a layer. When using Sequential models, prefer using an `Input(shape)` object as the first layer in the model instead.\n",
211
+ " super().__init__(activity_regularizer=activity_regularizer, **kwargs)\n"
212
+ ]
213
+ },
214
+ {
215
+ "name": "stdout",
216
+ "output_type": "stream",
217
+ "text": [
218
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m9s\u001b[0m 770ms/step - accuracy: 0.1250 - loss: 92.2737\n",
219
+ "Epoch 1: val_loss improved from inf to 11.10916, saving model to audio_classification.keras\n",
220
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m1s\u001b[0m 14ms/step - accuracy: 0.2303 - loss: 55.0404 - val_accuracy: 0.7282 - val_loss: 11.1092\n",
221
+ "Epoch 2/100\n",
222
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 24ms/step - accuracy: 0.7500 - loss: 10.0197\n",
223
+ "Epoch 2: val_loss improved from 11.10916 to 8.67415, saving model to audio_classification.keras\n",
224
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.7144 - loss: 12.5967 - val_accuracy: 0.7282 - val_loss: 8.6742\n",
225
+ "Epoch 3/100\n",
226
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.6875 - loss: 17.6288\n",
227
+ "Epoch 3: val_loss improved from 8.67415 to 4.57907, saving model to audio_classification.keras\n",
228
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6595 - loss: 14.6687 - val_accuracy: 0.7282 - val_loss: 4.5791\n",
229
+ "Epoch 4/100\n",
230
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.6875 - loss: 7.3426\n",
231
+ "Epoch 4: val_loss improved from 4.57907 to 3.37468, saving model to audio_classification.keras\n",
232
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6161 - loss: 8.2100 - val_accuracy: 0.7282 - val_loss: 3.3747\n",
233
+ "Epoch 5/100\n",
234
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.5938 - loss: 6.5427\n",
235
+ "Epoch 5: val_loss improved from 3.37468 to 2.81896, saving model to audio_classification.keras\n",
236
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6420 - loss: 6.6541 - val_accuracy: 0.7282 - val_loss: 2.8190\n",
237
+ "Epoch 6/100\n",
238
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.7500 - loss: 4.3348\n",
239
+ "Epoch 6: val_loss improved from 2.81896 to 2.11174, saving model to audio_classification.keras\n",
240
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6473 - loss: 5.6300 - val_accuracy: 0.7282 - val_loss: 2.1117\n",
241
+ "Epoch 7/100\n",
242
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━���━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.5312 - loss: 8.7231\n",
243
+ "Epoch 7: val_loss improved from 2.11174 to 1.68834, saving model to audio_classification.keras\n",
244
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.5790 - loss: 5.9270 - val_accuracy: 0.7282 - val_loss: 1.6883\n",
245
+ "Epoch 8/100\n",
246
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.5938 - loss: 6.1949\n",
247
+ "Epoch 8: val_loss improved from 1.68834 to 1.39033, saving model to audio_classification.keras\n",
248
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6495 - loss: 4.2499 - val_accuracy: 0.7282 - val_loss: 1.3903\n",
249
+ "Epoch 9/100\n",
250
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 24ms/step - accuracy: 0.7812 - loss: 2.6101\n",
251
+ "Epoch 9: val_loss improved from 1.39033 to 1.19555, saving model to audio_classification.keras\n",
252
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6688 - loss: 3.7850 - val_accuracy: 0.7282 - val_loss: 1.1956\n",
253
+ "Epoch 10/100\n",
254
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 22ms/step - accuracy: 0.5938 - loss: 3.2523\n",
255
+ "Epoch 10: val_loss improved from 1.19555 to 1.19277, saving model to audio_classification.keras\n",
256
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6258 - loss: 3.3801 - val_accuracy: 0.7282 - val_loss: 1.1928\n",
257
+ "Epoch 11/100\n",
258
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 22ms/step - accuracy: 0.6250 - loss: 2.2360\n",
259
+ "Epoch 11: val_loss improved from 1.19277 to 1.07271, saving model to audio_classification.keras\n",
260
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6525 - loss: 3.1927 - val_accuracy: 0.7282 - val_loss: 1.0727\n",
261
+ "Epoch 12/100\n",
262
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.6562 - loss: 4.2252\n",
263
+ "Epoch 12: val_loss improved from 1.07271 to 1.03591, saving model to audio_classification.keras\n",
264
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6514 - loss: 3.0835 - val_accuracy: 0.7282 - val_loss: 1.0359\n",
265
+ "Epoch 13/100\n",
266
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.6562 - loss: 3.3775\n",
267
+ "Epoch 13: val_loss did not improve from 1.03591\n",
268
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.6636 - loss: 2.7042 - val_accuracy: 0.7282 - val_loss: 1.0713\n",
269
+ "Epoch 14/100\n",
270
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.6562 - loss: 3.2629\n",
271
+ "Epoch 14: val_loss did not improve from 1.03591\n",
272
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.6326 - loss: 2.5686 - val_accuracy: 0.7282 - val_loss: 1.0756\n",
273
+ "Epoch 15/100\n",
274
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 26ms/step - accuracy: 0.6562 - loss: 3.7654\n",
275
+ "Epoch 15: val_loss improved from 1.03591 to 1.01481, saving model to audio_classification.keras\n",
276
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6663 - loss: 2.3855 - val_accuracy: 0.7282 - val_loss: 1.0148\n",
277
+ "Epoch 16/100\n",
278
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.6250 - loss: 2.1722\n",
279
+ "Epoch 16: val_loss improved from 1.01481 to 1.00236, saving model to audio_classification.keras\n",
280
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6389 - loss: 1.9125 - val_accuracy: 0.7282 - val_loss: 1.0024\n",
281
+ "Epoch 17/100\n",
282
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.4062 - loss: 2.8053\n",
283
+ "Epoch 17: val_loss improved from 1.00236 to 1.00158, saving model to audio_classification.keras\n",
284
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.5887 - loss: 2.0436 - val_accuracy: 0.7282 - val_loss: 1.0016\n",
285
+ "Epoch 18/100\n",
286
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.6562 - loss: 2.1894\n",
287
+ "Epoch 18: val_loss did not improve from 1.00158\n",
288
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.6451 - loss: 1.7902 - val_accuracy: 0.7282 - val_loss: 1.0104\n",
289
+ "Epoch 19/100\n",
290
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 24ms/step - accuracy: 0.6875 - loss: 1.2992\n",
291
+ "Epoch 19: val_loss improved from 1.00158 to 0.98988, saving model to audio_classification.keras\n",
292
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.6305 - loss: 1.8078 - val_accuracy: 0.7282 - val_loss: 0.9899\n",
293
+ "Epoch 20/100\n",
294
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 25ms/step - accuracy: 0.5938 - loss: 1.9796\n",
295
+ "Epoch 20: val_loss did not improve from 0.98988\n",
296
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.6325 - loss: 1.6304 - val_accuracy: 0.7282 - val_loss: 1.0112\n",
297
+ "Epoch 21/100\n",
298
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.7500 - loss: 1.0891\n",
299
+ "Epoch 21: val_loss did not improve from 0.98988\n",
300
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7144 - loss: 1.2691 - val_accuracy: 0.7282 - val_loss: 1.0299\n",
301
+ "Epoch 22/100\n",
302
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 22ms/step - accuracy: 0.6875 - loss: 1.4959\n",
303
+ "Epoch 22: val_loss did not improve from 0.98988\n",
304
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.6914 - loss: 1.5631 - val_accuracy: 0.7282 - val_loss: 1.0585\n",
305
+ "Epoch 23/100\n",
306
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.7500 - loss: 1.3707\n",
307
+ "Epoch 23: val_loss did not improve from 0.98988\n",
308
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7303 - loss: 1.2198 - val_accuracy: 0.7282 - val_loss: 1.0398\n",
309
+ "Epoch 24/100\n",
310
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.6562 - loss: 2.6802\n",
311
+ "Epoch 24: val_loss did not improve from 0.98988\n"
312
+ ]
313
+ },
314
+ {
315
+ "name": "stdout",
316
+ "output_type": "stream",
317
+ "text": [
318
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.6958 - loss: 1.6243 - val_accuracy: 0.7282 - val_loss: 1.0526\n",
319
+ "Epoch 25/100\n",
320
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.6250 - loss: 1.3346\n",
321
+ "Epoch 25: val_loss did not improve from 0.98988\n",
322
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.6771 - loss: 1.2211 - val_accuracy: 0.7282 - val_loss: 1.0524\n",
323
+ "Epoch 26/100\n",
324
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.5312 - loss: 2.1067\n",
325
+ "Epoch 26: val_loss did not improve from 0.98988\n",
326
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.6778 - loss: 1.6074 - val_accuracy: 0.7282 - val_loss: 1.0376\n",
327
+ "Epoch 27/100\n",
328
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 25ms/step - accuracy: 0.5312 - loss: 1.5559\n",
329
+ "Epoch 27: val_loss did not improve from 0.98988\n",
330
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.6782 - loss: 1.4016 - val_accuracy: 0.7282 - val_loss: 1.0039\n",
331
+ "Epoch 28/100\n",
332
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8125 - loss: 0.6590\n",
333
+ "Epoch 28: val_loss did not improve from 0.98988\n",
334
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7852 - loss: 0.9609 - val_accuracy: 0.7282 - val_loss: 0.9953\n",
335
+ "Epoch 29/100\n",
336
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.5938 - loss: 1.5004\n",
337
+ "Epoch 29: val_loss did not improve from 0.98988\n",
338
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7568 - loss: 1.1752 - val_accuracy: 0.7282 - val_loss: 1.0424\n",
339
+ "Epoch 30/100\n",
340
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.6250 - loss: 1.5044\n",
341
+ "Epoch 30: val_loss did not improve from 0.98988\n",
342
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7052 - loss: 1.1279 - val_accuracy: 0.7282 - val_loss: 1.0329\n",
343
+ "Epoch 31/100\n",
344
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.5938 - loss: 1.9206\n",
345
+ "Epoch 31: val_loss did not improve from 0.98988\n",
346
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7117 - loss: 1.3123 - val_accuracy: 0.7282 - val_loss: 1.0477\n",
347
+ "Epoch 32/100\n",
348
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8438 - loss: 0.7891\n",
349
+ "Epoch 32: val_loss did not improve from 0.98988\n",
350
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7742 - loss: 0.9474 - val_accuracy: 0.7282 - val_loss: 1.0726\n",
351
+ "Epoch 33/100\n",
352
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.8750 - loss: 0.7086\n",
353
+ "Epoch 33: val_loss did not improve from 0.98988\n",
354
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7596 - loss: 1.1175 - val_accuracy: 0.7282 - val_loss: 1.0487\n",
355
+ "Epoch 34/100\n",
356
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8750 - loss: 0.6791\n",
357
+ "Epoch 34: val_loss did not improve from 0.98988\n",
358
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7631 - loss: 1.0853 - val_accuracy: 0.7282 - val_loss: 1.0172\n",
359
+ "Epoch 35/100\n",
360
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.7188 - loss: 0.9913\n",
361
+ "Epoch 35: val_loss did not improve from 0.98988\n",
362
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7497 - loss: 1.0620 - val_accuracy: 0.7282 - val_loss: 0.9953\n",
363
+ "Epoch 36/100\n",
364
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.8125 - loss: 0.6980\n",
365
+ "Epoch 36: val_loss did not improve from 0.98988\n",
366
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7718 - loss: 0.9211 - val_accuracy: 0.7282 - val_loss: 1.0045\n",
367
+ "Epoch 37/100\n",
368
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8125 - loss: 0.6965\n",
369
+ "Epoch 37: val_loss did not improve from 0.98988\n",
370
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7637 - loss: 0.8684 - val_accuracy: 0.7282 - val_loss: 1.0102\n",
371
+ "Epoch 38/100\n",
372
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.9062 - loss: 0.6313\n",
373
+ "Epoch 38: val_loss did not improve from 0.98988\n",
374
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8126 - loss: 0.8493 - val_accuracy: 0.7282 - val_loss: 0.9966\n",
375
+ "Epoch 39/100\n",
376
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.6875 - loss: 1.1845\n",
377
+ "Epoch 39: val_loss did not improve from 0.98988\n",
378
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7415 - loss: 1.0281 - val_accuracy: 0.7282 - val_loss: 1.0113\n",
379
+ "Epoch 40/100\n",
380
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.8125 - loss: 1.2241\n",
381
+ "Epoch 40: val_loss did not improve from 0.98988\n",
382
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7883 - loss: 1.0743 - val_accuracy: 0.7282 - val_loss: 1.0338\n",
383
+ "Epoch 41/100\n",
384
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8438 - loss: 0.8787\n",
385
+ "Epoch 41: val_loss did not improve from 0.98988\n",
386
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8031 - loss: 0.8833 - val_accuracy: 0.7282 - val_loss: 1.0117\n",
387
+ "Epoch 42/100\n",
388
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.7188 - loss: 1.1767\n",
389
+ "Epoch 42: val_loss did not improve from 0.98988\n",
390
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7720 - loss: 0.9778 - val_accuracy: 0.7282 - val_loss: 0.9994\n",
391
+ "Epoch 43/100\n",
392
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.8125 - loss: 0.6985\n",
393
+ "Epoch 43: val_loss did not improve from 0.98988\n",
394
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7837 - loss: 0.9230 - val_accuracy: 0.7282 - val_loss: 1.0094\n",
395
+ "Epoch 44/100\n",
396
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8750 - loss: 0.4748\n",
397
+ "Epoch 44: val_loss did not improve from 0.98988\n",
398
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7915 - loss: 0.9233 - val_accuracy: 0.7282 - val_loss: 1.0357\n",
399
+ "Epoch 45/100\n",
400
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.8750 - loss: 0.6965\n",
401
+ "Epoch 45: val_loss did not improve from 0.98988\n",
402
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8135 - loss: 0.8862 - val_accuracy: 0.7282 - val_loss: 1.0525\n",
403
+ "Epoch 46/100\n",
404
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.7500 - loss: 0.7676\n",
405
+ "Epoch 46: val_loss did not improve from 0.98988\n",
406
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7945 - loss: 0.8852 - val_accuracy: 0.7282 - val_loss: 1.0297\n",
407
+ "Epoch 47/100\n",
408
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.8438 - loss: 0.5471\n",
409
+ "Epoch 47: val_loss did not improve from 0.98988\n",
410
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7881 - loss: 0.8314 - val_accuracy: 0.7282 - val_loss: 0.9968\n",
411
+ "Epoch 48/100\n",
412
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.8750 - loss: 0.5049\n",
413
+ "Epoch 48: val_loss did not improve from 0.98988\n",
414
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8032 - loss: 0.8580 - val_accuracy: 0.7282 - val_loss: 0.9949\n",
415
+ "Epoch 49/100\n",
416
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.7188 - loss: 0.9730\n",
417
+ "Epoch 49: val_loss did not improve from 0.98988\n",
418
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7768 - loss: 0.9519 - val_accuracy: 0.7282 - val_loss: 0.9950\n",
419
+ "Epoch 50/100\n"
420
+ ]
421
+ },
422
+ {
423
+ "name": "stdout",
424
+ "output_type": "stream",
425
+ "text": [
426
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.7812 - loss: 1.1724\n",
427
+ "Epoch 50: val_loss improved from 0.98988 to 0.98715, saving model to audio_classification.keras\n",
428
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.7758 - loss: 1.0263 - val_accuracy: 0.7282 - val_loss: 0.9871\n",
429
+ "Epoch 51/100\n",
430
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.7500 - loss: 1.0490\n",
431
+ "Epoch 51: val_loss improved from 0.98715 to 0.98223, saving model to audio_classification.keras\n",
432
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.7547 - loss: 1.0155 - val_accuracy: 0.7282 - val_loss: 0.9822\n",
433
+ "Epoch 52/100\n",
434
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.7500 - loss: 1.2764\n",
435
+ "Epoch 52: val_loss did not improve from 0.98223\n",
436
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7949 - loss: 0.9619 - val_accuracy: 0.7282 - val_loss: 0.9835\n",
437
+ "Epoch 53/100\n",
438
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.7500 - loss: 1.3644\n",
439
+ "Epoch 53: val_loss improved from 0.98223 to 0.97895, saving model to audio_classification.keras\n",
440
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.7872 - loss: 1.0066 - val_accuracy: 0.7282 - val_loss: 0.9789\n",
441
+ "Epoch 54/100\n",
442
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.8125 - loss: 0.7706\n",
443
+ "Epoch 54: val_loss improved from 0.97895 to 0.96493, saving model to audio_classification.keras\n",
444
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.8054 - loss: 0.8880 - val_accuracy: 0.7282 - val_loss: 0.9649\n",
445
+ "Epoch 55/100\n",
446
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.6875 - loss: 1.0864\n",
447
+ "Epoch 55: val_loss improved from 0.96493 to 0.95673, saving model to audio_classification.keras\n",
448
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.7823 - loss: 0.8469 - val_accuracy: 0.7282 - val_loss: 0.9567\n",
449
+ "Epoch 56/100\n",
450
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8125 - loss: 1.0572\n",
451
+ "Epoch 56: val_loss improved from 0.95673 to 0.95037, saving model to audio_classification.keras\n",
452
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.7951 - loss: 0.8759 - val_accuracy: 0.7282 - val_loss: 0.9504\n",
453
+ "Epoch 57/100\n",
454
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.7812 - loss: 1.0491\n",
455
+ "Epoch 57: val_loss did not improve from 0.95037\n",
456
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8068 - loss: 0.8893 - val_accuracy: 0.7282 - val_loss: 0.9679\n",
457
+ "Epoch 58/100\n",
458
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 22ms/step - accuracy: 0.8438 - loss: 0.8212\n",
459
+ "Epoch 58: val_loss did not improve from 0.95037\n",
460
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8067 - loss: 0.8464 - val_accuracy: 0.7282 - val_loss: 0.9785\n",
461
+ "Epoch 59/100\n",
462
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.7812 - loss: 0.8074\n",
463
+ "Epoch 59: val_loss did not improve from 0.95037\n",
464
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8110 - loss: 0.7783 - val_accuracy: 0.7282 - val_loss: 0.9657\n",
465
+ "Epoch 60/100\n",
466
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.9375 - loss: 0.5068\n",
467
+ "Epoch 60: val_loss did not improve from 0.95037\n",
468
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8376 - loss: 0.7360 - val_accuracy: 0.7282 - val_loss: 0.9663\n",
469
+ "Epoch 61/100\n",
470
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.7500 - loss: 0.8686\n",
471
+ "Epoch 61: val_loss did not improve from 0.95037\n",
472
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8111 - loss: 0.7876 - val_accuracy: 0.7282 - val_loss: 0.9632\n",
473
+ "Epoch 62/100\n",
474
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8125 - loss: 0.7523\n",
475
+ "Epoch 62: val_loss did not improve from 0.95037\n",
476
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8128 - loss: 0.7946 - val_accuracy: 0.7282 - val_loss: 0.9733\n",
477
+ "Epoch 63/100\n",
478
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.7812 - loss: 0.8872\n",
479
+ "Epoch 63: val_loss did not improve from 0.95037\n",
480
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8017 - loss: 0.7635 - val_accuracy: 0.7282 - val_loss: 0.9592\n",
481
+ "Epoch 64/100\n",
482
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 22ms/step - accuracy: 0.6562 - loss: 1.1472\n",
483
+ "Epoch 64: val_loss improved from 0.95037 to 0.94993, saving model to audio_classification.keras\n",
484
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.7825 - loss: 0.8701 - val_accuracy: 0.7282 - val_loss: 0.9499\n",
485
+ "Epoch 65/100\n",
486
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.7188 - loss: 0.8833\n",
487
+ "Epoch 65: val_loss improved from 0.94993 to 0.94609, saving model to audio_classification.keras\n",
488
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.7906 - loss: 0.8554 - val_accuracy: 0.7282 - val_loss: 0.9461\n",
489
+ "Epoch 66/100\n",
490
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.7188 - loss: 0.9959\n",
491
+ "Epoch 66: val_loss improved from 0.94609 to 0.92950, saving model to audio_classification.keras\n",
492
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.7973 - loss: 0.8486 - val_accuracy: 0.7282 - val_loss: 0.9295\n",
493
+ "Epoch 67/100\n",
494
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.6562 - loss: 1.2690\n",
495
+ "Epoch 67: val_loss did not improve from 0.92950\n",
496
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7939 - loss: 0.8727 - val_accuracy: 0.7282 - val_loss: 0.9404\n",
497
+ "Epoch 68/100\n",
498
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.8125 - loss: 0.7729\n",
499
+ "Epoch 68: val_loss did not improve from 0.92950\n",
500
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8011 - loss: 0.8859 - val_accuracy: 0.7282 - val_loss: 0.9327\n",
501
+ "Epoch 69/100\n",
502
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.7812 - loss: 0.7983\n",
503
+ "Epoch 69: val_loss did not improve from 0.92950\n",
504
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8240 - loss: 0.7458 - val_accuracy: 0.7282 - val_loss: 0.9319\n",
505
+ "Epoch 70/100\n",
506
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 22ms/step - accuracy: 0.8125 - loss: 0.7055\n",
507
+ "Epoch 70: val_loss improved from 0.92950 to 0.92735, saving model to audio_classification.keras\n",
508
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.8036 - loss: 0.7931 - val_accuracy: 0.7282 - val_loss: 0.9274\n",
509
+ "Epoch 71/100\n",
510
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.7812 - loss: 0.8137\n",
511
+ "Epoch 71: val_loss improved from 0.92735 to 0.92599, saving model to audio_classification.keras\n",
512
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.8023 - loss: 0.7893 - val_accuracy: 0.7282 - val_loss: 0.9260\n",
513
+ "Epoch 72/100\n",
514
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8438 - loss: 0.6523\n",
515
+ "Epoch 72: val_loss improved from 0.92599 to 0.92060, saving model to audio_classification.keras\n",
516
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.8132 - loss: 0.7949 - val_accuracy: 0.7282 - val_loss: 0.9206\n",
517
+ "Epoch 73/100\n",
518
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 22ms/step - accuracy: 0.8125 - loss: 0.8290\n",
519
+ "Epoch 73: val_loss did not improve from 0.92060\n",
520
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 4ms/step - accuracy: 0.8300 - loss: 0.7700 - val_accuracy: 0.7282 - val_loss: 0.9273\n"
521
+ ]
522
+ },
523
+ {
524
+ "name": "stdout",
525
+ "output_type": "stream",
526
+ "text": [
527
+ "Epoch 74/100\n",
528
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.8438 - loss: 0.7377\n",
529
+ "Epoch 74: val_loss did not improve from 0.92060\n",
530
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8034 - loss: 0.8272 - val_accuracy: 0.7282 - val_loss: 0.9297\n",
531
+ "Epoch 75/100\n",
532
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.8125 - loss: 0.8221\n",
533
+ "Epoch 75: val_loss did not improve from 0.92060\n",
534
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8141 - loss: 0.8071 - val_accuracy: 0.7282 - val_loss: 0.9264\n",
535
+ "Epoch 76/100\n",
536
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.8438 - loss: 0.6241\n",
537
+ "Epoch 76: val_loss did not improve from 0.92060\n",
538
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8313 - loss: 0.7314 - val_accuracy: 0.7282 - val_loss: 0.9241\n",
539
+ "Epoch 77/100\n",
540
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8125 - loss: 0.7567\n",
541
+ "Epoch 77: val_loss improved from 0.92060 to 0.91963, saving model to audio_classification.keras\n",
542
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.8189 - loss: 0.7449 - val_accuracy: 0.7282 - val_loss: 0.9196\n",
543
+ "Epoch 78/100\n",
544
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.7812 - loss: 0.7788\n",
545
+ "Epoch 78: val_loss did not improve from 0.91963\n",
546
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8098 - loss: 0.7512 - val_accuracy: 0.7282 - val_loss: 0.9206\n",
547
+ "Epoch 79/100\n",
548
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.7188 - loss: 0.9284\n",
549
+ "Epoch 79: val_loss did not improve from 0.91963\n",
550
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7784 - loss: 0.8218 - val_accuracy: 0.7282 - val_loss: 0.9208\n",
551
+ "Epoch 80/100\n",
552
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8438 - loss: 0.5897\n",
553
+ "Epoch 80: val_loss improved from 0.91963 to 0.91612, saving model to audio_classification.keras\n",
554
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.7774 - loss: 0.7675 - val_accuracy: 0.7282 - val_loss: 0.9161\n",
555
+ "Epoch 81/100\n",
556
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.7500 - loss: 0.8273\n",
557
+ "Epoch 81: val_loss improved from 0.91612 to 0.91471, saving model to audio_classification.keras\n",
558
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.8109 - loss: 0.7768 - val_accuracy: 0.7282 - val_loss: 0.9147\n",
559
+ "Epoch 82/100\n",
560
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8125 - loss: 0.7718\n",
561
+ "Epoch 82: val_loss did not improve from 0.91471\n",
562
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8297 - loss: 0.6972 - val_accuracy: 0.7282 - val_loss: 0.9167\n",
563
+ "Epoch 83/100\n",
564
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8125 - loss: 0.7190\n",
565
+ "Epoch 83: val_loss did not improve from 0.91471\n",
566
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8061 - loss: 0.7488 - val_accuracy: 0.7282 - val_loss: 0.9196\n",
567
+ "Epoch 84/100\n",
568
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.7812 - loss: 0.9175\n",
569
+ "Epoch 84: val_loss did not improve from 0.91471\n",
570
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8036 - loss: 0.7816 - val_accuracy: 0.7282 - val_loss: 0.9226\n",
571
+ "Epoch 85/100\n",
572
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.8125 - loss: 0.7815\n",
573
+ "Epoch 85: val_loss did not improve from 0.91471\n",
574
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8151 - loss: 0.7643 - val_accuracy: 0.7282 - val_loss: 0.9201\n",
575
+ "Epoch 86/100\n",
576
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.9062 - loss: 0.4617\n",
577
+ "Epoch 86: val_loss did not improve from 0.91471\n",
578
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8332 - loss: 0.6944 - val_accuracy: 0.7282 - val_loss: 0.9173\n",
579
+ "Epoch 87/100\n",
580
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.8125 - loss: 0.8332\n",
581
+ "Epoch 87: val_loss did not improve from 0.91471\n",
582
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8034 - loss: 0.7576 - val_accuracy: 0.7282 - val_loss: 0.9181\n",
583
+ "Epoch 88/100\n",
584
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.9062 - loss: 0.4584\n",
585
+ "Epoch 88: val_loss did not improve from 0.91471\n",
586
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8133 - loss: 0.7535 - val_accuracy: 0.7282 - val_loss: 0.9193\n",
587
+ "Epoch 89/100\n",
588
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8438 - loss: 0.7613\n",
589
+ "Epoch 89: val_loss did not improve from 0.91471\n",
590
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8365 - loss: 0.7370 - val_accuracy: 0.7282 - val_loss: 0.9191\n",
591
+ "Epoch 90/100\n",
592
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8750 - loss: 0.7454\n",
593
+ "Epoch 90: val_loss did not improve from 0.91471\n",
594
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8401 - loss: 0.6890 - val_accuracy: 0.7282 - val_loss: 0.9179\n",
595
+ "Epoch 91/100\n",
596
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.7812 - loss: 0.8456\n",
597
+ "Epoch 91: val_loss did not improve from 0.91471\n",
598
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8015 - loss: 0.7629 - val_accuracy: 0.7282 - val_loss: 0.9177\n",
599
+ "Epoch 92/100\n",
600
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.6562 - loss: 1.1234\n",
601
+ "Epoch 92: val_loss did not improve from 0.91471\n",
602
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7767 - loss: 0.8235 - val_accuracy: 0.7282 - val_loss: 0.9155\n",
603
+ "Epoch 93/100\n",
604
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.8750 - loss: 0.6304\n",
605
+ "Epoch 93: val_loss improved from 0.91471 to 0.90965, saving model to audio_classification.keras\n",
606
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.8300 - loss: 0.6764 - val_accuracy: 0.7282 - val_loss: 0.9097\n",
607
+ "Epoch 94/100\n",
608
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 19ms/step - accuracy: 0.8438 - loss: 0.7582\n",
609
+ "Epoch 94: val_loss improved from 0.90965 to 0.90822, saving model to audio_classification.keras\n",
610
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.8243 - loss: 0.7305 - val_accuracy: 0.7282 - val_loss: 0.9082\n",
611
+ "Epoch 95/100\n",
612
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8438 - loss: 0.6170\n",
613
+ "Epoch 95: val_loss improved from 0.90822 to 0.90646, saving model to audio_classification.keras\n",
614
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.8193 - loss: 0.6996 - val_accuracy: 0.7282 - val_loss: 0.9065\n",
615
+ "Epoch 96/100\n",
616
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.8125 - loss: 0.5991\n",
617
+ "Epoch 96: val_loss improved from 0.90646 to 0.90536, saving model to audio_classification.keras\n",
618
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.8026 - loss: 0.7276 - val_accuracy: 0.7282 - val_loss: 0.9054\n",
619
+ "Epoch 97/100\n",
620
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 21ms/step - accuracy: 0.8125 - loss: 0.7258\n",
621
+ "Epoch 97: val_loss did not improve from 0.90536\n",
622
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8285 - loss: 0.7067 - val_accuracy: 0.7282 - val_loss: 0.9055\n",
623
+ "Epoch 98/100\n",
624
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 23ms/step - accuracy: 0.7812 - loss: 0.7810\n",
625
+ "Epoch 98: val_loss improved from 0.90536 to 0.90286, saving model to audio_classification.keras\n",
626
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 5ms/step - accuracy: 0.8163 - loss: 0.7018 - val_accuracy: 0.7282 - val_loss: 0.9029\n"
627
+ ]
628
+ },
629
+ {
630
+ "name": "stdout",
631
+ "output_type": "stream",
632
+ "text": [
633
+ "Epoch 99/100\n",
634
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.7188 - loss: 1.1051\n",
635
+ "Epoch 99: val_loss did not improve from 0.90286\n",
636
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.8075 - loss: 0.7663 - val_accuracy: 0.7282 - val_loss: 0.9030\n",
637
+ "Epoch 100/100\n",
638
+ "\u001b[1m 1/13\u001b[0m \u001b[32m━\u001b[0m\u001b[37m━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[1m0s\u001b[0m 20ms/step - accuracy: 0.8125 - loss: 0.6395\n",
639
+ "Epoch 100: val_loss did not improve from 0.90286\n",
640
+ "\u001b[1m13/13\u001b[0m \u001b[32m━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[37m\u001b[0m \u001b[1m0s\u001b[0m 3ms/step - accuracy: 0.7935 - loss: 0.7532 - val_accuracy: 0.7282 - val_loss: 0.9033\n",
641
+ "Test accuracy: 72.82%\n"
642
+ ]
643
+ }
644
+ ],
645
+ "source": [
646
+ "\n",
647
+ "\n",
648
+ "# Define the model\n",
649
+ "model = Sequential()\n",
650
+ "\n",
651
+ "model.add(Dense(256, input_shape=(40,)))\n",
652
+ "model.add(Activation('relu'))\n",
653
+ "model.add(Dropout(0.5))\n",
654
+ "\n",
655
+ "model.add(Dense(128))\n",
656
+ "model.add(Activation('relu'))\n",
657
+ "model.add(Dropout(0.5))\n",
658
+ "\n",
659
+ "model.add(Dense(64))\n",
660
+ "model.add(Activation('relu'))\n",
661
+ "model.add(Dropout(0.5))\n",
662
+ "\n",
663
+ "model.add(Dense(len(categories)))\n",
664
+ "model.add(Activation('softmax'))\n",
665
+ "\n",
666
+ "# Compile the model\n",
667
+ "model.compile(loss='categorical_crossentropy', metrics=['accuracy'], optimizer='adam')\n",
668
+ "\n",
669
+ "# Train the model\n",
670
+ "num_epochs = 100\n",
671
+ "num_batch_size = 32\n",
672
+ "\n",
673
+ "checkpointer = ModelCheckpoint(filepath='audio_classification.keras', \n",
674
+ " verbose=1, save_best_only=True)\n",
675
+ "\n",
676
+ "history = model.fit(X_train, y_train, batch_size=num_batch_size, epochs=num_epochs, validation_data=(X_test, y_test), callbacks=[checkpointer], verbose=1)\n",
677
+ "\n",
678
+ "# Evaluate the model\n",
679
+ "test_accuracy = model.evaluate(X_test, y_test, verbose=0)\n",
680
+ "print(f'Test accuracy: {test_accuracy[1] * 100:.2f}%')\n"
681
+ ]
682
+ },
683
+ {
684
+ "cell_type": "code",
685
+ "execution_count": 10,
686
+ "id": "0559c8e5",
687
+ "metadata": {},
688
+ "outputs": [],
689
+ "source": [
690
+ "# Save the model\n",
691
+ "model.save('infant_cry_classification_model.keras')"
692
+ ]
693
+ },
694
+ {
695
+ "cell_type": "code",
696
+ "execution_count": null,
697
+ "id": "171ff113",
698
+ "metadata": {},
699
+ "outputs": [],
700
+ "source": []
701
+ }
702
+ ],
703
+ "metadata": {
704
+ "kernelspec": {
705
+ "display_name": "Python 3 (ipykernel)",
706
+ "language": "python",
707
+ "name": "python3"
708
+ },
709
+ "language_info": {
710
+ "codemirror_mode": {
711
+ "name": "ipython",
712
+ "version": 3
713
+ },
714
+ "file_extension": ".py",
715
+ "mimetype": "text/x-python",
716
+ "name": "python",
717
+ "nbconvert_exporter": "python",
718
+ "pygments_lexer": "ipython3",
719
+ "version": "3.9.12"
720
+ }
721
+ },
722
+ "nbformat": 4,
723
+ "nbformat_minor": 5
724
+ }
infant_cry_classification_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a8db8bed389d49d40ffe3fadfe3fa7642b365a955d38771e26e78ebd267af8d
3
+ size 666864
infant_cry_classification_model.keras ADDED
Binary file (666 kB). View file