import pickle import numpy as np import tensorflow as tf from sklearn.preprocessing import LabelEncoder from sklearn.metrics import accuracy_score from tensorflow.keras.callbacks import EarlyStopping early_stop = EarlyStopping(patience=2, restore_best_weights=True) dataDictTrain = pickle.load(open('../processed_data/trainData.pickle', 'rb')) cleanedData = [] cleanedLabels = [] for i, item in enumerate(dataDictTrain['data']): if isinstance(item, (np.ndarray, list)) and len(item) == 42: cleanedData.append(np.array(item, dtype=np.float32)) cleanedLabels.append(dataDictTrain['labels'][i]) xTrain = np.array(cleanedData) yTrainRaw = np.array(cleanedLabels) dataDictTest = pickle.load(open('../processed_data/testData.pickle', 'rb')) xTestRaw = np.array(dataDictTest['data'], dtype=np.float32) yTestRaw = np.array(dataDictTest['labels']) xTrain = xTrain.reshape(-1, 42, 1) xTest = xTestRaw.reshape(-1, 42, 1) labelEncoder = LabelEncoder() labelEncoder.fit(yTrainRaw) yTrainEncoded = labelEncoder.transform(yTrainRaw) yTestEncoded = labelEncoder.transform(yTestRaw) #second data set for finetuning dataDictTrain2 = pickle.load(open('../processed_data/fineTuneData.pickle', 'rb')) cleanedData2 = [] cleanedLabels2 = [] for i, item in enumerate(dataDictTrain2['data']): if isinstance(item, (np.ndarray, list)) and len(item) == 42: cleanedData2.append(np.array(item, dtype=np.float32)) cleanedLabels2.append(dataDictTrain2['labels'][i]) xTrain2 = np.stack(cleanedData2) yTrainRaw2= np.array(cleanedLabels2) xTrain2 = xTrain2.reshape(-1, 42, 1) labelEncoder2 = LabelEncoder() labelEncoder2.fit(yTrainRaw2) yTrainEncoded2 = labelEncoder2.transform(yTrainRaw2) model = tf.keras.models.Sequential([ tf.keras.layers.Conv1D(32, kernel_size=3, activation='relu', input_shape=(42, 1)), tf.keras.layers.MaxPooling1D(pool_size=2), tf.keras.layers.Conv1D(64, kernel_size=3, activation='relu'), tf.keras.layers.Flatten(), tf.keras.layers.Dense(64, activation='relu'), tf.keras.layers.Dense(len(labelEncoder.classes_), activation='softmax') ]) model.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['accuracy']) model.fit(xTrain, yTrainEncoded, epochs=15, batch_size=32, validation_split=0.2) for layer in model.layers[:-2]: layer.trainable = False model.compile(optimizer=tf.keras.optimizers.Adam(1e-5),loss='sparse_categorical_crossentropy',metrics=['accuracy']) yTrainEncoded2 = labelEncoder2.transform(yTrainRaw2) model.fit(xTrain2, yTrainEncoded2, epochs=10, batch_size=32, validation_split=0.2, callbacks = [early_stop]) testLoss, testAccuracy = model.evaluate(xTest, yTestEncoded) print(f"\nTest accuracy: {testAccuracy * 100:.2f}%") predictions = model.predict(xTest) predictedIndex = np.argmax(predictions, axis=1) predictedLabels = labelEncoder.inverse_transform(predictedIndex) print("\nPredictions:") for i, pred in enumerate(predictedLabels): print(f"Image {i}: Predicted = {pred}, Actual = {yTestRaw[i]}") model.save("detectLettersModel.keras") with open("labelEncoder.pickle", "wb") as file: pickle.dump(labelEncoder, file)