handsUp-backend / ai_model /models /detectLettersModel.py
mutarisi
Add large model files with Git LFS
f968273
import pickle
import numpy as np
import tensorflow as tf
from sklearn.preprocessing import LabelEncoder
from sklearn.metrics import accuracy_score
from tensorflow.keras.callbacks import EarlyStopping
early_stop = EarlyStopping(patience=2, restore_best_weights=True)
dataDictTrain = pickle.load(open('../processed_data/trainData.pickle', 'rb'))
cleanedData = []
cleanedLabels = []
for i, item in enumerate(dataDictTrain['data']):
if isinstance(item, (np.ndarray, list)) and len(item) == 42:
cleanedData.append(np.array(item, dtype=np.float32))
cleanedLabels.append(dataDictTrain['labels'][i])
xTrain = np.array(cleanedData)
yTrainRaw = np.array(cleanedLabels)
dataDictTest = pickle.load(open('../processed_data/testData.pickle', 'rb'))
xTestRaw = np.array(dataDictTest['data'], dtype=np.float32)
yTestRaw = np.array(dataDictTest['labels'])
xTrain = xTrain.reshape(-1, 42, 1)
xTest = xTestRaw.reshape(-1, 42, 1)
labelEncoder = LabelEncoder()
labelEncoder.fit(yTrainRaw)
yTrainEncoded = labelEncoder.transform(yTrainRaw)
yTestEncoded = labelEncoder.transform(yTestRaw)
#second data set for finetuning
dataDictTrain2 = pickle.load(open('../processed_data/fineTuneData.pickle', 'rb'))
cleanedData2 = []
cleanedLabels2 = []
for i, item in enumerate(dataDictTrain2['data']):
if isinstance(item, (np.ndarray, list)) and len(item) == 42:
cleanedData2.append(np.array(item, dtype=np.float32))
cleanedLabels2.append(dataDictTrain2['labels'][i])
xTrain2 = np.stack(cleanedData2)
yTrainRaw2= np.array(cleanedLabels2)
xTrain2 = xTrain2.reshape(-1, 42, 1)
labelEncoder2 = LabelEncoder()
labelEncoder2.fit(yTrainRaw2)
yTrainEncoded2 = labelEncoder2.transform(yTrainRaw2)
model = tf.keras.models.Sequential([
tf.keras.layers.Conv1D(32, kernel_size=3, activation='relu', input_shape=(42, 1)),
tf.keras.layers.MaxPooling1D(pool_size=2),
tf.keras.layers.Conv1D(64, kernel_size=3, activation='relu'),
tf.keras.layers.Flatten(),
tf.keras.layers.Dense(64, activation='relu'),
tf.keras.layers.Dense(len(labelEncoder.classes_), activation='softmax')
])
model.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['accuracy'])
model.fit(xTrain, yTrainEncoded, epochs=15, batch_size=32, validation_split=0.2)
for layer in model.layers[:-2]:
layer.trainable = False
model.compile(optimizer=tf.keras.optimizers.Adam(1e-5),loss='sparse_categorical_crossentropy',metrics=['accuracy'])
yTrainEncoded2 = labelEncoder2.transform(yTrainRaw2)
model.fit(xTrain2, yTrainEncoded2, epochs=10, batch_size=32, validation_split=0.2, callbacks = [early_stop])
testLoss, testAccuracy = model.evaluate(xTest, yTestEncoded)
print(f"\nTest accuracy: {testAccuracy * 100:.2f}%")
predictions = model.predict(xTest)
predictedIndex = np.argmax(predictions, axis=1)
predictedLabels = labelEncoder.inverse_transform(predictedIndex)
print("\nPredictions:")
for i, pred in enumerate(predictedLabels):
print(f"Image {i}: Predicted = {pred}, Actual = {yTestRaw[i]}")
model.save("detectLettersModel.keras")
with open("labelEncoder.pickle", "wb") as file:
pickle.dump(labelEncoder, file)