HNU-SE-materials / machine-learning /notes /HandWritingNeurlNet-DRN.py
Deerainy's picture
Upload 5 files
7262f3e verified
import numpy as np
from sklearn.datasets import fetch_openml
from sklearn.model_selection import train_test_split
import numpy
import sklearn
import time
from datetime import timedelta
import pickle
import os
from PIL import Image
import tkinter as tk
from tkinter import filedialog
print("NumPy版本:", numpy.__version__)
print("Scikit-learn版本:", sklearn.__version__)
class NeuralNetwork:
def __init__(self, input_size, hidden_size1, hidden_size2, output_size):
# 初始化网络参数
# 如果初始权重过大,输入到激活函数(如 sigmoid)的值可能会非常大或非常小,
# 导致激活函数的输出接近其极限值(0 或 1)。这会使得梯度接近于零,从而导致梯度消失问题,阻碍网络的学习
# 从标准正态分布(均值为 0,标准差为 1)中随机采样的浮点数
self.weights1 = np.random.randn(input_size, hidden_size1) * 0.1
self.bias1 = np.zeros((1, hidden_size1))
self.weights2 = np.random.randn(hidden_size1, hidden_size2) * 0.1
self.bias2 = np.zeros((1, hidden_size2))
self.weights3 = np.random.randn(hidden_size2, output_size) * 0.1
self.bias3 = np.zeros((1, output_size))
def sigmoid(self, x):
return 1 / (1 + np.exp(-x))
def sigmoid_derivative(self, x):
return x * (1 - x)
def softmax(self, x):
exp_x = np.exp(x - np.max(x, axis=1, keepdims=True))
return exp_x / np.sum(exp_x, axis=1, keepdims=True)
def forward(self, X):
# 前向传播
self.layer1 = self.sigmoid(np.dot(X, self.weights1) + self.bias1)
self.layer2 = self.sigmoid(np.dot(self.layer1, self.weights2) + self.bias2)
# 输出层用的是softmax,公式是a=e^z/sum(e^z)
self.output = self.softmax(np.dot(self.layer2, self.weights3) + self.bias3)
return self.output
def backward(self, X, y, learning_rate):
batch_size = X.shape[0]
# 输出层误差 δ^(L) = ∂C/∂a^(L)
# 对于softmax输出层的简化形式,公式是δ=a-y
delta3 = self.output - y # 对于softmax输出层的简化形式
# 隐藏层2误差 δ^(L-1) = ((w^(L))^T * δ^(L)) * σ'(z^(L-1))
delta2 = np.dot(delta3, self.weights3.T) * self.sigmoid_derivative(self.layer2)
# 隐藏层1误差 δ^(L-2) = ((w^(L-1))^T * δ^(L-1)) * σ'(z^(L-2))
delta1 = np.dot(delta2, self.weights2.T) * self.sigmoid_derivative(self.layer1)
# 权重更新公式:w^(L) = w^(L) - η * δ^(L) * (a^(L-1))^T,使用 mini-batch 梯度下降
self.weights3 -= learning_rate * np.dot(self.layer2.T, delta3) / batch_size
self.weights2 -= learning_rate * np.dot(self.layer1.T, delta2) / batch_size
self.weights1 -= learning_rate * np.dot(X.T, delta1) / batch_size
# np.dot(self.layer2.T, delta3)返回的是 累加的梯度总和,不是平均值。
# 而标准的梯度下降更新,其实是想用 每个样本的平均梯度,让不同 batch size 的训练过程有更一致的步长
# 偏置更新公式:b^(L) = b^(L) - η * δ^(L)
self.bias3 -= learning_rate * np.sum(delta3, axis=0, keepdims=True) / batch_size
self.bias2 -= learning_rate * np.sum(delta2, axis=0, keepdims=True) / batch_size
self.bias1 -= learning_rate * np.sum(delta1, axis=0, keepdims=True) / batch_size
def save_model(self, filename):
"""保存模型参数"""
model_params = {
'weights1': self.weights1,
'weights2': self.weights2,
'weights3': self.weights3,
'bias1': self.bias1,
'bias2': self.bias2,
'bias3': self.bias3
}
with open(filename, 'wb') as f:
pickle.dump(model_params, f)
def load_model(self, filename):
"""加载模型参数"""
with open(filename, 'rb') as f:
model_params = pickle.load(f)
self.weights1 = model_params['weights1']
self.weights2 = model_params['weights2']
self.weights3 = model_params['weights3']
self.bias1 = model_params['bias1']
self.bias2 = model_params['bias2']
self.bias3 = model_params['bias3']
def predict(self, X):
"""预测单个图像"""
# 确保输入是2D数组
if len(X.shape) == 1:
X = X.reshape(1, -1)
# 归一化,将每个像素的值归一化到 [0, 1] 的范围内
X = X / 255.0
# 前向传播
output = self.forward(X)
# 返回预测的数字
return np.argmax(output, axis=1)[0]
def preprocess_data():
# 加载MNIST数据集
X, y = fetch_openml('mnist_784', version=1, return_X_y=True, as_frame=False)
# 数据归一化
X = X / 255.0
# 将标签转换为one-hot编码
y_onehot = np.zeros((y.shape[0], 10))
for i in range(y.shape[0]):
y_onehot[i, int(y[i])] = 1
# 分割训练集和测试集
X_train, X_test, y_train, y_test = train_test_split(X, y_onehot, test_size=0.2, random_state=42)
return X_train, X_test, y_train, y_test
def main():
# 初始化参数
input_size = 784
hidden_size1 = 30
hidden_size2 = 60
output_size = 10
learning_rate = 0.1
epochs = 50
batch_size = 128
# 加载和预处理数据
X_train, X_test, y_train, y_test = preprocess_data()
# 创建神经网络
nn = NeuralNetwork(input_size, hidden_size1, hidden_size2, output_size)
print("开始训练...")
start_time = time.time()
total_batches = len(X_train) // batch_size
# 训练网络
for epoch in range(epochs):
epoch_start = time.time()
# 随机打乱训练数据
indices = np.random.permutation(len(X_train))
X_train = X_train[indices]
y_train = y_train[indices]
# 批量训练
for i in range(0, len(X_train), batch_size):
if i % (batch_size * 20) == 0:
print(f'\rEpoch {epoch + 1}/{epochs} - 批次进度: {i//batch_size}/{total_batches}', end='')
X_batch = X_train[i:i + batch_size]
y_batch = y_train[i:i + batch_size]
nn.forward(X_batch)
nn.backward(X_batch, y_batch, learning_rate)
epoch_time = time.time() - epoch_start
if (epoch + 1) % 5 == 0:
train_predictions = np.argmax(nn.forward(X_train), axis=1)
train_true = np.argmax(y_train, axis=1)
train_accuracy = np.mean(train_predictions == train_true)
test_predictions = np.argmax(nn.forward(X_test), axis=1)
test_true = np.argmax(y_test, axis=1)
test_accuracy = np.mean(test_predictions == test_true)
print(f'\nEpoch {epoch + 1}/{epochs}')
print(f'每轮用时: {timedelta(seconds=int(epoch_time))}')
print(f'训练集准确率: {train_accuracy:.4f}')
print(f'测试集准确率: {test_accuracy:.4f}')
total_time = time.time() - start_time
print(f'\n训练完成!总用时: {timedelta(seconds=int(total_time))}')
# 在训练完成后保存模型
print("保存模型...")
nn.save_model('mnist_model.pkl')
# 添加新的函数用于预测
def predict_digit(image_data):
"""
预测单个手写数字
image_data: 784维的numpy数组(28x28像素)
"""
# 创建模型实例
nn = NeuralNetwork(784, 30, 60, 10)
# 加载训练好的模型
nn.load_model('mnist_model.pkl')
# 进行预测
return nn.predict(image_data)
def load_image(image_path):
"""
加载并处理图片文件
image_path: 图片文件的路径
返回: 处理后的图片数据(784维numpy数组)
"""
# 打开图片
img = Image.open(image_path)
# 转换为灰度图
img = img.convert('L')
# 调整大小为28x28像素
img = img.resize((28, 28))
# 转换为numpy数组并归一化
img_array = np.array(img)
# 反转颜色(确保白底黑字)
img_array = 255 - img_array
# 展平为一维数组
img_array = img_array.reshape(784)
return img_array
def select_image():
try:
root = tk.Tk()
root.withdraw() # 隐藏主窗口
root.attributes('-topmost', True) # 确保对话框在最前面
file_path = filedialog.askopenfilename(
title='选择手写数字图片',
filetypes=[
('图片文件', '*.png *.jpg *.jpeg *.bmp *.gif'),
('所有文件', '*.*')
]
)
root.destroy() # 确保完全关闭Tk窗口
return file_path if file_path else None
except Exception as e:
print(f"选择文件时出错:{str(e)}")
return None
def select_image_alternative():
"""备选的文件选择方法"""
print("\n请直接输入图片文件的完整路径:")
file_path = input().strip()
if os.path.exists(file_path):
return file_path
print("文件不存在!")
return None
# 使用示例
if __name__ == "__main__":
if input("是否需要重新训练模型?(y/n): ").lower() == 'y':
main()
else:
if not os.path.exists('mnist_model.pkl'):
print("\n错误:找不到模型文件 'mnist_model.pkl'")
print("请先训练模型(输入 'y' 进行训练)再进行预测。")
exit()
try:
print("\n选择预测模式:")
print("1. 使用MNIST测试集样本")
print("2. 使用本地图片文件")
choice = input("请选择(1/2):")
nn = NeuralNetwork(784, 30, 60, 10)
nn.load_model('mnist_model.pkl')
if choice == '1':
# 使用MNIST测试集
print("\n加载测试数据...")
X, y = fetch_openml('mnist_784', version=1, return_X_y=True, as_frame=False)
test_samples = X[:10]
true_labels = y[:10]
print("\n预测演示:")
for i, (sample, true_label) in enumerate(zip(test_samples, true_labels)):
predicted = nn.predict(sample)
print(f"样本 {i+1}: 预测值 = {predicted}, 实际值 = {true_label}")
elif choice == '2':
print("\n准备打开文件选择对话框...")
try:
print("\n请在弹出的对话框中选择图片文件...")
image_path = select_image()
if not image_path:
print("\n使用备选方法...")
image_path = select_image_alternative()
if not image_path:
print("未能获取有效的图片文件路径")
exit()
print(f"已选择文件:{image_path}")
img_data = load_image(image_path)
predicted = nn.predict(img_data)
print(f"\n预测结果:这个数字是 {predicted}")
except Exception as e:
print(f"\n选择文件过程中发生错误:{str(e)}")
print("请确保系统支持文件对话框操作。")
else:
print("无效的选择")
except Exception as e:
print(f"\n发生错误:{str(e)}")
print("请确保已经完成模型训练,并且模型文件正确保存。")