File size: 12,006 Bytes
7262f3e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
import numpy as np
from sklearn.datasets import fetch_openml
from sklearn.model_selection import train_test_split
import numpy
import sklearn
import time
from datetime import timedelta
import pickle
import os
from PIL import Image
import tkinter as tk
from tkinter import filedialog

print("NumPy版本:", numpy.__version__)
print("Scikit-learn版本:", sklearn.__version__)

class NeuralNetwork:
    def __init__(self, input_size, hidden_size1, hidden_size2, output_size):
        # 初始化网络参数
        # 如果初始权重过大,输入到激活函数(如 sigmoid)的值可能会非常大或非常小,
        # 导致激活函数的输出接近其极限值(0 或 1)。这会使得梯度接近于零,从而导致梯度消失问题,阻碍网络的学习
        # 从标准正态分布(均值为 0,标准差为 1)中随机采样的浮点数
        self.weights1 = np.random.randn(input_size, hidden_size1) * 0.1
        self.bias1 = np.zeros((1, hidden_size1))
        self.weights2 = np.random.randn(hidden_size1, hidden_size2) * 0.1
        self.bias2 = np.zeros((1, hidden_size2))
        self.weights3 = np.random.randn(hidden_size2, output_size) * 0.1
        self.bias3 = np.zeros((1, output_size))

    def sigmoid(self, x):
        return 1 / (1 + np.exp(-x))

    def sigmoid_derivative(self, x):
        return x * (1 - x)

    def softmax(self, x):
        exp_x = np.exp(x - np.max(x, axis=1, keepdims=True))
        return exp_x / np.sum(exp_x, axis=1, keepdims=True)

    def forward(self, X):
        # 前向传播
        self.layer1 = self.sigmoid(np.dot(X, self.weights1) + self.bias1)
        self.layer2 = self.sigmoid(np.dot(self.layer1, self.weights2) + self.bias2)
        # 输出层用的是softmax,公式是a=e^z/sum(e^z)
        self.output = self.softmax(np.dot(self.layer2, self.weights3) + self.bias3)
        return self.output

    def backward(self, X, y, learning_rate):
        batch_size = X.shape[0]
        
        # 输出层误差 δ^(L) = ∂C/∂a^(L)
        # 对于softmax输出层的简化形式,公式是δ=a-y
        delta3 = self.output - y  # 对于softmax输出层的简化形式
        
        # 隐藏层2误差 δ^(L-1) = ((w^(L))^T * δ^(L)) * σ'(z^(L-1))
        delta2 = np.dot(delta3, self.weights3.T) * self.sigmoid_derivative(self.layer2)
        
        # 隐藏层1误差 δ^(L-2) = ((w^(L-1))^T * δ^(L-1)) * σ'(z^(L-2))
        delta1 = np.dot(delta2, self.weights2.T) * self.sigmoid_derivative(self.layer1)

        # 权重更新公式:w^(L) = w^(L) - η * δ^(L) * (a^(L-1))^T,使用 mini-batch 梯度下降
        self.weights3 -= learning_rate * np.dot(self.layer2.T, delta3) / batch_size
        self.weights2 -= learning_rate * np.dot(self.layer1.T, delta2) / batch_size
        self.weights1 -= learning_rate * np.dot(X.T, delta1) / batch_size
        # np.dot(self.layer2.T, delta3)返回的是 累加的梯度总和,不是平均值。
        # 而标准的梯度下降更新,其实是想用 每个样本的平均梯度,让不同 batch size 的训练过程有更一致的步长

        # 偏置更新公式:b^(L) = b^(L) - η * δ^(L)
        self.bias3 -= learning_rate * np.sum(delta3, axis=0, keepdims=True) / batch_size
        self.bias2 -= learning_rate * np.sum(delta2, axis=0, keepdims=True) / batch_size
        self.bias1 -= learning_rate * np.sum(delta1, axis=0, keepdims=True) / batch_size

    def save_model(self, filename):
        """保存模型参数"""
        model_params = {
            'weights1': self.weights1,
            'weights2': self.weights2,
            'weights3': self.weights3,
            'bias1': self.bias1,
            'bias2': self.bias2,
            'bias3': self.bias3
        }
        with open(filename, 'wb') as f:
            pickle.dump(model_params, f)
    
    def load_model(self, filename):
        """加载模型参数"""
        with open(filename, 'rb') as f:
            model_params = pickle.load(f)
        self.weights1 = model_params['weights1']
        self.weights2 = model_params['weights2']
        self.weights3 = model_params['weights3']
        self.bias1 = model_params['bias1']
        self.bias2 = model_params['bias2']
        self.bias3 = model_params['bias3']
    
    def predict(self, X):
        """预测单个图像"""
        # 确保输入是2D数组
        if len(X.shape) == 1:
            X = X.reshape(1, -1)
        # 归一化,将每个像素的值归一化到 [0, 1] 的范围内
        X = X / 255.0
        # 前向传播
        output = self.forward(X)
        # 返回预测的数字
        return np.argmax(output, axis=1)[0]

def preprocess_data():
    # 加载MNIST数据集
    X, y = fetch_openml('mnist_784', version=1, return_X_y=True, as_frame=False)
    
    # 数据归一化
    X = X / 255.0
    
    # 将标签转换为one-hot编码
    y_onehot = np.zeros((y.shape[0], 10))
    for i in range(y.shape[0]):
        y_onehot[i, int(y[i])] = 1
    
    # 分割训练集和测试集
    X_train, X_test, y_train, y_test = train_test_split(X, y_onehot, test_size=0.2, random_state=42)
    return X_train, X_test, y_train, y_test

def main():
    # 初始化参数
    input_size = 784
    hidden_size1 = 30
    hidden_size2 = 60
    output_size = 10
    learning_rate = 0.1
    epochs = 50
    batch_size = 128

    # 加载和预处理数据
    X_train, X_test, y_train, y_test = preprocess_data()

    # 创建神经网络
    nn = NeuralNetwork(input_size, hidden_size1, hidden_size2, output_size)

    print("开始训练...")
    start_time = time.time()
    total_batches = len(X_train) // batch_size

    # 训练网络
    for epoch in range(epochs):
        epoch_start = time.time()
        
        # 随机打乱训练数据
        indices = np.random.permutation(len(X_train))
        X_train = X_train[indices]
        y_train = y_train[indices]

        # 批量训练
        for i in range(0, len(X_train), batch_size):
            if i % (batch_size * 20) == 0:
                print(f'\rEpoch {epoch + 1}/{epochs} - 批次进度: {i//batch_size}/{total_batches}', end='')
            
            X_batch = X_train[i:i + batch_size]
            y_batch = y_train[i:i + batch_size]
            
            nn.forward(X_batch)
            nn.backward(X_batch, y_batch, learning_rate)

        epoch_time = time.time() - epoch_start
        
        if (epoch + 1) % 5 == 0:
            train_predictions = np.argmax(nn.forward(X_train), axis=1)
            train_true = np.argmax(y_train, axis=1)
            train_accuracy = np.mean(train_predictions == train_true)

            test_predictions = np.argmax(nn.forward(X_test), axis=1)
            test_true = np.argmax(y_test, axis=1)
            test_accuracy = np.mean(test_predictions == test_true)

            print(f'\nEpoch {epoch + 1}/{epochs}')
            print(f'每轮用时: {timedelta(seconds=int(epoch_time))}')
            print(f'训练集准确率: {train_accuracy:.4f}')
            print(f'测试集准确率: {test_accuracy:.4f}')

    total_time = time.time() - start_time
    print(f'\n训练完成!总用时: {timedelta(seconds=int(total_time))}')

    # 在训练完成后保存模型
    print("保存模型...")
    nn.save_model('mnist_model.pkl')

# 添加新的函数用于预测
def predict_digit(image_data):
    """

    预测单个手写数字

    image_data: 784维的numpy数组(28x28像素)

    """
    # 创建模型实例
    nn = NeuralNetwork(784, 30, 60, 10)
    # 加载训练好的模型
    nn.load_model('mnist_model.pkl')
    # 进行预测
    return nn.predict(image_data)

def load_image(image_path):
    """

    加载并处理图片文件

    image_path: 图片文件的路径

    返回: 处理后的图片数据(784维numpy数组)

    """
    # 打开图片
    img = Image.open(image_path)
    # 转换为灰度图
    img = img.convert('L')
    # 调整大小为28x28像素
    img = img.resize((28, 28))
    # 转换为numpy数组并归一化
    img_array = np.array(img)
    # 反转颜色(确保白底黑字)
    img_array = 255 - img_array
    # 展平为一维数组
    img_array = img_array.reshape(784)
    return img_array

def select_image():
    try:
        root = tk.Tk()
        root.withdraw()  # 隐藏主窗口
        root.attributes('-topmost', True)  # 确保对话框在最前面
        
        file_path = filedialog.askopenfilename(
            title='选择手写数字图片',
            filetypes=[
                ('图片文件', '*.png *.jpg *.jpeg *.bmp *.gif'),
                ('所有文件', '*.*')
            ]
        )
        
        root.destroy()  # 确保完全关闭Tk窗口
        return file_path if file_path else None
        
    except Exception as e:
        print(f"选择文件时出错:{str(e)}")
        return None

def select_image_alternative():
    """备选的文件选择方法"""
    print("\n请直接输入图片文件的完整路径:")
    file_path = input().strip()
    if os.path.exists(file_path):
        return file_path
    print("文件不存在!")
    return None

# 使用示例
if __name__ == "__main__":
    if input("是否需要重新训练模型?(y/n): ").lower() == 'y':
        main()
    else:
        if not os.path.exists('mnist_model.pkl'):
            print("\n错误:找不到模型文件 'mnist_model.pkl'")
            print("请先训练模型(输入 'y' 进行训练)再进行预测。")
            exit()
            
        try:
            print("\n选择预测模式:")
            print("1. 使用MNIST测试集样本")
            print("2. 使用本地图片文件")
            choice = input("请选择(1/2):")
            
            nn = NeuralNetwork(784, 30, 60, 10)
            nn.load_model('mnist_model.pkl')
            
            if choice == '1':
                # 使用MNIST测试集
                print("\n加载测试数据...")
                X, y = fetch_openml('mnist_784', version=1, return_X_y=True, as_frame=False)
                test_samples = X[:10]
                true_labels = y[:10]
                
                print("\n预测演示:")
                for i, (sample, true_label) in enumerate(zip(test_samples, true_labels)):
                    predicted = nn.predict(sample)
                    print(f"样本 {i+1}: 预测值 = {predicted}, 实际值 = {true_label}")
            
            elif choice == '2':
                print("\n准备打开文件选择对话框...")
                try:
                    print("\n请在弹出的对话框中选择图片文件...")
                    image_path = select_image()
                    if not image_path:
                        print("\n使用备选方法...")
                        image_path = select_image_alternative()
                    if not image_path:
                        print("未能获取有效的图片文件路径")
                        exit()
                    
                    print(f"已选择文件:{image_path}")
                    img_data = load_image(image_path)
                    predicted = nn.predict(img_data)
                    print(f"\n预测结果:这个数字是 {predicted}")
                except Exception as e:
                    print(f"\n选择文件过程中发生错误:{str(e)}")
                    print("请确保系统支持文件对话框操作。")
            
            else:
                print("无效的选择")
                
        except Exception as e:
            print(f"\n发生错误:{str(e)}")
            print("请确保已经完成模型训练,并且模型文件正确保存。")