RRFRRF2 commited on
Commit
5dc9a8b
·
1 Parent(s): ff6f5b5

feat: add TextRNN

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. TextCNN-THUCNews/Classification/scripts/verify_embedding_order.py +0 -235
  2. TextRNN-THUCNews/Classification/README.md +110 -0
  3. TextRNN-THUCNews/Classification/dataset/THUCNews/data/class.txt +10 -0
  4. TextRNN-THUCNews/Classification/dataset/THUCNews/data/dev.txt +0 -0
  5. TextRNN-THUCNews/Classification/dataset/THUCNews/data/embedding_SougouNews.npz +3 -0
  6. TextRNN-THUCNews/Classification/dataset/THUCNews/data/embedding_Tencent.npz +3 -0
  7. TextRNN-THUCNews/Classification/dataset/THUCNews/data/test.txt +0 -0
  8. TextRNN-THUCNews/Classification/dataset/THUCNews/data/train.txt +0 -0
  9. TextRNN-THUCNews/Classification/dataset/THUCNews/data/vocab.pkl +3 -0
  10. TextRNN-THUCNews/Classification/dataset/THUCNews/saved_dict/model.ckpt +0 -0
  11. TextRNN-THUCNews/Classification/dataset/index.json +0 -0
  12. TextRNN-THUCNews/Classification/dataset/info.json +15 -0
  13. TextRNN-THUCNews/Classification/dataset/labels.npy +3 -0
  14. TextRNN-THUCNews/Classification/epochs/epoch_1/embeddings.npy +3 -0
  15. TextRNN-THUCNews/Classification/epochs/epoch_1/model.pt +3 -0
  16. TextRNN-THUCNews/Classification/epochs/epoch_1/predictions.npy +3 -0
  17. TextRNN-THUCNews/Classification/epochs/epoch_10/embeddings.npy +3 -0
  18. TextRNN-THUCNews/Classification/epochs/epoch_10/model.pt +3 -0
  19. TextRNN-THUCNews/Classification/epochs/epoch_10/predictions.npy +3 -0
  20. TextRNN-THUCNews/Classification/epochs/epoch_11/embeddings.npy +3 -0
  21. TextRNN-THUCNews/Classification/epochs/epoch_11/model.pt +3 -0
  22. TextRNN-THUCNews/Classification/epochs/epoch_11/predictions.npy +3 -0
  23. TextRNN-THUCNews/Classification/epochs/epoch_12/embeddings.npy +3 -0
  24. TextRNN-THUCNews/Classification/epochs/epoch_12/model.pt +3 -0
  25. TextRNN-THUCNews/Classification/epochs/epoch_12/predictions.npy +3 -0
  26. TextRNN-THUCNews/Classification/epochs/epoch_13/embeddings.npy +3 -0
  27. TextRNN-THUCNews/Classification/epochs/epoch_13/model.pt +3 -0
  28. TextRNN-THUCNews/Classification/epochs/epoch_13/predictions.npy +3 -0
  29. TextRNN-THUCNews/Classification/epochs/epoch_14/embeddings.npy +3 -0
  30. TextRNN-THUCNews/Classification/epochs/epoch_14/model.pt +3 -0
  31. TextRNN-THUCNews/Classification/epochs/epoch_14/predictions.npy +3 -0
  32. TextRNN-THUCNews/Classification/epochs/epoch_15/embeddings.npy +3 -0
  33. TextRNN-THUCNews/Classification/epochs/epoch_15/model.pt +3 -0
  34. TextRNN-THUCNews/Classification/epochs/epoch_15/predictions.npy +3 -0
  35. TextRNN-THUCNews/Classification/epochs/epoch_16/embeddings.npy +3 -0
  36. TextRNN-THUCNews/Classification/epochs/epoch_16/model.pt +3 -0
  37. TextRNN-THUCNews/Classification/epochs/epoch_16/predictions.npy +3 -0
  38. TextRNN-THUCNews/Classification/epochs/epoch_17/embeddings.npy +3 -0
  39. TextRNN-THUCNews/Classification/epochs/epoch_17/model.pt +3 -0
  40. TextRNN-THUCNews/Classification/epochs/epoch_17/predictions.npy +3 -0
  41. TextRNN-THUCNews/Classification/epochs/epoch_18/embeddings.npy +3 -0
  42. TextRNN-THUCNews/Classification/epochs/epoch_18/model.pt +3 -0
  43. TextRNN-THUCNews/Classification/epochs/epoch_18/predictions.npy +3 -0
  44. TextRNN-THUCNews/Classification/epochs/epoch_19/embeddings.npy +3 -0
  45. TextRNN-THUCNews/Classification/epochs/epoch_19/model.pt +3 -0
  46. TextRNN-THUCNews/Classification/epochs/epoch_19/predictions.npy +3 -0
  47. TextRNN-THUCNews/Classification/epochs/epoch_2/embeddings.npy +3 -0
  48. TextRNN-THUCNews/Classification/epochs/epoch_2/model.pt +3 -0
  49. TextRNN-THUCNews/Classification/epochs/epoch_2/predictions.npy +3 -0
  50. TextRNN-THUCNews/Classification/epochs/epoch_20/embeddings.npy +3 -0
TextCNN-THUCNews/Classification/scripts/verify_embedding_order.py DELETED
@@ -1,235 +0,0 @@
1
- # coding: UTF-8
2
- """
3
- 验证embedding保存顺序与原始数据集顺序一致性的脚本
4
- """
5
- import os
6
- import sys
7
- import numpy as np
8
- import yaml
9
- import torch
10
- from tqdm import tqdm
11
-
12
- # 添加当前目录到路径
13
- sys.path.append(os.path.dirname(os.path.abspath(__file__)))
14
-
15
- from model import Model, Config
16
- from dataset_utils import build_dataset, build_iterator
17
-
18
-
19
- def verify_embedding_order(config_path, epoch_num=1):
20
- """
21
- 验证指定epoch的embedding顺序是否与原始数据集一致
22
-
23
- Args:
24
- config_path: 配置文件路径
25
- epoch_num: 要验证的epoch数
26
- """
27
- print("=" * 60)
28
- print("验证Embedding保存顺序与原始数据集一致性")
29
- print("=" * 60)
30
-
31
- # 加载配置
32
- with open(config_path, 'r', encoding='utf-8') as f:
33
- train_config = yaml.safe_load(f)
34
-
35
- # 初始化配置
36
- config = Config(train_config['dataset_path'], train_config.get('embedding', 'random'))
37
- config.batch_size = train_config.get('batch_size', 128)
38
-
39
- # 构建数据集
40
- print("1. 构建数据集...")
41
- vocab, train_data, dev_data, test_data = build_dataset(config, train_config.get('use_word', False))
42
- config.n_vocab = len(vocab)
43
-
44
- print(f" 训练数据集大小: {len(train_data)}")
45
- print(f" 验证数据集大小: {len(dev_data)}")
46
- print(f" 测试数据集大小: {len(test_data)}")
47
- print(f" 总数据集大小: {len(train_data) + len(dev_data) + len(test_data)}")
48
-
49
- # 检查保存的embedding文件是否存在
50
- epochs_dir = train_config.get('epochs_dir', '../epochs')
51
- epoch_dir = os.path.join(epochs_dir, f'epoch_{epoch_num}')
52
- embeddings_path = os.path.join(epoch_dir, 'embeddings.npy')
53
- predictions_path = os.path.join(epoch_dir, 'predictions.npy')
54
- dataset_info_path = os.path.join(epoch_dir, 'dataset_info.json')
55
-
56
- if not os.path.exists(embeddings_path):
57
- print(f"❌ 找不到embedding文件: {embeddings_path}")
58
- print("请先运行训练脚本生成embedding文件")
59
- return False
60
-
61
- print(f"2. 加载epoch {epoch_num}的embedding文件...")
62
- saved_embeddings = np.load(embeddings_path)
63
- saved_predictions = np.load(predictions_path)
64
-
65
- print(f" 保存的embedding形状: {saved_embeddings.shape}")
66
- print(f" 保存的prediction形状: {saved_predictions.shape}")
67
-
68
- # 加载数据集信息
69
- dataset_info = []
70
- if os.path.exists(dataset_info_path):
71
- import json
72
- with open(dataset_info_path, 'r', encoding='utf-8') as f:
73
- dataset_info = json.load(f)
74
- print(" 数据集信息:")
75
- for info in dataset_info:
76
- print(f" {info['name']}: {info['size']} 个样本, 索引 {info['start_idx']}-{info['end_idx']-1}")
77
- else:
78
- print(" ⚠️ 未找到dataset_info.json,假设只有训练数据")
79
-
80
- # 加载模型
81
- print("3. 加载模型...")
82
- device = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')
83
- config.device = device
84
-
85
- model = Model(config).to(device)
86
- model_path = os.path.join(epoch_dir, 'model.pt')
87
- model.load_state_dict(torch.load(model_path, map_location=device))
88
- model.eval()
89
-
90
- # 重新生成embedding(使用新的迭代器,按train/dev/test顺序)
91
- print("4. 重新生成embedding以验证顺序...")
92
-
93
- # 按顺序处理三个数据集
94
- datasets = [
95
- ('train', train_data),
96
- ('dev', dev_data),
97
- ('test', test_data)
98
- ]
99
-
100
- fresh_features = []
101
- fresh_predictions = []
102
-
103
- with torch.no_grad():
104
- for dataset_name, dataset in datasets:
105
- print(f" 重新生成 {dataset_name} 数据集特征...")
106
-
107
- data_iter = build_iterator(dataset, config)
108
- dataset_features = []
109
- dataset_predictions = []
110
-
111
- for texts, labels in tqdm(data_iter, desc=f"重新生成{dataset_name}特征"):
112
- features = model.feature(texts)
113
- predictions = model.get_prediction(texts)
114
-
115
- dataset_features.append(features)
116
- dataset_predictions.append(predictions)
117
-
118
- if dataset_features:
119
- dataset_embeddings = np.vstack(dataset_features)
120
- dataset_preds = np.vstack(dataset_predictions)
121
-
122
- fresh_features.append(dataset_embeddings)
123
- fresh_predictions.append(dataset_preds)
124
-
125
- print(f" {dataset_name} 重新生成形状: {dataset_embeddings.shape}")
126
-
127
- fresh_embeddings = np.vstack(fresh_features)
128
- fresh_predictions_array = np.vstack(fresh_predictions)
129
-
130
- print(f" 重新生成的总embedding形状: {fresh_embeddings.shape}")
131
- print(f" 重新生成的总prediction形状: {fresh_predictions_array.shape}")
132
-
133
- # 比较两次生成的结果
134
- print("5. 比较结果...")
135
-
136
- # 检查形状是否一致
137
- if saved_embeddings.shape != fresh_embeddings.shape:
138
- print("❌ 形状不一致!")
139
- print(f" 保存的: {saved_embeddings.shape}")
140
- print(f" 重新生成的: {fresh_embeddings.shape}")
141
- return False
142
-
143
- # 检查数值是否一致(允许小的浮点误差)
144
- embedding_diff = np.abs(saved_embeddings - fresh_embeddings)
145
- max_embedding_diff = np.max(embedding_diff)
146
- mean_embedding_diff = np.mean(embedding_diff)
147
-
148
- prediction_diff = np.abs(saved_predictions - fresh_predictions_array)
149
- max_prediction_diff = np.max(prediction_diff)
150
- mean_prediction_diff = np.mean(prediction_diff)
151
-
152
- print(f" Embedding最大差异: {max_embedding_diff:.10f}")
153
- print(f" Embedding平均差异: {mean_embedding_diff:.10f}")
154
- print(f" Prediction最大差异: {max_prediction_diff:.10f}")
155
- print(f" Prediction平均差异: {mean_prediction_diff:.10f}")
156
-
157
- # 判断是否一致(允许1e-6的误差)
158
- tolerance = 1e-6
159
- embedding_consistent = max_embedding_diff < tolerance
160
- prediction_consistent = max_prediction_diff < tolerance
161
-
162
- if embedding_consistent and prediction_consistent:
163
- print("✅ 验证通过!Embedding和Prediction的保存顺序与原始数据集完全一致")
164
-
165
- # 额外验证:检查不同数据集的样本对应关系
166
- print("\n6. 额外验证各数据集样本的对应关系...")
167
-
168
- # 合并所有数据集用于验证
169
- all_data = train_data + dev_data + test_data
170
-
171
- # 验证前几个样本和各数据集边界处的样本
172
- test_indices = []
173
-
174
- # 添加前5个样本
175
- test_indices.extend(range(min(5, len(all_data))))
176
-
177
- # 如果有数据集信息,添加边界处的样本
178
- if dataset_info:
179
- for info in dataset_info:
180
- # 添加每个数据集开始和结束的样本
181
- start_idx = info['start_idx']
182
- end_idx = info['end_idx'] - 1
183
- test_indices.extend([start_idx, end_idx])
184
-
185
- # 去重并排序
186
- test_indices = sorted(list(set(test_indices)))
187
- test_indices = [i for i in test_indices if i < len(all_data)]
188
-
189
- for i in test_indices:
190
- original_label = all_data[i][1]
191
-
192
- # 从保存的prediction中获取预测结果
193
- saved_pred_logits = saved_predictions[i]
194
- saved_pred_class = np.argmax(saved_pred_logits)
195
-
196
- # 确定样本属于哪个数据集
197
- dataset_name = "unknown"
198
- if dataset_info:
199
- for info in dataset_info:
200
- if info['start_idx'] <= i < info['end_idx']:
201
- dataset_name = info['name']
202
- break
203
-
204
- print(f" 样本{i}: {dataset_name}数据集, 原始标签={original_label}, 预测类别={saved_pred_class}")
205
-
206
- return True
207
- else:
208
- print("❌ 验证失败!数据顺序可能不一致")
209
- if not embedding_consistent:
210
- print(f" Embedding不一致,最大差异: {max_embedding_diff}")
211
- if not prediction_consistent:
212
- print(f" Prediction不一致,最大差异: {max_prediction_diff}")
213
- return False
214
-
215
-
216
- def main():
217
- """主函数"""
218
- import argparse
219
-
220
- parser = argparse.ArgumentParser(description='验证embedding保存顺序')
221
- parser.add_argument('--config', type=str, default='train.yaml', help='配置文件路径')
222
- parser.add_argument('--epoch', type=int, default=1, help='要验证的epoch数')
223
-
224
- args = parser.parse_args()
225
-
226
- success = verify_embedding_order(args.config, args.epoch)
227
-
228
- if success:
229
- print("\n🎉 验证成功!可以放心使用保存的embedding进行可视化分析。")
230
- else:
231
- print("\n⚠️ 验证失败!请检查训练脚本的embedding保存逻辑。")
232
-
233
-
234
- if __name__ == '__main__':
235
- main()
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
TextRNN-THUCNews/Classification/README.md ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # TextRNN 可视化实验脚本
2
+
3
+ 基于原始 Chinese-Text-Classification-Pytorch 仓库重构的 TextRNN 训练脚本,专门用于可视化实验。
4
+
5
+ ## 目录结构
6
+
7
+ ```
8
+ Classification/
9
+ ├── scripts/ # 脚本文件
10
+ │ ├── model.py # 模型定义,包含feature、get_prediction、prediction函数
11
+ │ ├── train.py # 训练脚本,支持多卡训练
12
+ │ ├── train.yaml # 训练配置文件
13
+ │ ├── dataset_utils.py # 数据集处理工具
14
+ │ └── get_label.py # 标签提取脚本
15
+ ├── dataset/ # 数据集文件
16
+ │ ├── train.txt # 训练数据
17
+ │ ├── dev.txt # 验证数据
18
+ │ ├── test.txt # 测试数据
19
+ │ ├── class.txt # 类别列表
20
+ │ ├── vocab.pkl # 词汇表
21
+ │ └── labels.npy # 提取的标签
22
+ └── epochs/ # 按epoch存放模型文件和特征向量
23
+ ├── epoch_1/
24
+ │ ├── model.pt # 模型权重
25
+ │ ├── embeddings.npy # 特征向量
26
+ │ └── predictions.npy # 预测值
27
+ └── epoch_2/
28
+ └── ...
29
+ ```
30
+
31
+ ## 功能说明
32
+
33
+ ### 1. model.py
34
+ - **Model类**: TextRNN模型实现
35
+ - **feature()**: 提取中间层特征向量(dropout层输出),用于可视化
36
+ - **get_prediction()**: 获取模型最终层输出向量(logits)
37
+ - **prediction()**: 根据中间特征向量预测结果
38
+
39
+ ### 2. train.py
40
+ - 支持多GPU训练
41
+ - 每个epoch自动保存模型、特征向量、预测值到 `epochs/epoch_N/`
42
+ - 支持配置文件驱动训练
43
+ - 实时显示训练进度和验证结果
44
+
45
+ ### 3. dataset_utils.py
46
+ - 数据集加载和预处理
47
+ - 词汇表构建
48
+ - 数据迭代器实现
49
+
50
+ ### 4. get_label.py
51
+ - 提取数据集标签并保存为 `labels.npy`
52
+ - 生成类别名称映射文件
53
+
54
+ ## 使用方法
55
+
56
+ ### 1. 准备数据集
57
+ 将THUCNews数据集放入 `dataset/` 目录:
58
+ ```bash
59
+ # 数据格式:每行一个样本,用tab分隔文本和标签
60
+ text1\t0
61
+ text2\t1
62
+ ...
63
+ ```
64
+
65
+ ### 2. 提取标签
66
+ ```bash
67
+ cd scripts
68
+ python get_label.py --config train.yaml --output ../dataset
69
+ ```
70
+
71
+ ### 3. 训练模型
72
+ ```bash
73
+ cd scripts
74
+ python train.py --config train.yaml
75
+ ```
76
+
77
+ ### 4. 配置文件说明
78
+ 编辑 `scripts/train.yaml` 来调整训练参数:
79
+ ```yaml
80
+ dataset_path: "../dataset" # 数据集路径
81
+ num_epochs: 20 # 训练轮数
82
+ batch_size: 128 # 批次大小
83
+ learning_rate: 0.001 # 学习率
84
+ use_word: false # false=字符级,true=词级
85
+ epochs_dir: "../epochs" # 模型保存路径
86
+ ```
87
+
88
+ ## 可视化数据
89
+
90
+ 训练完成后,每个epoch的数据保存在 `epochs/epoch_N/` 中:
91
+ - `model.pt`: 模型权重文件
92
+ - `embeddings.npy`: 特征向量矩阵 (N_samples, feature_dim)
93
+ - `predictions.npy`: 预测值矩阵 (N_samples, num_classes)
94
+
95
+ 这些数据可以直接用于可视化分析,如t-SNE降维、特征分布分析等。
96
+
97
+ ## 多GPU训练
98
+
99
+ 脚本自动检测可用GPU数量并启用多GPU训练:
100
+ ```python
101
+ # 自动使用所有可用GPU
102
+ if torch.cuda.device_count() > 1:
103
+ model = nn.DataParallel(model)
104
+ ```
105
+
106
+ ## 依赖要求
107
+
108
+ ```bash
109
+ pip install torch numpy scikit-learn tqdm pyyaml
110
+ ```
TextRNN-THUCNews/Classification/dataset/THUCNews/data/class.txt ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ finance
2
+ realty
3
+ stocks
4
+ education
5
+ science
6
+ society
7
+ politics
8
+ sports
9
+ game
10
+ entertainment
TextRNN-THUCNews/Classification/dataset/THUCNews/data/dev.txt ADDED
The diff for this file is too large to render. See raw diff
 
TextRNN-THUCNews/Classification/dataset/THUCNews/data/embedding_SougouNews.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:529917fbfd438697be8019befe41eb12e2b818815cac568a2fb5703782daa198
3
+ size 6339482
TextRNN-THUCNews/Classification/dataset/THUCNews/data/embedding_Tencent.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c3d59d109f48f119a3e04bddd7b8ac4c11245d11c57335f6be2de7ba7130b28
3
+ size 4148151
TextRNN-THUCNews/Classification/dataset/THUCNews/data/test.txt ADDED
The diff for this file is too large to render. See raw diff
 
TextRNN-THUCNews/Classification/dataset/THUCNews/data/train.txt ADDED
The diff for this file is too large to render. See raw diff
 
TextRNN-THUCNews/Classification/dataset/THUCNews/data/vocab.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:512b4100ae5e0063c8c5bfe44c85d2b1062b89d5abbdab6c2f76d80b2803cfd4
3
+ size 75018
TextRNN-THUCNews/Classification/dataset/THUCNews/saved_dict/model.ckpt ADDED
File without changes
TextRNN-THUCNews/Classification/dataset/index.json ADDED
The diff for this file is too large to render. See raw diff
 
TextRNN-THUCNews/Classification/dataset/info.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "TextRNN",
3
+ "classes": [
4
+ "finance",
5
+ "realty",
6
+ "stocks",
7
+ "education",
8
+ "science",
9
+ "society",
10
+ "politics",
11
+ "sports",
12
+ "game",
13
+ "entertainment"
14
+ ]
15
+ }
TextRNN-THUCNews/Classification/dataset/labels.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e892bdf6c821c255196e0fe5685ee33b556a721fdd539da3a1dd6f9feec0d63a
3
+ size 1600128
TextRNN-THUCNews/Classification/epochs/epoch_1/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ced258f666dfe08beb05794317c82636986a072b8bbb7a327b095f82d6fbc9e
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_1/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af42adafecf13f704e20d984d2464f8e2c0e6b06978fb3035ddf194be19ece4d
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_1/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:586e942f7522ec49c13b176138fca3d7aece66f834138d0b94aacabe309f50b5
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_10/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c7830d8730057c6a5b1e35e246f1c69fcf11267df4d376b0125a0da7e939ba0
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_10/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fdbf1e0a62a75b5fee6567201bd2bd643ee0ce141dbb2885216ac2b69323306
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_10/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad3561a873cc88f6157e2d95df0b0aaa8fff35c8af364a66d4d4b45788cb95f7
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_11/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7405cdf19eba953c74968888cc51dd96ad82f377ae59e29091df6c21423338d6
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_11/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb626a2d0bcbc9baea2b11a2017e6559cbe0a6d027ac53c39e4a6b7dbf2c2243
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_11/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f356f26baa2ca648c92fdcc03f340954d8f395cfeacf741475431ae6900af6f7
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_12/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:763a9ea12b1b8eb818fcf55a478b8ca917bf408296d6e11af4c51d631bf24b31
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_12/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0873e22c2605eea54c37eea4bc45f89e92232709571e4a4a4ed07afd372eb493
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_12/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8fd2dc147262e56f2ebcc11c8a1d7aa7aafbe94bda1fde3e223489cfb7e4510
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_13/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61ac32ada3873d4fe294891a2d35301390c07e323e8390545f3665b5e22aaf43
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_13/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ce9d0bf9f214259192880e3a13f7e513a65bad36681ade8ee684d96633a06bd
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_13/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b72e77f92520a0d123e5c01db3d7888edcff9de41336cb66b156649885076c5a
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_14/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:061eec71009d60a74b2ea23f99be06426073980f8bf3b467e718d5b9dfdb8bb6
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_14/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69ebd3299c43022eb852be999160529450ae013b2f43a4455b7dda17f91955c7
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_14/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7655ec6033a80fbc97462be73a35105d1255b24f29b969a57e3dac866c7fa0c
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_15/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2356a84943ccb786a84ea243f7fff205cb8edeb4e26ea54f67ac1377653bb288
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_15/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de2f76ce0e2446a12bffc0bdee82014d02eb3ddae1f6d2f7c67602aa431161fe
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_15/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a95046d0eb6502d3ce3256a26910997718371cf2baa5b9a422fb8ebd22a07c4a
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_16/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ad9deb4a4eb1cfe410978ba5529a8e9a5a713e4904c4445e63d0b3d9373291e
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_16/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40ffb1dc7257441e06901787c640f93e42f00c1da7d9238f7307f933abb369db
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_16/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ac2b25a238e30d630d9ccc115c07368a866e1e3705e70b85ad70c157ba2d9a8
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_17/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f9bea60e2c3ff1dfbbc30fec3368d3969fb2683b8a78202d64713ffaf21eb40
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_17/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:447ff29eea476cc509d51eb81bfa2163281ba07832f0b804a503eb0ba3ab5c3e
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_17/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fba1137f46d7929b4bca54e4f1dd147f47ecfa03efb6d64f1818477d431a7c0
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_18/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fad328d9e087ffdfb15a0b18ef203c22646ba0ed800173d6047e9e6262a644a
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_18/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34a86d4b1ec5ce1e71e802b68a42df269ccc3d469080fd6d8bb2df33c17a802d
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_18/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:006d49643d280160ce5650d0fc1ad0f11c703da261ed19c40b10ea32d1c76a64
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_19/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4445add223cef2b0963bdf4c480ceb4e59d3f648424c1e9eaa3c8b50cbdf8e83
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_19/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5d3e6dfc43405efbb3338a92247cc35a00008d5ab278862a75f22aeabfc2bba
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_19/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d477a57f5f8a5dc2e9fec377c4329399793d33db2a4914b4bdd6228c79f8e9f
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_2/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d70bd9955c63d41e0f7ae0ec5e2c4b4fe3b19aaab3202462fb3c102f6326728
3
+ size 204800128
TextRNN-THUCNews/Classification/epochs/epoch_2/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81562553e0b520b7c35a86fcb5274904dd154e929c92cf68a0c58d660029f6fe
3
+ size 9070272
TextRNN-THUCNews/Classification/epochs/epoch_2/predictions.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78ef1a8e4f54afc661c67d7dcfa70ac6dd1a0fc1b075b6c17599b8d5cefdb4b3
3
+ size 8000128
TextRNN-THUCNews/Classification/epochs/epoch_20/embeddings.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c15e922106177642d4e162b33c027ae6eab6fa0e0235199fa2a31a56d8c23760
3
+ size 204800128