feat: add TextRNN
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- TextCNN-THUCNews/Classification/scripts/verify_embedding_order.py +0 -235
- TextRNN-THUCNews/Classification/README.md +110 -0
- TextRNN-THUCNews/Classification/dataset/THUCNews/data/class.txt +10 -0
- TextRNN-THUCNews/Classification/dataset/THUCNews/data/dev.txt +0 -0
- TextRNN-THUCNews/Classification/dataset/THUCNews/data/embedding_SougouNews.npz +3 -0
- TextRNN-THUCNews/Classification/dataset/THUCNews/data/embedding_Tencent.npz +3 -0
- TextRNN-THUCNews/Classification/dataset/THUCNews/data/test.txt +0 -0
- TextRNN-THUCNews/Classification/dataset/THUCNews/data/train.txt +0 -0
- TextRNN-THUCNews/Classification/dataset/THUCNews/data/vocab.pkl +3 -0
- TextRNN-THUCNews/Classification/dataset/THUCNews/saved_dict/model.ckpt +0 -0
- TextRNN-THUCNews/Classification/dataset/index.json +0 -0
- TextRNN-THUCNews/Classification/dataset/info.json +15 -0
- TextRNN-THUCNews/Classification/dataset/labels.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_1/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_1/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_1/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_10/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_10/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_10/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_11/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_11/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_11/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_12/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_12/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_12/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_13/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_13/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_13/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_14/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_14/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_14/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_15/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_15/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_15/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_16/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_16/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_16/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_17/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_17/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_17/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_18/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_18/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_18/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_19/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_19/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_19/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_2/embeddings.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_2/model.pt +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_2/predictions.npy +3 -0
- TextRNN-THUCNews/Classification/epochs/epoch_20/embeddings.npy +3 -0
TextCNN-THUCNews/Classification/scripts/verify_embedding_order.py
DELETED
|
@@ -1,235 +0,0 @@
|
|
| 1 |
-
# coding: UTF-8
|
| 2 |
-
"""
|
| 3 |
-
验证embedding保存顺序与原始数据集顺序一致性的脚本
|
| 4 |
-
"""
|
| 5 |
-
import os
|
| 6 |
-
import sys
|
| 7 |
-
import numpy as np
|
| 8 |
-
import yaml
|
| 9 |
-
import torch
|
| 10 |
-
from tqdm import tqdm
|
| 11 |
-
|
| 12 |
-
# 添加当前目录到路径
|
| 13 |
-
sys.path.append(os.path.dirname(os.path.abspath(__file__)))
|
| 14 |
-
|
| 15 |
-
from model import Model, Config
|
| 16 |
-
from dataset_utils import build_dataset, build_iterator
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
def verify_embedding_order(config_path, epoch_num=1):
|
| 20 |
-
"""
|
| 21 |
-
验证指定epoch的embedding顺序是否与原始数据集一致
|
| 22 |
-
|
| 23 |
-
Args:
|
| 24 |
-
config_path: 配置文件路径
|
| 25 |
-
epoch_num: 要验证的epoch数
|
| 26 |
-
"""
|
| 27 |
-
print("=" * 60)
|
| 28 |
-
print("验证Embedding保存顺序与原始数据集一致性")
|
| 29 |
-
print("=" * 60)
|
| 30 |
-
|
| 31 |
-
# 加载配置
|
| 32 |
-
with open(config_path, 'r', encoding='utf-8') as f:
|
| 33 |
-
train_config = yaml.safe_load(f)
|
| 34 |
-
|
| 35 |
-
# 初始化配置
|
| 36 |
-
config = Config(train_config['dataset_path'], train_config.get('embedding', 'random'))
|
| 37 |
-
config.batch_size = train_config.get('batch_size', 128)
|
| 38 |
-
|
| 39 |
-
# 构建数据集
|
| 40 |
-
print("1. 构建数据集...")
|
| 41 |
-
vocab, train_data, dev_data, test_data = build_dataset(config, train_config.get('use_word', False))
|
| 42 |
-
config.n_vocab = len(vocab)
|
| 43 |
-
|
| 44 |
-
print(f" 训练数据集大小: {len(train_data)}")
|
| 45 |
-
print(f" 验证数据集大小: {len(dev_data)}")
|
| 46 |
-
print(f" 测试数据集大小: {len(test_data)}")
|
| 47 |
-
print(f" 总数据集大小: {len(train_data) + len(dev_data) + len(test_data)}")
|
| 48 |
-
|
| 49 |
-
# 检查保存的embedding文件是否存在
|
| 50 |
-
epochs_dir = train_config.get('epochs_dir', '../epochs')
|
| 51 |
-
epoch_dir = os.path.join(epochs_dir, f'epoch_{epoch_num}')
|
| 52 |
-
embeddings_path = os.path.join(epoch_dir, 'embeddings.npy')
|
| 53 |
-
predictions_path = os.path.join(epoch_dir, 'predictions.npy')
|
| 54 |
-
dataset_info_path = os.path.join(epoch_dir, 'dataset_info.json')
|
| 55 |
-
|
| 56 |
-
if not os.path.exists(embeddings_path):
|
| 57 |
-
print(f"❌ 找不到embedding文件: {embeddings_path}")
|
| 58 |
-
print("请先运行训练脚本生成embedding文件")
|
| 59 |
-
return False
|
| 60 |
-
|
| 61 |
-
print(f"2. 加载epoch {epoch_num}的embedding文件...")
|
| 62 |
-
saved_embeddings = np.load(embeddings_path)
|
| 63 |
-
saved_predictions = np.load(predictions_path)
|
| 64 |
-
|
| 65 |
-
print(f" 保存的embedding形状: {saved_embeddings.shape}")
|
| 66 |
-
print(f" 保存的prediction形状: {saved_predictions.shape}")
|
| 67 |
-
|
| 68 |
-
# 加载数据集信息
|
| 69 |
-
dataset_info = []
|
| 70 |
-
if os.path.exists(dataset_info_path):
|
| 71 |
-
import json
|
| 72 |
-
with open(dataset_info_path, 'r', encoding='utf-8') as f:
|
| 73 |
-
dataset_info = json.load(f)
|
| 74 |
-
print(" 数据集信息:")
|
| 75 |
-
for info in dataset_info:
|
| 76 |
-
print(f" {info['name']}: {info['size']} 个样本, 索引 {info['start_idx']}-{info['end_idx']-1}")
|
| 77 |
-
else:
|
| 78 |
-
print(" ⚠️ 未找到dataset_info.json,假设只有训练数据")
|
| 79 |
-
|
| 80 |
-
# 加载模型
|
| 81 |
-
print("3. 加载模型...")
|
| 82 |
-
device = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')
|
| 83 |
-
config.device = device
|
| 84 |
-
|
| 85 |
-
model = Model(config).to(device)
|
| 86 |
-
model_path = os.path.join(epoch_dir, 'model.pt')
|
| 87 |
-
model.load_state_dict(torch.load(model_path, map_location=device))
|
| 88 |
-
model.eval()
|
| 89 |
-
|
| 90 |
-
# 重新生成embedding(使用新的迭代器,按train/dev/test顺序)
|
| 91 |
-
print("4. 重新生成embedding以验证顺序...")
|
| 92 |
-
|
| 93 |
-
# 按顺序处理三个数据集
|
| 94 |
-
datasets = [
|
| 95 |
-
('train', train_data),
|
| 96 |
-
('dev', dev_data),
|
| 97 |
-
('test', test_data)
|
| 98 |
-
]
|
| 99 |
-
|
| 100 |
-
fresh_features = []
|
| 101 |
-
fresh_predictions = []
|
| 102 |
-
|
| 103 |
-
with torch.no_grad():
|
| 104 |
-
for dataset_name, dataset in datasets:
|
| 105 |
-
print(f" 重新生成 {dataset_name} 数据集特征...")
|
| 106 |
-
|
| 107 |
-
data_iter = build_iterator(dataset, config)
|
| 108 |
-
dataset_features = []
|
| 109 |
-
dataset_predictions = []
|
| 110 |
-
|
| 111 |
-
for texts, labels in tqdm(data_iter, desc=f"重新生成{dataset_name}特征"):
|
| 112 |
-
features = model.feature(texts)
|
| 113 |
-
predictions = model.get_prediction(texts)
|
| 114 |
-
|
| 115 |
-
dataset_features.append(features)
|
| 116 |
-
dataset_predictions.append(predictions)
|
| 117 |
-
|
| 118 |
-
if dataset_features:
|
| 119 |
-
dataset_embeddings = np.vstack(dataset_features)
|
| 120 |
-
dataset_preds = np.vstack(dataset_predictions)
|
| 121 |
-
|
| 122 |
-
fresh_features.append(dataset_embeddings)
|
| 123 |
-
fresh_predictions.append(dataset_preds)
|
| 124 |
-
|
| 125 |
-
print(f" {dataset_name} 重新生成形状: {dataset_embeddings.shape}")
|
| 126 |
-
|
| 127 |
-
fresh_embeddings = np.vstack(fresh_features)
|
| 128 |
-
fresh_predictions_array = np.vstack(fresh_predictions)
|
| 129 |
-
|
| 130 |
-
print(f" 重新生成的总embedding形状: {fresh_embeddings.shape}")
|
| 131 |
-
print(f" 重新生成的总prediction形状: {fresh_predictions_array.shape}")
|
| 132 |
-
|
| 133 |
-
# 比较两次生成的结果
|
| 134 |
-
print("5. 比较结果...")
|
| 135 |
-
|
| 136 |
-
# 检查形状是否一致
|
| 137 |
-
if saved_embeddings.shape != fresh_embeddings.shape:
|
| 138 |
-
print("❌ 形状不一致!")
|
| 139 |
-
print(f" 保存的: {saved_embeddings.shape}")
|
| 140 |
-
print(f" 重新生成的: {fresh_embeddings.shape}")
|
| 141 |
-
return False
|
| 142 |
-
|
| 143 |
-
# 检查数值是否一致(允许小的浮点误差)
|
| 144 |
-
embedding_diff = np.abs(saved_embeddings - fresh_embeddings)
|
| 145 |
-
max_embedding_diff = np.max(embedding_diff)
|
| 146 |
-
mean_embedding_diff = np.mean(embedding_diff)
|
| 147 |
-
|
| 148 |
-
prediction_diff = np.abs(saved_predictions - fresh_predictions_array)
|
| 149 |
-
max_prediction_diff = np.max(prediction_diff)
|
| 150 |
-
mean_prediction_diff = np.mean(prediction_diff)
|
| 151 |
-
|
| 152 |
-
print(f" Embedding最大差异: {max_embedding_diff:.10f}")
|
| 153 |
-
print(f" Embedding平均差异: {mean_embedding_diff:.10f}")
|
| 154 |
-
print(f" Prediction最大差异: {max_prediction_diff:.10f}")
|
| 155 |
-
print(f" Prediction平均差异: {mean_prediction_diff:.10f}")
|
| 156 |
-
|
| 157 |
-
# 判断是否一致(允许1e-6的误差)
|
| 158 |
-
tolerance = 1e-6
|
| 159 |
-
embedding_consistent = max_embedding_diff < tolerance
|
| 160 |
-
prediction_consistent = max_prediction_diff < tolerance
|
| 161 |
-
|
| 162 |
-
if embedding_consistent and prediction_consistent:
|
| 163 |
-
print("✅ 验证通过!Embedding和Prediction的保存顺序与原始数据集完全一致")
|
| 164 |
-
|
| 165 |
-
# 额外验证:检查不同数据集的样本对应关系
|
| 166 |
-
print("\n6. 额外验证各数据集样本的对应关系...")
|
| 167 |
-
|
| 168 |
-
# 合并所有数据集用于验证
|
| 169 |
-
all_data = train_data + dev_data + test_data
|
| 170 |
-
|
| 171 |
-
# 验证前几个样本和各数据集边界处的样本
|
| 172 |
-
test_indices = []
|
| 173 |
-
|
| 174 |
-
# 添加前5个样本
|
| 175 |
-
test_indices.extend(range(min(5, len(all_data))))
|
| 176 |
-
|
| 177 |
-
# 如果有数据集信息,添加边界处的样本
|
| 178 |
-
if dataset_info:
|
| 179 |
-
for info in dataset_info:
|
| 180 |
-
# 添加每个数据集开始和结束的样本
|
| 181 |
-
start_idx = info['start_idx']
|
| 182 |
-
end_idx = info['end_idx'] - 1
|
| 183 |
-
test_indices.extend([start_idx, end_idx])
|
| 184 |
-
|
| 185 |
-
# 去重并排序
|
| 186 |
-
test_indices = sorted(list(set(test_indices)))
|
| 187 |
-
test_indices = [i for i in test_indices if i < len(all_data)]
|
| 188 |
-
|
| 189 |
-
for i in test_indices:
|
| 190 |
-
original_label = all_data[i][1]
|
| 191 |
-
|
| 192 |
-
# 从保存的prediction中获取预测结果
|
| 193 |
-
saved_pred_logits = saved_predictions[i]
|
| 194 |
-
saved_pred_class = np.argmax(saved_pred_logits)
|
| 195 |
-
|
| 196 |
-
# 确定样本属于哪个数据集
|
| 197 |
-
dataset_name = "unknown"
|
| 198 |
-
if dataset_info:
|
| 199 |
-
for info in dataset_info:
|
| 200 |
-
if info['start_idx'] <= i < info['end_idx']:
|
| 201 |
-
dataset_name = info['name']
|
| 202 |
-
break
|
| 203 |
-
|
| 204 |
-
print(f" 样本{i}: {dataset_name}数据集, 原始标签={original_label}, 预测类别={saved_pred_class}")
|
| 205 |
-
|
| 206 |
-
return True
|
| 207 |
-
else:
|
| 208 |
-
print("❌ 验证失败!数据顺序可能不一致")
|
| 209 |
-
if not embedding_consistent:
|
| 210 |
-
print(f" Embedding不一致,最大差异: {max_embedding_diff}")
|
| 211 |
-
if not prediction_consistent:
|
| 212 |
-
print(f" Prediction不一致,最大差异: {max_prediction_diff}")
|
| 213 |
-
return False
|
| 214 |
-
|
| 215 |
-
|
| 216 |
-
def main():
|
| 217 |
-
"""主函数"""
|
| 218 |
-
import argparse
|
| 219 |
-
|
| 220 |
-
parser = argparse.ArgumentParser(description='验证embedding保存顺序')
|
| 221 |
-
parser.add_argument('--config', type=str, default='train.yaml', help='配置文件路径')
|
| 222 |
-
parser.add_argument('--epoch', type=int, default=1, help='要验证的epoch数')
|
| 223 |
-
|
| 224 |
-
args = parser.parse_args()
|
| 225 |
-
|
| 226 |
-
success = verify_embedding_order(args.config, args.epoch)
|
| 227 |
-
|
| 228 |
-
if success:
|
| 229 |
-
print("\n🎉 验证成功!可以放心使用保存的embedding进行可视化分析。")
|
| 230 |
-
else:
|
| 231 |
-
print("\n⚠️ 验证失败!请检查训练脚本的embedding保存逻辑。")
|
| 232 |
-
|
| 233 |
-
|
| 234 |
-
if __name__ == '__main__':
|
| 235 |
-
main()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
TextRNN-THUCNews/Classification/README.md
ADDED
|
@@ -0,0 +1,110 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# TextRNN 可视化实验脚本
|
| 2 |
+
|
| 3 |
+
基于原始 Chinese-Text-Classification-Pytorch 仓库重构的 TextRNN 训练脚本,专门用于可视化实验。
|
| 4 |
+
|
| 5 |
+
## 目录结构
|
| 6 |
+
|
| 7 |
+
```
|
| 8 |
+
Classification/
|
| 9 |
+
├── scripts/ # 脚本文件
|
| 10 |
+
│ ├── model.py # 模型定义,包含feature、get_prediction、prediction函数
|
| 11 |
+
│ ├── train.py # 训练脚本,支持多卡训练
|
| 12 |
+
│ ├── train.yaml # 训练配置文件
|
| 13 |
+
│ ├── dataset_utils.py # 数据集处理工具
|
| 14 |
+
│ └── get_label.py # 标签提取脚本
|
| 15 |
+
├── dataset/ # 数据集文件
|
| 16 |
+
│ ├── train.txt # 训练数据
|
| 17 |
+
│ ├── dev.txt # 验证数据
|
| 18 |
+
│ ├── test.txt # 测试数据
|
| 19 |
+
│ ├── class.txt # 类别列表
|
| 20 |
+
│ ├── vocab.pkl # 词汇表
|
| 21 |
+
│ └── labels.npy # 提取的标签
|
| 22 |
+
└── epochs/ # 按epoch存放模型文件和特征向量
|
| 23 |
+
├── epoch_1/
|
| 24 |
+
│ ├── model.pt # 模型权重
|
| 25 |
+
│ ├── embeddings.npy # 特征向量
|
| 26 |
+
│ └── predictions.npy # 预测值
|
| 27 |
+
└── epoch_2/
|
| 28 |
+
└── ...
|
| 29 |
+
```
|
| 30 |
+
|
| 31 |
+
## 功能说明
|
| 32 |
+
|
| 33 |
+
### 1. model.py
|
| 34 |
+
- **Model类**: TextRNN模型实现
|
| 35 |
+
- **feature()**: 提取中间层特征向量(dropout层输出),用于可视化
|
| 36 |
+
- **get_prediction()**: 获取模型最终层输出向量(logits)
|
| 37 |
+
- **prediction()**: 根据中间特征向量预测结果
|
| 38 |
+
|
| 39 |
+
### 2. train.py
|
| 40 |
+
- 支持多GPU训练
|
| 41 |
+
- 每个epoch自动保存模型、特征向量、预测值到 `epochs/epoch_N/`
|
| 42 |
+
- 支持配置文件驱动训练
|
| 43 |
+
- 实时显示训练进度和验证结果
|
| 44 |
+
|
| 45 |
+
### 3. dataset_utils.py
|
| 46 |
+
- 数据集加载和预处理
|
| 47 |
+
- 词汇表构建
|
| 48 |
+
- 数据迭代器实现
|
| 49 |
+
|
| 50 |
+
### 4. get_label.py
|
| 51 |
+
- 提取数据集标签并保存为 `labels.npy`
|
| 52 |
+
- 生成类别名称映射文件
|
| 53 |
+
|
| 54 |
+
## 使用方法
|
| 55 |
+
|
| 56 |
+
### 1. 准备数据集
|
| 57 |
+
将THUCNews数据集放入 `dataset/` 目录:
|
| 58 |
+
```bash
|
| 59 |
+
# 数据格式:每行一个样本,用tab分隔文本和标签
|
| 60 |
+
text1\t0
|
| 61 |
+
text2\t1
|
| 62 |
+
...
|
| 63 |
+
```
|
| 64 |
+
|
| 65 |
+
### 2. 提取标签
|
| 66 |
+
```bash
|
| 67 |
+
cd scripts
|
| 68 |
+
python get_label.py --config train.yaml --output ../dataset
|
| 69 |
+
```
|
| 70 |
+
|
| 71 |
+
### 3. 训练模型
|
| 72 |
+
```bash
|
| 73 |
+
cd scripts
|
| 74 |
+
python train.py --config train.yaml
|
| 75 |
+
```
|
| 76 |
+
|
| 77 |
+
### 4. 配置文件说明
|
| 78 |
+
编辑 `scripts/train.yaml` 来调整训练参数:
|
| 79 |
+
```yaml
|
| 80 |
+
dataset_path: "../dataset" # 数据集路径
|
| 81 |
+
num_epochs: 20 # 训练轮数
|
| 82 |
+
batch_size: 128 # 批次大小
|
| 83 |
+
learning_rate: 0.001 # 学习率
|
| 84 |
+
use_word: false # false=字符级,true=词级
|
| 85 |
+
epochs_dir: "../epochs" # 模型保存路径
|
| 86 |
+
```
|
| 87 |
+
|
| 88 |
+
## 可视化数据
|
| 89 |
+
|
| 90 |
+
训练完成后,每个epoch的数据保存在 `epochs/epoch_N/` 中:
|
| 91 |
+
- `model.pt`: 模型权重文件
|
| 92 |
+
- `embeddings.npy`: 特征向量矩阵 (N_samples, feature_dim)
|
| 93 |
+
- `predictions.npy`: 预测值矩阵 (N_samples, num_classes)
|
| 94 |
+
|
| 95 |
+
这些数据可以直接用于可视化分析,如t-SNE降维、特征分布分析等。
|
| 96 |
+
|
| 97 |
+
## 多GPU训练
|
| 98 |
+
|
| 99 |
+
脚本自动检测可用GPU数量并启用多GPU训练:
|
| 100 |
+
```python
|
| 101 |
+
# 自动使用所有可用GPU
|
| 102 |
+
if torch.cuda.device_count() > 1:
|
| 103 |
+
model = nn.DataParallel(model)
|
| 104 |
+
```
|
| 105 |
+
|
| 106 |
+
## 依赖要求
|
| 107 |
+
|
| 108 |
+
```bash
|
| 109 |
+
pip install torch numpy scikit-learn tqdm pyyaml
|
| 110 |
+
```
|
TextRNN-THUCNews/Classification/dataset/THUCNews/data/class.txt
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
finance
|
| 2 |
+
realty
|
| 3 |
+
stocks
|
| 4 |
+
education
|
| 5 |
+
science
|
| 6 |
+
society
|
| 7 |
+
politics
|
| 8 |
+
sports
|
| 9 |
+
game
|
| 10 |
+
entertainment
|
TextRNN-THUCNews/Classification/dataset/THUCNews/data/dev.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
TextRNN-THUCNews/Classification/dataset/THUCNews/data/embedding_SougouNews.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:529917fbfd438697be8019befe41eb12e2b818815cac568a2fb5703782daa198
|
| 3 |
+
size 6339482
|
TextRNN-THUCNews/Classification/dataset/THUCNews/data/embedding_Tencent.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8c3d59d109f48f119a3e04bddd7b8ac4c11245d11c57335f6be2de7ba7130b28
|
| 3 |
+
size 4148151
|
TextRNN-THUCNews/Classification/dataset/THUCNews/data/test.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
TextRNN-THUCNews/Classification/dataset/THUCNews/data/train.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
TextRNN-THUCNews/Classification/dataset/THUCNews/data/vocab.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:512b4100ae5e0063c8c5bfe44c85d2b1062b89d5abbdab6c2f76d80b2803cfd4
|
| 3 |
+
size 75018
|
TextRNN-THUCNews/Classification/dataset/THUCNews/saved_dict/model.ckpt
ADDED
|
File without changes
|
TextRNN-THUCNews/Classification/dataset/index.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
TextRNN-THUCNews/Classification/dataset/info.json
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model": "TextRNN",
|
| 3 |
+
"classes": [
|
| 4 |
+
"finance",
|
| 5 |
+
"realty",
|
| 6 |
+
"stocks",
|
| 7 |
+
"education",
|
| 8 |
+
"science",
|
| 9 |
+
"society",
|
| 10 |
+
"politics",
|
| 11 |
+
"sports",
|
| 12 |
+
"game",
|
| 13 |
+
"entertainment"
|
| 14 |
+
]
|
| 15 |
+
}
|
TextRNN-THUCNews/Classification/dataset/labels.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e892bdf6c821c255196e0fe5685ee33b556a721fdd539da3a1dd6f9feec0d63a
|
| 3 |
+
size 1600128
|
TextRNN-THUCNews/Classification/epochs/epoch_1/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8ced258f666dfe08beb05794317c82636986a072b8bbb7a327b095f82d6fbc9e
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_1/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:af42adafecf13f704e20d984d2464f8e2c0e6b06978fb3035ddf194be19ece4d
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_1/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:586e942f7522ec49c13b176138fca3d7aece66f834138d0b94aacabe309f50b5
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_10/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c7830d8730057c6a5b1e35e246f1c69fcf11267df4d376b0125a0da7e939ba0
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_10/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1fdbf1e0a62a75b5fee6567201bd2bd643ee0ce141dbb2885216ac2b69323306
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_10/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ad3561a873cc88f6157e2d95df0b0aaa8fff35c8af364a66d4d4b45788cb95f7
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_11/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7405cdf19eba953c74968888cc51dd96ad82f377ae59e29091df6c21423338d6
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_11/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fb626a2d0bcbc9baea2b11a2017e6559cbe0a6d027ac53c39e4a6b7dbf2c2243
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_11/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f356f26baa2ca648c92fdcc03f340954d8f395cfeacf741475431ae6900af6f7
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_12/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:763a9ea12b1b8eb818fcf55a478b8ca917bf408296d6e11af4c51d631bf24b31
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_12/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0873e22c2605eea54c37eea4bc45f89e92232709571e4a4a4ed07afd372eb493
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_12/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f8fd2dc147262e56f2ebcc11c8a1d7aa7aafbe94bda1fde3e223489cfb7e4510
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_13/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:61ac32ada3873d4fe294891a2d35301390c07e323e8390545f3665b5e22aaf43
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_13/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2ce9d0bf9f214259192880e3a13f7e513a65bad36681ade8ee684d96633a06bd
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_13/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b72e77f92520a0d123e5c01db3d7888edcff9de41336cb66b156649885076c5a
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_14/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:061eec71009d60a74b2ea23f99be06426073980f8bf3b467e718d5b9dfdb8bb6
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_14/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:69ebd3299c43022eb852be999160529450ae013b2f43a4455b7dda17f91955c7
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_14/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e7655ec6033a80fbc97462be73a35105d1255b24f29b969a57e3dac866c7fa0c
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_15/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2356a84943ccb786a84ea243f7fff205cb8edeb4e26ea54f67ac1377653bb288
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_15/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:de2f76ce0e2446a12bffc0bdee82014d02eb3ddae1f6d2f7c67602aa431161fe
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_15/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a95046d0eb6502d3ce3256a26910997718371cf2baa5b9a422fb8ebd22a07c4a
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_16/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8ad9deb4a4eb1cfe410978ba5529a8e9a5a713e4904c4445e63d0b3d9373291e
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_16/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:40ffb1dc7257441e06901787c640f93e42f00c1da7d9238f7307f933abb369db
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_16/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6ac2b25a238e30d630d9ccc115c07368a866e1e3705e70b85ad70c157ba2d9a8
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_17/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9f9bea60e2c3ff1dfbbc30fec3368d3969fb2683b8a78202d64713ffaf21eb40
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_17/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:447ff29eea476cc509d51eb81bfa2163281ba07832f0b804a503eb0ba3ab5c3e
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_17/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2fba1137f46d7929b4bca54e4f1dd147f47ecfa03efb6d64f1818477d431a7c0
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_18/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5fad328d9e087ffdfb15a0b18ef203c22646ba0ed800173d6047e9e6262a644a
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_18/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:34a86d4b1ec5ce1e71e802b68a42df269ccc3d469080fd6d8bb2df33c17a802d
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_18/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:006d49643d280160ce5650d0fc1ad0f11c703da261ed19c40b10ea32d1c76a64
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_19/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4445add223cef2b0963bdf4c480ceb4e59d3f648424c1e9eaa3c8b50cbdf8e83
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_19/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a5d3e6dfc43405efbb3338a92247cc35a00008d5ab278862a75f22aeabfc2bba
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_19/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4d477a57f5f8a5dc2e9fec377c4329399793d33db2a4914b4bdd6228c79f8e9f
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_2/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4d70bd9955c63d41e0f7ae0ec5e2c4b4fe3b19aaab3202462fb3c102f6326728
|
| 3 |
+
size 204800128
|
TextRNN-THUCNews/Classification/epochs/epoch_2/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:81562553e0b520b7c35a86fcb5274904dd154e929c92cf68a0c58d660029f6fe
|
| 3 |
+
size 9070272
|
TextRNN-THUCNews/Classification/epochs/epoch_2/predictions.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:78ef1a8e4f54afc661c67d7dcfa70ac6dd1a0fc1b075b6c17599b8d5cefdb4b3
|
| 3 |
+
size 8000128
|
TextRNN-THUCNews/Classification/epochs/epoch_20/embeddings.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c15e922106177642d4e162b33c027ae6eab6fa0e0235199fa2a31a56d8c23760
|
| 3 |
+
size 204800128
|