yangwang825 commited on
Commit
aeb35d0
·
verified ·
1 Parent(s): 6c9e76f

Create prepare.py

Browse files
Files changed (1) hide show
  1. prepare.py +103 -0
prepare.py ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import random
2
+ from glob import glob
3
+
4
+ from datasets import load_dataset, Dataset, DatasetDict
5
+
6
+ # Get all the token files
7
+ token_files = glob('tokenized/*.tokens')
8
+ total_files = len(token_files)
9
+
10
+ print(f"Found {total_files} token files")
11
+
12
+ # Set the split sizes
13
+ train_size = 23
14
+ dev_size = 8
15
+ test_size = 8
16
+
17
+ # Ensure we have enough files
18
+ if total_files < (train_size + dev_size + test_size):
19
+ print(f"Warning: Not enough files ({total_files}) for the requested split sizes.")
20
+ # Adjust sizes proportionally if needed
21
+ total_requested = train_size + dev_size + test_size
22
+ train_size = int(total_files * (train_size / total_requested))
23
+ dev_size = int(total_files * (dev_size / total_requested))
24
+ test_size = total_files - train_size - dev_size
25
+
26
+ # Randomly shuffle the files
27
+ random.seed(42) # For reproducibility
28
+ random.shuffle(token_files)
29
+
30
+ # Split the files
31
+ train_files = token_files[:train_size]
32
+ dev_files = token_files[train_size:train_size + dev_size]
33
+ test_files = token_files[train_size + dev_size:train_size + dev_size + test_size]
34
+
35
+ # Function to process a list of files and return data points
36
+ def process_files(file_list):
37
+ result = []
38
+ for file in file_list:
39
+ tokens = []
40
+ ner_tags = []
41
+
42
+ with open(file, 'r') as f:
43
+ for line in f:
44
+ line = line.strip()
45
+
46
+ # If empty line, append current document and reset
47
+ if not line:
48
+ if tokens: # Only append if there are tokens (avoid empty entries)
49
+ result.append({
50
+ "tokens": tokens,
51
+ "ner_tags": ner_tags,
52
+ "file_name": file # Include file name for reference
53
+ })
54
+ tokens = []
55
+ ner_tags = []
56
+ continue
57
+
58
+ # Split line into components
59
+ parts = line.split()
60
+
61
+ # Ensure we have at least 3 parts (token, POS, NER)
62
+ if len(parts) >= 3:
63
+ token = parts[0]
64
+ ner_tag = parts[2] # NER tag is the 3rd column
65
+
66
+ tokens.append(token)
67
+ ner_tags.append(ner_tag)
68
+
69
+ # Don't forget to add the last document if it exists
70
+ if tokens:
71
+ result.append({
72
+ "tokens": tokens,
73
+ "ner_tags": ner_tags,
74
+ "file_name": file # Include file name for reference
75
+ })
76
+
77
+ return result
78
+
79
+ # Process each split
80
+ train_data = process_files(train_files)
81
+ dev_data = process_files(dev_files)
82
+ test_data = process_files(test_files)
83
+
84
+ # Create datasets for each split
85
+ train_dataset = Dataset.from_list(train_data)
86
+ dev_dataset = Dataset.from_list(dev_data)
87
+ test_dataset = Dataset.from_list(test_data)
88
+
89
+ # Create a DatasetDict with train, dev, and test splits
90
+ dataset_dict = DatasetDict({
91
+ "train": train_dataset,
92
+ "validation": dev_dataset, # Using 'validation' as it's more standard
93
+ "test": test_dataset
94
+ })
95
+
96
+ # Print some statistics
97
+ print(f"Train split: {len(train_data)} files")
98
+ print(f"Validation split: {len(dev_data)} files")
99
+ print(f"Test split: {len(test_data)} files")
100
+ print(f"Dataset features: {train_dataset.features}")
101
+
102
+ # Uncomment to push to Hub
103
+ dataset_dict.push_to_hub('extraordinarylab/malware-text-db')