Spaces:
Sleeping
Sleeping
LorenzoBioinfo
commited on
Commit
·
bce9d7f
1
Parent(s):
9c9babe
Comment datapreparation
Browse files- src/data_preparation.py +4 -4
src/data_preparation.py
CHANGED
|
@@ -44,11 +44,11 @@ def tokenize_function(examples):
|
|
| 44 |
max_length=128,
|
| 45 |
)
|
| 46 |
|
| 47 |
-
tweet_tokenized = tweet_eval.map(tokenize_function, batched=True)
|
| 48 |
-
youtube_tokenized = youtube.map(tokenize_function, batched=True)
|
| 49 |
|
| 50 |
-
tweet_tokenized.save_to_disk(os.path.join(PROCESSED_DIR, "tweet_eval_tokenized"))
|
| 51 |
-
youtube_tokenized.save_to_disk(os.path.join(PROCESSED_DIR, "youtube_tokenized"))
|
| 52 |
|
| 53 |
def prepare_tweet_eval(tokenizer, output_path):
|
| 54 |
print("Scarico e preparo il dataset Tweet Eval...")
|
|
|
|
| 44 |
max_length=128,
|
| 45 |
)
|
| 46 |
|
| 47 |
+
#tweet_tokenized = tweet_eval.map(tokenize_function, batched=True)
|
| 48 |
+
#youtube_tokenized = youtube.map(tokenize_function, batched=True)
|
| 49 |
|
| 50 |
+
#tweet_tokenized.save_to_disk(os.path.join(PROCESSED_DIR, "tweet_eval_tokenized"))
|
| 51 |
+
#youtube_tokenized.save_to_disk(os.path.join(PROCESSED_DIR, "youtube_tokenized"))
|
| 52 |
|
| 53 |
def prepare_tweet_eval(tokenizer, output_path):
|
| 54 |
print("Scarico e preparo il dataset Tweet Eval...")
|