File size: 3,669 Bytes
4335375
 
 
 
 
 
 
6bebb3f
 
 
 
7aba31f
 
 
 
 
 
 
 
4335375
6bebb3f
 
 
7aba31f
 
 
 
 
 
4335375
 
 
 
 
 
7aba31f
 
 
 
 
 
 
 
 
 
 
29f370a
 
 
7aba31f
 
 
 
 
 
29f370a
7aba31f
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
---
license: cc-by-4.0
language:
- en
size_categories:
- 100M<n<1B
configs:
- config_name: channels_clean
  data_files: "channels_clean.parquet"
- config_name: grammy_raw
  data_files: "grammy_raw.parquet"
- config_name: grammy_metadata
  data_files: "grammy_metadata.parquet"
- config_name: grammy_channels
  data_files: "grammy_channels.parquet"
- config_name: grammy_metadata_extended
  data_files: "grammy_metadata_extended.parquet"
- config_name: grammy_timeseries
  data_files: "grammy_timeseries.parquet"
- config_name: music_metadata
  data_files: "music_metadata.parquet"
- config_name: music_video_ids
  data_files: "music_video_ids.parquet"
- config_name: music_comments
  data_files: "music_comments.parquet"
- config_name: cf_item_factors
  data_files: "CF_item_factors.parquet"
- config_name: cf_user_factors
  data_files: "CF_user_factors.parquet"
---

# Preprocessed Data for the ADA Project 2025

## By DataCookers

### General Datasets

* **channels_clean.parquet**: A cleaned version of the `df_channels_en` split from the Youniverse Dataset. This dataset corrects parsing errors where missing values in the **Category** column caused subsequent columns to shift one position to the left.
* **grammy_raw.parquet**: The foundational dataset containing Grammy winners and nominees (1965–2024), sourced from [Kaggle](https://www.kaggle.com/datasets/johnpendenque/grammy-winners-and-nominees-from-1965-to-2024).
* **grammy_metadata.parquet**: A merged dataset combining the *grammy_raw* data with the `yt_metadata_en` split from Youniverse. It is further enriched with web-scraped song lyrics.
* **grammy_channels.parquet**: A filtered subset of `df_channels_en` containing only those YouTube channels that have published a Grammy-nominated or winning song.
* **grammy_metadata_extended.parquet**: An expansion of the metadata containing all videos belonging to the channels identified in *grammy_channels* (not just the Grammy-winning videos).
* **grammy_timeseries.parquet**: Temporal data (time series) specifically associated with the channels found in the *grammy_channels* dataset.
* **music_metadata.parquet**: A subset of the original `yt_metadata_en` Youniverse split, filtered to include only entries classified under the **Music** Category.
* **music_video_ids.parquet**: A lightweight dataset containing the specific video IDs filtered from the `yt_metadata_en` split, strictly for videos listed in the **Music** category.
* **music_comments.parquet**: A collection of user comments obtained by filtering `youtube_comments.tsv.gz`. It retains only comments posted on videos present in the *music_video_ids* dataset.

### Collaborative Filtering Data

The following files represent the output of Matrix Factorization performed on the `merged_comments.tsv.gz` split. 

**Methodology:** 1.  We constructed a sparse interaction matrix of shape `(users, items)` with binary entries (**0** or **1**), where a **1** indicates the user commented on a specific video.
2.  We performed Matrix Factorization using the Alternating Least Squares (ALS) algorithm via the `implicit` library.

**Resulting Files:**

* **CF_item_factors.parquet**: The latent space vectors representing every `video_id` (Item Factors).
* **CF_user_factors.parquet**: The latent space vectors representing every `author` (User Factors).
* **CF_als_model.pkl**: The serialized weights of the trained ALS model.
* **CF_user_id_map.pkl**: A dictionary mapping the **original** `author` (string) to the integer ID used during ALS model training.
* **CF_item_id_map.pkl**: A dictionary mapping the **original** `video_id` (string) to the integer ID used during ALS model training.