Ishaank18 commited on
Commit
2a89c68
·
verified ·
1 Parent(s): ce4106c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -65,18 +65,18 @@ This dataset contains **913 linguistic features** extracted from movie screenpla
65
  from datasets import load_dataset
66
 
67
  # Load a single feature group
68
- ds = load_dataset("YOUR_USERNAME/screenplay-features", data_files="train/base.parquet")
69
  df = ds['train'].to_pandas()
70
 
71
  # Load multiple groups for training
72
- ds = load_dataset("YOUR_USERNAME/screenplay-features",
73
  data_files={
74
  "train": ["train/base.parquet", "train/gc_polarity.parquet", "train/emotional.parquet"]
75
  })
76
  df = ds['train'].to_pandas()
77
 
78
  # Load all splits for evaluation
79
- ds = load_dataset("YOUR_USERNAME/screenplay-features",
80
  data_files={
81
  "train": "train/gc_polarity.parquet",
82
  "validation": "validation/gc_polarity.parquet",
@@ -90,7 +90,7 @@ ds = load_dataset("YOUR_USERNAME/screenplay-features",
90
  import pandas as pd
91
 
92
  # From HuggingFace URL
93
- df = pd.read_parquet("hf://datasets/YOUR_USERNAME/screenplay-features/train/base.parquet")
94
 
95
  # Or if you have the repo cloned locally
96
  df = pd.read_parquet("train/base.parquet")
@@ -105,7 +105,7 @@ from feature_cache.load_hf import load_groups
105
  X, y = load_groups(
106
  groups=["base", "gc_polarity", "emotional", "rst"],
107
  split="train",
108
- hf_repo="YOUR_USERNAME/screenplay-features"
109
  )
110
 
111
  # Load features only (no labels)
@@ -113,7 +113,7 @@ X = load_groups(
113
  groups=["base", "gc_polarity"],
114
  split="test",
115
  include_label=False,
116
- hf_repo="YOUR_USERNAME/screenplay-features"
117
  )
118
  ```
119
 
 
65
  from datasets import load_dataset
66
 
67
  # Load a single feature group
68
+ ds = load_dataset("Ishaank18/screenplay-features", data_files="train/base.parquet")
69
  df = ds['train'].to_pandas()
70
 
71
  # Load multiple groups for training
72
+ ds = load_dataset("Ishaank18/screenplay-features",
73
  data_files={
74
  "train": ["train/base.parquet", "train/gc_polarity.parquet", "train/emotional.parquet"]
75
  })
76
  df = ds['train'].to_pandas()
77
 
78
  # Load all splits for evaluation
79
+ ds = load_dataset("Ishaank18/screenplay-features",
80
  data_files={
81
  "train": "train/gc_polarity.parquet",
82
  "validation": "validation/gc_polarity.parquet",
 
90
  import pandas as pd
91
 
92
  # From HuggingFace URL
93
+ df = pd.read_parquet("hf://datasets/Ishaank18/screenplay-features/train/base.parquet")
94
 
95
  # Or if you have the repo cloned locally
96
  df = pd.read_parquet("train/base.parquet")
 
105
  X, y = load_groups(
106
  groups=["base", "gc_polarity", "emotional", "rst"],
107
  split="train",
108
+ hf_repo="Ishaank18/screenplay-features"
109
  )
110
 
111
  # Load features only (no labels)
 
113
  groups=["base", "gc_polarity"],
114
  split="test",
115
  include_label=False,
116
+ hf_repo="Ishaank18/screenplay-features"
117
  )
118
  ```
119