|
|
""" |
|
|
Fetch data from HuggingFace dataset undertheseanlp/UVB-v0.1 |
|
|
- Get 5,000 high-quality sentences from fiction books |
|
|
- Get 5,000 high-quality sentences from non-fiction books |
|
|
""" |
|
|
|
|
|
import re |
|
|
from os.path import dirname, join |
|
|
|
|
|
from datasets import load_dataset |
|
|
from underthesea import sent_tokenize, text_normalize |
|
|
|
|
|
|
|
|
|
|
|
FICTION_GENRES = { |
|
|
"Fiction", "Novels", "Romance", "Fantasy", "Science Fiction", |
|
|
"Mystery", "Thriller", "Horror", "Historical Fiction", "Literary Fiction", |
|
|
"Adventure", "Crime", "Suspense", "Drama", "Short Stories" |
|
|
} |
|
|
|
|
|
|
|
|
NON_FICTION_GENRES = { |
|
|
"Non Fiction", "Nonfiction", "History", "Biography", "Autobiography", |
|
|
"Self Help", "Psychology", "Philosophy", "Science", "Politics", |
|
|
"Economics", "Business", "Education", "Travel", "Memoir", |
|
|
"Essays", "Reference", "Health", "Religion", "Spirituality" |
|
|
} |
|
|
|
|
|
|
|
|
def clean_text(text): |
|
|
"""Remove formatting and clean text.""" |
|
|
|
|
|
text = text_normalize(text) |
|
|
|
|
|
text = re.sub(r'^#+\s+', '', text, flags=re.MULTILINE) |
|
|
|
|
|
text = re.sub(r'\*+', '', text) |
|
|
|
|
|
text = re.sub(r'^-+$', '', text, flags=re.MULTILINE) |
|
|
|
|
|
text = re.sub(r'\[([^\]]+)\]\([^)]+\)', r'\1', text) |
|
|
|
|
|
text = re.sub(r'\n{2,}', '\n', text) |
|
|
|
|
|
lines = [line.strip() for line in text.split('\n')] |
|
|
text = '\n'.join(lines) |
|
|
return text |
|
|
|
|
|
|
|
|
def is_high_quality_sentence(sent): |
|
|
"""Check if sentence is high quality for UD annotation.""" |
|
|
sent = sent.strip() |
|
|
|
|
|
if not sent: |
|
|
return False, sent |
|
|
|
|
|
|
|
|
if len(sent) < 30: |
|
|
return False, sent |
|
|
if len(sent) > 250: |
|
|
return False, sent |
|
|
|
|
|
|
|
|
words = sent.split() |
|
|
if len(words) < 5: |
|
|
return False, sent |
|
|
if len(words) > 40: |
|
|
return False, sent |
|
|
|
|
|
|
|
|
if not sent[0].isupper(): |
|
|
return False, sent |
|
|
|
|
|
|
|
|
if not sent.rstrip()[-1] in '.!?…"»': |
|
|
return False, sent |
|
|
|
|
|
|
|
|
if sum(1 for c in sent if c.isupper()) > len(sent) * 0.3: |
|
|
return False, sent |
|
|
|
|
|
|
|
|
if not re.search(r'[àáảãạăắằẳẵặâấầẩẫậèéẻẽẹêếềểễệìíỉĩịòóỏõọôốồổỗộơớờởỡợùúủũụưứừửữựỳýỷỹỵđ]', sent, re.IGNORECASE): |
|
|
return False, sent |
|
|
|
|
|
|
|
|
num_digits = sum(1 for c in sent if c.isdigit()) |
|
|
if num_digits > len(sent) * 0.15: |
|
|
return False, sent |
|
|
|
|
|
|
|
|
if re.match(r'^(Chương|Phần|Mục|Điều|\d+\.|\([a-z]\))', sent): |
|
|
return False, sent |
|
|
|
|
|
|
|
|
if re.search(r'(http|www\.|@|\.com|\.vn)', sent, re.IGNORECASE): |
|
|
return False, sent |
|
|
|
|
|
|
|
|
punct_count = sum(1 for c in sent if c in '.,;:!?-–—()[]{}""\'\'«»') |
|
|
if punct_count > len(words) * 1.5: |
|
|
return False, sent |
|
|
|
|
|
|
|
|
if '...' in sent[:-5]: |
|
|
return False, sent |
|
|
|
|
|
|
|
|
quote_count = sent.count('"') + sent.count('"') + sent.count('"') |
|
|
if quote_count > 4: |
|
|
return False, sent |
|
|
|
|
|
return True, sent |
|
|
|
|
|
|
|
|
def classify_book(genres): |
|
|
"""Classify book as fiction or non-fiction based on genres.""" |
|
|
if not genres: |
|
|
return None |
|
|
|
|
|
genres_set = set(genres) |
|
|
|
|
|
is_fiction = bool(genres_set & FICTION_GENRES) |
|
|
is_non_fiction = bool(genres_set & NON_FICTION_GENRES) |
|
|
|
|
|
if is_fiction and not is_non_fiction: |
|
|
return "fiction" |
|
|
elif is_non_fiction and not is_fiction: |
|
|
return "non-fiction" |
|
|
elif is_fiction and is_non_fiction: |
|
|
|
|
|
fiction_count = len(genres_set & FICTION_GENRES) |
|
|
non_fiction_count = len(genres_set & NON_FICTION_GENRES) |
|
|
return "fiction" if fiction_count > non_fiction_count else "non-fiction" |
|
|
|
|
|
return None |
|
|
|
|
|
|
|
|
def extract_sentences_from_book(content, max_sentences=500): |
|
|
"""Extract high-quality sentences from book content.""" |
|
|
content = clean_text(content) |
|
|
sentences = sent_tokenize(content) |
|
|
|
|
|
valid_sentences = [] |
|
|
for sent in sentences: |
|
|
is_valid, cleaned_sent = is_high_quality_sentence(sent) |
|
|
if is_valid: |
|
|
valid_sentences.append(cleaned_sent) |
|
|
if len(valid_sentences) >= max_sentences: |
|
|
break |
|
|
|
|
|
return valid_sentences |
|
|
|
|
|
|
|
|
def fetch_and_process(): |
|
|
print("Loading UVB-v0.1 dataset from HuggingFace...") |
|
|
ds = load_dataset("undertheseanlp/UVB-v0.1", split="train") |
|
|
|
|
|
print(f"Total books in dataset: {len(ds)}") |
|
|
|
|
|
|
|
|
fiction_books = [] |
|
|
non_fiction_books = [] |
|
|
|
|
|
for book in ds: |
|
|
genres = book.get("genres", []) |
|
|
rating = book.get("goodreads_rating", 0) or 0 |
|
|
num_ratings = book.get("goodreads_num_ratings", 0) or 0 |
|
|
|
|
|
|
|
|
quality_score = rating * min(num_ratings / 100, 10) |
|
|
|
|
|
book_type = classify_book(genres) |
|
|
book_info = { |
|
|
"title": book["title"], |
|
|
"content": book["content"], |
|
|
"rating": rating, |
|
|
"num_ratings": num_ratings, |
|
|
"quality_score": quality_score, |
|
|
"genres": genres |
|
|
} |
|
|
|
|
|
if book_type == "fiction": |
|
|
fiction_books.append(book_info) |
|
|
elif book_type == "non-fiction": |
|
|
non_fiction_books.append(book_info) |
|
|
|
|
|
print(f"Fiction books: {len(fiction_books)}") |
|
|
print(f"Non-fiction books: {len(non_fiction_books)}") |
|
|
|
|
|
|
|
|
fiction_books.sort(key=lambda x: x["quality_score"], reverse=True) |
|
|
non_fiction_books.sort(key=lambda x: x["quality_score"], reverse=True) |
|
|
|
|
|
|
|
|
print("\nExtracting sentences from fiction books...") |
|
|
fiction_sentences = [] |
|
|
for i, book in enumerate(fiction_books): |
|
|
if len(fiction_sentences) >= 5000: |
|
|
break |
|
|
sentences = extract_sentences_from_book(book["content"]) |
|
|
for sent in sentences: |
|
|
if len(fiction_sentences) >= 5000: |
|
|
break |
|
|
fiction_sentences.append(sent) |
|
|
print(f" [{i+1}/{len(fiction_books)}] {book['title'][:50]} - {len(sentences)} sentences (total: {len(fiction_sentences)})") |
|
|
|
|
|
|
|
|
print("\nExtracting sentences from non-fiction books...") |
|
|
non_fiction_sentences = [] |
|
|
for i, book in enumerate(non_fiction_books): |
|
|
if len(non_fiction_sentences) >= 5000: |
|
|
break |
|
|
sentences = extract_sentences_from_book(book["content"]) |
|
|
for sent in sentences: |
|
|
if len(non_fiction_sentences) >= 5000: |
|
|
break |
|
|
non_fiction_sentences.append(sent) |
|
|
print(f" [{i+1}/{len(non_fiction_books)}] {book['title'][:50]} - {len(sentences)} sentences (total: {len(non_fiction_sentences)})") |
|
|
|
|
|
print(f"\nFiction sentences collected: {len(fiction_sentences)}") |
|
|
print(f"Non-fiction sentences collected: {len(non_fiction_sentences)}") |
|
|
|
|
|
|
|
|
all_sentences = fiction_sentences[:5000] + non_fiction_sentences[:5000] |
|
|
print(f"Total sentences: {len(all_sentences)}") |
|
|
|
|
|
|
|
|
output_dir = dirname(dirname(__file__)) |
|
|
output_file = join(output_dir, "sentences_uvb.txt") |
|
|
|
|
|
with open(output_file, "w", encoding="utf-8") as f: |
|
|
for i, sent in enumerate(all_sentences, 1): |
|
|
source = "fiction" if i <= len(fiction_sentences[:5000]) else "non-fiction" |
|
|
f.write(f"{i}\t{source}\t{sent}\n") |
|
|
|
|
|
print(f"\nSaved to: {output_file}") |
|
|
|
|
|
|
|
|
print("\nSample fiction sentences:") |
|
|
for i, sent in enumerate(fiction_sentences[:3], 1): |
|
|
print(f" {i}. {sent[:100]}...") |
|
|
|
|
|
print("\nSample non-fiction sentences:") |
|
|
for i, sent in enumerate(non_fiction_sentences[:3], 1): |
|
|
print(f" {i}. {sent[:100]}...") |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
fetch_and_process() |
|
|
|