| """ |
| Step 2: BERTopic + UMAP clustering on GPU. |
| Run this on RUNPOD (2x RTX 5090, 64GB VRAM). |
| |
| Input: embeddings.npz + doc_metadata.jsonl (from Step 1) |
| Output: bertopic_results.jsonl (doc_id -> topic assignments + labels) |
| topic_info.json (topic descriptions) |
| umap_coords.npz (2D coordinates for visualization) |
| |
| Install: pip install bertopic cuml-cu12 hdbscan umap-learn plotly |
| (or: pip install bertopic[all] cuml-cu12) |
| """ |
|
|
| import json |
| import time |
| import numpy as np |
|
|
| |
|
|
| WORKSPACE = "/workspace" |
| EMBEDDINGS_FILE = f"{WORKSPACE}/embeddings.npz" |
| METADATA_FILE = f"{WORKSPACE}/doc_metadata.jsonl" |
| OUTPUT_DIR = WORKSPACE |
|
|
| |
| MIN_TOPIC_SIZE = 50 |
| NR_TOPICS = "auto" |
| UMAP_N_NEIGHBORS = 15 |
| UMAP_N_COMPONENTS = 5 |
| UMAP_MIN_DIST = 0.0 |
| UMAP_METRIC = "cosine" |
|
|
| |
| VIZ_N_COMPONENTS = 2 |
| VIZ_N_NEIGHBORS = 15 |
|
|
|
|
| def main(): |
| t_start = time.time() |
|
|
| |
| print("Loading embeddings...") |
| data = np.load(EMBEDDINGS_FILE) |
| embeddings = data["embeddings"] |
| doc_ids = data["doc_ids"] |
| print(f" Shape: {embeddings.shape}, dtype: {embeddings.dtype}") |
| print(f" Memory: {embeddings.nbytes / 1e9:.2f} GB") |
|
|
| print("Loading metadata...") |
| metadata = {} |
| with open(METADATA_FILE) as f: |
| for line in f: |
| d = json.loads(line) |
| metadata[d["id"]] = d |
| print(f" Documents: {len(metadata)}") |
|
|
| |
| try: |
| from cuml.manifold import UMAP as cuUMAP |
| print("\nUsing GPU-accelerated UMAP (cuML)") |
| umap_model = cuUMAP( |
| n_neighbors=UMAP_N_NEIGHBORS, |
| n_components=UMAP_N_COMPONENTS, |
| min_dist=UMAP_MIN_DIST, |
| metric=UMAP_METRIC, |
| random_state=42, |
| ) |
| USE_GPU = True |
| except ImportError: |
| from umap import UMAP |
| print("\nUsing CPU UMAP (cuML not available)") |
| umap_model = UMAP( |
| n_neighbors=UMAP_N_NEIGHBORS, |
| n_components=UMAP_N_COMPONENTS, |
| min_dist=UMAP_MIN_DIST, |
| metric=UMAP_METRIC, |
| random_state=42, |
| low_memory=True, |
| ) |
| USE_GPU = False |
|
|
| |
| try: |
| from cuml.cluster import HDBSCAN as cuHDBSCAN |
| print("Using GPU-accelerated HDBSCAN (cuML)") |
| hdbscan_model = cuHDBSCAN( |
| min_cluster_size=MIN_TOPIC_SIZE, |
| min_samples=10, |
| gen_min_span_tree=True, |
| prediction_data=True, |
| ) |
| except ImportError: |
| from hdbscan import HDBSCAN |
| print("Using CPU HDBSCAN") |
| hdbscan_model = HDBSCAN( |
| min_cluster_size=MIN_TOPIC_SIZE, |
| min_samples=10, |
| gen_min_span_tree=True, |
| prediction_data=True, |
| ) |
|
|
| |
| from bertopic import BERTopic |
| from bertopic.vectorizers import ClassTfidfTransformer |
| from sklearn.feature_extraction.text import CountVectorizer |
|
|
| |
| |
| |
| |
|
|
| print("\nPreparing document texts from metadata...") |
| |
| |
| docs = [] |
| for doc_id in doc_ids: |
| meta = metadata.get(int(doc_id), {}) |
| section = meta.get("section", "unknown") |
| path = meta.get("path", "") |
| fname = path.split("/")[-1] if path else "" |
| docs.append(f"{section} {fname}") |
|
|
| vectorizer = CountVectorizer(stop_words="english", ngram_range=(1, 2)) |
| ctfidf = ClassTfidfTransformer(reduce_frequent_words=True) |
|
|
| print("\nInitializing BERTopic...") |
| topic_model = BERTopic( |
| umap_model=umap_model, |
| hdbscan_model=hdbscan_model, |
| vectorizer_model=vectorizer, |
| ctfidf_model=ctfidf, |
| nr_topics=NR_TOPICS, |
| top_n_words=10, |
| verbose=True, |
| calculate_probabilities=False, |
| ) |
|
|
| |
| print(f"\nFitting BERTopic on {len(embeddings)} documents...") |
| t_fit = time.time() |
| topics, probs = topic_model.fit_transform(docs, embeddings=embeddings) |
| print(f"Fit complete in {(time.time() - t_fit) / 60:.1f} minutes") |
|
|
| |
| topic_info = topic_model.get_topic_info() |
| print(f"\nTopics discovered: {len(topic_info) - 1}") |
| print(f"Outlier documents (topic -1): {(np.array(topics) == -1).sum()}") |
| print("\nTop 20 topics:") |
| print(topic_info.head(20).to_string()) |
|
|
| |
| print("\nComputing 2D UMAP projection for visualization...") |
| t_viz = time.time() |
| try: |
| if USE_GPU: |
| viz_umap = cuUMAP( |
| n_neighbors=VIZ_N_NEIGHBORS, |
| n_components=VIZ_N_COMPONENTS, |
| min_dist=0.1, |
| metric=UMAP_METRIC, |
| random_state=42, |
| ) |
| else: |
| from umap import UMAP |
| viz_umap = UMAP( |
| n_neighbors=VIZ_N_NEIGHBORS, |
| n_components=VIZ_N_COMPONENTS, |
| min_dist=0.1, |
| metric=UMAP_METRIC, |
| random_state=42, |
| low_memory=True, |
| ) |
| coords_2d = viz_umap.fit_transform(embeddings) |
| if hasattr(coords_2d, "to_numpy"): |
| coords_2d = coords_2d.to_numpy() |
| coords_2d = np.array(coords_2d, dtype=np.float32) |
| print(f"2D projection complete in {(time.time() - t_viz) / 60:.1f} minutes") |
| except Exception as e: |
| print(f"2D projection failed: {e}") |
| coords_2d = np.zeros((len(embeddings), 2), dtype=np.float32) |
|
|
| |
| print("\nSaving results...") |
|
|
| |
| results_path = f"{OUTPUT_DIR}/bertopic_results.jsonl" |
| with open(results_path, "w") as f: |
| for i, doc_id in enumerate(doc_ids): |
| meta = metadata.get(int(doc_id), {}) |
| record = { |
| "document_id": int(doc_id), |
| "source_section": meta.get("section", ""), |
| "topic_id": int(topics[i]), |
| "umap_x": float(coords_2d[i][0]), |
| "umap_y": float(coords_2d[i][1]), |
| } |
| f.write(json.dumps(record) + "\n") |
| print(f" {results_path} ({len(doc_ids)} records)") |
|
|
| |
| topic_info_path = f"{OUTPUT_DIR}/topic_info.json" |
| topic_details = {} |
| for topic_id in topic_info["Topic"].unique(): |
| if topic_id == -1: |
| topic_details[-1] = {"label": "Outlier", "words": [], "count": int((np.array(topics) == -1).sum())} |
| continue |
| words = topic_model.get_topic(topic_id) |
| topic_details[int(topic_id)] = { |
| "label": "_".join([w for w, _ in words[:3]]), |
| "words": [{"word": w, "score": float(s)} for w, s in words[:10]], |
| "count": int((np.array(topics) == topic_id).sum()), |
| } |
| with open(topic_info_path, "w") as f: |
| json.dump(topic_details, f, indent=2) |
| print(f" {topic_info_path} ({len(topic_details)} topics)") |
|
|
| |
| coords_path = f"{OUTPUT_DIR}/umap_coords.npz" |
| np.savez_compressed(coords_path, coords=coords_2d, doc_ids=doc_ids, topics=np.array(topics)) |
| print(f" {coords_path}") |
|
|
| |
| model_path = f"{OUTPUT_DIR}/bertopic_model" |
| topic_model.save(model_path, serialization="safetensors", save_ctfidf=True) |
| print(f" {model_path}/") |
|
|
| |
| total_time = (time.time() - t_start) / 60 |
| print(f"\n{'='*60}") |
| print(f"BERTopic clustering complete!") |
| print(f" Documents: {len(doc_ids):,}") |
| print(f" Topics found: {len(topic_details) - 1}") |
| print(f" Outliers: {(np.array(topics) == -1).sum():,}") |
| print(f" Total time: {total_time:.1f} minutes") |
| print(f" GPU used: {USE_GPU}") |
| print(f"\nFiles to transfer back to Hetzner:") |
| print(f" scp {results_path} {topic_info_path} {coords_path} hetzner:/var/www/research/runpod/") |
| print(f"{'='*60}") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|