Datasets:
License:
File size: 2,205 Bytes
34395ff f9b9f5b 34395ff f9b9f5b 34395ff f9b9f5b 34395ff f9b9f5b 34395ff | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 | ---
license: mit
language:
- zh
- en
- de
- fr
task_categories:
- feature-extraction
- text-classification
tags:
- embeddings
- sociology
- retrieval
- sentence-transformers
- numpy
- qwen3
pretty_name: THETA Embeddings
---
# THETA-embeddings
Pre-computed embeddings generated by [THETA](https://huggingface.co/CodeSoulco/THETA), a domain-specific embedding model fine-tuned on Qwen3-Embedding for sociology and social science texts.
## Description
This dataset contains dense vector embeddings produced under three settings:
- **zero_shot:** Embeddings from the base Qwen3-Embedding model without fine-tuning
- **supervised:** Embeddings from the LoRA-adapted model trained with label-guided contrastive learning
- **unsupervised:** Embeddings from the LoRA-adapted model trained with SimCSE
## Repository Structure
```
CodeSoulco/THETA-embeddings/
├── 0.6B/
│ ├── zero_shot/
│ ├── supervised/
│ └── unsupervised/
└── 4B/
├── zero_shot/
├── supervised/
└── unsupervised/
```
## Embedding Details
| Model | Dimension | Format |
|---|---|---|
| Qwen3-Embedding-0.6B | 896 | `.npy` |
| Qwen3-Embedding-4B | 2560 | `.npy` |
**Source Datasets:** germanCoal, FCPB, socialTwitter, hatespeech, mental_health
## How to Use
```python
import numpy as np
# Load pre-computed embeddings
embeddings = np.load("0.6B/zero_shot/germanCoal_zero_shot_embeddings.npy")
print(embeddings.shape) # (num_samples, 896)
```
Or download via `huggingface_hub`:
```python
from huggingface_hub import hf_hub_download
import numpy as np
path = hf_hub_download(
repo_id="CodeSoulco/THETA-embeddings",
filename="0.6B/supervised/socialTwitter_supervised_embeddings.npy",
repo_type="dataset"
)
embeddings = np.load(path)
```
## Related
- **Model (LoRA weights):** [CodeSoulco/THETA](https://huggingface.co/CodeSoulco/THETA)
## License
This dataset is released under the **MIT License**.
## Citation
```bibtex
@misc{theta2026,
title={THETA: Textual Hybrid Embedding--based Topic Analysis},
author={CodeSoul},
year={2026},
publisher={Hugging Face},
url={https://huggingface.co/CodeSoulco/THETA}
}
``` |