Datasets:
File size: 5,378 Bytes
a25f42f | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 | ---
license: cc-by-4.0
language:
- pt
pretty_name: PT-BR SciELO Articles (Brazilian Open-Access Research)
size_categories:
- 100K<n<1M
task_categories:
- text-generation
tags:
- pt-br
- brazilian-portuguese
- academic
- scientific
- scielo
- research
- pretraining
configs:
- config_name: default
data_files:
- split: train
path: data/*.parquet
---
# PT-BR SciELO Articles (Brazilian Open-Access Research)
Part of the **MagTina350m pretrain corpus release** by [Dataseek](https://dataseek.com.br)
under the Magestic.ai brand. This is one of nine silver-layer datasets that fed
[`dataseek/magtina350m-base`](https://huggingface.co/dataseek/magtina350m-base).
## Summary
154 K full-text Brazilian Portuguese research and review articles from SciELO Brazil (post-2010), spanning health sciences, social sciences, humanities and engineering. Avg ~34 KB per article — the largest academic-prose corpus in the MagTina350m mix.
## Source and collection method
Source: SciELO `articlemeta` API → fulltext HTML scrape → HTML→text → PT-only gate → year ≥ 2010 → doctype ∈ {research-article, review-article}.
**ETL script (in the MagTina1B repository):** [`scripts/etl/24_scielo_articles_v1.py`](https://huggingface.co/dataseek/magtina350m-base) *(public release of the ETL scripts is on the roadmap; until then the data card below documents the recipe in full).*
## Filters and deduplication
The following filters were applied before this dataset reached its silver
(release-ready) state:
- doctype ∈ {research-article, review-article}
- year ≥ 2010
- lang = pt
- len(text) ≥ 500 chars
Global URL-normalised deduplication was applied across all web-derived corpora
(`webpages`, `news`, `blogs`) so the same article does not appear twice across
those three datasets.
## Schema
| Column | Type | Description |
|---|---|---|
| `text` | `string` | Article fulltext. |
| `source` | `string` | Always 'scielo.br'. |
| `lang` | `string` | Language code (typically 'pt'). |
| `year` | `int32` | Publication year. |
| `doctype` | `string` | research-article | review-article. |
| `doc_id` | `string` | SciELO PID (links back to article). |
| `n_chars` | `int64` | Character count. |
Columns dropped at export (kept private as ETL internals): *none*
## Size statistics
| Metric | Value |
|---|---:|
| Rows | 154.2 K (154,218) |
| Characters | 5.29 B (5,291,892,295) |
| Estimated tokens (PT-BR, chars / 4.5) | 1.18 B |
| Compressed Parquet on disk | ~2.96 GB |
**Used in MagTina350m pretrain:** 1.176 B tokens
(6.8 % of MagTina350m's 17.39 B-token pretrain budget).
## How to load
```python
from datasets import load_dataset
ds = load_dataset("dataseek/ptbr-scielo", split="train", streaming=True)
for row in ds.take(5):
print(row["text"][:200])
```
Streaming is recommended for the larger configs. For the smaller datasets
(`ptbr-dou`, `ptbr-books-publicos`) eager loading is fine.
## Licensing
CC-BY 4.0 — the dominant license across SciELO Brazil (open access). Individual articles may carry CC-BY-NC variants; downstream users should honour the per-article licenses available via the SciELO API. Attribution by article DOI is required for redistribution.
**Upstream attribution:** SciELO Brazil — https://scielo.br/
## Citation
If you use this dataset, please cite both the upstream source and MagTina350m:
```bibtex
@misc{magtina350m_pretrain_2026,
title = {MagTina350m pretrain corpus — PT-BR SciELO Articles (Brazilian Open-Access Research)},
author = {Frasson, Ricardo and {Dataseek Team}},
year = 2026,
publisher = {Hugging Face},
url = {https://huggingface.co/datasets/dataseek/ptbr-scielo}
}
```
Please also honour the upstream license terms — for CC-BY-derived data,
attribution to the upstream creators is mandatory; for CC-BY-SA, downstream
derivatives must remain CC-BY-SA-compatible.
## Intended use
- Pre-training, continued pre-training, or domain-adapting of Brazilian Portuguese
language models.
- PT-BR NLP research where statistically representative public-web / academic /
legal / encyclopedic data is needed.
- Reproducing or improving on the MagTina350m result.
## Known limitations and PII statement
- **Text was NOT PII-scrubbed.** URLs, emails, phone numbers and personal names
that occurred in the source data may still be present. We strip zero-width
characters and normalise Unicode but we do not run an NER pass.
- **Crawled data carries upstream biases** of CommonCrawl, Wikipedia, news outlets
and academic institutions present in the source. We have not audited these.
- **No safety filtering** beyond langid and basic alpha-ratio gates. Hate-speech,
spam and adult content present in the source remain unless caught incidentally.
- **Provenance preserved at row level.** Every row has either a `url`, `source` or
`doc_id` column that points back to upstream — this is intentional, so consumers
can re-license, redact or filter.
## Related releases
- **Model:** [`dataseek/magtina350m-base`](https://huggingface.co/dataseek/magtina350m-base) (354.6 M params, pretrained on this corpus + 8 sibling datasets)
- **Instruct model:** [`dataseek/magtina350m-instruct`](https://huggingface.co/dataseek/magtina350m-instruct)
- **Sibling datasets:** see `dataseek/ptbr-*` for all nine corpora
## License
[cc-by-4.0](https://spdx.org/licenses/cc-by-4.0.html)
|