File size: 2,912 Bytes
bbc372c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
eb6c1dd
 
 
 
 
bbc372c
eb6c1dd
 
 
 
 
 
 
 
 
d937543
eb6c1dd
 
 
 
 
f0e4299
 
eb6c1dd
 
a07cf52
eb6c1dd
 
 
 
 
 
 
 
 
 
 
 
 
 
a07cf52
eb6c1dd
a07cf52
eb6c1dd
 
 
 
 
 
 
 
a07cf52
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
---
dataset_info:
  features:
  - name: doc_id
    dtype: string
  - name: doc_ps
    list:
    - name: entities
      list:
      - name: end_offset
        dtype: int64
      - name: entity_id
        dtype: string
      - name: label
        dtype: string
      - name: start_offset
        dtype: int64
      - name: text
        dtype: string
    - name: p_id
      dtype: int64
    - name: p_text
      dtype: string
  splits:
  - name: selective
    num_bytes: 850523
    num_examples: 129
  - name: total
    num_bytes: 936826
    num_examples: 129
  download_size: 918085
  dataset_size: 1787349
configs:
- config_name: default
  data_files:
  - split: selective
    path: data/selective-*
  - split: total
    path: data/total-*
license: mit
task_categories:
- token-classification
language:
- pt
---

# Segundo HAREM

According to [Linguateca](https://www.linguateca.pt/) - a repository for Natural Language resources in Portuguese - HAREM is a "joint evaluation in the area of Named Entity Recognition in Portuguese".

The sole goal of this dataset is to make [Segundo HAREM](https://www.linguateca.pt/HAREM/) generally available through the huggingface hub and the datasets library.

Intended use: this dataset may be used to train and/or evaluate Named Entity Recognition models in Portuguese.

The script used to convert the original xml file containing the annotations of Segundo HAREM to json, and subsequently to a dataset, was derived from [this fork](https://github.com/fabiocapsouza/harem_preprocessing).

If you'd like to reproduce the creation of the dataset, you can check [my own fork](https://github.com/marquesafonso/harem_preprocessing) from which this dataset was created.

You may find the authors and relevant work of this dataset in the citations section below. All credit is theirs.

Note: The first version of HAREM is already available in the huggingface hub [here](https://huggingface.co/datasets/Linguateca/harem)


## Citations
```
@inproceedings{freitas-etal-2010-second,
    title = "Second {HAREM}: Advancing the State of the Art of Named Entity Recognition in {P}ortuguese",
    author = "Freitas, Cl{\'a}udia  and
      Mota, Cristina  and
      Santos, Diana  and
      Oliveira, Hugo Gon{\c{c}}alo  and
      Carvalho, Paula",
    booktitle = "Proceedings of the Seventh International Conference on Language Resources and Evaluation ({LREC}'10)",
    month = may,
    year = "2010",
    address = "Valletta, Malta",
    publisher = "European Language Resources Association (ELRA)",
    url = "http://www.lrec-conf.org/proceedings/lrec2010/pdf/412_Paper.pdf"
}
```

```
@misc{fabiocapsouza2020,
  author = {fabiocapsouza},
  title = {harem_preprocessing},
  year = {2020},
  publisher = {GitHub},
  journal = {GitHub repository},
  howpublished = {\url{https://github.com/fabiocapsouza/harem_preprocessing}},
  commit = {30083a542390094d84ec45116085ab71ceb1442c}
}
```