File size: 4,844 Bytes
792a480
 
 
 
 
 
 
 
 
8448a7b
792a480
0a602e9
 
792a480
 
 
0a602e9
792a480
 
8448a7b
 
792a480
8448a7b
792a480
 
 
 
 
 
 
 
 
 
 
0a602e9
792a480
 
 
 
0a602e9
792a480
8448a7b
 
0a602e9
8448a7b
0a602e9
 
 
 
 
 
 
 
 
8448a7b
0a602e9
 
 
 
 
 
 
d8c0857
0a602e9
 
d8c0857
0a602e9
 
d8c0857
0a602e9
d8c0857
8448a7b
 
 
d8c0857
8448a7b
d8c0857
 
8448a7b
d8c0857
 
 
 
 
8448a7b
 
d8c0857
 
8448a7b
 
d8c0857
 
 
8448a7b
d8c0857
 
 
 
 
 
 
 
 
8448a7b
 
0a602e9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124

---
license: mit
language:
- en
tags:
- OntoLearner
- ontology-learning
- events
pretty_name: Events
---

<div align="center">
  <img  src="https://raw.githubusercontent.com/sciknoworg/OntoLearner/main/images/logo.png"  alt="OntoLearner"
    style="display: block; margin: 0 auto; width: 500px; height: auto;">
  <h1 style="text-align: center; margin-top: 1em;">Events Domain Ontologies</h1>
  <a href="https://github.com/sciknoworg/OntoLearner"><img src="https://img.shields.io/badge/GitHub-OntoLearner-blue?logo=github" /></a>
</div>



## Overview
The events domain encompasses the structured representation and semantic modeling of occurrences in time, including their temporal, spatial, and contextual attributes. This domain is pivotal in knowledge representation as it facilitates the interoperability and integration of event-related data across diverse systems, enabling precise scheduling, planning, and historical analysis. By providing a framework for understanding and linking events, this domain supports advanced applications in areas such as artificial intelligence, information retrieval, and decision support systems.

## Ontologies
| Ontology ID | Full Name | Classes | Properties | Last Updated |
|-------------|-----------|---------|------------|--------------|
| Conference | Conference Ontology (Conference) | 42 | 52 | 2016/04/30|
| iCalendar | iCalendar Vocabulary (iCalendar) | 54 | 49 | 2004/04/07|
| LODE | Linking Open Descriptions of Events (LODE) | 1 | 7 | 2020-10-31|

## Dataset Files
Each ontology directory contains the following files:
1. `<ontology_id>.<format>` - The original ontology file
2. `term_typings.json` - A Dataset of term-to-type mappings
3. `taxonomies.json` - Dataset of taxonomic relations
4. `non_taxonomic_relations.json` - Dataset of non-taxonomic relations
5. `<ontology_id>.rst` - Documentation describing the ontology


## Usage
These datasets are intended for ontology learning research and applications. Here's how to use them with OntoLearner:

First of all, install the `OntoLearner` library via PiP:

```bash
pip install ontolearner
```

**How to load an ontology or LLM4OL Paradigm tasks datasets?**
``` python
from ontolearner import Conference

ontology = Conference()

# Load an ontology.
ontology.load()  

# Load (or extract) LLMs4OL Paradigm tasks datasets
data = ontology.extract()
```


**How use the loaded dataset for LLM4OL Paradigm task settings?**
``` python
# Import core modules from the OntoLearner library
from ontolearner import Conference, LearnerPipeline, train_test_split

# Load the Conference ontology, which contains concepts related to wines, their properties, and categories
ontology = Conference()
ontology.load()  # Load entities, types, and structured term annotations from the ontology
data = ontology.extract()

# Split into train and test sets
train_data, test_data = train_test_split(data, test_size=0.2, random_state=42)

# Initialize a multi-component learning pipeline (retriever + LLM)
# This configuration enables a Retrieval-Augmented Generation (RAG) setup
pipeline = LearnerPipeline(
    retriever_id='sentence-transformers/all-MiniLM-L6-v2',      # Dense retriever model for nearest neighbor search
    llm_id='Qwen/Qwen2.5-0.5B-Instruct',                        # Lightweight instruction-tuned LLM for reasoning
    hf_token='...',                                             # Hugging Face token for accessing gated models
    batch_size=32,                                              # Batch size for training/prediction if supported
    top_k=5                                                     # Number of top retrievals to include in RAG prompting
)

# Run the pipeline: training, prediction, and evaluation in one call
outputs = pipeline(
    train_data=train_data,
    test_data=test_data,
    evaluate=True,              # Compute metrics like precision, recall, and F1
    task='term-typing'          # Specifies the task
                                # Other options: "taxonomy-discovery" or "non-taxonomy-discovery"
)

# Print final evaluation metrics
print("Metrics:", outputs['metrics'])

# Print the total time taken for the full pipeline execution
print("Elapsed time:", outputs['elapsed_time'])

# Print all outputs (including predictions)
print(outputs)
```

For more detailed documentation, see the [![Documentation](https://img.shields.io/badge/Documentation-ontolearner.readthedocs.io-blue)](https://ontolearner.readthedocs.io)


## Citation

If you find our work helpful, feel free to give us a cite.


```bibtex
@inproceedings{babaei2023llms4ol,
  title={LLMs4OL: Large language models for ontology learning},
  author={Babaei Giglou, Hamed and D’Souza, Jennifer and Auer, S{\"o}ren},
  booktitle={International Semantic Web Conference},
  pages={408--427},
  year={2023},
  organization={Springer}
}
```