hamedbabaeigiglou's picture
minor update to readme
b1e516f verified
metadata
license: mit
language:
  - en
tags:
  - OntoLearner
  - ontology-learning
  - ecology-and-environment
pretty_name: Ecology And Environment
OntoLearner

Ecology And Environment Domain Ontologies

Overview

The ecology and environment domain encompasses the structured representation of knowledge pertaining to ecological systems, diverse environmental contexts, biomes, and the principles of sustainability science. This domain is pivotal in facilitating the integration, sharing, and analysis of complex environmental data, thereby advancing our understanding of ecological interactions and promoting informed decision-making for sustainable development. By providing a formalized framework, it supports interdisciplinary research and policy-making aimed at addressing global environmental challenges.

Ontologies

Ontology ID Full Name Classes Properties Last Updated
ENVO Environment Ontology (ENVO) 9309 208 2024-07-01
OEO The Open Energy Ontology (OEO) 0 0 03/2025
SWEET Semantic Web for Earth and Environment Technology Ontology (SWEET) 10240 358 July 14, 2022

Dataset Files

Each ontology directory contains the following files:

  1. <ontology_id>.<format> - The original ontology file
  2. term_typings.json - A Dataset of term-to-type mappings
  3. taxonomies.json - Dataset of taxonomic relations
  4. non_taxonomic_relations.json - Dataset of non-taxonomic relations
  5. <ontology_id>.rst - Documentation describing the ontology

Usage

These datasets are intended for ontology learning research and applications. Here's how to use them with OntoLearner:

First of all, install the OntoLearner library via PiP:

pip install ontolearner

How to load an ontology or LLM4OL Paradigm tasks datasets?

from ontolearner import ENVO

ontology = ENVO()

# Load an ontology.
ontology.load()  

# Load (or extract) LLMs4OL Paradigm tasks datasets
data = ontology.extract()

How use the loaded dataset for LLM4OL Paradigm task settings?

# Import core modules from the OntoLearner library
from ontolearner import ENVO, LearnerPipeline, train_test_split

# Load the ENVO ontology, which contains concepts related to wines, their properties, and categories
ontology = ENVO()
ontology.load()  # Load entities, types, and structured term annotations from the ontology
data = ontology.extract()

# Split into train and test sets
train_data, test_data = train_test_split(data, test_size=0.2, random_state=42)

# Initialize a multi-component learning pipeline (retriever + LLM)
# This configuration enables a Retrieval-Augmented Generation (RAG) setup
pipeline = LearnerPipeline(
    retriever_id='sentence-transformers/all-MiniLM-L6-v2',      # Dense retriever model for nearest neighbor search
    llm_id='Qwen/Qwen2.5-0.5B-Instruct',                        # Lightweight instruction-tuned LLM for reasoning
    hf_token='...',                                             # Hugging Face token for accessing gated models
    batch_size=32,                                              # Batch size for training/prediction if supported
    top_k=5                                                     # Number of top retrievals to include in RAG prompting
)

# Run the pipeline: training, prediction, and evaluation in one call
outputs = pipeline(
    train_data=train_data,
    test_data=test_data,
    evaluate=True,              # Compute metrics like precision, recall, and F1
    task='term-typing'          # Specifies the task
                                # Other options: "taxonomy-discovery" or "non-taxonomy-discovery"
)

# Print final evaluation metrics
print("Metrics:", outputs['metrics'])

# Print the total time taken for the full pipeline execution
print("Elapsed time:", outputs['elapsed_time'])

# Print all outputs (including predictions)
print(outputs)

For more detailed documentation, see the Documentation

Citation

If you find our work helpful, feel free to give us a cite.

@inproceedings{babaei2023llms4ol,
  title={LLMs4OL: Large language models for ontology learning},
  author={Babaei Giglou, Hamed and D’Souza, Jennifer and Auer, S{\"o}ren},
  booktitle={International Semantic Web Conference},
  pages={408--427},
  year={2023},
  organization={Springer}
}