Upload folder using huggingface_hub
Browse files
README.md
CHANGED
|
@@ -41,7 +41,7 @@ base_model: Scentience-OVL-Classifiers-Base
|
|
| 41 |
|
| 42 |
This repository is a foundational series of multimodal joint classifier models trained on olfaction, vision, and language data.
|
| 43 |
It is meant as a quick start on loading the olfaction-vision-language models and getting the probability/logits of the presence of observed chemical compounds in a visual scene given a set of aroma descriptors.
|
| 44 |
-
For example, given an input image and a set of observed aromas (fruity, musky, etc),
|
| 45 |
|
| 46 |
Based on the original series of [embeddings models here](https://huggingface.co/kordelfrance/Olfaction-Vision-Language-Embeddings), these models are built specifically for prototyping and exploratory tasks within AR/VR, robotics, and embodied artificial intelligence.
|
| 47 |
Analogous to how CLIP and SigLIP embeddings give vision-language relationships, our embeddings models here give olfaction-vision-language (OVL) relationships.
|
|
@@ -50,7 +50,7 @@ Whether these models are used for better vision-scent navigation with drones, tr
|
|
| 50 |
We especially hope these models encourage the community to contribute to building standardized datasets and evaluation protocols for olfaction-vision-language learning.
|
| 51 |
|
| 52 |
## Models
|
| 53 |
-
We offer
|
| 54 |
- (1) `ovlc-gat`: The OVL base model built around a graph-attention network. This model is optimal for online tasks where accuracy is paramount and inference time is not as critical.
|
| 55 |
- (2) `ovlc-base`: The original OVL base model optimized for faster inference and edge-based robotics. This model is optimized for export to common frameworks that run on Android, iOS, Rust, and others.
|
| 56 |
|
|
|
|
| 41 |
|
| 42 |
This repository is a foundational series of multimodal joint classifier models trained on olfaction, vision, and language data.
|
| 43 |
It is meant as a quick start on loading the olfaction-vision-language models and getting the probability/logits of the presence of observed chemical compounds in a visual scene given a set of aroma descriptors.
|
| 44 |
+
For example, given an input image and a set of observed aromas (fruity, musky, etc), these models can give the probability that acetone is present.
|
| 45 |
|
| 46 |
Based on the original series of [embeddings models here](https://huggingface.co/kordelfrance/Olfaction-Vision-Language-Embeddings), these models are built specifically for prototyping and exploratory tasks within AR/VR, robotics, and embodied artificial intelligence.
|
| 47 |
Analogous to how CLIP and SigLIP embeddings give vision-language relationships, our embeddings models here give olfaction-vision-language (OVL) relationships.
|
|
|
|
| 50 |
We especially hope these models encourage the community to contribute to building standardized datasets and evaluation protocols for olfaction-vision-language learning.
|
| 51 |
|
| 52 |
## Models
|
| 53 |
+
We offer two olfaction-vision-language (OVL) classifier models with this repository:
|
| 54 |
- (1) `ovlc-gat`: The OVL base model built around a graph-attention network. This model is optimal for online tasks where accuracy is paramount and inference time is not as critical.
|
| 55 |
- (2) `ovlc-base`: The original OVL base model optimized for faster inference and edge-based robotics. This model is optimized for export to common frameworks that run on Android, iOS, Rust, and others.
|
| 56 |
|