taxa update to TOL-10M, retrained
#10
by egrace479 - opened
README.md
CHANGED
|
@@ -5,7 +5,7 @@ language:
|
|
| 5 |
- en
|
| 6 |
library_name: open_clip
|
| 7 |
model_name: BioCLIP
|
| 8 |
-
model_description: "Foundation model for the tree of life, built using CLIP architecture as a vision model for general organismal biology. It is trained on TreeOfLife-10M, our specially-created dataset covering over
|
| 9 |
tags:
|
| 10 |
- zero-shot-image-classification
|
| 11 |
- clip
|
|
@@ -13,6 +13,8 @@ tags:
|
|
| 13 |
- CV
|
| 14 |
- images
|
| 15 |
- animals
|
|
|
|
|
|
|
| 16 |
- species
|
| 17 |
- taxonomy
|
| 18 |
- rare species
|
|
@@ -30,11 +32,14 @@ datasets:
|
|
| 30 |
|
| 31 |
# Model Card for BioCLIP
|
| 32 |
|
|
|
|
|
|
|
|
|
|
| 33 |
<!--
|
| 34 |
This modelcard has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1). And further altered to suit Imageomics Institute needs -->
|
| 35 |
|
| 36 |
BioCLIP is a foundation model for the tree of life, built using CLIP architecture as a vision model for general organismal biology.
|
| 37 |
-
It is trained on [TreeOfLife-10M](https://huggingface.co/datasets/imageomics/TreeOfLife-10M), our specially-created dataset covering over
|
| 38 |
Through rigorous benchmarking on a diverse set of fine-grained biological classification tasks, BioCLIP consistently outperformed existing baselines by 16% to 17% absolute.
|
| 39 |
Through intrinsic evaluation, we found that BioCLIP learned a hierarchical representation aligned to the tree of life, which demonstrates its potential for robust generalizability.
|
| 40 |
|
|
@@ -47,7 +52,7 @@ Through intrinsic evaluation, we found that BioCLIP learned a hierarchical repre
|
|
| 47 |
BioCLIP is based on OpenAI's [CLIP](https://openai.com/research/clip).
|
| 48 |
We trained the model on [TreeOfLife-10M](https://huggingface.co/datasets/imageomics/TreeOfLife-10M) from OpenAI's ViT-B/16 checkpoint, using [OpenCLIP's](https://github.com/mlfoundations/open_clip) code.
|
| 49 |
BioCLIP is trained with the standard CLIP objective to imbue the model with an understanding, not just of different species, but of the hierarchical structure that relates species across the tree of life.
|
| 50 |
-
In this way, BioCLIP offers potential to aid biologists in discovery of new and related creatures, since it does not see the
|
| 51 |
|
| 52 |
|
| 53 |
- **Developed by:** Samuel Stevens, Jiaman Wu, Matthew J. Thompson, Elizabeth G. Campolongo, Chan Hee Song, David Edward Carlyn, Li Dong, Wasila M. Dahdul, Charles Stewart, Tanya Berger-Wolf, Wei-Lun Chao, and Yu Su
|
|
@@ -59,9 +64,11 @@ This model was developed for the benefit of the community as an open-source prod
|
|
| 59 |
|
| 60 |
### Model Sources
|
| 61 |
|
|
|
|
| 62 |
- **Repository:** [BioCLIP](https://github.com/Imageomics/BioCLIP)
|
| 63 |
-
- **Paper:** BioCLIP: A Vision Foundation Model for the Tree of Life
|
| 64 |
- **Demo:** [BioCLIP Demo](https://huggingface.co/spaces/imageomics/bioclip-demo)
|
|
|
|
| 65 |
|
| 66 |
## Uses
|
| 67 |
|
|
@@ -72,7 +79,7 @@ The ViT-B/16 vision encoder is recommended as a base model for any computer visi
|
|
| 72 |
### Direct Use
|
| 73 |
|
| 74 |
See the demo [here](https://huggingface.co/spaces/imageomics/bioclip-demo) for examples of zero-shot classification.
|
| 75 |
-
It can also be used in a few-shot setting with a KNN; please see [our paper](https://
|
| 76 |
|
| 77 |
|
| 78 |
## Bias, Risks, and Limitations
|
|
@@ -127,7 +134,7 @@ We tested BioCLIP on the following collection of 10 biologically-relevant tasks.
|
|
| 127 |
- [Birds 525](https://www.kaggle.com/datasets/gpiosenka/100-bird-species): We evaluated on the 2,625 test images provided with the dataset.
|
| 128 |
- [Rare Species](https://huggingface.co/datasets/imageomics/rare-species): A new dataset we curated for the purpose of testing this model and to contribute to the ML for Conservation community. It consists of 400 species labeled Near Threatened through Extinct in the Wild by the [IUCN Red List](https://www.iucnredlist.org/), with 30 images per species. For more information, see our dataset, [Rare Species](https://huggingface.co/datasets/imageomics/rare-species).
|
| 129 |
|
| 130 |
-
For more information about the contents of these datasets, see Table 2 and associated sections of [our paper](https://
|
| 131 |
|
| 132 |
### Metrics
|
| 133 |
|
|
@@ -137,7 +144,7 @@ We use top-1 and top-5 accuracy to evaluate models, and validation loss to choos
|
|
| 137 |
|
| 138 |
We compare BioCLIP to OpenAI's CLIP and OpenCLIP's LAION-2B checkpoint.
|
| 139 |
Here are the zero-shot classification results on our benchmark tasks.
|
| 140 |
-
Please see [our paper](https://
|
| 141 |
|
| 142 |
<table cellpadding="0" cellspacing="0">
|
| 143 |
<thead>
|
|
@@ -227,7 +234,7 @@ BioCLIP outperforms general-domain baselines by 17% on average for zero-shot.
|
|
| 227 |
|
| 228 |
### Model Examination
|
| 229 |
|
| 230 |
-
We encourage readers to see Section 4.6 of [our paper](https://
|
| 231 |
In short, BioCLIP forms representations that more closely align to the taxonomic hierarchy compared to general-domain baselines like CLIP or OpenCLIP.
|
| 232 |
|
| 233 |
|
|
@@ -238,14 +245,16 @@ In short, BioCLIP forms representations that more closely align to the taxonomic
|
|
| 238 |
```
|
| 239 |
@software{bioclip2023,
|
| 240 |
author = {Samuel Stevens and Jiaman Wu and Matthew J. Thompson and Elizabeth G. Campolongo and Chan Hee Song and David Edward Carlyn and Li Dong and Wasila M. Dahdul and Charles Stewart and Tanya Berger-Wolf and Wei-Lun Chao and Yu Su},
|
| 241 |
-
doi = {
|
| 242 |
-
month =
|
| 243 |
title = {BioCLIP},
|
| 244 |
-
|
| 245 |
-
year = {2023}
|
| 246 |
}
|
| 247 |
```
|
| 248 |
|
|
|
|
|
|
|
|
|
|
| 249 |
Please also cite our paper:
|
| 250 |
|
| 251 |
```
|
|
@@ -288,7 +297,9 @@ Please also consider citing OpenCLIP, iNat21 and BIOSCAN-1M:
|
|
| 288 |
|
| 289 |
## Acknowledgements
|
| 290 |
|
| 291 |
-
The authors would like to thank Josef Uyeda, Jim Balhoff, Dan Rubenstein, Hank Bart, Hilmar Lapp, Sara Beery, and colleagues from the Imageomics Institute and the OSU NLP group for their valuable feedback. We also thank the BIOSCAN-1M team and the iNaturalist team for making their data available and easy to use, and Jennifer
|
|
|
|
|
|
|
| 292 |
|
| 293 |
The [Imageomics Institute](https://imageomics.org) is funded by the US National Science Foundation's Harnessing the Data Revolution (HDR) program under [Award #2118240](https://www.nsf.gov/awardsearch/showAward?AWD_ID=2118240) (Imageomics: A New Frontier of Biological Information Powered by Knowledge-Guided Machine Learning). Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.
|
| 294 |
|
|
|
|
| 5 |
- en
|
| 6 |
library_name: open_clip
|
| 7 |
model_name: BioCLIP
|
| 8 |
+
model_description: "Foundation model for the tree of life, built using CLIP architecture (ViT-B/16) as a vision model for general organismal biology. It is trained on TreeOfLife-10M, our specially-created dataset covering over 390K taxa--the most biologically diverse ML-ready dataset available at release."
|
| 9 |
tags:
|
| 10 |
- zero-shot-image-classification
|
| 11 |
- clip
|
|
|
|
| 13 |
- CV
|
| 14 |
- images
|
| 15 |
- animals
|
| 16 |
+
- plants
|
| 17 |
+
- fungi
|
| 18 |
- species
|
| 19 |
- taxonomy
|
| 20 |
- rare species
|
|
|
|
| 32 |
|
| 33 |
# Model Card for BioCLIP
|
| 34 |
|
| 35 |
+
If you are looking for the original BioCLIP model presented in the paper, please see [Revision 7b4abf1](https://huggingface.co/imageomics/bioclip/tree/7b4abf1f6ee747c15de00c7d28a5e62990b5dabc) (the most accurate documentation for the original version at BioCLIP [Revision ce901ab](https://huggingface.co/imageomics/bioclip/tree/ce901ab3c6a913f9e9ef94ce6d27761069f4f01c)), trained on TreeOfLife-10M [Revision ffa2a31](https://huggingface.co/datasets/imageomics/TreeOfLife-10M/tree/ffa2a318a1396f2f9e456ba171d3b5b5d8b4f051).
|
| 36 |
+
|
| 37 |
+
|
| 38 |
<!--
|
| 39 |
This modelcard has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1). And further altered to suit Imageomics Institute needs -->
|
| 40 |
|
| 41 |
BioCLIP is a foundation model for the tree of life, built using CLIP architecture as a vision model for general organismal biology.
|
| 42 |
+
It is trained on [TreeOfLife-10M](https://huggingface.co/datasets/imageomics/TreeOfLife-10M), our specially-created dataset covering over 390K taxa--the most biologically diverse ML-ready dataset available at its release.
|
| 43 |
Through rigorous benchmarking on a diverse set of fine-grained biological classification tasks, BioCLIP consistently outperformed existing baselines by 16% to 17% absolute.
|
| 44 |
Through intrinsic evaluation, we found that BioCLIP learned a hierarchical representation aligned to the tree of life, which demonstrates its potential for robust generalizability.
|
| 45 |
|
|
|
|
| 52 |
BioCLIP is based on OpenAI's [CLIP](https://openai.com/research/clip).
|
| 53 |
We trained the model on [TreeOfLife-10M](https://huggingface.co/datasets/imageomics/TreeOfLife-10M) from OpenAI's ViT-B/16 checkpoint, using [OpenCLIP's](https://github.com/mlfoundations/open_clip) code.
|
| 54 |
BioCLIP is trained with the standard CLIP objective to imbue the model with an understanding, not just of different species, but of the hierarchical structure that relates species across the tree of life.
|
| 55 |
+
In this way, BioCLIP offers potential to aid biologists in discovery of new and related creatures, since it does not see the 394K different taxa as distinct classes, but as part of an interconnected hierarchy.
|
| 56 |
|
| 57 |
|
| 58 |
- **Developed by:** Samuel Stevens, Jiaman Wu, Matthew J. Thompson, Elizabeth G. Campolongo, Chan Hee Song, David Edward Carlyn, Li Dong, Wasila M. Dahdul, Charles Stewart, Tanya Berger-Wolf, Wei-Lun Chao, and Yu Su
|
|
|
|
| 64 |
|
| 65 |
### Model Sources
|
| 66 |
|
| 67 |
+
- **Homepage:** [BioCLIP Page](https://imageomics.github.io/bioclip/) and [BioCLIP Ecosystem Site](https://imageomics.github.io/bioclip-ecosystem/)
|
| 68 |
- **Repository:** [BioCLIP](https://github.com/Imageomics/BioCLIP)
|
| 69 |
+
- **Paper:** [BioCLIP: A Vision Foundation Model for the Tree of Life](https://openaccess.thecvf.com/content/CVPR2024/papers/Stevens_BioCLIP_A_Vision_Foundation_Model_for_the_Tree_of_Life_CVPR_2024_paper.pdf)
|
| 70 |
- **Demo:** [BioCLIP Demo](https://huggingface.co/spaces/imageomics/bioclip-demo)
|
| 71 |
+
<!-- [arXiv](https://doi.org/10.48550/arXiv.2311.18803) -->
|
| 72 |
|
| 73 |
## Uses
|
| 74 |
|
|
|
|
| 79 |
### Direct Use
|
| 80 |
|
| 81 |
See the demo [here](https://huggingface.co/spaces/imageomics/bioclip-demo) for examples of zero-shot classification.
|
| 82 |
+
It can also be used in a few-shot setting with a KNN; please see [our paper](https://openaccess.thecvf.com/content/CVPR2024/papers/Stevens_BioCLIP_A_Vision_Foundation_Model_for_the_Tree_of_Life_CVPR_2024_paper.pdf) for details for both few-shot and zero-shot settings without fine-tuning.
|
| 83 |
|
| 84 |
|
| 85 |
## Bias, Risks, and Limitations
|
|
|
|
| 134 |
- [Birds 525](https://www.kaggle.com/datasets/gpiosenka/100-bird-species): We evaluated on the 2,625 test images provided with the dataset.
|
| 135 |
- [Rare Species](https://huggingface.co/datasets/imageomics/rare-species): A new dataset we curated for the purpose of testing this model and to contribute to the ML for Conservation community. It consists of 400 species labeled Near Threatened through Extinct in the Wild by the [IUCN Red List](https://www.iucnredlist.org/), with 30 images per species. For more information, see our dataset, [Rare Species](https://huggingface.co/datasets/imageomics/rare-species).
|
| 136 |
|
| 137 |
+
For more information about the contents of these datasets, see Table 2 and associated sections of [our paper](https://openaccess.thecvf.com/content/CVPR2024/papers/Stevens_BioCLIP_A_Vision_Foundation_Model_for_the_Tree_of_Life_CVPR_2024_paper.pdf).
|
| 138 |
|
| 139 |
### Metrics
|
| 140 |
|
|
|
|
| 144 |
|
| 145 |
We compare BioCLIP to OpenAI's CLIP and OpenCLIP's LAION-2B checkpoint.
|
| 146 |
Here are the zero-shot classification results on our benchmark tasks.
|
| 147 |
+
Please see [our paper](https://openaccess.thecvf.com/content/CVPR2024/papers/Stevens_BioCLIP_A_Vision_Foundation_Model_for_the_Tree_of_Life_CVPR_2024_paper.pdf) for few-shot results.
|
| 148 |
|
| 149 |
<table cellpadding="0" cellspacing="0">
|
| 150 |
<thead>
|
|
|
|
| 234 |
|
| 235 |
### Model Examination
|
| 236 |
|
| 237 |
+
We encourage readers to see Section 4.6 of [our paper](https://openaccess.thecvf.com/content/CVPR2024/papers/Stevens_BioCLIP_A_Vision_Foundation_Model_for_the_Tree_of_Life_CVPR_2024_paper.pdf).
|
| 238 |
In short, BioCLIP forms representations that more closely align to the taxonomic hierarchy compared to general-domain baselines like CLIP or OpenCLIP.
|
| 239 |
|
| 240 |
|
|
|
|
| 245 |
```
|
| 246 |
@software{bioclip2023,
|
| 247 |
author = {Samuel Stevens and Jiaman Wu and Matthew J. Thompson and Elizabeth G. Campolongo and Chan Hee Song and David Edward Carlyn and Li Dong and Wasila M. Dahdul and Charles Stewart and Tanya Berger-Wolf and Wei-Lun Chao and Yu Su},
|
| 248 |
+
doi = {<update-on-generation>},
|
| 249 |
+
month = mar,
|
| 250 |
title = {BioCLIP},
|
| 251 |
+
year = {2026}
|
|
|
|
| 252 |
}
|
| 253 |
```
|
| 254 |
|
| 255 |
+
Note that this version is updated from the original BioCLIP model presented in the paper, please see [Revision 7b4abf1](https://huggingface.co/imageomics/bioclip/tree/7b4abf1f6ee747c15de00c7d28a5e62990b5dabc) (the most accurate documentation for the original version at BioCLIP [Revision ce901ab](https://huggingface.co/imageomics/bioclip/tree/ce901ab3c6a913f9e9ef94ce6d27761069f4f01c)), trained on TreeOfLife-10M [Revision ffa2a31](https://huggingface.co/datasets/imageomics/TreeOfLife-10M/tree/ffa2a318a1396f2f9e456ba171d3b5b5d8b4f051). This updated version of the model uses an updated TreeOfLife-10M which resolves taxonomic alignment issues discovered in the first version. The taxonomic resolution was completed using [TaxonoPy](https://github.com/Imageomics/TaxonoPy), which was developed for [TreeOfLife-200M](https://huggingface.co/datasets/imageomics/TreeOfLife-200M).
|
| 256 |
+
|
| 257 |
+
|
| 258 |
Please also cite our paper:
|
| 259 |
|
| 260 |
```
|
|
|
|
| 297 |
|
| 298 |
## Acknowledgements
|
| 299 |
|
| 300 |
+
The authors would like to thank Josef Uyeda, Jim Balhoff, Dan Rubenstein, Hank Bart, Hilmar Lapp, Sara Beery, and colleagues from the Imageomics Institute and the OSU NLP group for their valuable feedback. We also thank the BIOSCAN-1M team and the iNaturalist team for making their data available and easy to use, and Jennifer Hammock at EOL for her invaluable help in accessing EOL’s images.
|
| 301 |
+
|
| 302 |
+
Additionally, we thank Ziheng Zhang and Jianyang Gu for running the training and evaluation of the revised model (with the taxonomic fixes) while they were training [BioCAP](https://huggingface.co/imageomics/biocap).
|
| 303 |
|
| 304 |
The [Imageomics Institute](https://imageomics.org) is funded by the US National Science Foundation's Harnessing the Data Revolution (HDR) program under [Award #2118240](https://www.nsf.gov/awardsearch/showAward?AWD_ID=2118240) (Imageomics: A New Frontier of Biological Information Powered by Knowledge-Guided Machine Learning). Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.
|
| 305 |
|