taxa update to TOL-10M, retrained

#10
by egrace479 - opened
Files changed (1) hide show
  1. README.md +24 -13
README.md CHANGED
@@ -5,7 +5,7 @@ language:
5
  - en
6
  library_name: open_clip
7
  model_name: BioCLIP
8
- model_description: "Foundation model for the tree of life, built using CLIP architecture as a vision model for general organismal biology. It is trained on TreeOfLife-10M, our specially-created dataset covering over 450K taxa--the most biologically diverse ML-ready dataset available to date."
9
  tags:
10
  - zero-shot-image-classification
11
  - clip
@@ -13,6 +13,8 @@ tags:
13
  - CV
14
  - images
15
  - animals
 
 
16
  - species
17
  - taxonomy
18
  - rare species
@@ -30,11 +32,14 @@ datasets:
30
 
31
  # Model Card for BioCLIP
32
 
 
 
 
33
  <!--
34
  This modelcard has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1). And further altered to suit Imageomics Institute needs -->
35
 
36
  BioCLIP is a foundation model for the tree of life, built using CLIP architecture as a vision model for general organismal biology.
37
- It is trained on [TreeOfLife-10M](https://huggingface.co/datasets/imageomics/TreeOfLife-10M), our specially-created dataset covering over 450K taxa--the most biologically diverse ML-ready dataset available to date.
38
  Through rigorous benchmarking on a diverse set of fine-grained biological classification tasks, BioCLIP consistently outperformed existing baselines by 16% to 17% absolute.
39
  Through intrinsic evaluation, we found that BioCLIP learned a hierarchical representation aligned to the tree of life, which demonstrates its potential for robust generalizability.
40
 
@@ -47,7 +52,7 @@ Through intrinsic evaluation, we found that BioCLIP learned a hierarchical repre
47
  BioCLIP is based on OpenAI's [CLIP](https://openai.com/research/clip).
48
  We trained the model on [TreeOfLife-10M](https://huggingface.co/datasets/imageomics/TreeOfLife-10M) from OpenAI's ViT-B/16 checkpoint, using [OpenCLIP's](https://github.com/mlfoundations/open_clip) code.
49
  BioCLIP is trained with the standard CLIP objective to imbue the model with an understanding, not just of different species, but of the hierarchical structure that relates species across the tree of life.
50
- In this way, BioCLIP offers potential to aid biologists in discovery of new and related creatures, since it does not see the 454K different taxa as distinct classes, but as part of an interconnected hierarchy.
51
 
52
 
53
  - **Developed by:** Samuel Stevens, Jiaman Wu, Matthew J. Thompson, Elizabeth G. Campolongo, Chan Hee Song, David Edward Carlyn, Li Dong, Wasila M. Dahdul, Charles Stewart, Tanya Berger-Wolf, Wei-Lun Chao, and Yu Su
@@ -59,9 +64,11 @@ This model was developed for the benefit of the community as an open-source prod
59
 
60
  ### Model Sources
61
 
 
62
  - **Repository:** [BioCLIP](https://github.com/Imageomics/BioCLIP)
63
- - **Paper:** BioCLIP: A Vision Foundation Model for the Tree of Life ([arXiv](https://doi.org/10.48550/arXiv.2311.18803))
64
  - **Demo:** [BioCLIP Demo](https://huggingface.co/spaces/imageomics/bioclip-demo)
 
65
 
66
  ## Uses
67
 
@@ -72,7 +79,7 @@ The ViT-B/16 vision encoder is recommended as a base model for any computer visi
72
  ### Direct Use
73
 
74
  See the demo [here](https://huggingface.co/spaces/imageomics/bioclip-demo) for examples of zero-shot classification.
75
- It can also be used in a few-shot setting with a KNN; please see [our paper](https://doi.org/10.48550/arXiv.2311.18803) for details for both few-shot and zero-shot settings without fine-tuning.
76
 
77
 
78
  ## Bias, Risks, and Limitations
@@ -127,7 +134,7 @@ We tested BioCLIP on the following collection of 10 biologically-relevant tasks.
127
  - [Birds 525](https://www.kaggle.com/datasets/gpiosenka/100-bird-species): We evaluated on the 2,625 test images provided with the dataset.
128
  - [Rare Species](https://huggingface.co/datasets/imageomics/rare-species): A new dataset we curated for the purpose of testing this model and to contribute to the ML for Conservation community. It consists of 400 species labeled Near Threatened through Extinct in the Wild by the [IUCN Red List](https://www.iucnredlist.org/), with 30 images per species. For more information, see our dataset, [Rare Species](https://huggingface.co/datasets/imageomics/rare-species).
129
 
130
- For more information about the contents of these datasets, see Table 2 and associated sections of [our paper](https://doi.org/10.48550/arXiv.2311.18803).
131
 
132
  ### Metrics
133
 
@@ -137,7 +144,7 @@ We use top-1 and top-5 accuracy to evaluate models, and validation loss to choos
137
 
138
  We compare BioCLIP to OpenAI's CLIP and OpenCLIP's LAION-2B checkpoint.
139
  Here are the zero-shot classification results on our benchmark tasks.
140
- Please see [our paper](https://doi.org/10.48550/arXiv.2311.18803) for few-shot results.
141
 
142
  <table cellpadding="0" cellspacing="0">
143
  <thead>
@@ -227,7 +234,7 @@ BioCLIP outperforms general-domain baselines by 17% on average for zero-shot.
227
 
228
  ### Model Examination
229
 
230
- We encourage readers to see Section 4.6 of [our paper](https://doi.org/10.48550/arXiv.2311.18803).
231
  In short, BioCLIP forms representations that more closely align to the taxonomic hierarchy compared to general-domain baselines like CLIP or OpenCLIP.
232
 
233
 
@@ -238,14 +245,16 @@ In short, BioCLIP forms representations that more closely align to the taxonomic
238
  ```
239
  @software{bioclip2023,
240
  author = {Samuel Stevens and Jiaman Wu and Matthew J. Thompson and Elizabeth G. Campolongo and Chan Hee Song and David Edward Carlyn and Li Dong and Wasila M. Dahdul and Charles Stewart and Tanya Berger-Wolf and Wei-Lun Chao and Yu Su},
241
- doi = {10.57967/hf/1511},
242
- month = nov,
243
  title = {BioCLIP},
244
- version = {v0.1},
245
- year = {2023}
246
  }
247
  ```
248
 
 
 
 
249
  Please also cite our paper:
250
 
251
  ```
@@ -288,7 +297,9 @@ Please also consider citing OpenCLIP, iNat21 and BIOSCAN-1M:
288
 
289
  ## Acknowledgements
290
 
291
- The authors would like to thank Josef Uyeda, Jim Balhoff, Dan Rubenstein, Hank Bart, Hilmar Lapp, Sara Beery, and colleagues from the Imageomics Institute and the OSU NLP group for their valuable feedback. We also thank the BIOSCAN-1M team and the iNaturalist team for making their data available and easy to use, and Jennifer Hammack at EOL for her invaluable help in accessing EOL’s images.
 
 
292
 
293
  The [Imageomics Institute](https://imageomics.org) is funded by the US National Science Foundation's Harnessing the Data Revolution (HDR) program under [Award #2118240](https://www.nsf.gov/awardsearch/showAward?AWD_ID=2118240) (Imageomics: A New Frontier of Biological Information Powered by Knowledge-Guided Machine Learning). Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.
294
 
 
5
  - en
6
  library_name: open_clip
7
  model_name: BioCLIP
8
+ model_description: "Foundation model for the tree of life, built using CLIP architecture (ViT-B/16) as a vision model for general organismal biology. It is trained on TreeOfLife-10M, our specially-created dataset covering over 390K taxa--the most biologically diverse ML-ready dataset available at release."
9
  tags:
10
  - zero-shot-image-classification
11
  - clip
 
13
  - CV
14
  - images
15
  - animals
16
+ - plants
17
+ - fungi
18
  - species
19
  - taxonomy
20
  - rare species
 
32
 
33
  # Model Card for BioCLIP
34
 
35
+ If you are looking for the original BioCLIP model presented in the paper, please see [Revision 7b4abf1](https://huggingface.co/imageomics/bioclip/tree/7b4abf1f6ee747c15de00c7d28a5e62990b5dabc) (the most accurate documentation for the original version at BioCLIP [Revision ce901ab](https://huggingface.co/imageomics/bioclip/tree/ce901ab3c6a913f9e9ef94ce6d27761069f4f01c)), trained on TreeOfLife-10M [Revision ffa2a31](https://huggingface.co/datasets/imageomics/TreeOfLife-10M/tree/ffa2a318a1396f2f9e456ba171d3b5b5d8b4f051).
36
+
37
+
38
  <!--
39
  This modelcard has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1). And further altered to suit Imageomics Institute needs -->
40
 
41
  BioCLIP is a foundation model for the tree of life, built using CLIP architecture as a vision model for general organismal biology.
42
+ It is trained on [TreeOfLife-10M](https://huggingface.co/datasets/imageomics/TreeOfLife-10M), our specially-created dataset covering over 390K taxa--the most biologically diverse ML-ready dataset available at its release.
43
  Through rigorous benchmarking on a diverse set of fine-grained biological classification tasks, BioCLIP consistently outperformed existing baselines by 16% to 17% absolute.
44
  Through intrinsic evaluation, we found that BioCLIP learned a hierarchical representation aligned to the tree of life, which demonstrates its potential for robust generalizability.
45
 
 
52
  BioCLIP is based on OpenAI's [CLIP](https://openai.com/research/clip).
53
  We trained the model on [TreeOfLife-10M](https://huggingface.co/datasets/imageomics/TreeOfLife-10M) from OpenAI's ViT-B/16 checkpoint, using [OpenCLIP's](https://github.com/mlfoundations/open_clip) code.
54
  BioCLIP is trained with the standard CLIP objective to imbue the model with an understanding, not just of different species, but of the hierarchical structure that relates species across the tree of life.
55
+ In this way, BioCLIP offers potential to aid biologists in discovery of new and related creatures, since it does not see the 394K different taxa as distinct classes, but as part of an interconnected hierarchy.
56
 
57
 
58
  - **Developed by:** Samuel Stevens, Jiaman Wu, Matthew J. Thompson, Elizabeth G. Campolongo, Chan Hee Song, David Edward Carlyn, Li Dong, Wasila M. Dahdul, Charles Stewart, Tanya Berger-Wolf, Wei-Lun Chao, and Yu Su
 
64
 
65
  ### Model Sources
66
 
67
+ - **Homepage:** [BioCLIP Page](https://imageomics.github.io/bioclip/) and [BioCLIP Ecosystem Site](https://imageomics.github.io/bioclip-ecosystem/)
68
  - **Repository:** [BioCLIP](https://github.com/Imageomics/BioCLIP)
69
+ - **Paper:** [BioCLIP: A Vision Foundation Model for the Tree of Life](https://openaccess.thecvf.com/content/CVPR2024/papers/Stevens_BioCLIP_A_Vision_Foundation_Model_for_the_Tree_of_Life_CVPR_2024_paper.pdf)
70
  - **Demo:** [BioCLIP Demo](https://huggingface.co/spaces/imageomics/bioclip-demo)
71
+ <!-- [arXiv](https://doi.org/10.48550/arXiv.2311.18803) -->
72
 
73
  ## Uses
74
 
 
79
  ### Direct Use
80
 
81
  See the demo [here](https://huggingface.co/spaces/imageomics/bioclip-demo) for examples of zero-shot classification.
82
+ It can also be used in a few-shot setting with a KNN; please see [our paper](https://openaccess.thecvf.com/content/CVPR2024/papers/Stevens_BioCLIP_A_Vision_Foundation_Model_for_the_Tree_of_Life_CVPR_2024_paper.pdf) for details for both few-shot and zero-shot settings without fine-tuning.
83
 
84
 
85
  ## Bias, Risks, and Limitations
 
134
  - [Birds 525](https://www.kaggle.com/datasets/gpiosenka/100-bird-species): We evaluated on the 2,625 test images provided with the dataset.
135
  - [Rare Species](https://huggingface.co/datasets/imageomics/rare-species): A new dataset we curated for the purpose of testing this model and to contribute to the ML for Conservation community. It consists of 400 species labeled Near Threatened through Extinct in the Wild by the [IUCN Red List](https://www.iucnredlist.org/), with 30 images per species. For more information, see our dataset, [Rare Species](https://huggingface.co/datasets/imageomics/rare-species).
136
 
137
+ For more information about the contents of these datasets, see Table 2 and associated sections of [our paper](https://openaccess.thecvf.com/content/CVPR2024/papers/Stevens_BioCLIP_A_Vision_Foundation_Model_for_the_Tree_of_Life_CVPR_2024_paper.pdf).
138
 
139
  ### Metrics
140
 
 
144
 
145
  We compare BioCLIP to OpenAI's CLIP and OpenCLIP's LAION-2B checkpoint.
146
  Here are the zero-shot classification results on our benchmark tasks.
147
+ Please see [our paper](https://openaccess.thecvf.com/content/CVPR2024/papers/Stevens_BioCLIP_A_Vision_Foundation_Model_for_the_Tree_of_Life_CVPR_2024_paper.pdf) for few-shot results.
148
 
149
  <table cellpadding="0" cellspacing="0">
150
  <thead>
 
234
 
235
  ### Model Examination
236
 
237
+ We encourage readers to see Section 4.6 of [our paper](https://openaccess.thecvf.com/content/CVPR2024/papers/Stevens_BioCLIP_A_Vision_Foundation_Model_for_the_Tree_of_Life_CVPR_2024_paper.pdf).
238
  In short, BioCLIP forms representations that more closely align to the taxonomic hierarchy compared to general-domain baselines like CLIP or OpenCLIP.
239
 
240
 
 
245
  ```
246
  @software{bioclip2023,
247
  author = {Samuel Stevens and Jiaman Wu and Matthew J. Thompson and Elizabeth G. Campolongo and Chan Hee Song and David Edward Carlyn and Li Dong and Wasila M. Dahdul and Charles Stewart and Tanya Berger-Wolf and Wei-Lun Chao and Yu Su},
248
+ doi = {<update-on-generation>},
249
+ month = mar,
250
  title = {BioCLIP},
251
+ year = {2026}
 
252
  }
253
  ```
254
 
255
+ Note that this version is updated from the original BioCLIP model presented in the paper, please see [Revision 7b4abf1](https://huggingface.co/imageomics/bioclip/tree/7b4abf1f6ee747c15de00c7d28a5e62990b5dabc) (the most accurate documentation for the original version at BioCLIP [Revision ce901ab](https://huggingface.co/imageomics/bioclip/tree/ce901ab3c6a913f9e9ef94ce6d27761069f4f01c)), trained on TreeOfLife-10M [Revision ffa2a31](https://huggingface.co/datasets/imageomics/TreeOfLife-10M/tree/ffa2a318a1396f2f9e456ba171d3b5b5d8b4f051). This updated version of the model uses an updated TreeOfLife-10M which resolves taxonomic alignment issues discovered in the first version. The taxonomic resolution was completed using [TaxonoPy](https://github.com/Imageomics/TaxonoPy), which was developed for [TreeOfLife-200M](https://huggingface.co/datasets/imageomics/TreeOfLife-200M).
256
+
257
+
258
  Please also cite our paper:
259
 
260
  ```
 
297
 
298
  ## Acknowledgements
299
 
300
+ The authors would like to thank Josef Uyeda, Jim Balhoff, Dan Rubenstein, Hank Bart, Hilmar Lapp, Sara Beery, and colleagues from the Imageomics Institute and the OSU NLP group for their valuable feedback. We also thank the BIOSCAN-1M team and the iNaturalist team for making their data available and easy to use, and Jennifer Hammock at EOL for her invaluable help in accessing EOL’s images.
301
+
302
+ Additionally, we thank Ziheng Zhang and Jianyang Gu for running the training and evaluation of the revised model (with the taxonomic fixes) while they were training [BioCAP](https://huggingface.co/imageomics/biocap).
303
 
304
  The [Imageomics Institute](https://imageomics.org) is funded by the US National Science Foundation's Harnessing the Data Revolution (HDR) program under [Award #2118240](https://www.nsf.gov/awardsearch/showAward?AWD_ID=2118240) (Imageomics: A New Frontier of Biological Information Powered by Knowledge-Guided Machine Learning). Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.
305