update doi and arXiv link to paper
Browse files
README.md
CHANGED
|
@@ -37,20 +37,6 @@ datasets:
|
|
| 37 |
- BIOSCAN-1M
|
| 38 |
- EOL
|
| 39 |
---
|
| 40 |
-
<!--
|
| 41 |
-
Image with caption (jpg or png):
|
| 42 |
-
||
|
| 43 |
-
|:--|
|
| 44 |
-
|**Figure #.** [Image of <>](https://huggingface.co/imageomics/<model-repo>/raw/main/<filepath>) <caption description>.|
|
| 45 |
-
-->
|
| 46 |
-
|
| 47 |
-
<!--
|
| 48 |
-
Notes on styling:
|
| 49 |
-
|
| 50 |
-
To render LaTex in your README, wrap the code in `\\(` and `\\)`. Example: \\(\frac{1}{2}\\)
|
| 51 |
-
|
| 52 |
-
Escape underscores ("_") with a "\". Example: image\_RGB
|
| 53 |
-
-->
|
| 54 |
|
| 55 |
# Model Card for BioCAP
|
| 56 |
|
|
@@ -76,7 +62,7 @@ Compared with [BioCLIP](https://imageomics.github.io/bioclip/), BioCAP improves
|
|
| 76 |
|
| 77 |
- **Homepage:** https://imageomics.github.io/biocap
|
| 78 |
- **Repository:** [BioCAP](https://github.com/Imageomics/biocap)
|
| 79 |
-
- **Paper:** [BioCAP: Exploiting synthetic captions beyond labels in biological foundation models]()
|
| 80 |
- **Demo:** [BioCAP]()
|
| 81 |
|
| 82 |
## Uses
|
|
@@ -155,11 +141,11 @@ For text-image retrieval tasks, we used:
|
|
| 155 |
* [Cornell Bird](https://www.birds.cornell.edu/home/): A paired image–text dataset we collected from the [Macaulay Library](https://www.macaulaylibrary.org). It contains naturalistic bird photographs paired with descriptive text.
|
| 156 |
* [PlantID](https://plantid.net/Home.aspx): A paired dataset we collected from [PlantID](https://plantid.net/Home.aspx). It provides plant photographs and associated textual descriptions for evaluating retrieval in botanical domains.
|
| 157 |
|
| 158 |
-
**Note:** More details regarding the evaluation implementation can be referred to in the [paper](). Dataset access code and the CSVs for the last two text-image retrieval tasks are provided in the [evaluation section of the BioCAP Pipeline](https://github.com/Imageomics/biocap/blob/main/BioCAP-pipeline.md#evaluation-data).
|
| 159 |
|
| 160 |
|
| 161 |
### Results
|
| 162 |
-
We show the zero-shot classification and text-image retrieval task results here. For more detailed results, please check the [paper]().
|
| 163 |
<table cellpadding="0" cellspacing="0">
|
| 164 |
<thead>
|
| 165 |
<tr>
|
|
@@ -241,7 +227,7 @@ We show the zero-shot classification and text-image retrieval task results here.
|
|
| 241 |
<tr>
|
| 242 |
<td>BioCLIP</td>
|
| 243 |
<td>58.8</td>
|
| 244 |
-
<td
|
| 245 |
<td>34.9</td>
|
| 246 |
<td>20.5</td>
|
| 247 |
<td>31.7</td>
|
|
@@ -389,15 +375,23 @@ It took 30hrs to complete the training of 50 epochs.
|
|
| 389 |
title = {{BioCAP}},
|
| 390 |
url = {https://huggingface.co/imageomics/biocap},
|
| 391 |
version = {1.0.0},
|
| 392 |
-
doi = {},
|
| 393 |
publisher = {Hugging Face},
|
| 394 |
year = {2025}
|
| 395 |
}
|
| 396 |
```
|
| 397 |
Please also cite our paper:
|
| 398 |
```
|
| 399 |
-
@article{
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 400 |
}
|
|
|
|
| 401 |
```
|
| 402 |
|
| 403 |
Also consider citing OpenCLIP and BioCLIP:
|
|
@@ -444,14 +438,6 @@ Our research is also supported by resources from the [Ohio Supercomputer Center]
|
|
| 444 |
|
| 445 |
Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.
|
| 446 |
|
| 447 |
-
<!-- ## Glossary -->
|
| 448 |
-
|
| 449 |
-
<!-- [optional] If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 450 |
-
|
| 451 |
-
<!-- ## More Information -->
|
| 452 |
-
|
| 453 |
-
<!-- [optional] Any other relevant information that doesn't fit elsewhere. -->
|
| 454 |
-
|
| 455 |
## Model Card Authors
|
| 456 |
|
| 457 |
Ziheng Zhang
|
|
|
|
| 37 |
- BIOSCAN-1M
|
| 38 |
- EOL
|
| 39 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
|
| 41 |
# Model Card for BioCAP
|
| 42 |
|
|
|
|
| 62 |
|
| 63 |
- **Homepage:** https://imageomics.github.io/biocap
|
| 64 |
- **Repository:** [BioCAP](https://github.com/Imageomics/biocap)
|
| 65 |
+
- **Paper:** [BioCAP: Exploiting synthetic captions beyond labels in biological foundation models](https://arxiv.org/abs/2510.20095)
|
| 66 |
- **Demo:** [BioCAP]()
|
| 67 |
|
| 68 |
## Uses
|
|
|
|
| 141 |
* [Cornell Bird](https://www.birds.cornell.edu/home/): A paired image–text dataset we collected from the [Macaulay Library](https://www.macaulaylibrary.org). It contains naturalistic bird photographs paired with descriptive text.
|
| 142 |
* [PlantID](https://plantid.net/Home.aspx): A paired dataset we collected from [PlantID](https://plantid.net/Home.aspx). It provides plant photographs and associated textual descriptions for evaluating retrieval in botanical domains.
|
| 143 |
|
| 144 |
+
**Note:** More details regarding the evaluation implementation can be referred to in the [paper](https://arxiv.org/abs/2510.20095). Dataset access code and the CSVs for the last two text-image retrieval tasks are provided in the [evaluation section of the BioCAP Pipeline](https://github.com/Imageomics/biocap/blob/main/BioCAP-pipeline.md#evaluation-data).
|
| 145 |
|
| 146 |
|
| 147 |
### Results
|
| 148 |
+
We show the zero-shot classification and text-image retrieval task results here. For more detailed results, please check the [paper](https://arxiv.org/abs/2510.20095).
|
| 149 |
<table cellpadding="0" cellspacing="0">
|
| 150 |
<thead>
|
| 151 |
<tr>
|
|
|
|
| 227 |
<tr>
|
| 228 |
<td>BioCLIP</td>
|
| 229 |
<td>58.8</td>
|
| 230 |
+
<td>6.1</td>
|
| 231 |
<td>34.9</td>
|
| 232 |
<td>20.5</td>
|
| 233 |
<td>31.7</td>
|
|
|
|
| 375 |
title = {{BioCAP}},
|
| 376 |
url = {https://huggingface.co/imageomics/biocap},
|
| 377 |
version = {1.0.0},
|
| 378 |
+
doi = {10.57967/hf/6794},
|
| 379 |
publisher = {Hugging Face},
|
| 380 |
year = {2025}
|
| 381 |
}
|
| 382 |
```
|
| 383 |
Please also cite our paper:
|
| 384 |
```
|
| 385 |
+
@article{zhang2025biocap,
|
| 386 |
+
title = {Bio{CAP}: Exploiting Synthetic Captions Beyond Labels in Biological Foundation Models},
|
| 387 |
+
author = {Ziheng Zhang and Xinyue Ma and Arpita Chowdhury and Elizabeth G Campolongo and Matthew J Thompson and Net Zhang and Samuel Stevens and Hilmar Lapp and Tanya Berger-Wolf and Yu Su and Wei-Lun Chao and Jianyang Gu},
|
| 388 |
+
year = {2025},
|
| 389 |
+
eprint = {2510.20095},
|
| 390 |
+
archivePrefix={arXiv},
|
| 391 |
+
primaryClass={cs.CV},
|
| 392 |
+
url={https://arxiv.org/abs/2510.20095}
|
| 393 |
}
|
| 394 |
+
|
| 395 |
```
|
| 396 |
|
| 397 |
Also consider citing OpenCLIP and BioCLIP:
|
|
|
|
| 438 |
|
| 439 |
Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.
|
| 440 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 441 |
## Model Card Authors
|
| 442 |
|
| 443 |
Ziheng Zhang
|