ZihengZ commited on
Commit
af8db7a
·
1 Parent(s): ddadf05

update doi and arXiv link to paper

Browse files
Files changed (1) hide show
  1. README.md +14 -28
README.md CHANGED
@@ -37,20 +37,6 @@ datasets:
37
  - BIOSCAN-1M
38
  - EOL
39
  ---
40
- <!--
41
- Image with caption (jpg or png):
42
- |![Figure #](https://huggingface.co/imageomics/<model-repo>/resolve/main/<filepath>)|
43
- |:--|
44
- |**Figure #.** [Image of <>](https://huggingface.co/imageomics/<model-repo>/raw/main/<filepath>) <caption description>.|
45
- -->
46
-
47
- <!--
48
- Notes on styling:
49
-
50
- To render LaTex in your README, wrap the code in `\\(` and `\\)`. Example: \\(\frac{1}{2}\\)
51
-
52
- Escape underscores ("_") with a "\". Example: image\_RGB
53
- -->
54
 
55
  # Model Card for BioCAP
56
 
@@ -76,7 +62,7 @@ Compared with [BioCLIP](https://imageomics.github.io/bioclip/), BioCAP improves
76
 
77
  - **Homepage:** https://imageomics.github.io/biocap
78
  - **Repository:** [BioCAP](https://github.com/Imageomics/biocap)
79
- - **Paper:** [BioCAP: Exploiting synthetic captions beyond labels in biological foundation models]()
80
  - **Demo:** [BioCAP]()
81
 
82
  ## Uses
@@ -155,11 +141,11 @@ For text-image retrieval tasks, we used:
155
  * [Cornell Bird](https://www.birds.cornell.edu/home/): A paired image–text dataset we collected from the [Macaulay Library](https://www.macaulaylibrary.org). It contains naturalistic bird photographs paired with descriptive text.
156
  * [PlantID](https://plantid.net/Home.aspx): A paired dataset we collected from [PlantID](https://plantid.net/Home.aspx). It provides plant photographs and associated textual descriptions for evaluating retrieval in botanical domains.
157
 
158
- **Note:** More details regarding the evaluation implementation can be referred to in the [paper](). Dataset access code and the CSVs for the last two text-image retrieval tasks are provided in the [evaluation section of the BioCAP Pipeline](https://github.com/Imageomics/biocap/blob/main/BioCAP-pipeline.md#evaluation-data).
159
 
160
 
161
  ### Results
162
- We show the zero-shot classification and text-image retrieval task results here. For more detailed results, please check the [paper]().
163
  <table cellpadding="0" cellspacing="0">
164
  <thead>
165
  <tr>
@@ -241,7 +227,7 @@ We show the zero-shot classification and text-image retrieval task results here.
241
  <tr>
242
  <td>BioCLIP</td>
243
  <td>58.8</td>
244
- <td><b>6.1</b></td>
245
  <td>34.9</td>
246
  <td>20.5</td>
247
  <td>31.7</td>
@@ -389,15 +375,23 @@ It took 30hrs to complete the training of 50 epochs.
389
  title = {{BioCAP}},
390
  url = {https://huggingface.co/imageomics/biocap},
391
  version = {1.0.0},
392
- doi = {},
393
  publisher = {Hugging Face},
394
  year = {2025}
395
  }
396
  ```
397
  Please also cite our paper:
398
  ```
399
- @article{
 
 
 
 
 
 
 
400
  }
 
401
  ```
402
 
403
  Also consider citing OpenCLIP and BioCLIP:
@@ -444,14 +438,6 @@ Our research is also supported by resources from the [Ohio Supercomputer Center]
444
 
445
  Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.
446
 
447
- <!-- ## Glossary -->
448
-
449
- <!-- [optional] If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
450
-
451
- <!-- ## More Information -->
452
-
453
- <!-- [optional] Any other relevant information that doesn't fit elsewhere. -->
454
-
455
  ## Model Card Authors
456
 
457
  Ziheng Zhang
 
37
  - BIOSCAN-1M
38
  - EOL
39
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
 
41
  # Model Card for BioCAP
42
 
 
62
 
63
  - **Homepage:** https://imageomics.github.io/biocap
64
  - **Repository:** [BioCAP](https://github.com/Imageomics/biocap)
65
+ - **Paper:** [BioCAP: Exploiting synthetic captions beyond labels in biological foundation models](https://arxiv.org/abs/2510.20095)
66
  - **Demo:** [BioCAP]()
67
 
68
  ## Uses
 
141
  * [Cornell Bird](https://www.birds.cornell.edu/home/): A paired image–text dataset we collected from the [Macaulay Library](https://www.macaulaylibrary.org). It contains naturalistic bird photographs paired with descriptive text.
142
  * [PlantID](https://plantid.net/Home.aspx): A paired dataset we collected from [PlantID](https://plantid.net/Home.aspx). It provides plant photographs and associated textual descriptions for evaluating retrieval in botanical domains.
143
 
144
+ **Note:** More details regarding the evaluation implementation can be referred to in the [paper](https://arxiv.org/abs/2510.20095). Dataset access code and the CSVs for the last two text-image retrieval tasks are provided in the [evaluation section of the BioCAP Pipeline](https://github.com/Imageomics/biocap/blob/main/BioCAP-pipeline.md#evaluation-data).
145
 
146
 
147
  ### Results
148
+ We show the zero-shot classification and text-image retrieval task results here. For more detailed results, please check the [paper](https://arxiv.org/abs/2510.20095).
149
  <table cellpadding="0" cellspacing="0">
150
  <thead>
151
  <tr>
 
227
  <tr>
228
  <td>BioCLIP</td>
229
  <td>58.8</td>
230
+ <td>6.1</td>
231
  <td>34.9</td>
232
  <td>20.5</td>
233
  <td>31.7</td>
 
375
  title = {{BioCAP}},
376
  url = {https://huggingface.co/imageomics/biocap},
377
  version = {1.0.0},
378
+ doi = {10.57967/hf/6794},
379
  publisher = {Hugging Face},
380
  year = {2025}
381
  }
382
  ```
383
  Please also cite our paper:
384
  ```
385
+ @article{zhang2025biocap,
386
+ title = {Bio{CAP}: Exploiting Synthetic Captions Beyond Labels in Biological Foundation Models},
387
+ author = {Ziheng Zhang and Xinyue Ma and Arpita Chowdhury and Elizabeth G Campolongo and Matthew J Thompson and Net Zhang and Samuel Stevens and Hilmar Lapp and Tanya Berger-Wolf and Yu Su and Wei-Lun Chao and Jianyang Gu},
388
+ year = {2025},
389
+ eprint = {2510.20095},
390
+ archivePrefix={arXiv},
391
+ primaryClass={cs.CV},
392
+ url={https://arxiv.org/abs/2510.20095}
393
  }
394
+
395
  ```
396
 
397
  Also consider citing OpenCLIP and BioCLIP:
 
438
 
439
  Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.
440
 
 
 
 
 
 
 
 
 
441
  ## Model Card Authors
442
 
443
  Ziheng Zhang