Update README.md
Browse files
README.md
CHANGED
|
@@ -6,10 +6,8 @@ pipeline_tag: image-classification
|
|
| 6 |
|
| 7 |
### Model Description
|
| 8 |
|
| 9 |
-
|
| 10 |
-
|
| 11 |
-
The utility of this model is to be used in single-cell analysis of microscopic imaging.
|
| 12 |
-
|
| 13 |
|
| 14 |
- **Developed by:** Vidit Agrawal, John Peters, Juan Caicedo
|
| 15 |
- **Shared by:** [Caicedo Lab]
|
|
@@ -53,20 +51,30 @@ Use the code below to get started with the model.
|
|
| 53 |
|
| 54 |
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 55 |
|
| 56 |
-
|
|
|
|
| 57 |
|
| 58 |
### Training Procedure
|
| 59 |
|
| 60 |
-
We have utilized the DINO model.
|
| 61 |
|
| 62 |
#### Preprocessing
|
| 63 |
|
| 64 |
-
|
| 65 |
|
| 66 |
## Evaluation
|
| 67 |
|
| 68 |
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 69 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 70 |
|
| 71 |
#### Summary
|
| 72 |
|
|
@@ -74,20 +82,21 @@ specify all 6 benchmarks, and point to them. Point to the paper?
|
|
| 74 |
|
| 75 |
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
| 76 |
|
| 77 |
-
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
| 78 |
-
|
| 79 |
- **Hardware Type:** Nvidia RTX A6000
|
| 80 |
-
- **Hours used:**
|
| 81 |
- **Cloud Provider:** Private Infrastructure
|
| 82 |
- **Compute Region:** Private Infrastructure
|
| 83 |
-
- **Carbon Emitted:**
|
| 84 |
|
| 85 |
## Technical Specifications
|
| 86 |
|
| 87 |
|
| 88 |
The model is a ViT Small trained on 2500 Nvidia A6000 GPU hours. The model was trained on a multi-node system with 2 nodes, each containing 7 GPUs.
|
| 89 |
|
| 90 |
-
|
|
|
|
|
|
|
|
|
|
| 91 |
|
| 92 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 93 |
|
|
|
|
| 6 |
|
| 7 |
### Model Description
|
| 8 |
|
| 9 |
+
Morphem is a self supervised learning framework trained with the DINO Bag of Channels recipe on the entire CHAMMI-75 dataset.
|
| 10 |
+
It serves as a benchmark for performance for self-supervised models.
|
|
|
|
|
|
|
| 11 |
|
| 12 |
- **Developed by:** Vidit Agrawal, John Peters, Juan Caicedo
|
| 13 |
- **Shared by:** [Caicedo Lab]
|
|
|
|
| 51 |
|
| 52 |
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 53 |
|
| 54 |
+
MorphEm was pre-trained on the entire CHAMMI-75 pre-training data.
|
| 55 |
+
The CHAMMI-75 dataset consists of 75 heterogenous studies and 2.8 million multi-channel images.
|
| 56 |
|
| 57 |
### Training Procedure
|
| 58 |
|
| 59 |
+
We have utilized the self-supervised learning framework called DINO. We pre-trained a model which inputs a single channel one at a time. For evaluation, you would concatenate each channel specifically.
|
| 60 |
|
| 61 |
#### Preprocessing
|
| 62 |
|
| 63 |
+
We used three transforms mainly for preprocessing: SaturationNoiseInjector(), SelfImageNormalize(), Resize(224,224)
|
| 64 |
|
| 65 |
## Evaluation
|
| 66 |
|
| 67 |
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 68 |
+
We have evaluated this model on 6 different benchmarks. The model is highly competitive in most of them. The benchmarks are listed below:
|
| 69 |
+
|
| 70 |
+
1. CHAMMI
|
| 71 |
+
2. HPAv23
|
| 72 |
+
3. Jump-CP
|
| 73 |
+
4. IDR0017
|
| 74 |
+
5. CELLPHIE
|
| 75 |
+
6. RBC-MC
|
| 76 |
+
|
| 77 |
+
More details can be found in the paper:
|
| 78 |
|
| 79 |
#### Summary
|
| 80 |
|
|
|
|
| 82 |
|
| 83 |
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
| 84 |
|
|
|
|
|
|
|
| 85 |
- **Hardware Type:** Nvidia RTX A6000
|
| 86 |
+
- **Hours used:** 2352
|
| 87 |
- **Cloud Provider:** Private Infrastructure
|
| 88 |
- **Compute Region:** Private Infrastructure
|
| 89 |
+
- **Carbon Emitted:** 304 kg CO2
|
| 90 |
|
| 91 |
## Technical Specifications
|
| 92 |
|
| 93 |
|
| 94 |
The model is a ViT Small trained on 2500 Nvidia A6000 GPU hours. The model was trained on a multi-node system with 2 nodes, each containing 7 GPUs.
|
| 95 |
|
| 96 |
+
## Citation
|
| 97 |
+
|
| 98 |
+
Can be cited as the following:
|
| 99 |
+
|
| 100 |
|
| 101 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 102 |
|