Vidit2003 commited on
Commit
9649db9
·
verified ·
1 Parent(s): 1c4692d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -13
README.md CHANGED
@@ -6,10 +6,8 @@ pipeline_tag: image-classification
6
 
7
  ### Model Description
8
 
9
- This model is the best model released from the CHAMMI-75:pre-training multi-channel models with heterogeneous microscopy images paper currently under review at a conference.
10
- This model was pre-trained on the entire CHAMMI-75 dataset using the bag of channels method. It has acheived state of the art in 7/8 self-supervised learning benchmarks.
11
- The utility of this model is to be used in single-cell analysis of microscopic imaging.
12
-
13
 
14
  - **Developed by:** Vidit Agrawal, John Peters, Juan Caicedo
15
  - **Shared by:** [Caicedo Lab]
@@ -53,20 +51,30 @@ Use the code below to get started with the model.
53
 
54
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
55
 
56
- [More Information Needed]
 
57
 
58
  ### Training Procedure
59
 
60
- We have utilized the DINO model.
61
 
62
  #### Preprocessing
63
 
64
- [More Information Needed]
65
 
66
  ## Evaluation
67
 
68
  <!-- This section describes the evaluation protocols and provides the results. -->
69
- specify all 6 benchmarks, and point to them. Point to the paper?
 
 
 
 
 
 
 
 
 
70
 
71
  #### Summary
72
 
@@ -74,20 +82,21 @@ specify all 6 benchmarks, and point to them. Point to the paper?
74
 
75
  <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
76
 
77
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
78
-
79
  - **Hardware Type:** Nvidia RTX A6000
80
- - **Hours used:** 2500
81
  - **Cloud Provider:** Private Infrastructure
82
  - **Compute Region:** Private Infrastructure
83
- - **Carbon Emitted:** 324 kg CO2
84
 
85
  ## Technical Specifications
86
 
87
 
88
  The model is a ViT Small trained on 2500 Nvidia A6000 GPU hours. The model was trained on a multi-node system with 2 nodes, each containing 7 GPUs.
89
 
90
- <!-- ## Citation -->
 
 
 
91
 
92
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
93
 
 
6
 
7
  ### Model Description
8
 
9
+ Morphem is a self supervised learning framework trained with the DINO Bag of Channels recipe on the entire CHAMMI-75 dataset.
10
+ It serves as a benchmark for performance for self-supervised models.
 
 
11
 
12
  - **Developed by:** Vidit Agrawal, John Peters, Juan Caicedo
13
  - **Shared by:** [Caicedo Lab]
 
51
 
52
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
53
 
54
+ MorphEm was pre-trained on the entire CHAMMI-75 pre-training data.
55
+ The CHAMMI-75 dataset consists of 75 heterogenous studies and 2.8 million multi-channel images.
56
 
57
  ### Training Procedure
58
 
59
+ We have utilized the self-supervised learning framework called DINO. We pre-trained a model which inputs a single channel one at a time. For evaluation, you would concatenate each channel specifically.
60
 
61
  #### Preprocessing
62
 
63
+ We used three transforms mainly for preprocessing: SaturationNoiseInjector(), SelfImageNormalize(), Resize(224,224)
64
 
65
  ## Evaluation
66
 
67
  <!-- This section describes the evaluation protocols and provides the results. -->
68
+ We have evaluated this model on 6 different benchmarks. The model is highly competitive in most of them. The benchmarks are listed below:
69
+
70
+ 1. CHAMMI
71
+ 2. HPAv23
72
+ 3. Jump-CP
73
+ 4. IDR0017
74
+ 5. CELLPHIE
75
+ 6. RBC-MC
76
+
77
+ More details can be found in the paper:
78
 
79
  #### Summary
80
 
 
82
 
83
  <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
84
 
 
 
85
  - **Hardware Type:** Nvidia RTX A6000
86
+ - **Hours used:** 2352
87
  - **Cloud Provider:** Private Infrastructure
88
  - **Compute Region:** Private Infrastructure
89
+ - **Carbon Emitted:** 304 kg CO2
90
 
91
  ## Technical Specifications
92
 
93
 
94
  The model is a ViT Small trained on 2500 Nvidia A6000 GPU hours. The model was trained on a multi-node system with 2 nodes, each containing 7 GPUs.
95
 
96
+ ## Citation
97
+
98
+ Can be cited as the following:
99
+
100
 
101
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
102