File size: 2,959 Bytes
5669f76
 
 
 
 
 
 
772b482
 
 
84af9b7
d656729
 
 
 
84af9b7
 
6bff041
84af9b7
 
 
168ec29
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5669f76
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
---
license: mit
language:
- en
base_model:
- google/vit-base-patch16-224
---

# Research paper: Segmentation of Wheat Rust Disease Using Co-Salient Feature Extraction
can be accessed via https://doi.org/10.3390/agriengineering7020023 

RustCoSeg is a is a novel two-stage pipeline that first classifies wheat leaf images using a Vision Transformer, followed by 
segmentation using a Co-Salient Object Detection-inspired architecture, effectively isolating rust-infected regions across related samples.

Base models include:
--ViT based classification and DCFM based co-salient feature extraction for segmenting out wheat rust disease.

# ViT
Vision Transformer. 


# DCFM
The official repo of the paper `Democracy Does Matter: Comprehensive Feature Mining for Co-Salient Object Detection`.

## Environment Requirement
create enviroment and intall as following:
`pip install -r requirements.txt`

## Data Format
  trainset: CoCo-SEG
  
  testset: CoCA, CoSOD3k, Cosal2015
  
  Put the [CoCo-SEG](https://drive.google.com/file/d/1GbA_WKvJm04Z1tR8pTSzBdYVQ75avg4f/view), [CoCA](http://zhaozhang.net/coca.html), [CoSOD3k](http://dpfan.net/CoSOD3K/) and [Cosal2015](https://drive.google.com/u/0/uc?id=1mmYpGx17t8WocdPcw2WKeuFpz6VHoZ6K&export=download) datasets to `DCFM/data` as the following structure:
  ```
  DCFM
     β”œβ”€β”€ other codes
     β”œβ”€β”€ ...
     β”‚ 
     └── data
           
           β”œβ”€β”€ CoCo-SEG (CoCo-SEG's image files)
           β”œβ”€β”€ CoCA (CoCA's image files)
           β”œβ”€β”€ CoSOD3k (CoSOD3k's image files)
           └── Cosal2015 (Cosal2015's image files)
  ```
  
## Trained model

trained model can be downloaded from [papermodel](https://drive.google.com/file/d/1cfuq4eJoCwvFR9W1XOJX7Y0ttd8TGjlp/view?usp=sharing).

Run `test.py` for inference.

The evaluation tool please follow: https://github.com/zzhanghub/eval-co-sod


<!-- USAGE EXAMPLES -->
## Usage
Download pretrainde backbone model [VGG](https://drive.google.com/file/d/1Z1aAYXMyJ6txQ1Z9N7gtxLOIai4dxrXd/view?usp=sharing).

Run `train.py` for training.

## Prediction results
The co-saliency maps of DCFM can be found at [preds](https://drive.google.com/file/d/1wGeNHXFWVSyqvmL4NIUmEFdlHDovEtQR/view?usp=sharing).

## Reproduction
reproductions by myself on 2080Ti can be found at [reproduction1](https://drive.google.com/file/d/1vovii0RtYR_EC0Y2zxjY_cTWKWM3WaxP/view?usp=sharing) and [reproduction2](https://drive.google.com/file/d/1YPOKZ5kBtmZrCDhHpP3-w1GMVR5BfDoU/view?usp=sharing).

reprodution by myself on TITAN X can be found at [reproduction3](https://drive.google.com/file/d/1bnGFtRTYkVXqI2dcjeWFRDXnqqbUUBJr/view?usp=sharing).

## Others
The code is based on [GCoNet](https://github.com/fanq15/GCoNet).
I've added a validation part to help select the model for closer results. This validation part is based on [GCoNet_plus](https://github.com/ZhengPeng7/GCoNet_plus). You can try different evaluation metrics to select the model.