mzkhawaja commited on
Commit
da7e600
·
verified ·
1 Parent(s): e5595f5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -1
README.md CHANGED
@@ -23,4 +23,45 @@ SceneSeg performs robustly across challenging weather and lighting conditions, i
23
  SceneSeg performs out of the box on roads across the world without any parameter tuning. SceneSeg provides self-driving cars with a core
24
  safety layer, helping to address 'long-tail' edge cases which plauge object-level detectors.
25
 
26
- <img src="https://github.com/autowarefoundation/autoware_vision_pilot/blob/main/Media/SceneSeg_GIF_Rain.gif">
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  SceneSeg performs out of the box on roads across the world without any parameter tuning. SceneSeg provides self-driving cars with a core
24
  safety layer, helping to address 'long-tail' edge cases which plauge object-level detectors.
25
 
26
+ During training, SceneSeg estimates three semantic classes
27
+
28
+ - `Foreground Objects`
29
+ - `Background Elements`
30
+ - `Drivable Road Surface`
31
+
32
+ However, during inference, we only use the outputs from the **`Foreground Objects`** class.
33
+
34
+ ## Watch the explainer video
35
+ Please click the video link to play - [***Video link***](https://drive.google.com/file/d/1riGlT3Ct-O1Y2C0DqxemwWS233dJrY7F/view?usp=sharing)
36
+
37
+ ## Performance Results
38
+ SceneSeg was trained on a diverse dataset comprised of multiple open-source datasets, including ACDC, MUSES, IDDAW, Mapillary Vistas and the Comma10K datset. These datasets provide challenging training data covering a wide range of countries, road types, lighting conditions and weather conditions. The BDD100K dataset was not used during training and served as a broad and diverse test set.
39
+
40
+ Mean Intersection Over Union (mIoU) scores are provided for both validation and test data. Validation results are provided for each of the datasets which comprise the complete validation set, alongside the results for the entire validation set, which are presented in the Cross Dataset column. Per-class mIoU scores are provided, alongside mIoU averaged across classes, as well as an Overall mIoU score which calculates the mIoU between the full multi-class prediction and multi-class ground truth.
41
+
42
+ ### Validation Set Performance - mIoU Scores
43
+ || Cross Dataset | Mapillary| MUSES | ACDC | IDDAW | Comma10K |
44
+ |--------|---------------|------------------|-------|------|-------|----------|
45
+ | Overall | **90.7** | 91.1 | 83.7 | 89.3 | 87.2 | **92.5** |
46
+ | Background Objects | **93.5** | 93.7 | 89.1 | 93.2 | 90.0 | **95.1** |
47
+ | Foreground Objects | **58.2** | **60.9** | 35.7 | 46.9 | 58.6 | 58.9 |
48
+ | Drivable Road Surface | **84.2** | 85.7 | 70.8 | 74.4 | 81.8 | **86.3** |
49
+ | Class Average | **78.6** | **80.1** | 65.2 | 71.5 | 76.8 | **80.1** |
50
+
51
+ ### Test Set Performance - mIoU Scores
52
+ || BDD100K |
53
+ |-|---------|
54
+ | Overall | **91.5** |
55
+ | Background Objects | **94.3** |
56
+ | Foreground Objects | **69.8** |
57
+ | Drivable Road Surface | **71.3** |
58
+ | Class Average | **78.5** |
59
+
60
+ ### Inference Speed
61
+ Inference speed tests were performed on a laptop equipped with an RTX3060 Mobile Gaming GPU, and an AMD Ryzen 7 5800H CPU. The SceneSeg network comprises a total of 223.43 Billion Floating Point Operations.
62
+
63
+ #### FP32 Precision
64
+ At FP32 precision, SceneSeg achieved 18.1 Frames Per Second inference speed
65
+
66
+ #### FP16 Precision
67
+ At FP16 precision, SceneSeg achieved 26.7 Frames Per Second inference speed