Abstract

This work proposes a semantic segmentation network that produces high-quality uncertainty estimates in a single forward pass. We exploit general representations from foundation models and unlabelled datasets through a Masked Image Modeling (MIM) approach, which is robust to augmentation hyper-parameters and simpler than previous techniques. For neural networks used in safety-critical applications, bias in the training data can lead to errors; therefore it is crucial to understand a network’s limitations at run time and act accordingly. To this end, we test our proposed method on a number of test domains including the SAX Segmentation benchmark, which includes labelled test data from dense urban, rural and off-road driving domains. The proposed method consistently outperforms uncertainty estimation and Out-of-Distribution (OoD) techniques on this difficult benchmark

Paper: https://ieeexplore.ieee.org/document/10610398

Bibtex

@INPROCEEDINGS{williams2024,
  author={Williams, David S. W. and Gadd, Matthew and Newman, Paul and De Martini, Daniele},
  booktitle={2024 IEEE International Conference on Robotics and Automation (ICRA)}, 
  title={Masked γ-SSL: Learning Uncertainty Estimation via Masked Image Modeling}, 
  year={2024},
  volume={},
  number={},
  pages={16192-16198},
  keywords={Training;Uncertainty;Semantic segmentation;Neural networks;Estimation;Training data;Benchmark testing;Segmentation;Scene Understanding;Introspection;Performance Assessment;Deep Learning;Autonomous Vehicles},
  doi={10.1109/ICRA57147.2024.10610398}}