Building a Trustworthy Deep Learning Model for Urban-Scene Image Segmentation: Robustness and Uncertainty Analysis

Abstract: Urban scene image segmentation typically identifies components such as pedestrians, sidewalks, streets, and bus stops within city environments. This technology is crucial for understanding and modeling complex urban systems, aiding city planners and policymakers in designing inclusive policies and infrastructures to support underrepresented groups, such as individuals with vision impairments. It also plays a critical role in autonomous vehicles, traffic management, environmental planning, and public safety. Urban scene image segmentation is a well-studied field, with many publicly available datasets for research. Deep learning techniques perform well on these databases. However, these segmentation models are vulnerable to adversarial perturbations—minor input modifications deliberately designed to deceive the model. Specifically, the ability of models trained on one dataset to generalize effectively to adversarial or noise-disturbed datasets is crucial but often insufficient. It is critical to develop robust models capable of withstanding such adversarial attacks. In this project, we aim to assess cutting-edge segmentation models’ robustness and generalization capabilities under adversarial attacks by evaluating their uncertainty and segmentation performance under adversarial perturbations. Furthermore, we will develop a novel contrastive learning diffusion model, ContraDiff, for trustworthy urban-scene image segmentation. The proposed method will utilize a contrastive learning mechanism with latent diffusion approaches to improve the models’ robustness and generalization ability under adversarial attacks.

Location: Kean University, University of California, Santa Cruz

Duration: 9/1/2024 to 8/31/2025

People

  • Principal Investigator: Dr. Meng Xu
  • Co-Principal Investigator: Dr. Yuyin Zhou (University of California, Santa Cruz)
  • Undergraduate Student Researchers:
    • Cesar Marte

Publications

  • To be updated …

Acknowledgements

This project is funded by the CAHSI-Google Institutional Research program.