Uni-EPM: A Unified Extensible Perception Model Without Labeling Everything

Yilin Gao, Shiyi Mu, Shugong Xu*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Multi-task perception system to simultaneously perceive various kinds of objects is essential for autonomous driving. Existing perception frameworks always rely on multi-labeled datasets, which encompass labels for all pertinent objects, thereby constraining their adaptability to leverage specialized, task-oriented datasets. This approach hinders the efficient utilization of abundant but focused data. Furthermore, stacking multiple expert networks to address these perception objectives inevitably introduces additional computational overhead. To address this limitation, we propose Uni-EPM (Unified Extensible Perception Model), with a novel training framework for multi-task perception using task prompt selection to decouple tasks, which enables perceiving traffic signs and traffic lights in addition to lane lines and traffic elements from existing task-specific datasets without re-labeling. To the best of our knowledge, Uni-EPM is the first model can do this in the field of autonomous driving. By introducing the parameter-sharing decoder among tasks, we alleviate the problems of stacking task heads, including significant parameter increase, etc. Uni-EPM achieves state-of-the-art results in multi-task algorithms without substantial increase in parameters, which also demonstrates comparable performance to existing standalone models. The efficiency of the design is validated through comprehensive ablation experiments and results.

Original languageEnglish
JournalIEEE Transactions on Intelligent Transportation Systems
DOIs
Publication statusAccepted/In press - 2024
Externally publishedYes

Keywords

  • Multi-task
  • panoptic driving perception
  • unified framework

Fingerprint

Dive into the research topics of 'Uni-EPM: A Unified Extensible Perception Model Without Labeling Everything'. Together they form a unique fingerprint.

Cite this