Prompt Generation for Enhanced Camouflaged Object Detection in Low-Altitude Economy

Research output: Chapter in Book or Report/Conference proceedingConference Proceedingpeer-review

Abstract

To ensure the safety of both aircraft and operators in low-altitude economy (LAE) activities, precise environmental perception capabilities are essential for effective collision prevention. However, achieving accurate perception remains challenging, particularly when obstacles are camouflaged and visually blended into their surroundings, making detection difficult, even with the robust foundation model, the Segment Anything Model (SAM). Although SAM's prompt-based strategy improves its performance in the Camouflaged Object Detection (COD) task, its reliance on limited prompts introduces new challenges. Instead of manually annotating prompts, our work introduces a multimodal learning approach that utilizes a Vision-Language Model (VLM) to automatically generate mask prompts. By integrating visual and textual information, this work generates high-quality prompts that significantly enhance the performance of SAM in identifying camouflaged objects. Experimental results demonstrate that the proposed method achieves an average improvement of 13% over the baseline SAM across three COD benchmark datasets.

Original languageEnglish
Title of host publication2025 IEEE 101st Vehicular Technology Conference, VTC 2025-Spring 2025 - Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798331531478
DOIs
Publication statusPublished - 2025
Event101st IEEE Vehicular Technology Conference, VTC 2025-Spring 2025 - Oslo, Norway
Duration: 17 Jun 202520 Jun 2025

Publication series

NameIEEE Vehicular Technology Conference
ISSN (Print)1550-2252

Conference

Conference101st IEEE Vehicular Technology Conference, VTC 2025-Spring 2025
Country/TerritoryNorway
CityOslo
Period17/06/2520/06/25

Keywords

  • Camouflaged Object Detection
  • Low-altitude Economy
  • Multimodal Learning
  • Prompt Generation
  • Segment Anything Model

Cite this