Conditional Polarization Guidance for Camouflaged Object Detection

2026-03-31Computer Vision and Pattern Recognition

Computer Vision and Pattern Recognition
AI summary

The authors present CPGNet, a new system to detect objects that blend very well with their backgrounds, called camouflaged objects. They use information from polarized light to guide the learning of regular color images in a smarter way, making the detection easier and more accurate without making the model too complex. Their method includes a special way to focus on edges and refine details, helping the system see hidden objects better. Tests show that their approach works better than previous methods on different datasets.

Camouflaged Object DetectionPolarization CuesRGB FeaturesFeature FusionHigh-frequency ComponentsEdge GuidanceIterative Feedback DecoderHierarchical Representation Learning
Authors
QIfan Zhang, Hao Wang, Xiangrong Qin, Ruijie Li
Abstract
Camouflaged object detection (COD) aims to identify targets that are highly blended with their backgrounds. Recent works have shown that the optical characteristics of polarization cues play a significant role in improving camouflaged object detection. However, most existing polarization-based approaches depend on complex visual encoders and fusion mechanisms, leading to increased model complexity and computational overhead, while failing to fully explore how polarization can explicitly guide hierarchical RGB representation learning. To address these limitations, we propose CPGNet, an asymmetric RGB-polarization framework that introduces a conditional polarization guidance mechanism to explicitly regulate RGB feature learning for camouflaged object detection. Specifically, we design a lightweight polarization interaction module that jointly models these complementary cues and generates reliable polarization guidance in a unified manner. Unlike conventional feature fusion strategies, the proposed conditional guidance mechanism dynamically modulates RGB features using polarization priors, enabling the network to focus on subtle discrepancies between camouflaged objects and their backgrounds. Furthermore, we introduce a polarization edge-guided frequency refinement strategy that enhances high-frequency components under polarization constraints, effectively breaking camouflage patterns. Finally, we develop an iterative feedback decoder to perform coarse-to-fine feature calibration and progressively refine camouflage prediction. Extensive experiments on polarization datasets across multiple tasks, along with evaluations on non-polarization datasets, demonstrate that CPGNet consistently outperforms state-of-the-art methods.