Detecting camouflaged objects is crucial in various applications such as military surveillance, wildlife conservation, and in search and rescue operations. However, the limited availability of camouflaged object data poses a significant challenge in developing accurate detection models. This paper proposes a quasi-synthetic data generation by image compositing combined with attention-based deep learning-based harmonization methodology to generate feature-enriched realistic images for camouflaged objects under varying scenarios. In our work, we developed a diverse set of images to simulate different environmental conditions, including lighting, shadows, fog, dust, and snow, to test our proposed methodology. The intention of generating such photo-realistic images is to increase the robustness of the model with the additional benefit of data augmentation for training our camouflaged object detection model(COD). Furthermore, we evaluate our approach using state-of-the-art object detection models and demonstrate that training with our quasi-synthetic images can significantly improve the detection accuracy of camouflaged objects under varying conditions. Additionally, to test the real operational performance of the developed models, we deployed the models on resource-constrained edge devices for real-time object detection to validate the performance of the trained model on quasi-synthetic data compared to the synthetic data generated by conventional neural style transfer architecture.
Camouflage is the art of deception which is often used in the animal world. It is also used on the battlefield to hide military assets. Camouflaged objects hide within their environments by taking on colors and textures that are similar to their surroundings. In this work, we explore the classification and localization of camouflaged enemy assets including soldiers. In this paper we address two major challenges: a) how to overcome the paucity of domain-specific labeled data and b) how to perform camouflage object detection using edge devices. To address the first challenge, we develop a deep neural style transfer model that blends content images of objects such as soldiers, tanks, and mines/improvised explosive devices with style images depicting deserts, jungles, and snow-covered regions. To address the second challenge, we develop combined depth-guided deep neural network models that combine image features with depth features. Previous research suggests that depth features not only contain local information about object geometry but also provide information on the position, and shape for camouflaged object identification and localization. In this work, we use precomputed monocular method for the generation of the depth maps. The novel fusion-based architecture provides an efficient representation learning space for object detection. In addition, we perform ablation studies to measure the performance of depth versus RGB in detecting camouflaged objects. We also demonstrate how such as model can be deployed in edge devices for real-time object identification and localization.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.