Paper
9 September 2019 Virtual multi-modal object detection and classification with deep convolutional neural networks
Nikolaos Mitsakos, Manos Papadakis
Author Affiliations +
Abstract
In this paper we demonstrate how the post-processing of gray-scale images with algorithms which have a singularity enhancement effect can assume the role of auxiliary modalities, as in the case where an intelligent system fuses information from multiple physical modalities. We show that as in multimodal AI-fusion, “virtual” multimodal inputs can improve the performance of object detection. We design, implement and test a novel Convolutional Neural Network architecture, based on the Faster R-CNN network for multiclass object detection and classification. Our architecture combines deep feature representations of the input images, generated by networks trained independently on physical and virtual imaging modalities. Using an analog of the ROC curve, the Average Recall over Precision curve, we show that the fusion of certain virtual modality inputs, capable of enhancing singularities and neutralizing illumination, improve recognition accuracy.
© (2019) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Nikolaos Mitsakos and Manos Papadakis "Virtual multi-modal object detection and classification with deep convolutional neural networks", Proc. SPIE 11138, Wavelets and Sparsity XVIII, 1113805 (9 September 2019); https://doi.org/10.1117/12.2529233
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Sensors

Machine learning

Target detection

Data modeling

Image enhancement

Network architectures

Performance modeling

Back to Top