Explaining the complicated mechanisms underlying the image classification capabilities of deep Convolutional Neural Network (CNN) models, continue to pose huge challenges within the field of computer vision. To address this concern, numerous interpretability methods have been devised, aimed at clarifying the image classification process. These methods include approaches such as sensitivity maps, which involve computing the gradients of class outputs with respect to input images, and techniques like class activation mapping (CAM). Furthermore, the incorporation of noise into input images has emerged as an effective strategy for augmenting visualization quality and removing noise. In this paper, we propose two key contributions: the introduction of a novel approach that injects noise into network weights to enhance visualization, involving image gradient updates and average gradient computations; and a new indicator for evaluating interpretability - the center of gravity, and comprehensive experiments were conducted on multiple datasets and different deep neural network models. In the subsequent experimental sections, we demonstrate that our method achieves superior visualization quality and can be combined with other interpretability methods to enhance their performance.
Point cloud registration is a fundamental problem in computer vision and robotics, and has been widely used in various applications, including 3D reconstruction, simultaneous localization and mapping, and autonomous driving. Over the last decades, numerous researchers have devoted themselves to tackling this challenging problem. With the success of deep learning in geometric vision tasks, various types of deep learning-based point cloud registration methods have been proposed, and some of them exhibit good performance. However, their performance degrades rapidly when the overlap between point clouds is reduced. In this paper, we propose a shape completion-guided registration network for face point cloud registration, which takes face shape as a prior knowledge to guide the network to learn a complete face features at the low overlap. Another noteworthy point is that we use the Regularized Projective Manifold Gradient (RPMG) layer to bridge the gap between the neural network output measured in Euclidean space and the rotation parameters described in the SO(3) manifold. In addition, our proposed method infers the translation parameters from the two complementary point clouds and the predicted rotation rather than estimates them directly. For the problem of low overlap point cloud registration, our method can effectively improve the registration accuracy compared to the current state-of-the-art methods. The rotation error and translation error are, respectively, less than 2 degrees and 2mm on the FaceScape-based low overlap face registration dataset.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.