KEYWORDS: 3D modeling, Performance modeling, RGB color model, Feature extraction, Data modeling, 3D image processing, Pattern recognition, Overfitting, Education and training, Cameras
Methods for recognizing 3D models from multiple views have shown remarkable performance in 3D model recognition. The primary challenge faced by these methods is integrating features from multiple views into a single, highly descriptive 3D shape descriptor. This paper proposes a Drop-based Multi-view Recognition Network (DropNet) for accurate recognition of multiple views of 3D models. The proposed DropNet first extracts multi-view features using a backbone network, and then randomly discards individual view features through the DropView module, which makes the number of view combinations that can be learned by the network grow at the level of combinatorial numbers, which can effectively improve the network's ability to recognize multi-views, and at the same time, improve the robustness of the network. Subsequently, the remaining view feature values are randomly discarded by the DropPoint module, which further enhances the robustness of the network and effectively prevents overfitting. The experimental results on benchmark datasets show that the DropNet outperforms several recent approaches in terms of multi-view recognition performance.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.