Summary
3D shape recognition has drawn much attention in recent years. Despite the amazing progress on view-based 3D feature description, previous multi-view based methods suffer from a burden in computa-tion efficiency compared with point cloud based methods. To overcome the limitation, we propose a novel light-weight multi-view based network built on parameterized-view-learning mechanism, PVLNet, which can achieve the state-of-the-art performance with only 1/10 FLOPs compared with previous multi-view based methods. Guided by the parameterized-view-learning mechanism, the views are directly built as parameters of PVLNet which can be automatically optimized by gradient descent. A simplified differentiable depth map generator is used to ensure the gradient propagation when generating depth images from view parameters. Then multi-view features extracted by CNNs are aggregated by global max-pooling. Our experimental results on ModelNet40 and ScanObjectNN demonstrate the superior per-formance of the proposed method. The visualization of the networks attention further interprets the ef-ficiency of our PVLNet.