PDF  PubReader

Zeng* , Liu* , Li* , Che** , and Wang***: Convolutional Neural Network Based Multi-feature Fusion for Non-rigid 3D Model Retrieval

Hui Zeng* , Yanrong Liu* , Siqi Li* , JianYong Che** and Xiuqing Wang***

Convolutional Neural Network Based Multi-feature Fusion for Non-rigid 3D Model Retrieval

Abstract: This paper presents a novel convolutional neural network based multi-feature fusion learning method for non-rigid 3D model retrieval, which can investigate the useful discriminative information of the heat kernel signature (HKS) descriptor and the wave kernel signature (WKS) descriptor. At first, we compute the 2D shape distributions of the two kinds of descriptors to represent the 3D model and use them as the input to the networks. Then we construct two convolutional neural networks for the HKS distribution and the WKS distribution separately, and use the multi-feature fusion layer to connect them. The fusion layer not only can exploit more discriminative characteristics of the two descriptors, but also can complement the correlated information between the two kinds of descriptors. Furthermore, to further improve the performance of the description ability, the cross-connected layer is built to combine the low-level features with high-level features. Extensive experiments have validated the effectiveness of the designed multi-feature fusion learning method.

Keywords: Convolutional Neural Network , HKS , Multi-Feature Fusion , Non-rigid 3D Model , WKS

1. Introduction

In recent years, with the rapid development of the computer technology and the multimedia technology, more and more 3D models have been used in many research fields such as face recognition, object recognition, self-driving car, virtual reality, biology and 3D game. Efficient 3D model retrieval has becomes a research hot spot in the field of computer vision. Generally, 3D model retrieval includes the following three steps: model preprocessing, feature extraction, and similarity matching [1]. Among these steps, the 3D feature extraction is the most key step and it plays a decisive role in retrieval results. So it has attracted more and more researchers' attentions.

According to the type of the 3D models, the existing 3D feature extraction methods can be divided into two categories: rigid 3D model based method and non-rigid 3D model based method. Most of the existing methods are designed for the rigid 3D model, such as 3D shape contexts descriptor [2], local surface patch (LSP) descriptor [3], THRIFT descriptor [4], spin image (SI) descriptor [5,6], normal histogram (NormHist) [7], rotational projection statistics (RoPS) descriptor [8], fast point feature histogram (FPFH) [9], signature histograms of orientations (SHOT) [10,11], and so on. Although the above methods have achieved good retrieval results, they are designed for rigid 3D models and not suitable for non-rigid 3D models. All the above features are not invariant to the non-rigid deformations of the 3D models. To solve this problem, the researchers began to study the 3D feature extraction method for 3D non-rigid model. Sun et al. [12] proposed the heat kernel signature (HKS) descriptor to describe the local characteristics of the non-rigid 3D models. It is based on diffusion scale-space analysis and characterized by the heat transfer process of the 3D surface. The HKS descriptor is invariant under isometric deformations and stable under perturbations of the model. It has achieved good performance in the application of non-rigid 3D model retrieval [13-15]. However, it is sensitive to the scale changes of the 3D model. Aubry et al. [16] proposed the wave kernel signature (WKS) descriptor to describe the non-rigid 3D model, which describes the average probability of quantum mechanics at a position on a non-rigid 3D model surface. The WKS descriptor explains the relationship between the points on the different spatial scales and the rest of the model surface, and it has a better discriminative ability than the HKS descriptor. Up to now, the HKS descriptor and the WKS descriptor have been used in non-rigid 3D model retrieval separately, and the relationships between them have not been fully investigated. To further improve the retrieval performance, it is necessary to study the effective method to fusion the two kinds of descriptors.

Recently, a large number of deep learning methods have been proposed for feature extraction in the field of computer vision, such as auto-encoder, Convolutional Neural Network (CNN), restricted Boltzmann machine (RBM), Deep Belief Networks (DBN), etc. Many experiments have proved that the deep learning based feature extraction methods are more effective. The most widely used deep learning methods in image analysis field is the CNN, which was proposed in 1998 by LeCun et al. [17] to identify handwritten numbers. In 2012, the CNN's accuracy beyond the second nearly 10% in the competition ImageNet. An important feature of the CNN is that it is very similar to our visual system, and the feature is extracted hierarchically. The high-level feature is a combination of low-level features. From low to high, the features are more and more abstract, and this is more conducive to the representation of semantics. Compared with the previous manual designed features, CNN can be automatically extracted more appropriate features, which can greatly improve the recognition performance. So the CNN has been used in more and more areas. For example, the deep learning methods have been used for 3D shape retrieval. Xie et al. [18] proposed a deep shape descriptor for 3D shape retrieval. Firstly, the multiscale shape distribution features are computed. Then a set of discriminative auto-encoders are trained to extract high-level shape features at different scales. Finally, the outputs from hidden layers of the auto-encoders are concatenated to for the shape descriptor. According to our knowledge, the CNNbased 3D retrieval methods have not been fully investigated. In this paper, we proposed a CNN based multi-feature fusion learning method for non-rigid 3D model retrieval. It can combine the effective discriminative information of the HKS descriptor and the WKS descriptor, which not only includes the time domain information of the HKS descriptor but also makes full use of the frequency domain information of the WKS descriptor. Our experimental results have testified that our proposed method can improve the retrieval performance than the single descriptor based method and other state-of-theart methods.

The rest of the paper is organized as follows. In Section 2, the related works including the HKS descriptor and the WKS descriptor are reviewed. Section 3 gives our proposed CNN-based multi- feature fusion learning method. The non-rigid 3D model retrieval experimental results are presented in Section 4, and some concluding remarks are listed in Section 5.

2. Related Works

For the non-rigid 3D shape retrieval, it is extremely important to select the appropriate feature descriptor. Due to the wide existence of non-rigid deformation, the extracted features must have strong robustness. In this paper, we select the HKS descriptor and the WKS descriptor to describe the 3D local structures, which can describe the 3D local patch from different views and complement with each other.

2.1 HKS Descriptor

Sun et al. [12] proposed the HKS descriptor to describe the local structure of the 3D model. The HKS descriptor is constructed based on the heat diffusion process of the surface of the model, which abandoned the heat kernel's spatial information and only left the time domain information. Considering a 3D model M as a Riemannian manifold, there is the following heat diffusion equation:

(1)
[TeX:] $$\left( \Delta _ { M } + \frac { \partial } { \partial t } \right) u ( x , t ) = 0$$

where, M is the positive semi-definite Laplace-Beltrami operator of M and t is time parameter. The solution u(x,t) of Eq. (1) is the amount of heat on the surface at point x in time t. When the [TeX:] $$u ( x , 0 ) = \delta ( x - y )$$ is defined as the initial conditions, the solution set of the heat diffusion equation is called heat kernel kt(x,y), which can be written as follows:

(2)
[TeX:] $$k _ { t } ( x , y ) = \sum _ { k = 1 } ^ { \infty } e ^ { - \lambda , t } \phi _ { i } ( x ) \phi _ { i } ( y )$$

where, x and y are the points of the 3D model, [TeX:] $$\lambda _ { i } \geq 0 \text { is the } i ^ { \text { th } }$$ eigenvalue and [TeX:] $$\phi _ { i } ( x ) \text { is the } i ^ { \text { th } }$$ eigenfunction of the Laplace-Beltrami operator [TeX:] $$\Delta _ { M } , \text { satisfying } \Delta _ { M } \phi _ { i } = \lambda _ { i } \phi _ { i }$$.The heat kernel signature of the point x of the 3D model at time t can be expressed as [12]:

(3)
[TeX:] $$h ( x , t ) = k _ { t } ( x , x ) = \sum _ { k = 1 } ^ { \infty } e ^ { - \lambda , t } \phi _ { i } ^ { 2 } ( x )$$

Then the HKS descriptor of the point x can be obtained by computing its corresponding heat kernel signatures at time sequence. The HKS feature has many excellent characteristics: isometric invariance, multi-scale, and time parameters can be used for representing the small distortion on the model. The heat kernel can be regarded as the transfer density function of the Brownian motion on the fluid, so the local distortion of the model surface will not cause much influence on the heat kernel. In addition, HKS features also have some shortcomings. For example, the HKS descriptor is simplified by the heat kernel. Although the search efficiency is improved, but the experimental results show that the heat kernel can be affected by the spatial domain. So discarding the spatial domain information is a limitation of HKS characteristics. Secondly, the HKS descriptor is not invariant to the scale of the model. The HKS descriptor mainly contains the low frequency information of the 3D model, ignoring the high frequency information, so it is not suitable for high precision matching. Furthermore, the time parameter t of the HKS descriptor is not directly related to the intrinsic attributes of the 3D model itself. Therefore, for the HKS descriptor, there are some limitations for describing the 3D models.

2.2 WKS Descriptor

Aubry et al. [16] proposed the WKS descriptor for characterizing pints on non-rigid 3D shapes. Given a model M, the WKS descriptor is represented by measuring the average probability of the particles to be measured at each vertex at an energy level. The energy of the particles is related to the frequency, so the effects of different frequencies can be clearly distinguished. The difference between the WKS descriptor and the HKS descriptor is that the WKS descriptor uses the following Schrodinger equation instead of the heat diffusion equation:

(4)
[TeX:] $$\frac { \partial \Psi } { \partial t } ( x , t ) = i \Delta \Psi ( x , t )$$

where Δ is the Laplace–Beltrami operator of the 3D model, (x,t) is the wave function. Then the WKS descriptor can be defined by the following formula:

(5)
[TeX:] $$W K S ( x , E ) = \sum _ { k = 0 } ^ { \infty } \phi _ { k } ^ { 2 } ( x ) f _ { E } ^ { 2 } \left( E _ { k } \right)$$

where [TeX:] $$f _ { E } ^ { 2 } \left( E _ { k } \right)$$ is the energy probability distribution. In order to select the appropriate energy distribution, let [TeX:] $$f _ { E } ^ { 2 }$$ be the Gaussian distribution, the energy scale [TeX:] $$e = \log \left( E _ { k } \right)$$, there are following formula:

(6)
[TeX:] $$\left\{ \begin{array} { l } { W K S ( x , \cdot ) : \mathrm { R } \rightarrow R }, \\ { W K S ( x , e ) = C _ { e } \sum _ { k = 0 } ^ { \infty } \phi _ { k } ^ { 2 } ( x ) \exp \left( \frac { - \left( e - \log E _ { k } \right) ^ { 2 } } { 2 \sigma ^ { 2 } } \right) } \end{array} \right.$$

where [TeX:] $$C _ { e } = \left( \sum _ { k = 0 } ^ { \infty } \exp \left( \frac { - \left( e - \log E _ { k } \right) ^ { 2 } } { 2 \sigma ^ { 2 } } \right) \right) ^ { - 1 }$$

Like the HKS descriptor, the WKS descriptor also can be looked as an application of a set of filters with the frequency responses [TeX:] $$f _ { E } ^ { 2 } \left( E _ { k } \right)$$. But the HKS descriptor only uses low-pass filters, and the WKS descriptor uses different frequencies to separate different scales. Compared with the HKS descriptor, the WKS descriptor describes the 3D local patch from a different view. The WKS descriptor not only contains low-frequency information, but also contains high-frequency information. The WKS descriptor is invariant to the non-rigid transformations, and it is stable under perturbations of the 3D shape. So the WKS descriptor is suitable for analyzing 3D shapes undergoing non-rigid deformations.

3. Convolutional Neural Network-Based Multi-feature Fusion

In this section, we firstly introduce the extraction method of the HKS distribution and the WKS distribution. Then the architecture of the multi-feature fusion learning networks and the network optimization are described.

3.1 The HKS Distribution and the WKS Distribution

For each vertex of the 3D model, we can calculate its corresponding HKS descriptor and corresponding WKS descriptor separately. Because the numbers of the vertices of different 3D models are different, the HKS descriptors and the WKS descriptors can't be used as the input of the CNN directly. In this paper, we use the shape distribution to describe the 3D model [19], which refers to a probability distribution sampled from a shape function and can be used as the input of the CNN. For the HKS descriptor, we use the multi-scale shape distribution proposed by Xie et al. [18] to obtain the input of the network.

It is a statistics probability distribution of the HKS descriptor of the 3D model. For each scale, we compute the histogram of the HKS descriptor to form the shape distribution. The detailed construction method can be found in [18]. For the WKS descriptor, we can construct the multi-scale shape distribution at each energy using similar method. Here, different scale denotes different energy. In this paper, The HKS multi-scale shape distribution matrix is 128×96, and the number of the diffusion times is 96 and the number of the discrete HKS descriptor values is 128. The WKS multi-scale shape distribution matrix is also 128×96, and the number of the energy values is 96 and the number of the discrete WKS descriptor values is 128. All the above parameters are determined by experiments. Fig. 1 shows the HKS multi-scale shape distributions of the human models and the ant models with different poses. From Fig. 1 we can see that the multi-scale shape distributions are different for different classes, and the 3D models of the same class have similar multi-scale shape distributions. Although these 3D models of the same class have different postures and the details of the shape distributions are different, their main features have been captured by the multi-scale shape distributions.

Fig. 2 shows the WKS multi-scale shape distributions of the human models and the ant models with different poses. From Fig. 2 we can see that the distributions have not clear differences for different models. This is because that the WKS descriptor is represented by the average probabilities of quantum particles of different energy levels. The scale of the WKS multi-scale shape distribution denotes the energy, and the 3D models with different postures of the same class usually correspond different energies. So it is necessary to extract more discriminative features from the HKS multi-scale shape distributions and the WKS multi-scale shape distributions, and study the effective fusion method to further improve the retrieval performance.

Fig. 1.

The HKS multi-scale shape distributions of human models (a) and ant models (b).
1.png

Fig. 2.

The WKS multi-scale shape distributions of human models (a) and ant models (b).
2.png
3.2 The Architecture of the Multi-feature Fusion Learning

Inspired by the multi-modal deep learning method for RGB-D object recognition [20], we proposed a conventional neural network based multi-feature fusion learning architecture for non-rigid 3D model retrieval. Our proposed multi-feature fusion learning architecture includes two CNNs and a multifeature fusion layer. At first, two CNNs are built to learn the HKS and WKS features, and the inputs of them are the HKS and WKS distributions. As shown in Fig. 3, each CNN consists of three convolutional layers (C1, C2, C3), three pooling layers (S1, S2, S3), one cross-connected layer and two fully-connected layers. Then the last pool layer of each CNNs is combined with the penultimate pool layer into a crossconnected layer, which can fully utilize the characteristics of hidden layer to further improve the retrieval performance. The cross-connected layer is composed of S2 layer and S3 layer. Finally, the multi-feature fusion layer is used to fuse the two descriptor of the fully-connected layers. It combines the two kinds of features with two feature transformation matrix Q1 and Q2.

Fig. 3.

Our multi-feature fusion network, where “conv”, ”pool”, “cc”, ”fc1” and “fc2” represent “convolutional”, “mean-pooling”, “cross-connected layers”, “first fully-connected layer of the HKS/WKS feature” and “second fully-connected layer of the HKS/WKS feature” respectively.
3.png

Let A be the output of the second fully-connected layer of the HKS feature, where [TeX:] $$A = \left[ a _ { 1 } , a _ { 2 } , \cdots , a _ { N } \right] \in R ^ { M \times N }$$, where M is the dimension of the activations and N is the number of the training samples. Similarly, B is the output of the second fully-connected layer of the WKS feature, and [TeX:] $$B = \left[ b _ { 1 } , b _ { 2 } , \cdots , b _ { N } \right] \in R ^ { M \times N }$$. The transformation matrixes of the two kinds of features are Q1 and Q2, where [TeX:] $$a _ { i } ^ { \prime } = Q _ { 1 } a _ { i } , b _ { i } ^ { \prime } = Q _ { 2 } b _ { i } , \text { and } a _ { i } ^ { \prime } \text { and } b _ { i } ^ { \prime }$$ are weighted by k1 and k2 to form the final learned features. In the process of learning, we simultaneously optimize the matrixes 1Q , 2 Q and the weights k1 and k2.

3.3 Network Optimization

In order to learn Q1 and Q2 for the two kinds of features to obtain better representations, we consider the discriminative information and related information of the two kinds of features. Our objective is to learn the fusion features to minimize the distances between same-class samples and maximize the distances between different-class samples. So the objective function is defined as follows:

(7)
[TeX:] $$\begin{array} { l } { \min _ { \left\{ Q _ { 1 } , Q _ { 2 } , k _ { 1 } , k _ { 2 } \right\} } F = k _ { 1 } D _ { 1 } \left( Q _ { 1 } \right) + k _ { 2 } D _ { 2 } \left( Q _ { 2 } \right) + \lambda C \left( Q _ { 1 } , Q _ { 2 } \right) } \\ { \text { subject to } k _ { 1 } + k _ { 2 } = 1 , k _ { 1 } \geq 0 , k _ { 2 } \geq 0 , \lambda > 0 } \end{array}$$

where D1 and D2 represent the discriminative terms, C denotes the correlated term, and λ is the weight between the discriminative terms and the related term. Here the discriminative term D1 is defined as:

(8)
[TeX:] $$D _ { 1 } \left( Q _ { 1 } \right) = \sum _ { i j } h \left( t _ { 1 } - y _ { i j } \left( u _ { 1 } - d _ { Q _ { 1 } } \left( a _ { i } , a _ { j } \right) \right) \right)$$

where [TeX:] $$h ( x ) = \max ( 0 , x ) , \text { and } d _ { Q _ { 1 } }$$ represents the distance of transformed feature [TeX:] $$Q _ { 1 } a _ { i } \text { and } Q _ { 1 } a _ { j }$$. The distance [TeX:] $$d _ { Q _ { 1 } }$$ can be computed as:

(9)
[TeX:] $$d _ { Q _ { 1 } } \left( a _ { i } , a _ { j } \right) = \left( Q _ { 1 } a _ { i } - Q _ { 1 } a _ { j } \right) ^ { T } \cdot \left( Q _ { 1 } a _ { i } - Q _ { 1 } a _ { j } \right)$$

If ai and aj are the features from the same class, the distance [TeX:] $$d _ { Q _ { 1 } }$$ should be smaller than a given threshold [TeX:] $$u _ { 1 } - t _ { 1 } \left( u _ { 1 } > t _ { 1 } > 0 \right) . \text { If } a _ { i } \text { and } a _ { j }$$ are the features from different classes, the distance [TeX:] $$d _ { Q _ { 1 } }$$ should be larger than a given threshold [TeX:] $$u _ { 1 } + t _ { 1 }$$. So we can conclude that the distances of the same and different class have the following constraint:

(10)
[TeX:] $$y _ { i j } \left( u _ { 1 } - d _ { Q _ { t } } \left( a _ { i } , a _ { j } \right) > t _ { 1 }\right.$$

For the label yij, we define it as follows: when ai and aj are from the same class, yij = 1. when ai and aj are from different classes, yij = -1. For the WKS feature bi, we have similar definitions and constraints.

The correlated term is used to exploit the complementary information of the two kinds of features, and it is defined by the difference of pairwise distances between the transformed HKS features and the transformed WKS features. The correlated term can be described as:

(11)
[TeX:] $$C \left( Q _ { 1 } , Q _ { 2 } \right) = \sum _ { i j } \left( \sqrt { d _ { Q _ { 1 } } \left( a _ { i } , a _ { j } \right) } - \sqrt { d _ { Q _ { 2 } } \left( b _ { i } , b _ { j } \right) } \right) ^ { 2 }$$

For the two 3D models from the same class, their corresponding distance [TeX:] $$d _ { Q _ { 1 } }$$ is small and their corresponding distance [TeX:] $$d _ { Q _ { 2 } }$$ is also small. For the two 3D models from different classes, their corresponding distance [TeX:] $$d _ { Q _ { 1 } }$$ is large and their corresponding distance [TeX:] $$d _ { Q _ { 2 } }$$ is also large. So we can minimize the correlated term [TeX:] $$C \left( Q _ { 1 } , Q _ { 2 } \right)$$ to optimize the transform matrixes [TeX:] Q _ { 1 } \text { and } Q _ { 2 }.

In this paper, we use an alternating optimization approach to obtain the optimal solution for Equation (7). At first, Q1 and Q2 are fixed while k1 and k2 are optimized. Secondly, k1 and k2 are fixed while Q1 and Q2 are optimized. Furthermore, in order to avoid suboptimal results and increase the non-linearity, we adopt the strategy proposed in [20] and modify the objective function to:

(12)
[TeX:] $$\begin{array} { l } { \min _ { \left\{ Q _ { 1 } , Q _ { 2 } , k _ { 1 } , k \right\} } F = k _ { 1 } ^ { p } D _ { 1 } \left( Q _ { 1 } \right) + k _ { 2 } ^ { p } D _ { 2 } \left( Q _ { 2 } \right) + \lambda C \left( Q _ { 1 } , Q _ { 2 } \right) } \\ { \text { subject to } k _ { 1 } + k _ { 2 } = 1 , k _ { 1 } \geq 0 , k _ { 2 } \geq 0 , \lambda > 0 } \end{array}$$

where p > 1 . In Eq. (12), we use [TeX:] $$k _ { 1 } ^ { p } \text { and } k _ { 2 } ^ { p }$$ instead of k1 and k2, this will balance the weights of the two kinds of features. Then we can construct the following Lagrangian function:

(13)
[TeX:] $$L ( k , \eta ) = k _ { 1 } ^ { p } D _ { 1 } + k _ { 2 } ^ { p } D _ { 2 } + \lambda C - \eta \left( k _ { 1 } + k _ { 2 } - 1 \right)$$

By setting [TeX:] $$\frac { \partial L ( k , \eta ) } { \partial k } \text { and } \frac { \partial L ( k , \eta ) } { \partial \eta }$$ to 0, the weight [TeX:] $$k _ { m } ( m = 1,2 )$$ can be updated as:

(14)
[TeX:] $$k _ { m } = \frac { \left( \frac { 1 } { D _ { m } } \right) ^ { \frac { 1 } { p - 1 } } } { \sum _ { m = 1 } ^ { 2 } \left( \frac { 1 } { D _ { m } } \right) ^ { \frac { 1 } { p - 1 } } }$$

And then use the back propagation algorithm to update the transform matrix, Q1 and Q2. The derivative of Q1 can be calculated as follows:

(15)
[TeX:] $$\frac { \partial F } { \partial Q _ { 1 } } = 2 Q _ { 1 } \left[ k _ { 1 } ^ { p } \sum _ { i j } y _ { i j } h ^ { \prime } \left( t _ { 1 } - y _ { i j } \left( u _ { 1 } - d _ { Q _ { 1 } } \left( a _ { i } , a _ { j } \right) \right) \right) A _ { i j } ^ { 1 } + \lambda \sum _ { i j } \left( 1 - \sqrt { \frac { d _ { Q _ { 2 } } \left( b _ { i } , b _ { j } \right) } { d _ { Q _ { 1 } } \left( a _ { i } , a _ { j } \right) } } \right) A _ { i j } ^ { 1 } \right]$$

where [TeX:] $$A _ { i j } ^ { 1 } = \left( a _ { i } - a _ { j } \right) \left( a _ { i } - a _ { j } \right) ^ { T }$$. So the transform matrix Q1 can be updated using the following equation:

(16)
[TeX:] $$Q _ { 1 } = Q _ { 1 } - \beta \frac { \partial F } { \partial Q _ { 1 } }$$

The updating method of the transform matrix Q2 is similar to Q1.

In the process of the back-propagation, the derivatives of D and C for ai and bi is used. For ai, they can be described as follows:

(17)
[TeX:] $$\frac { \partial D _ { 1 } } { \partial a _ { j } } = \sum _ { j } y _ { i j } \left( Q _ { 1 } ^ { T } Q _ { 1 } + Q _ { 1 } Q _ { 1 } ^ { T } \right) \left( a _ { i } - a _ { j } \right) h ^ { \prime } \left( t _ { 1 } - y _ { i j } \left( u _ { 1 } - d _ { Q _ { 1 } } \left( a _ { i } , a _ { j } \right) \right) \right)$$

(18)
[TeX:] $$\frac { \partial C } { \partial a _ { j } } = \sum _ { j } \left( Q _ { 1 } ^ { T } Q _ { 1 } + Q _ { 1 } Q _ { 1 } ^ { T } \right) \left( a _ { i } - a _ { j } \right) \sqrt { \frac { d Q _ { 1 } \left( a _ { i } , a _ { j } \right) - d Q _ { 2 } \left( b _ { i } , b _ { j } \right) } { d Q _ { 1 } \left( a _ { i } , a _ { j } \right) } }$$

Likewise for bi to compute the derivatives of D and C. The proposed multi-feature fusion learning method can be listed as follows:

(i) For each training sample, compute the HKS descriptor and the WKS descriptor of each vertex;

(ii) Compute the HKS multi-scale shape distribution and the WKS multi-scale shape distribution of each 3D model;

(iii) Randomly initialize each conventional neural network, and then pre-training to get pretrained ai and bi.respectively.

iv) Perform an alternating method to optimize Qi, Q2, ki and k2 : Fix Q1 and Q2, firstly, update k1 and k2, and then fix k1 and k2, update Q1 and Q2.

(v) Perform back-propagation: fix Q1, Q2, k1 and k2, update the parameters of the conventional neural networks with gradient descent method.

(vi) Repeat (iv)–(v) until convergence or reach the maximum number of iterations.

4. Experimental Results

In this paper, the McGill 3D shape benchmark is used for non-rigid 3D model retrieval experiments to evaluate the performance of our proposed method [20]. This benchmark contains 255 non-rigid 3D models from 10 different categories, including: ant, crab, spectacle, hand, human, octopus, plier, snake, spider and teddy-bear. Each category has 20–30 3D models. Among these models, there are rotational transformation, scale transformation and non-rigid deformation. In our non-rigid 3D model retrieval experiments, 15 3D models per class are randomly selected for training and the others for querying.

Some example 3D models are shown in Fig. 4. The experimental environment was i7-6700 CPU 3.40 GHz 12.0G memory Lenovo computer with MATLAB R2014a. To evaluate the retrieval performance, we use the following measures: nearest neighbor (NN), the first tier (FT), the second tier (ST) and the discounted cumulative gain (DCG).

In this experiments, the HKS descriptor and the WKS descriptor of each vertex are firstly computed. Secondly, the HKS multi-scale shape distribution and the WKS multi-scale shape distribution are calculated, which can be used as the inputs of the two neural networks. Then the two conventional neural networks and the fusion layer are built. Table 1 gives a detailed description of each conventional neural network, including the type of each layer, the size of the convolutional kernel, the stride and the output size of each layer. The activation function of all convolutional layers and fully-connected layers are ReLU. One of the advantages of the ReLU function is that it can learn the features faster. The other advantage is the biology of rationality, which is unilateral. Compared with sigmoid function and tanh function, the ReLU function conforms to the characteristics of biological neurons. In this paper, the pre-training methods are used to initialize the networks, which can make the network converge faster. The two conventional neural networks are trained independently with the HKS distribution and the WKS distribution. The learning rate is set to 0.05, Qm is initialized as an identity of 1000×1000, and k is initialized as [0.5, 0.5]. The [TeX:] $$u _ { 1 } , u _ { 2 } , t _ { 1 } , t _ { 2 } , p , \beta , \lambda$$ were set as 700, 500, 70, 50, 2, 0.0003, 0.009, respectively. In the process of learning, we use the stochastic gradient descent to update the parameters.

Fig. 4.

Example 3D models of the McGill 3D shape benchmark.
4.png

Table 1.

Description of each conventional neural network
Layer Type Patch size Stride

Number of

feature maps

Output size
x Input 128×96
C1 Convolution 5×5 1 6 124×92
S1 Mean pooling 2×2 2 6 62×46
C2 Convolution 5×5 1 12 58×42
S2 Mean pooling 2×2 2 12 29×21
C3 Convolution 2×2 1 16 28×20
S3 Mean pooling 2×2 2 16 14×10
cc Cross-connected 1 9548
fc1 Fully-connected 1 3000
fc2 Fully-connected 1 1000
ff Feature fusion layer 1 2000

To validate the effectiveness of the multi-feature fusion, we compare the retrieval results of our proposed method with single feature based method. For the HKS descriptor, we firstly compute the HKS multi-scale shape distribution. Then the conventional neural network is built and its structure is the same as the single conventional neural network used in the multi-feature fusion learning method. Likewise for the WKS descriptor. Table 2 gives the retrieval results of the single-feature based methods and the multi-feature fusion method. From Table 2 we can see that compared with the single-feature based methods, our proposed multi-feature fusion learning method has better performance with the NN, FT, ST, and DCG measures. It shows that the fusion layer can learn more useful information. So compared with single-feature based method, using the HKS descriptor and the WKS descriptor simultaneously and our proposed conventional neural network based fusion method can improve the retrieval results effectively.

Then we compare our proposed method to the covariance descriptor based method [21], the graphbased method [22], the PCA-based VLAT method [23], the hybrid BOW method [24], hybrid 2D/3D method [25], the CBoFHKS method [26] and the discriminative auto-encoder based shape descriptor (DASD) method [18]. Table 3 gives the retrieval results of our proposed multi-feature fusion learning method and other methods. From Table 3 we can see that our proposed method has the best performance with the FT, ST, and DCG measures and have comparable performance with NN measures. So our proposed method is more robust to non-rigid deformations and it has achieved very competitive results compared with other methods.

Table 2.

Retrieval results compared with single-feature based methods
NN FT ST DCG
HKS 0.819 0.622 0.744 0.827
WKS 0.914 0.775 0.866 0.914
Multi-feature fusion learning method 0.971 0.905 0.981 0.963

Table 3.

Retrieval results compared with other methods
Methods NN FT ST DCG
Covariance method [21] 0.977 0.732 0.818 0.937
Graph-based method [22] 0.976 0.741 0.911 0.933
PCA-based VLAT [23] 0.969 0.658 0.781 0.894
Hybrid BOW [24] 0.957 0.635 0.790 0.886
Hybrid 2D/3D [25] 0.925 0.557 0.698 0.850
CBoFHKS [26] 0.901 0.778 0.876 0.891
DASD [18] 0.988 0.782 0.834 0.955
Multi-feature fusion learning method 0.971 0.905 0.981 0.963

5. Conclusion

In this paper, we have proposed a novel multi-feature fusion learning method for non-rigid 3D model retrieval. Firstly, the HKS descriptor and the WKS descriptor are computed. Then the corresponding HKS multi-scale shape distribution and the WKS multi-scale shape distribution are constructed to be used as the inputs of the conventional neural networks. Finally the conventional neural networks based multi-feature fusion learning framework is built to obtain the fusion feature. The contribution of this paper is that we use the cross-connected layer to combine the low-level features with high-level features, and the fusion layer can learn not only the discriminative characteristics of the two kinds of descriptors but also the correlated information between them. So our proposed fusion feature can make full use of the effective information of the HKS descriptor and the WKS descriptor, and it has achieved a better retrieval performance.

Acknowledgement

This paper is supported by the National Natural Science Foundation of China (No. 61375010 and 61503224), the Fundamental Research Funds for the China Central Universities of USTB (No. FRF-BD- 16-005A), Beijing Key Discipline Development Program (No. XK100080537).

Biography

Hui Zeng
https://orcid.org/0000-0002-4137-7424

She received B.S. and M.S. degrees from Shandong University in 2001 and 2004, respectively, and received the Ph.D. degree from National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences in 2007. She is currently an associate professor at School of Automation and Electrical Engineering, University of Science and Technology Beijing, China. Her main research interests include computer vision, pattern recognition and machine learning.

References

  • 1 Y. Matsuda, N. Miura, A. Nagasaka, H. Kiyomizu, T. Miyatake, "Finger-vein authentication based on deformation-tolerant feature-point matching," Machine Vision and Applications, 2016, vol. 27, no. 2, pp. 237-250. doi:[[[10.1007/s00138-015-0745-3]]]
  • 2 A. Frome, D. Huber, R. Kolluri, T. Bulow, J. Malik, "Recognizing objects in range data using regional point descriptors," in Proceedings of the European Conference on Computer Vision, Prague, Czech Republic, 2004;pp. 224-237. doi:[[[10.1007/978-3-540-24672-5_18]]]
  • 3 H. Chen, B. Bhanu, "3D free-form object recognition in range images using local surface patches," Pattern Recognition Letters, 2007, vol. 28, no. 10, pp. 1252-1262. doi:[[[10.1016/j.patrec.2007.02.009]]]
  • 4 A. Flint, A. Dick, A. Van den Hengel, "Local 3D structure recognition in range images," IET Computer Vision, 2008, vol. 2, no. 4, pp. 208-217. doi:[[[10.1049/iet-cvi:20080037]]]
  • 5 A. E. Johnson, M. Hebert, "Surface matching for object recognition in complex three-dimensional scenes," Image and Vision Computing, 1998, vol. 16, no. 9-10, pp. 635-651. doi:[[[10.1016/S0262-8856(98)00074-2]]]
  • 6 A. E. Johnson, M. Hebert, "Using spin images for efficient object recognition in cluttered 3D scenes," IEEE Transactions on Pattern Analysis and Machine Intelligence, 1999, vol. 21, no. 5, pp. 443-449. doi:[[[10.1109/34.765655]]]
  • 7 G. Hetzel, B. Leibe, P. Levi, B. Schiele, "3D object recognition from range images using local feature histograms," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Kauai, HI, 2001;pp. 394-399. doi:[[[10.1109/CVPR.2001.990988]]]
  • 8 Y. Guo, F. Sohel, M. Bennamoun, M. Lu, J. Wan, "Rotational projection statistics for 3D local surface description and object recognition," International Journal of Computer Vision, 2013, vol. 105, no. 1, pp. 63-86. doi:[[[10.1007/s11263-013-0627-y]]]
  • 9 R. B. Rusu, N. Blodow, M. Beetz, "Fast point feature histograms (FPFH) for 3D registration," in Proceedings of the IEEE International Conference on Robotics and Automation, Kobe, Japan, 2009;pp. 3212-3217. doi:[[[10.1109/ROBOT.2009.5152473]]]
  • 10 F. Tombari, S. Salti, L. Di Stefano, "Unique signatures of histograms for local surface description," in Proceedings of the European Conference on Computer Vision, Crete, Greece, 2010;pp. 356-369. doi:[[[10.1007/978-3-642-15558-1_26]]]
  • 11 S. Salti, F. Tombari, L. Di Stefano, "SHOT: unique signatures of histograms for surface and texture description," Computer Vision Image Understanding, 2014, vol. 125, no. 8, pp. 251-264. doi:[[[10.1016/j.cviu.2014.04.011]]]
  • 12 J. Sun, M. Ovsjanikov, L. Guibas, "A concise and provably informative multi‐scale signature based on heat diffusion," Computer Graphics Forum, 2009, vol. 28, no. 5, pp. 1383-1392. doi:[[[10.1111/j.1467-8659.2009.01515.x]]]
  • 13 A. M. Bronstein, M. M. Bronstein, R. Kimmel, M. Mohmoudi, G. Sapiro, "A Gromov-Hausdorff framework with diffusion geometry for topologically-robust non-rigid shape matching," International Journal of Computer Vision, 2010, vol. 89, no. 2-3, pp. 266-286. doi:[[[10.1007/s11263-009-0301-6]]]
  • 14 Z. Lian, A. Godil, B. Bustos, M. Daoudi, J. Hermans, S. Kawamura, et al., "A comparison of methods for non-rigid 3D shape retrieval," Pattern Recognition, 2013, vol. 46, no. 1, pp. 449-461. doi:[[[10.1016/j.patcog.2012.07.014]]]
  • 15 A. M. Bronstein, M. M. Bronstein, L. J. Guibas, M. Ovsjanikow, "Shape Google: geometric words and expressions for invariant shape retrieval," ACM Transactions on Graphics, 2011, vol. 30, no. 1, pp. 623-636. doi:[[[10.1145/1899404.1899405]]]
  • 16 M. Aubry, U. Schlickewei, D. Cremers, "The wave kernel signature: a quantum mechanical approach to shape analysis," in Proceedings of the IEEE International Conference on Computer Vision Workshops, Barcelona, Spain, 2011;pp. 1626-1633. doi:[[[10.1109/ICCVW.2011.6130444]]]
  • 17 Y. LeCun, L. Bottou, Y. Bengio, P. Haffner, "Gradient-based learning applied to document recognition," in Proceedings of the IEEE, 1998;vol. 86, no. 11, pp. 2278-2324. doi:[[[10.1109/5.726791]]]
  • 18 J. Xie, G. Dai, F. Zhu, E. K. Wong, Y. Fang, "DeepShape: deep-learned shape descriptor for 3D shape retrieval," IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017, vol. 39, no. 7, pp. 1335-1345. doi:[[[10.1109/TPAMI.2016.2596722]]]
  • 19 R. Osada, T. Funkhouser, B. Chazelle, D. Dokin, "Shape distributions," ACM Transactions on Graphics, 2002, vol. 21, no. 4, pp. 807-832. doi:[[[10.1145/571647.571648]]]
  • 20 A. Wang, J. Lu, J. Cai, T. J. Cham, "Large-margin multi-modal deep learning for RGB-D object recognition," and G. Wang IEEE Transactions on Multimedia, 2015, vol. 17, no. 11, pp. 1887-1898. doi:[[[10.1109/TMM.2015.2476655]]]
  • 21 H. Tabia, H. Laga, D. Picard, P. H. Gosselin, "Covariance descriptors for 3D shape matching and retrieval," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, 2014;pp. 4185-4192. doi:[[[10.1109/CVPR.2014.533]]]
  • 22 A. Agathos, I. Pratikakis, P . Papadakis, S. J. Perantonis, P . N. Azariadis, N. S. Sapidis, "Retrieval of 3D articulated objects using a graph-based representation," in Proceedings of the Eurographics Workshop on 3D Object Retrieval, Munich, Germany, 2009;pp. 29-36. doi:[[[10.2312/3DOR/3DOR09/029-036]]]
  • 23 H. Tabia, D. Picard, H. Laga, P . H. Gosselin, "Compact vectors of locally aggregated tensors for 3D shape retrieval," in Proceedings of the Eurographics Workshop on 3D Object Retrieval, Girona, Spain, 2013;pp. 17-24. doi:[[[10.2312/3DOR/3DOR13/017-024]]]
  • 24 P . Papadakis, I. Pratikakis, T. Theoharis, G. Passalis, S. Perantonis, "3D object retrieval using an efficient and compact hybrid shape descriptor," in Proceedings of the Eurographics Workshop on 3D Object Retrieval, Crete, Greece, 2008;pp. 9-16. doi:[[[10.2312/3DOR/3DOR08/009-016]]]
  • 25 G. Lavoue, "Combination of bag-of-words descriptors for robust partial shape retrieval," The Visual Computer, 2012, vol. 28, no. 9, pp. 931-942. doi:[[[10.1007/s00371-012-0724-x]]]
  • 26 Z. Lian, A. Godil, T. Fabry, T. Furuya, J. Hermans, R. Ohbuchi, et al., "SHREC'10 Track: non-rigid 3D shape retrieval," in Proceedings of the Eurographics Workshop on 3D Object Retrieval, Zurich, Switzerland, 2015;pp. 107-120. doi:[[[10.2312/3DOR/3DOR10/101-108]]]