Next Article in Journal
A Scheduling Solution for Robotic Arm-Based Batching Systems with Multiple Conveyor Belts
Previous Article in Journal
Multiset-Trie Data Structure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MixFormer: A Self-Attentive Convolutional Network for 3D Mesh Object Recognition

Mobile Network Application Technology Laboratory, School of Information Science and Engineering, Ningbo University, 818 Fenghua Road, Ningbo 315211, China
*
Author to whom correspondence should be addressed.
Algorithms 2023, 16(3), 171; https://doi.org/10.3390/a16030171
Submission received: 22 January 2023 / Revised: 26 February 2023 / Accepted: 16 March 2023 / Published: 21 March 2023

Abstract

:
3D mesh as a complex data structure can provide effective shape representation for 3D objects, but due to the irregularity and disorder of the mesh data, it is difficult for convolutional neural networks to be directly applied to 3D mesh data processing. At the same time, the extensive use of convolutional kernels and pooling layers focusing on local features can cause the loss of spatial information and dependencies of low-level features. In this paper, we propose a self-attentive convolutional network MixFormer applied to 3D mesh models. By defining 3D convolutional kernels and vector self-attention mechanisms applicable to 3D mesh models, our neural network is able to learn 3D mesh model features. Combining the features of convolutional networks and transformer networks, the network can focus on both local detail features and long-range dependencies between features, thus achieving good learning results without stacking multiple layers and saving arithmetic overhead compared to pure transformer architectures. We conduct classification and semantic segmentation experiments on SHREC15, SCAPE, FAUST, MIT, and Adobe Fuse datasets. Experimental results show that the network can achieve 96.7% classification and better segmentation results by using fewer parameters and network layers.

1. Introduction

Compared with 2D data, 3D data contains rich spatial information and target details, providing the possibility to achieve more detailed computer vision tasks. At the same time, with the development of 3D vision technology, the cost of acquiring and processing 3D data is getting lower and lower, thus giving rise to emerging application areas such as autonomous driving [1,2,3], augmented reality [4,5], and robotics. In recent years, how to process 3D data with the help of deep learning methods [6,7,8], which are more mature in the field of 2D image research [9,10,11,12,13,14], has received a great deal of attention from scholars. However, unlike a 2D image where pixels are uniformly distributed on a 2D grid, a 3D grid is a collection of connected relationships between a series of points and edges in 3D space. This means that the 3D grid model does not have a regular 3D spatial representation, while the input of data is disordered. The structural differences between 2D images and 3D grid models make it impossible to directly apply the more mature neural network design in the field of 2D image research to 3D mesh models.
To overcome the problem of irregularity and disorder in 3D grid models, some studies have pioneered attempts by designing the input format of 3D grid model data and 3D convolution and 3D pooling methods [15,16,17,18]. Feng et al. and Hanocka et al. [16,17] used grid or grid edge as the basic unit to define the standard input format of 3D grid models and, based on the input features, for 3D grid models design convolutional classification networks based on the input features. However, although the convolution and pooling process can weigh the combination of lower-level features to form higher-level features, the spatial information and dependencies of the lower-level features in the higher-level features are lost [19]. Especially in 3D space, the 3D model is richer in location structure information and the local–local interrelationships are closer.
To preserve inter-feature dependencies, inspired by the self-attentive mechanisms that have been successful in the natural language domain and in the 2D image domain, work such as point cloud transformer (PCT) [20], point transformer (PT) [21], and transformer (METRO) [22] in a 3D point cloud model analysis using transformer architecture. However, these methods applied to the 3D domain encode the input in a way that does not design a self-attentive mechanism based on the local features of the 3D model and are not as interpretable as tokens in NLP. At the same time, such transformer networks, which are stripped from the original convolutional architecture, tend to retain the multi-layered characteristics of convolutional networks, which often implies a huge computational overhead when combined with the self-attentive mechanism with O(n2) time complexity in a transformer [23].
Based on the characteristics of convolutional networks and transformer networks, this paper proposes a self-attentive convolutional network MixFormer for 3D mesh models. The shape features of 3D mesh models are extracted as the semantic information input to the transformer module through the front 3D mesh convolutional module. To ensure that the 3D mesh convolution module can effectively extract the feature information of the 3D mesh model, a 3D convolution kernel is defined in this paper. The introduction of the 3D mesh transformer module enables the network to learn the global association among the 3D mesh shape features, which makes up for the deficiency of the convolution operation in learning long-distance features. And by introducing a vector-type self-noticing mechanism in the transformer network, the spatial information of the 3D grid model is introduced, which makes the transformer module better applicable to the 3D grid model.
In this paper, we conducted classification experiments on SHREC15 and Manifold40 datasets, and segmentation experiments on SCAPE, FAUST, and MIT datasets, and demonstrated the effectiveness of each module on 3D feature learning by ablation experiments. The experiments show that MixFormer can achieve good learning results without stacking multiple layers and with less computational overhead by learning local feature information through the front 3D grid convolution module and establishing dependencies between features by the 3D grid transformer.
In summary, the main contributions of this paper are as follows.
(1)
A 3D convolutional kernel applicable to 3D mesh models is designed to enable the network to extract feature information on 3D grid models with irregularities and disorder.
(2)
The vector-based transformer module is designed to better learn the dependencies between features through a vector-based self-attentive mechanism and a learnable position encoding.
(3)
The shape features extracted by the convolutional network are processed by using the self-attentive mechanism with a pyramidal structure, so that the network can fully establish global feature dependencies while extracting feature information more accurately.

2. Related Work

2.1. 3D Convolutional Network

The mesh of a 3D mesh model consists of a connection relation between vertices and edges with irregular distributions, so it is not possible to use the 2D convolution method directly on the 3D mesh model, which means that a suitable 3D convolution operation needs to be defined for the structure of the 3D mesh model. Feng et al. proposed MeshNet [16], which firstly, defines an input format for the mesh: for each mesh, extract its centroid, the centroid-to-vertex vector, the normal vector of the mesh, and the neighboring mesh coordinates as its input features, and then design the CNN classification network. Hanocka et al. proposed MeshCNN [17], which defines the convolutional neighborhood in terms of edges and defines a formula to transform the four edges of the neighborhood to ensure the invariance of the convolutional operation. Moreover, the dihedral angle, the interior angle of two faces, and the edge length ratio of two faces above the base are used as five-dimensional input features, and the pooling is defined by edge folding in the pooling operation.
The above method follows the 2D convolutional approach in processing 3D mesh models by aggregating local features captured by the convolutional kernel through pooling operations, thus facilitating the extraction of high-level features by the posterior network [12]. Although convolutional operations can effectively capture local information, vision tasks such as object detection [24,25,26,27,28,29], instance segmentation [30,31,32], and key-point detection often require the establishment of long-range dependencies [33], and convolution-based architectures often require stacking multiple layers in order to aggregate local features and improve the performance of convolutional backbone networks [10,34]. However, although the convolutional, pooling process can form higher-level features by weighting the combination of lower-level features, it loses the spatial information and dependencies of the lower-level features in the higher-level features [19]. Three-dimensional grid models form spatial surface features through the connection relationship of points and edges, and the loss of spatial location information often has a significant impact on the extraction of three-dimensional target recognition [12]. For example, in a human model, two meshes that are similar in spatial location may be distributed on different fingers, and if we only focus on the local area, we lose its self-contained semantic information. Therefore, a mechanism for modeling based on global (non-local) dependencies may be a more robust and scalable solution.

2.2. Transformer

Establishing long-range dependencies is not only important for 3D target feature learning, but also for Natural Language Processing (NLP). In recent years, the transformer has become increasingly popular in NLP based on the matching mechanism and parallelizability of the encoding. This has now become a standard tool in NLP in the form of a transformer [35], with prominent examples being the GPT [36,37] and BERT [38,39] models.
Noting the excellent ability of the self-attention to establish long-range dependencies, several research workers have tried to apply the self-attention to the 2D image domain and the three-dimensional vision domain. In the 2D image domain, a simple way to use the self-attention is to replace the convolutional layer with the multi-head self-attention (MHSA) layer proposed in the transformer [35]. SASA [40], AACN [41], SANet [42], Axial-SASA [43], etc., introduce various forms of a transformer (local, global, axial, vector) by replacing the original convolutional layers in the ResNet [10] backbone network. On the other hand, methods such as vision transformer (VIT) [44] and Swin transformer [45] segment images into non-overlapping and overlapping blocks, and then by linearly stacking transformer blocks, while ensuring the interpretability of the input tokens, the feature learning of the image is achieved. In the field of 3D vision, a point cloud transformer (PCT) [20] proposes a point-based transformer, which learns features through a vector representation of the transformer. Point transformer (PT) [21] enhances the potential feature representation of the input to better capture the local features in the point cloud. local features. A mesh transformer (METRO) [22], on the other hand, performs human shape reconstruction on 3D mesh models. However, none of these methods applied to the 3D domain encodes the input in a way that is as interpretable as token in NLP. At the same time, such transformer networks derived from the original convolutional architecture tend to retain the multi-layered nature of convolutional networks, which often implies a huge computational overhead when combined with the self-attention of O ( n 2 ) time complexity in a transformer [23].
Inspired by the above work, in order to better establish the global dependency of feature information in the network and at the same time reduce the number of network parameters so that the network can be better applied to 3D mesh models, three innovations are proposed in this paper: (1) a 3D convolutional kernel is designed for 3D mesh models, which enables the network to extract feature information on 3D mesh models with complexity and disorder; (2) a vector-type transformer block is designed to better learn the dependencies among the features through the vector-based self-attention and the learnable position coding; (3) the shape features extracted by the convolutional network are processed by the self-attention with the pyramidal structure so that the network can fully establish the global feature dependencies while extracting the feature information more accurately.

3. MixFormer

In this paper, we propose a vector self-attention convolutional network, MixFormer, applied to 3D mesh models from the perspective of local feature learning to establish global feature dependencies. The specific network model design is shown in Figure 1, which can be divided into the following two major blocks.
  • 3D mesh convolution block: For the input 3D mesh data, the 3D mesh convolution block defines the local surface by polynomial expression, and acquires the 3D convolution kernel by aggregating the local surface features through clustering operation. Then the data is reduced by 3D sampling to get the shape feature representation of 3D mesh model data.
  • 3D mesh transformer block: for the shape feature representation of the 3D mesh model data obtained after processing by the convolution block, the spatial feature information of the 3D mesh model is introduced through the learnable position encoding, and the global dependency of its high-level semantic features is established using the vector self-attention, so as to realize the downstream tasks such as classification and semantic segmentation of the 3D mesh models.
Compared with previous methods, this method has the following advantages: (1) A learnable 3D convolutional kernel template is introduced to cope with the problem that traditional convolutional operations are difficult to apply to 3D mesh models. (2) The transformer block, which is suitable for establishing long-distance dependencies, is introduced to learn the association between global features and make up for the deficiency of the convolution block in learning long-distance features. (3) The combination of the convolution block and the transformer block, which learns the shape features in the 3D mesh model through the pre-convolution operation and then establishes the association between the features through the transformer block, makes MixFormer achieve good classification and segmentation results without building a multi-level network architecture. Compared with other network models applied to 3D mesh models, the number of parameters and computational effort is significantly reduced.

3.1. 3D Mesh Convolution Block

To perform convolutional operations on surface features of complex 3D mesh models, a polynomial convolutional kernel is designed in this paper. The method is described as follows.
Taking the vertices of the 3D mesh model as the center of the window, we first perform a breadth-first search on each vertex to obtain its K-neighborhood, and call the 3D mesh surface formed by the vertices and interconnected edges contained in its K-neighborhood a local surface window. For the local surface window, we define the polynomial representation equation of the local surface and fit the surface by weight learning to obtain the polynomial expression of the local surface. The obtained polynomials for local surfaces are then clustered to extract the 25 surface shapes commonly found in the model as the base convolution kernels. The similarity of each local surface to the basic surface convolution kernel is obtained by convolving each local surface using the basic surface convolution kernel. This is used as the local surface feature of the vertex neighborhood.

3.1.1. Mesh Local Surface Representation

First of all, since the same model in the 3D model undergoes translation or rotation, its polynomial expression parameters are changed accordingly. In order to eliminate the effect of translation and rotation on the features, the grid window needs to be orthogonalized so that its window center is the origin of the new coordinate system (translation) and its normal vector is the z-axis of the new coordinate system (rotation).
Second, for each vertex’s K-neighborhood, we have the following definition.
V win - i   = v j V i - n e i , j = 1,2 , , K - 1
E win - i   = v a , v b v a , v b V win - i  
Here V i - n e i denotes the set of vertices contained in the K-neighborhood of vertex v i , V win - i   is the set of vertices contained in the window with vertex v i as the window center, and E win - i   is the set of edges contained in the window with vertex v i as the window center.
Since the local window contains only a small amount of 3D surface mesh when the local window is designed small, the ability to aggregate its neighborhood information is limited. Moreover, when the window is designed larger, the window shape becomes complex, and the original 3D coordinates of the vertices alone are not sufficient to describe the mesh features. Therefore, geodesic distances are introduced in related works. However, the calculation of geodesic distance often requires a lot of computational resources, so block distance d is chosen as the approximation of geodesic distance.
Therefore, the fitting function of the local window is shown in Equation (3).
F v c θ = z - θ 0 + θ 1 x + θ 2 y + θ 3 d + θ 4 x 2 + θ 5 y 2 + θ 6 d 2 + θ 7 x y + θ 8 x d + θ 9 y d
Let the fitting function F v c θ of the local window surface equal to 0, train the learnable parameters θ i therein, and fit the local surface polynomial. The function F v c θ obtained after fitting is the approximate representation of the local window.
In the training process, the loss function is the mean loss:
L f = 1 K v V   F v c θ
In order to obtain the basis surfaces in the model as the base convolution kernel, the obtained local window expression features need to be clustered. In this paper, it is divided into 25 types of basis surfaces. The feature representation of the base surface obtained by clustering is the base convolution kernel. Finally, we use the base convolution kernel to convolve each positive-definite local surface to obtain the similarity between the local surface and the base surface as its shape features.

3.1.2. Similarity Measure

In order to perform the clustering operation on the local window, the similarity definition of the local surface polynomials is needed to evaluate the clustering effect. To define the similarity between local surfaces, in this paper, the average distance between surfaces is used as a measure of the similarity between surfaces. For this purpose, combined with Equation (3), we have the following equation:
D i s t S i , S t = v j S i   | F v j θ t |
where the surface polynomials F v i θ and F v t θ corresponding to the local surface S i and the target surface S t (the center of clustering), respectively. | F v j θ t | denotes the distance between the points of the local surface S i and the target surface S t . Since Equation (5) only represents the one-way distance, to reduce loss, we define a two-way difference metric based on Equation (5).
D I F ( S i , S t ) = 1 2 D i s t S i , S t + D i s t S t , S i
For each local spatial surface of the 3D mesh model, the polynomial representation function is F v c θ . We assume that the local space surfaces in the 3D mesh model obey a Gaussian distribution on the data with variance σ and mean μ. Then, for any local space surface S i , the probability that the surface belongs to the target surface is Equation (7).
P S i θ , σ = 1 2 π σ 2 exp - v i S i   ( F i v i | θ - μ ) 2 2 σ 2

3.1.3. Definition of 3D Convolution Operations

In the convolution operation of 2D images, the 2D convolution kernel convolves the local image, the essence of which can be understood as computing the similarity between the 2D convolution kernel and the local image. In Section 3.1.2, we define the similarity function P S i θ , σ , which is used to perform the clustering operation on the local surface windows to obtain the standard surface windows (clustering centers). The standard surface window obtained by the clustering operation can be used as a 3D convolution kernel for the 3D convolution operation. The 3D convolution operation is the process of calculating the 3D convolution kernel and the local surface window, so the similarity measure can still be used in Equation (7).
To make the similarity function P S i θ , σ closer to the common convolution operation, taking the logarithm of the full probability equation yields.
ln P S i θ , σ = - 1 2 σ 2 × v i S   1 × ( F v i | θ - μ ) 2 - 1 2 l n 2 π σ 2
This formula can be converted into a convolution formula:
Y = W X + b
X represents the feature vector ( x 1 , x 2 , , x k ) and x i represents the feature at vertex v i . For example 0, 1 can represent the presence or absence of vertex v i on a local surface, where W represents the weight vector ( w 1 , w 2 , , w k ) , and b represents the bias corresponding to the feature vector.
w i = - 1 2 σ 2 × v i S   1 × ( F v i | θ - μ ) 2
b = - 1 2 l n 2 π σ 2

3.1.4. Three-Dimensional Sampling

After the convolution operation by the polynomial convolution kernel, each vertex aggregates the local surface features of its neighborhood. However, this introduces a new problem, as each vertex contains the local surface features of its neighborhood, so there is a large amount of data redundancy. To reduce the amount of data in the network, 3D sampling can be performed based on the vertices.
Unlike the 3D point cloud models, the 3D mesh models represent the contour features of the model by the combination of vertices and triangulated facets, and the vertices and triangulated facets are also not uniformly distributed on the model, showing a partly dense and partly sparse feature. Therefore, if the vertices are to be sampled, the farthest sampling method, which is commonly used in point cloud models to cover all points in space as uniformly as possible, is not very suitable for 3D mesh models. To solve this problem, we found that Poisson disk sampling can achieve uniform sampling according to the model contour.
The algorithm for Poisson disk sampling is shown in Algorithm 1.
Algorithm 1: Poisson disk sampling
Input: Input vector x i , desired number of samples N
Output: Output vector y j
  • Build a kd tree for samples
  • Allocate   a   heap   S i for each sample to store the weights w i
  • Assign   initialized   weights   to   each   sample .   w i = j w i j
                    w i j = 1 d ^ i j 2 A 3 4 2 N 3 8
  • While   number   of   samples   >   desired :
       Si pull the top sample from heap
       For each sample S i around S j
       Remove w i j from w i
       Update the heap position of w i

3.2. 3D Mesh Transformer Block

Through the 3D convolution operation, we extract the local shape features of the 3D mesh model. In order to establish the dependencies between the features, the local shape features of the 3D mesh model need to be input into the 3D mesh transformer block. The core structure of the 3D mesh transformer block is shown in Figure 2.
Considering the disorder of data in the 3D mesh model, i.e., the model features of the vertices in the 3D mesh are independent of the input order, an MLP network with shared weights is introduced in the first layer of the network to eliminate the effect of input order and extract the 3D mesh model features, feature mapping is performed for local shape features, and then the 3D mesh is mapped through multiple transformer layer and down layers to The global features of the model are then learned through multiple transformer layers and down layers. Finally, the features of the model are aggregated by a global maximum pooling operation and the model type is predicted by an MLP layer. For semantic segmentation operation, the transformer up layer is introduced to restore features to predict the semantic labels of mesh. The internal implementation of the transformer layer, down layer, and up layer is shown in Figure 3.

3.2.1. Vector Self-Attention

The core of the model is the transformer layer containing residual blocks. vector self-attentiveness enables the exchange between local features and the acquisition of associations between global features. By linear mapping, the dimensionality of the data computed by the vector-based self-attention can be reduced.
The input of the traditional transformer’s self-attention contains three matrices, Q R n × D k , K R m × D k , and V R m × D v . Q , K , V represent query, key, and value. K , V is like a key–value relationship, which is a one-to-one correspondence.
A t t e n t i o n ( Q , K , V ) = s o f t m a x Q K + p o s V
Self-attentions can be divided into two categories: scalar self-attentions and vector self-attentions.
The scalar self-attention can be expressed as Equation (13).
o i = x i X   ψ q i k j + p o s v j
Corresponding to each input feature x i , o i is the output feature. ψ is a normalization function, such as softmax. pos is a position encoding function. q, k, and v represent the query, key, and value generated by the corresponding x i . It can be expressed as Equation (14).
q i = Q x i    k i = K x i    v i = V x i
Q , K , and V are affine transformations, such as linear mappings or multilayer perceptrons.
In vector self-attentions, the representation of the self-attentive functions differs, like in Equation (15).
o i = x j X ( i )   ψ α ( φ q i , k j ) + p o s ( v j + p o s )
X ( i ) is the set of vertices representing the vertices adjacent to x i , and φ is a correlation function, e.g., +, −, ×, or ⊙. α is a mapping function that maps a self-attentive vector to a self-attentive feature, e.g., MLP.
For the 3D mesh model, the 3D mesh transformer layer used in this paper is based on the vector transformer design. The obtained self-attention function is as follows.
o i = x j X ( i )   ψ α ( q i k j ) + p o s ( v j + p o s )
X ( i ) is the set representing the vertices adjacent to x i (in this experiment it is the nearest neighbor k vertices), and the local self-attention mechanism is applied in the local neighborhood around each vertex. The correlation function φ is chosen as the Hadamard product, and the mapping α function is an MLP with two linear layers and a Relu layer.

3.2.2. Learnable Position Encoding

In the self-attention, positional encoding can mark the relative position relationship between elements and introduce more spatial information. Therefore, positional encoding is often introduced to enhance the model effect in natural language processing or 2D image processing. In the 3D model, its original 3D coordinate information can reflect the position relationship between elements, in order to eliminate the influence of the coordinate system and coordinate scale on the results. We introduce a learnable position encoding method.
p o s = β ( p i - p j )
p i and p j denote the 3D coordinates corresponding to the vertices v i and v j of the 3D mesh model. β is an MLP with two linear layers and a Relu layer.
After processing by the transformer layer, new semantic features are obtained after performing local feature fusion. To reduce the number of parameters, a down layer is introduced to sample tokens. First, the input is sampled, and for processing convenience, the sampling function here is the farthest point sampling. Since some feature information is lost after sampling, the KNN method is introduced. The sampled tokens that aggregate the feature information of their K-neighborhoods are obtained by an MLP layer consisting of normalization and Relu and a K-neighborhood local maximum pooling layer.

3.2.3. Transformer down Layer

To extract the low-level features into high-level features and reduce the feature dimension, we design the transformer down layer. The transformer down layer is shown in Figure 3b. Where X represents the input feature vector and P represents the input vertex set. Xnew represents the output feature vector and Pnew represents the output feature vector. First, by sampling, we select a well-distributed subset of vertices Pnew, Pnew ∈ P. To pool the feature vector X associated with P onto the feature vector Xnew associated with Pnew, we use KNN for P and then max-pool each point in P from the K neighboring points in P.

3.2.4. Transformer up Layer

For the semantic segmentation task, we combine the main modules of MixFormer with the U-net network. To decode the feature information extracted by the network, we designed the transformer up layer to map features from the set of downsampled input points P to its superset Pnew. The transformer up layer is shown in Figure 3c. For this purpose, each input point feature is processed by a linear layer, followed by batch normalization and Relu, and then the features are mapped to a higher resolution point set Pnew by trilinear interpolation. Finally, the interpolated features from the previous decoder level are provided with the corresponding encoder-level feature summaries by a jump connection.

3.3. Time Complexity Analysis

At the end of this chapter, we will discuss the time complexity of the methods in this paper.
In the mesh CNN block, we use Equation (7) to perform the similarity measure between the 3D convolution kernel and the local surfaces, which is calculated only with respect to the number of points in the local surfaces, and the number of points in the local surfaces is a custom constant K, so the similarity calculation time is a constant t . To obtain a specified number k of standard surfaces as 3D convolution kernels by clustering, all n local surfaces need to be traversed to calculate similarity. The time complexity is O I n k t . In order to perform 3D convolution operation on a 3D mesh model using 3D convolution kernels, each convolution kernel needs to calculate similarity with all local surfaces, so the time complexity of convolution operation is O n k t . In summary, without considering the operations of the downsampling and local surface division, the time complexity of the 3D convolution module is O I n k t + n k t , where I, k, and t can be regarded as constants, so the time complexity can be simplified to O n .
In the mesh transformer block, the time complexity is mainly focused on the operations of the self-attention. For the self-attention operation of layer L, the input token number is N/4L, and the feature dimension d is 64*2L−1. Therefore, the time complexity of self-attention is O((N/4L)2d), where d and 4L can be regarded as constants, so the time complexity can be simplified to O(N2).

4. Results and Discussion

In this section, we verify the effectiveness of our method in two applications: 3D mesh model classification and surface semantic segmentation. The network design details of the experiments are shown in Figure 4. We also have done ablation experiments for the sampling method, for the related function, and for the position encoding. We also tried different parameters in the mesh transformer block to explore the effect of the number of transformer layers and transformer output layer dimension on the experimental results.
The experimental host configuration is NVIDIA GeForce RTX 2080 Ti graphics card, Intel(R) Core(TM) i5-9600 CPU processor, and 32G RAM.

4.1. Model Parameters

The network model is built based on the Pytorch framework with Pytorch version 1.8.1. The 3D mesh model data is first convolved by the 3D mesh convolution block, and the convolved surface window is taken with the vertex as the center, and the neighborhood range is 152 vertices close to the vertex. The vertices with the convolved surface information aggregated to the neighborhood are sampled by Poisson disk to take 10% of the points, which are then mapped to the feature space by the MLP layer and input to the 3D mesh transformer block.

4.2. Classification Experiments

4.2.1. Classification Experiments Based on SHREC15 Dataset

The SHREC15 dataset is derived from SHREC11 and SHREC14 and contains both rigid and non-rigid models. The SHREC15 dataset contains 50 categories with 24 models in each category, for a total of 1200 3D mesh models. The maximum mesh resolution in the dataset is 60,210 vertices, and the average mesh resolution is 21,141 vertices. In the ratio of 3:1, 18 3D mesh models are randomly selected from each category as the training set and the rest as the test set.
In order to verify the classification effect of MixFormer proposed in this paper, this paper compares the traditional manual feature-based classification method SPH with the 3D convolution-based classification methods MeshNet and MeshCNN. The accuracy and parametric number comparison of each method for classification in the SHREC15 dataset are given in Table 1.
Due to the large number of classes included in SHREC15, only the class classification accuracies and the average accuracy of each method for all real columns for 11 of these models are shown in Table 1. It can be seen that the classification effect of the MixFormer proposed in this paper is better than other comparison methods, and the classification accuracy is 96.7%, which is 5 percentage points better than the optimal method MeshCNN among the comparison methods. This shows that the MixFormer proposed in this paper does have a good ability to learn features on 3D grid models and can achieve good classification results without deep network architecture
Table 2 compares the time and space complexity of the methods in this paper with those of other classification methods. The column Parmas shows the total number of parameters of the network, and the column FLOPs shows the number of floating point operations performed on each input sample, representing the spatial and temporal complexity, respectively. By learning local feature information through the front 3D grid convolution module and establishing dependencies between features through the 3D grid transformer, MixFormer has a stronger feature learning capability, so good learning results can be achieved without stacking multiple layers and with less computational overhead.

4.2.2. Classification Results Based on the Manifold40 Dataset

Manifold40 [46] is derived from the ModelNet40 dataset, which is a new dataset contributed by Hu et al. after repairing the models in ModelNet40 to tight manifolds. In the Manifold40 dataset, 12,300 models with 40 categories are included. Manifold40 is more challenging due to the reconstruction error and simplification distortion in the Manifolds dataset.
To further validate the classification effect of the proposed MixFormer proposed in this paper, experiment was also conducted on the Manifold40 dataset and compared with the vertex-based methods: PointNet++ and PCT, and the mesh feature-based methods: MeshNet, MeshWalker, and SubdivNet. The accuracy of the classification of each method on the Manifold40 dataset is given in Table 3.
From the comparison of the point cloud methods, it can be seen that the transformer-based PCT method works better than the PointNet++ based on multilayer convolutional networks, which implies that it may be important to obtain the dependencies of global features by introducing the transformer structure. From the comparison applied to the 3D grid model, it can be seen that the method in this paper improves by 2.1 percentage points over the SubdivNet method, which is the best among the compared methods, further indicating that the self-supervised convolutional architecture proposed by the method in this paper can learn the 3D mesh model features better.

4.3. Semantic Segmentation Experiments

To validate the feature learning effect of the proposed MixFormer in this paper, we used 370 models from SCAPE, FAUST, MIT, and Adobe Fuse as training data and 18 models from the human category in the SHREC07 dataset as test data. All models were segmented by Maron [49] et al. into eight categories of labels: head, torso, thigh, forearm, hand thigh, calf, and foot.
We combined the main module of MixFormer with the U-Net network to use the acquired model features for downstream tasks, such as semantic segmentation. Limited by the memory size of the graphics card, we first downsampled the model to 1024 local surfaces during the training process, and then predicted the human model. The semantic segmentation effect is shown in Figure 5.
From the segmentation visualization results, it can be seen that the proposed MixFormer has good feature extraction ability and the ability to establish inter-feature dependencies, and can clearly delineate all parts of the human body and achieve good segmentation results. On the sampled model, mIoU is 0.849. However, some errors do exist at the joint demarcation line, which may be due to the fact that the marker of the model is not particularly fine and the joint demarcation line is difficult to define from the semantic point of view.

4.4. Ablation Experiments

To verify the effectiveness of the sampling method in MixFormer and the feature extraction method in transformer block, the following sets of ablation experiments are done for the sampling method, the correlation function, and the position encoding, respectively.

4.4.1. Sampling Method Validity

For the sampling methods, we selected random point sampling, farthest point sampling, and Poisson disk sampling for comparison. The experimental results are shown in Table 4. Compared with the other sampling methods, the sampling points selected by the Poisson disk can better reflect the surface contour characteristics of the original 3D mesh model. As far as the classification results are concerned, the Poisson disk sampling is used for data preprocessing to obtain better classification results for the classification network, 96.7%, which is 3.6 percentage points higher than the optimal method of farthest point sampling among the compared methods.

4.4.2. Related Function

In the scalar transformer, the similarity is often calculated by dot product operation. In this paper, we adopt a vector transformer, which uses addition, subtraction, Hadamard product, or splicing as the correlation function, and obtains the vector output by the correlation function φ, which is used to measure the similarity between vectors. The experimental results in Table 5 show that compared with the traditional dot product operation of the scalar transformer, better classification results are often achieved by using the vector transformer, which indicates that the vector transformer may be more suitable than the scalar transformer for extracting the spatial features of the 3D mesh model. When the Hadamard product is used as the correlation function, a good classification result can be obtained with a relatively low number of parameters and arithmetic power consumption.

4.4.3. Position Encoding

For location coding, the following experiments were also conducted in this paper to compare several different location coding methods. As can be seen in Table 6, the classification accuracy of the model is 92.79% when no location encoding is introduced. When the conventional relative location encoding is introduced, the classification accuracy of the model improves to 94.08%, indicating that location encoding can indeed model the location relationship between tokens and introduce spatial information. The classification accuracy is further hinted at when we introduce the learnable location encoding, indicating that the learnable location encoding method can indeed better model the relative location and dependency relationships among tokens. To further explore the sensitivity of each part to location encoding, we acted location encoding on the attention part and feature part separately, and the experimental results showed that the attention part was more sensitive to location encoding. In this network, the learnable location encoding acting on the attention part and feature part together can achieve the optimal feature learning effect of 96.76%, which is nearly 4% higher than that without the introduction of location encoding.

4.4.4. Number of Mesh Transformer Blocks

For the number of mesh transformer blocks, this paper tries different numbers of transformer blocks to explore their effects on the experimental results. From Table 7, we can see that the model performs best when the number of blocks is 2. When there are fewer transformer blocks, the network may not have enough deep-learning features. When there are more transformer blocks, the number of tokens input to the deep transformer block decreases and the token semantic information is blurred, and too many parameters are also introduced, which may reduce the network performance.

4.4.5. Feature Dimensions of the Output Layer

We also tried different output layer feature dimensions to explore their effects on the model performance, and the experimental results are shown in Table 8. The network performance is optimal when the output layer feature dimension is 256.

5. Conclusions

Experiments show that the proposed MixFormer in this paper can achieve 96.7% classification accuracy on the dataset SHREC15, which is better than the rest of the methods applied to 3D mesh models, demonstrating the classification capability of the network. Moreover, we made a simple attempt at its semantic segmentation effect on the 3D mesh model to further demonstrate the feature learning capability of the network through a downstream task. Finally, ablation experiments were also performed to explore the effect of various factors on the network performance. In subsequent studies, its application to the semantic segmentation task of 3D mesh models or the introduction of a transformer decoder module for tasks such as unsupervised model generation can be further investigated in depth.

Author Contributions

Conceptualization, L.H., J.Z. and Y.C.; methodology, L.H., J.Z. and Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

Supported by General Program of National Natural Science Foundation of China (No. 62071260 and No. 62006131). Supported by General Program of National Natural Science Foundation of Zhejiang Province (No. LZ22F020001).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Hazard, C.; Bhagat, A.; Buddharaju, B.R.; Liu, Z.; Shao, Y.; Lu, L.; Omari, S.; Cui, H. Importance Is in Your Attention: Agent Importance Prediction for Autonomous Driving. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 2531–2534. [Google Scholar] [CrossRef]
  2. Klingner, M.; Muller, K.; Mirzaie, M.; Breitenstein, J.; Termohlen, J.-A.; Fingscheidt, T. On the Choice of Data for Efficient Training and Validation of End-to-End Driving Models. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 4802–4811. [Google Scholar] [CrossRef]
  3. Wang, J.; Li, X.; Sullivan, A.; Abbott, L.; Chen, S. PointMotionNet: Point-Wise Motion Learning for Large-Scale LiDAR Point Clouds Sequences. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 4418–4427. [Google Scholar] [CrossRef]
  4. Grishchenko, I.; Ablavatski, A.; Kartynnik, Y.; Raveendran, K.; Grundmann, M. Attention Mesh: High-Fidelity Face Mesh Prediction in Real-Time. arXiv 2020, arXiv:2006.10962. [Google Scholar]
  5. Cohn, B.A.; Maselli, A.; Ofek, E.; Gonzalez-Franco, M. SnapMove: Movement Projection Mapping in Virtual Reality. In Proceedings of the 2020 IEEE International Conference on Artificial Intelligence and Virtual Reality (AIVR), Utrecht, The Netherlands, 14–18 December 2020; pp. 74–81. [Google Scholar] [CrossRef]
  6. Zamanakos, G.; Tsochatzidis, L.; Amanatiadis, A.; Pratikakis, I. A Comprehensive Survey of LIDAR-Based 3D Object Detection Methods with Deep Learning for Autonomous Driving. Comput. Graph. 2021, 99, 153–181. [Google Scholar] [CrossRef]
  7. Han, X.-F.; Laga, H.; Bennamoun, M. Image-Based 3D Object Reconstruction: State-of-the-Art and Trends in the Deep Learning Era. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 1578–1604. [Google Scholar] [CrossRef] [Green Version]
  8. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep Learning for 3D Point Clouds: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4338–4364. [Google Scholar] [CrossRef]
  9. Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar] [CrossRef] [Green Version]
  10. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  11. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2020, arXiv:1905.11946. [Google Scholar]
  12. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  13. Chen, Y.; Yang, J.; Luo, L.; Zhang, H.; Qian, J.; Tai, Y.; Zhang, J. Adaptive Noise Dictionary Construction via IRRPCA for Face Recognition. Pattern Recognit. 2016, 59, 26–41. [Google Scholar] [CrossRef]
  14. Qian, J.; Wong, W.K.; Zhang, H.; Xie, J.; Yang, J. Joint Optimal Transport with Convex Regularization for Robust Image Classification. IEEE Trans. Cybern. 2022, 52, 1553–1564. [Google Scholar] [CrossRef] [PubMed]
  15. Milano, F.; Loquercio, A.; Rosinol, A.; Scaramuzza, D.; Carlone, L. Primal-Dual Mesh Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Red Hook, NY, USA, 6–12 December 2020; Volume 33, pp. 952–963. [Google Scholar]
  16. Feng, Y.; Feng, Y.; You, H.; Zhao, X.; Gao, Y. MeshNet: Mesh Neural Network for 3D Shape Representation. Proc. AAAI Conf. Artif. Intell. 2019, 33, 8279–8286. [Google Scholar] [CrossRef] [Green Version]
  17. Hanocka, R.; Hertz, A.; Fish, N.; Giryes, R.; Fleishman, S.; Cohen-Or, D. MeshCNN: A Network with an Edge. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef] [Green Version]
  18. Xu, H.; Dong, M.; Zhong, Z. Directionally Convolutional Networks for 3D Shape Segmentation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2698–2707. [Google Scholar]
  19. Sabour, S.; Frosst, N.; Hinton, G.E. Dynamic Routing between Capsules. Adv. Neural Inf. Process. Syst. 2017, 30, 3859–3869. [Google Scholar]
  20. Guo, M.-H.; Cai, J.-X.; Liu, Z.-N.; Mu, T.-J.; Martin, R.R.; Hu, S.-M. Pct: Point Cloud Transformer. Comput. Vis. Media 2021, 7, 187–199. [Google Scholar] [CrossRef]
  21. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.H.; Koltun, V. Point Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 16259–16268. [Google Scholar]
  22. Lin, K.; Wang, L.; Liu, Z. End-to-End Human Pose and Mesh Reconstruction with Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1954–1963. [Google Scholar]
  23. Hua, W.; Dai, Z.; Liu, H.; Le, Q. Transformer Quality in Linear Time. In Proceedings of the International Conference on Machine Learning, PMLR, Baltimore, MD, USA, 17–23 July 2022; pp. 9099–9117. [Google Scholar]
  24. Li, Y.; Mao, H.; Girshick, R.; He, K. Exploring Plain Vision Transformer Backbones for Object Detection. arXiv 2022, arXiv:2203.16527. [Google Scholar]
  25. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Computer Vision—ECCV 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar] [CrossRef] [Green Version]
  27. Ren, S.; He, K.; Girshick, R.; Zhang, X.; Sun, J. Object Detection Networks on Convolutional Feature Maps. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1476–1481. [Google Scholar] [CrossRef] [Green Version]
  28. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar] [CrossRef] [Green Version]
  29. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
  30. Dai, J.; He, K.; Sun, J. Instance-Aware Semantic Segmentation via Multi-Task Network Cascades. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3150–3158. [Google Scholar] [CrossRef] [Green Version]
  31. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar] [CrossRef]
  32. Hariharan, B.; Arbeláez, P.; Girshick, R.; Malik, J. Simultaneous Detection and Segmentation. In Computer Vision—ECCV 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2014; Volume 8695, pp. 297–312. [Google Scholar]
  33. Hu, H.; Gu, J.; Zhang, Z.; Dai, J.; Wei, Y. Relation Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3588–3597. [Google Scholar]
  34. Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y.; Lin, H.; Zhang, Z.; Sun, Y.; He, T.; Mueller, J.; Manmatha, R.; et al. ResNeSt: Split-Attention Networks. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 2735–2745. [Google Scholar] [CrossRef]
  35. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  36. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A. Language Models Are Few-Shot Learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  37. Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language Models Are Unsupervised Multitask Learners. OpenAI Blog 2019, 1, 9. [Google Scholar]
  38. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. Roberta: A Robustly Optimized Bert Pretraining Approach. arXiv 2019, arXiv:190711692. [Google Scholar]
  39. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the NAACL-HLT; Association for Computational Linguistics: Minneapolis, MN, USA, 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
  40. Ramachandran, P.; Parmar, N.; Vaswani, A.; Bello, I.; Levskaya, A.; Shlens, J. Stand-Alone Self-Attention in Vision Models. Adv. Neural Inf. Process. Syst. 2019, 32, 1–13. [Google Scholar]
  41. Bello, I.; Zoph, B.; Vaswani, A.; Shlens, J.; Le, Q.V. Attention Augmented Convolutional Networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3286–3295. [Google Scholar]
  42. Zhao, H.; Jia, J.; Koltun, V. Exploring Self-Attention for Image Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10076–10085. [Google Scholar]
  43. Wang, H.; Zhu, Y.; Green, B.; Adam, H.; Yuille, A.; Chen, L.-C. Axial-Deeplab: Stand-Alone Axial-Attention for Panoptic Segmentation. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 108–126. [Google Scholar]
  44. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:201011929. [Google Scholar]
  45. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  46. Hu, S.-M.; Liu, Z.-N.; Guo, M.-H.; Cai, J.-X.; Huang, J.; Mu, T.-J.; Martin, R.R. Subdivision-Based Mesh Convolution Networks. ACM Trans. Graph. 2022, 41, 25. [Google Scholar] [CrossRef]
  47. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  48. Lahav, A.; Tal, A. MeshWalker: Deep Mesh Understanding by Random Walks. ACM Trans. Graph. 2020, 39, 263. [Google Scholar] [CrossRef]
  49. Maron, H.; Galun, M.; Aigerman, N.; Trope, M.; Dym, N.; Yumer, E.; Kim, V.G.; Lipman, Y. Convolutional Neural Networks on Surfaces via Seamless Toric Covers. ACM Trans. Graph. 2017, 36, 71. [Google Scholar] [CrossRef]
Figure 1. MixFormer network structure and 3D target processing flow. MixFormer can be viewed as two blocks: a mesh CNN block and mesh transformer block. A mesh CNN block can acquire local surface features of the 3D mesh model and use them as Token input to the mesh transformer block. The block learns global features by building dependencies between local surface features. The learned features can be used for downstream tasks such as classification and segmentation.
Figure 1. MixFormer network structure and 3D target processing flow. MixFormer can be viewed as two blocks: a mesh CNN block and mesh transformer block. A mesh CNN block can acquire local surface features of the 3D mesh model and use them as Token input to the mesh transformer block. The block learns global features by building dependencies between local surface features. The learned features can be used for downstream tasks such as classification and segmentation.
Algorithms 16 00171 g001
Figure 2. 3D mesh transformer block.
Figure 2. 3D mesh transformer block.
Algorithms 16 00171 g002
Figure 3. Internal implementation of transformer layer, transformer down layer, and transformer up layer.
Figure 3. Internal implementation of transformer layer, transformer down layer, and transformer up layer.
Algorithms 16 00171 g003
Figure 4. MixFormer segmentation network (up) and classification network (down).
Figure 4. MixFormer segmentation network (up) and classification network (down).
Algorithms 16 00171 g004
Figure 5. 3D human model torso structure semantic segmentation effect visualization (each part of the human body is marked with the following colors. Head: orange, torso: yellow, arms: green, forearms: light blue, hands: purple, thighs: root color, calves: dark blue, and feet: red).
Figure 5. 3D human model torso structure semantic segmentation effect visualization (each part of the human body is marked with the following colors. Head: orange, torso: yellow, arms: green, forearms: light blue, hands: purple, thighs: root color, calves: dark blue, and feet: red).
Algorithms 16 00171 g005
Table 1. Classification accuracy of different methods on SHREC15.
Table 1. Classification accuracy of different methods on SHREC15.
MethodClass Accuracy/%Accuracy/%
AlienAntsCatDog1Dog2ManSharkSantaPliersGlassesCamel
SPH87.486.290.489.386.789.190.289.487.189.988.188.2
MeshNet89.589.689.691.490.590.890.189.888.091.490.390.4
MeshCNN91.291.492.190.293.791.692.790.591.890.490.391.7
Ours10010010087.510010010010010010075.096.7
Table 2. Comparison of the classification accuracy of different pooling sampling methods.
Table 2. Comparison of the classification accuracy of different pooling sampling methods.
MethodParmas/MFLOPs/109Accuracy/%
SPH2.44.488.2
MeshNet4.35.190.4
MeshCNN1.35.091.7
Ours0.81.296.7
Table 3. Classification accuracy of different methods on Manifold40.
Table 3. Classification accuracy of different methods on Manifold40.
MethodAccuracy/%
PointNet++ [47]87.9
PCT [20]92.4
MeshNet [16]88.4
MeshWalker [48]90.5
SubdivNet [46]91.5
Ours93.6
Table 4. Classification accuracy of different sampling methods.
Table 4. Classification accuracy of different sampling methods.
MethodAccuracy/%
Random sampling92.76
Farthest point sampling93.17
Poisson disc sampling96.76
Table 5. Classification accuracy of different correlation functions.
Table 5. Classification accuracy of different correlation functions.
Related   Function   φ Parmas/MFLOPs/109Accuracy/%
Add0.751.0293.17
Subtraction0.751.0296.33
Had. product0.751.0296.76
Concatenation0.801.2494.33
Dot product0.842.4293.28
Table 6. Classification accuracy with different Position encoding.
Table 6. Classification accuracy with different Position encoding.
Position EncodingParmas/MFLOPs/109Accuracy/%
None0.641.0292.79
Absolute0.681.0294.08
Relative for attention0.751.0295.44
Relative for feature0.751.0294.17
Relative for both0.751.0296.76
Table 7. Effect of the number of mesh transformer blocks on the classification accuracy.
Table 7. Effect of the number of mesh transformer blocks on the classification accuracy.
Number of Transformer BlocksAccuracy/%
192.7
296.7
387.4
475.2
Table 8. Classification accuracy with different dimensions of the output layer.
Table 8. Classification accuracy with different dimensions of the output layer.
Output Layer DimensionAccuracy/%
12893.4
25696.7
51292.6
102487.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, L.; Zhao, J.; Chen, Y. MixFormer: A Self-Attentive Convolutional Network for 3D Mesh Object Recognition. Algorithms 2023, 16, 171. https://doi.org/10.3390/a16030171

AMA Style

Huang L, Zhao J, Chen Y. MixFormer: A Self-Attentive Convolutional Network for 3D Mesh Object Recognition. Algorithms. 2023; 16(3):171. https://doi.org/10.3390/a16030171

Chicago/Turabian Style

Huang, Lingfeng, Jieyu Zhao, and Yu Chen. 2023. "MixFormer: A Self-Attentive Convolutional Network for 3D Mesh Object Recognition" Algorithms 16, no. 3: 171. https://doi.org/10.3390/a16030171

APA Style

Huang, L., Zhao, J., & Chen, Y. (2023). MixFormer: A Self-Attentive Convolutional Network for 3D Mesh Object Recognition. Algorithms, 16(3), 171. https://doi.org/10.3390/a16030171

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop