Next Article in Journal
A Multivariable Study of a Traveling Ionosphere Disturbance Using the Arecibo Incoherent Scatter Radar
Next Article in Special Issue
Hybrid Offset Position Encoding for Large-Scale Point Cloud Semantic Segmentation
Previous Article in Journal
Characteristics of R2019 Processing of MODIS Sea Surface Temperature at High Latitudes
Previous Article in Special Issue
CLOUDSPAM: Contrastive Learning On Unlabeled Data for Segmentation and Pre-Training Using Aggregated Point Clouds and MoCo
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AIFormer: Adaptive Interaction Transformer for 3D Point Cloud Understanding

1
School of Computer Science and Technology, Tongji University, Shanghai 201804, China
2
School of Computer Engineering, Jiangsu Ocean University, Lianyungang 222005, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(21), 4103; https://doi.org/10.3390/rs16214103
Submission received: 13 September 2024 / Revised: 26 October 2024 / Accepted: 31 October 2024 / Published: 2 November 2024

Abstract

:
Recently, significant advancements have been made in 3D point cloud analysis by leveraging transformer architecture in 3D space. However, it remains challenging to effectively implement local and global learning within irregular and sparse structures of 3D point clouds. This paper presents the Adaptive Interaction Transformer (AIFormer), a novel hierarchical transformer architecture designed to enhance 3D point cloud analysis by fusing local and global features through the adaptive interaction of features. Specifically, AIFormer mainly consists of several stacked AIFormer Blocks. Each AIFormer module employs the Local Relation Aggregation Module and the Global Context Aggregation Module, respectively, to extract local details of relationships within the reference point and long-range dependencies between reference points. Then, the local and global features are fused using the Adaptive Interaction Module for adaptive interaction to optimize the point representation. Additionally, the AIFormer Block further designs geometric relation functions and contextual relative semantic encoding to enhance local and global feature extraction capabilities, respectively. Extensive experiments on three popular 3D point cloud datasets verify that AIFormer achieves state-of-the-art or comparable performances. Our comprehensive ablation study further validates the effectiveness and soundness of the AIFormer design.

1. Introduction

In recent years, with the development of deep learning and the widespread availability of 3D acquisition devices, 3D point cloud data has played an increasingly important role in various 3D applications, such as autonomous driving, robotics, and augmented reality. Unlike 2D images, the 3D point cloud is a discrete set of points with irregular, sparse, and non-uniform distribution, which hampers the direct application of mature convolutional methods to process the 3D point cloud. Therefore, exploring advanced methods for 3D point cloud analysis is a significant and challenging task.
As a fundamental yet challenging task, various 3D point cloud analysis methods have been proposed to address this problem, including projection-based [1,2,3,4] and voxel-based [5,6,7,8] methods. They transform irregular 3D point clouds into regular grids or voxels to be processed with mature convolution methods. However, such transformations inevitably lead to information loss. Therefore, point-based methods [9,10,11,12,13] operating directly on the raw point cloud have become popular. These methods retain the detailed spatial information of the 3D point cloud, but the Multi-Layer Perceptrons (MLPs) used to capture point features are limited in establishing long-range dependencies, resulting in insignificant performance in large-scale 3D scenes.
With ViT [14] successfully dominating the field of 2D visual understanding, extending the Transformer-like architecture to 3D is a natural idea. PCT [15] and PTv1 [16] are the first concurrent methods to utilize the Transformer architecture for 3D point cloud analysis. PCT employs global attention on the entire point set, leading to unacceptable O ( N 2 ) memory consumption. PTv1, on the other hand, adopts a hierarchical aggregation paradigm similar to PointNet++ [17], which utilizes a local self-attention mechanism to capture the local region features and progressively expands the size of the local region along the network hierarchy, achieving an effective balance between memory usage and computational efficiency. Several follow-up approaches, such as PVT [18] and FastPointTransformer [19], explore enhancement to local region representation to improve the capability of point modeling.
However, Transformer-based architecture extensions of 3D do not yield the same superior performance as 2D tasks. We analyze and identify the key issues: point features only rely on the interactions of points in a local region and lack actual long-range dependencies. Some existing methods [20,21,22,23,24,25] expand the receptive field of a point to enhance the long-range dependence of the point. For example, ST [20] densely samples nearby points and sparsely samples distant points to increase the receptive field of the points. However, this approach necessitates a sophisticated and complex design to manage the varying number of points across different windows. LRP [22] optimizes the pooling operator with dilated maximal pooling to provide networks with large adaptive receptive fields with few parameters. This method is lightweight but has limited ability to capture features. CDFormer [23] introduces a long-range contextual feature capture mechanism for local features and then utilizes cross-attention to achieve short- and long-range contextual communication of point clouds, which improves the global awareness of point cloud features but incurs higher computational and storage costs.
In this paper, we propose an Adaptive Interaction Transformer (AIFormer), which achieves local region feature extraction and modeling long-range dependencies via local relation aggregation and global semantic aggregation, respectively, and then fuses the local and global features with adaptive interaction to improve the representation of points. Specifically, we first use a grid sample to obtain the reference point set that contains a variable number of nearest-neighbor points. Subsequently, we use the Local Relation Aggregation Module and Global Context Aggregation Module, respectively, to extract local details of relationships within the reference point and long-range dependencies between reference points. Finally, the Adaptive Interaction Module achieves adaptive interaction fusion between local and global features. To further enhance feature learning, we designed the geometric relation function for the Local Relation Aggregation Module, which is more suitable for extracting high-frequency geometric features in the local region of the reference point. Additionally, contextual relative semantic encoding serves as position encoding in the Global Context Aggregation Module to capture richer position and semantic information. We refer to the transformer block comprising the above three main modules as Adaptive Interaction Transformer Block (AIFormer Block), as shown in Figure 1, and we build the hierarchical AIFormer architecture by stacking the AIFormer Block. Extensive experiments on three popular 3D point cloud datasets verify that AIFormer achieves state-of-the-art or comparable performances. Comprehensive ablation studies further validate the soundness and effectiveness of the AIFormer design. Our main contributions can be summarized as follows:
  • We propose the Adaptive Interaction Transformer Block by extracting local relational features within the local region of a reference point and modeling long-range dependencies between reference points, respectively. Then the local and global features are fused with Adaptive Interaction. AIFormer achieves effective local and global semantic capture of points.
  • We propose enhancement approaches for local and global features that facilitate the information communication of points. The former extracts high-frequency geometric features from the local region of a reference point using the geometric relation function, while the latter captures richer positional and semantic information through contextual relative semantic encoding as position encoding.
  • We propose a hierarchical transformer architecture based on the AIFormer Block, called AIFormer, which is demonstrated effectiveness via extensive experiments and analysis, and achieves state-of-the-art or comparable performance in several 3D point cloud segmentation task.

2. Related Work

2.1. 3D Point Cloud Analysis

With the popularity of 3D point cloud sensors, the deep learning-based 3D point cloud analysis has been widely studied [26], and the existing 3D point cloud analysis methods can be roughly categorized into three: voxel-based methods, projection-based methods, and point-based methods. The voxel-based methods [5,6,7,8,27,28,29] voxelize the 3D space and then use 3D convolution on the voxels. The earlier method [8] uses dense 3D convolution, which produces cubic memory consumption. To improve computational efficiency, methods such as MinkowskiNet [6] and SSCNs [7] proposed sparse tensors and developed open-source libraries to provide memory-efficient generalized processing operators for 3D voxels. The projection-based methods [1,2,3,4,30,31,32,33,34,35,36,37] project the 3D point cloud to the 2D mesh in order to use standard 2D convolutions such as PointCNN [2] learning the X -transformation from the raw data and then applying the typical element-wise product and sum operations of the typical convolution operator. DFAMNet [31] maps the 3D point cloud to the image and extracts the features using a 2D grid, and then the features are fused by a dual fusion attention mechanism. Voxel-based and projection-based methods suffer from spatial information loss due to quantization or projection operations. In contrast, point-based methods operate directly on the raw data without information loss. PointNet [9] and PointNet++ [17] were pioneering work in point-based methods, proposing Multi-Layer Perceptron (MLP) and hierarchical local feature aggregation structures to capture local features, and this became the paradigm for follow-up point-based methods [10,11,12,13,38,39,40,41]. For example, RandLA-Net [10] utilizes a random sampling strategy to improve computational and memory significantly and then proposes a local feature aggregation module to keep geometric details efficiently. PointASNL [13] adaptively adjusts the sampling points in the local range by reweighting the neighboring points and fuses the information of local points and the entire point cloud in multi-scale with the Local-Nonlocal module.

2.2. Transformer Architectures

In recent years, as the Transformer architecture [42] has dominated the natural language processing tasks, it has also been used to explore the realm of vision tasks [43] with equal success. ViT [14] divides the image into non-overlapping local patches, treats each patch as a token, and extracts the image features directly using a Transformer encoder. Building upon the success of ViT, various architectures with multi-scale resolution or incorporate local-global information [44,45,46,47,48,49,50] have been proposed to address the issue of low-resolution output in dense prediction and the high computational burden of global self-attention. For instance, PVT [48] devises a pyramid hierarchical shrinking architecture to explore multi-resolution features and utilize spatial-reduction attention to save memory. At the same time, Swin Transformer [46] employs window-based attention to introduce a shifted window strategy in successive transform blocks to enable cross-window communication. MixFormer [45] considers the feature fusion between channels and spatial dimensions and designs their bidirectional interaction to provide complementary clues. Further, some approaches proposed local-global feature fusion to improve the performance, such as Twins-SVT [51] leverages interleaved local and global attention for higher throughput and significant performance. SepViT [52] proposed a depthwise separable self-attention to facilitate local-global information interaction within and among the windows.
Our work is inspired by FAT [53], which uses the human visual system to inspire a bidirectional adaptation process to local and global information in a context-aware manner. Both it and our goal is to explore the bidirectional interaction of local and global information. However, FAT is based on 2D image design, and employs classical self-attention for local information processing, while our approach utilizes relational functions for extracting local information.

2.3. Point Cloud Transformers

As the Transformer architecture has shown powerful capabilities in modeling long-range dependencies, it is a natural idea to explore [54,55] the application to non-regular 3D point cloud data. PTv1 [16] and PCT [15] explore the Transformer architecture for 3D point cloud data at the same period. PCT performs global attention on the entire 3D point cloud and achieves decent performance in object classification and shape segmentation, but it suffers from high memory. PTv1 introduces local attention, called vector attention, for each point, which reduces memory consumption and extends the network to large-scale outdoor scene segmentation. Follow-up variants like FastPointTransformer [19] and PTv2 [56] have introduced voxel hashing or grouped vector attention to improve network efficiency and performance. Additionally, several methods [18,57,58] have projected point features into regular features to use more mature 2D transformer technologies. Other approaches [20,59,60,61] have extended the 2D transformer architecture to 3D. For instance, ST [20] adopted the Swin Transformer, facilitating cross-window communication for successive windows and proposing a stratified strategy to extend the receptive field.
Our work explores the modeling and interaction of local and global information of points, alongside many recent studies [19,21,22,23,24,25,62,63,64,65,66]. For example, PointCAT [65] introduces two separate cross-attention transformer branches combined with multi-scale features for point representation. APPT [25] introduces global pivot attention to extract global features and expand the effective receptive field. An asymmetric parallel structure is also designed to integrate local and global information effectively.

3. Method

We first review typical point-based transformer architectures and overview the architecture of AIFormer in Section 3.1. Then, we propose the Adaptive Interaction Transformer Block in Section 3.2, which mainly consists of the Local Relation Aggregation Module, Global Context Aggregation Module, and Adaptive Interaction Module. Finally, we introduce the relevant components of AIFormer in Section 3.3.

3.1. Review Point-Based Transformer

Formally, given an unordered 3D point cloud set X = ( P , F ) with N points, where the coordinate set P = { p i | i = 1 , 2 , , N } R N × 3 denotes the ( x , y , z ) position feature of each point p i , and the feature set F = { f i | i = 1 , 2 , , N } R N × c ) denotes the extra features, which may contain in the corresponding point p i , such as ( r , g , b ) colors, normal vectors, strengths, etc., and c denote the channels of f i .
Then, the nearest-neighbor points set M ( x i ) = { ( p j , f j ) | p j Neighborhood ( p i ) } is constructed for the point x i in 3D space. Neighborhood ( ) denotes the nearest neighbor algorithm, such as KNN, Ball Query, etc. We treat the nearest-neighbor set M ( x i ) as a local region of the point x i . Then, we feed M ( x i ) into individual Linear layers or MLPs to generate the query, key, and value, respectively. The formulation is represented as follows:
Q = Linear q ( M ( x i ) ) , K = Linear k ( M ( x i ) ) , V = Linear v ( M ( x i ) ) ,
here, Q R K × d , K R K × d , V R K × d , K denotes the number of nearest-neighbor points of M ( x i ) and d represents the channels of feature. Then, we can represent the self-attention (SA) operation on a 3D point cloud as:
W a t t n = Softmax ( Q K / d ) ,
f x i = A ( W a t t n V ) ,
where W a t t n R K × K denotes the attention map, A denotes the SUM operation to aggregate the features of the local region. The f x i R d is the output feature vector of M ( x i ) , aggregated to the point x i . For the sake of clarity, we omit bias in Softmax ( ) and describe details in Section 3.2.2. Note that the above equations only show the calculation process for a single local region, and all local regions work in the same manner independently.
Review that the self-attention operation in point-based Transformer, which can only capture features within the nearest-neighbor points set M ( x i ) . This operation makes establishing a favorable long-distance dependency on the reference point difficult, even if the number of nearest-neighbor points is increased.
To address this problem, we propose the Adaptive Interaction Transformer. As shown in Figure 1a, our network architecture utilizes a U-Net structure with skip connections, which contains four levels of encoders and decoders. Specifically, before the Encoder, the Initial Point Embedding layer is used to embed the local information of the raw points. Then, the backbone is stacked by a Downsample or Upsample block and several AIFormer Blocks. The Downsample or Upsample block provides a reference point set, while the AIFormer Block processes and fuses local and global semantic features of the points in an adaptive interaction manner. Following the backbone, the Segmentation Head (Seg. Head) block predicts the category label of each point. Detailed parameters of the model architecture are described in Section 4.2. In the next subsections, we introduce the modules of the AIFormer Block in detail.

3.2. Adaptive Interaction Transformer Block

As shown in Figure 1b, the AIFormer Block consists of three modules: the Local Relation Aggregation Module (LRA) to capture the local relational features of the point, the Global Context Aggregation Module (GCA) to capture the long-range semantic information of the point, and the Adaptive Interaction Module (AI) that provides adaptive interaction for local and global features.

3.2.1. Local Relation Aggregation Module

Considering the balance between computation and memory, we sample a subset of point cloud set X as the reference point set R . we take the reference point set R as the center point and use the entire point cloud set X in the current stage as the query set to construct the nearest-neighbor point set M l o c a l ( R ) R L × K local × d , where L is the number of reference points, and K l o c a l is the number of nearest-neighbor points in each local region in the LRA.
By observing the local regions, we find two simple facts: First, local regions contain a large amount of high-frequency geometric information in 3D, i.e., low-level geometric relation features. Second, local regions naturally have a large amount of overlap between them.
Based on the above two simple facts, we take the convolution instead of complex self-attention operation as the feature extraction unit for local regions. The convolution has an inductive bias that benefits the extraction of high-frequency local information. The natural overlap between local regions facilitates cross-region communication. We formulate the convolution operation for the local region as follows:
f r i l o c a l = A ( ϕ ( r j ) ) , r j M l o c a l ( r i ) ,
where ϕ denotes the convolution operation and f r i R d is the output vector after aggregation of the local region features.
However, we realized that the naive convolution operation ignores the rich geometric relation expression in the local region [17]. Therefore, we design local relational convolutions to obtain inductive local representations for more robust representation.
Figure 2a illustrates the proposed LRA. Specifically, for each local region M l o c a l ( r i ) , we explicitly encode the low-level geometric relation g i j between the reference point r i and its nearest-neighbor points with a position vector. We obtain an explicit representation of the low-level geometric details of the local region in 3D space. Then, instead of naive convolution, we use geometric relation function θ to map g i j to high dimensions to explore high-level semantic relations. This process can be described as:
g ^ i j = θ ( g i j ) ,
f r i l o c a l = A ( g ^ i j x j ) ,
where θ is the geometric relation function, achieved through PointNet-like MLPs, and ⊙ is the Hadamard product. The details of the low-level geometric relation g i j are discussed in Section 4.4.1.
To conclude, the Local Relation Aggregation Module is capable of summarizing the high-level semantic relations within the reference point set R = ( P R , F R l o c a l ) . It is important to note that the semantic representation at this stage remains localized and does not encompass long-range information out of the local region. To address this, we employ the Global Context Aggregation Module to facilitate long-range semantic awareness of the points.

3.2.2. Global Context Aggregation Module

Like the LRA module, modeling long-range semantic dependencies also requires the construction of local regions. In the Global Context Aggregation Module. We take the reference point set R as the center point, and use the reference point set as the query set instead of the entire point cloud set X to construct the nearest-neighbor point set M g l o b a l ( R ) R L × K global × d . L represents the number of reference points, and K g l o b a l denotes the number of nearest-neighbor points in GCA.
Each nearest-neighbor point represents a local region in the Local Relation Aggregation Module. So, the reference point at Global Context Aggregation Module obtains the long-range semantic information whose perception range is far beyond the Local Relation Aggregation Module. In this way, we achieve long-distance dependencies of points within the local region of the reference point and also express the entire point set in terms of a small subset, i.e., the reference point set. Then, as in Equation (1), we project the local region to query, key and value, which can be expressed as:
Q = Linear q ( M g l o b a l ( r i ) ) , K = Linear k ( M g l o b a l ( r i ) ) , V = Linear v ( M g l o b a l ( r i ) ) .
Typically, the bias in Equation (2) is usually to add relative position information. Such as, Swin Transformer [46] utilizes relative position encoding to improve the model performance, ST [20] captures fine-grained position information in 3D by learning contextual relative position encoding, and CDFormer [23] proposes context-aware position encoding, which computes the relative position differences with the current feature interaction, which further enhances the positional cues.
Inspired by these works, we propose contextual relative semantic encoding (CRSE). Specifically, CRSE is calculated by the difference between relative positions and relative features as bias . For given any local region M g l o b a l ( r i ) , we denote the p r i R 1 × 3 is a coordinates vector of r i , and p r j R K global × 3 is the coordinate vector of the nearest-neighbor points. Thus, we give the relative position difference Δ p r i = p r i p r j , where Δ p r i R K global × 3 denotes the relative distance difference between the center point of the the local region and its nearest-neighbor points. Similarly, the relative feature difference defined as Δ f r i = f r i f r j , where Δ f r i R K global × 3 denotes the relative feature difference between the center point of the the local region and its nearest-neighbor points. Then, concatenated along the channel dimensions to obtain the relative semantic difference Δ r i = Δ p r i Δ f r i . CRSE considers both relative position information and feature information of the point r i with its nearest-neighbor points compared to previous work, which adaptively enhances the position and semantic information of the local region. Then the self-attention operation with CRSE can be represented as:
bias = ( Q ) ( α ( Δ r i ) ) + ( K ) ( β ( Δ r i ) ) ,
W a t t n = Softmax ( ( Q K + bias ) / d ) ,
f r i g l o b a l = A ( ( W a t t n ) ( V + γ ( Δ r i ) ) ) ,
where α , β , γ is a trainable function that maps relative semantic differences Δ r i to the same dimension as Q , K , V . W a t t n R K × K denotes the attention map, A denotes the SUM function, and f r i g l o b a l R d denotes the output vector summarized to r i after aggregation of the nearest-neighbor point set M l o c a l ( r i ) . Figure 2 illustrates the proposed Global Context Aggregation Module.
The above equation is the output for a reference point, and the final output is obtained as R = ( P R , F R g l o b a l ) after summarization. Using the Global Context Aggregation Module, the reference point set R achieves the semantic representation of long-range dependencies.

3.2.3. Adaptive Interaction Module

For each reference point r i , after the LRA and the GCA, it ideally captures both the detailed relational features of the local region within the reference point and the long-range semantic dependencies between the reference points are included. However, the features of F R l o c a l and F R g l o b a l still suffer from insufficient interaction.
As shown in Figure 2c, we designed the Adaptive Interaction Module to implement global-to-local and local-to-global cross-interaction to achieve full fusion of the two, which can be represented as follows:
f ^ r i l o c a l = f r i l o c a l φ ( f r i g l o b a l ) ,
f ^ r i g l o b a l = f r i g l o b a l φ ( f r i l o c a l ) ,
f ^ r i = ψ ( f ^ r i l o c a l f ^ r i g l o b a l ) ,
where φ is the adaptive function to normalize the feature, here we use Sigmoid ( ) , ⊙ is the Hadamard product, and ψ is the interaction function, we use Linear layer to achieve interaction fusion of the two features. Equation (11) represents the global-to-local feature interaction, f ^ r i l o c a l R d is the output of the local relation feature after it gets interacted with the global semantic feature. Similarly, Equation (12) represents local-to-global feature interactions, yielding f ^ r i g l o b a l R d . Equation (13) represents the process of adaptive interaction, and f ^ r i indicates the fusion of local and global features to obtain the final output. Summarizing the output features of all reference points yields R ^ = ( P R , F ^ R ) .

3.3. Relevant Components

3.3.1. Initial Point Embedding

Typically, using linear to project the original features prior to input into the transformer block to a uniform high dimensional facilitates the model to efficiently process and learning the sophisticated patterns of the input data. However, we found poor performance after embedding the original features by linear or Position Encoding [16] in the Initial Point Embedding layer. Inspired by the ST [20] employ KPConv [11] in the Initial Point Embedding layer to summarize the features of the local region of each point at the finest level. It yields decent improvements while suffering only minimal additional computation. For different datasets, we set the initial receptive field of KPConv as the corresponding grid size in Table 1 and set the maximum number of nearest neighbor points as 16. Meanwhile, to enhance the position information and avoid the interference of other features, we only use ( x , y , z ) coordinates as the input. The number of channels is also set to a quarter of the first AIFormer Block following the setting of ST. With this setup, we achieve a performance improvement far beyond Linear, while the architectural impact of this change is minimal. We demonstrate the effectiveness of this modification in our experiments (Section 4.4.6 and Section 4.4.5). It also demonstrates the core idea in our paper that a larger receptive field has a more powerful characterization.

3.3.2. Downsample and Upsample

Traditional point-based Transformer architectures [16,20,56] typically utilize downsampling blocks to reduce the scale of the point cloud. For instance, PTv2 [56] introduced Partition-based Pooling, aggregating points within the same grid into a point. This type of downsampling block integrates reference point sampling, nearest-neighbor point querying, and feature aggregation, thereby improving pooling efficiency but resulting in losing point features. In contrast, AIFormer provides a simpler sampling strategy and avoids any loss of detail in the 3D point cloud. We provide only the reference point set in the Downsample Block, and integrate the nearest-neighbor point query and feature aggregation operations in the AIFormer Block.
Specifically, in the Downsample Block, as shown in Figure 3, inspired by Partition-based Pooling, the 3D point cloud space is partitioned into non-overlapping grids based on the ( x , y , z ) coordinate of the 3D point cloud set X . Then record the indexes of all points in each non-empty grid and construct the mapping table. For each grid, we take the point that is closest to the center is referred to as the reference point. Collecting all the reference point is called a reference point set R . In our implementation, the grid size for each stage is determined by the sampling multipliers and the grid size of the previous stage. In this way, both the precise coordinates of the points are preserved and the overlapping area of the receptive field is effectively controlled.
In the Upsample Block, as shown in Figure 3, unlike common practices such as interpolation, we have implemented a more straightforward and precise upsampling operation by index mapping. We use the mapping table saved in the Downsample Block to map the features from the reference point to other points in the same grid via index lookups.
It is worth mentioning that index lookup is the method with the lowest time complexity of O ( 1 ) . The space complexity is O ( N ) . Therefore, it has negligible time consumption and RAM space consumption.

4. Experiments

4.1. Data and Metric

Dataset Description. We conducted 3D point cloud semantic segmentation experiments on two indoor scene datasets, S3DIS [67] and ScanNetv2 [68], and an outdoor scene dataset, SemanticKITTI [69].
  • The Stanford Large-Scale 3D Indoor Spaces (S3DIS) is one of the notable indoor datasets. The dataset focuses on indoor environments, scanning six large indoor areas in three different buildings. These data are primarily structured as RGB-D imagery, and converted into point clouds. S3DIS offers detailed semantic labels, consisting mainly of thirteen common object categories (e.g., walls, floors, chairs, tables, etc.). Typically, Areas 1–4 and 6 are used as the training set, while Area 5 serves as the test set. We follow this convention to obtain results that can be fairly evaluated with existing methods.
  • ScanNetv2 is a richly annotated dataset of 3D scans of indoor environments. It covers a wider variety of indoor environments, extending from educational and office spaces to residential rooms and public spaces. ScanNetv2 includes semantic annotations for over 2.5 million segments in more than 1500 scans, across hundreds of different spaces, containing more than 17.5 million annotation points. Each point is assigned a semantic label from 20 categories (e.g., show curtain, refrigerators, picture, etc.). The dataset is divided into three parts: 1201 scenes for the training, 312 scenes for the validation, and 100 scenes for online testing. Due to its high-quality point cloud annotations and challenging scenes, it is widely used for tasks such as 3D semantic segmentation, 3D object recognition, and other forms of scene understanding. Owing to its high-quality point cloud annotations and the variety of challenging indoor scenes, ScanNetv2 is widely used for tasks such as 3D semantic segmentation, 3D object recognition, and other forms of scene understanding.
  • SemanticKITTI is a dataset specifically focused on outdoor large-scale scenes. The dataset is an extension of the KITTI Vision Benchmarking Suite, using LiDAR point clouds collected from a vehicle moving through urban and rural areas, but with dense semantic annotation of the dataset’s 3D point clouds. Different in scale and classes from the indoor scenes dataset, the data provided consists mainly of nineteen common outdoor environments classes. Not only traffic participants are included, but also functional classes for ground such as parking lots and sidewalks. Typically, sequences 0–7 and 9–10 are used for training, sequence 8 for validation, and sequences 11–21 for online testing. The dataset is primarily used to evaluate semantic segmentation and other tasks in autonomous driving scenes.
Evaluation Metrics. For our experimental dataset, we employed three evaluation metrics: overall accuracy (OA), mean Accuracy (mAcc), and mean intersection over union (mIoU):
OA = TP + TN TP + FP + FN + TN ,
mAcc = 1 C i = 1 C TP i + TN i TP i + FP i + FN i + TN i ,
mIoU = 1 C i = 1 C TP i TP i + FP i + FN i ,
where C is the number of classes, TP is the true positive sample, TN is the true negative sample, FP is the false positive sample, and FN is the false negative sample.

4.2. Experiment Settings

Data Preprocessing and Augmentation. For the S3DIS dataset, we adopted the data preprocessing and data augmentation from PTv1 [16], using ( x , y , z ) coordinates and ( r , g , b ) colors as inputs fed into the network. For ScanNetv2, we followed the data preprocessing and data augmentation of ST [20] and used ( x , y , z ) coordinates and normal vectors as inputs. Meanwhile, SemanticKITTI uses ( x , y , z ) coordinates and strength as inputs, and we followed the data augmentation of SPVNAS [5]. During training, the maximum number of points fed into the network is 80,000 for S3DIS, 100,000 for ScanNetv2, and 120,000 for SemanticKITTI, respectively. Table 1 provides the data augmentation implementation details.
Implementation Details. The main architecture is shown in Figure 1a, and the main hyper-parameter is available in Table 2. There are four stages of encoder and decoder with block depths of [2, 2, 6, 2] and [1, 1, 1, 1], respectively. Specifically, for ScanNetv2, the point number is much larger. Therefore, we equip the four stages of the encoder with additional block depths of [3, 9, 3, 3].
The sampling multipliers for the downsample stage are [×3.5, ×3.0, ×2.5, ×2.0], respectively. The sampling multiplier is the grid size scaling ratio over the previous stage. Take the grid size of ScanNetv2 in Table 1 as an example, the basic grid size is 0.02 m. According to the sampling multiplier, the downsample stage sizes are [0.07, 0.21, 0.525, 1.05] m.
Experimental Environment. All experiments in this paper are performed on a single node machine with four RTX 3090 GPUs, 345G RAM, and Intel(R) Xeon(R) Silver 4210R CPU. Pytorch version is 1.12.1 and CUDA version is 11.3.

4.3. Experiment Result and Analysis

4.3.1. Evaluation on S3DIS

We spent about 18 h training S3DIS, with an average training latency of 314 ms for a single sample and an average inference latency of 259 ms. However, we must emphasize that the time required may vary from device to device or load to load. Table 3 demonstrates the results of AIFormer compared with previous methods. Our method outperforms the point-based transformer methods [15,16,19], achieving improvements of 0.3%, 9.4%, and 0.4%, respectively, and reaches 70.7% on mIoU. AIFormer also outperforms other representative methods, including voxel-based method [6], projection-based method [2], point-based method [10,13], exhibiting significant performance advantages.
The qualitative results of indoor semantic segmentation are visualized in Figure 4. AIFormer leverages both local and global features of points, utilizing adaptive interactive fusion to enhance feature representation. This approach significantly improves the model’s ability to predict small objects in challenging 3D scenes accurately. For example, in the first and second rows of Figure 4, AIFormer is relatively complete in predicting the category of windows and doors, while the comparison methods are relatively poor in semantic segmentation. In the third row, the AIFormer achieves more complete predict the categories of bookcase in this sample compared to the PCT and PTv1.

4.3.2. Evaluation on ScanNetv2

ScanNetv2 contains more categories and scenarios than S3DIS, presenting more difficult challenges. We spent about 14 h training ScanNetv2, with an average training latency of 613 ms for a single sample and an average inference latency of 189 ms. And we summarize the results of comparing AIFormer with existing methods in Table 4. Our method maintains a competitive advantage over the majority of approaches and achieves up to 74.9% on mIoU. AIFormer achieves the second highest accuracy in 8 out of 20 categories, while PTv3 [71], the newest model with significant current performance, dominates the highest accuracy in almost all categories. However, it is worth noting that the performance is significantly improved by greatly expanding the receptive field from 16 to 1024 points, a useful but more brute force approach. In contrast, we indirectly, capture the local region first and then use the local feature to establish the global dependency, which is a more elegant way. On the other hand, PTv3 implements many new techniques and is difficult to deploy compared to other approaches. In contrast, our approach has a simple overall architecture and is easy to deploy.
Figure 5 visualizes the qualitative semantic segmentation results on ScanNetv2. In the first and second rows, AIFormer accurately segments irregularly overlapping objects by capturing short- and long-range features and an adaptive interactive fusion mechanism. It is worth mentioning that the third row, where the objects are chaotically placed and spatially very close, poses a challenge for existing methods with limited receptive fields to distinguish their semantic information. However, AIFormer still predicts the cabinet category more accurately. At the same time, we also note that points with the ground truth labeled “table” were incorrectly predicted as “cabinet”. We attribute the misprediction to an overly large receptive field that allows the point to capture a large number of features from the surrounding space, while the subdivision category fails to achieve sufficient convergence in training and is an issue we need to improve upon further.
By analyzing both quantitative and qualitative results and integrating findings from two indoor datasets, ScanNetv2 and S3DIS, AIFormer effectively achieves the capture of local and global information, models detailed structural information through adaptive interaction, and achieves superior segmentation performance for small objects such as chair, sofa, window, etc., which validates the effectiveness of AIFormer.

4.3.3. Evaluation on SemanticKITTI

S3DIS and ScanNetv2 offer uniformly distributed indoor 3D point clouds, yet we have also evaluated AIFormer’s versatility in outdoor environments with non-uniform distributions. We spent about 20 h training SemanticKITTI, with an average training latency of 369 ms for a single sample and an average inference latency of 143 ms. And we summarize in Table 5 the results of comparing AIFormer with relevant methods, and our method has the same competitive advantage. For instance, compared to the classical Cylinder3D [27] and RangeViT [73], AIFormer improves mIoU metrics by 0.2% and 4%, respectively. Although DFAMNet [31] performs better, it relies on a more sophisticated architecture and requires additional image data to form a pseudo-point cloud. In contrast, AIFormer operates solely on 3D point cloud data and is designed to be simple and lightweight. Our method also achieves surpassing or approaching performance compared to other methods. The qualitative results of the outdoor scene segmentation are visualized in Figure 6. As can be seen within the red rectangle, AIFormer takes full advantage of capturing long-range dependencies at small-sized points, reducing a large number of false predictions compared to other methods.
Modeling effective segmentation models in non-uniformly distributed 3D spaces presents a considerable challenge. Through quantitative and qualitative analysis, it is demonstrated that by expanding the receptive field of a point in AIFormer and then performing local and global feature fusion in an adaptive interactive manner, the points can extract more effective semantic information. The superior performance demonstrates the effectiveness of the AIFormer under different environmental conditions.

4.4. Ablation Study

Since test set of ScanNetv2 requires online submission, we selected validation set of ScanNetv2 as the baseline for conducting the ablation experiments. This setting is followed throughout in this subsection, if not otherwise specified.

4.4.1. Low-Level Geometric Relation g i j

The key to the Local Relation Aggregation Module is learning from low-level geometric relations, thereby how to define g i j is a critical issue. We explore four intuitive definitions to verify that g i j can reflect the geometric relations through flexible definitions, whose results are summarized in Table 6. The data indicates that using only the 3D Euclidean distance (Exp. 1), or the coordinates p i of the reference point p j and its nearest-neighbor point p j (Exp. 2) as g i j , the accuracy can reach 74.3% and 74.5%, respectively. This underscores the efficacy of our LRA, which surpasses self-attention in learning high-level semantic relations from low-level geometric relations within local regions. Notably, as we incorporate additional relations such as coordinate differences, there is a corresponding increase in performance, reaching up to 75.5% (Exp. 4). These findings demonstrate that explicit definitions of geometric relations in local regions with small receptive fields can significantly enhance the exploration and understanding of geometric relations, leading to improved model performance.

4.4.2. Geometric Relation Function θ

The performance and parameters of the geometric relation function θ deployed with different layers are summarised in Table 7. It is found that the best performance of 75.52% is achieved with three layers of MLP. On the contrary, the performance decreases slightly when the number of layers is increased further. There might be too many layers that make it difficult for the network to converge. It is worth noting that a decent performance of 75.21% is achieved with only two layers of θ . This validates the power of the geometric relation function in capturing low-level set relationships in the local region.
Furthermore, combining ID 2 in Table 6 and 1 layer of MLP in Table 7, we can roughly regard it as a simple implementation of Equation (4), i.e., without additional processing of points in the local region, only using the coordinates of the center point and the nearest-neighbor points, and then extracting the features by using the convolution operator. Observation of the data reveals that the performance results are not satisfactory. By improving the low-level geometric relations, we observe that the performance of the model is further improved, which proves our view.

4.4.3. Contextual Relative Semantic Encoding

We evaluate the impact of using only relative position difference Δ p r i , only relative feature difference Δ f r i and using relative semantic difference Δ r i as bias on the Global Context Aggregation Module, respectively. Furthermore, we also evaluate the performance of the global aggregation module without using position encoding. As shown in Table 8, without position encoding comes with unsatisfactory performance, which is consistent with the observation of ST [20]. Next, using Δ p r i or Δ f r i as bias increases the performance by 0.2% and 0.8%, respectively, when compared to the model without position encoding. Further, using Δ r i performance reaches 75.5%, achieving superior performance over ST and PTv2 [56]. This experiment shows that Contextual Relative Semantic Encoding (CRSE) effectively enhances the position and feature information. CRSE-equipped Global Context Aggregation Module significantly improves the ability to capture global semantic relations.

4.4.4. Efficacy of Adaptive Interaction Processing

The core operation of the Adaptive Interaction Module is the adaptive interaction process between global and local features. First, we validate the effectiveness of the adaptive function φ by comparing two typical normalization methods. As demonstrated in Table 9, results obtained without normalization of local and global features were unsatisfactory, which shows the importance of the adaptive function. Further, considering the performance and computational efficiency of the Sigmoid ( ) and Softmax ( ) methods, we choose Sigmoid ( ) as the adaptive function φ more appropriately.
Note that Adaptive Interaction goes through the local and global features and returns their adaptive weight vector. This vector, which is non-trainable and generated directly by the features themselves, avoids the introduction of additional parameters. Instead of employing linear fusion, the Hadamard product of the features is computed, aligning more closely with the mechanisms of attention. In Table 9, we further show the performance impact of different feature interaction methods, showing that our proposed Adaptive Interaction method outperforms traditional linear approaches to feature fusion.

4.4.5. Efficacy of Initial Point Embedding

To intuitively explore the impact of Initial Point Embedding, we show the results of different point embedding methods on training loss and validation mIoU in Figure 7. It is clear that using KPConv [11] as a point embedding method increases 0.5% and 1.1% compared to using PTv1 [16] and linear projection, respectively. Observing the training loss curve, it is found that KPConv converges faster and remains relatively stable without significant fluctuations during training. The impact of Initial Point Embedding on overall performance is further explored in Table 10, and this minor change in architecture resulted in a 1.6% improvement.

4.4.6. Module Design

We conduct extensive ablation studies of the modules introduced in AIFormer to verify the effectiveness of the modules. As shown in Table 10, we ablate different modules as follow: PointEmb. (Initial Point Embedding), DataAug. (Data Augmentation), LRA (Local Relation Aggregation Module), GCA (Global Context Aggregation Module), and AI (Adaptive Interaction Module).
Experiment (1) verifies the effectiveness of the LRA module, which is similar to a direct extension of PointNet++ [17] to the Transformer architecture. Meanwhile, Experiment (2) evaluates the effectiveness of the GCA module, similar to those used in PTv1 [16] and PCT [15]. Together, these experiments establish the benchmark for our model. In Experiment (4), benefiting from the AI, our benchmark results are improved by 5.8% and 3.7%, respectively. Compared to Experiment (3), the mIoU increased from 72.1% to 73.8%, demonstrating the significant role of the adaptive interaction module in enhancing long-range awareness of points. Experiments (5) and (6) validate the effects of the Initial Point Embedding and Data Augmentation strategy, respectively. The observed improvements in mIoU from Experiment (4) demonstrate the nontrivial role of proper data processing in enhancing model performance. The comparison of Experiments (7) and (8) further highlights the effectiveness and reasonableness of our AIFormer architecture.

5. Conclusions

This paper introduces the AIFormer for 3D point cloud analysis, which mainly consists of stacked AIFormer Blocks. The AIFormer Blocks achieve effective capturing of local and global features through the included Local Relation Aggregation Module and Global Context Aggregation Module, and adaptive fusion of the local and global features by using the Adaptive Interaction Module to optimize the point representation in a feature interaction manner. Furthermore, the AIFormer Block further designs geometric relation functions and contextual relative semantic encoding to enhance local and global feature extraction capabilities. We conducted 3D semantic segmentation and extensive ablation experiments on S3DIS, ScanNetv2, and SemanticKITTI to demonstrate the superiority of AIFormer and the soundness of each design. Further, we will explore the potential of AIFormer for other 3D point cloud processing tasks.

Author Contributions

Conceptualization, X.C. and S.Z.; methodology, X.C.; software, X.C.; validation, X.C. and S.Z.; formal analysis, X.C.; investigation, X.C. and S.Z.; resources, S.Z.; data curation, X.C.; writing—original draft preparation, X.C.; writing—review and editing, X.C. and S.Z.; visualization, X.C. and S.Z.; supervision, S.Z. and H.D.; project administration, S.Z. and H.D.; funding acquisition, S.Z. and H.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key R&D Program of China 2023YFC3806000 and 2023YFC3806002, in part by the National Natural Science Foundation of China under Grant 61936014, in part by Shanghai Municipal Science and Technology Major Project No. 2021SHZDZX0100, in part by the Shanghai Science and Technology Innovation Action Plan Project 22511105300 and in part by Fundamental Research Funds for the Central Universities.

Data Availability Statement

Publicly available datasets were analyzed in this study. The SemanticKITTI can be found here (https://semantic-kitti.org/ (accessed on 5 May 2024)). The S3DIS dataset was obtained based on the Stanford Large-Scale 3D Indoor Spaces Dataset by Matterport Camera (https://cvgl.stanford.edu/resources.html (accessed on 5 May 2024)). The ScanNetv2 can be found here (https://kaldir.vc.in.tum.de/scannet_benchmark/) (accessed on 5 May 2024).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Milioto, A.; Vizzo, I.; Behley, J.; Stachniss, C. RangeNet++: Fast and Accurate LiDAR Semantic Segmentation. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), The Venetian Macao, Macau, China, 4–8 November 2019; pp. 4213–4220. [Google Scholar]
  2. Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. PointCNN: Convolution On X-Transformed Points. In Proceedings of the 31th Advances in Neural Information Processing Systems (NeurIPS), Montréal, QC, Canada, 2–8 December 2018. [Google Scholar]
  3. Jiang, L.; Zhao, H.; Liu, S.; Shen, X.; Fu, C.-W.; Jia, J. Hierarchical Point-Edge Interaction Network for Point Cloud Semantic Segmentation. In Proceedings of the 17th IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 10432–10440. [Google Scholar]
  4. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic Graph CNN for Learning on Point Clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef]
  5. Tang, H.; Liu, Z.; Zhao, S.; Lin, Y.; Lin, J.; Wang, H.; Han, S. Searching Efficient 3D Architectures with Sparse Point-Voxel Convolution. In Proceedings of the European Conference on Computer Vision (ECCV), Virtual, 23–28 August 2020; pp. 685–702. [Google Scholar]
  6. Choy, C.; Gwak, J.; Savarese, S. 4D Spatio-Temporal ConvNets: Minkowski Convolutional Neural Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 3070–3079. [Google Scholar]
  7. Graham, B.; Engelcke, M.; van der Maaten, L. 3D Semantic Segmentation with Submanifold Sparse Convolutional Networks. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 9224–9232. [Google Scholar]
  8. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D ShapeNets: A Deep Representation for Volumetric Shapes. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1912–1920. [Google Scholar]
  9. Qi, C.R.; Su, H.; Kaichun, M.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  10. Hu, Q.; Yang, B.; Xie, L.; Rosa, S.; Guo, Y.; Wang, Z.; Trigoni, N.; Markham, A. Learning Semantic Segmentation of Large-Scale Point Clouds with Random Sampling. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 8338–8354. [Google Scholar] [CrossRef] [PubMed]
  11. Thomas, H.; Qi, C.R.; Deschaud, J.E.; Marcotegui, B.; Goulette, F.; Guibas, L.J. KPConv: Flexible and Deformable Convolution for Point Clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6410–6419. [Google Scholar]
  12. Wu, W.; Qi, Z.; Fuxin, L. PointConv: Deep Convolutional Networks on 3D Point Clouds. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 9613–9622. [Google Scholar]
  13. Yan, X.; Zheng, C.; Li, Z.; Wang, S.; Cui, S. PointASNL: Robust Point Clouds Processing Using Nonlocal Neural Networks with Adaptive Sampling. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 5588–5597. [Google Scholar]
  14. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual, 3–7 May 2021. [Google Scholar]
  15. Guo, M.; Cai, J.; Liu, Z.; Mu, T.; Martin, R.R.; Hu, S. PCT: Point Cloud Transformer. Comp. Vis. Media 2021, 7, 187–199. [Google Scholar] [CrossRef]
  16. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.; Koltun, V. Point Transformer. In Proceedings of the 19th IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 16239–16248. [Google Scholar]
  17. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep hierarchical feature learning on point sets in a metric space. In Proceedings of the 30th Advances in Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017; pp. 5105–5114. [Google Scholar]
  18. Zhang, C.; Wan, H.; Shen, X.; Wu, Z. PVT: Point-Voxel Transformer for Point Cloud Learning. arXiv 2021, arXiv:2108.06076. [Google Scholar] [CrossRef]
  19. Park, C.; Jeong, Y.; Cho, M.; Park, J. Fast Point Transformer. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 16928–16937. [Google Scholar]
  20. Lai, X.; Liu, J.; Jiang, L.; Wang, L.; Zhao, H.; Liu, S.; Qi, X.; Jia, J. Stratified Transformer for 3D Point Cloud Segmentation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 8490–8499. [Google Scholar]
  21. Duan, L.; Zhao, S.; Xue, N.; Gong, M.; Xia, G.-S.; Tao, D. ConDaFormer: Disassembled Transformer with Local Structure Enhancement for 3D Point Cloud Understanding. In Proceedings of the 36th Advances in Neural Information Processing Systems (NeurIPS), New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
  22. Li, X.-L.; Guo, M.-H.; Mu, T.-J.; Martin, R.R.; Hu, S.-M. Long Range Pooling for 3D Large-Scale Scene Understanding. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 20–22 June 2023; pp. 10300–10311. [Google Scholar]
  23. Qiu, H.; Yu, B.; Tao, D. Collect-and-Distribute Transformer for 3D Point Cloud Analysis. arXiv 2023, arXiv:2306.01257. [Google Scholar]
  24. He, Y.; Yu, H.; Yang, Z.; Liu, X.; Sun, W.; Mian, A. Full Point Encoding for Local Feature Aggregation in 3-D Point Clouds. IEEE Trans. Neural Netw. Learn. Syst. 2024. early access. [Google Scholar] [CrossRef]
  25. Li, H.; Zheng, T.; Chi, Z.; Yang, Z.; Wang, W.; Wu, B.; Lin, B.; Cai, D. APPT: Asymmetric Parallel Point Transformer for 3D Point Cloud Understanding. arXiv 2023, arXiv:2303.17815. [Google Scholar]
  26. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep Learning for 3D Point Clouds: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4338–4364. [Google Scholar] [CrossRef]
  27. Zhu, X.; Zhou, H.; Wang, T.; Hong, F.; Ma, Y.; Li, W.; Li, H.; Lin, D. Cylindrical and Asymmetrical 3D Convolution Networks for LiDAR Segmentation. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 9934–9943. [Google Scholar]
  28. Yan, X.; Gao, J.; Li, J.; Zhang, R.; Li, Z.; Huang, R.; Cui, S. Sparse Single Sweep LiDAR Point Cloud Segmentation via Learning Contextual Shape Priors from Scene Completion. In Proceedings of the 35th AAAI Conference on Artificial Intelligence (AAAI), Virtually, 2–9 February 2021; pp. 3101–3109. [Google Scholar]
  29. Chen, Y.; Liu, J.; Zhang, X.; Qi, X.; Jia, J. LargeKernel3D: Scaling up Kernels in 3D Sparse CNNs. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 20–22 June 2023; pp. 13488–13498. [Google Scholar]
  30. Xu, C.; Wu, B.; Wang, Z.; Zhan, W.; Vajda, P.; Keutzer, K.; Tomizuka, M. SqueezeSegV3: Spatially-Adaptive Convolution for Efficient Point-Cloud Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Virtual, 23–28 August 2020; pp. 1–19. [Google Scholar]
  31. Li, M.; Wang, G.; Zhu, M.; Li, C.; Liu, H.; Pan, X.; Long, Q. DFAMNet: Dual Fusion Attention Multi-Modal Network for Semantic Segmentation on LiDAR Point Clouds. Appl. Intell. 2024, 54, 3169–3180. [Google Scholar] [CrossRef]
  32. Puy, G.; Boulch, A.; Marlet, R. Using a Waffle Iron for Automotive Point Cloud Semantic Segmentation. In Proceedings of the 20th IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 3356–3366. [Google Scholar]
  33. Zhang, Y.; Zhou, Z.; David, P.; Yue, X.; Xi, Z.; Gong, B.; Foroosh, H. PolarNet: An Improved Grid Representation for Online LiDAR Point Clouds Semantic Segmentation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 9598–9607. [Google Scholar]
  34. Lin, Y.; Yan, Z.; Huang, H.; Du, D.; Liu, L.; Cui, S.; Han, X. FPConv: Learning Local Flattening for Point Convolution. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 4292–4301. [Google Scholar]
  35. Hu, W.; Zhao, H.; Jiang, L.; Jia, J.; Wong, T.-T. Bidirectional Projection Network for Cross Dimension Scene Understanding. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14368–14377. [Google Scholar]
  36. Peng, B.; Wu, X.; Jiang, L.; Chen, Y.; Zhao, H.; Tian, Z.; Jia, J. OA-CNNs: Omni-Adaptive Sparse CNNs for 3D Semantic Segmentation. arXiv 2024, arXiv:2403.14418. [Google Scholar]
  37. Fan, Y.-C.; Liao, K.-Y.; Xiao, Y.-S.; Lu, M.-H.; Yan, W.-Z. 3D Point Cloud Semantic Segmentation System Based on Lightweight FPConv. IEEE Access 2023, 11, 31767–31777. [Google Scholar] [CrossRef]
  38. Gong, J.; Xu, J.; Tan, X.; Song, H.; Qu, Y.; Xie, Y.; Ma, L. Omni-Supervised Point Cloud Segmentation via Gradual Receptive Field Component Reasoning. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 11668–11677. [Google Scholar]
  39. Qian, G.; Li, Y.; Peng, H.; Mai, J.; Hammoud, H.; Elhoseiny, M.; Ghanem, B. PointNeXt: Revisiting PointNet++ with Improved Training and Scaling Strategies. In Proceedings of the 35th Advances in Neural Information Processing Systems (NeurIPS), New Orleans, LA, USA, 28 November–9 December 2022; pp. 23192–23204. [Google Scholar]
  40. Kang, X.; Wang, C.; Chen, X. Region-Enhanced Feature Learning for Scene Semantic Segmentation. IEEE Trans. Multimed. 2023. early access. [Google Scholar] [CrossRef]
  41. Wei, M.; Wei, Z.; Zhou, H.; Hu, F.; Si, H.; Chen, Z.; Zhu, Z.; Qiu, J.; Yan, X.; Guo, Y.; et al. AGConv: Adaptive Graph Convolution on 3D Point Clouds. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 9374–9392. [Google Scholar] [CrossRef] [PubMed]
  42. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Polosukhin, I. Attention is all you need. In Proceedings of the 30th Advances in Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  43. Khan, S.; Naseer, M.; Hayat, M.; Zamir, S.W.; Khan, F.S.; Shah, M. Transformers in Vision: A Survey. ACM Comput. Surv. 2022, 54, 1–41. [Google Scholar] [CrossRef]
  44. Han, Q.; Fan, Z.; Dai, Q.; Sun, L.; Cheng, M.-M.; Liu, J.; Wang, J. On the Connection between Local Attention and Dynamic Depth-Wise Convolution. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual, 25–29 April 2022. [Google Scholar]
  45. Chen, Q.; Wu, Q.; Wang, J.; Hu, Q.; Hu, T.; Ding, E.; Cheng, J.; Wang, J. MixFormer: Mixing Features across Windows and Dimensions. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5239–5249. [Google Scholar]
  46. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the 19th IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  47. Gu, J.; Kwon, H.; Wang, D.; Ye, W.; Li, M.; Chen, Y.-H.; Lai, L.; Chandra, V.; Pan, D.Z. Multi-Scale High-Resolution Vision Transformer for Semantic Segmentation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 12084–12093. [Google Scholar]
  48. Wang, W.; Xie, E.; Li, X.; Fan, D.-P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions. In Proceedings of the 19th IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 548–558. [Google Scholar]
  49. Tu, Z.; Talebi, H.; Zhang, H.; Yang, F.; Milanfar, P.; Bovik, A.; Li, Y. MaxViT: Multi-Axis Vision Transformer. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 459–479. [Google Scholar]
  50. Dong, X.; Bao, J.; Chen, D.; Zhang, W.; Yu, N.; Yuan, L.; Chen, D.; Guo, B. CSWin Transformer: A General Vision Transformer Backbone with Cross-Shaped Windows. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June; pp. 12124–12134.
  51. Chu, X.; Tian, Z.; Wang, Y.; Zhang, B.; Ren, H.; Wei, X.; Xia, H.; Shen, C. Twins: Revisiting the Design of Spatial Attention in Vision Transformers. In Proceedings of the 34th Advances in Neural Information Processing Systems (NeurIPS), Virtual, 6–14 December 2021; pp. 9355–9366. [Google Scholar]
  52. Li, W.; Wang, X.; Xia, X.; Wu, J.; Li, J.; Xiao, X.; Zheng, M.; Wen, S. SepViT: Separable Vision Transformer. arXiv 2022, arXiv:2203.15380. [Google Scholar]
  53. Fan, Q.; Huang, H.; Zhou, X.; He, R. Lightweight Vision Transformer with Bidirectional Interaction. In Proceedings of the 37th Advances in Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 9–15 December 2024. [Google Scholar]
  54. Lahoud, J.; Cao, J.; Khan, F.S.; Cholakkal, H.; Anwer, R.M.; Khan, S.; Yang, M.-H. 3D Vision with Transformers: A Survey. arXiv 2024, arXiv:2208.04309. [Google Scholar]
  55. Lu, D.; Xie, Q.; Wei, M.; Xu, L.; Li, J. Transformers in 3D Point Clouds: A Survey. arXiv 2022, arXiv:2205.07417. [Google Scholar]
  56. Wu, X.; Lao, Y.; Jiang, L.; Liu, X.; Zhao, H. Point Transformer V2: Grouped Vector Attention and Partition-Based Pooling. In Proceedings of the 35th Advances in Neural Information Processing Systems (NeurIPS), New Orleans, LA, USA, 28 November–9 December 2022; pp. 33330–33342. [Google Scholar]
  57. Liu, Z.; Yang, X.; Tang, H.; Yang, S.; Han, S. FlatFormer: Flattened Window Attention for Efficient Point Cloud Transformer. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 20–22 June 2023; pp. 1200–1211. [Google Scholar]
  58. Fan, L.; Pang, Z.; Zhang, T.; Wang, Y.-X.; Zhao, H.; Wang, F.; Wang, N.; Zhang, Z. Embracing Single Stride 3D Object Detector with Sparse Transformer. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 8458–8468. [Google Scholar]
  59. Xiang, P.; Wen, X.; Liu, Y.-S.; Zhang, H.; Fang, Y.; Han, Z. Retro-FPN: Retrospective Feature Pyramid Network for Point Cloud Semantic Segmentation. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 20–22 June 2023; pp. 17780–17792. [Google Scholar]
  60. Yang, Y.; Guo, Y.; Xiong, J.; Liu, Y.; Pan, H.; Wang, P.; Tong, X.; Guo, B. Swin3D: A Pretrained Transformer Backbone for 3D Indoor Scene Understanding. arXiv 2023, arXiv:2304.06906. [Google Scholar]
  61. Hui, L.; Yang, H.; Cheng, M.; Xie, J.; Yang, J. Pyramid Point Cloud Transformer for Large-Scale Place Recognition. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 6078–6087. [Google Scholar]
  62. Yang, J.; Zhang, Q.; Ni, B.; Li, L.; Liu, J.; Zhou, M.; Tian, Q. Modeling Point Clouds with Self-Attention and Gumbel Subset Sampling. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 3318–3327. [Google Scholar]
  63. Ai, D.; Xu, C.; Zhang, X.; Ai, Y.; Bai, Y.; Liu, Y. ASSA-Net: Semantic Segmentation Network for Point Clouds Based on Adaptive Sampling and Self-Attention. In Proceedings of the 2023 5th International Conference on Natural Language Processing (ICNLP), Guangzhou, China, 24–26 March 2023; pp. 60–64. [Google Scholar]
  64. Zhang, C.; Wan, H.; Shen, X.; Wu, Z. PatchFormer: An Efficient Point Transformer with Patch Attention. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11789–11798. [Google Scholar]
  65. Yang, X.; Jin, M.; He, W.; Chen, Q. PointCAT: Cross-Attention Transformer for Point Cloud. arXiv 2023, arXiv:2304.03012. [Google Scholar]
  66. Huang, Z.; Zhao, Z.; Li, B.; Han, J. LCPFormer: Towards Effective 3D Point Cloud Analysis via Local Context Propagation in Transformers. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 4985–4996. [Google Scholar] [CrossRef]
  67. Armeni, I.; Sener, O.; Zamir, A.R.; Jiang, H.; Brilakis, I.; Fischer, M.; Savarese, S. 3D Semantic Parsing of Large-Scale Indoor Spaces. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1534–1543. [Google Scholar]
  68. Rozenberszki, D.; Litany, O.; Dai, A. Language-Grounded Indoor 3D Semantic Segmentation in the Wild. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 125–141. [Google Scholar]
  69. Behley, J.; Garbade, M.; Milioto, A.; Quenzel, J.; Behnke, S.; Stachniss, C.; Gall, J. SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences. In Proceedings of the 17th IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9296–9306. [Google Scholar]
  70. Tang, L.; Zhan, Y.; Chen, Z.; Yu, B.; Tao, D. Contrastive Boundary Learning for Point Cloud Segmentation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 8479–8489. [Google Scholar]
  71. Wu, X.; Jiang, L.; Wang, P.-S.; Liu, Z.; Liu, X.; Qiao, Y.; Ouyang, W.; He, T.; Zhao, H. Point Transformer V3: Simpler, Faster, Stronger. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 4840–4851. [Google Scholar]
  72. Tao, A.; Duan, Y.; Wei, Y.; Lu, J.; Zhou, J. SegGroup: Seg-Level Supervision for 3D Instance and Semantic Segmentation. IEEE Trans. Image Process. 2022, 31, 4952–4965. [Google Scholar] [CrossRef] [PubMed]
  73. Kong, L.; Liu, Y.; Chen, R.; Ma, Y.; Zhu, X.; Li, Y.; Hou, Y.; Qiao, Y.; Liu, Z. Rethinking Range View Representation for LiDAR Segmentation. In Proceedings of the 120th IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 228–240. [Google Scholar]
Figure 1. (a) Framework Overview. (b) Structure of Adaptive Interaction Transformer Block (AIFormer Block). Best viewed in color.
Figure 1. (a) Framework Overview. (b) Structure of Adaptive Interaction Transformer Block (AIFormer Block). Best viewed in color.
Remotesensing 16 04103 g001
Figure 2. Illustration of the Local Relation Aggregation Module, the Global Context Aggregation Module and the Adaptive Interaction Module, respectively. Best viewed in color.
Figure 2. Illustration of the Local Relation Aggregation Module, the Global Context Aggregation Module and the Adaptive Interaction Module, respectively. Best viewed in color.
Remotesensing 16 04103 g002
Figure 3. Illustration of the Downsample and Upsample, best viewed in color.
Figure 3. Illustration of the Downsample and Upsample, best viewed in color.
Remotesensing 16 04103 g003
Figure 4. Visual comparison of our model with other methods [15,16] on S3DIS. Differences in semantic segmentation results are highlighted using red box for clarity.
Figure 4. Visual comparison of our model with other methods [15,16] on S3DIS. Differences in semantic segmentation results are highlighted using red box for clarity.
Remotesensing 16 04103 g004
Figure 5. Visual comparison of our model with other methods [20,59] on ScanNetv2. Note that black indicates ignored labels, and differences in semantic segmentation results are highlighted using red box for clarity.
Figure 5. Visual comparison of our model with other methods [20,59] on ScanNetv2. Note that black indicates ignored labels, and differences in semantic segmentation results are highlighted using red box for clarity.
Remotesensing 16 04103 g005
Figure 6. Visual comparison of our model with other methods [10,27] on SemanticKITTI. Note that black indicates ignored labels.
Figure 6. Visual comparison of our model with other methods [10,27] on SemanticKITTI. Note that black indicates ignored labels.
Remotesensing 16 04103 g006
Figure 7. Validation mIoU and training loss curves of different point embedding methods on the ScanNetv2.
Figure 7. Validation mIoU and training loss curves of different point embedding methods on the ScanNetv2.
Remotesensing 16 04103 g007
Table 1. Data augmentation settings.
Table 1. Data augmentation settings.
DatasetDrop PointsRotateFlipScaleJitterDistortChromaticGrid Size
ScanNetv20.02 m
S3DIS 0.02 m
SemanticKITTI 0.04 m
Table 2. Main hyper-parameter settings.
Table 2. Main hyper-parameter settings.
DatasetEpochLeraning RateWeight DecaySchedulerOptimizerBatch Size
ScanNetV212000.0010.02CosineAdamW16
S3DIS12000.0060.01OneCycleLRAdamW12
SemanticKITTI800.0020.005CosineAdamW8
Table 3. Semantic segmentation results evaluated on S3DIS [67] (Area 5). The best results are presented in bold, the second-best results are underlined, and “-” means unknown.
Table 3. Semantic segmentation results evaluated on S3DIS [67] (Area 5). The best results are presented in bold, the second-best results are underlined, and “-” means unknown.
MethodmIoU (%)mAcc (%)OA (%)CeilingFloorWallBeamColumWindowDoorTableChairSofaBookcaseBoardClutter
PointNet [9]41.149.0-88.897.369.80.13.946.310.852.658.940.35.926.433.2
PointNet++ [17]57.363.5-91.396.978.70.016.054.931.983.574.667.249.354.245.9
PointCNN [2]57.363.985.992.398.279.40.017.622.862.180.674.466.731.762.156.7
PointWeb [3]60.366.687.092.098.579.40.021.159.734.876.388.346.969.364.952.5
MinkwskiNet [6]65.471.7-91.898.786.20.034.148.962.489.981.674.947.274.458.6
PointConv [12]58.364.785.492.896.377.00.018.247.754.387.972.861.665.933.949.3
KPConv [11]65.470.9-92.697.381.40.016.554.569.590.180.274.666.463.758.1
PointWeb [3]61.968.387.291.598.281.40.023.365.340.075.587.758.567.865.649.7
PointASNL [13]62.668.587.794.398.479.10.026.755.266.283.386.847.668.356.452.1
PCT [15]61.367.7-92.598.480.60.019.361.648.076.585.246.267.767.952.2
PTv1 [16]70.476.5-94.098.586.30.038.063.474.389.182.474.380.276.059.3
PointNeXt [39]70.576.890.694.298.584.40.037.759.374.083.191.677.477.278.860.6
CBL [70]69.475.290.693.998.484.20.037.057.771.991.781.877.875.669.162.9
FastPointTrans. [19]70.377.9-94.298.086.00.253.861.277.381.389.460.172.880.458.9
Our70.775.990.591.798.384.10.028.063.575.581.792.180.380.182.561.2
Table 4. Quantitative results of our proposed method and state-of-the-art LiDAR semantic segmentation methods on the test set of ScanNetv2 [68]. The best results are presented in bold, and the second-best results are underlined. The scores are obtained from the official leaderboard of ScanNetv2 when available, otherwise from the respective paper.
Table 4. Quantitative results of our proposed method and state-of-the-art LiDAR semantic segmentation methods on the test set of ScanNetv2 [68]. The best results are presented in bold, and the second-best results are underlined. The scores are obtained from the official leaderboard of ScanNetv2 when available, otherwise from the respective paper.
MethodmIoU (%)BathtubBedBookshelfCabinetChairCounterCurtainDeskDoorFloorOther FurniturePictureRefrigeratorShower CurtainSinkSofaTableToiletWallWindow
PointNet++ [17]33.958.447.845.825.636.025.024.727.826.167.618.311.721.214.536.434.623.254.852.325.2
PointConv [12]66.670.378.175.165.583.047.176.947.453.795.147.527.963.569.867.575.155.381.680.670.3
KPConv [11]68.484.775.878.464.781.447.377.260.559.493.545.018.158.780.569.078.561.488.281.963.2
PointWeb [3]61.872.966.864.759.776.641.468.052.052.594.643.221.549.359.963.861.757.089.780.660.5
MinkowskiNet [6]73.685.981.883.270.984.052.185.366.064.395.154.428.673.189.367.577.268.387.485.272.7
FPConv [34]63.978.576.071.360.379.839.253.460.352.494.845.725.053.872.359.869.661.487.279.956.7
PointASNL [13]66.678.175.969.964.482.247.577.956.450.495.342.820.358.675.466.175.358.890.281.364.2
BPNet [35]74.990.981.881.175.283.948.584.267.364.495.752.830.577.385.978.881.869.391.685.672.3
RFCR [38]70.288.974.581.367.281.849.381.562.361.094.747.024.959.484.870.577.964.689.282.361.1
CBL [70]70.576.977.580.968.782.043.981.266.159.194.551.517.163.385.672.079.666.888.984.768.9
SegGroup [72]62.781.874.770.160.276.438.562.949.050.893.140.920.156.472.561.869.253.987.379.454.8
ST [20]74.790.180.384.575.784.651.282.569.664.595.657.626.274.486.174.277.070.589.986.073.4
LargeKernel [29]74.091.082.080.674.085.254.582.659.464.395.554.126.372.385.877.576.767.893.384.869.4
REFL-Net [40]72.973.782.376.672.685.246.886.568.463.495.356.529.777.377.477.774.966.690.785.070.8
Retro-FPN [59]74.484.280.076.774.083.654.191.467.262.695.855.227.277.788.669.680.167.494.185.871.7
PTv3 [71]79.494.181.385.178.289.059.791.669.671.397.963.538.479.390.782.179.069.696.790.380.5
Ours74.973.880.984.177.083.053.890.967.867.996.054.533.276.779.474.278.669.691.988.077.0
Table 5. Quantitative results of our proposed method and state-of-the-art LiDAR semantic segmentation methods on the test set of SemanticKITTI [69]. The best results are presented in bold, and the second-best results are underlined. The scores are obtained from the official leaderboard of SemanticKITTI when available, otherwise from the respective paper.
Table 5. Quantitative results of our proposed method and state-of-the-art LiDAR semantic segmentation methods on the test set of SemanticKITTI [69]. The best results are presented in bold, and the second-best results are underlined. The scores are obtained from the official leaderboard of SemanticKITTI when available, otherwise from the respective paper.
MethodmIoU (%)CarBicycleMotorcycleTruckOther-VehiclePersonBicyclistMotorcyclistRoadParkingSidewalkOther-GroundBuildingFenceVegetationTrunkTerrainPoleTraffic-Sign
PointNet [9]14.646.31.30.30.10.80.20.20.061.615.835.71.441.412.931.04.617.62.43.7
PointNet++ [17]20.153.71.90.20.90.20.91.00.072.018.741.85.662.316.946.513.830.06.08.9
KPConv [11]58.896.032.042.533.444.361.561.611.888.861.372.731.695.064.284.869.269.156.447.4
RangeNet++ [1]52.291.425.734.425.723.038.338.84.891.865.075.227.887.458.680.555.164.647.955.9
RandLANet [10]50.394.019.821.442.738.747.548.84.690.456.967.915.581.149.778.360.359.044.238.1
SqueezeSegV3 [30]55.992.538.736.529.633.045.646.220.191.763.474.826.489.059.482.058.765.449.658.9
SPVNAS [5]66.497.351.550.859.858.865.765.243.790.267.675.216.991.365.986.173.471.064.266.9
PointASNL [13]46.887.957.625.139.029.234.257.60.087.424.374.31.883.143.984.152.270.657.836.9
PolarNet [33]54.393.840.330.122.928.543.240.25.690.861.774.421.790.061.384.065.567.851.857.5
Cylinder3D [27]67.897.167.664.050.858.673.967.936.091.465.175.532.391.066.585.471.868.562.665.6
JS3C-Net [28]66.095.859.352.954.346.069.565.439.988.961.972.131.992.570.884.569.867.960.768.7
WaffleIron [32]67.396.562.364.155.248.770.477.829.690.569.575.924.691.868.185.470.869.662.065.2
RangeViT [73]64.095.455.843.529.842.163.958.238.193.170.280.032.592.069.085.370.671.260.864.7
DFAMNet [31]69.096.754.580.895.568.579.791.70.294.350.681.84.191.865.889.473.476.863.151.9
Ours68.096.665.459.856.253.674.669.337.389.667.774.223.491.868.985.973.971.465.567.4
Table 6. The results (%) of four intuitive low-level geometric relations g i j on the ScanNetv2. The best results are presented in bold.
Table 6. The results (%) of four intuitive low-level geometric relations g i j on the ScanNetv2. The best results are presented in bold.
IDLow-Level Relation  g ij ChannelsmIoU (%)
(1) ( Ed ) 174.3
(2) ( p i , p j ) 674.5
(3) ( Ed , p i p j ) 475.0
(4) ( Ed , p i p j , p i , p j ) 1075.5
Table 7. The performance and parameters of different layers of the geometric relation function θ on the ScanNetv2. The best results are presented in bold.
Table 7. The performance and parameters of different layers of the geometric relation function θ on the ScanNetv2. The best results are presented in bold.
MLP Layer12345
Layer Params (M)2.064.116.178.2210.27
mIoU (%)74.9275.2175.5275.5175.31
Table 8. Ablation study of Contextual Relative Semantic Encoding on the ScanNetv2. The best results are presented in bold.
Table 8. Ablation study of Contextual Relative Semantic Encoding on the ScanNetv2. The best results are presented in bold.
MethodmIoU (%)
without bias 74.1
Δ p r i only74.3
Δ f r i only74.9
Δ r i 75.5
ST [20]74.3
PTv2 [23]75.4
Table 9. The results(%) of different designs on adaptive interaction processing on the ScanNetv2 and S3DIS. The best results are presented in bold.
Table 9. The results(%) of different designs on adaptive interaction processing on the ScanNetv2 and S3DIS. The best results are presented in bold.
FunctionMethodScanNetv2S3DIS
φ without normalization73.268.4
Softmax ( ) 75.370.3
Sigmoid ( ) 75.570.7
ψ add + Linear 74.268.6
cat + Linear 74.769.9
Adaptive Interaction75.570.7
Table 10. Ablation study of AIFormer architecture design on the ScanNetv2. The best results are presented in bold.
Table 10. Ablation study of AIFormer architecture design on the ScanNetv2. The best results are presented in bold.
IDPointEmb.DataAug.LRAGCAAImIoU (%)
(1) 67.8
(2) 71.8
(3) 72.1
(4) 73.8
(5) 74.1
(6) 73.9
(7) 74.5
(8)75.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chu, X.; Zhao, S.; Dai, H. AIFormer: Adaptive Interaction Transformer for 3D Point Cloud Understanding. Remote Sens. 2024, 16, 4103. https://doi.org/10.3390/rs16214103

AMA Style

Chu X, Zhao S, Dai H. AIFormer: Adaptive Interaction Transformer for 3D Point Cloud Understanding. Remote Sensing. 2024; 16(21):4103. https://doi.org/10.3390/rs16214103

Chicago/Turabian Style

Chu, Xutao, Shengjie Zhao, and Hongwei Dai. 2024. "AIFormer: Adaptive Interaction Transformer for 3D Point Cloud Understanding" Remote Sensing 16, no. 21: 4103. https://doi.org/10.3390/rs16214103

APA Style

Chu, X., Zhao, S., & Dai, H. (2024). AIFormer: Adaptive Interaction Transformer for 3D Point Cloud Understanding. Remote Sensing, 16(21), 4103. https://doi.org/10.3390/rs16214103

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop