Next Article in Journal
Error-Resistant Movement Detection Algorithm for the Elderly with Smart Mirror
Next Article in Special Issue
Special Issue on Advanced Wireless Sensor Networks for Emerging Applications
Previous Article in Journal
DECIDE: A Deterministic Mixed Quantum-Classical Dynamics Approach
Previous Article in Special Issue
Management of Distributed Renewable Energy Resources with the Help of a Wireless Sensor Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Dense Features for Point Cloud Registration Using a Graph Attention Network

Cho Chun Shik Graduate School of Mobility, Korea Advanced Institute of Science and Technology (KAIST), 291 Daehak-ro, Yuseong-gu, Daejeon 34141, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(14), 7023; https://doi.org/10.3390/app12147023
Submission received: 6 June 2022 / Revised: 2 July 2022 / Accepted: 10 July 2022 / Published: 12 July 2022
(This article belongs to the Special Issue Advanced Wireless Sensor Networks for Emerging Applications)

Abstract

:
Point cloud registration is a fundamental task in many applications such as localization, mapping, tracking, and reconstruction. Successful registration relies on extracting robust and discriminative geometric features. Though existing learning-based methods require high computing capacity for processing a large number of raw points at the same time, computational capacity limitation is not an issue thanks to powerful parallel computing process using GPU. In this paper, we introduce a framework that efficiently and economically extracts dense features using a graph attention network for point cloud matching and registration (DFGAT). The detector of the DFGAT is responsible for finding highly reliable key points in large raw data sets. The descriptor of the DFGAT takes these keypoints combined with their neighbors to extract invariant density features in preparation for the matching. The graph attention network (GAT) uses the attention mechanism that enriches the relationships between point clouds. Finally, we consider this as an optimal transport problem and use the Sinkhorn algorithm to find positive and negative matches. We perform thorough tests on the KITTI dataset and evaluate the effectiveness of this approach. The results show that this method with the efficiently compact keypoint selection and description can achieve the best performance matching metrics and reach the highest success ratio of 99.88% registration in comparison with other state-of-the-art approaches.

1. Introduction

Efficient keypoint matching is required for different applications such as 3D object identification, simultaneous localization and mapping (SLAM), sparse depth odometry, 3D shape retrieval, and range image/point cloud registration [1,2,3,4]. All of these techniques include keypoint recognition and their matching by employing element descriptors to find genuine keypoint correspondences. To find salient/interesting points on 3D point clouds, a variety of 3D keypoint detectors are available in the literature [5,6]. Following the detection of keypoints on 3D point clouds, feature descriptors are used to match them. Feature descriptors are multi-dimensional vectors that encode information in the vicinity of a keypoint. In most cases, the feature descriptors are matched by computing Euclidean distances in their high-dimensional vector spaces, which is a computationally expensive procedure. Furthermore, the memory footprint and computing requirements for feature descriptor matching grow with the size of the feature descriptor.
Different handheld depth sensors such as Microsoft Kinect [7], Asus Xtion Pro Live [8], Intel RealSense camera [9], and Google Tango [10] have made it possible to access 3D data. Since then, there has been an increase in consumer applications that use 3D sensors and interpret dense depth data on mobile devices for vision and robotics. Project Tango, a Google ATAP project that can give online 3D posture estimates and detailed depth data from a mobile device, has been in the spotlight. The accessibility of these transportable 3D data acquisition devices necessitates the development of applications with a low memory footprint and low computing power requirements.
Wireless sensor networks (WSNs) are being applied not only in the traditional way with a limited number, but also allow connecting a large number of high-quality sensor network nodes for practical problems as shown in Figure 1. Technical issues of the WSNs are broad, including spectrum sensing [11], broadband data transmission [12], and recharging rechargeable WSNs [13,14]. Due to their rapid development in recent years, deep learning techniques allow us to improve the performance of sensor networks used in communication protocols, intelligent transportation systems, and smart charging schemes, especially for vehicles [15,16]. To date, light detection and ranging (LiDAR) is considered one of the most outstanding sensor systems helping autonomous vehicles with perception of their surroundings. LiDAR scans the objects creating sets of data called point clouds, which are continuously aligned with each other to create 3D models through point cloud registration.
3D point cloud registration plays a fundamental role in robotics applications. Many systems use it for applications of autonomous driving, such as LiDAR-based SLAM [17,18], or archaeological inspection via 3D model reconstruction. Given two partially overlapped sets of 3D points, the goal of registration is finding the best Euclidean relative transformation to align these fragments according to a common frame, thus obtaining a successful point cloud registration. Existing solutions can be categorized into two main methods. Methods that iteratively minimize the distance between nearest-neighboring points with good initialization, e.g., iterative closest point (ICP) [19] and its variants [20,21,22] are popular approaches. Other methods randomly sample corresponding points through robust descriptor matching, e.g., using random sample consensus (RANSAC) [23]. Unfortunately, the key challenge of finding 3D-3D data correspondences is non-trivial.
This problem is particularly challenging because (1) different point clouds can be captured with different angles, and (2) the raw 3D scans are noisy and have various densities; thus, descriptor-based matching often increases robustness across different scenarios. To compute compact 3D descriptors, there are one-stage [24,25,26,27,28] or two-stage [29,30,31] methods. The former directly encodes the local geometric information of raw points (e.g., collection of coordinates). On the other hand, the latter firstly determines a local reference frame (LRF) from the points to transform the whole patch into suitable representations; they then encode the information of these new representations into a generalized descriptor.
Recent descriptors based on deep learning [25,26,32] have outperformed their handcrafted counterparts [29,33] because of the availability of large-scale labeled datasets. The methods based on one-stage learning can achieve rotation invariance by encoding the local geometric information with a set of geometric relationships, and then by learning descriptors via a deep network to achieve generalization ability and permutation invariance. In two-step methods, the LRFs can be computed and transformed into a voxel grid, where each voxel contains the information of inside points. Finally, descriptors can be encoded from the voxelization by using ConvNet [34].
The SpinNet [35] is a new neural architecture which projects the point cloud into a cylindrical space; then the descriptor is invariably extracted rotationally by a special convolutional neural layer. Inspired by the success of GAT for computer vision tasks, Shi et al. [4] proposed a multiplex dynamic graph attention network matching method (MDGAT-matcher). It uses a novel and flexible graph network architecture on the point cloud and achieves the enriched feature representation by recovering local information. Although MDGAT-matcher finds high-quality data associations, its architecture provides space for improvements. Without employing the rotation-invariant description method, it limits the robustness and distinctiveness of the learned descriptor. In this study, we aim to tackle these challenges in a joint learning framework that is able to not only accurately detect keypoints from 3D point clouds, but also derive compact and representative descriptors for matching.
To this end, we find inspiration in the MDGAT-matcher [4] for a GAT-based two-stage descriptor for point cloud registration. However, the extension of our work is non-trivial. A reliable network for the extraction of robustness features against noise and occlusion is needed instead of using hand-crafted counterparts. In this work, we adopt joint learning of dense detection and description of 3d local (D3Feat) [28] to extract dense features from the kernel point convolution (KPConv) network [36] and achieve density invariance on irregular 3D point clouds. Based on these density-invariant saliency scores, we detect keypoints in their local neighborhood. Finally, these dense features are further fed into the attention network, DFGAT, which fully exploits the local information and generates a generalizable and distinctive feature vector.
Overall, the proposed DFGAT has the following key features:
  • It is the learning pipeline using joint keypoints descriptors and detectors, is density invariant, and generalized across all point clouds.
  • It can build the global spatial relationship and enrich characteristics for matching features.
  • It helps to save time and computing resources but still achieves the highest matching scales and success ratio in the registration experiment.

2. Related Work

2.1. Keypoint Detector

Hand-crafted 3D keypoint detectors are still commonly used even as the use of deep learning to generate keypoint descriptors is becoming more frequent [32,37]. A comprehensive review and evaluation of such common approaches can be found in [38]. The work in [33] detects outstanding keypoints with a large variation in the principal direction. Local surface patches [39] and shape indexes [40] define a keypoint based on the global extremum of its principal curvature in a neighborhood. Laplace–Beltrami Scale-space (LBSS) [41] computes the saliency by applying a Laplace–Beltrami operator on increasing supports for each point. Typically, the local spatial characteristics greatly influence the design of traditional methods applied to point clouds. Hence, the performances when these detectors are applied to the real world in 3D domains may be impaired by influences such as noise, varying densities, and occlusion. To improve detection ability, a recent study used an unsupervised stable interest point (USIP) [42], which introduces an unsupervised method to learn points that are likely to repeat and localize correctly. Additionally, this detector is more robust because of its learning ability from data disturbances. Nonetheless, its performance is significantly degraded with a limited number of keypoints. This motivates the use of joint detection and a description pipeline to identify stable 3D local keypoints for matching.

2.2. Local Descriptors

Handcrafted descriptors consist of two groups: one-stage methods and two-stage methods. Typically, one-stage (Non-LRF-based) descriptors compute descriptors from point representations that are defined between pairs of points using 3D point coordinates only [43] and/or surface normals [29]. Point-pair features (PPF) [44] and fast point feature histograms (FPFH) [24] build an oriented histogram by using pairwise combinations of surface normals. They provide accurate feature matching with reasonable computational efficiency. The lack of essential geometrical information for the local frame is the main disadvantage of these methods. The two-stage (based on the LRF) descriptors, such as intrinsic shape signatures (ISS) descriptor [33], signature of histograms of orientations (SHOT) descriptor [30] and rotational projection statistics descriptor [45], are not only able to characterize the geometric patterns of the local support region but also effectively exploit the 3D spatial attributes. However, the LRF-based methods inherently introduce rotation errors, sacrificing the feature robustness.

2.3. Graph Attention Network

Recently, the development of the GAT [46] contributes to simplifying CNN into graph representations. Even though there are often common traits with on-set deep learning, the GAT can aggregate more powerful features along the edges and nodes. Previously, Zanfir et al. [47] presented an application of deep neural networks for finding the similarity matrix in graph matching. Wang et al. [48] introduced another GAT-based network for the image domain. Its framework consists of a feature extractor in the image, a graph embedding component, an affinity metric function, and a permutation prediction. Typically, the graph matching problem is mathematically formulated as a quadratic assignment problem, which consists of finding the assignment that maximizes an objective function of local compatibilities (a linear term) and structural compatibilities (a quadratic term) [47,49]. Many algorithms have been proposed to find a locally optimal solution for the above matching problem. One popular method is to use an approximation of the objective function. By using the Hungarian method proposed by Kuhn [50], the problem can be solved for minimizing total cost (or) maximizing the profit. As an approximate version of the Hungarian method, a Sinkhorn network [51] is developed for linear assignment learning with a predefined assignment cost which is designated to enforce doubly-stochastic regulation on any non-negative square matrix. However, very few studies have focused on using GAT to solve the point cloud registration problem. The proposed methodology using GAT improves performance but still saves the computational cost of the registration point cloud.
In the MDGAT-matcher [4], a dynamic GAT was introduced to embed the geometry feature from the FPFH descriptor to create more robust feature for the matching process. Because of the limitations of handcrafted descriptors, their results are sensitive and rotation-variant to rigid transformation in Euclidian space. Recently, Xuyang et al. [28] proposed D3Feat which is invariant in density, representative, and has outstanding adaptive capability across untrained scenarios. However, it either requires the calculation of the point density or relies on random points selection to achieve rotation robustness. This is often unfeasible and is insufficient when there are limited computing resources. However, the proposed method not only selectively exploits the rich spatial and contextual information of the D3Feat, but also takes advantage of the flexible GAT to find high-quality point-wise correspondences in 3D point cloud feature matching.
The main focus of this study is to use a learning-based approach to overcome the limitations of manual methods. For indoor datasets, early works such as 3dMatch [25] use a 3D convolutional network to learn geometric features by training on positive and negative pairs. The fully convolutional geometric features approach (FCGF) [27] used a fully convolutional network for descriptors and achieved rotation invariance. The SpinNet approach [35] used a spatial point transformer to project the input point to a cylindrical space followed by a series of powerful neural networks. It learns rotation-invariant, compact, and highly descriptive local representations. Similar to the FCGF, the D3Feat approach uses a fully convolutional network [28], but unlike FCGC, the D3Feat extracts global descriptors using dense deep features through a KPConv backbone [36]. Yew and Lee [52] used a weakly-supervised network to extract features in outdoor scenes. However, these descriptors are the result of random or full sampling. To overcome the above drawbacks, an effective combination of a keypoint and a descriptor is proposed to provide better registration performance.

3. Proposed Methodology

The overview of our DFGAT model is illustrated in Figure 2 and it consists of three main parts: the keypoint detector in Section 3.1, the descriptor encoder in Section 3.2, and the dynamic GAT in Section 3.3.

3.1. Keypoint Encoder

Instead of randomly choosing the local points as the input, we use a detector inspired by the D3Feat detector [28] to find the keypoints with local-max scores in the neighborhood within a radius. We denote a query point cloud from the training dataset as Q 3 and the dense feature matrix F d , where d is the dimension of feature. The condition of a point p i Q to be a keypoint of point cloud Q is given by:
{ k = arg max t F i : k t i = arg max j N p i Q F j : k k
where F i : k is the k -th channels ( k = 1 ,   ,   d ) of the i -th descriptor corresponding with point p i Q detected by the D3Feat detector, and N p i Q is the set of radius neighborhood points of p i Q . The maximum channel is chosen before comparing with the same channel of local neighborhood points. Xuyang et al. [28] used two local scores: Channel max score and Density-invariant saliency score. More details about these scores can be found in their work [28]. Both of these are used to build the final score map for each input of the 3D point cloud. We select the desired number of keypoints with the highest scores s i Q . Inspired by the SuperGlue [53], we encode each D3Feat keypoint position with its detection score into a higher dimension feature by using a multi-layer perceptron (MLP). The keypoint encoder provides a better signature position and keeps spatial information around a keypoint in a form as follows:
p o s i Q = M L P p o s ( D 3 F e a t ( p i Q ,   s i Q ) )
where p o s i Q is the encoded information about position and score of keypoint p i Q .

3.2. Descriptor Encoder

Inspired by the KPConv [36] in 2D convolution, Xuyang et al. [28] performed feature extraction on irregular point clouds. They added a new normalization term to ensure that the convolution formulation is robust against spatial density. Given the set of features F i : n , the convolution by kernel g at point p i Q is given by:
( F i : n g ) = 1 | N p i Q | p j N p i Q g ( p j p i Q ) f j
where p j is a supporting point, and f j is the corresponding feature.
Inspired by the MDGAT-matcher [53], we combine the D3Feat of the i -th feature F i : with the output of the keypoint encoder to fully exploit spatial clues, and turn them into a new descriptor d i Q as follows:
d i Q = M L P d e s c ( D 3 F e a t ( F i : Q ,   p o s i Q ) )

3.3. Dynamic Graph Attention Network

After the feature encoder, we apply a dynamic graph attention network, inspired by the GAT [54], to exploit relationship information for matching. In contrast to the traditional static architecture, the MDGAT-matcher [4] showed that a dynamic graph can deal with variable sizes of inputs and focus on the parts that matter most to make decisions efficiently. Furthermore, this technique operates a specific aggregator over a fixed-size neighborhood and yields impressive performance across a large-scale dataset.

3.3.1. Graph Attention Layer

The GAT starts with a high-dimensional state for each node and computes at each layer an updated representation by simultaneously aggregating messages across all given attentions for all nodes. We use a single layer throughout the architecture and the following is a specific description of this graph attention layer. The input to our layer is a set of embeddings from the previous step which are used as node features, where the number of nodes is the number of keypoints, and the number of features in each node is the dimension of the descriptor vector according to Equation (4). To transform the input features into higher-level features, at least one learnable linear transformation is required. Moreover, our graph includes two types of attention: self-attentions and cross-attentions. Self-attention is a connection mechanism between keypoints in the same point cloud that utilizes surrounding information to increase their representation. On the other hand, cross-attentions connect keypoints between two-point clouds to inspect the valid connection of respective points. The layer produces a new set of node features as its output.

3.3.2. Dynamic Graph Update

We inject the graph structure into the mechanism by performing masked attention. Within the same layer, for a query node d i Q in query point cloud Q and all neighborhoods source nodes d j S in source point cloud S, the network computes three attention coefficients: query q i , key k j , and value v j , which are used for later graph updating and attentional propagation. To that end, a shared linear transformation, parametrized by weight matrices W 1 , 2 , 3 and biases b 1 , 2 , 3 , is applied to every query node, as follows:
q i = W 1 d i Q + b 1 [ k j v j ] = [ W 2 W 3 ] d j S + [ b 2 b 3 ]
We normalize the coefficients between different nodes by using the softmax function for easy comparison, i.e., α i j = softmax j ( q i k j ) . This weight shows how important the node d j S is to the node d i Q .
To stabilize the learning process, we extend our mechanism to employ multi-head attention, similarly to Vaswani et al. [55]. Specifically, K independent attention mechanisms execute the transformation, and then their features are concatenated, resulting in the output feature representation. To avoid the bad influence of many dissimilar neighbor source nodes, we only extract certain K edges for each query node thanks to dynamic graph attention weight α i j . A decay strategy is adopted to reduce the links when the layers pay more attention to a specific area of interest.

3.3.3. Attentional Propagation

Once the graph updating is completed, attention propagation is performed on each node in the new graph via the message passing process as follows:
d i Q = d i Q + message
When the aggregation is completed to the final layer, the current descriptors are passed through an additional linear layer with weight matrix W 4 and bias b 4 :
d i Q = W 4 d i Q + b 4

3.4. Gap Loss

To optimize feature embeddings, different loss formulas, such as contrastive [27] and triplet [32] are used. Similar works relevant to data metric learning consider effective negative exploitation [27]. The model is trained to pull similar features together while pushing dissimilar features apart. Considering a set of triplets of input histograms p i Q , it includes an anchor point in Q , a true connection in S that is known to be similar to the anchor, and a false connection in S that is known to be dissimilar, indexed by i , p , n , respectively. We select anchor points and a mining group per scene. A positive matching p is the matching with minimum Euclidean distance and shorter than the threshold value, whereas the negative matching n is any of non-positive matching for the keypoint in the anchor point cloud. In addition, we use the pairwise loss for the mined quadruplet ( p i , n i , p j , n j ) and form a gap loss. Given the number of keypoints in these two clouds M and N , the ground-truth correspondences matrix M { 0 , 1 } ( M + 1 ) × ( N + 1 ) are generated by projecting the keypoints to the other point cloud. Specifically, for each keypoint in the point cloud, it is either connected with a matched keypoint or with the non-matched keypoint (the so-called “dustbin”) of the other point cloud. Consequently, the matching scores between any keypoint and the dustbin are filled in the last row and column of correspondences matrix M . For a generalized graph matching, we create the predicted soft score assignment matrix P ¯ [ 0 , 1 ] ( M + 1 ) × ( N + 1 ) using the Sinkhorn algorithm [53]. The loss is then calculated as:
L o s s = i = 1 M log ( n = 1 N + 1 [ ( z i ) + + 1 ] ) + j = 1 N log ( m = 1 M + 1 [ ( z j ) + + 1 ] )
where z i = log r i + log   n i + η , z j =   log c j + log n j + η , and ( z i ) + = max ( z i , 0 ) . Specifically, r i = ( i , n ) P P ¯ i n M i n represents the true connection assignment value for keypoint p i in point cloud Q and P is a set of all positive pairs in extracted features in a minibatch. Meanwhile, n i = ( i , n ) N P ¯ i n M i n refers to the false connection assignment value for p i and N is a negative subset of features in a minibatch that will be used for negative mining. The gap loss aims to expand the margin between the positive match with all the other negative matches at least a margin away η . The second terms in the loss function for keypoint p j in point cloud S are computed in a similar manner.

4. Implemented Details

4.1. Dataset

We perform several evaluations of our framework with the KITTI odometry dataset [56]. This is a large point cloud dataset obtained by the Velodyne-64 device while driving in diverse situations. This dataset includes 11 sequences of real driving situations. Inspired by the preprocessing data method in the MDGAT [4], we use sequences 0 to 9 for the training set, except sequence 8 for the validation set, and 10 for the test set. We follow D3Feat method [28] to utilize only scan couples with more than 30% similarity.

4.2. Parameter

We use the D3Feat model [28] pre-trained with outdoor LiDAR [56] and the dataset prepared by 3DFeat-Net [52] for keypoints selection and description. Specifically, we use 23,190 point clouds in the KITTI dataset that have been preprocessed to 256 desired keypoints/point cloud to evaluate the model. For each selective keypoint, we extract its descriptor by using the D3Feat pre-trained model. The output size of the keypoint encoder and descriptor encoder are 128. For the architecture of the GAT, we use L = 9 layers of alternating multi-head self-attention and cross-attention with four attention heads and perform T = 20 Sinkhorn iterations [51]. For the self-attention, we use fully-connected graph for the first five layers before changing the number of k nearest neighbors in deeper layers. More specifically, in the last four layers, we set k 6 s e l f = k 7 s e l f = 128 , k 8 s e l f = k 9 s e l f = 64 . In terms of cross-attention, we keep all fully connected layers, where k 1 9 c r o s s = 256 . The model is implemented in the Pytorch framework and NVIDIA GTX 3090.

4.3. Training

Training is performed on the KITTI odometry dataset [56]. The ground-truth results obtained by GPS are also included. To reduce the noise in each pair of scans, we apply ICP to refine the alignment. If registration fails or the number of matching voxels is less than 1k, we filtered that pair out of the preprocessed dataset. In addition, we only select pairs with a minimum distance of 10 m away from each other for training and testing. By doing this, we yield more samples, which are 17,920 for training, 4070 for validation, and 1200 for testing. The model is optimized using the Adam method [57] with a batch size of 64. The learning rate is set at 0.001 during the first 15 epochs and decreases to 0.0001 until convergence.

5. Experiments

We validate the DFGAT on the outdoor dataset. Additionally, we compare our evaluations to handcrafted descriptors. Our approach can better distinguish the features in registration on the KITTI odometry dataset.

5.1. Evaluation Metrics

To measure the matching performance, we compute the matching Precision, Accuracy, Recall and F1-score from the ground-truth correspondences. According to registration ability, we use the indicators proposed in [58,59]: relative rotation error (RRE), relative translation error (RTE), and success rate.
RTE and RRE are defined as:
R T E = | T ^ T * | R R E = a r c c o s T r ( R ^ T R * ) 1 2
where T * and R * denote the ground-truth translation and rotation, respectively. Meanwhile, T ^ is the estimated translation and R ^ is the estimated rotation matrix after registration.

5.2. Performance

The matching is evaluated as follows. For fair comparisons, we refer to the results of the group of methods including USIP detector [42], MDGAT descriptor and SuperGlue matcher. Specifically, we re-check SuperGlue matcher with several threshold assignments algorithm and threshold-free. After that, we run MDGAT-matcher instead of the combination between MDGAT descriptor and SuperGlue matcher. As show in Table 1 and Figure 3, the DFGAT achieves the best performance in matching ability with about 80% for Precision, 89% for Accuracy, 76% for Recall and 0.783 for the F1-score. Although the USIP can detect keypoints by using an unsupervised learning network, it struggles to predict keypoints with a limited number of data and density-invariant requirements. The further reason is that the USIP method is not integrated with the descriptions framework which limits the full exploitation of the geometric characteristics of the point cloud.
We further evaluate the registration ability of the DFGAT as shown in Table 2. We compare the metric scores achieved against several strong baselines: the hand-crafted methods RANSAC [23], FPFH [24], SHOT [30], and USIP. For the learning-based method, theMDGAT + SuperGlue [4] is selected. As shown in Table 2, our network achieves the lowest registration failure rate of 0.117% among all the baselines and has a comparable inlier ratio to that of the MDGAT-matcher. The reason is that our descriptor can create better learnable features, and after the Sinkhorn algorithm assignment, the scores matrices have many more positive pairs. For fair comparison with MDGAT, our approach uses the number of desired keypoints 256 and has large improvements in RTE, RRE, and the success ratio. In addition, as shown in Table 3, our model can achieve the lowest score (0.24°) on the RRE scale and the highest percentage success (99.88%), while the D3Feat still remains the best in RTE with 6.90 cm. It is noteworthy that our approach uses only 256 keypoints for each point cloud, whereas in other methods, they originally used no less than 1024 keypoints. In terms of training time, our model takes only 40 epochs in comparison with 1000 epochs for the MDGAT-matcher because we use better strategy in the keypoints selection and description.

5.3. Ablation Study

We perform ablation studies on three key components of our framework: the descriptor encoder, the GAT, and the loss function.

5.3.1. Ablation of Descriptor Encoder

In this experiment, we compare four different combinations of descriptor encoders. The first option is the PointNet [60] encoder, working on separate points sampled by the nearest neighbor search. The second design takes the output of descriptor decoder and performs max pooling to extract global features. We then use an MLP to embed this information into a global descriptor as the input of GAT. We refer to this descriptor as Global Desc. The third design is the one proposed by Shi et al. [4] named MDGAT. This design is inspired by the USIP detector [42] and FPFH descriptor [24] to enrich the feature representation before being fed into the GAT.
We train different networks with the above encoders. The results in Table 4 demonstrate that our descriptor with most generalized structure can achieve the best performance. The result is that the density-invariant feature descriptor inspired by D3Feat allows the network to cope with the inherent density of point cloud variations. Furthermore, the dynamic GAT with two types of attention enhances the performance of a powerful and flexible matching mechanism. For Global Desc., we can see that concatenating global information too early in the descriptor encoding process does not improve the matching efficiency but even reduces the discriminant ability. The same effect applies to PointNet, which extracts the overall interface using maximum aggregation and loses local information.

5.3.2. Ablation of GAT

In this study, we evaluate several variations of a dynamic GAT with different decay strategies as the network goes deeper into the last layers. We introduce the DFGAT variant 1 using k 1 7 s e l f = 128 for the first 7 layers and k 8 , 9 s e l f = 64 for the last 2 layers in self-attention. For cross-attention, we maintain full connection for all 9 layers with k 1 9 c r o s s = 256 . We refer to the other design using the first 5 fully connected layers, while 6,7 layers have k 6 , 7 s e l f = k 6 , 7 c r o s s = 128 , and the last 2 layers k 8 , 9 s e l f = k 8 , 9 c r o s s = 64 , as for DFGAT variant 2.
As shown in Table 5, compared with two above variant versions, our gradual decay strategy performs best. The reason could be that first examining the large observations and then gradually focusing on specific interests becomes more natural. Specifically, self-attention starts from the full connected layer and k gradually decreases in the last layers. In addition, the cross-attention will always maintain the full connected layer to ensure the generalization information between connections.

5.3.3. Ablation of Loss Function

In this ablation study, we evaluate the gap loss against the SuperGlue loss [53] and the triplet loss [32] using the hardest negative match. We retrain our matcher on three loss functions and show the results in Table 6. In comparison with the SuperGlue loss, the gap loss is more suitable for the matching highly density point clouds problem. In addition, gap loss outperforms triplet loss due to its ability to use all the negative matches more efficiently to improve generalization.

6. Discussion

Robotics and autonomous driving both face a basic challenge with point cloud registration. It is used in medical imaging, panorama stitching, object localization and recognition, motion estimates, 3-D reconstruction, and object recognition. Common methods for registration with unknown or uncertain correspondences depend on the existence of an initial hypothesis for the unknown transformation. These methods could stop working at any time and return estimates that are arbitrarily far from the transformation of the ground truth. In our proposed study, the keypoint detector, feature descriptor, and the dynamic graph attention network are the three components of our proposed DFGAT. The detector uses deep learning to achieve repeatability, specificity, and computing efficiency in its capacity to identify key locations. The learned descriptor can preserve intricate local geometric patterns and is rotation invariant. The graph architecture focuses on useful information with self and cross-attention. To demonstrate the efficiency and uniqueness of our network, experiments using the KITTI odometry are carried out. Our methodology surpasses previous registration methods with a success ratio of almost 99.88 % and achieves the best results in evaluation matching metrics when compared with existing methods.

7. Conclusions

In this study, we present a GAT-based framework for point cloud registration. Our proposed DFGAT includes three components: a keypoint detector, a feature descriptor, and a dynamic graph attention network. The detector is deep learning-based so that it can detect key points with repeatability, particularity, and computational efficiency. The learned descriptor is rotation invariant, descriptive, and able to preserve complex local geometric patterns. The graph architecture with self- and cross-attention focuses on the useful information. Experiments with the KITTI odometry are conducted to show the effectiveness and distinctiveness of our network in comparison with the state-of-the-art approaches. Among existing matching methods, including both learned and traditional ones, our approach achieves the best results in evaluation matching metrics and outperforms other registration approaches with a success ratio of 99.88% and the lowest failure rate of 0.117%. The results suggest that our approach outperforms similar existing methods and achieves substantial improvements. An interesting avenue for future work is to integrate detector, descriptor, and graph networks into an end-to-end architecture.

Author Contributions

Conceptualization, Q.-V.L.-D.; data curation, Q.-V.L.-D. and H.J.; formal analysis, Q.-V.L.-D., S.H.N. and H.J.; investigation, Q.-V.L.-D.; methodology, Q.-V.L.-D. and S.H.N.; software, Q.-V.L.-D.; supervision, Q.-V.L.-D. and S.H.N.; visualization, Q.-V.L.-D. and S.H.N.; writing—original draft, Q.-V.L.-D.; writing—review & editing, Q.-V.L.-D., S.H.N. and H.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Korean government (MSIT) (No. 2020-0-00440).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work was supported by an Institute for Information Communications Technology Promotion (IITP) grant funded by the Korean government (MSIT) (No. 2020-0-00440, Development of Artificial Intelligence Technology that continuously improves itself as the situation changes in the real world).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

GATGraph attention network
DFGATDense features graph attention network
SLAMSimultaneous localization and mapping
WSNsWireless sensor networks
LiDARLight detection and ranging
ICPIterative closest point
RANSACRandom sample consensus
LRFLocal reference frame
D3FeatDense detection and description of 3d local
KPConvKernel point convolution
LBSSLaplace-Beltrami Scale-space
USIPUnsupervised stable interest point
PPFPoint-pair features
FPFHFast point feature histograms
ISSIntrinsic shape signatures
SHOTSignature of histograms of orientations
FCGFFully convolutional geometric features approach
MLPMulti-layer perceptron

References

  1. Aldoma, A.; Marton, Z.C.; Tombari, F.; Wohlkinger, W.; Potthast, C.; Zeisl, B.; Rusu, R.B.; Gedikli, S.; Vincze, M. Tutorial: Point cloud library: Three-dimensional object recognition and 6 dof pose estimation. IEEE Robot. Autom. Mag. 2012, 19, 80–91. [Google Scholar] [CrossRef]
  2. Guo, Y.; Bennamoun, M.; Sohel, F.; Lu, M.; Wan, J. 3D object recognition in cluttered scenes with local surface features: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2270–2287. [Google Scholar] [CrossRef] [PubMed]
  3. Endres, F.; Hess, J.; Engelhard, N.; Sturm, J.; Cremers, D.; Burgard, W. An evaluation of the RGB-D SLAM system. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation, Saint Paul, MN, USA, 14–18 May 2012; pp. 1691–1696. [Google Scholar]
  4. Shi, C.; Chen, X.; Huang, K.; Xiao, J.; Lu, H.; Stachniss, C. Keypoint matching for point cloud registration using multiplex dynamic graph attention networks. IEEE Robot. Autom. Lett. 2021, 6, 8221–8228. [Google Scholar] [CrossRef]
  5. Prakhya, S.M.; Liu, B.; Lin, W. Detecting keypoint sets on 3D point clouds via Histogram of Normal Orientations. Pattern Recognit. Lett. 2016, 83, 42–48. [Google Scholar] [CrossRef]
  6. Rodolà, E.; Albarelli, A.; Cremers, D.; Torsello, A. A simple and effective relevance-based point sampling for 3D shapes. Pattern Recognit. Lett. 2015, 59, 41–47. [Google Scholar] [CrossRef] [Green Version]
  7. Zhang, Z. Microsoft kinect sensor and its effect. IEEE Multimed. 2012, 19, 4–10. [Google Scholar] [CrossRef] [Green Version]
  8. Kundu, A.S.; Mazumder, O.; Dhar, A.; Bhaumik, S. Occupancy grid map generation using 360° scanning xtion pro live for indoor mobile robot navigation. In Proceedings of the 2016 IEEE First International Conference on Control, Measurement and Instrumentation (CMI), Kolkata, India, 8–10 January 2016; pp. 464–468. [Google Scholar]
  9. Keselman, L.; Iselin Woodfill, J.; Grunnet-Jepsen, A.; Bhowmik, A. Intel realsense stereoscopic depth cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1–10. [Google Scholar]
  10. Hyyppä, J.; Virtanen, J.P.; Jaakkola, A.; Yu, X.; Hyyppä, H.; Liang, X. Feasibility of Google Tango and Kinect for crowdsourcing forestry information. Forests 2017, 9, 6. [Google Scholar] [CrossRef] [Green Version]
  11. Hong, E.; Kim, K.; Har, D. Spectrum sensing by parallel pairs of cross-correlators and comb filters for OFDM systems with pilot tones. IEEE Sens. J. 2012, 12, 2380–2383. [Google Scholar] [CrossRef]
  12. Kim, H.; Hong, E.; Ahn, C.; Har, D. A pilot symbol pattern enabling data recovery without side information in PTS-based OFDM systems. IEEE Trans. Broadcasting 2011, 57, 307–312. [Google Scholar] [CrossRef]
  13. Moraes, C.; Myung, S.; Lee, S.; Har, D. Distributed sensor nodes charged by mobile charger with directional antenna and by energy trading for balancing. Sensors 2017, 17, 122. [Google Scholar] [CrossRef] [Green Version]
  14. Moraes, C.; Har, D. Charging distributed sensor nodes exploiting clustering and energy trading. IEEE Sens. J. 2016, 17, 546–555. [Google Scholar] [CrossRef]
  15. Kim, T.; Vecchietti, L.F.; Choi, K.; Lee, S.; Har, D. Machine learning for advanced wireless sensor networks: A review. IEEE Sens. J. 2020, 21, 12379–12397. [Google Scholar] [CrossRef]
  16. Seo, M.; Vecchietti, L.F.; Lee, S.; Har, D. Rewards prediction-based credit assignment for reinforcement learning with sparse binary rewards. IEEE Access 2019, 7, 118776–118791. [Google Scholar] [CrossRef]
  17. Stachniss, C.; Leonard, J.J.; Thrun, S. Simultaneous localization and mapping. In Springer Handbook of Robotics; Springer: Berlin/Heidelberg, Germany, 2016; pp. 1153–1176. [Google Scholar]
  18. Schonberger, J.L.; Frahm, J.M. Structure-from-motion revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4104–4113. [Google Scholar]
  19. Besl, P.J.; McKay, N.D. Method for registration of 3-D shapes. In Sensor Fusion IV: Control Paradigms and Data Structures; SPIE: Bellingham, WA, USA, 1992; pp. 586–606. [Google Scholar]
  20. Bouaziz, S.; Tagliasacchi, A.; Pauly, M. Sparse iterative closest point. In Computer Graphics Forum; Blackwell Publishing Ltd.: Oxford, UK, 2013; pp. 113–123. [Google Scholar]
  21. Rusinkiewicz, S.; Levoy, M. Efficient variants of the ICP algorithm. In Proceedings of the Third International Conference on 3-D Digital Imaging and Modeling, Quebec City, QC, Canada, 28 May–1 June 2001; pp. 145–152. [Google Scholar]
  22. Segal, A.; Haehnel, D.; Thrun, S. Generalized-icp. In Proceedings of the Robotics: Science and Systems, Seattle, WA, USA, 28 June–1 July 2009; p. 435. [Google Scholar]
  23. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  24. Rusu, R.B.; Blodow, N.; Beetz, M. Fast point feature histograms (FPFH) for 3D registration. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
  25. Zeng, A.; Song, S.; Nießner, M.; Fisher, M.; Xiao, J.; Funkhouser, T. 3dmatch: Learning the matching of local 3d geometry in range scans. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017; p. 4. [Google Scholar]
  26. Deng, H.; Birdal, T.; Ilic, S. Ppfnet: Global context aware local features for robust 3d point matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 195–205. [Google Scholar]
  27. Choy, C.; Park, J.; Koltun, V. Fully convolutional geometric features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 8958–8966. [Google Scholar]
  28. Bai, X.; Luo, Z.; Zhou, L.; Fu, H.; Quan, L.; Tai, C.L. D3feat: Joint learning of dense detection and description of 3d local features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6359–6367. [Google Scholar]
  29. Johnson, A.E.; Hebert, M. Using spin images for efficient object recognition in cluttered 3D scenes. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 433–449. [Google Scholar] [CrossRef] [Green Version]
  30. Salti, S.; Tombari, F.; Di Stefano, L. SHOT: Unique signatures of histograms for surface and texture description. Comput. Vis. Image Underst. 2014, 125, 251–264. [Google Scholar] [CrossRef]
  31. Gojcic, Z.; Zhou, C.; Wegner, J.D.; Wieser, A. The perfect match: 3d point cloud matching with smoothed densities. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5545–5554. [Google Scholar]
  32. Khoury, M.; Zhou, Q.Y.; Koltun, V. Learning compact geometric features. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 153–161. [Google Scholar]
  33. Zhong, Y. Intrinsic shape signatures: A shape descriptor for 3d object recognition. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, Kyoto, Japan, 27 September–4 October 2009; pp. 689–696. [Google Scholar]
  34. Tian, Y.; Fan, B.; Wu, F. L2-net: Deep learning of discriminative patch descriptor in euclidean space. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 661–669. [Google Scholar]
  35. Ao, S.; Hu, Q.; Yang, B.; Markham, A.; Guo, Y. Spinnet: Learning a general surface descriptor for 3d point cloud registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11753–11762. [Google Scholar]
  36. Thomas, H.; Qi, C.R.; Deschaud, J.E.; Marcotegui, B.; Goulette, F.; Guibas, L.J. Kpconv: Flexible and deformable convolution for point clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 6411–6420. [Google Scholar]
  37. Deng, H.; Birdal, T.; Ilic, S. Ppf-foldnet: Unsupervised learning of rotation invariant 3d local descriptors. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 602–618. [Google Scholar]
  38. Tombari, F.; Salti, S.; Di Stefano, L. Performance evaluation of 3D keypoint detectors. Int. J. Comput. Vis. 2013, 102, 198–220. [Google Scholar] [CrossRef]
  39. Chen, H.; Bhanu, B. 3D free-form object recognition in range images using local surface patches. Pattern Recognit. Lett. 2007, 28, 1252–1262. [Google Scholar] [CrossRef]
  40. Dorai, C.; Jain, A.K. COSMOS-A representation scheme for 3D free-form objects. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 1115–1130. [Google Scholar] [CrossRef] [Green Version]
  41. Unnikrishnan, R.; Hebert, M. Multi-scale interest regions from unorganized point clouds. In Proceedings of the 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  42. Li, J.; Lee, G.H. Usip: Unsupervised stable interest point detection from 3d point clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 361–370. [Google Scholar]
  43. Cirujeda, P.; Mateo, X.; Dicente, Y.; Binefa, X. MCOV: A covariance descriptor for fusion of texture and shape features in 3D point clouds. In Proceedings of the 2014 2nd International Conference on 3D Vision, Tokyo, Japan, 8–11 December 2014; pp. 551–558. [Google Scholar]
  44. Rusu, R.B.; Blodow, N.; Marton, Z.C.; Beetz, M. Aligning point cloud views using persistent feature histograms. In Proceedings of the 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, Nice, France, 22–26 September 2008; pp. 3384–3391. [Google Scholar]
  45. Guo, Y.; Sohel, F.; Bennamoun, M.; Lu, M.; Wan, J. Rotational projection statistics for 3D local surface description and object recognition. Int. J. Comput. Vis. 2013, 105, 63–86. [Google Scholar] [CrossRef] [Green Version]
  46. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Zanfir, A.; Sminchisescu, C. Deep learning of graph matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2684–2693. [Google Scholar]
  48. Wang, R.; Yan, J.; Yang, X. Learning combinatorial embedding networks for deep graph matching. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3056–3065. [Google Scholar]
  49. Caetano, T.S.; McAuley, J.J.; Cheng, L.; Le, Q.V.; Smola, A.J. Learning graph matching. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 1048–1058. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Kuhn, H.W. The Hungarian method for the assignment problem. Nav. Res. Logist. Q. 1955, 2, 83–97. [Google Scholar] [CrossRef] [Green Version]
  51. Sinkhorn, R. A relationship between arbitrary positive matrices and doubly stochastic matrices. Ann. Math. Stat. 1964, 35, 876–879. [Google Scholar] [CrossRef]
  52. Yew, Z.J.; Lee, G.H. 3dfeat-net: Weakly supervised local 3d features for point cloud registration. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 607–623. [Google Scholar]
  53. Sarlin, P.E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4938–4947. [Google Scholar]
  54. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  55. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Processing Syst. 2017, 30, 5998–6008. [Google Scholar]
  56. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  57. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  58. Elbaz, G.; Avraham, T.; Fischer, A. 3D point cloud registration for localization using a deep neural network auto-encoder. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4631–4640. [Google Scholar]
  59. Ma, Y.; Guo, Y.; Zhao, J.; Lu, M.; Zhang, J.; Wan, J. Fast and accurate registration of structured point clouds with small overlaps. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1–9. [Google Scholar]
  60. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
Figure 1. Autonomous vehicles and their wireless sensor networks.
Figure 1. Autonomous vehicles and their wireless sensor networks.
Applsci 12 07023 g001
Figure 2. DFGAT pipeline. The architecture is made up of three major components: keypoint encoder (Section 3.1), descriptor encoder (Section 3.2), GAT (Section 3.3).
Figure 2. DFGAT pipeline. The architecture is made up of three major components: keypoint encoder (Section 3.1), descriptor encoder (Section 3.2), GAT (Section 3.3).
Applsci 12 07023 g002
Figure 3. Visualization of matching results on KITTI. The first column are keypoints detected and matched using our method and MDGAT. The second column are ground-truth results without raw point clouds while the third column represents ground-truth results including raw point clouds. Our method performs better in terms of keypoints matching with more correct pairs (green) and fewer wrong pairs (red).
Figure 3. Visualization of matching results on KITTI. The first column are keypoints detected and matched using our method and MDGAT. The second column are ground-truth results without raw point clouds while the third column represents ground-truth results including raw point clouds. Our method performs better in terms of keypoints matching with more correct pairs (green) and fewer wrong pairs (red).
Applsci 12 07023 g003
Table 1. Matching performance.
Table 1. Matching performance.
MatcherPrecisionAccuracyRecallF1-Score
MDGAT Desc. + SuperGlue(0.1)54.376.757.70.559
MDGAT Desc. + SuperGlue(0.2)57.675.849.20.531
MDGAT Desc. + SuperGlue(0.3)62.272.739.40.482
MDGAT Desc. + SuperGlue(0)57.772.235.40.439
MDGAT (Desc. + matcher)70.386.769.40.698
DFGAT80.490.276.10.782
Table 2. Registration performance.
Table 2. Registration performance.
MatcherFailure RateInlier Ratio
FPFH + RANSAC8.3718.77
SHOT + RANSAC5.4018.21
USIP + RANSAC1.4132.20
MDGAT Desc. + SuperGlue0.5836.19
MDGAT (Desc. + matcher)1.3337.54
DFGAT0.11734.93
Table 3. Registration results on KITTI.
Table 3. Registration results on KITTI.
MethodRTE (cm)Standard Deviation (cm)RRE (°)Standard Deviation (°)Success (%)
3DFeat-Net25.926.20.570.4695.97
FCGF9.521.300.300.2896.57
D3Feat6.900.300.240.0699.81
SpinNet9.880.500.470.0999.10
MDGAT-matcher16.0064.90.6083.5698.67
DFGAT9.7021.00.241.2299.88
Table 4. Ablation of descriptor encoder.
Table 4. Ablation of descriptor encoder.
Descriptor EncoderPrecisionAccuracyRecallF1-Score
PointNet59.480.362.30.608
Global Desc.79.189.775.20.771
MDGAT70.386.769.40.698
DFGAT80.490.276.10.782
Table 5. Ablation of GAT.
Table 5. Ablation of GAT.
MatcherPrecisionAccuracyRecallF1-Score
DFGAT variant 177.989.676.30.771
DFGAT variant 278.289.173.60.758
DFGAT80.490.276.10.782
Table 6. Ablation of loss function.
Table 6. Ablation of loss function.
LossPrecisionAccuracyRecallF1-Score
SuperGlue loss54.675.747.60.509
Triplet loss56.377.851.60.538
Gap loss80.490.276.10.782
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lai-Dang, Q.-V.; Nengroo, S.H.; Jin, H. Learning Dense Features for Point Cloud Registration Using a Graph Attention Network. Appl. Sci. 2022, 12, 7023. https://doi.org/10.3390/app12147023

AMA Style

Lai-Dang Q-V, Nengroo SH, Jin H. Learning Dense Features for Point Cloud Registration Using a Graph Attention Network. Applied Sciences. 2022; 12(14):7023. https://doi.org/10.3390/app12147023

Chicago/Turabian Style

Lai-Dang, Quoc-Vinh, Sarvar Hussain Nengroo, and Hojun Jin. 2022. "Learning Dense Features for Point Cloud Registration Using a Graph Attention Network" Applied Sciences 12, no. 14: 7023. https://doi.org/10.3390/app12147023

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop