Next Article in Journal
A Reliability Analysis of a MEMS Flow Sensor with an Accelerated Degradation Test
Previous Article in Journal
The Design and Fabrication of Large-Area Under-Screen Fingerprint Sensors with Optimized Aperture and Microlens Structures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiattention Mechanism 3D Object Detection Algorithm Based on RGB and LiDAR Fusion for Intelligent Driving

State Key Laboratory of Automotive Simulation and Control, Jilin University, Changchun 130022, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(21), 8732; https://doi.org/10.3390/s23218732
Submission received: 13 September 2023 / Revised: 11 October 2023 / Accepted: 19 October 2023 / Published: 26 October 2023
(This article belongs to the Section Vehicular Sensing)

Abstract

:
This paper proposes a multimodal fusion 3D target detection algorithm based on the attention mechanism to improve the performance of 3D target detection. The algorithm utilizes point cloud data and information from the camera. For image feature extraction, the ResNet50 + FPN architecture extracts features at four levels. Point cloud feature extraction employs the voxel method and FCN to extract point and voxel features. The fusion of image and point cloud features is achieved through regional point fusion and voxel fusion methods. After information fusion, the Coordinate and SimAM attention mechanisms extract fusion features at a deep level. The algorithm’s performance is evaluated using the DAIR-V2X dataset. The results show that compared to the Part-A2 algorithm; the proposed algorithm improves the mAP value by 7.9% in the BEV view and 7.8% in the 3D view at IOU = 0.5 (cars) and IOU = 0.25 (pedestrians and cyclists). At IOU = 0.7 (cars) and IOU = 0.5 (pedestrians and cyclists), the mAP value of the SECOND algorithm is improved by 5.4% in the BEV view and 4.3% in the 3D view, compared to other comparison algorithms.

1. Introduction

With the continuous development of intelligent driving technology, the requirements for environmental awareness performance are constantly improving. Therefore, 3D object detection has received great attention [1,2,3]. However, single-modal 3D object detection alone cannot handle such a complex scene. Therefore, the research focuses on multimodal three-dimensional object detection, a key method to improve the environment perception of intelligent driving.
Convolutional Neural Network (CNN)-based techniques have performed well on detection datasets of images [4]. Significant achievements have been made in 2D target detection in images [4,5,6,7]. However, these methods cannot be directly applied to 3D detection due to the different input modes. Compared to images, LiDAR can precisely locate objects in 3D space, while 3D point clouds can provide detailed geometry and capture the 3D structure of the scene. Detection techniques based on LiDAR data are generally superior to camera-based 3D detection techniques [4,5,6,7,8,9]. Point clouds, conversely, are irregular and cannot be directly processed by powerful deep learning models, such as convolutional neural networks. This presents an excellent challenge for effective feature learning. Some of these methods convert 3D point clouds into depth maps and bird’s-eye-view (BEV) maps by manual processing and then process them in a 2D CNN manner for vehicle detection and classification [8]. However, the manually extracted features must fully utilize the information from the point cloud and may lead to performance degradation when detecting fewer points or variable geometry objects. There are also approaches [9] that use a 2D detector to generate a 2D detection frame on the image, transform the 2D detection frame into a proposed region in 3D space, and then use the PointNet [10] architecture for target detection on the point cloud. However, this approach relies heavily on the performance of the 2D target detector and cannot take advantage of the 3D information to generate robust bounding boxes.
Recent research in 3D target detection has primarily focused on utilizing end-to-end trainable neural networks that can directly process point cloud data without requiring manual feature extraction, as seen in the case of BEV maps. Vote3D [11] uses a sliding window on the sparse volume of a three-dimensional voxel grid to detect objects. The hand-crafted geometric features are extracted on each volume and fed into the SVM classifier [12]. Vote3Deep [13] also uses voxel representations of point clouds but uses 3D convolutional neural networks to extract the features of each individual [14]. The main problem with voxel representation is efficiency, as 3D voxel grids typically have high dimensions. In contrast, VeloFCN [15] projects a 3D point cloud onto the front view to obtain a 2D depth map. The vehicle is then detected using a 2D CNN on the depth map. Qi et al. [16] developed a neural network architecture that directly utilizes point clouds as input and output class labels. This allows for the learning of representations from raw data. However, this approach cannot be applied to target detection and localization due to limitations within the network architecture and the high storage cost. Zhou and Tuzel [17] proposed VoxelNet, which involves voxelizing the point cloud and employing a series of voxel feature encoding (VFE) layers to overcome this issue. This processing allows the VoxelNet network to directly extract the point cloud within the voxel using 3D convolution. Another model proposed by S. Shi et al., PointRCNN [14], offers superior performance in 3D target detection of point clouds compared to the two-stage image target detection network, Fast-RCNN [18]. PointRCNN requires further clarification regarding the point cloud pooling strategy despite its advantages. The pooling of different proposals may lead to pooling the same set of points, resulting in the loss of geometric information encoding. To address this, a new point cloud pooling operation for regions of interest, Part-A2 [19] networks, was proposed, which retains all information from nonempty and empty voxels within the proposals, eliminating ambiguity from previous point cloud pooling strategies. Although these methods have demonstrated improved performance, they all rely solely on a single modality, namely point cloud data. RGB images, on the other hand, offer denser texture, color, and additional information than point clouds, suggesting that both modalities can be leveraged to enhance detection performance.
This paper aims to solve the abovementioned issues by implementing a fusion method that combines LiDAR point cloud and RGB image for 3D target detection. The process involves utilizing ResNet50 as the backbone network for image feature extraction and the FPN structure to gather multilevel features from the images. Point cloud and image early and late fusion are achieved through regional point fusion and voxel fusion. The Coordinate and SimAM attention mechanisms further process the information extracted after point cloud image fusion. Finally, features are outputted using the SECOND-FPN structure. The network structure framework depicted in Figure 1 demonstrates the approach proposed in this paper.

2. Related Work

This section provides an overview of 3D object detection techniques for vehicles, pedestrians, and cyclists in the context of autonomous driving and ADAS systems. The research in this field has seen significant advancements in recent years. It can be categorized into three main approaches: camera-based sensor, LiDAR sensor, and camera and LiDAR multimode fusion.

2.1. Target Detection Based on Camera Sensor

Various methods have been developed to estimate 3D bounding boxes using 2D image information [20,21,22,23,24]. For example, geometric constraints between three-dimensional and two-dimensional boundary boxes are used to restore the pose of three-dimensional objects in object detection [25,26]. Similarly, refs. [27,28,29] leverage the similarity between 3D objects and CAD models to estimate the attitude of 3D object detection. Chen et al. [30,31] propose an energy function to represent the three-dimensional geometric framework and score predefined 3D boxes. Recent studies [32,33] have also explored using stereo images to enhance the performance of 3D object detection in stereoscopic cameras. However, these methods often need more precise depth information, resulting in coarse 3D detection results that are susceptible to changes in appearance.

2.2. Target Detection Based on LiDAR Sensor

The research on the use of LiDAR for three-dimensional target detection has attracted widespread attention. The manual feature method was initially successfully adopted [34,35,36,37,38], but only in scenes with clear texture information and comprehensive 3D data. Subsequently, voxel mesh occupation is introduced to represent three-dimensional point clouds [11,13,39], and three-dimensional bounding box calculation is realized by three-dimensional convolution. However, these methods are computationally and memory-intensive. To address this problem, a method based on the BEV (Bird’s-Eye-View) feature map is proposed [40,41]. This method assumes that the point cloud has a sparse vertical height, which may not hold in most cases.
Another approach to 3D object detection is to use a two-stage object detection network. In the first phase, region proposals are generated. Then, in the second phase, point clouds and associated semantic features within the candidate region are utilized to improve the accuracy of the 3D bounding box. Some methods utilize sophisticated 2D detectors to obtain the corresponding image’s two-dimensional region of interest (ROI). Inverse projection transforms these regions into three-dimensional spaces, forming a cone-shaped three-dimensional point cloud [9,42]. Finally, using these conical 3D point clouds as input, PointNet/ConvNet extracts features of interest within the area.

2.3. Target Detection Based on Camera and LiDAR Fusion

The existing research on multimode information fusion for target detection using LiDAR and RGB image data is limited [42,43]. To address the challenge of effectively fusing multiview features, Zhang et al. [44] proposed a multiview feature adaptive fusion framework for 3D object detection. Chen et al. [1] introduced the Multi-View 3D Target Detection network (MV3D), which takes in LiDAR and image data and combines regional features to generate 3D bounding boxes. While this method achieves promising results through multimode fusion, it suffers from issues, such as point cloud information loss and late-stage fusion of multimode information, which restrict the exchange of information between different data modes. Ku et al. [45] proposed a multimode fusion network incorporating regional features to overcome these limitations. By designing RPN structures with high-resolution feature mapping, this network achieves better detection results and improves its performance in detecting and classifying small objects.
In a different approach, Qi et al. proposed Frustum PointNets [46], a 3D target detection method that integrates LiDAR and image data. This model initially employs a 2D detector to generate a 2D detection box on the image, which is then converted into a proposed region in 3D space. Finally, the PointNet architecture performs target detection on the point cloud. However, this approach primarily focuses on utilizing image information and only partially exploits the potential of both data sources [46]. We adopted an attention-based multimode fusion strategy to address this limitation and enhance the exchange of information between multiple modes at the early stages.

3. Proposed Method

3.1. Image Feature Extraction Framework

In image processing, the performance of convolutional neural networks improves with increasing depth to extract more advanced features. However, traditional convolutional neural networks will encounter problems, such as network decay, gradient explosion, and gradient disappearing with increasing depth, resulting in a decline in network performance with increasing depth. To overcome these problems, this paper adopted ResNet50 [47], which employs a residual module as the backbone network for image extraction (shown in Figure 2). The residual module’s introduction solves the network recession issue, while adding a BN layer addresses gradient disappearance and explosion. ResNet50 consists of five stages, namely Stage 0 to Stage 5. The initial stage performs more straightforward tasks, primarily preprocessing the input data. The subsequent four stages (stage 1 to stage 4) comprise bottleneck structures that extract high-level semantic features from the image.
To enhance the ability to integrate low-level details and high-level semantics, it is necessary to expand the sensory field of the bottom layer and improve the detection performance of small targets. This paper adopted the multiscale feature pyramid structure of the FPN in order to fuse the low-level detail information and the high-level semantic information, thereby increasing the sensory field of the bottom layer and enhancing the detection performance of small targets. The FPN structure, as depicted in Figure 2, consists of three lines: the self-low-upward, the self-top-downward, and the lateral link. The low-up module continuously pools the forward propagation feature maps, resulting in four feature maps of different sizes. The top-down module upsamples the small-size feature maps and performs splicing and fusion operations with the prominent feature maps from the low-up process. The horizontal link adjusts the output of different feature maps using a 1 × 1 convolutional kernel with 256 channels, facilitating subsequent fusion. Finally, the final prediction output is performed on the four feature maps, forming a multilevel and multiscale feature pyramid structure.

3.2. Point Cloud Feature Extraction

3.2.1. Point Cloud Voxelization

In order to facilitate feature extraction, this paper adopted a method similar to VoxelNet [17] to perform voxelization, grouping, and random extraction operations on the sparse point cloud throughout the space. Voxelization divides the 3D space into equally spaced voxels, with voxel sizes defined based on the range of the point cloud along the Z, Y, and X directions denoted as D, H, and W, respectively. Points are then grouped based on the voxel where they are located. Due to factors such as distance, occlusion, and sparsity, the number of points in each voxel can vary, as shown in the leftmost voxelization of Figure 3.
In this paper, we address the issue of a memory burden on the computational platform that arises from processing the high-resolution LiDAR point cloud, which consists of millions of points. Directly processing these point clouds can overwhelm the computational platform due to the high number of points. Another challenge is the variable density of the point cloud throughout the spatial height, which can impact the accuracy of the detection results. We propose randomly extracting a fixed number of point clouds in each voxel to mitigate these issues. This strategy reduces the memory burden on the computational platform and helps balance the voxel distribution, resulting in improved training diversification.

3.2.2. Point Cloud Feature Extraction

In the point cloud feature extraction network, which is shown in Figure 3, first p i = x i , y i , z i , r i T represents the coordinates of each point in the voxel. The voxel has four elements that represent the X, Y, and Z coordinates and reflectance. Before feature extraction, the initial feature of each point cloud is represented by the point coordinates and the center position relative to the point coordinates. This initial feature can be expressed as p i = x i , y i , z i , r i , x i v x , y i v y , z i v z T , where v x , v y , v z denotes the center position coordinates of voxels [44]. The FCN feature extraction network then extracts the features of the points inside each voxel. After the feature extraction of each point inside the voxel is completed, the features are extracted as the voxel features using the maximum pooling method in the channel corresponding to each point. Finally, each voxel’s extracted point features and voxel features are spliced together as the final features. The FCN network comprises a linear layer, a batch normalization (BN) layer, and an activation function layer (ReLU). All the nonempty voxels are encoded in the same form and share all the parameter sets in the FCN network. The FCN structure transforms the input point cloud data into high-dimensional features. This structure encodes point interactions within voxels, allowing the final feature representation to learn to describe shape information. Therefore, point cloud features are extracted by stacking three layers of this structure in a point cloud feature extraction network.

3.3. Multimodal Fusion

This paper proposes two fusion techniques to improve the performance of 3D target detection in infrastructure view by extending the VoxelNet framework to fuse point cloud and image data. As mentioned, the VoxelNet [17] model is based on a single modality. However, this study enhances it by adding a multimodal fusion scheme, improving the network’s performance.

3.3.1. Regional Point Fusion

This early fusion technique utilizes image features to aggregate 3D point clouds, enhancing the contextual information, as illustrated in Figure 4. The method initially employs ResNet50 and FPN structure extraction networks to extract high-level feature maps from images with multilevel semantic encoding. These feature maps are then utilized to project each 3D point onto different layers’ feature maps using a calibration matrix, thereby identifying the corresponding position on the image for each point. Subsequently, a 3 × 3 convolution is applied to extract centralized features of the small region associated with each point. These features are combined with the point cloud features derived from the previous feature extraction network. Following the splicing of features, a set of FCN layer network structures is employed for further processing. Ultimately, these processed features are employed in the subsequent detection stage.
The advantage of this approach is the ability to connect multilevel image features to point cloud features at an early stage. This provides information on the location of the image corresponding to each point and feature within a small area of that point. Subsequently, the network can learn helpful information in both modalities through the FCN layer.

3.3.2. Regional Voxel Fusion

Region voxel fusion employs a relatively late fusion strategy compared to region point fusion features. After the 3D convolution has extracted the features, the 3D space is transformed into a 2D space. Subsequently, an expansion convolution with an expansion coefficient of 1 is used to further the fusion of a more extensive range of information to increase the sensory field. After region point fusion, semantic features of the image are attached to the voxel level. Each voxel contains both point cloud and image features within it. In order to fully consider the information interaction between voxel contexts, this paper adopted three 3D convolutions for the fusion and extraction of regional features on 3D information.
Regional voxel fusion is a relatively late fusion strategy but offers certain advantages. Firstly, this approach enables the fusion of features projected from the image onto the point cloud at the voxel level, thereby enhancing the combination of feature information in multiple regions near the point cloud. Secondly, expanding convolution can enhance the receptive field and facilitate the detection of smaller objects.

3.4. Fusion Feature Extraction and Attention Mechanism

3.4.1. Coordinate Attention Mechanism

After fusing point cloud and image information, this paper utilized the Coordinate attention mechanism for enhancing the feature map processing, as depicted in Figure 5.
This paper proposes an attention mechanism that considers both channel aspect information and position information in the feature map’s horizontal and vertical dimensions. To ensure that spatial information is not compressed into the channel and to enable spatial interaction of the captured information, average pooling is performed separately in the X-axis and Y-axis directions. The pooled results are then concatenated and subjected to convolution operation, allowing interactions between the positions in the two axes. Once the information interaction is completed, the weights are sparsely computed along the two axes, resulting in a final feature that retains positional- and channel-specific information.

3.4.2. SimAM Attention Mechanism

After completing the fusion feature extraction with the Coordinate attention mechanism, this paper focused on utilizing the FPN structure in the point cloud for two-level feature extraction. This approach ensures that the deeper-level feature map contains richer semantic information, while the shallow-level feature map preserves more complete geometric details. Figure 6 demonstrates the SimAM parameter-free attention mechanism employed in this study to assess the significance of each neuron in the network. This mechanism accomplishes the differentiation by defining the energy function’s form, where neurons with higher energy functions are assigned greater weights due to their increased importance. In comparison, those with lower energy functions are assigned lower weights.

4. Loss Functions

This paper firstly parameterized the 3D truth frame as follows: ( x c g , y c g , z c g , l g , w g , h g , θ g ) . x c g , y c g , z c g represent the center of the 3D truth frame, l g , ω g , h g represent the length, width, and height of the 3D truth frame, and θ g represents the rotation angle along the Z-axis. At the same time, the paper parameterized the Anchor designed by ourselves in the target detection as ( x c a , y c a , z c a , l a , w a , h a , θ a ) and defined seven residual coefficients of regression. These coefficients represent the offset relative to the center coordinate, the elongation or shortening ratio of the three dimensions of length, width, and height, as well as the deviation around the direction of the Z-axis. The calculations of these seven coefficients are shown as follows.
Δ x = x c g x c a d a , Δ y = y c g y c a d a , Δ z = z c g z c a h a , Δ l = log ( l g l a ) , Δ w = log ( w g w a ) , Δ h = log ( h g h a ) , Δ θ = θ g θ a
In the designed anchor frame, Equation (1) l a represents the diagonal length of the 3D frame base. In order to estimate the oriented 3D detection frame directly, Δ x , Δ y , and the diagonal of the 3D frame d a were normalized. This differs from the method provided by Li et al. [15]. Finally, the loss function is defined in this paper, as shown below.
L = α 1 N p o s i L c l s ( p i p o s , 1 ) + β 1 N n e g j L c l s ( p j n e g , 0 ) + 1 N p o s i L r e g ( u i , u i * )
In Equation (2), N p o s represents the number of positive anchor frames and N n e g represents the number of negative anchor frames. p i p o s represents the probability that the i th anchor frame is predicted to be a true label, and p j n e g represents the probability that the j anchor frame is predicted to be a false label. u i represents the bounding box regression parameter of the i th anchor box, and u i * represents the bounding box regression parameter of the i th anchor box corresponding to the truth box. L c l s denotes cross-entropy loss, α , β denote balanced positive and negative sample parameters, and L r e g denotes regression parameters.

5. Experimental Section and Discussion

5.1. Datasets

This paper proposes a multimodal fusion with multiple attention mechanisms for a 3D target detection algorithm. The algorithm will be evaluated on the DAIR-V2X dataset, which contains 15,285 image data and 15,285 frames of point cloud data. The dataset is further decomposed into a training set and a validation set in the ratio of 7:3. After the split, the training set consists of 10,700 samples, and the validation set consists of 4585. The evaluation will analyze the effectiveness of the proposed multimodal approach by comparing it with previously published methods for 3D target detection tasks. The evaluation will consider three difficulty levels, easy, medium, and hard, based on object size, visibility (occlusion), and truncation.

5.2. Data Enhancement

This study aims to solve the overfitting problem in the network training process by enhancing the data of both images and point clouds. We performed the following steps on the image: First, we resized the image to two different sizes, (640,192) and (2,560,768). Next, we used [102.42,117.36,124.58] as the mean and [1.0,1.0,1.0] as the regularized variance. Finally, we flipped the image horizontally with a flip ratio of 0.5. We flipped within the specified angle range for point clouds and zoomed in the [0.90,1.10] range. Like the image, the point cloud was flipped horizontally with a flip ratio of 0.5.
In addition, all ground truth boxes (bi) and the entire point cloud (M) were scaled globally. Specifically, we multiplied the XYZ coordinate and three-dimensional space of each bi and the XYZ coordinate of a point in M by a random variable evenly distributed in [0.95,1.05]. In addition, this paper introduces global scaling in image-based classification [48] and detection tasks [18] to enhance the network’s ability to detect targets of different sizes and distances, thereby improving overall robustness.

5.3. Experimental Seting

We conducted experiments on the 3D detection using the Pytorch framework in Ubuntu 18.04. The experiments were implemented on a computer with the following specifications: 1 NVIDIA RX3060 graphics card with 12 GB GPU memory, i9-7900X @3.60 GHz × 10 processors, and 64GB RAM. We used Adam with hyperparameters for training optimization and set to 0.9 and 0.999. The number of iterations was set to 500 K, with an initial learning rate of 0.0001. The learning rate decayed every 20 epochs with a decay rate of 0.9. Due to memory constraints, we conducted the experiments in four batches to obtain better experimental results with the available hardware, as larger batch sizes usually lead to more improvements. It is important to note that background should not be ignored in training and testing processes.

5.4. Experimental Parameters

5.4.1. Image Detection Networks

In this study, ResNet50 + FPN architecture was used for feature extraction in image object detection. The training dataset was the DAIR-V2X dataset, and data augmentation techniques were applied. During training, the shortest edge of the image was rescaled to 600 pixels. This paper used four scale anchors {4,8,16,32} and three aspect ratios {0.5,1,2} for the final output layer. If the intersection of the anchor point with the ground truth bounding box exceeded 0.7, the anchor point was marked as positive, and if the IOU was less than 0.3, the anchor point was marked as negative. The network was trained by stochastic gradient descent with a learning rate 0.0005 and a momentum of 0.9. In the multimodal fusion training process, once the image detection network training was completed, the parameters of the image network were frozen, and the weight coefficient of the image backbone network remained unchanged.

5.4.2. Point Cloud Detection Networks

This research focuses on point cloud detection using VoxelNet architecture. The point cloud was considered along the Z, Y, and X axes of [−3,1] × [−40,40] × [0,70.4] meters, respectively. To ensure accuracy, points projected outside the image boundaries were removed [25]. For voxel size, we chose vD = 0.05, vH = 0.05, vW = 0.1 m, and we obtained D’ = 80, H’ = 1600, W’ = 1408. In addition, we set T = 35 as the maximum number of random sampling points in each nonempty voxel. Two fully convolutional network (FCN) layers, FCN-1 (7, 32) and FCN-2 (32, 128). The final FCN mapped the output of VFE-2 to R128. Subsequently, feature extraction between point clouds was achieved using cubic convolutional fusion.

5.4.3. Multimodal Fusion

Two 128-dimensional FCN modules were used to extract features from the point cloud. These modules extracted features after fusion, projecting image features onto each corresponding point. Then, 3 layers of conv3D convolution were used to perform voxel fusion of the information obtained after the point fusion. Finally, the Regional Suggestion Network (RPN) structure outputted the final result.

5.5. DAIR-V2X Dataset Evaluation

This paper evaluated the detection performance using standard DAIR-V2X evaluation protocols (vehicle IOU = 0.7 and IOU = 0.5 and bicycle and pedestrian IOU = 0.5 and IOU = 0.25). Table 1, Table 2, Table 3 and Table 4 compare the algorithm presented in this article with the AP of commonly used algorithms in the 3D view and the BEV view. The results show that the three-dimensional target detection algorithm proposed in this paper significantly improves the detection performance compared with the commonly used algorithms. It is worth noting that the fusion effect is more pronounced in the 3D view score than in the BEV view scoring. It is worth mentioning that the proposed fusion technique outperforms the original voxel network through more powerful RPNs and is enhanced with additional data. In addition, this method has consistently outperformed the best-performing process in recent times. A sample test result of the proposed method is shown in Figure 7.

5.6. Performance Analysis of the AP Value Algorithm

As seen from Table 1, Table 2, Table 3 and Table 4, the algorithm proposed in this paper is the most advanced in detecting vehicles, pedestrians, and cyclists. It is better than the SECOND algorithm of single mode and the MV3D algorithm of multimode to a large extent. Especially in detecting small target pedestrians and cyclists in easy, medium, and difficult modes, the AP performance index of the algorithm proposed in this paper is improved by more than 7%. At the same time, in the case of IOU = 0.5, our algorithm can reach more than 60% in the BEV view and more than 50% in the 3D view. This is because the texture and color information provided by images in detecting small targets in this paper can compensate for the sparse point cloud. In vehicle detection, the AP performance index of the proposed algorithm can approach 80% when IOU = 0.7.

5.7. Stability Analysis

The algorithm’s stability is represented by calculating the mAP value of each method in this paper. A relatively high mAP value indicates better performance in detecting vehicles, pedestrians, and bicyclists, making the algorithm suitable for multisize, multitarget detections. Conversely, a low mAP value suggests that the algorithm is less effective in detecting one of the objects in the target detection. In Table 5, it can be observed that the proposed algorithm in this paper exhibits a 7.9% improvement in the mAP value in the BEV view and a 7.8% improvement in the 3D view compared to the Part-A2 network, which is an exceptional performer among the compared algorithms, with an IOU of 0.5 for cars and 0.25 for pedestrians and bicycles. Additionally, Table 6 demonstrates that when the IOU is set to 0.7 for cars and 0.5 for pedestrians and bicycles, the proposed algorithm shows a 5.4% improvement in the mAP value in the BEV view and a 4.3% improvement in the 3D view, compared to the SECOND algorithm, which is another remarkable performer among the compared algorithms.

5.8. Ablation Experiment

In order to explore the performance of the algorithm in a complex environment, we collected images and LiDAR data on rainy, foggy, and cloudy days in the DAIR-V2X dataset for model experiment verification. The experimental results are shown in Table 7 and Table 8.
As can be seen from Table 7 and Table 8, the proposed algorithm performs best in rainy, cloudy, and foggy environments in complex environments. Compared with conventional environments, the mAP performance of the BEV and 3D view decreases by 2% when IOU = 0.5 (vehicle) and IOU = 0.25 (pedestrian and bicycle). With an IOU of 0.7 (vehicle) and 0.5 (pedestrian and bicycle), the mAP performance of the BEV view and 3D view decreases by 3%, so we can see that the performance of our proposed algorithm is not significantly reduced in complex environments compared to other algorithms.
At the same time, to evaluate the camera’s performance under different viewing angles, the model we established was analyzed experimentally on the KITTI dataset, and the experimental results are shown in Table 9.
It can be seen from Table 9 that the algorithm we proposed also has the best performance on the KITTI dataset. The mAP value of the algorithm Part-A2 with a better performance than that of a single mode is increased by 1.27% and that of the MV3D algorithm with a multimode fusion framework is increased by 5.3%. The performance of our proposed algorithm on different datasets is relatively stable.
In order to consider the complexity and real-time performance, we tested each algorithm’s running time on a device equipped with an NVIDIA RX3060 graphics card and 12GB GPU memory. The experimental results are shown in Table 10.
From Table 10, the running time of the algorithm we proposed on the same device is not very dominant. However, the time is close to the existing multimode fusion algorithm, and no significant difference exists. This is related to the complexity of the model and the running time cost increases when the algorithm performance improves. In future research, we should continue exploring more efficient algorithms with the premise that the operating cost is unchanged.

6. Conclusions

The proposal presented in this paper is a multimodal fusion with multiple attention mechanisms for the 3D Target Detection Algorithm aimed at addressing the limitations of single-modal target detection. The first step involved utilizing the ResNet50 + FPN network framework to extract image features, resulting in the extraction of four-level features. Simultaneously, the point cloud feature extraction employed the voxel grid method and FCN to extract point and voxel features from each voxel. These extracted features were then considered the final features of the point cloud. Following this, regional point fusion and voxel fusion techniques combined the image and point cloud features. Once the fusion process was completed, the fused features underwent depth extraction using the SECOND network. Furthermore, the Coordinate attention mechanism and the SimAM attention mechanism were implemented during this stage. Finally, the RPN was applied to obtain the output. To validate and compare the proposed algorithm with other state-of-the-art algorithms, it was tested on the DAIR-V2X dataset. The results demonstrate that the proposed algorithm surpasses other algorithms regarding detection performance. However, due to the limitation of the 3D object detection dataset, the 3D object detection algorithm could not be studied under more severe conditions in this study. Therefore, in future research, datasets under harsh conditions will be collected by actual vehicles, types and scenarios of datasets will be enriched, and deep learning models based on Transformer will be built to improve the performance of 3D target detection.

Author Contributions

Conceptualization, X.Z., Y.W. and L.H.; software, X.Z.; validation, J.C. and B.W., data curation, Y.Z.; writing—original draft preparation, X.Z.; writing—review and editing, Y.W. and Y.Z.; visualization, J.C. and B.W.; supervision, L.H.; project administration, L.H.; funding acquisition, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported by the project on research and development of intelligent vehicle key technologies and industrialization projects based on new-energy vehicles under Grant TC210H02S and by the project on quantitative development and measurement technology research of expected functional safety based on vehicle–cloud collaboration under Grant 20220301012GX.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data in this paper is from DAIR-V2X’s Cooperative Vehicle Infrastructure3D Detection: https://thudair.baai.ac.cn/task-cooptest (accessed on 18 October 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3d object detection network for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1907–1915. [Google Scholar]
  2. Liang, M.; Yang, B.; Wang, S.; Urtasun, R. Deep Continuous Fusion for Multi-sensor 3D Object Detection. In Proceedings of the Computer Vision—ECCV 2018—15th European Conference, Munich, Germany, 8–14 September 2018; Volume 11220, pp. 663–678. [Google Scholar]
  3. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. PointPillars: Fast Encoders for Object Detection From Point Clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  4. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the 13th European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  5. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  6. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  7. Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A. You Only Look Once: Unifified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  8. Yang, B.; Luo, W.; Urtasun, R. PIXOR: Real-Time 3D Object Detection From Point Clouds. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2018), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7652–7660. [Google Scholar]
  9. Wang, Z.; Jia, K. Frustum convnet: Sliding frustums to aggregate local point-wise features for amodal 3d object detection. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 1742–1749. [Google Scholar]
  10. Garcia-Garcia, A.; Gomez-Donoso, F.; Garcia-Rodriguez, J.; Orts-Escolano, S.; Cazorla, M.; Azorin-Lopez, J. Pointnet: A 3d Convolutional Neural Network for Real-Time Object Class Recognition. In Proceedings of the 2016 International Joint Conference on Neural Networks, Vancouver, BC, Canada, 24–29 July 2016. [Google Scholar]
  11. Wang, D.Z.; Posner, I. Voting for Voting in Online Point Cloud Object Detection. In Proceedings of the Robotics: Science and Systems XI, Sapienza University of Rome, Rome, Italy, 13–17 July 2015. [Google Scholar]
  12. Suykens, J.A.K.; Vandewalle, J. Least Squares Support Vector Machine Classifiers. Neural Process. Lett. 1999, 9, 293–300. [Google Scholar] [CrossRef]
  13. Engelcke, M.; Rao, D.; Wang, D.Z.; Tong, C.H.; Posner, I. Vote3Deep: Fast object detection in 3D point clouds using efficient convolutional neural networks. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA 2017), Singapore, 29 May–3 June 2017; pp. 1355–1361. [Google Scholar]
  14. Tran, D.; Bourdev, L.D.; Fergus, R.; Torresani, L.; Paluri, M. Learning Spatiotemporal Features with 3D Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision 2015, Las Condes, Chile, 11–18 December 2015; pp. 4489–4497. [Google Scholar]
  15. Li, B.; Zhang, T.; Xia, T. Vehicle detection from 3d lidar using fully convolutional network. arXiv 2016, arXiv:1608.07916. [Google Scholar]
  16. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classifification and Segmentation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017), Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar]
  17. Zhou, Y.; Tuzel, O. VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2018), Salt Lake City, UT, USA, 18–22 June 2018; pp. 4490–4499. [Google Scholar]
  18. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  19. Shi, S.; Wang, X.; Li, H. PointRCNN: 3D Object Proposal Generation and Detection From Point Cloud. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  20. Shi, S.; Wang, Z.; Wang, X.; Li, H. Part-a2 net: 3d part-aware and aggregation neural network for object detection from point cloud. arXiv 2019, arXiv:1907.03670. [Google Scholar]
  21. Song, S.; Xiao, J. Deep sliding shapes for amodal 3D object detection in rgb-d images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 808–816. [Google Scholar]
  22. Xiang, Y.; Choi, W.; Lin, Y.; Savarese, S. Data-driven 3d voxel patterns for object category recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1903–1911. [Google Scholar]
  23. Xue, B.; Tong, N. Real-World ISAR Object Recognition Using Deep Multimodal Relation Learning. IEEE Trans. Cybern. 2020, 50, 4256–4267. [Google Scholar] [CrossRef] [PubMed]
  24. Timofte, R.; Zimmermann, K.; Van Gool, L. Multi-View Traffific Sign Detection, Recognition, and 3D Localisation. Mach. Vis. Appl. 2014, 25, 633–647. [Google Scholar] [CrossRef]
  25. Li, B.; Ouyang, W.; Sheng, L.; Zeng, X.; Wang, X. GS3D: An Effificient 3D Object Detection Framework for Autonomous Driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 1019–1028. [Google Scholar]
  26. Mousavian, A.; Anguelov, D.; Flynn, J.; Košecká, J. 3D bounding box estimation using deep learning and geometry. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5632–5640. [Google Scholar]
  27. Chabot, F.; Chaouch, M.; Rabarisoa, J.; Teuliere, C.; Chateau, T. Deep MANTA: A Coarse-to-fifine Many-task Network for Joint 2D and 3D Vehicle Analysis from Monocular Image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2040–2049. [Google Scholar]
  28. Zhu, M.; Derpanis, K.G.; Yang, Y.; Brahmbhatt, S.; Zhang, M.; Phillips, C.; Lecce, M.; Daniilidis, K. Single Image 3D Object Detection and Pose Estimation for Grasping. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA 2014), Hong Kong, China, 31 May–7 June 2014. [Google Scholar]
  29. Manhardt, F.; Kehl, W.; Gaidon, A. ROI-10D: Monocular Lifting of 2D Detection to 6D Pose and Metric Shape. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 2069–2078. [Google Scholar]
  30. Chen, X.; Kundu, K.; Zhang, Z.; Ma, H.; Fidler, S.; Urtasun, R. Monocular 3D Object Detection for Autonomous Driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2147–2156. [Google Scholar]
  31. Chen, X.; Kundu, K.; Zhu, Y.; Berneshawi, A.G.; Ma, H.; Fidler, S.; Urtasun, R. 3d object proposals for accurate object class detection. In Advances in Neural Information Processing Systems; NIPS: San Diego, CA, USA, 2015; pp. 424–432. [Google Scholar]
  32. Li, P.; Chen, X.; Shen, S. Stereo r-cnn based 3d object detection for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 7644–7652. [Google Scholar]
  33. Wang, Y.; Chao, W.L.; Garg, D.; Hariharan, B.; Campbell, M.; Weinberger, K.Q. Pseudo-LiDAR From Visual Depth Estimation: Bridging the Gap in 3D Object Detection for Autonomous Driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 8437–8445. [Google Scholar]
  34. Chua, C.S.; Jarvis, R. Point signatures: A new representation for 3d object recognition. Int. J. Comput. Vis. 1997, 25, 63–85. [Google Scholar] [CrossRef]
  35. Ba, J.; Mnih, V.; Kavukcuoglu, K.J. Multiple object recognition with visual attention. arXiv 2014, arXiv:1412.7755. [Google Scholar]
  36. Liu, Z.; Zhao, X.; Huang, T.; Hu, R.; Zhou, Y.; Bai, X. Tanet: Robust 3d object detection from point clouds with triple attention. In Proceedings of the AAAI Conference on Artifificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 11677–11684. [Google Scholar]
  37. Dorai, C.; Jain, A.K. COSMOS-A representation scheme for 3D free-form objects. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 1115–1130. [Google Scholar] [CrossRef]
  38. Tuzel, O.; Liu, M.-Y.; Taguchi, Y.; Raghunathan, A. Learning to rank 3d features. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Part I; pp. 520–535. [Google Scholar]
  39. Li, B. 3d fully convolutional network for vehicle detection in point cloud. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 1513–1518. [Google Scholar]
  40. Premebida, C.; Carreira, J.; Batista, J.; Nunes, U. Pedestrian detection combining rgb and dense lidar data. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Chicago, IL, USA, 14–18 September 2014; pp. 4112–4117. [Google Scholar]
  41. Gonzalez, A.; Villalonga, G.; Xu, J.; Vazquez, D.; Amores, J.; Lopez, A. Multiview random forest of local experts combining rgb and lidar data for pedestrian detection. In Proceedings of the 2015 IEEE Intelligent Vehicles Symposium (IV), Seoul, Republic of Korea, 28 June–1 July 2015; pp. 356–361. [Google Scholar]
  42. Qi, C.R.; Liu, W.; Wu, C.; Su, H.; Guibas, L.J. Frustum pointnets for 3d object detection from rgb-d data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 918–927. [Google Scholar]
  43. Zhang, Z.H.; Liang, Z.D.; Zhang, M.; Zhao, X.; Li, H.; Yang, M.; Tan, W.M.; Pu, S.L. RangeLVDet: Boosting 3D Object Detection in LIDAR with Range Image and RGB Image. IEEE Sens. J. 2022, 22, 1391–1403. [Google Scholar] [CrossRef]
  44. Zhang, Y.; Wu, H. 3D Object Detection Based on Multi-view Adaptive Fusion. In Proceedings of the 2022 IEEE Asia-Pacific Conference on Image Processing, Electronics and Computers (IPEC), Dalian, China, 14–16 April 2022; pp. 743–748. [Google Scholar]
  45. Ku, J.; Mozififian, M.; Lee, J.; Harakeh, A.; Waslander, S. Joint 3D Proposal Generation and Object Detection from View Aggregation. arXiv 2017, arXiv:1712.02294. [Google Scholar]
  46. Rozenberszki, D.; Litany, O.; Dai, A. UnScene3D: Unsupervised 3D Instance Segmentation for Indoor Scenes. arXiv 2023, arXiv:2303.14541. [Google Scholar]
  47. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  48. Yan, Y.; Mao, Y.; Li, B. Second: Sparsely embedded convolutional detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef] [PubMed]
  49. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  50. He, C.; Zeng, H.; Huang, J.; Hua, X.S.; Zhang, L. Structure aware single-stage 3d object detection from point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11873–11882. [Google Scholar]
  51. Vora, S.; Lang, A.H.; Helou, B.; Beijbom, O. PointPainting: Sequential Fusion for 3D Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
Figure 1. The framework of 3D object detection based on multimodal fusion. After the image and point cloud data pass through their respective feature extraction networks, the image and point cloud features are obtained, respectively. Then, the point cloud and image features are matched, and the multimodal information fusion operation is achieved by regional point fusion and voxel fusion. Finally, 3D target prediction output is produced using the Coordinate and SimAM attention mechanisms.
Figure 1. The framework of 3D object detection based on multimodal fusion. After the image and point cloud data pass through their respective feature extraction networks, the image and point cloud features are obtained, respectively. Then, the point cloud and image features are matched, and the multimodal information fusion operation is achieved by regional point fusion and voxel fusion. Finally, 3D target prediction output is produced using the Coordinate and SimAM attention mechanisms.
Sensors 23 08732 g001
Figure 2. Image feature extraction network framework. After passing through four bottleneck modules of ResNet50, image data are exported with four levels of feature graphs. The feature graphs at four levels are collected for feature stitching, and feature output is performed from four dimensions.
Figure 2. Image feature extraction network framework. After passing through four bottleneck modules of ResNet50, image data are exported with four levels of feature graphs. The feature graphs at four levels are collected for feature stitching, and feature output is performed from four dimensions.
Sensors 23 08732 g002
Figure 3. Point cloud voxelization and feature extraction. The feature extraction of each point in each voxel is carried out. Then, the maximum pooling operation is carried out on the horizontal channel to obtain the feature as the voxel feature. Combining the most point feature and the voxel feature represents the final feature.
Figure 3. Point cloud voxelization and feature extraction. The feature extraction of each point in each voxel is carried out. Then, the maximum pooling operation is carried out on the horizontal channel to obtain the feature as the voxel feature. Combining the most point feature and the voxel feature represents the final feature.
Sensors 23 08732 g003
Figure 4. The framework of regional point fusion. The point features are transformed into four-level image features utilizing coordinate transformation, and then the point features and image features are aligned.
Figure 4. The framework of regional point fusion. The point features are transformed into four-level image features utilizing coordinate transformation, and then the point features and image features are aligned.
Sensors 23 08732 g004
Figure 5. The schematic diagram of coordinated attention mechanism. The feature map is average pooled in the X and Y directions. The features in the two directions are spliced, different convolution is used to extract the features, and, finally, the feature addition is used to output.
Figure 5. The schematic diagram of coordinated attention mechanism. The feature map is average pooled in the X and Y directions. The features in the two directions are spliced, different convolution is used to extract the features, and, finally, the feature addition is used to output.
Sensors 23 08732 g005
Figure 6. The schematic diagram of SimAM attention mechanism. The feature map uses two FPN structures to extract their features, takes the minimum energy function as the goal to extract their respective weight coefficients, and then weights the features extracted by FPN. Finally, the two weighted features are spliced and output.
Figure 6. The schematic diagram of SimAM attention mechanism. The feature map uses two FPN structures to extract their features, takes the minimum energy function as the goal to extract their respective weight coefficients, and then weights the features extracted by FPN. Finally, the two weighted features are spliced and output.
Sensors 23 08732 g006
Figure 7. Diagram of algorithm test results.
Figure 7. Diagram of algorithm test results.
Sensors 23 08732 g007
Table 1. Performance comparison of AP value algorithms with IOU = 0.5 and IOU = 0.25 from the BEV perspective.
Table 1. Performance comparison of AP value algorithms with IOU = 0.5 and IOU = 0.25 from the BEV perspective.
MethodCarPedestrian Cyclist
IOU = 0.5IOU = 0.25IOU = 0.25
EasyMedHardEasyMedHardEasyMedHard
SECOND [49]85.586.183.973.364.163.670.365.864.3
PointPillars [3]85.583.983.766.955.354.862.859.858.7
Part-A2 [20]85.185.483.465.959.859.366.661.760.5
3D-SSD [50]84.283.982.0N/AN/AN/AN/AN/AN/A
MV3D [1]85.484.283.174.265.063.369.464.563.2
PointPaint [51]85.083.682.073.364.562.768.663.662.1
ContFusion [2]84.883.181.472.863.561.667.262.260.8
Ours88.486.686.581.571.070.771.670.870.9
Table 2. Performance comparison of AP value algorithms with IOU = 0.7 and IOU = 0.5 from the BEV perspective.
Table 2. Performance comparison of AP value algorithms with IOU = 0.7 and IOU = 0.5 from the BEV perspective.
MethodCarPedestrian Cyclist
IOU = 0.5IOU = 0.25IOU = 0.25
EasyMedHardEasyMedHardEasyMedHard
SECOND [49]81.983.083.657.549.348.965.059.658.7
PointPillars [3]82.080.978.457.746.345.850.555.153.4
Part-A2 [20]80.280.779.755.747.647.162.856.155.2
3D-SSD [50]79.379.479.5N/AN/AN/AN/AN/AN/A
MV3D [1]82.682.181.864.253.953.264.857.256.8
PointPaint [51]81.481.080.562.651.450.663.756.355.4
ContFusion [2]80.980.579.861.951.750.264.155.854.2
Ours84.883.583.870.260.259.966.064.463.1
Table 3. Performance comparison of AP value algorithms with IOU = 0.5 and IOU = 0.25 from the 3D perspective.
Table 3. Performance comparison of AP value algorithms with IOU = 0.5 and IOU = 0.25 from the 3D perspective.
MethodCarPedestrian Cyclist
IOU = 0.5IOU = 0.25IOU = 0.25
EasyMedHardEasyMedHardEasyMedHard
SECOND [49]82.983.683.372.963.662.470.265.664.3
PointPillars [3]83.083.581.166.755.054.162.759.857.9
Part-A2 [20]84.283.282.465.559.559.065.960.960.2
3D-SSD [50]82.381.972.5N/AN/AN/AN/AN/AN/A
MV3D [1]85.485.683.174.364.163.869.264.863.9
PointPaint [51]84.684.283.672.861.560.267.862.661.7
ContFusion [2]82.782.682.171.360.459.666.161.560.2
Ours87.886.283.881.170.670.370.570.670.7
Table 4. Performance comparison of AP value algorithms with IOU = 0.7 and IOU = 0.5 from the 3D perspective.
Table 4. Performance comparison of AP value algorithms with IOU = 0.7 and IOU = 0.5 from the 3D perspective.
MethodCarPedestrian Cyclist
IOU = 0.5IOU = 0.25IOU = 0.25
EasyMedHardEasyMedHardEasyMedHard
SECOND [49]75.873.970.952.843.442.963.256.555.7
PointPillars [3]75.673.670.852.540.640.156.550.849.0
Part-A2 [20]77.174.771.552.841.841.561.654.052.4
3D-SSD [50]76.171.065.7N/AN/AN/AN/AN/AN/A
MV3D [1]76.874.271.953.444.142.465.057.255.8
PointPaint [51]75.975.271.452.842.740.863.855.754.2
ContFusion [2]75.374.874.253.541.442.662.554.353.4
Ours79.976.173.562.951.751.463.458.757.6
Table 5. mAP results at IOU = 0.5 and IOU = 0.25.
Table 5. mAP results at IOU = 0.5 and IOU = 0.25.
MethodIOU = 0.5 (Car)
IOU = 0.25 (Pedestrian, Cyclist)
BEV3D
SECOND [49]66.072.1
PointPillars [3]67.967.1
Part-A2 [20]69.769.0
MV3D [1]72.173.4
PointPaint [51]71.470.8
ContFusion [2]70.169.4
Ours77.676.8
Table 6. mAP results at IOU = 0.7 and IOU = 0.5.
Table 6. mAP results at IOU = 0.7 and IOU = 0.5.
MethodIOU = 0.7 (Car)
IOU = 0.5 (Pedestrian, Cyclist)
BEV3D
SECOND [49]65.359.4
PointPillars [3]61.156.6
Part-A2 [20]62.958.6
MV3D [1]66.7611
PointPaint [51]65.462.1
ContFusion [2]65.860.0
Ours70.763.9
Table 7. mAP results in complex environments are IOU = 0.5 and IOU = 0.25.
Table 7. mAP results in complex environments are IOU = 0.5 and IOU = 0.25.
MethodIOU = 0.5 (Car)
IOU = 0.25 (Pedestrian, Cyclist)
BEV3D
SECOND [49]61.069.1
PointPillars [3]62.962.1
Part-A2 [20]66.766.0
MV3D [1]67.269.6
PointPaint [51]66.366.8
ContFusion [2]64.563.2
Ours75.674.8
Table 8. mAP results in complex environments are IOU = 0.7 and IOU = 0.5.
Table 8. mAP results in complex environments are IOU = 0.7 and IOU = 0.5.
MethodIOU = 0.7 (Car)
IOU = 0.5 (Pedestrian, Cyclist)
BEV3D
SECOND [49]55.347.4
PointPillars [3]52.149.6
Part-A2 [20]58.954.6
MV3D [1]61.154.8
PointPaint [51]60.253.2
ContFusion [2]58.252.0
Ours67.760.9
Table 9. mAP results on KITTI dataset.
Table 9. mAP results on KITTI dataset.
MethodmAP
SECOND [49]65.3
PointPillars [3]64.07
Part-A2 [20]68.33
MV3D [1]64.3
PointPaint [51]61.3
ContFusion [2]60.2
Ours69.6
Table 10. Comparison of the running time of the algorithm.
Table 10. Comparison of the running time of the algorithm.
MethodTime (s)
SECOND [49]0.046
PointPillars [3]0.27
Part-A2 [20]0.52
MV3D [1]0.33
PointPaint [51]0.31
ContFusion [2]0.28
Ours0.39
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; He, L.; Chen, J.; Wang, B.; Wang, Y.; Zhou, Y. Multiattention Mechanism 3D Object Detection Algorithm Based on RGB and LiDAR Fusion for Intelligent Driving. Sensors 2023, 23, 8732. https://doi.org/10.3390/s23218732

AMA Style

Zhang X, He L, Chen J, Wang B, Wang Y, Zhou Y. Multiattention Mechanism 3D Object Detection Algorithm Based on RGB and LiDAR Fusion for Intelligent Driving. Sensors. 2023; 23(21):8732. https://doi.org/10.3390/s23218732

Chicago/Turabian Style

Zhang, Xiucai, Lei He, Junyi Chen, Baoyun Wang, Yuhai Wang, and Yuanle Zhou. 2023. "Multiattention Mechanism 3D Object Detection Algorithm Based on RGB and LiDAR Fusion for Intelligent Driving" Sensors 23, no. 21: 8732. https://doi.org/10.3390/s23218732

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop