Next Article in Journal
Advances in MIMO Antenna Design for 5G: A Comprehensive Review
Previous Article in Journal
Minimal Information Data-Modelling (MID) and an Easily Implementable Low-Cost SHM System for Use on a Short-Span Bridge
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MInet: A Novel Network Model for Point Cloud Processing by Integrating Multi-Modal Information

School of Electronic Engineering, Beijing University of Post and Telecommunications, Beijing 100876, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(14), 6327; https://doi.org/10.3390/s23146327
Submission received: 26 May 2023 / Revised: 3 July 2023 / Accepted: 8 July 2023 / Published: 12 July 2023
(This article belongs to the Section Sensor Networks)

Abstract

:
Three-dimensional LiDAR systems that capture point cloud data enable the simultaneous acquisition of spatial geometry and multi-wavelength intensity information, thereby paving the way for three-dimensional point cloud recognition and processing. However, due to the irregular distribution, low resolution of point clouds, and limited spatial recognition accuracy in complex environments, inherent errors occur in classifying and segmenting the acquired target information. Conversely, two-dimensional visible light images provide real-color information, enabling the distinction of object contours and fine details, thus yielding clear, high-resolution images when desired. The integration of two-dimensional information with point clouds offers complementary advantages. In this paper, we present the incorporation of two-dimensional information to form a multi-modal representation. From this, we extract local features to establish three-dimensional geometric relationships and two-dimensional color relationships. We introduce a novel network model, termed MInet (Multi-Information net), which effectively captures features relating to both two-dimensional color and three-dimensional pose information. This enhanced network model improves feature saliency, thereby facilitating superior segmentation and recognition tasks. We evaluate our MInet architecture using the ShapeNet and ThreeDMatch datasets for point cloud segmentation, and the Stanford dataset for object recognition. The robust results, coupled with quantitative and qualitative experiments, demonstrate the superior performance of our proposed method in point cloud segmentation and object recognition tasks.

1. Introduction

As autonomous driving and unmanned systems increase in application, the importance of environmental modeling, recognition, segmentation, and other processing tasks has escalated. Additionally, high spatial precision in complex environments has become vital, particularly for autonomous driving [1]. Consequently, using different sensors to acquire multi-modal information has emerged as a prevalent trend, with devices such as RGB (Two-dimensional color space information of the target) cameras [2], light detection and ranging (LiDAR) sensors for distance measurement [3], depth cameras, and radar sensors becoming increasingly common. LiDAR has established itself as a key element of 3D perception systems, enabling direct spatial dimension measurements [4] and yielding accurate 3D representations. Mobile point clouds derived from LiDAR are applicable to various tasks, such as object detection, tracking, semantic segmentation, and point cloud recognition [5]. Moreover, the geometric information encapsulated in point clouds is fundamental to numerous applications, including road network management, building and urban planning, and 3D high-definition mapping for autonomous vehicles [6].
At present, established processing algorithms for LiDAR point clouds encompass PointNet [7], PointNet++ [8], GACNet [9], DGCNN [10], among others. However, feature extraction in these algorithms only applies to the three-dimensional geometric features of the target object, which is not enough to fully characterize the object, which makes further point cloud processing still a considerable challenge. Thus, the application of suitable characteristics proved to be a research challenge. The primary obstacle in point cloud segmentation and recognition is the selection of more suitable features. To address this challenge, we propose MInet (Multi-Information net model), a network architecture for multi-modal information recognition. Our approach incorporates the 2D color information features of objects and their 2D color normals to enrich the interpretation of objects. Acknowledging the significance of point cloud normals [11]—the feedback of neighborhood support, or k-neighbors—which provide point positional information, we innovatively suggest the “normals” of point color information in 2D features. Our method involves using point cloud data that combines 3D positional information (XYZ) and 2D color space information (RGB) [12] as input, thereby integrating a multi-modal information point cloud structure. We apply our proposed MInet to the standard ShapeNet dataset [13] and the complex environment dataset, ThreeDMatch [14], for segmentation tasks, and the Stanford3D [15] dataset for recognition tasks.
The principal contributions of this research include:
(1) The introduction of a novel 2D image feature, RGB normals, which estimates the RGB normals at a point based on the RGB values from neighboring points. This feature functions as a rendering effect in the surrounding environment, thereby further enhancing the collection of color information from neighboring points.
(2) A novel end-to-end MInet for LiDAR point cloud recognition tasks is proposed in this study. The innovative aspect of this network is that it enhances the feature extraction of 2D data on the basis of 3D data, and integrates the features of both as multi-element information features for subsequent point cloud processing.
(3) Comprehensive comparisons investigating the feasibility of multi-modal information in point cloud processing technology, thereby demonstrating the superior performance of the proposed architecture.
In Section 2, we outline the latest domestic and international research developments, as well as the problems that this experiment aims to address. In Section 3, we primarily elaborate on the network architecture of the proposed Mlnet and its method of implementation, inclusive of a detailed explanation of the loss function utilized in this experiment. In Section 4, we undertake a comparative analysis of the proposed Mlnet and other models, scrutinizing their robustness in segmentation and recognition across various datasets. Section 5 provides a summative conclusion and future prospects. Here, we reflect on the work completed in this paper, articulate the innovative aspects of this study, evaluate the advantages and shortcomings of the proposed scheme, and indicate directions for future research endeavors.

2. Related Works

The pioneering PointNet model, brought forward by Charles R. Qi et al. at Stanford University [7], revolutionized the handling of point clouds by processing them directly. Each point in the input point cloud is assigned a corresponding spatial encoding, which are subsequently combined to create a global point cloud feature. This novel approach has not only achieved commendable results in point cloud processing, but also has made profound contributions to the broader field. Building on this, Qi went on to develop an enhanced version, PointNet++ [8], which innovatively introduces a multi-level feature extraction structure, facilitating the efficient extraction and comparison of local and global features.
Furthering the discourse in this field, Klokov and Lempitsky [16] presented a novel deep learning architecture, Kd-network, specifically tailored for 3D model recognition tasks involving unstructured point cloud data. By employing multiplicative transformations and kd-tree partitions with shared parameters, their approach deftly sidesteps the scalability concerns that plague traditional grid-based convolution architectures. In various tasks, including shape classification, retrieval, and part segmentation, the Kd-network has shown a performance competitive with established methods.
In a similar vein, Li et al. proposed PointCNN [17], an elegantly simple yet universally applicable framework for point cloud feature learning. By applying X-transformations for point reordering and weighting, followed by operations akin to convolution, PointCNN is capable of effectively managing unordered and irregular point cloud data, achieving performance that either matches or surpasses the most advanced methods across a multitude of tasks and datasets.
Addressing the potential loss of graph structural information during transitions and the amalgamation of redundant information, Bo Wu offered a solution with his Dynamic Graph CNN (DGCNN) [10]. This network expands upon the conventional CNN by integrating a preprocessing layer known as EdgeConv (DGCL). It further enhances convolution layers by implementing a backward propagation optimization process, thereby mitigating information loss in graph transformations. Similarly, to navigate the inherent constraints in point cloud semantic segmentation due to isotropy of its features, Lei Wang put forth a novel Graph Attention Convolution (GAC) [9]. This approach allows the kernel to be dynamically tailored to fit the structure of objects, enabling GACNet to extract structured features of point clouds for fine-grained segmentation while preventing inter-object feature contamination. Zhang proposed an SVM classifier based on color and shape features for target classification and recognition [18].
Despite these strides in point cloud processing, challenges persist when dealing with diverse complex environments or situations demanding higher accuracy, often requiring complex parameter tuning. This process is not only time-intensive but also involves considerable technical effort to generate desired processing results from filtered point cloud data. To address this, our study proposes a multi-information-based deep point cloud processing algorithm. The enhanced neural network model uniquely separates the 3D geometric and 2D color information of the target object from the input data, forming distinct point cloud features after processing. Empirical results indicate that the proposed multi-information processing approach demonstrates robust performance across various environmental conditions and requires minimal manual intervention. Consequently, our user-friendly and efficient method stands as a promising algorithmic model in this field.

3. Methods

This experiment capitalizes on multi-modal information for object feature extraction, and implements a novel local feature descriptor to enhance the accuracy and robustness of this extraction process. In this context, we propose a network model dubbed MInet, which accepts parallel inputs of multi-modal information. We utilize two network architectures to separately extract information from the three-dimensional and two-dimensional features of the target point cloud. In the following section, we will detail the feature extraction process for each dimension separately.

3.1. 3D Information Feature Extraction

Initially, the model ingests a dataset formatted as ( N 3 ) , where the three-dimensional coordinates ( x , y , z ) of the target point cloud are fed into the model. Once the dataset is prepared, a sampling process is conducted. For this process, we use the SG module (sample and grouping module) to sample and group data sets. We employ the Farthest Point Sampling (FPS) [19] algorithm (i.e., the sampling process in the SG module of the flowchart in Figure 1), which affords superior coverage of the entire point set compared to other sampling algorithms. A total of N points are selected as samples. To ensure the assembled sample points meet the criteria of compactness and rationality, we set N = 2048 for segmentation tasks and N = 4096 for recognition tasks. Subsequently, we group the sampled points. During this grouping process, we utilize the Ball Query method to generate N local regions (i.e., the grouping process in the SG module of the flowchart in Figure 1), from which local features are extracted. Given the possibility of encountering non-uniform point clouds during point cloud processing, we employ a multi-scale grouping strategy for local feature extraction. For each central point, we operate at three different scales by constructing three regions around each central point. Each region boasts a unique radius and a different number of points. For any given central point, three disparate scale regions are processed with three sets of MLPs and convolution kernels, specifically (32, 64, 128), (64, 128, 256), and (64, 128, 256). Following this, all the local features are amalgamated into global features, as depicted in Figure 2. The global feature extraction process and the overall three-dimensional information feature extraction process are shown in Figure 1.

3.2. Two-dimensional Information Feature Extraction

Building on the foundation of two-dimensional information, we propose the utilization of color normal vectors to encapsulate the two-dimensional color information. Color normals, serving as a specific type of normal vector, embody the influence of the surrounding colors on a point’s color information. This study also carried out the process of obtaining color normal vector for two-dimensional color features. We interpret the task of normal vector estimation as the projection of a normal vector tangent to the surface. Consequently, the problem of normal vector estimation is reformulated as a least-squares plane fitting estimation problem. Through an examination of the eigenvectors and eigenvalues of a covariance matrix (also referred to as Principal Component Analysis–PCA [20]), we simplify this problem. The covariance matrix is generated by the nearest neighbors of the query point. For each point p i , we define the covariance matrix C as follows:
C = 1 k i = 1 k ( p i p ¯ ) ( p i p ¯ ) T , C v j = λ j v j , j { 0 , 1 , 2 }
where
k is the nearest k points to p i point;
p ¯ is the center of the nearest neighbor;
λ j is the j-th eigenvalue;
v j is the j-th eigenvalue vector.
Based on the neighborhood support estimation from the ( r , g , b ) values of each point, the color normal of the point, representing the normal vector of the two-dimensional color information, is determined. The designed structure for two-dimensional feature extraction is illustrated in Figure 3. It takes a set of points with d in dimensions as input data. The input undergoes three different convolutional processes with kernel sizes of 16, 32, and 64, respectively. The results obtained from the three convolutions are then integrated, followed by the extraction of two-dimensional information features using MLP layers with three kernels of size 64, 32, and num_class. This process generates the final global feature representation. When only ( r , g , b ) information is available, d in = 3 . When both ( r , g , b ) and ( r , g , b ) color normals are considered, d in = 6 . In the flowchart shown in Figure 3, the entire process of two-dimensional information feature extraction for our experiment is presented.

3.3. Object Feature

In previous studies, two-dimensional features are often used to discriminate data processed with three-dimensional features. However, relying on a single-dimensional feature has proven to be inadequate in addressing all potential scenarios. Consequently, in this study, we propose a method that concurrently processes the target using both two-dimensional and three-dimensional features. This methodology bolsters the robustness of our experiment in terms of target segmentation and recognition. Based on this, our experiment proposes the process framework shown in Figure 4, combining two-dimensional and three-dimensional features to form unique point cloud characteristics. This lays the groundwork for subsequent point cloud processing tasks.

3.4. Loss Function and Evaluation

This model does not employ feature transformation matrices; hence, this study resorts to the cross-entropy loss function [21] for our experimental loss function. Cross-entropy serves as a measure to gauge the divergence between two distinct probability distributions over the same random variable. Within the context of machine learning, it quantifies the disparity between the true probability distribution and the predicted probability distribution. The expression for the loss function is provided as follows:
H ( p , q ) = i = 1 n p ( x i ) log ( q ( x i ) )
In order to provide a quantitative comparison and analysis of the classification results for LiDAR point clouds, we employ several evaluation metrics. These include overall accuracy (OA) [22], mean intersection over union (mIoU) [23], recall [24], precision [25], and the f1 score [26].
The metrics are defined as follows:
O A = t p + t n t p + t n + f p + f n
m I o U = t p t p + f p + f n
p r e c i s i o n = t p t p + f p
r e c a l l = t p t p + f n
f 1 - score = 2 p r e c i s i o n r e c a l l p r e c i s i o n + r e c a l l
All the experiments were implemented in the ubuntu18.04 environment using a single GPU. A total of 200 epochs were used with 0.1 momentum and 0.001 initial learning rate as training parameters. Set batch_size to 8 for hardware performance. i7-12700, 2.10 GHz processor and RTX 3080 Ti graphics card were used in the experiment. Results were evaluated by overall accuracy (OA), average crossover and association (mIoU), recall, precision, and f1 scores.

4. Result and Discussion

4.1. Data Introduction

To obtain the requisite multi-modal information, we blend two-dimensional images and three-dimensional point clouds. Rapid acquisition of three-dimensional point clouds is possible through LiDAR, while visible light cameras supply the two-dimensional images. Moreover, training datasets that sufficiently represent complex environments are necessary. Summarizing, we have chosen the ShapeNet [13] and 3DMatch [14] datasets as input data to appraise the segmentation performance of our model. ShapeNet, developed by Princeton University, is a large-scale 3D model dataset encompassing various model types, including furniture, animals, vehicles, and buildings. Each 3D model is supplemented with metadata such as classification labels, poses, and materials. Distinctly, ShapeNet segregates geometric and semantic information, permitting researchers to utilize these models conveniently for specific problems. 3DMatch is an extensive dataset collaboratively developed by researchers from ETH Zurich, Stanford, and Princeton University. Comprising pairs of 3D point clouds from three disparate sensors, it is employed to evaluate the effectiveness of various point cloud matching algorithms. This dataset finds wide applications in fields like 3D reconstruction, SLAM, map creation, and robot navigation. Moreover, we employ the Stanford [15] standard dataset to verify our model’s recognition performance. This dataset, commonly used for more demanding semantic scene segmentation tasks, consists of six indoor areas with 2.79 billion points scanned from 271 rooms across three different buildings. All points are labeled with 13 semantic categories, including board, chair, and other common objects. Rooms are subdivided into 1 m × 1 m blocks. We represent the point clouds using a nine-dimensional vector ( ( x , y , z ) , ( r , g , b ) , and normalized coordinates).

4.2. Experiment Result

4.2.1. Result of Shapenet Dataset Segmentation

In our study, we performed a rigorous evaluation of our network’s performance on the ShapeNet dataset using well-established metrics, namely, category-average Intersection over Union (mIoU) and instance mIoU, facilitating direct comparison with other state-of-the-art methods. We set the bar high, benchmarking our network against a wide range of robust models such as PointNet, PointNet++, DGCNN, RGCNN [27] and PCT [28]. The empirical results obtained by MInet and other competing models were graphically visualized to enable an intuitive and direct assessment of their performance. Table 1, provides a concise yet comprehensive snapshot of the part segmentation results produced by our network in comparison to other models. Notably, our results underscore the proficiency of our network in extracting and encapsulating local information, especially when handling multi-faceted data. This is in stark contrast to other methodologies that confine themselves to the acquisition of three-dimensional information.
To compare the various models, we evaluated the results yielded by other methodologies, and juxtaposed them against the results obtained from MInet.The visual representation of the trained data of MInet and pointnet++ is shown in Figure 5. A noteworthy observation is the visually apparent discrepancy in the airplane segmentation results in the last row. PointNet++ exhibited minor shortcomings in accurately segmenting the airplane wings, whereas our method demonstrated a distinct ability to segregate all the parts of the airplane accurately. Moreover, the inability of PointNet++ to incorporate a two-dimensional representation of the objects resulted in an indistinct separation of the backrest and legs of the chair in the chair segmentation scenario presented in the second row. In stark contrast, our method leverages a synergistic combination of two-dimensional and three-dimensional information to seize the most critical features of the chair, yielding accurate segmentation results. The fidelity of our results, the true value of the data entered, coupled with a thorough comparison of the visualization results, unequivocally substantiates the superior accuracy and robustness of our network. This claim is further reinforced visually in Table 1, which exhibits the high mIoU value of 89.3% achieved by our network model, thereby further showcasing its efficacy and precision.

4.2.2. Result of 3DMatch Dataset Segmentation

For 3DMatch data in a complex environment, we added the normal vector of two-dimensional color information into the input data, and conducted training tests on MInet, pointnet++, pointnet, dgcnn and gacnet using the specified training parameters on the 3DMatch data set.
Table 2 presents the segmentation outcomes for different networks on the 3DMatch dataset. An analysis of the training and evaluation of several methods—MInet, PointNet, PointNet++, DGCNN, and GACNet—suggests that the MInet approach proposed in our study boasts the highest level of robustness. Notably, its accuracy demonstrates an enhancement of 18.8% and 17.9% in comparison with PointNet++ and DGCNN, respectively. Figure 6 visualizes the comparison between the training values generated by MInet and the ground truth values, revealing superior performance of our trained model across various complex environments, such as living rooms, staircases, and bedrooms. This superiority is predominantly attributed to MInet’s focus on the multi-faceted information of the target object. In complex settings, a singular focus on three-dimensional geometric data is insufficient to fully represent the target object. However, by incorporating multi-informational aspects, such as color data and color normals, we can extract more specific and detailed target features.
During the testing phase, we observed that our model performed best in environments with a wide range of colors, and it maintains high robustness even in environments with minimal color variations. Table 3 and Table 4, respectively, offer examples of four types of targets within environments featuring larger and smaller color spans, reinforcing the versatility and adaptability of our approach.

4.2.3. Result of Stanford 3D Dataset Recognition

In our approach to model training, we strategically extracted 4096 points at random from each block. Notably, the test points were spared from any sampling process. We deployed a comprehensive six-fold cross-validation, gauging the efficacy of our model based on the average mIoU results and the Overall Accuracy (OA). Our methodology faithfully followed the same protocols of random scaling, rotation, and point expansion. The visualizations demonstrating the Stanford 3D test images can be viewed in Figure 7.
In Figure 7, we show a sharp contrast between MInet and PointNet++, which can highlight the advantages of the model proposed in this experiment in comparison to the actual situation. Careful examination of the selected examples—roof, podium and table—shows that our MInet is superior to existing models today. This advantage is clearly demonstrated in the roof recognition example depicted in the first row, where PointNet++ falls short in identifying the lights and four gray bricks on the roof. On the contrary, our model offers a distinct recognition result, primarily due to its enhanced sensitivity to variations in color regions. Moreover, in the bottom row that showcases table recognition, our model exhibits a sophisticated understanding of the varied colors of the tablecloth, while PointNet++ lags slightly behind in this regard. The superior performance of our proposed MInet is visually corroborated by the data presented in Table 5, where it outshines other models in terms of both mIoU and OA, registering scores of 57.2% and 87.1%, respectively. This underscores the effectiveness of our multi-information approach in handling complex point cloud recognition tasks.

5. Conclusions

In this research, we introduced a novel model named MInet, which assimilates multi-modal information extraction techniques based on a multi-spectral LiDAR system. This model is designed for the execution of point cloud segmentation and recognition tasks within diverse environments. MInet can effectively capture the local geometric relationship between the points obtained from the non-uniformly sampled data, and enhance the extracted features by loading multiple information and utilizing the two-dimensional and three-dimensional characteristics of the target object. The evaluation results elucidate that our MInet attains a mIoU of 0.893 on the ShapeNet dataset, while demonstrating recognition performance metrics of 0.855 OA, 0.922 mIoU on the 3DMatch dataset. Furthermore, the experimental outcomes showcase MInet’s superiority over other models in recognition tasks, in the semantic segmentation results of the Stanford 3D dataset, 0.572 of mIoU and 0.871 of OA were obtained.
MInet possesses the ability to extract abundant two-dimensional information from images while maintaining three-dimensional coordinate data, considerably facilitating semantic segmentation. This model introduces an innovative approach of utilizing multi-modal information to process targets, employing both two-dimensional and three-dimensional features. This breaks away from the prevailing practice of using two-dimensional features solely as a discriminator for three-dimensional ones [31]. In practical applications, most data collection and processing efforts employ multiple sensors. Our Mlnet, through the use of multi-modal information for data processing, aligns well with multi-sensor data collection and further enhances the accuracy of segmentation and recognition. Looking ahead, our objective is to extend our work to larger datasets and a wider variety of scenarios. This endeavor is aimed at ensuring that our methods are more comprehensive and efficient in real-world applications.

Author Contributions

Conceptualization, Y.W. and Z.D.; methodology, Y.W.; software, Y.W.; validation, Y.W., Z.D. and Y.Z.; formal analysis, Y.W.; investigation, T.L.; resources, Y.W.; data curation, Y.W.; writing—original draft preparation, Y.W.; writing—review and editing, Z.D.; visualization, X.S.; supervision, J.W.; project administration, Y.Z.; funding acquisition, X.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, Z.; Cai, Y.; Wang, H.; Chen, L.; Gao, H.; Jia, Y.; Li, Y. Robust target recognition and tracking of self-driving cars with radar and camera information fusion under severe weather conditions. IEEE Trans. Intell. Transp. Syst. 2021, 23, 6640–6653. [Google Scholar] [CrossRef]
  2. Jiang, J.; Liu, D.; Gu, J.; Süsstrunk, S. What is the space of spectral sensitivity functions for digital color cameras? In Proceedings of the 2013 IEEE Workshop on Applications of Computer Vision (WACV), Clearwater Beach, FL, USA, 15–17 January 2013; pp. 168–179. [Google Scholar]
  3. Villa, F.; Severini, F.; Madonini, F.; Zappa, F. SPADs and SiPMs arrays for long-range high-speed light detection and ranging (LiDAR). Sensors 2021, 21, 3839. [Google Scholar] [CrossRef] [PubMed]
  4. Cheng, L.; Chen, S.; Liu, X.; Xu, H.; Wu, Y.; Li, M.; Chen, Y. Registration of laser scanning point clouds: A review. Sensors 2018, 18, 1641. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Schumann, O.; Hahn, M.; Dickmann, J.; Wöhler, C. Semantic segmentation on radar point clouds. In Proceedings of the 2018 21st International Conference on Information Fusion (FUSION), Cambridge, UK, 10–13 July 2018; pp. 2179–2186. [Google Scholar]
  6. Ilci, V.; Toth, C. High definition 3D map creation using GNSS/IMU/LiDAR sensor integration to support autonomous vehicle navigation. Sensors 2020, 20, 899. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  8. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  9. Wang, L.; Huang, Y.; Hou, Y.; Zhang, S.; Shan, J. Graph attention convolution for point cloud semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 10296–10305. [Google Scholar]
  10. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef] [Green Version]
  11. OuYang, D.; Feng, H.Y. On the normal vector estimation for point cloud data from smooth surfaces. Comput.-Aided Des. 2005, 37, 1071–1079. [Google Scholar] [CrossRef]
  12. Bala, R.; Sharma, G.; Monga, V.; Van de Capelle, J.P. Two-dimensional transforms for device color correction and calibration. IEEE Trans. Image Process. 2005, 14, 1172–1186. [Google Scholar] [CrossRef] [PubMed]
  13. Chang, A.X.; Funkhouser, T.; Guibas, L.; Hanrahan, P.; Huang, Q.; Li, Z.; Savarese, S.; Savva, M.; Song, S.; Su, H.; et al. Shapenet: An information-rich 3d model repository. arXiv 2015, arXiv:1512.03012. [Google Scholar]
  14. Zeng, A.; Song, S.; Nießner, M.; Fisher, M.; Xiao, J.; Funkhouser, T. 3dmatch: Learning local geometric descriptors from rgb-d reconstructions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1802–1811. [Google Scholar]
  15. Luo, Y.; Mi, Z.; Tao, W. DeepDT: Learning geometry from Delaunay triangulation for surface reconstruction. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 2277–2285. [Google Scholar]
  16. Klokov, R.; Lempitsky, V. Escape from cells: Deep kd-networks for the recognition of 3d point cloud models. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 863–872. [Google Scholar]
  17. Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. Pointcnn: Convolution on x-transformed points. arXiv 2018, arXiv:1801.07791. [Google Scholar]
  18. Zhang, C.; Zhang, K.; Ge, L.; Zou, K.; Wang, S.; Zhang, J.; Li, W. A method for organs classification and fruit counting on pomegranate trees based on multi-features fusion and support vector machine by 3D point cloud. Sci. Hortic. 2021, 278, 109791. [Google Scholar] [CrossRef]
  19. Eldar, Y.; Lindenbaum, M.; Porat, M.; Zeevi, Y.Y. The farthest point strategy for progressive image sampling. IEEE Trans. Image Process. 1997, 6, 1305–1315. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Bryant, F.B.; Yarnold, P.R. Principal-Components Analysis and Exploratory and Confirmatory Factor Analysis; American Psychological Association: Washington, DC, USA, 1995. [Google Scholar]
  21. Pang, T.; Xu, K.; Dong, Y.; Du, C.; Chen, N.; Zhu, J. Rethinking softmax cross-entropy loss for adversarial robustness. arXiv 2019, arXiv:1905.10626. [Google Scholar]
  22. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  23. Turpin, A.; Scholer, F. User performance versus precision measures for simple search tasks. In Proceedings of the 29th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Seattle, WA, USA, 6–11 August 2006; pp. 11–18. [Google Scholar]
  24. Wang, D.Z.; Posner, I. Voting for voting in online point cloud object detection. In Proceedings of the Robotics: Science and Systems, Rome, Italy, 13–17 July 2015; Volume 1, pp. 10–15. [Google Scholar]
  25. Quan, S.; Yang, J. Compatibility-guided sampling consensus for 3-d point cloud registration. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7380–7392. [Google Scholar] [CrossRef]
  26. Guo, Z.; Du, S.; Li, M.; Zhao, W. Exploring GIS knowledge to improve building extraction and change detection from VHR imagery in urban areas. Int. J. Image Data Fusion 2016, 7, 42–62. [Google Scholar] [CrossRef]
  27. Te, G.; Hu, W.; Zheng, A.; Guo, Z. Rgcnn: Regularized graph cnn for point cloud segmentation. In Proceedings of the 26th ACM International Conference on Multimedia, Seoul, Republic of Korea, 22–26 October 2018; pp. 746–754. [Google Scholar]
  28. Guo, M.H.; Cai, J.X.; Liu, Z.N.; Mu, T.J.; Martin, R.R.; Hu, S.M. Pct: Point cloud transformer. Comput. Vis. Media 2021, 7, 187–199. [Google Scholar] [CrossRef]
  29. Engelmann, F.; Kontogianni, T.; Hermans, A.; Leibe, B. Exploring spatial context for 3D semantic segmentation of point clouds. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 716–724. [Google Scholar]
  30. Jing, W.; Zhang, W.; Li, L.; Di, D.; Chen, G.; Wang, J. AGNet: An attention-based graph network for point cloud classification and segmentation. Remote Sens. 2022, 14, 1036. [Google Scholar] [CrossRef]
  31. Endres, F.; Hess, J.; Sturm, J.; Cremers, D.; Burgard, W. 3-D mapping with an RGB-D camera. IEEE Trans. Robot. 2013, 30, 177–187. [Google Scholar] [CrossRef]
Figure 1. Three-dimensional feature extraction diagram. The SG module is FPS sample and grouping. (For three different scale regions of the same central point, each of the three sets of MLP, the convolution kernel is (32, 64, 128), (64, 128, 256), and (64, 128, 256), and then the three-dimensional features are obtained by concat operation).
Figure 1. Three-dimensional feature extraction diagram. The SG module is FPS sample and grouping. (For three different scale regions of the same central point, each of the three sets of MLP, the convolution kernel is (32, 64, 128), (64, 128, 256), and (64, 128, 256), and then the three-dimensional features are obtained by concat operation).
Sensors 23 06327 g001
Figure 2. Local feature extraction (The local features are obtained by extracting features from three different scales).
Figure 2. Local feature extraction (The local features are obtained by extracting features from three different scales).
Sensors 23 06327 g002
Figure 3. Two-dimensional feature extraction diagram (Three different convolution processes were carried out on the input, and the convolution kernels were 16, 32, and 64, respectively. After concat operation was carried out on the results obtained from the convolution processing, three MLP layers with convolution kernels 64, 32, num_class were used to extract two-dimensional information features and obtain the final two-dimensional features).
Figure 3. Two-dimensional feature extraction diagram (Three different convolution processes were carried out on the input, and the convolution kernels were 16, 32, and 64, respectively. After concat operation was carried out on the results obtained from the convolution processing, three MLP layers with convolution kernels 64, 32, num_class were used to extract two-dimensional information features and obtain the final two-dimensional features).
Sensors 23 06327 g003
Figure 4. Object feature extraction (The multivariate features of the target are obtained through concat and MLP layers).
Figure 4. Object feature extraction (The multivariate features of the target are obtained through concat and MLP layers).
Sensors 23 06327 g004
Figure 5. Visualization of segmentation results based on ShapeNet.
Figure 5. Visualization of segmentation results based on ShapeNet.
Sensors 23 06327 g005
Figure 6. Visualization of segmentation results based on 3Dmatch.
Figure 6. Visualization of segmentation results based on 3Dmatch.
Sensors 23 06327 g006
Figure 7. Visualization of different methods on S3DIS, from left to right: True value, MInet, Pointnet++.
Figure 7. Visualization of different methods on S3DIS, from left to right: True value, MInet, Pointnet++.
Sensors 23 06327 g007
Table 1. Segmentation results based on ShapeNet data (%).
Table 1. Segmentation results based on ShapeNet data (%).
ModelMiouPlaneBagCapCarChairEar PhoneGuitarKnifeLampLaptopMotor BikeMugPistolRocketSkate BoardTable
shapes269076558983758697873921547451202184283661525271
Pointnet82.582.277.681.373.988.372.090.284.779.794.064.391.780.157.171.879.5
Point net++84.581.576.187.576.989.974.090.585.683.695.368.294.681.460.674.682.2
RGCNN83.879.882.392.174.888.773.290.887.982.895.563.495.260.444.172.479.9
DGCNN84.783.582.986.277.390.174.390.787.082.395.265.894.480.663.074.085.1
PCT85.882.877.488.878.291.275.391.886.984.996.669.595.982.761.975.983.5
MInet89.393.987.692.697.195.673.197.792.491.697.688.395.788.171.479.479.7
Table 2. Parameter comparison of training 3DMatch data of different models (%).
Table 2. Parameter comparison of training 3DMatch data of different models (%).
ModelAccuracymIoURecallPrecisionf1-Score
Pointnet71.681.480.481.576.7
Pointnet++72.080.576.778.579.2
Gacnet72.585.777.982.678.7
Dgcnn72.483.373.973.973.9
MInet85.592.289.189.189.0
Table 3. The mIoU values of four types of objects in the environment with large color span (%).
Table 3. The mIoU values of four types of objects in the environment with large color span (%).
BDapt2Browncs2ChairsBDcopyroom
Pointnet++71.199.688.093.5
Gacnet68.296.783.192.9
Dgcnn75.097.580.092.9
MInet99.899.697.897.9
Table 4. The mIoU values of four types of objects in the environment with small color span (%).
Table 4. The mIoU values of four types of objects in the environment with small color span (%).
RedikitchenPumpkin-OffOfficeBDoffice1
Pointnet++71.673.979.583.4
Gacnet67.765.468.370.4
Dgcnn66.067.165.971.4
MInet80.776.281.885.7
Table 5. Comparison results of model recognition.
Table 5. Comparison results of model recognition.
ModelmIoU (%)OA (%)
Pointnet42.778.1
Pointnet++52.382.2
Dgcnn54.883.3
G+RCU [29]44.680.7
AGnet [30]58.285.1
MInet57.287.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Zuo, Y.; Du, Z.; Song, X.; Luo, T.; Hong, X.; Wu, J. MInet: A Novel Network Model for Point Cloud Processing by Integrating Multi-Modal Information. Sensors 2023, 23, 6327. https://doi.org/10.3390/s23146327

AMA Style

Wang Y, Zuo Y, Du Z, Song X, Luo T, Hong X, Wu J. MInet: A Novel Network Model for Point Cloud Processing by Integrating Multi-Modal Information. Sensors. 2023; 23(14):6327. https://doi.org/10.3390/s23146327

Chicago/Turabian Style

Wang, Yuhao, Yong Zuo, Zhihua Du, Xiaohan Song, Tian Luo, Xiaobin Hong, and Jian Wu. 2023. "MInet: A Novel Network Model for Point Cloud Processing by Integrating Multi-Modal Information" Sensors 23, no. 14: 6327. https://doi.org/10.3390/s23146327

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop