Next Article in Journal
Estimates of Crop Yield Anomalies for 2022 in Ukraine Based on Copernicus Sentinel-1, Sentinel-3 Satellite Data, and ERA-5 Agrometeorological Indicators
Previous Article in Journal
Model Predictive Control for Speed-Dependent Active Suspension System with Road Preview Information
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Lunar Lineament Extraction Method Based on Improved UNet++ and YOLOv5

1
Faculty of Innovation Engineering, Macau University of Science and Technology, Avenida Wai Long, Taipa, Macau, China
2
The State Key Laboratory of Lunar and Planetary Sciences, Macau University of Science and Technology, Avenida Wai Long, Taipa, Macau, China
3
The Department of Intelligent Manufacturing, Wuyi University, Jiangmen 529020, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(7), 2256; https://doi.org/10.3390/s24072256
Submission received: 24 February 2024 / Revised: 26 March 2024 / Accepted: 28 March 2024 / Published: 1 April 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Lineament is a unique geological structure. The study of Lunar lineament structure has great significance on understanding its history and evolution of Lunar surface. However, the existing geographic feature extraction methods are not suitable for the extraction of Lunar lineament structure. In this paper, a new lineament extraction method is proposed based on improved-UNet++ and YOLOv5. Firstly, new lineament dataset is created containing lineaments structure based on CCD data from LROC. At same time the residual blocks are replaced with the VGG blocks in the down sample part of the UNet++ with adding the attention block between each layer. Secondly, the improved-UNet++ and YOLO networks are trained to execute the object detection and semantic segmentation of lineament structure respectively. Finally, a polygon-match strategy is proposed to combine the results of object detection and semantic segmentation. The experiment result indicate that this new method has relatively better and more stable performance compared with current mainstream networks and the original UNet++ network in the instance segmentation of lineament structure. Additionally, the polygon-match strategy is able to perform preciser edge detail in the instance segmentation of lineament structure result.

1. Introduction

The word lineaments was originally created in the early 20th century, referring to the significant tectonic units in the crust with the formation of minierals, active faults, underground water controls, earthquakes and geomorphology on the earth [1]. As the development of deep space exploration, the academia also found that lineament is one of the common geological structures on the surface of terrestrial planets and their satellites in solar system. Although it has imaging similarity with the lineaments on the earth, the formation mechanism of lineaments on the surface of terrestrial planets and their satellites in solar system is different from that on the earth. Most often, this term refers to linear features on the surfaces of other planets. Lineament structures are usually considered to be closely associated with volcanic activity on the surfaces of terrestrial planets or their satellites. Based on its imaging features and cause mechanism, the lineaments structure contains following main types: ridges, rill, faults, vaills, catena and fractures. Ridges is a complex lineament structure that extend longer and have lower altitudes on the lunar plain. The widely accepted view is that wrinkle ridges are formed by volcanic erosion and tectonic action [2]. Rill and fault are different manifestations under the same causative mechanism [3,4]. These types of lineament structure unusually are associated with lava activity on the Lunar surface [5,6]. Catena is a chain structure consists of regular arrangement and combination of pits in certain interval. It’s usually in a nearly straight line shape. The causes and mechanisms of catena are diverse including surface ring collapsing [7], magma activity under the moon’s surface [8], and meteorite impact [9], they are classified based on the consistent in imaging.
Lunar surface structures are of great significance for understanding and reconstructing the evolution of the lunar geological structure, and the morphology and distribution characteristics of linear structures such as lunar ridges and creeks are closely related to the dynamical tectonic movements in the Moon. The limited number of samples and the difficulty of field surveys make remote sensing the most important means of planetary science research in this area [10]. On the other hand, within the academic community, attention has mainly been focused on the construction of a classification system for linear structures and the study of the causes of different classes of linear structures based on their dynamical mechanisms [11]. Based on the current situation, we designed a new lineament extraction method based on improved-UNet++ and YOLOv5, which aims to improve the detection accuracy and efficiency of lineaments structures on the Lunar surface based on present samples. The main contributions of this paper are as follows:
  • A new Lunar lineaments dataset, which is one of the few datasets of its kind, is created. This dataset contains lineament structures based on Charge-Coupled Device (CCD) data from the Lunar Reconnaissance Orbiter Camera (LROC). With a total of 1000 manually collected Lunar lineaments samples, this dataset provides a unique and valuable resource for lunar research. The scarcity of such datasets is due to the complexity and time-consuming nature of the data collection and processing.
  • A new lineament extraction method based on improved-UNet++ and YOLOv5 is proposed. It is able to extract lineament structure with relatively better and more stable performance compared with current mainstream networks and the original UNet++ network. The experiment results show that our method increase the accuracy from 0.64 to 0.67, precision from 0.432 to 0.679, maAP@50 from 0.58 to 0.60 and maAP@50:95 from 0.34 to 0.38, IoU from 0.58 to 0.69, mean pixel accuracy from 0.92 to 0.94.
  • A polygon-match strategy is proposed to extract lineament structure with perciser edge detail. It is able to perform preciser edge detail in the instance segmentation of lineament structure result.
The rest of this paper is organized as follows. Section 2 introdues the related work. Section 3 describes the proposed method. Section 4 presents the experimental results. Section 5 concludes the paper.

2. Related Work

The early studies of lineament structure classification and identification on the earth show that due to the complexity of atmospheric and surface conditions it usually relies on multiple sources of data including Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER), Shuttle Radar Topography Mission (SRTM), Digital Elevation Model (DEM), Synthetic-aperture radar (SAR), combining with variety of technic including manual semi-automated and automated algorithms [12]. As a result, single digital image processing might not be competent to produce accurate detection result. The situation is less complicated on the Lunar surface. The Lunar surface is lack of atmosphere and water, it is easier for lunar exploration satellite to collect clear DEM and CCD data [13,14]. The existing lineament structure extraction methods on Lunar surface are mainly based on manual interpretation [5,6,15] and semi-automated algorithms. The former method is usually time-consuming and laborious. It relies heavily on experience and judgment. Often this experience requires a very deep understanding of geology and planetary dynamics. In most situation, this identification is difficult to popularize. The latter category of algorithms is usually based on the characteristics of lineament structure, such as the linear shape, the difference in brightness and the difference in texture. The existing algorithms are mainly based on the above characteristics to extract lineament structure. In early One of the ridge extraction algorithms is based on combination of series digital image processing methods [16]. Due to the specificity of functional design, these semi-automatic algorithms usually perform poorly on some images with complex geological information. This type of lineament extraction method is usually restricted by its specialization ability. It usually has relatively worse performance when dealing with multiple types of landform in same region. Little research have been done to improve the performance of Lunar lineament extraction algorithms. Our previous work focus on classification of lineament structure [17], we did not go further to extract the lineament structure.
The deep learning based method has been widely used in the field of image processing feild [18,19,20]. Especially the YOLO and UNet series networks, it has achieved great success in the field of image classification, object detection and instance segmentation. It has been proved that convolutional neural networks have satisfied result on crater detection based on DEM and CCD remote sensing data. For example, the FasterRCNN [21] has already been applied in object detection with higher efficiency compared with previous work, MSA-YOLO [22] is desgined to object detection in DIOR dataset, RSI-YOLOv5 [23] is applied in remote sensing dataset such as DOTA and NWPU-VHR with outperform results, and a Graph Neural Network (GNN) [24] is desgined to execute the semantic segmentation in satellite image data. It can be concluded that deep learning method has great potential in remote sensing data processing. However, the deep learning methods are mainly applied in crater detection. Very limited attention is attracted to lineament structure identification. Previous work shows that it already made some progress on earth lineament structure [12,25]. As the development of deep space exploration, the Lunar surface exploration satellite has collected relatively considerable amount of high-resolution CCD data. The existing application is able to classify and detect lineament structure from Mars and Moon [17,26]. However, present deep-learning based methods are mainly focusing on crater detection [27,28,29], very few research are related with extraction of lineament structure on the Lunar surface based on deep learning networks due to the lack of such datasets.
In this paper, we subtly navigate through this research gap by proposing an efficient and accurate lineament extraction method based on YOLOv5 and improved-UNet++. This method, specifically designed for the task of lineament structure extraction, leverages the power of YOLOv5’s object detection capabilities and the improved UNet++’s superior performance in semantic segmentation tasks. To support our approach, we have developed a lineament dataset, consisting of high-resolution Lunar surface images with annotated lineament structures. This dataset serves as a valuable resource for training and validating our model. Our experimental results show promising performance of our method, making it a practical tool for processing large amounts of Lunar surface data. These results subtly hint at the untapped potential of deep learning for this complex task, opening up new possibilities for future research in Lunar surface exploration.

3. Dataset and Methodology

3.1. Dataset Construction

In this experiment, we choose the lineament structure ridge, rills, catena and valles as the content of the dataset. These four types of lineament structure occupied the majority of the known lineament structure on the Lunar surface shown below (Figure 1).
As shown in Figure 2, in this experiment, we manually labeled the edge of the mentioned lineament structures based on the Lunar Orbital Data Explorer [31]. The manual results are converted into normalized polygon coordinate points through our pre-process program. Those points can be easily converted into object detection labels and binarized mask for semantic segmentation. 1000 lineament samples in total are collected containing all types of mentioned lineament structure for the experiment dataset. Based on the 8:1:1 ratio for training, validating and testing datasets, 800 samples for training set, 100 samples for validation set, 100 samples for test set respectively. The training set and validation set are used to train the improved-UNet++ and YOLOv5 network. The test set is used to evaluate the performance of the proposed method and other instance segmentation algorithms.

3.2. Diagram of Algorithm

In this section, we will introduce the proposed algorithm in detail. In the first part, a general view will be given to the proposed algorithm. It will show the lineament extraction algorithm based on instance segmentation contributed by improved-UNet++, YOLOv5, and polygon-match strategy. The following parts will introduce the dataset construction, related network architectures and polygon-match strategy algorithm in detail.
The diagram of algorithm is shown in Figure 3. The proposed method mainly consists of three parts: dataset construction network architecture and polygon-match strategy algorithm. The dataset construction part is mainly responsible for the construction of lineament dataset suitable for object detection and semantic segmentation. It consists 2 primary parts: manual labeled process enables to generate the contour of lineaments in json form, dataset pre-process is able to convert previous result into text file form for object detection and binarized mask for semantic segmentation. The network architecture part is mainly responsible for the training of improved-UNet++ and YOLOv5 network. The ’polygon-match’ strategy algorithm part is mainly responsible for the combination of the results of object detection and semantic segmentation. This part is allowed to load trained models from previous part. The final lineament extraction result is generated based on the combination of the results of object detection and semantic segmentation with the correction of polygon-match strategy algorithm.

3.3. Network Architecture

3.3.1. YOLOv5 Network Architecture

YOLOv5 [32] network is one of the most popular target detection algorithm. Its speed and accuracy outperform its seniors and peers based on its architecture. YOLOv5 consists of three parts: backbone, neck and head. The backbone is usually a convolutional neural network, responsible for extraction of image features. The neck is usually a combination of convolutional layers and upsample layers, responsible for the combination of features extracted by the backbone. The head is usually a combination of convolutional layers and detection layers responsible for the detection of targets. Concat module is used to combine the features of different scales in neck and head, which allowed YOLOv5 to stitch together feature maps at different scales to be spliced. It allows the model to detect targets at different scales simultaneously. The YOLOv5 network architecture is shown in Figure 4. The YOLOv5 network is able to extract features of different scales and combine them to achieve better detection results. It introduces SPPF and C3 into the network architecture innovatively. The SPPF shown in Figure 5 is able to extract features of different scales. The C3 shown in Figure 5 is able to improve the stability and generalization ability of the model. Furthermore, the detect layer is applied as the output layer, it is able to normalize sigmoid output and convert it into bounding boxes location. Also, it engaged Non-Maximum Suppression(NMS), which allowed YOLOv5 to eliminate overlapping bounding boxes. Based on its net engaged modules, the YOLOv5 network is able to achieve better detection results with fewer parameters and faster speed compared with YOLOv4 [33] and YOLOv3 [34].

3.3.2. Improved-UNet++ and UNet++ Network Architecture

The UNet++ [35] network is one of the most popular semantic segmentation algorithm applied in medical imaging field. In this experiment, we tried to improve the performance of UNet++ by replace the residual blocks with VGG blocks and adding attention blocks in downsampling layers. This type of layer is usually considered as the encoder responsible for lowering the calculation complexity and extracting the distinctive features. The upsampling layers are usually considered as the decoder responsible for the combination of features extracted by the encoder and the recovery of the original image size. Compared with the VGG block, it can be found that a ’shortcut’ path is added in the residual block, which is the key to improve the stability of the network [36]. It has great performance on avoiding the gradient disappearance problem caused by the increase of network depth based on its structure shown in Figure 6 and Figure 7. It follows principles shown below, where x l is the input of the lth layer, F is the residual block, w F is the weight of the residual block, L is the total number of layers in the network, ε is the loss. This type of linear connection between layers guarantees that the loss can be always none-zero value.
x l + 1 = F ( x l , w F ) + x l
x L = x l + i = l L 1 F ( x i , w F )
𝜕 ε 𝜕 x l = 𝜕 ε 𝜕 x L ( 1 + 𝜕 𝜕 x l i = l L 1 F ( x i , w i ) )
Furthermore, an attention block CBAM(Convolutional block attention module) [37] is added between each layer in the downsampling part of the network, based on its lightweight and versatile features. The attention block is usually considered as a module to improve the network’s ability to extract features. This type of attention model is a combination of channel and spatial. It is implemented based on the following process shown in Figure 8. The output result from previous convolutional layer will first pass through a channel attention module to calculate the weighted result, and then pass through a spatial attention module, and finally the weighted result will be calculated. This type of attention model block is designed based on mathematic principles below. 7 × 7 is the size of convolutional core in spatial attention block. It shows that 7 × 7 has better performance compared with general 3 × 3 core [37].
M c h a n n e l b l o c k ( F ) = σ ( M L P ( A v g P o o l ( F ) ) + M L P ( M a x P o o l ( F ) ) )
= σ ( W 1 ( W 0 ( F a v g c ) ) + W 1 ( W 0 ( F m a x c ) ) )
M s p a t i a l b l o c k ( F ) = σ ( f 7 x 7 ( [ A v g P o o l ( F ) ; M a x P o o l ( F ) ] ) )
= σ ( f 7 x 7 ( F a v g s ; F m a x s ) )

3.4. Polygon-Match Strategy Algorithm

The polygon-match strategy algorithm is mainly responsible for matching the semantic segmentation result with object detection result. The semantic segmentation developed by UNet series networks is based on classification of each pixel with relatively weak understanding of the overall situation. This character usually lead to some random noise on the binarized mask output shown Figure 9 below. From the binarized mask, it can be seen that the binarized mask is not able to accurately outline the edge of the lineament structure, shown in the visualized result. Faced with this situation, we designed the polygon-match strategy algorithm to select the most proper semantic segmentation result based on the object detection result (bounding boxes) from pre-trained YOLOv5 network. This strategy aims to convert the binarized mask result into a list of polygons. For each image, the polygon-match contains following steps:
The polygon-match is designed based on the excellent performance of YOLOv5 in object detection. As shown in Algorithm 1, it is able to detect the area containing lineament structure with high possibility and accuracy. At same time, it is also able to guarantee that each bounding box contains only one lineament structure unit in most situation. As a result, the polygon-match strategy algorithm is able to perform preciser edge detail in the instance segmentation of lineament structure result.
Algorithm 1 Polygon-match Algorithm
1:
while there are bounding boxes do
2:
   for each polygon in the list do
3:
       calculate the area of the polygon
4:
       for each bounding box do
5:
          calculate the IoU of the polygon and the bounding box
6:
          if the polygon is completely contained in the bounding box then
7:
              record the maximum area of the polygon in the bounding box
8:
          end if
9:
          find the polygon with the largest IoU to the bounding box
10:
          record the polygon with larger area between the largest polygon contained in the box and the polygon with the largest IoU
11:
       end for
12:
   end for
13:
   append the selected polygon into a list
14:
end while
15:
return the bounding box list and the selected polygon

4. Experiment Results

4.1. Experiment Environment Configuration

This experiment realized based on following hardware and software settings:
  • GPU: A4000 x4, RTX3080 x1, RTX3080Ti x1
  • CPU: Intel(R) Xeon(R) Gold 6254 CPU @ 3.10 GHz x1,
    Intel(R) Core 10-10900k CPU @ 3.70GHz x2
  • Memory: 16 GB DDR4 3200 MHz x4, 32 GB DDR4 3200 MHz x2
  • PyTorch: 2.0.1
  • CUDA: 12.3
  • Batch Size: 16 (for YOLO networks), 4 (for UNet++ networks)
  • Epoch: 300
  • Learning Rate: 0.001
  • Optimizer: SGD (for YOLO networks), Adam (for UNet++ networks)

4.2. Evaluation Metrics

In this experiment, we use the following metrics to evaluate the performance of the proposed method and other related algorithms objectively:
  • FLOPS: (Floating Point Operations Per Second): Usually, this metric is used to measure the complexity of a model and its time consumption. It is calculated based on the number of floating-point operations in the network and the time consumed by the network.
  • IoU (Intersection over Union): IoU is usually used to measure the accuracy of the model. It is calculated based on the intersection area and the union area of the prediction and the ground truth.
  • Recall: The ratio of the number of correctly predicted positive samples to the total number of positive samples in the dataset. It is usually used to measure the ability of the model to detect positive samples.
  • Precision: The ratio of the number of correctly predicted positive samples to the total number of positive samples predicted by the model. It is usually used to measure the ability of the model to detect positive samples.
  • mAP50: The average precision of the model when the IoU is greater than 0.5. mAP is the average of the average precision (AP) of all categories. For each category, AP is the average of precision at different recall levels. It is usually used to measure the accuracy of the model.
  • mAP@50:95: The average precision of the model when the IoU is greater than 0.5 and less than 0.95. It is usually used to measure the accuracy of the model.

4.3. Experiment Results

In this experiment, in order to ensure the accuracy of the training model and the true generalization ability, we randomly separated the train dataset validation dataset and test dataset with the ration 8:1:1 for 5 times. The average results of 5 times are used for evaluation metrics calculation. Furthermore, in order to verify the effectiveness of our improvement for our UNet++,we desgined an ablation experiment. We trained the original UNet++, orgininal UNet++ with CBAM, original UNet++ with replacing the residual blocks in downsampling layers and our improved-UNet++ with the same data set partitioning and training strategy. The experiment results are shown in Table 1. It can be seen that the improved-UNet++ has relatively better and more stable performance compared with the original UNet++ network in the instance segmentation of lineament structure. At same time, the original UNet++ with CBAM has relatively worse performance compared with the original UNet++ network. It shows that the CBAM is not suitable for the instance segmentation of lineament structure. However, the combination of combination of CBAM and replacing the residual blocks in downsampling layers is able to improve the performance of the original UNet++ network.
Additionally, the polygon-match strategy is able to perform preciser edge detail in the instance segmentation of lineament structure result. The result of the polygon-match strategy algorithm is shown in Figure 9. It is able to improve the IoU from 0.37 to 0.6601, based on our experiment result.
For the comparison of the proposed method and other related algorithms, we trained FCN [38], Faster-RCNN [39], SegNet [40], Deeplabv3 [41] YOLOv5-seg [42] and YOLOv8-seg [43] with the same strategy and evaluate metrics calculation method. The experiment results are shown in Table 2. It can be seen that the proposed method has relatively better and more stable performance compared with other related algorithms in the instance segmentation of lineament structure. It has the relatively higher value of mAP@50 and mAP@50:95. In most situation, if a model has a high mAP in any IoU threshold, it means that the model performs very well in object detection or semantic segmentation tasks, being able to accurately detect the presence of objects and having good detection performance across various categories of objects, shown in Figure 10.
The experiment results show that our method have relatively better and more stable performance compared with other related algorithms in the instance segmentation of lineament structure. It can be seen that our method have the relatively higher value of mAP@50, mAP@50:95 mIoU and mean pixel accuracy in all tested thresholds accroding to Table 2 and Figure 11. At same time, it also shows that our ’polygon-match’ strategy algorithm is able to perform preciser edge detail in the instance segmentation of lineament structure result accroding to Figure 9.

5. Conclusions

A new Lunar lineaments structure landform extraction method is propoesed based on improved-UNet++ and YOLOv5 with polygon-match strategy algorithm. The method introduces a possible solution for the instance segmentation of Lunar lineament structure on the Lunar surface. First, a reliable Lunar lineament structure dataset is established based on the Lunar Orbital Data Explorer. Second, the improved-UNet++ and YOLOv5 networks are trained based on our new constructed dataset. Third, the polygon-match strategy algorithm is designed to improve the accuracy of the instance segmentation of lineament structure. Finally, the proposed method is compared with other related algorithms. The experiment results show that the proposed method has relatively better and more stable performance compared with other related algorithms in the instance segmentation of lineament structure. The proposed method has the relatively higher value of mAP@50 and mAP@50:95 compared with other related algorithms in the instance segmentation of lineament structure. The proposed method is able to extract Lunar lineament structure with preciser edge detail. It is able to provide a new solution for the instance segmentation of Lunar lineament structure on the Lunar surface. The proposed method has the potential to be applied in the geological exploration of Lunar surface, its accurate and efficient identification will contribute to the understanding of the geological history of the Moon and the exploration of the Moon’s resources and provide a reference for more accurate and detailed mapping of specific features of the Moon.
However, the proposed method still has some limitations. It contains following aspects. Firstly, we haven’t expended our dataset to other types of landform or lineament structure in other planet and test the performance of our improved-UNet++ and YOLOv5 networks. Due to the various and complexity geological activity in other planet, the lineament structure may has great difference compared with Lunar lineament. The generalization is still under verification. Secondly, an important prerequisite of our ’polygon match’ strategy algorithm is that the bounding boxes from YOLOv5 network should contain only correspond to one major lineament structure unit in most situation. However, in some situation, the bounding boxes may contain more than one lineament structure unit with similar size. The ’polygon match’ strategy algorithm is not able to perform well in this situation. The improvement of the ’polygon match’ strategy algorithm is still under consideration.
In the future, the data enhancement will be engaged in the future work. As the booming development of generative model [44], diffusion model [45], GAN model and other generative model will be applied in our future work [46], aiming to generate high quality and diversity lineament structure samples including building the lineament structure dataset from other planets. Furthermore, facing with restricted samples on some planets, for instance, Venus, where only few samples are recorded, we will try to generate more samples from the existing samples. At same time, we also plan to focus on establishing a standard to evaluate the result of generative results, which will guarantee the reality of the generated samples. Our final goal is to establish a lineament structure dataset with great quantity, high diversity and high quality. Meanwhile, we will also expend our dataset to other types of landform, aiming to improve the generalization ability of the proposed method. Simultaneously, we will also focus on the improvement of the network architecture, aiming to improve the performance of the proposed method, which aims to increase the ability of object detection and semantic segmentation.

Author Contributions

Conceptualization, P.Y.; Methodology, P.Y.; Investigation, P.Y.; Writing—Original Draft, P.Y. and J.L.; Writing—Review and Editing, P.Y. and J.L.; Supervision, Y.Z. and X.T.; Project Administration, X.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Wuyi University Hong Kong Macao Joint Research and Development Fund under Grant 2022WGALH19; in part by the Science and Technology Development Fund of Macau (grant number 0038/2020/A1).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kattenborn, T.; Leitloff, J.; Schiefer, F.; Hinz, S. Review on Convolutional Neural Networks (CNN) in vegetation remote sensing. ISPRS J. Photogramm. Remote Sens. 2021, 173, 24–49. [Google Scholar] [CrossRef]
  2. Young, R.A.; Brennan, W.J.; Wolfe, R.W.; Nichols, D.J. Volcanism in the lunar maria. Apollo 17: Prelim. Sci. Rep. 1973, 330, 31-1–31-11. [Google Scholar]
  3. McGill, G.E. Attitude of fractures bounding straight and arcuate lunar rilles. Icarus 1971, 14, 53–58. [Google Scholar] [CrossRef]
  4. Lucchitta, B.K.; Watkins, J.A. Age of graben systems on the Moon. In Proceedings of the 9th Lunar and Planetary Science Conference, Houston, TX, USA, 13–17 March 1978; Volume 3; (A79-39253 16–91); Pergamon Press, Inc.: New York, NY, USA, 1978; Volume 9, pp. 3459–3472. [Google Scholar]
  5. Cruikshank, D.; Wood, C. Lunar rilles and Hawaiian volcanic features: Possible analogues. Moon 1972, 3, 412–447. [Google Scholar] [CrossRef]
  6. Watters, T.R.; Robinson, M.S.; Banks, M.E.; Tran, T.; Denevi, B.W. Recent extensional tectonics on the Moon revealed by the Lunar Reconnaissance Orbiter Camera. Nat. Geosci. 2012, 5, 181–185. [Google Scholar] [CrossRef]
  7. Mège, D.; Cook, A.; Garel, E.; Lagabrielle, Y.; Cormier, M.H. Surface collapse and volcanic rifting on Mars. In Proceedings of the Lunar and Planetary Science Conference, Houston, TX, USA, 11–15 March 2002; p. 2042. [Google Scholar]
  8. Melosh, H.J.; Schenk, P. Split comets and the origin of crater chains on Ganymede and Callisto. Nature 1993, 365, 731–733. [Google Scholar] [CrossRef]
  9. Ferrill, D.A.; Wyrick, D.Y.; Morris, A.P.; Sims, D.W.; Franklin, N.M. Dilational fault slip and pit chain formation on Mars. GsA Today 2004, 14, 4. [Google Scholar] [CrossRef]
  10. Li, K.; Chen, J.; Tarolli, P.; Sofia, G.; Feng, Z.; Jing, L. Geomorphometric multi-scale analysis for the automatic detection of linear structures on the lunar surface. Earth Sci. Front. 2014, 21, 212–222. [Google Scholar]
  11. Lin, L.; Jianzhong, L.; Li, Z.; Jinzhu, J.; Dijun, G.; Jingwen, L. Research on the classification system of lunar lineaments. Acta Petrol. Sin. 2017, 33, 3285–3301. [Google Scholar]
  12. Ahmadi, H.; Pekkan, E. Fault-based geological lineaments extraction using remote sensing and GIS—A review. Geosciences 2021, 11, 183. [Google Scholar] [CrossRef]
  13. Smith, D.E.; Zuber, M.T.; Neumann, G.A.; Lemoine, F.G.; Mazarico, E.; Torrence, M.H.; McGarry, J.F.; Rowlands, D.D.; Head, J.W., III; Duxbury, T.H.; et al. Initial observations from the lunar orbiter laser altimeter (LOLA). Geophys. Res. Lett. 2010, 37, L18204. [Google Scholar] [CrossRef]
  14. Vondrak, R.; Keller, J.; Chin, G.; Garvin, J. Lunar Reconnaissance Orbiter (LRO): Observations for lunar exploration and science. Space Sci. Rev. 2010, 150, 7–22. [Google Scholar] [CrossRef]
  15. Strom, R.G. Analysis of Lunar Lineaments. i-Tectonic Maps of the Moon. 1964. Available online: https://ntrs.nasa.gov/citations/19650040560 (accessed on 2 February 2024).
  16. Tamililakkiya, V.; Vani, K. Feature Extraction from Lunar Images. In Proceedings of the International Conference on Digital Image Processing and Information Technology, Tirunelveli, India, 23–25 September 2011; Springer: Berlin/Heidelberg, Germany, 2011; pp. 34–43. [Google Scholar]
  17. Yan, P.; Su, Y.; Tian, X. Classification of Mars Lineament and Non-Lineament Structure Based on ResNet50. In Proceedings of the 2020 IEEE International Conference on Advances in Electrical Engineering and Computer Applications (AEECA), Dalian, China, 25–27 August 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 437–441. [Google Scholar]
  18. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  19. Yuan, X.; Zhu, J.; Lei, H.; Peng, S.; Wang, W.; Li, X. Duplex-Hierarchy Representation Learning for Remote Sensing Image Classification. Sensors 2024, 24, 1130. [Google Scholar] [CrossRef] [PubMed]
  20. Raniga, D.; Amarasingam, N.; Sandino, J.; Doshi, A.; Barthelemy, J.; Randall, K.; Robinson, S.A.; Gonzalez, F.; Bollard, B. Monitoring of Antarctica’s Fragile Vegetation Using Drone-Based Remote Sensing, Multispectral Imagery and AI. Sensors 2024, 24, 1063. [Google Scholar] [CrossRef] [PubMed]
  21. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  22. Su, Z.; Yu, J.; Tan, H.; Wan, X.; Qi, K. MSA-YOLO: A Remote Sensing Object Detection Model Based on Multi-Scale Strip Attention. Sensors 2023, 23, 6811. [Google Scholar] [CrossRef] [PubMed]
  23. Li, Z.; Yuan, J.; Li, G.; Wang, H.; Li, X.; Li, D.; Wang, X. RSI-YOLO: Object detection method for remote sensing images based on improved YOLO. Sensors 2023, 23, 6414. [Google Scholar] [CrossRef]
  24. Kavran, D.; Mongus, D.; Žalik, B.; Lukač, N. Graph Neural Network-Based Method of Spatiotemporal Land Cover Mapping Using Satellite Imagery. Sensors 2023, 23, 6648. [Google Scholar] [CrossRef]
  25. Aghaee, A.; Shamsipour, P.; Hood, S.; Haugaard, R. A convolutional neural network for semi-automated lineament detection and vectorisation of remote sensing data using probabilistic clustering: A method and a challenge. Comput. Geosci. 2021, 151, 104724. [Google Scholar] [CrossRef]
  26. Li, Z.; Yan, P.; Liang, J.; Tian, X. Automatic DEM data classification of Mars lineament structure. In Proceedings of the International Conference on Computer Application and Information Security (ICCAIS 2021), Wuhan, China, 18–19 December 2021; SPIE: Cergy, France, 2022; Volume 12260, pp. 474–479. [Google Scholar]
  27. Lee, C. Automated crater detection on Mars using deep learning. Planet. Space Sci. 2019, 170, 16–28. [Google Scholar] [CrossRef]
  28. Hsu, C.Y.; Li, W.; Wang, S. Knowledge-driven GeoAI: Integrating spatial knowledge into multi-scale deep learning for Mars Crater detection. Remote Sens. 2021, 13, 2116. [Google Scholar] [CrossRef]
  29. Li, W.; Zhou, B.; Hsu, C.Y.; Li, Y.; Ren, F. Recognizing terrain features on terrestrial surface using a deep learning model: An example with crater detection. In Proceedings of the 1st Workshop on Artificial Intelligence and Deep Learning for Geographic Knowledge Discovery, Los Angeles, CA, USA, 7–10 November 2017; pp. 33–36. [Google Scholar]
  30. Robinson, M.S.; Brylow, S.; Tschimmel, M.; Humm, D.; Lawrence, S.; Thomas, P.; Denevi, B.W.; Bowman-Cisneros, E.; Zerr, J.; Ravine, M.; et al. Lunar reconnaissance orbiter camera (LROC) instrument overview. Space Sci. Rev. 2010, 150, 81–124. [Google Scholar] [CrossRef]
  31. Baranyi, P. PDS Geosciences Node at Washington University in St. Louis. Available online: https://ode.rsl.wustl.edu/moon/index.aspx (accessed on 2 February 2024).
  32. Wu, W.; Liu, H.; Li, L.; Long, Y.; Wang, X.; Wang, Z.; Li, J.; Chang, Y. Application of local fully Convolutional Neural Network combined with YOLO v5 algorithm in small target detection of remote sensing image. PLoS ONE 2021, 16, e0259283. [Google Scholar] [CrossRef] [PubMed]
  33. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  34. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  35. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: Redesigning Skip Connections to Exploit Multiscale Features in Image Segmentation. IEEE Trans. Med. Imaging 2019, 39, 1856–1867. [Google Scholar] [CrossRef] [PubMed]
  36. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  37. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  38. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  39. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  40. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  41. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  42. Jocher, G.; Chaurasia, A.; Stoken, A.; Borovec, J.; Kwon, Y.; Michael, K.; Fang, J.; Yifu, Z.; Wong, C.; Montes, D.; et al. ultralytics/yolov5: V7. 0-yolov5 sota Realtime Instance Segmentation; Zenodo: Geneva, Switzerland, 2022. [Google Scholar]
  43. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLOv8. 2023. Available online: https://docs.ultralytics.com/ (accessed on 2 February 2024).
  44. Bandi, A.; Adapa, P.V.S.R.; Kuchi, Y.E.V.P.K. The power of generative ai: A review of requirements, models, input–output formats, evaluation metrics, and challenges. Future Internet 2023, 15, 260. [Google Scholar] [CrossRef]
  45. Croitoru, F.A.; Hondru, V.; Ionescu, R.T.; Shah, M. Diffusion models in vision: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10850–10869. [Google Scholar] [CrossRef]
  46. Cao, Y.; Li, S.; Liu, Y.; Yan, Z.; Dai, Y.; Yu, P.S.; Sun, L. A comprehensive survey of ai-generated content (aigc): A history of generative ai from gan to chatgpt. arXiv 2023, arXiv:2303.04226. [Google Scholar]
Figure 1. Images of Lineament Structure on the Lunar Surface used in this experiment are captured by Lunar Reconnaissance Orbiter Camera (LROC) [30] at a resolution of 100 m/pix. These images are porvided by NASA/JPL-Caltech. Images above are the typical lineament structure on the Lunar surface. The lineament structure ridge, rills, catena and vallis are the main types of lineament structure on the Lunar surface.
Figure 1. Images of Lineament Structure on the Lunar Surface used in this experiment are captured by Lunar Reconnaissance Orbiter Camera (LROC) [30] at a resolution of 100 m/pix. These images are porvided by NASA/JPL-Caltech. Images above are the typical lineament structure on the Lunar surface. The lineament structure ridge, rills, catena and vallis are the main types of lineament structure on the Lunar surface.
Sensors 24 02256 g001
Figure 2. The Visualized Results of the Dataset in this Experiment. The images above visualize the object detection and semantic segmentation results of the lineament structure from the bounding box and binarized mask respectively. The former is used to train the YOLOv5 network and the latter is used to train the improved-UNet++ network.
Figure 2. The Visualized Results of the Dataset in this Experiment. The images above visualize the object detection and semantic segmentation results of the lineament structure from the bounding box and binarized mask respectively. The former is used to train the YOLOv5 network and the latter is used to train the improved-UNet++ network.
Sensors 24 02256 g002
Figure 3. Diagram of Algorithm The marked sections can be regarded as our contribution in this experiment. The contribution can be concluded as following three parts: New lineament dataset construction, UNet++ improvement, and polygon-match strategy algorithm desgin.
Figure 3. Diagram of Algorithm The marked sections can be regarded as our contribution in this experiment. The contribution can be concluded as following three parts: New lineament dataset construction, UNet++ improvement, and polygon-match strategy algorithm desgin.
Sensors 24 02256 g003
Figure 4. Architecture of YOLOv5.
Figure 4. Architecture of YOLOv5.
Sensors 24 02256 g004
Figure 5. Architecture of C3 and SPPF.
Figure 5. Architecture of C3 and SPPF.
Sensors 24 02256 g005
Figure 6. Architecture of UNet++ and Improved UNet++. UNet++ is able to drop the layers, when the model already learns enough features. This type of feature allows the network to have much higher efficiency due to the reduction of the number of parameters from the dropped layers [35].
Figure 6. Architecture of UNet++ and Improved UNet++. UNet++ is able to drop the layers, when the model already learns enough features. This type of feature allows the network to have much higher efficiency due to the reduction of the number of parameters from the dropped layers [35].
Sensors 24 02256 g006
Figure 7. Architecture of VGG Block and Residual Block. The VGG module usually consists of several 3 × 3 convolutional layers and a 2 × 2 max-pooling layer, with the step size and padding of all convolutional layers set to 1, making the outputs of the convolutional layers have the same height and width as the inputs. The residual module, on the other hand, consists of two or more convolutional layers and a cross-layer connection, which connects the inputs directly to the layers behind them, thus forming a “short circuit”.
Figure 7. Architecture of VGG Block and Residual Block. The VGG module usually consists of several 3 × 3 convolutional layers and a 2 × 2 max-pooling layer, with the step size and padding of all convolutional layers set to 1, making the outputs of the convolutional layers have the same height and width as the inputs. The residual module, on the other hand, consists of two or more convolutional layers and a cross-layer connection, which connects the inputs directly to the layers behind them, thus forming a “short circuit”.
Sensors 24 02256 g007
Figure 8. Architecture of Attention Block. Attention block is a combination of channel and spatial attention. It is able to improve the network’s ability to extract features. Due to its felxibility and lightweight, it is widely used in convlolutional layers with very less cost increasing the feature extraction ability at same time [37].
Figure 8. Architecture of Attention Block. Attention block is a combination of channel and spatial attention. It is able to improve the network’s ability to extract features. Due to its felxibility and lightweight, it is widely used in convlolutional layers with very less cost increasing the feature extraction ability at same time [37].
Sensors 24 02256 g008
Figure 9. Result of Polygon Match Strategy Algorithm. The binarized mask on the left is the result without polygon-match strategy, result processed by polygon-match strategy and the ground truth. The right part are the visualized results of the binarized masks on the left respectively.
Figure 9. Result of Polygon Match Strategy Algorithm. The binarized mask on the left is the result without polygon-match strategy, result processed by polygon-match strategy and the ground truth. The right part are the visualized results of the binarized masks on the left respectively.
Sensors 24 02256 g009
Figure 10. Comparison Experiment Results with Other Related Algorithms. It can be seen that less irrelevant pixels are detected in the result of the proposed method compared with other related algorithms. At same time it also shows clearer edge and bound box in the instance segmentation and object detection respectively of lineament structure result.
Figure 10. Comparison Experiment Results with Other Related Algorithms. It can be seen that less irrelevant pixels are detected in the result of the proposed method compared with other related algorithms. At same time it also shows clearer edge and bound box in the instance segmentation and object detection respectively of lineament structure result.
Sensors 24 02256 g010
Figure 11. The Comparison of the Proposed Method and Other Related Algorithms in mean IoU and mean pixel accuracy under different IoU threshold and ROC curves. The IoU and mean pixel accuracy shown on the right are under threshold 0.5, which is the most common metric to evaluate the authenticity of samples. At same time, the ROC curves is calulated under the same threshold.
Figure 11. The Comparison of the Proposed Method and Other Related Algorithms in mean IoU and mean pixel accuracy under different IoU threshold and ROC curves. The IoU and mean pixel accuracy shown on the right are under threshold 0.5, which is the most common metric to evaluate the authenticity of samples. At same time, the ROC curves is calulated under the same threshold.
Sensors 24 02256 g011
Table 1. Comparison Experiment Results of Improved-UNet++ and other UNet++ Networks.
Table 1. Comparison Experiment Results of Improved-UNet++ and other UNet++ Networks.
ModelPRmAP@50mAP@50:95
UNet++0.43260.64520.58430.3493
Residual Block + UNet++0.4930.66420.60520.3786
CBAM + UNet++0.400.390.05110.0979
Ours (Residual Block + CBAM + UNet++)0.6790.67660.60810.3887
Table 2. Comparison Experiment Results with Other Related Networks.
Table 2. Comparison Experiment Results with Other Related Networks.
ModelPRmAP@50mAP@50:95ParametersFLOPS
FCN0.4130.2820.2130.086102,760,448136.2 G
Faster-RCNN0.4810.3440.2960.10319,234,523239 G
SegNet0.43770.4510.32480.142229,461,472165.2 G
Deeplabv3+0.45260.49180.39430.192342,004,074173.82 G
UNet++0.43260.64520.58430.34939,207,47256.88 G
YOLOv5-seg0.6230.4770.4640.1477,398,42225.7 G
YOLOv8-seg0.6230.4770.4640.1473,258,25912.0 G
Ours0.6790.67660.60810.388712,688,56873.81 G
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yan, P.; Liang, J.; Tian, X.; Zhai, Y. A New Lunar Lineament Extraction Method Based on Improved UNet++ and YOLOv5. Sensors 2024, 24, 2256. https://doi.org/10.3390/s24072256

AMA Style

Yan P, Liang J, Tian X, Zhai Y. A New Lunar Lineament Extraction Method Based on Improved UNet++ and YOLOv5. Sensors. 2024; 24(7):2256. https://doi.org/10.3390/s24072256

Chicago/Turabian Style

Yan, Pengcheng, Jiarui Liang, Xiaolin Tian, and Yikui Zhai. 2024. "A New Lunar Lineament Extraction Method Based on Improved UNet++ and YOLOv5" Sensors 24, no. 7: 2256. https://doi.org/10.3390/s24072256

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop