Abstract

In order to improve the visual navigation performance in complex environment, a robust visual navigation method for substation inspection robot is proposed in this paper. Based on the robustness of hexagonal cone model to light changes, this method can solve the squeezing problem of navigation path in complex environment and reduce the interference caused by external light factors. Based on HM preprocessed images, semantic segmentation is carried out with deep convolutional neural network to obtain global features, local features, and multiscale information of images, so as to effectively improve the network recognition accuracy. The results show that the images after HM color space transformation and grayscale reconstruction can compress the color space while preserving the edge details, which is beneficial to the semantic segmentation network for further scene road recognition. Because the original structure of the network is not adjusted and the corresponding preprocessing layer is added, the size of the network model is relatively increased, but the reasoning speed of the original network is significantly improved, which is 16.4% on average.

1. Introduction

As a direction for the development of future energy networks, smart grids have been extended to existing power grid control networks [1]. Intelligent and efficient management technologies will be used to implement automation, integration, centralization, and intelligence in power grid management, such as power generation, transmission, distribution, and energy consumption [2]. As an important connection hub in the power supply network, the substation needs regular inspection in order to find potential problems in time and maintain them in time to ensure the safe operation of the power grid [3]. As an important carrier of automatic monitoring of substation operation, robot can be widely used in the automatic monitoring of substation environment [4]. The magnetic path navigation method is simple and reliable and has high navigation accuracy, but this method uses a magnetic path for navigation, which requires a change in the layout of the substation. There are cargo and potential safety hazards during actual work [5]. Inertial navigation and wireless positioning methods require the installation of wireless devices for signal transmission and reception in the work environment. Although the installation is convenient, the decentralized wireless devices lead to poor recognition stability and adaptability and are prone to cumulative errors [6, 7]. Although GPS navigation method can work without adjusting the working scene, it has poor accuracy and low accuracy. In particular, in the power system environment, electromagnetic interference is particularly strong, and the well cannot meet the accuracy requirements of substation patrol inspection [8]. Figure 1 shows a manufacturing technology of inspection robot based on improved LSTM+CNN algorithm.

2. Literature Review

The inspection system of electric power inspection robot is a scientific and technological achievement integrating multiple disciplines. In addition to the traditional and relatively perfect technologies such as machinery, electricians and electronics, communication, automation, and sensors, it also adds deep learning neural network to improve the ability of system fault identification [9, 10]. The power inspection robot system is mainly composed of image detection and recognition unit, automatic obstacle avoidance unit, data monitoring unit, and data transmission and storage unit. The image detection system is a deep learning neural network based on PyTorch framework. It uses a neural network model similar to Yolov3 to dynamically collect image information, extract the characteristics of the image at any time, and judge whether there is fault information in the power system. Using this technology, the power inspection robot can directly warn the instrument panel, indicator lights, open fire, smoke, and other faults. At the same time, the video will be transmitted to the Alibaba cloud platform for backup, which can be retrieved by the inspectors at any time. It can also directly follow the real-time video and conduct video remote detection synchronously with the machine [11, 12]. Compared with the early inspection robot, the advantage of this power inspection robot is to increase the image detection function of convolutional neural network. Flexible neural networks are widely used in the field of visual target detection. Transient neural networks can automatically retrieve functional information from images, learn specific information from different lesion information, and continuously optimize their identification accuracy to effectively detect substations [13, 14]. The foreground vision device is used to capture the road image in real time, the image processing technology is used to identify the road surface to determine the position relationship between the inspection robot and the navigation line, and then the deep learning algorithm of the inspection robot walking and detection task is realized through the bottom motion control, which has attracted extensive attention in the existing visual path navigation methods [15]. Compared with other methods mentioned above, visual navigation method not only overcomes the disadvantages of low accuracy and poor adaptability of these methods in power system scene but also has the advantages of simplicity, reliability, accuracy, convenient transformation, and installation [16].

Based on this, we offer a reliable method of visual navigation for substation control robots to improve visual navigation performance in complex environments. This method uses the robustness of the gecko model (HM) to change the light, which solves the problem of compressing the navigation path in complex environments and reduces interference caused by external lighting factors. In order to effectively improve the accuracy of network recognition, semantic segmentation with a deep neural network (DCNN) based on a preprocessed image of HM is obtained, which provides information on global features, local features, and multiscale images.

3. Research Methods

3.1. Color Space Conversion

Good road recognition ability is the key to the self-adaptive navigation of the substation inspection robot. Only by controlling the inspection robot on the basis of identifying the effective road surface can the inspection robot effectively complete the inspection task and avoid entering the restricted area or areas that are not suitable for driving. Road recognition based on camera images needs to take into account various complex situations such as strong light, shadows, and rainy days. Define the maximum value of the three component values of , , and of each pixel as and the minimum value as , and then, there are

For the three components in HM, component represents color brightness, and is expressed as

component represents color purity, and then, represents

3.2. Gray Image Reconstruction

Each component of the Hopfield Model (HM) has different characteristics. , , and components of different weights are reassembled to restore the image to minimize the effect of lighting changes in the image [17, 18]. In terms of features, the navigation path properties on components and are essentially opposite; the normalization method is considered to enhance and preliminarily extract the navigation path feature information. First, inverse image processing is performed on the component, as shown in

Then, the inverse image is normalized, where the normalized range is , where is the minimum value of and is the maximum value of , and then, the corresponding coefficient matrix is obtained, as shown in

This operation avoids the forced enlargement and reduction of the image in the normalization process. To recreate the gray image, add the components to the components and then multiply by a coefficient matrix according to

Analyze equation (7) and select the components of the reconstructed gray image and double the result. Consider effectively eliminating shadow interference in the image by increasing the weight of the components while restoring the gray image. This allows you to restore a gray image without the interference of external objects such as strong light, shadows, and surface water that highlight the road surface [19].

3.3. Image Acquisition under Dark Light Conditions

In the dark light environment, due to the light supplement defect of the point light source of the ordinary light source, the image acquisition of the inspection robot has light spots or concentrated light points, which brings the influence of image acquisition. Based on the HM, the image components are decomposed and used to detect the brightness of a simple light source, and the illumination of the collected image under the condition of dark light is appropriately adjusted. If the average brightness in the component is defined as , then equation (8) is shown.

where is the saturation value of component and is the number of pixels in the saturation value. Define the average illuminance of the component as SP, followed by

where is the saturation value in the component and is the pixel count in the saturation value. In combination with equations (8) and (9), the conditions for judging whether the image is too bright or too dark are (a) and and (b) and .

According to the above results, the inspection robot analyzes the brightness information in the current image and determines whether the supplementary light intensity needs to be adjusted so that the light intensity can reach the same illuminance during path navigation in different road sections, so as to ensure the consistency and stability of the image collected by the camera [20, 21].

3.4. HM-DCNN

For a traditional CNN extension, the DCNN structure still consists of a circulation layer, a sampling layer, a consolidation layer, and a fully connected layer. Different from other conventional road semantic segmentation networks, HM-DCNN proposed in this paper performs semantic segmentation of the image through deep learning after image gray preprocessing, adding the adjustment of different light source intensity and supplementary preprocessing layer. In addition, because the preprocessed image is more concise and unified, the complexity of the network and the scale of target training parameters can be greatly reduced. The specific algorithm flow of the improved HM-DCNN is as follows. Based on the HM, complete the image acquisition and corresponding gray image reconstruction under the light conditions of different light sources, and adjust the fill light intensity appropriately. Through the first CNN, the global location is carried out to find the feature point region of continuous features and local features, and then, the maximum rectangular region of the target and surrounding objects is output. The convolutional layer is a key component of the HM-DCNN, and its primary function is to calculate the multiplication of the receiving field point and the rotation of the filter (or core) that can be studied. After the conversion operation, a nonlinear sample is performed in the aggregation layer to reduce the dimensionality of the data [22, 23]. The most common integration strategy is the most consolidated and average consolidation. Max pooling takes the maximum value from the candidates, while the average aggregation selects the average from the estimated candidates. Here, the maximum aggregation method is chosen to minimize the calculated average displacement due to the rotation layer parameter error in order to preserve the image structure information as much as possible. The function map obtained after the sample below is then sent to the activation function to process the nonlinear conversion. High-level grounding is done through fully connected layers. Nerve cells in this layer are involved in all the activations of the previous layer. The loss layer is usually the last layer of the DCNN and specifies how to punish the difference between the predicted and actual label during network training. Distribution points are strongly separated by a sigmoid layer. Here, a complete circulatory neural network based on the BVLC Caffe was selected to generate the HM-DCNN. The network structure is shown in Figure 2. There are a total of five circulation layers between the input and output layers. The network uses the Linear Unit (ReLU) as a function of nonlinear activation and finally identifies the distribution points of semantic segmentation through the sigmoid and builds the corresponding DCNN architecture. Table 1 shows the parameters for generating the corresponding circulation network according to the functional definition of the different layers of the DCNN [24].

Due to grayscale reconstruction preprocessing in the inspection robot, global gain normalization is applied to the processed images to further reduce the intensity changes in the images. This gain is calculated by aligning the calculated signal envelope using the median filter. To calculate the signal envelope, the image is filtered through a low-pass Gaussian core. DCNN is highly resistant to lighting changes, but the combination of image gray recovery and normal processing makes the signal dynamic range more uniform, which can further improve processing accuracy and merging speed.

4. Result Analysis

4.1. HM Fill Light Experiment and Image Preprocessing

To test the effectiveness of the proposed HM-DCNN network application in combination with the gray reconstruction in the actual substation scene, the substation path map collected by the substation control robot was used to train the network and test the network recognition accuracy. The network was trained in stochastic gradient drop (SGD), and the learning speed was set to 0.01, the pulse parameter to 0.9, and the weight loss to 0.0005. In order to make the inspection robot still run well in the dark environment, it is particularly important to supplement the light appropriately. Scene A is the road condition during the day, scene B is the adaptive fill light condition, and scene C is the nonfill light condition. The calculation results of and are shown in Table 2.

From the results in Table 2, it can be seen that the differences between the and components are clear in different lighting conditions. According to different difference values, setting the corresponding threshold can achieve appropriate light compensation. In order to analyze the influence of HM on the network recognition effect in the scene, the traditional edge extraction processing alone has poor adaptability in the case of high illumination difference and scene feature brightness and will lose a lot of details. The image after HM color space conversion and gray reconstruction can compress the color space and greatly retain the edge detail features, which is conducive to further scene pavement recognition by semantic segmentation network.

4.2. HM-DCNN Semantic Segmentation Experiment

In the preparation process of HM-DCNN training samples, in order to avoid the overfitting of training, improve the robustness of the network, and better extract the core features, a series of adjustments are made to the sample data, including illumination, color saturation, rotation, and clipping.

As a user-defined layer of other reasoning network preprocessing, HM can be combined with a variety of semantic segmentation networks. The processed image is used for gray reconstruction, which is used as the input of each comparison network and compared with the network results trained with the original image directly as the input. The effects of different illumination conditions on the experimental results are shown in Table 3.

As shown in Figure 3, the recognition effects of several commonly used semantic segmentation networks are compared and analyzed through experiments. Under normal illumination, DCNN has high recognition effect, but under dim illumination and supplementary illumination, the recognition effect decreases significantly. Through the comparison of each network after adding HM treatment, although the improvement effect under normal light is not significant, the recognition rate of the dark environment of the original network has been greatly improved, especially when combined with DCNN.

As shown in Table 4, the original structure of the network is not adjusted and the corresponding preprocessing layer is added, which makes the size of the network model increase relatively, but the reasoning speed of the original network is significantly improved, with an average of 16.4%.

5. Conclusion

This document proposes a new method of reliable visual navigation of substation control robots in complex road environments, which has the advantages of strong anti-interference ability, simple operation, high precision, and good stability. This method fully considers the road characteristics and working conditions of the substation and uses the insensitivity of HM color space to light change, shadow, and interference to reconstruct the gray image of the collected image; therefore, the specifics of the navigation path are reflected in more detail. In addition, DCNN is used to train and recognize images processed on a grayscale to improve network recognition accuracy. Because the HM reduces the amount of color space, the first-ever DCNN can be used in substation control robots with poor processing capabilities. The results of the experiment show that the network recognition results show a clear advantage in combination with the HM gray reconstruction compared to the existing pure training network. At the same time, the HM-DCNN method proposed in this paper can stabilize the average processing accuracy of DCNN from 91% to 93%, have the best effect on the entire network, and fully meet the actual needs of the substation situation.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.