Next Article in Journal
Enhancing SDN WISE with Slicing Over TSCH
Previous Article in Journal
Impact of Experimental Parameters on Cell–Cell Force Spectroscopy Signature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Scale-Invariant Multidirectional License Plate Detection with the Network Combining Indirect and Direct Branches

1
School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
2
USTB-EEasyTech Joint Lab of Artificial Intelligence, University of Science and Technology Beijing, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(4), 1074; https://doi.org/10.3390/s21041074
Submission received: 6 January 2021 / Revised: 24 January 2021 / Accepted: 2 February 2021 / Published: 4 February 2021
(This article belongs to the Section Intelligent Sensors)

Abstract

:
As the license plate is multiscale and multidirectional in the natural scene image, its detection is challenging in many applications. In this work, a novel network that combines indirect and direct branches is proposed for license plate detection in the wild. The indirect detection branch performs small-sized vehicle plate detection with high precision in a coarse-to-fine scheme using vehicle–plate relationships. The direct detection branch detects the license plate directly in the input image, reducing false negatives in the indirect detection branch due to the miss of vehicles’ detection. We propose a universal multidirectional license plate refinement method by localizing the four corners of the license plate. Finally, we construct an end-to-end trainable network for license plate detection by combining these two branches via post-processing operations. The network can effectively detect the small-sized license plate and localize the multidirectional license plate in real applications. To our knowledge, the proposed method is the first one that combines indirect and direct methods into an end-to-end network for license plate detection. Extensive experiments verify that our method outperforms the indirect methods and direct methods significantly.

1. Introduction

License plate detection (LPD) plays an essential role in many practical applications, including electronic toll collection, traffic surveillance, and enforcement. When the image acquisition conditions (shooting distance and angle) are restricted, such as the parking toll, the LPD task is almost completely solved. However, if the image is captured in the wild, it remains challenging due to various sizes, orientations, and backgrounds. Figure 1 illustrates some license plate (LP) examples in real scenarios.
Recent LPD methods can be roughly divided into direct and indirect ways. Direct methods directly localize the license plate in the input image with handcrafted features [1,2,3,4], deep-learning features [5,6,7,8,9], or license plate recognition system [10,11]. However, detecting small-sized license plates is challenging since they only occupy a relatively small area in the whole image. Indirect methods detect the license plate using the vehicle’s proposal [12,13,14,15,16,17] or the vehicle head region [18]. The vehicle head region is manually defined as the smallest region enclosing the headlights and tires. The indirect methods can reduce the detection area and background noises, which is favorable to small-sized license plate detection. However, when the vehicle fails to be detected due to severe occlusion or nonuniform illumination, it will fail to localize the license plate.
To overcome these problems, we propose a novel network composed of an indirect branch and a direct branch. The indirect detection branch can approximately localize the license plate based on the spatial relationships between the license plate and the vehicle. Then it can refine the license plate in the local region. This way, it can significantly reduce the detection area and mitigate the adverse effects of the background noises, which is favorable to small-sized license plate detection. The direct detection branch can reduce false negatives in the indirect detection branch due to the miss of vehicles’ detection. We combine the indirect and direct branches to construct an end-to-end trainable network for license plate detection. The detection results of two branches are merged by post-processing operations, such as non-maximum suppression (NMS). Extensive experiments show that our method outperforms both the direct approach and the indirect approach.
Moreover, many methods [6,9,10,11,16,18] do not consider the orientation of the license plate, which is only applicable to specific scenarios, such as parking charges and vehicle access/exit management. When it comes to more complex scenarios, such as road scenes, if we regard the tilted license plate as the horizontal direction, it may cause errors in the subsequent license plate recognition [19,20,21,22]. Although Dong et al. [14,23] propose to detect the multidirectional license plate, these methods are very complicated due to adopting multiple separate models.
We propose to detect the multidirectional license plate by localizing the four corners of the license plate to reduce complexity. It can be easily implemented by integrating the corner prediction module into the two branches mentioned above, with no extra models. In this way, the whole detection network is still in an end-to-end trainable manner, as shown in our open-source codes [24].
Our main contributions can be summarized as:
  • We propose a novel network that combines indirect and direct branches for license plate detection in the wild. The indirect detection branch utilizes vehicle–plate relation and can precisely locate the license plate in a coarse-to-fine scheme. The direct detection branch localizes the license plate in the input image directly, reducing false negatives in the indirect detection branch due to the miss of vehicles’ detection.
  • We propose to detect the multidirectional license plate by localizing the four corners of the license plate. This universal detection module can be easily integrated into standard detection networks.
  • Notably, the whole model is constructed in an end-to-end trainable manner. By utilizing the post-processing operations, such as NMS, the final detection results are obtained by merging the indirect and direct branches. Hence, the whole model benefits from joint learning of all tasks. To our knowledge, our model is the first one that combines indirect and direct methods into an end-to-end network for license plate detection.
The rest of this paper is organized as follows. Related work is described in Section 2. In Section 3, we describe our method in detail. Section 4 presents comparative experiments and analyses. A short discussion is presented in Section 5. The final remarks are presented in Section 6.

2. Related Work

2.1. Direct License Plate Detection

The following methods propose to detect the license plate in the input image directly. Jun et al. [25] present a morphology-based method for LPD by extracting contrast features. To solve illumination variation and background interference, Tian et al. [1] propose an Adaboost algorithm combined with a color differential model, which can detect the license plate in a coarse-to-fine manner. The literature [4,26] propose to use the edge and texture features for license plate detection. Zhou et al. [2] propose to localize the license plate by principal visual word, discovery, and local feature matching. Li et al. [3] propose a component-based method for license plate detection. This method detects candidate characters first, then constructs the spatial relationships of characters using conditional random field (CRF), and finally estimates the whole license plate. Yuan et al. [6] apply dense filters to extract all the possible candidate LP regions and then preserve true positive LPs using a cascaded classifier. Rabiah et al. [7] propose a YOLO-inspired adaptive solution with optimized parameters to enhance LPD performance. In literature [8], the license plate features from the bottom and high levels of the CNN network are extracted and integrated to achieve precise and real-time detection. Chen et al. [9] propose to detect the license plate in a separate branch to avoid the suppression effects caused by the vehicle. Xu et al. [10] use multi-level CNN features to detect multi-scale license plates. Li et al. [11] utilize Faster R-CNN [27] to detect the license plate, where the scales and shapes of anchors are designed to fit the license plate. However, these approaches are prone to fail small-sized license plates because the license plate only occupies a relatively small input image area.

2.2. Indirect License Plate Detection

The following methods propose to detect the license plate via vehicle–plate relation. In this way, it enables the model to focus on the potential location of the license plate and reduce disturbing background noises, which can improve the detection performance of small-sized license plates. Kim et al. [12] use R-CNN [28] to detect the vehicles first, then localize the license plate inside each vehicle. Fu et al. [13] apply the region proposal network (RPN) [27] to generate candidate vehicle proposals and then detect the license plate based on the convolutional features of the vehicle. The literature [14,15] propose a two-stage YOLOv2 [29] method for accurate license plate detection. The first stage detects the vehicle and the second stage detects the license plate in the detected vehicle region. Rayson et al. [16] utilize YOLOv2 [29] to detect all the possible vehicles and then localize all the license plates in the vehicle patches simultaneously. Sergio et al. [18] propose to detect the vehicle firstly, then detect the vehicle head region in the vehicle, and localize the license plate in each vehicle head region finally. Chen et al. [17] propose estimating the approximate location of the license plate based on the offset between the center of the license plate and the vehicle, then refine the quadrilateral bounding box of the license plate in the local region. However, these methods will inevitably fail to localize the license plate if the vehicle fails to be detected. Our method combines the advantages of both indirect LPD approaches and direct LPD approaches, where it can detect small-sized license plates via vehicle–plate relation and reduce false-negative license plates caused by wrongly detected vehicles.

2.3. Multidirectional License Plate Detection

Xie et al. [30] propose to predict the rotation angle for multidirectional license plate detection based on modified YOLO [31]. Han et al. [5] propose to detect the license plate with a parallelogram by predicting three corners of the license plate. Tian et al. [32] adopt a semantic segmentation network for candidate license plate extraction and then refine the oriented bounding box of the license plate. All of the methods above regard the oblique license plate as a parallelogram. However, in real scenarios, a highly oblique license plate is an arbitrary quadrangle due to perspective transformation. Dong et al. [23] present to extract license plate candidates with RPN [27] and then use R-CNN [28] to localize the four corners of the license plate. The literature [14,33] employ spatial transformer networks (STN) [34] to obtain the affine transformation parameters of the license plate and transform the oblique license plate into a horizontal direction. However, the literature [14,23] are complicated due to adopting several separate models; moreover, they demand large-scale training data for STN. Our method can localize the quadrilateral bounding box of the license plate in an end-to-end manner, with no need for large-scale training data.

3. Materials and Methods

We propose a novel network for license plate detection, which can effectively detect the small-sized license plate and accurately localize the multidirectional license plate in real applications. The overall architecture is described in Section 3.1. The indirect detection branch can precisely detect the small-sized license plate, as described in Section 3.2. The direct detection branch can reduce the false-negative license plate in the indirect detection branch due to incorrectly detected vehicles, as described in Section 3.3. The whole network is constructed in an end-to-end trainable manner, as described in Section 3.4. The detection results of these two detection branches are merged by post-processing operations, such as NMS, as described in Section 3.5.

3.1. Overall Architecture

The overall architecture is illustrated in Figure 2. The network is constructed with two detection branches, i.e., indirect detection branch and direct detection branch. In the indirect detection branch, the approximate location and size of the license plate are predicted at the ALPD stage, where the center of the license plate (green circle) is obtained based on the offset (purple arrow) between the center of the license plate and the vehicle (orange circle). Moreover, the probability of the vehicle containing a license plate (red number) is predicted simultaneously. At the LREA stage, the local region of LP is obtained by expanding the LP region, and all the expanded LP regions (green dashed rectangle) are resized and aggregated into feature patches via differentiable region of interest (RoI) warping [35] for batch operation. At the MLPR stage, the quadrilateral (red circle) and horizontal (green rectangle) bounding boxes of the license plate are detected simultaneously in the local region of LP. In the direct detection branch, the license plate is directly detected in the input image at the DLPD stage. The DLPD and ALPD modules share the same backbone network but different detection head networks. Finally, the detection results of two branches are merged by post-processing operations, such as NMS. The network can be trained in an end-to-end manner, where the red arrows denote the backpropagation gradients.

3.2. Indirect Detection Branch

The indirect detection branch predicts the approximate location of the license plate utilizing spatial vehicle–plate relationships firstly (Section 3.2.1), then estimates the local region by expanding the LP region followed by an aggregation operation (Section 3.2.2), and refines the quadrilateral and horizontal bounding boxes of the license plate in the local region finally (Section 3.2.3). This multi-level design enables the model to focus on the potential location of the license plate and reduce the disturbing background noises.

3.2.1. Approximate License Plate Detection (ALPD)

At this stage, the approximate location of the license plate is estimated according to the vehicle–plate relation. At first, the vehicle is detected, so the center of the vehicle is determined. After that, the location of the license plate is obtained based on the offset between the center of the license plate and the vehicle. Meanwhile, the size of the license plate is directly predicted in the input image. According to the center and size, the license plate is approximately detected. In addition, the probability of the vehicle containing a license plate is predicted simultaneously. As shown in Figure 2, the location and size of the license plate are not accurate in general cases because the license plate only occupies a relatively small area in the large input image.
The ALPD module is based on SSD [36] for multi-task learning, which is the same as SSD512 [36] except for the training objective. The training objective of the ALPD module is defined as Equation (1), including five losses: vehicle classification loss L c l s ( c ) , vehicle regression loss L r e g ( p , g ) , offset loss L o f f ( p , g ) , LP size loss L s i z e ( p , g ) , and containing-LP loss L c o n _ l p ( p , g ) .
L 1 ( c , p , g ) = 1 N v L c l s ( c ) + L r e g ( p , g ) + L o f f ( p , g ) + L s i z e ( p , g ) + L c o n _ l p ( p , g ) ,
where N v is the number of matched anchor boxes with the ground-truth vehicles, c is the vehicle presence confidence, p is the predicted parameters, and g is the ground-truth parameters.
The training objective of vehicle detection is derived from SSD [36], including classification loss (i.e., Equation (2)) and regression loss (i.e., Equation (3)). The classification loss is the softmax loss over categories ζ { v e h i c l e , b a c k g r o u n d } . The regression loss is the smooth L1 loss [37] of the foreground category ζ + = v e h i c l e , which regresses to offsets for the center ( c x , c y ) , width ( w ) , and height ( h ) of the matched anchor box.
L c l s ( c ) = i = 1 N v ζ log c i ζ c i ζ = e x p c i ^ ζ ζ e x p c i ^ ζ ,
L r e g ( p , g ) = i = 1 N v m c x , c y , w , h I i j ζ + S m o o t h L 1 p i m g j m ,
where I i j ζ + 0 , 1 is the indicator of whether the ith anchor box matches the jth ground-truth box.
The offset and size losses are the smooth L1 loss between the predicted parameters ( p ) and the ground-truth parameters ( g ) based on the matched anchor boxes, as shown in Equations (4) and (5). The vehicle must contain a license plate; otherwise, the losses L o f f ( p , g ) and L s i z e ( p , g ) are 0 by setting g j + = 0 . This way, it can avoid learning false-positive predictions during training.
L o f f ( p , g ) = i = 1 N v m o f f x , o f f y I i j ζ + g j + S m o o t h L 1 p i m g j m ,
L s i z e ( p , g ) = i = 1 N v m l p w , l p h I i j ζ + g j + S m o o t h L 1 p i m g j m ,
where o f f x and o f f y are the offsets between the center of the license plate and the vehicle in x-direction and y-direction, l p w and l p h are the width and height of the license plate, and g j + 0 , 1 is the indicator of whether the jth vehicle contains a license plate.
Moreover, the probability of the vehicle containing a license plate can be used to reduce false positives of the license plate. A license plate will be detected only when the probability is greater than a certain threshold, and the threshold is empirically set to 0.5. During training, the vehicles with very small-sized or invisible license plates (occlusion, far shooting-distance, etc.) are regarded as without license plates; otherwise, the vehicles are considered as containing a license plate. The containing-LP probability is optimized by the binary cross-entropy loss (i.e., Equation (6)).
L c o n _ l p ( p , g ) = i = 1 N v g j + · log σ p i + + 1 g j + · log 1 σ p i + ,
where σ is a sigmoid function to limit the predicted containing-LP probability p i + 0 , 1 in case of loss divergence.

3.2.2. Local Region Estimation and Aggregation (LREA)

After the ALPD stage, there is a large deviation between the predicted license plate and the ground truth. We should make a further refinement to get more precise detection results, i.e., fine-tuning the license plate in the local region around the license plate. Based on the center and size of the license plate, we obtain the local region by merely expanding the license plate region with a preset ratio, enclosing the license plate and little background. The license plate occupies a relatively larger area in the local region than in the input image so that the subsequent refinement network can get more precise detection results.
There are many license plate regions obtained from different vehicles simultaneously. All the region features are extracted from the first convolutional layer, ensuring the whole network is constructed in an end-to-end manner. The first convolutional layer preserves the same size as the input image, which retains sufficient spatial information to detect small-sized license plates. Furthermore, all the LP regions are resized and aggregated via differentiable RoI warping [35] for batch operation, ensuring all the license plates are detected simultaneously to reduce the running time.

3.2.3. Multidirectional License Plate Refinement (MLPR)

In the local region, the quadrilateral and horizontal bounding boxes of the license plate are detected simultaneously. The quadrilateral bounding box is obtained by regressing the four corners of the license plate based on the matched anchor box, as illustrated in Figure 3. The matched anchor box is determined by the intersection over union (IoU) with the horizontal ground-truth box. The horizontal bounding box is used for NMS because of the fast computing speed. Compared with the ALPD module, the detection results of the MLPR module are more accurate.
The MLPR module has only 6 convolutional layers because the LPD task in the local region is relatively simple. Please refer to our open-source codes [24] for more details. The training objective of the MLPR module is defined as Equation (7), including three parts: LP classification loss L c l s ( c ) , LP regression loss L r e g ( p , g ) , and LP corner loss L c o r n e r ( p , g ) .
L 2 ( c , p , g ) = 1 N l p L c l s ( c ) + L r e g ( p , g ) + L c o r n e r ( p , g ) ,
where N l p is the number of matched anchor boxes with the horizontal LP ground-truth boxes, c is the LP presence confidence, p is the predicted LP parameter, and g is the LP ground-truth parameter.
The losses of the horizontal bounding box L c l s ( c ) and L r e g ( p , g ) are the same as vehicle detection except for the foreground category being LP, as shown in Equation (2) and Equation (3). As shown in Equation (8), the corner loss of the quadrilateral bounding box is the smooth L1 loss of the foreground category ζ + = l i c e n s e p l a t e , which regresses to offsets between the center of the matched anchor box and the four corners of the license plate.
L c o r n e r ( p , g ) = i = 1 N l p m t l x , t l y , t r x , t r y , b r x , b r y , b l x , b l y I i j ζ + S m o o t h L 1 p i m g j m ,
where m t l x , t l y , t r x , t r y , b r x , b r y , b l x , b l y are the four corners of the license plate, i.e., top-left, top-right, bottom-right, and bottom-left.

3.3. Direct Detection Branch

Direct License Plate Detection (DLPD)
The DLPD module can directly detect the license plate in the input image. In this way, small-sized license plates can not always be detected. However, in some cases, when the license plate fails to be detected in the indirect detection branch due to incorrectly detected vehicles, the DLPD module can reduce the false-negative license plate. The DLPD module is similar to the MLPR module described in Section 3.2.3. One significant difference is that the license plate is directly detected in the input image, not in the local region of the license plate. In addition, the backbone network of the DLPD module is the same as SSD [36] with 25 convolutional layers; the backbone network of the MLPR module only consists of 6 convolutional layers, as described in Section 3.2.3.
According to [9], it is difficult to effectively detect the vehicle and license plate simultaneously due to their subordinate relationships. This issue is caused by feature interaction between the vehicle and license plate in the traditional anchor-based detection method, such as SSD [36]. To solve this problem, we construct two separate detection branches for the DLPD and ALPD modules, respectively, as shown in Figure 2. These two modules share the same backbone network (i.e., VGG-16 [38] and extra layers) but different head networks. Please refer to our open-source codes [24] for more details. In this way, we can eliminate the adverse effects on the license plate caused by the vehicle.
Similar to the MLPR module, the training objective of the DLPD module is defined as Equation (9), including LP classification loss L c l s ( c ) , LP regression loss L r e g ( p , g ) , and LP corner loss L c o r n e r ( p , g ) .
L 3 ( c , p , g ) = 1 N l p L c l s ( c ) + L r e g ( p , g ) + L c o r n e r ( p , g ) .

3.4. End-to-End Trainable Detection Network

By integrating the indirect and direct detection branches, we develop an end-to-end trainable network for license plate detection, which can effectively detect the small-sized license plate and accurately localize the multidirectional license plate in real applications. Combining Equations (1), (7), and (9), the loss of the whole network is shown in Equation (10), where α and β are simply set to 1 to balance these loss terms.
L = L 1 ( c , p , g ) + α L 2 ( c , p , g ) + β L 3 ( c , p , g ) .
Figure 4 illustrates the loss changes during training, including L 1 and L 2 of the indirect detection branch as well as L 3 of the direct detection branch. During end-to-end training, the ALPD module can be optimized to detect the vehicle and approximate location of the license plate. Meanwhile, the license plate can be directly detected in the input image by the DLPD module. After training for some iterations, the MLPR module starts to refine the location of the license plate in the local region; then, the entire network will be optimized simultaneously. Specifically, during the first few training iterations, L 1 and L 3 go down, and L 2 remains zero because the untrained ALPD module can not estimate the location of the license plate; then, L 2 goes up dramatically because the ALPD module can approximately localize the license plate, and the MLPR module starts learning to regress the four corners of the license plate in the local region; finally, the total loss L goes down steadily because the indirect and direct detection branches are optimized simultaneously.

3.5. Post Processing

Figure 5 illustrates the post-processing operations. We can filter the most useless detection results by thresholding the confidence predicted by the network. After threshold filtering, the post-processing module can merge the detection results from two detection branches via NMS, removing duplicate detections. Instead of the quadrilateral bounding box, the horizontal bounding box of the license plate is used for NMS because of its faster computing speed. As shown in Section 4.7, the final detection results are mainly from the indirect detection branch because of its ability to detect small-sized license plates. In some cases, the direct detection branch can reduce the false-negative license plate in the indirect detection branch due to incorrectly detected vehicles. In this way, the network can detect the license plate with both high Precision and Recall rates.

4. Results

The backbone network of the DLPD and ALPD modules follows SSD512 [36], which is initialized with the ILSVRC CLS-LOC dataset [39]. The backbone network of the MLPR module is initialized with the Xavier initializer [40]. Following SSD [36], we adopt the data augmentation and hard negative mining strategies for model robustness. We train the model for 60 K iterations using Adam [41] with initial learning rate 10 4 , 0.9 β 1 momentum, 0.99 β 2 momentum, 5 × 10 4 weight decay, and batch size 32. The learning rate is decreased by 10 times at the 20K and 40K iterations.

4.1. Datasets

TILT720. We use a driving recorder to capture road videos with a resolution of 720 × 1280 , including the scenes of residential areas, highways, and expressways. After keyframe extraction and deduplication, we get 1033 valid images. We carefully annotate all the visible vehicles and license plates, including their subordinate relationships. The vehicle is annotated with the top-left and bottom-right points, forming a horizontal bounding box. The license plate is annotated with the four corners, forming a quadrilateral bounding box. The horizontal bounding box of the license plate is the minimal horizontal bounding rectangle of the quadrilateral bounding box. For simplicity, we name this dataset TILT720 (mulTidirectional lIcense pLate deTection dataset 720P). All the images are randomly divided into the training-validation set and test set in the proportion of 9:1.
TILT1080. Similar to the TILT720, we obtain the TILT1080 with another driving recorder. The TILT1080 contains 4112 images, and all the images have a size of 1080 × 1920 . All the images are randomly divided into the training-validation set and test set in the proportion of 9:1.

4.2. Evaluation Protocols

We adopt the Average Precision (AP) to evaluate the horizontal bounding box. Specifically, we use the 11-points computation of VOC2007 [42] with different IoU thresholds (i.e., 0.5 and 0.75). As shown in Figure 6a, the IoU is calculated between two horizontal boxes, i.e., IoU = C 1 A 1 + B 1 C 1 .
Moreover, we adopt the Precision, Recall, and F 1 -score to evaluate the quadrilateral bounding box. With the confidence threshold 0.5, a quadrilateral bounding box is correct only when its IoU with the quadrilateral ground-truth box is greater than a certain threshold. As shown in Figure 6b, the IoU is calculated between two quadrilateral boxes, i.e., IoU = C 2 A 2 + B 2 C 2 .

4.3. Ablation Study

As shown in Table 1, we adopt the ALPD module as the benchmark model. The ALPD module is described in Section 3.2.1, which is the first step of the indirect detection branch and can approximately estimate the license plate in the input image. The module only achieves very low AP on all the test sets, especially for the IoU threshold 0.75. After only adding the MLPR module, the detection performance worsens because the license plate is refined in the region that cannot completely enclose the license plate. According to [17], we further add the LREA module, where the license plate region is expanded to 3 times. In this way, the license plate can be refined in the local region that can completely enclose the license plate with a little background. The ALPD, LREA, and MLPR modules assemble the indirect detection branch, improving the AP by 10%–20% with different IoU thresholds compared with the ALPD module.
The DLPD module can directly detect the license plate in the input image, which achieves comparable performance with the indirect detection branch with a small IoU threshold; however, with a large IoU threshold, the performance is much lower. The DLPD module cannot accurately localize the license plate in the large input image because of more background noises, making it difficult to detect the small-sized license plate.
Combining the indirect and direct detection branches, we get the whole detection network, which achieves higher AP on all the test sets with different IoU thresholds. The network can detect the small-sized license plate via vehicle–plate relation and reduce the false-negative license plate caused by incorrectly detected vehicles.
Moreover, the ALPD module, the indirect detection branch, and the whole detection network have almost the same vehicle detection performance as the vanilla SSD [36]. This way, it proves our method can continuously improve the license plate detection performance while maintaining the vehicle detection performance [43,44,45,46,47].

4.4. Evaluation of Horizontal Bounding Box

We do not consider the orientation of the license plate and calculate the AP based on the detected horizontal bounding box in this subsection. We compare [9,14,17,27,29,36,48] with our proposed method. The backbone network and input size of Faster R-CNN [27], SSD [36], and the method in [17] are the same as our method, while the settings of methods [9,14,29,48] remain unchanged. Except for [14] (The authors released models for license plate detection at https://github.com/sergiomsilva/alpr-unconstrained (accessed on 3 February 2021).), all other methods are trained with the trainval set of TILT720 and TILT1080, respectively. As shown in Table 2, our method achieves the best performance for all the test sets and IoU thresholds. Moreover, as shown in Figure 7, our method has the best performance considering the area under the curve (AUC) and achieves the highest Recall rate according to the Recall-axis. SSD [36] can directly detect the license plate in the input image and achieve comparable performance with our method with a small IoU threshold. However, with a large IoU threshold, SSD [36] significantly lags because the background noises from the large input image can disturb the detection of the license plate. The method [17] can significantly improve the AP by detecting the license plate in the local region around the license plate, which can greatly reduce the background noises; nevertheless, it will inevitably fail the license plate if the vehicle fails to be correctly detected. Our method combines the advantages of method [17] and SSD [36] by integrating two detection branches, i.e., indirect branch and direct branch. Same as [17], the indirect detection branch can detect the license plate in the local region. Furthermore, the direct detection branch can reduce the false-negative license plate in the indirect detection branch due to incorrectly detected vehicles. In this way, our method achieves higher Precision and Recall rates compared with the method in [17] and SSD [36].

4.5. Evaluation of Multidirectional License Plate

We calculate the Precision, Recall, and F 1 -score based on the predicted quadrilateral bounding box. For the methods [9,27,29,36,48] that can only detect the horizontal bounding box, we only compare the best SSD [36] with our proposed method. As shown in Table 3, our method achieves the best F 1 -score for all the test sets with different IoU thresholds. SSD [36] achieves relatively poor performance, because the detection results of SSD [36] have very low IoU with the quadrilateral ground-truth box. Furthermore, like the DLPD module, we upgrade SSD [36] and make it capable of directly detecting the four corners of the license plate in the input image (SSD+FC). SSD+FC can achieve much better performance than the vanilla SSD [36], especially for the large IoU threshold. However, SSD+FC suffers low Recall because of the background noises. As shown in Section 4.4, our method combines the advantages of method [17] and SSD+FC, and can precisely detect the multidirectional license plate with a higher Recall rate.

4.6. Evaluation of Small-Sized License Plate

According to the height of the license plate, we divide the test set into three parts, i.e., small, medium, and large. To avoid large deviation, we define the height of the multidirectional license plate as H Q b b o x in Figure 8b. As for our datasets, we define the small LP with h e i g h t ( 0 , 16 ] pixels, the medium LP with h e i g h t ( 16 , 32 ] pixels, and the large LP with h e i g h t ( 32 , + ) pixels.
We use the Recall to evaluate the multiscale detection performance. As shown in Table 4, our method achieves the best Recall rate for almost all the sizes of different datasets with different IoU thresholds. Compared with the benchmark models SSD [36] and SSD+FC, our method achieves a large performance improvement, especially for the small- and medium-sized license plate. Same as [17], our method can effectively detect the small-sized license plate in the local region, which greatly improves the Recall rate. Furthermore, the direct detection branch can reduce the false-negative license plate due to incorrectly detected vehicles, which further improves the Recall rate based on method [17]. However, the post-processing module may remove true-positive predictions (i.e., 30.53 vs. 29.47 of TILT720), and this is what we should improve in future work.

4.7. Qualitative Results

Some qualitative detection results are illustrated in Figure 9. The license plate can be detected via vehicle–plate relation in the indirect detection branch, especially for the small-sized license plate. However, when many vehicles are close to each other, some vehicles may be detected with a large deviation, as shown in the first two images. In addition, in some cases, the vehicle fails to be detected due to boundary truncation, as shown in the third image. In these cases, the license plate cannot be detected in the indirect detection branch. Meanwhile, the license plate can be directly detected in the input image in the direct detection branch. However, due to the disturbing background noises, the direct detection branch can only detect relatively large and horizontal license plate.
By combing these two detection branches with post-processing operations, such as NMS, we get the final detection results. As can be seen, these two detection branches are complementary to each other. The indirect detection branch can detect most of the license plates; in some cases, the direct detection branch can reduce the false-negative license plate in the indirect detection branch due to undetected vehicles or vehicles with large deviations.

5. Discussions

In summary, we have verified the effectiveness of our proposed method to detect multiscale and multidirectional license plates. The indirect detection branch can detect most license plates via vehicle–plate relation. The direct detection branch can reduce false-negative license plates when the vehicle is wrongly detected in the indirect detection branch. Both detection branches can detect multidirectional license plates by regressing the four corners of the license plate. After tilt correction, we can improve the license plate recognition performance [19,20,21,22]. The license plate information can be applied to barrier access control [20,22], vehicle target detection [49], vehicle re-identification [50], etc. Moreover, the location of the license plate can be used for vehicle trajectory prediction [51] via license plate detection and tracking.
However, in some cases, the proposed method fails to detect the license plate, and Figure 10 illustrates some failed examples. In these cases, both the indirect and direct detection branches fail to detect the license plate, especially the indirect detect branch. As shown in Figure 10a, the license plate of the middle pick-up truck is undetected due to various illuminations caused by the mirror reflection on the front windshield, which may be caused by never seeing such images during training. As shown in Figure 10b, the license plate of the leftmost black vehicle is undetected due to the miss of vehicles’ detection. The two close vehicles are detected with only one box due to vehicle occlusion or boundary truncation. We will improve the vehicle detection performance to enhance the LPD performance in future work. As shown in Figure 10c, the left two license plates are undetected due to various orientations, which may be caused by a large deviation of the center offset between the vehicle and license plate. In this case, the approximate location of the license plate is wrongly estimated, so the next refinement stage cannot localize the license plate.

6. Conclusions

We propose an end-to-end trainable network for license plate detection, which can effectively detect the small-sized license plate and accurately localize the multidirectional license plate in real applications. The network is composed of two detection branches, i.e., indirect branch and direct branch. The indirect detection branch can detect the license plate via vehicle–plate relation in a coarse-to-fine scheme. The direct detection branch can directly detect the license plate in the input image. All these branches can detect multidirectional license plates by regressing the four corners of the license plate. The final detection results are obtained by merging these two detection branches via post-processing operations, such as NMS. To our knowledge, our proposed method is the first one that combines indirect and direct methods into an end-to-end network for license plate detection. Experiments show that the indirect detection branch can detect most license plates, especially the small-sized license plate. The direct detection branch can reduce the false-negative license plate in the indirect detection branch due to incorrectly detected vehicles. In this way, our proposed method achieves both high Precision and Recall rates.

Author Contributions

Conceptualization, S.-L.C. and C.Y.; methodology, S.-L.C., J.-W.M. and C.Y.; software, S.-L.C.; validation, S.-L.C., Q.L. and J.-W.M.; formal analysis, S.-L.C.; investigation, S.-L.C.; resources, S.-L.C. and C.Y.; data curation, S.-L.C., Q.L. and J.-W.M.; writing—original draft preparation, S.-L.C.; writing—review and editing, Q.L.; visualization, S.-L.C.; supervision, C.Y.; project administration, S.-L.C.; funding acquisition, C.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China grant number 61806017, the Beijing Municipal Natural Science Foundation grant number 4194084, the China Postdoctoral Science Foundation grant number 2018M641199, and the Fundamental Research Funds for the Central Universities grant number FRF-TP-18-060A1.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Acknowledgments

The authors gratefully acknowledge Shu Tian, Feng Chen, and Xu-Cheng Yin for their time and invaluable suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LPLicense plate
LPDLicense plate detection
CRFConditional random field
RPNRegion proposal network
STNSpatial transformer networks
IoUIntersection over union
RoIRegion of interest
SSDSingle shot multibox detector
YOLOYou only look once
VGGVisual geometry group
NMSNon-maximum suppression
DLPDDirect license plate detection
ALPDApproximate license plate detection
LREALocal region estimation and aggregation
MLPRMultidirectional license plate refinement
FCFour corners

References

  1. Tian, Y.; Song, J.; Zhang, X.; Shen, P.; Zhang, L.; Gong, W.; Wei, W.; Zhu, G. An algorithm combined with color differential models for license-plate location. Neurocomputing 2016, 212, 22–35. [Google Scholar] [CrossRef]
  2. Zhou, W.; Li, H.; Lu, Y.; Tian, Q. Principal visual word discovery for automatic license plate detection. IEEE Trans. Image Process. 2012, 21, 4269–4279. [Google Scholar] [CrossRef] [PubMed]
  3. Li, B.; Tian, B.; Li, Y.; Wen, D. Component-based license plate detection using conditional random field Model. IEEE Trans. Intell. Transp. Syst. 2013, 14, 1690–1699. [Google Scholar] [CrossRef]
  4. Al-Ghaili, A.M.; Mashohor, S.; Ramli, A.R.; Ismail, A. Vertical-edge-based car-license-plate detection method. IEEE Trans. Veh. Technol. 2013, 62, 26–38. [Google Scholar] [CrossRef] [Green Version]
  5. Han, J.; Yao, J.; Zhao, J.; Tu, J.; Liu, Y. Multi-Oriented and Scale-Invariant License Plate Detection Based on Convolutional Neural Networks. Sensors 2019, 19, 1175. [Google Scholar] [CrossRef] [Green Version]
  6. Yuan, Y.; Zou, W.; Zhao, Y.; Wang, X.; Hu, X.; Komodakis, N. A robust and efficient approach to license plate detection. IEEE Trans. Image Process. 2017, 26, 1102–1114. [Google Scholar] [CrossRef]
  7. Al-qudah, R.; Suen, C.Y. Enhancing YOLO deep networks for the detection of license plates in complex scenes. In Proceedings of the Second International Conference on Data Science, E-Learning and Information Systems (DATA), Dubai, United Arab Emirates, 2–5 December 2018; pp. 1–6. [Google Scholar] [CrossRef]
  8. Pu, D.; Gu, N.; Zhang, X. A robust and real-time approach for license plate detection. In Proceedings of the International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD), Huangshan, China, 28–30 July 2018; pp. 1–7. [Google Scholar] [CrossRef]
  9. Chen, S.; Yang, C.; Ma, J.; Chen, F.; Yin, X. Simultaneous end-to-end vehicle and license plate detection with multi-branch attention neural network. IEEE Trans. Intell. Transp. Syst. 2020, 21, 3686–3695. [Google Scholar] [CrossRef]
  10. Xu, Z.; Yang, W.; Meng, A.; Lu, N.; Huang, H.; Ying, C.; Huang, L. Towards end-to-end license plate detection and recognition: A large dataset and baseline. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 261–277. [Google Scholar] [CrossRef]
  11. Li, H.; Wang, P.; Shen, C. Towards end-to-end car license plates detection and recognition with deep neural networks. IEEE Trans. Intell. Transp. Syst. 2019, 20, 1126–1136. [Google Scholar] [CrossRef]
  12. SG, K.; HG, J.; HI, K. Deep-learning-based license plate detection method using vehicle region extraction. Electron. Lett. 2017, 53, 1034–1036. [Google Scholar]
  13. Fu, Q.; Shen, Y.; Guo, Z. License plate detection using deep cascaded convolutional neural networks in complex scenes. In Proceedings of the 24th International Conference on Neural Information Processing (ICONIP), Guangzhou, China, 14–18 November 2017; pp. 696–706. [Google Scholar] [CrossRef]
  14. Silva, S.M.; Jung, C.R. License plate detection and recognition in unconstrained scenarios. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 593–609. [Google Scholar] [CrossRef]
  15. Yonetsu, S.; Iwamoto, Y.; Chen, Y. Two-stage YOLOv2 for accurate license-plate detection in complex scenes. In Proceedings of the IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 11–13 January 2019; pp. 1–4. [Google Scholar] [CrossRef]
  16. Laroca, R.; Severo, E.; Zanlorensi, L.A.; Oliveira, L.S.; Gonçalves, G.R.; Schwartz, W.R.; Menotti, D. A robust real-time automatic license plate recognition based on the YOLO detector. In Proceedings of the International Joint Conference on Neural Network (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–10. [Google Scholar] [CrossRef] [Green Version]
  17. Chen, S.L.; Tian, S.; Ma, J.W.; Liu, Q.; Yang, C.; Chen, F.; Yin, X.C. End-to-end trainable network for degraded license plate detection via vehicle-plate relation mining. arXiv 2020, arXiv:2010.14266v1. [Google Scholar]
  18. Silva, S.M.; Jung, C.R. Real-time brazilian license plate detection and recognition using deep convolutional neural networks. In Proceedings of the Conference on Graphics, Patterns and Images (SIBGRAPI), Niterói, Brazil, 17–20 October 2017; pp. 55–62. [Google Scholar] [CrossRef]
  19. Jin, L.; Xian, H.; Bie, J.; Sun, Y.; Hou, H.; Niu, Q. License Plate Recognition Algorithm for Passenger Cars in Chinese Residential Areas. Sensors 2012, 12, 8355–8370. [Google Scholar] [CrossRef]
  20. Ullah, F.; Anwar, H.; Shahzadi, I.; Rehman, A.U.; Mehmood, S.; Niaz, S.; Awan, K.M.; Khan, A.; Kwak, D. Barrier Access Control Using Sensors Platform and Vehicle License Plate Characters Recognition. Sensors 2019, 19, 3015. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Do Vale Dalarmelina, N.; Teixeira, M.A.; Meneguette, R.I. A Real-Time Automatic Plate Recognition System Based on Optical Character Recognition and Wireless Sensor Networks for ITS. Sensors 2020, 20, 55. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Islam, K.T.; Raj, R.G.; Islam, S.M.S.; Wijewickrema, S.N.R.; Hossain, M.S.; Razmovski, T.; O’Leary, S.J. A Vision-Based Machine Learning Method for Barrier Access Control Using Vehicle License Plate Authentication. Sensors 2020, 20, 3578. [Google Scholar] [CrossRef]
  23. Dong, M.; He, D.; Luo, C.; Liu, D.; Zeng, W. A CNN-Based Approach for Automatic License Plate Recognition in the Wild. In Proceedings of the British Machine Vision Conference (BMVC), London, UK, 4–7 September 2017. [Google Scholar]
  24. Chen, S.L. Open Source Code for the Network Combining Indirect and Direct Branches. Github (2020). Available online: https://github.com/chensonglu/LPD-indirect-direct (accessed on 3 February 2021).
  25. Hsieh, J.; Yu, S.; Chen, Y. Morphology-based license plate detection in images of differently illuminated and oriented cars. J. Electron. Imaging 2002, 11, 507–516. [Google Scholar] [CrossRef]
  26. Zheng, D.; Zhao, Y.; Wang, J. An efficient method of license plate location. Pattern Recognit. Lett. 2005, 26, 2431–2438. [Google Scholar] [CrossRef]
  27. Ren, S.; He, K.; Girshick, R.B.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proceedings of the Annual Conference on Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  28. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar] [CrossRef] [Green Version]
  29. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar] [CrossRef] [Green Version]
  30. Xie, L.; Ahmad, T.; Jin, L.; Liu, Y.; Zhang, S. A new CNN-based method for multi-directional car license plate detection. IEEE Trans. Intell. Transp. Syst. 2018, 19, 507–517. [Google Scholar] [CrossRef]
  31. Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A. You Only Look Once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef] [Green Version]
  32. Tian, J.; Wang, G.; Liu, J. Semantic region proposals for adaptive license plate detection in open environment. J. Electron. Imaging 2019, 28, 023017. [Google Scholar] [CrossRef]
  33. Lu, N.; Yang, W.; Meng, A.; Xu, Z.; Huang, H.; Huang, L. Automatic Recognition for Arbitrarily Tilted License Plate. In Proceedings of the 2nd International Conference on Video and Image Processing (ICVIP), Hong Kong, China, 29–31 December 2018; pp. 23–28. [Google Scholar] [CrossRef]
  34. Jaderberg, M.; Simonyan, K.; Zisserman, A.; Kavukcuoglu, K. Spatial Transformer Networks. In Proceedings of the Annual Conference on Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 7–12 December 2015; pp. 2017–2025. [Google Scholar]
  35. Dai, J.; He, K.; Sun, J. Instance-aware semantic segmentation via multi-task network cascades. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3150–3158. [Google Scholar] [CrossRef] [Green Version]
  36. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.E.; Fu, C.Y.; Berg, A.C. SSD: Single shot multiBox detector. In Proceedings of the 14th European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar] [CrossRef] [Green Version]
  37. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  38. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  39. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  40. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics (AISTATS), Chia Laguna Resort, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  41. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  42. Everingham, M.; Gool, L.J.V.; Williams, C.K.I.; Winn, J.M.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
  43. Wu, Y.; Deng, L.; He, W. BwimNet: A Novel Method for Identifying Moving Vehicles Utilizing a Modified Encoder-Decoder Architecture. Sensors 2020, 20, 7170. [Google Scholar] [CrossRef]
  44. Shieh, J.; Haq, M.A.; Karam, S.; Chondro, P.; Gao, D.; Ruan, S. Continual Learning Strategy in One-Stage Object Detection Framework Based on Experience Replay for Autonomous Driving Vehicle. Sensors 2020, 20, 6777. [Google Scholar] [CrossRef]
  45. Stuparu, D.; Ciobanu, R.; Dobre, C. Vehicle Detection in Overhead Satellite Images Using a One-Stage Object Detection Model. Sensors 2020, 20, 6485. [Google Scholar] [CrossRef]
  46. Barbosa, R.C.; Ayub, M.S.; Rosa, R.L.; Rodríguez, D.Z.; Wuttisittikulkij, L. Lightweight PVIDNet: A Priority Vehicles Detection Network Model Based on Deep Learning for Intelligent Traffic Lights. Sensors 2020, 20, 6218. [Google Scholar] [CrossRef]
  47. Chen, X.; Chang, C.; Yu, C.; Chen, Y. A Real-Time Vehicle Detection System under Various Bad Weather Conditions Based on a Deep Learning Model without Retraining. Sensors 2020, 20, 5731. [Google Scholar] [CrossRef] [PubMed]
  48. Liao, M.; Shi, B.; Bai, X.; Wang, X.; Liu, W. TextBoxes: A fast text detector with a single deep neural network. In Proceedings of the 31st AAAI Conference on Artificial Intelligence (AAAI), San Francisco, CA, USA, 4–9 February 2017; pp. 4161–4167. [Google Scholar]
  49. Wang, H.; He, S.; Yu, J.; Wang, L.; Liu, T. Research and Implementation of Vehicle Target Detection and Information Recognition Technology Based on NI myRIO. Sensors 2020, 20, 1765. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Chen, X.; Sui, H.; Fang, J.; Zhou, M.; Wu, C. Multi-Proxy Constraint Loss for Vehicle Re-Identification. Sensors 2020, 20, 5142. [Google Scholar] [CrossRef] [PubMed]
  51. Zhang, Z.; Liu, H.; Rai, L.; Zhang, S. Vehicle Trajectory Prediction Method Based on License Plate Information Obtained from Video-Imaging Detectors in Urban Road Environment. Sensors 2020, 20, 1258. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. License plates with various sizes, orientations, and backgrounds in real scenarios. All the recognizable license plates are manually covered with a yellow ellipsoid to protect privacy.
Figure 1. License plates with various sizes, orientations, and backgrounds in real scenarios. All the recognizable license plates are manually covered with a yellow ellipsoid to protect privacy.
Sensors 21 01074 g001
Figure 2. A thumbnail of the overall architecture is shown in the top-left corner (DLPD: Direct License Plate Detection; ALPD: Approximate License Plate Detection; LREA: Local Region Estimation and Aggregation; MLPR: Multidirectional License Plate Refinement). All the recognizable license plates are manually covered with a yellow ellipsoid to protect privacy.
Figure 2. A thumbnail of the overall architecture is shown in the top-left corner (DLPD: Direct License Plate Detection; ALPD: Approximate License Plate Detection; LREA: Local Region Estimation and Aggregation; MLPR: Multidirectional License Plate Refinement). All the recognizable license plates are manually covered with a yellow ellipsoid to protect privacy.
Sensors 21 01074 g002
Figure 3. The four corners of the license plate are regressed based on the offsets (red arrow) from the center (orange circle) of the matched anchor box (dashed orange rectangle). The matched anchor box is determined by the intersection over union (IoU) with the horizontal ground-truth box (solid green rectangle). The negative anchor box (dashed gray rectangle) is neglected due to low IoU. The license plate is manually covered with a yellow ellipsoid to protect privacy.
Figure 3. The four corners of the license plate are regressed based on the offsets (red arrow) from the center (orange circle) of the matched anchor box (dashed orange rectangle). The matched anchor box is determined by the intersection over union (IoU) with the horizontal ground-truth box (solid green rectangle). The negative anchor box (dashed gray rectangle) is neglected due to low IoU. The license plate is manually covered with a yellow ellipsoid to protect privacy.
Sensors 21 01074 g003
Figure 4. Training loss. L 1 ( c , p , g ) is the loss of the ALPD module in the indirect detection branch. L 2 ( c , p , g ) is the loss of the MLPR module in the indirect detection branch. L 3 ( c , p , g ) is the loss of the DLPD module in the direct detection branch. L is the total loss of the end-to-end network.
Figure 4. Training loss. L 1 ( c , p , g ) is the loss of the ALPD module in the indirect detection branch. L 2 ( c , p , g ) is the loss of the MLPR module in the indirect detection branch. L 3 ( c , p , g ) is the loss of the DLPD module in the direct detection branch. L is the total loss of the end-to-end network.
Sensors 21 01074 g004
Figure 5. Post-processing operations. The green rectangles are the possible detection results predicted by the network. After threshold filtering and non-maximum suppression (NMS), we can get the final detection results. The license plate is manually covered with a yellow ellipsoid to protect privacy.
Figure 5. Post-processing operations. The green rectangles are the possible detection results predicted by the network. After threshold filtering and non-maximum suppression (NMS), we can get the final detection results. The license plate is manually covered with a yellow ellipsoid to protect privacy.
Sensors 21 01074 g005
Figure 6. (a) IoU between two horizontal boxes. (b) IoU between two quadrilateral boxes.
Figure 6. (a) IoU between two horizontal boxes. (b) IoU between two quadrilateral boxes.
Sensors 21 01074 g006
Figure 7. The Precision–Recall curve of different methods. The title of each graph indicates the dataset and IoU threshold for testing. Our method achieves the best performance for all the test sets and IoU thresholds in terms of the area under the curve (AUC). Moreover, our method achieves the best Recall rate, according to the Recall-axis.
Figure 7. The Precision–Recall curve of different methods. The title of each graph indicates the dataset and IoU threshold for testing. Our method achieves the best performance for all the test sets and IoU thresholds in terms of the area under the curve (AUC). Moreover, our method achieves the best Recall rate, according to the Recall-axis.
Sensors 21 01074 g007
Figure 8. (a) The height of the horizontal bounding box H H b b o x . (b) The distance between the top-left corner and the straight line formed by the bottom-left and bottom-right corners H Q b b o x . In this work, the height of the license plate is defined as H Q b b o x . The license plate is manually covered with a yellow ellipsoid to protect privacy.
Figure 8. (a) The height of the horizontal bounding box H H b b o x . (b) The distance between the top-left corner and the straight line formed by the bottom-left and bottom-right corners H Q b b o x . In this work, the height of the license plate is defined as H Q b b o x . The license plate is manually covered with a yellow ellipsoid to protect privacy.
Sensors 21 01074 g008
Figure 9. Detection results. The orange rectangle represents the horizontal bounding box of the vehicle. The red quadrangle indicates the quadrilateral bounding box of the license plate. Our method merges the detection results of the indirect and direct detection branches. All the recognizable license plates are manually covered with a yellow ellipsoid to protect privacy.
Figure 9. Detection results. The orange rectangle represents the horizontal bounding box of the vehicle. The red quadrangle indicates the quadrilateral bounding box of the license plate. Our method merges the detection results of the indirect and direct detection branches. All the recognizable license plates are manually covered with a yellow ellipsoid to protect privacy.
Sensors 21 01074 g009
Figure 10. Failed examples. The orange rectangle represents the horizontal bounding box of the vehicle. The red quadrangle indicates the quadrilateral bounding box of the license plate. (a) Failed due to various illuminations. (b) Failed due to vehicle occlusion or boundary truncation. (c) Failed due to various orientations. All the recognizable license plates are manually covered with a yellow ellipsoid to protect privacy.
Figure 10. Failed examples. The orange rectangle represents the horizontal bounding box of the vehicle. The red quadrangle indicates the quadrilateral bounding box of the license plate. (a) Failed due to various illuminations. (b) Failed due to vehicle occlusion or boundary truncation. (c) Failed due to various orientations. All the recognizable license plates are manually covered with a yellow ellipsoid to protect privacy.
Sensors 21 01074 g010
Table 1. Ablation study of different datasets with different IoU thresholds. The values represent the Average Precision (AP) based on the horizontal bounding box.
Table 1. Ablation study of different datasets with different IoU thresholds. The values represent the Average Precision (AP) based on the horizontal bounding box.
MethodLREAMLPRDLPDIoU = 0.5IoU = 0.75
TILT720TILT1080TILT720TILT1080
ALPD 76.71%77.71%26.27%35.27%
Indirect 40.35%40.62%7.48%10.76%
89.19%87.67%54.51%56.92%
Direct 86.85%86.01%47.52%53.34%
Two-branch89.30%87.79%56.54%57.94%
Table 2. Comparative experiments of the horizontal bounding box. The values represent the AP based on the horizontal bounding box.
Table 2. Comparative experiments of the horizontal bounding box. The values represent the AP based on the horizontal bounding box.
MethodIoU = 0.5IoU = 0.75
TILT720TILT1080TILT720TILT1080
Faster R-CNN [27]81.65%73.88%13.63%14.29%
TextBoxes [48]69.67%67.56%37.24%38.66%
Method [14]74.67%64.78%42.67%38.61%
Method [9]84.05%82.05%45.35%53.42%
YOLOv2 [29]80.80%79.58%51.66%49.32%
SSD [36]86.63%86.34%47.06%53.88%
Method [17]89.19%87.67%54.51%56.92%
Ours (Direct)86.85%86.01%47.52%53.34%
Ours (Indirect)89.13%87.11%54.48%56.96%
Ours (Two-branch)89.30%87.79%56.54%57.94%
Table 3. Comparative experiments of the multidirectional license plate. The values are calculated based on the quadrilateral bounding box.
Table 3. Comparative experiments of the multidirectional license plate. The values are calculated based on the quadrilateral bounding box.
MethodTILT720 (IoU = 0.5/0.75)TILT1080 (IoU = 0.5/0.75)
PrecisionRecall F 1 -ScorePrecisionRecall F 1 -Score
SSD [36]98.66/65.1058.80/38.8073.68/48.6293.88/75.9240.38/30.0756.47/43.08
Method [14]88.79/53.2776.00/45.6081.90/49.1483.53/55.0868.97/45.4875.55/49.83
SSD+FC97.47/75.3261.60/47.6075.49/58.3397.57/84.6742.61/36.9859.32/51.48
Method [17]90.61/60.4188.80/59.2089.70/59.8088.17/61.5187.89/61.3288.03/61.42
Ours (Direct)98.69/82.3160.40/48.4074.94/60.9696.96/85.9544.00/39.0060.53/53.66
Ours (Indirect)88.93/60.8790.00/61.6089.46/61.2388.72/61.6587.78/61.0088.25/61.32
Ours (Two-branch)89.68/61.9090.40/62.4090.04/62.1587.85/62.0989.16/63.0288.50/62.55
Table 4. Comparative experiments of the multiscale license plate. The values represent the Recall based on the quadrilateral bounding box.
Table 4. Comparative experiments of the multiscale license plate. The values represent the Recall based on the quadrilateral bounding box.
MethodTILT720 (IoU = 0.5/0.75)TILT1080 (IoU = 0.5/0.75)
LargeMediumSmallLargeMediumSmall
SSD [36]88.46/69.2374.42/56.5929.47/6.3276.43/59.2445.28/34.4510.87/5.43
Method [14]92.31/88.4685.27/62.0258.95/11.5888.54/82.8078.74/54.3339.86/7.97
SSD+FC96.15/92.3177.52/65.1230.53/11.5880.25/73.2547.83/42.1311.59/6.88
Method [17]96.15/88.4698.45/78.2973.68/25.2699.36/86.6296.65/77.1765.22/17.75
Ours (Direct)96.15/92.3179.07/67.4425.26/10.5382.17/76.4350.20/45.4710.87/5.80
Ours (Indirect)96.15/88.4698.45/79.0776.84/30.5399.36/84.7196.65/77.1764.86/17.75
Ours (Two-branch)96.15/92.3198.45/80.6277.89/29.47100.00/86.6298.43/80.1265.94/18.12
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, S.-L.; Liu, Q.; Ma, J.-W.; Yang, C. Scale-Invariant Multidirectional License Plate Detection with the Network Combining Indirect and Direct Branches. Sensors 2021, 21, 1074. https://doi.org/10.3390/s21041074

AMA Style

Chen S-L, Liu Q, Ma J-W, Yang C. Scale-Invariant Multidirectional License Plate Detection with the Network Combining Indirect and Direct Branches. Sensors. 2021; 21(4):1074. https://doi.org/10.3390/s21041074

Chicago/Turabian Style

Chen, Song-Lu, Qi Liu, Jia-Wei Ma, and Chun Yang. 2021. "Scale-Invariant Multidirectional License Plate Detection with the Network Combining Indirect and Direct Branches" Sensors 21, no. 4: 1074. https://doi.org/10.3390/s21041074

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop