Next Article in Journal
Infrastructure-Wide and Intent-Based Networking Dataset for 5G-and-beyond AI-Driven Autonomous Networks
Previous Article in Journal
Monitoring Off-Shore Fishing in the Northern Indian Ocean Based on Satellite Automatic Identification System and Remote Sensing Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Autonomous Detection of Humans in Off-Limits Mountain Areas

System Engineering Department, Sejong University, Seoul 05006, Republic of Korea
Sensors 2024, 24(3), 782; https://doi.org/10.3390/s24030782
Submission received: 4 January 2024 / Revised: 15 January 2024 / Accepted: 24 January 2024 / Published: 25 January 2024
(This article belongs to the Section Intelligent Sensors)

Abstract

:
This paper is on the autonomous detection of humans in off-limits mountains. In off-limits mountains, a human rarely exists, thus human detection is an extremely rare event. Due to the advances in artificial intelligence, object detection–classification algorithms based on a Convolution Neural Network (CNN) can be used for this application. However, considering off-limits mountains, there should be no person in general. Thus, it is not desirable to run object detection–classification algorithms continuously, since they are computationally heavy. This paper addresses a time-efficient human detector system, based on both motion detection and object classification. The proposed scheme is to run a motion detection algorithm from time to time. In the camera image, we define a feasible human space where a human can appear. Once motion is detected inside the feasible human space, one enables the object classification, only inside the bounding box where motion is detected. Since motion detection inside the feasible human space runs much faster than an object detection–classification method, the proposed approach is suitable for real-time human detection with low computational loads. As far as we know, no paper in the literature used the feasible human space, as in our paper. The outperformance of our human detector system is verified by comparing it with other state-of-the-art object detection–classification algorithms (HOG detector, YOLOv7 and YOLOv7-tiny) under experiments. This paper demonstrates that the accuracy of the proposed human detector system is comparable to other state-of-the-art algorithms, while outperforming in computational speed. Our experiments show that in environments with no humans, the proposed human detector runs 62 times faster than YOLOv7 method, while showing comparable accuracy.

1. Introduction

The problem of human detection is to automatically locate people in an image or video sequence, and has been actively researched in the past decade [1,2]. Using an aerial platform, such as an Unmanned Airborne Vehicle (UAV), to perform human detection has been in the attention of researchers for a significant period of time [3].
Our paper is on the autonomous detection of a non-cooperative human in off-limits mountains. Note that we are not interested in the detection of a non-human, such as a wild animal. Our goal is to detect a human in off-limits mountains.
In off-limits mountains, a human rarely exists, thus human detection is an extremely rare event. Suppose one deploys surveillance cameras in armed border areas, in order to detect a non-cooperative human entering the area. Once a non-cooperative human is detected, then the surveillance cameras can let the system operator recognize the human automatically. Considering the maintenance of the surveillance system, it is desirable that the cameras are cheap and their operations are computationally time-efficient.
Due to the advances in artificial intelligence, object detection–classification algorithms based on a Convolution Neural Network (CNN) can be utilized for camera surveillance systems. In object detection–classification algorithms, an object, such as human, is detected and one needs to locate the object in the image. This implies that an object detection and object localization must be performed simultaneously.
There are many papers on object detection–classification algorithms. R-CNN methods [4,5] classified object proposals using deep CNN. R-CNN methods employed several innovations to improve training and testing speed, while also increasing detection accuracy. You Only Look Once (YOLO) [6,7] framed object detection as a regression problem to spatially separated bounding boxes and associated class probabilities. The authors of [8] showed that in an identical testing environment, YOLO-v3 outperforms Single Shot Detection (SSD), and Faster Region based Convolutional Neural Networks (Faster R-CNN), making it the best of the three algorithms. Reference [9] stated that both R-CNN and Fast R-CNN fail to perform real-time detection, but YOLO can perform real-time classification with good speed.
Recently, YOLOv7 and YOLOv7-tiny [10] were developed to improve the detection–classification performance as well as the inference speed of YOLO-based algorithms. The accuracy of YOLOv7 outperforms that of previous YOLOs, such as YOLOv5 or YOLOv6 [11]. However, recently developed object detection–classification algorithms require a Graphics Processing Unit (GPS) for fast processing and are computationally heavy.
Usually, no person enters off-limits mountains, except for personnel guarding the areas. Considering off-limits mountains, there should be no person in general. In off-limits mountains, there can be many wild animals, but detection of wild animals is not the scope of our paper. Therefore, it is not desirable to run object detection–classification algorithms, such as YOLO, continuously, since they are computationally heavy.
This paper addresses a time-efficient human detector system, based on both motion detection and object classification. Here, motion detection is to detect motion by comparing the previous frame with the current one by examining the pixel values. Moreover, object classification is to classify an object inside a given bounding box.
In our paper, we run a motion detection algorithm [12], which is computationally time-efficient. In practice, strong winds can generate the motion of an object. For instance, motion can be detected on various objects, such as tree leaves or squirrels on trees. Since our goal is to detect the motion of a human, the motion of a non-human can be considered as a false alarm for our motion detection algorithm.
In the camera image, we use a feasible human space where a human can appear. For instance, we may have a priori information on humans’ feasible routes. In this case, a feasible human space can be set on humans’ feasible routes. We do not use motion if it is detected outside the feasible human space.
Once motion is detected inside the feasible human space, one enables the object classification, only inside the bounding box where motion is detected. A motion detector in the feasible human space runs much faster than an object detection–classification method, such as YOLO-based methods. Thus, the proposed approach is effective for real-time human detection with low computational loads. As far as we know, no paper in the literature used the feasible human space, as in our paper.
Once motion is detected in the feasible human space, one sets a bounding box on the moving object, followed by classifying the moving object inside the box. For human classification inside the given bounding box, Support Vector Machine (SVM) can be applied to Histograms of Oriented Gradient (HOG) of an object. This method is called HOG detector [13]. However, HOG detector is only suitable for detecting a human with standing upright pose and is sensitive to bounding box size.
In our paper, MobileNetV2 [14] is applied as a CNN model structure for human classification inside the bounding box. Note that MobileNetV2 cannot localize an object in the image. The structure of MobileNetV2 is light, and its operation is fast. MobileNetV2 is selected in our paper, because we focus on time-efficient classification inside the given bounding box. We acknowledge that using MobileNetV2 is not our novel contribution. Applying other recent models, such as MobileNetV3 [15], can further outperform our classification result.
In practice, a human in off-limits mountains may move, while their body is partially occluded. The random erasing method [16] was developed to solve the image occlusion problems. In training the CNN, random erasing approach randomly selects a rectangle region in an image and erases its pixels with random values. In this process, training images with various levels of occlusion are generated, which makes the model robust to occlusion. The random erasing method is simple to implement, since we only need to train various occluded images. Our paper uses the random erasing method for training various occluded human images.
The outperformance of our human detector system is verified by comparing it with other state-of-the-art object detection–classification algorithms (HOG detector [13], YOLOv7 [10], and YOLOv7-tiny [10]) under experiments. This paper demonstrates that our human detector system is much more time-efficient than other state-of-the-art algorithms, while showing comparable accuracy. Our experiments further show that in environments with no humans, the proposed human detector runs 62 times faster than YOLOv7 method, while showing comparable accuracy.
The paper is organized as follows. Section 2 presents the literature review related to this paper. Section 3 describes the proposed human detector system. Section 4 presents experiments of the proposed human detector system. Section 5 presents the conclusions.

2. Literature Review

Reference [2] reviewed human detection in surveillance videos and its applications. Object detection could be performed using background subtraction, optical flow and spatio-temporal filtering techniques. Once detected, a moving object could be classified as a human being using shape-based, texture-based or motion-based features. Ref. [2] mentioned that texture-based features, such as HOG detector [13], outperform shape-based, or texture-based methods. Thus, we compare the performance of the proposed human detector with the HOG detector in [13].
In the camera image, we use a feasible human space where a human can appear. We do not use motion if it is detected outside the feasible human space. As far as we know, no paper in the literature used the feasible human space, as in our paper.
For solving the motion detection problem, the optical flow-based object detection technique uses characteristics of flow vectors of moving objects over time to detect moving regions in an image sequence [2,17,18]. Optical flow-based methods can be used to detect independently moving objects even in the presence of camera motion. However, a real-time implementation of optical flow requires a specialized hardware due to the complexity of the algorithm and moderately high frame rate for accurate measurements [2].
For solving the motion detection problem efficiently, various methods have been developed [19,20]. Since one requires cheap surveillance systems with static cameras, it is desirable to use a fast algorithm for motion detection. For motion detection, our paper uses the Gaussian Mixture-based Background Segmentation Algorithm (GMBS) in openCV [21]. We acknowledge that other motion detection algorithm can be used, as long as it runs fast.
For human classification inside a bounding box, Support Vector Machine (SVM) can be applied to a HOG or Scale-Invariant Feature Transform (SIFT) feature of an object [2,13,22,23]. In order to make up for the deficiency of single statistical feature, one can combine more features together for better detection performances [22,23]. However, these detectors are only suitable for detecting a human with standing upright pose and are sensitive to bounding box size.
In our paper, MobileNetV2 [14] is applied as a CNN model structure for human classification inside a bounding box at the feasible human space. In off-limits mountains, a hostile human can crawl, while not standing upright. We argue that CNN structures, such as MobileNetV2 [14] or MobileNetV3 [15], are more extendable than human classifiers in [2,13,22,23], which are specifically designed to classify a pedestrian who usually appears upright in video data.
For instance, for monitoring armed border areas using cameras, CNN structures can be trained to classify a hostile human with weapons, who can be dangerous in armed border area. We can train the CNN network, so that a human with weapons can be distinguished from a human with no weapons. By training CNN structures, we can handle detection of humans, while analyzing the human’s equipment. Human classifiers in [2,13,22,23], which are specifically designed to classify an upright pedestrian, cannot be applied for this equipment detection.
As another example, a lying human can be detected, as long as CNN structures are trained to classify a lying human. One can train CNN structures so that a lying human can be detected. Human classifiers in [2,13,22,23], which are specifically designed to classify an upright pedestrian, cannot be applied to classify a lying human.
The performance of our human detector system is verified by comparing it with other state-of-the-art object detection–classification algorithms (HOG detector [13], YOLOv7 [10], and YOLOv7-tiny [10] ) under experiments. We verify that the proposed human detector system is much more time- efficient than other state-of-the-art algorithms, while showing comparable accuracy.

3. Proposed Human Detector System

We address a time-efficient human detector system, based on both motion detection and object classification. At every  F > 0 frame, one detects a moving object by applying motion detection algorithms. Motion detection algorithms are computationally efficient. Once motion is detected, then one applies the object classification inside the ROI of the moving object.
For motion detection at every F frame, one applies the GMBS in openCV [21]. In openCV, the function name is createBackgroundSubtractorMOG2. The GMBS calculates the foreground mask performing a subtraction between the current frame and a background model, containing the static part of the scene (everything that can be considered as background given the characteristics of the observed scene).
By setting a sufficiently large F, we can further improve the time efficiency of the motion detector. However, setting a too large F may disturb the performance of the motion detector, since the static part of the scene may move as time goes on.
Once a moving object is detected, one generates a bounding box surrounding the moving object. The method of obtaining multiple bounding boxes surrounding moving objects is ConnectedComponentsWithStats function provided by OpenCV [21]. Due to clutter noise, many small bounding boxes may be falsely generated after ConnectedComponentsWithStats function is applied.
The connectedComponentsWithStats function requires the connectivity option of the function. One sets 8 as the connectivity, which implies that a cell is connected to all eight cells surrounding the cell. In Figure 1, the central cell is connected to all eight cells surrounding the cell. Here, a connected cell of the central cell is marked with an arrow.
The ConnectedComponentsWithStats function returns 2D coordinates of the box center, the width and height of the box, and the pixel area of the box. One then sorts all bounding boxes in the decreasing order of pixel area of each box. Let  S denote the set of sorted bounding boxes. Let  S [ n ] denote the n-th element (sorted bounding box) in  S S is sorted, such that the pixel area of  S [ m ] is bigger than that of  S [ n ] for any  m > n .
Once an object’s motion is detected, a bounding box surrounding the object can be detected. In the camera image, we define a feasible human space, where a human can appear in the image. We do not use a bounding box if its center is outside the feasible human space. Once a bounding box’s center is detected inside the feasible human space, one enables the object classification, only inside the bounding box. For instance, in Figure 2, a black rectangle indicates the feasible human space.
Let human bounding box indicate a box which may contain a human image. Starting from  S [ 1 ] , one checks if there is a bounding box satisfying the following human bounding box conditions:
  • The bounding box’s center exists in the feasible human space.
  • The pixel area of the bounding box is smaller than  δ m a x .
  • The pixel width of the bounding box is bigger than  δ m i n w .
  • The pixel height of the bounding box is bigger than  δ m i n h .
Here,  δ m a x δ m i n w , and  δ m i n h are set considering a viable human size in the feasible human space. For instance, we can make a human with various poses move in the feasible human space. Then, we can derive a feasible bounding box size for a human. We say that a human bounding box has a human size.
This detection continues until one checks a bounding box which does not have a sufficient number of cells inside it. One uses 300 pixels as this threshold, say  T h . This threshold  T h is set considering a feasible human size in the feasible human space.
If this threshold  T h is not set, then one needs to check every bounding box generated using the connectedComponentsWithStats function. However, this is time-consuming, since there can be a tiny bounding box. Thus, one sets a threshold for detection a human bounding box. See Algorithm 1 for our detection process.
Algorithm 1 Detect a human bounding box
1:
Let  S denote the set of sorted bounding boxes, returned by the connectedComponentsWithStats function;
2:
S is sorted, such that the pixel area of  S [ m ] is bigger than that of  S [ n ] for any  m > n ;
3:
Let N denote the number of all bounding boxes;
4:
n = 1 ;
5:
R O I = [ ] ;
6:
while the pixel area of  S [ n ] is bigger than  T h  do
7:
if  S [ n ] is a human bounding box then
8:
   R O I . a p p e n d ( S [ n ] ) ;
9:
end if
10:
n = n + 1;
11:
end while
Bounding boxes with human area size are set as Regions of Interest (ROIs). Each ROI is set as the object classification input image, so that the object classification is enabled inside the ROI. In Algorithm 1, a bounding box inside the set  R O I is used as a ROI for object classification. In Figure 2, a green rectangle indicates the ROI of the moving object (human) detected using the proposed method. In Figure 3, the GMBS and ConnectedComponentsWithStats are applied for detecting a moving object in Figure 2. In Figure 3, white area indicates the silhouette of a moving object. Once a moving object is detected in a frame, then one generates the ROI of the moving object.

3.1. Object Classification Inside a ROI

Once a ROI is found, the object classification is enabled inside the ROI. In this paper, MobileNetV2 [14] is applied as a CNN model structure for image classification. The structure of MobileNetV2 is light, and its operation is fast. MobileNetV2 is selected, because we focus on efficient computation.
This paper uses transfer learning of the CNN of MobileNetV2. The CNN’s weighted model, which was pre-trained with COCO dataset [24], is set as the CNN model. For training the network, this paper uses INRIA human dataset [25] with 1240 training images and 597 validation images.
In practice, a human in a border area may move, while their or her body is partially occluded in the woods. The random erasing method [16] was developed to solve the image occlusion problems. In training, random erasing approach randomly selects a rectangle region in an image and erases its pixels with random values. In this process, training images with various levels of occlusion are generated, which makes the model robust to occlusion. Our paper applies the random erasing method [16] to the human dataset, for handling the image occlusion problems.
The batch size was set to 16 as a learning parameter. Since this paper uses the transfer learning, the learning rate was set sufficiently low as  10 5 . Epoch is set to 20, thus the entire learning process repeats 20 times. The learning accuracy of the learning results is 98.67 percent, learning loss is 0.0375, validation accuracy is 98.66 percent, and validation loss is 0.0467. The weight model obtained as a result of learning is stored and used as a binary classifier.
We use a binary classifier, since we only need to classify whether an object is human or not. The classification layer is used to return the probability that the object is human. Let human probability denote the probability that the object in the ROI is human.
At each frame, we obtain one or more ROIs. Suppose that there are  N k ROIs at frame k. Let  R n ( k ) , where  n { 1 , 2 , , N k } , denote the n-th ROI at frame k. Let  p c n n ( R n ( k ) ) denote the human probability of  R n ( k ) , as the CNN is applied to the ROI  R n ( k ) . In the case where  p c n n ( R n ( k ) ) > t h r e s , we assume that  R n ( k ) contains a human. Here,  t h r e s is a tuning parameter, and it can be set as any value in the interval  [ 0 , 1 ] .

4. Test Experiments

For verification of our human detection system, the hardware specifications are as follows: 12th Gen Intel(R) Core(TM) i7-12700K 3.60 GHz, 32 GB. Recall that in the case where  p c n n ( R n ( k ) ) > t h r e s , we assume that  R n ( k ) contains a human. Here,  t h r e s is a tuning parameter, and it can be set as any value in the interval  [ 0 , 1 ] . In our experiments, we compare between two cases where  t h r e s = 0.6 and  t h r e s = 0.3 , respectively.
For detecting a human bounding box, we use  δ m i n w = 10, and  δ m i n h = 40 , and  δ m a x = 10,000 pixels in the experiments. This box size can be determined by measuring humans moving in the feasible human space. At every  F = 5 frame, one detects a moving object by applying motion detection algorithms.
We compare the proposed human detection system with HOG detector [13], YOLOv7 [10], and YOLOv7-tiny [10]. Both YOLOv7 and YOLOv7-tiny have weighted models, which are pre-trained with COCO dataset [10,24]. The training of these algorithms is conducted using the identical dataset, which is used in the proposed algorithm. See Section 3.1.
In our experiments,  P r o 3 indicates the case where our proposed human detection algorithm uses  t h r e s = 0.3 P r o 6 indicates the case where our proposed human detection algorithm uses  t h r e s = 0.6 . Both YOLOv7 (3) and YOLOv7-tiny (3) use  t h r e s = 0.3 . Both YOLOv7 (6) and YOLOv7-tiny (6) use  t h r e s = 0.6 . The HOG detector in [13] is only used for human detection.
To analyze the computation time of compared algorithms, we use frame-per-second ( F P S ). Here,  F P S is computed as total number of frames divided by total computation time. In the proposed human detection system, we enable the object classification (MobileNetV2), only inside a bounding box where motion is detected in the feasible object space. The computation time for objection classification (MobileNetV2) is evaluated as  F P S c . Here,  F P S c is computed as total number of objection classification frames divided by total objection classification time. Note that  F P S c is only used in the proposed human detection system.
A classification model performance evaluation index is used to measure the classification accuracy. Among all bounding boxes, True Positive (TP) denotes the number of boxes in which a human is classified as a human. Among all bounding boxes, False Positive (FP) denotes the number of boxes in which a non-human is classified as a human. Among all bounding boxes, False Negative (FN) denotes the number of boxes in which a non-human is classified as a non-human. Among all bounding boxes, True Negative (TN) denotes the number of boxes in which a human is classified as a non-human.
Based on the experiments, the classification accuracy is measured using
A c c u r a c y = T P + F N T P + T N + F P + F N .
It is desirable that  A c c u r a c y in (1) is as close to 1 as possible.
A bounding box can be generated at a non-human, and the box can be falsely classified as a human. This case is associated with  F P . In practice, a bounding box can be generated at a partial part of a human, and the box can be falsely classified as a non-human. This case is associated with  T N .
As far as we know, there is no open dataset for detecting a person in mountains. For testing our detection system, we made experiment videos, which include a human moving in mountains. As a human moves in mountains, he or she may be occluded due to trees or grass. This makes our experiments challenging, even for the state-of-the-art object detection–classification algorithms. Our test dataset can be provided once requested.

4.1. Video File 1 of a Person Moving in Mountains

Our human detection system was tested with a video file, which shows a person passing by in 1920 × 1080 resolution. The video file recorded a person moving in mountains. One human image in the video file is plotted in Figure 2. In Figure 2, a green box indicates the ROI of the moving human detected using the proposed method.
Table 1 shows the classification accuracy and computational load of the proposed human detection system, compared to other state-of-the-art methods. In this table,  F P S is used to indicate the computational speed.  F P S c is only applied for the proposed method. Table 1 verifies that the proposed method is comparable to other state-of-the-art YOLO-based methods considering  A c c u r a c y . In addition, the proposed method considerably outperforms YOLO-based methods considering  F P S (computational speed). See that the accuracy of HOG detector is considerably lower than other methods, since the HOG detector does not use CNN structures.

4.2. Video File 2 of a Person Moving in Mountains

Our human detection system was tested with a video file, which shows a person moving in mountains. One human image in the video file is plotted in Figure 4. Once a moving object is detected, then one generates the ROI of the moving object. In Figure 4, a green box indicates the moving human occluded by trees.
Table 2 shows the classification accuracy and computational load of the proposed human detection system, compared to other methods. Table 2 demonstrates that the proposed method  P r o 3 outperforms other state-of-the-art methods considering  A c c u r a c y . Table 2 shows that  P r o 3 performs better than  P r o 6 , especially in experiments with occluded human. See Figure 4. The accuracy of HOG detector is considerably lower than other methods, since the HOG detector does not use CNN structures.

4.3. Video File 3 of a Person Moving in Mountains

Our human detection system was tested with a video file, which shows a person moving in mountains. One human image in the video file is plotted in Figure 5. Once a moving object is detected, then one generates the ROI of the moving object. In Figure 5, a green box indicates the moving human who is occluded by trees.
Table 3 shows the classification accuracy and computational load of the proposed human detection system, compared to other state-of-the-art methods. Table 3 verifies that the proposed method  P r o 3 is comparable to other state-of-the-art YOLO-based methods considering  A c c u r a c y . Table 3 shows that  P r o 3 performs better than  P r o 6 , especially in experiments with occluded human. See Figure 5. Table 3 further shows that the proposed method outperforms all other methods considering  F P S (computational speed). The accuracy of HOG detector is considerably lower than other methods, since the HOG detector does not use CNN structures.

4.4. Video File 4 of a Wild Animal in Mountains

For video surveillance in the off-limit area, IR cameras can be used to detect human activities. IR cameras are useful for detection of humans in dark and cluttered environments. In Video file 4, we consider IR cameras, and we test the case where there is a wild animal in the movie file. See Figure 6.
Considering the case where there is a wild animal, TP and FN are defined as follows. Among all bounding boxes, True Positive (TP) denotes the number of boxes in which a non-human is classified as a non-human. Among all bounding boxes, False Negative (FN) denotes the number of boxes in which a human is classified as a human. Since there is no human in Video file 4, FN = 0.
Table 4 shows the classification accuracy and computational load of the proposed human detection system, compared to other state-of-the-art methods. Table 4 shows that the proposed method  P r o 6 is comparable to  Y O L O v 7 ( 6 ) considering  A c c u r a c y . Table 4 verifies that the proposed method outperforms all other methods considering  F P S (processing speed). The HOG detector is not used in Table 4, since the HOG detector cannot be used to detect a non-human.

4.5. Video File 5 of No Person in Mountains

In Video file 5, we test the case where there is no human in the movie file. Moreover, there is no moving object in this Video file 5. For this scenario with no human, we use  H u m a n C o u n t as a comparison index. Here,  H u m a n C o u n t indicates the number of cases where a human is detected in the movie file. Considering the case where there is no human in the movie file, it is desirable that  H u m a n C o u n t is as close to 0 as possible.
We use a video file of no person in mountains, as plotted in Figure 7. Moreover, there is no moving object in this video file. Table 5 shows the classification accuracy and computational load of the proposed human detection system, compared to other methods. Since there is no human in the video file,  H u m a n C o u n t is zero in the video. Since motion detector runs much faster than an object detection–classification method (e.g., YOLO-based methods), the proposed scheme is suitable for real-time human detection with low computational loads.
Table 5 verifies that our human detector system is much more effective than other state-of-the-art algorithms, especially in the case where there are sparse or no humans. In off-limits mountains, a human rarely exists, hence human detection is extremely rare. Thus, the proposed scheme is preferred in off-limits mountains where there are sparse or no humans.

5. Conclusions

Our paper handles the case where the camera system in off-limits mountains detects humans. In off-limits mountains, a human rarely exists, thus human detection is an extremely rare event. We address a time-efficient human detector system, based on both motion detection and object classification.
In the camera image, we use a feasible human space where a human can appear. Our strategy is to enable the object classification, only inside a bounding box where motion is detected in the feasible human space. Since motion detection inside the feasible human space runs much faster than an object detection–classification method, the proposed approach is suitable for real-time human detection with low computational loads. As far as we know, no paper in the literature used the feasible human space, as in our paper.
Experiments showed that the proposed human detector system outperforms other state-of-the-art methods considering both FPS and Accuracy. This paper verifies that the accuracy of the proposed human detector system is comparable to other state-of-the-art algorithms, while outperforming in computational speed.
The proposed detector system can be extended to classify any moving object, such as animals. In the camera image, we can define a feasible object space where a moving object can appear. We enable the object classification, only inside a bounding box where motion is detected in the feasible object space. For object classification, we can train the CNN to classify any moving object inside a bounding box. Since motion detection inside the feasible object space runs much faster than an object detection–classification method (e.g., YOLO-based methods), the proposed approach is suitable for real-time object detection with low computational loads.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (Grant Number: 2022R1A2C1091682). This research was supported by the faculty research fund of Sejong university in 2023.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is available upon request.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Nguyen, D.T.; Li, W.; Ogunbona, P.O. Human detection from images and videos: A survey. Pattern Recognit. 2016, 51, 148–175. [Google Scholar] [CrossRef]
  2. Paul, M.; Haque, S.; Chakraborty, S. Human detection in surveillance videos and its applications—A review. EURASIP J. Adv. Signal Process. 2013, 2013, 176. [Google Scholar] [CrossRef]
  3. AlDahoul, N.; Sabri, A.Q.M.; Mansoor, A.M. Real-Time Human Detection for Aerial Captured Video Sequences via Deep Models. Comput. Intell. Neurosci. 2018, 2018, 1639561. [Google Scholar] [CrossRef] [PubMed]
  4. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  5. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  6. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  7. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  8. Srivastava, S.; Divekar, A.V.; Anilkumar, C.; Naik, I.; Kulkarni, V.; Pattabiraman, V. Comparative analysis of deep learning image detection algorithms. J. Big Data 2021, 8, 66. [Google Scholar] [CrossRef]
  9. Malhotra, P.; Garg, E. Object Detection Techniques: A Comparison. In Proceedings of the 2020 7th International Conference on Smart Structures and Systems (ICSSS), Chennai, India, 23–24 July 2020; pp. 1–4. [Google Scholar]
  10. Wang, C.; Bochkovskiy, A.; Liao, H. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 7464–7475. [Google Scholar]
  11. Rath, S.; Gupta, A. Performance Comparison of YOLO Object Detection Models—An Intensive Study. Website. 2024. Available online: https://learnopencv.com/performance-comparison-of-yolo-models/ (accessed on 15 January 2024).
  12. Piccardi, M. Background subtraction techniques: A review. In Proceedings of the 2004 IEEE International Conference on Systems, Man and Cybernetics (IEEE Cat. No.04CH37583), The Hague, The Netherlands, 10–13 October 2004; Volume 4, pp. 3099–3104. [Google Scholar]
  13. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  14. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  15. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for MobileNetV3. arXiv 2019, arXiv:1905.02244. [Google Scholar]
  16. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; Yang, Y. Random Erasing Data Augmentation. In Proceedings of the of AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 13001–13008. [Google Scholar]
  17. Candamo, J.; Shreve, M.; Goldgof, D.B.; Sapper, D.B.; Kasturi, R. Understanding Transit Scenes: A Survey on Human Behavior-Recognition Algorithms. IEEE Trans. Intell. Transp. Syst. 2010, 11, 206–224. [Google Scholar] [CrossRef]
  18. Ji, X.; Liu, H. Advances in View-Invariant Human Motion Analysis: A Review. IEEE Trans. Syst. Man, Cybern. Part C Appl. Rev. 2010, 40, 13–24. [Google Scholar]
  19. Ramadhan, D.I.; Sari, I.P.; Sari, L.O. Comparison of Background Subtraction, Sobel, Adaptive Motion Detection, Frame Differences, and Accumulative Differences Images on Motion Detection. SINERGI 2018, 22, 51–62. [Google Scholar] [CrossRef]
  20. Villa, C.E.; Caccia, M.; Sironi, L.; D’Alfonso, L.; Collini, M.; Rivolta, I.; Miserocchi, G.; Gorletta, T.; Zanoni, I.; Granucci, F.; et al. Accumulative Difference Image Protocol for Particle Tracking in Fluorescence Microscopy Tested in Mouse Lymphonodes. PLoS ONE 2010, 5, e12216. [Google Scholar] [CrossRef] [PubMed]
  21. openCV. Morphological Transformations. Website. 2021. Available online: https://opencv24-python-tutorials.readthedocs.io/en/latest/ (accessed on 15 January 2024).
  22. Gao, F.; Wang, C.; Li, C. A Combined Object Detection Method with Application to Pedestrian Detection. IEEE Access 2020, 8, 194457–194465. [Google Scholar] [CrossRef]
  23. Hu, Q.; Paisitkriangkrai, S.; Shen, C.; van den Hengel, A.; Porikli, F. Fast Detection of Multiple Objects in Traffic Scenes with a Common Detection Framework. IEEE Trans. Intell. Transp. Syst. 2016, 17, 1002–1014. [Google Scholar] [CrossRef]
  24. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  25. Cao, J.; Song, C.; Peng, S.; Song, S.; Zhang, X.; Shao, Y.; Xiao, F. Pedestrian Detection Algorithm for Intelligent Vehicles in Complex Scenarios. Sensors 2020, 20, 3646. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The central cell is connected to all 8 cells surrounding the cell. Here, a connected cell of the central cell is marked with an arrow.
Figure 1. The central cell is connected to all 8 cells surrounding the cell. Here, a connected cell of the central cell is marked with an arrow.
Sensors 24 00782 g001
Figure 2. Video file 1. A black rectangle indicates the feasible human space. Once a moving object is detected inside the feasible human space, then one generates the ROI of the moving object. A green rectangle indicates the ROI of the moving object (human) detected using the proposed method.
Figure 2. Video file 1. A black rectangle indicates the feasible human space. Once a moving object is detected inside the feasible human space, then one generates the ROI of the moving object. A green rectangle indicates the ROI of the moving object (human) detected using the proposed method.
Sensors 24 00782 g002
Figure 3. Video file 1. The GMBS and ConnectedComponentsWithStats are applied for detecting a moving object in Figure 2. White area indicates the silhouette of a moving object.
Figure 3. Video file 1. The GMBS and ConnectedComponentsWithStats are applied for detecting a moving object in Figure 2. White area indicates the silhouette of a moving object.
Sensors 24 00782 g003
Figure 4. Video file 2. Once a moving object is detected, then one generates the ROI of the moving object. A green box indicates the moving human occluded by trees.
Figure 4. Video file 2. Once a moving object is detected, then one generates the ROI of the moving object. A green box indicates the moving human occluded by trees.
Sensors 24 00782 g004
Figure 5. Video file 3. Once a moving object is detected, then one generates the ROI of the moving object. A green box indicates the moving human who is occluded by trees.
Figure 5. Video file 3. Once a moving object is detected, then one generates the ROI of the moving object. A green box indicates the moving human who is occluded by trees.
Sensors 24 00782 g005
Figure 6. Video file 4 of a wild animal in mountains.
Figure 6. Video file 4 of a wild animal in mountains.
Sensors 24 00782 g006
Figure 7. Video file 5 of no person in mountains.
Figure 7. Video file 5 of no person in mountains.
Sensors 24 00782 g007
Table 1. Video file 1 comparison ( A c c u r a c y and  F P S ).
Table 1. Video file 1 comparison ( A c c u r a c y and  F P S ).
Algorithm   Accuracy FPS (Total Human Frame Number H)   FPS c
  P r o 3 0.9914 (629)17
  P r o 6 0.9916 (629)17
YOLOv7 (3)0.943 (629)-
YOLOv7-tiny (3)0.999 (629)-
YOLOv7 (6)0.951 (629)-
YOLOv7-tiny (6)0.997 (629)-
HOG0.3119 (629)-
Table 2. Video file 2 comparison ( A c c u r a c y and  F P S ).
Table 2. Video file 2 comparison ( A c c u r a c y and  F P S ).
Algorithm   Accuracy FPS (Total Human Frame Number H)   FPS c
  P r o 3 0.9318 (775)18
  P r o 6 0.8716 (775)18
YOLOv7 (3)0.731 (775)-
YOLOv7-tiny (3)0.576 (775)-
YOLOv7 (6)0.741 (775)-
YOLOv7-tiny (6)0.89 (775)-
HOG023 (775)-
Table 3. Video file 3 comparison ( A c c u r a c y and  F P S ).
Table 3. Video file 3 comparison ( A c c u r a c y and  F P S ).
Algorithm   Accuracy FPS (Total Human Frame Number H)   FPS c
  P r o 3 0.9833 (339)19
  P r o 6 0.9338 (339)19
YOLOv7 (3)0.961 (339)-
YOLOv7-tiny (3)0.945 (339)-
YOLOv7 (6)0.981 (339)-
YOLOv7-tiny (6)0.9810 (339)-
HOG028 (339)-
Table 4. Video file 4 comparison ( A c c u r a c y and  F P S ).
Table 4. Video file 4 comparison ( A c c u r a c y and  F P S ).
Algorithm   Accuracy FPS (Total Frame Number)   FPS c
  P r o 3 0.8711 (308)19
  P r o 6 0.9213 (308)19
YOLOv7 (3)0.643 (308)-
YOLOv7-tiny (3)0.311 (308)-
YOLOv7 (6)0.973 (308)-
YOLOv7-tiny (6)0.678 (308)-
HOG---
Table 5. Video file 5 comparison ( H u m a n C o u n t and  F P S ).
Table 5. Video file 5 comparison ( H u m a n C o u n t and  F P S ).
Algorithm   HumanCount FPS (Total Human Frame Number H)   FPS c
  P r o 3 062 (0)15
  P r o 6 062 (0)16
YOLOv7 (3)01 (0)-
YOLOv7-tiny (3)06 (0)-
YOLOv7 (6)01 (0)-
YOLOv7-tiny (6)07 (0)-
HOG030 (0)-
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, J. Autonomous Detection of Humans in Off-Limits Mountain Areas. Sensors 2024, 24, 782. https://doi.org/10.3390/s24030782

AMA Style

Kim J. Autonomous Detection of Humans in Off-Limits Mountain Areas. Sensors. 2024; 24(3):782. https://doi.org/10.3390/s24030782

Chicago/Turabian Style

Kim, Jonghoek. 2024. "Autonomous Detection of Humans in Off-Limits Mountain Areas" Sensors 24, no. 3: 782. https://doi.org/10.3390/s24030782

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop