Skip to main content
Log in

Rep-YOLO: an efficient detection method for mine personnel

  • Research
  • Published:
Journal of Real-Time Image Processing Aims and scope Submit manuscript

Abstract

The detection of underground personnel is one of the key technologies in computer vision. However, this detection technique is susceptible to complex environments, resulting in low accuracy and slow speed. To accurately detect underground coal mine operators in complex environments, we combine the underground image features with K-means++ clustering anchor frames and propose a new Re-parameterization YOLO (Rep-YOLO) detection algorithm. First, the Criss-Cross-Vertical with Channel Attention (CVCA) mechanism is introduced at the end of the network to capture the Long-Range Dependencies (LRDs) in the image. This mechanism also emphasizes the significance of different channels to enhance image processing performance and improve the representation ability of the model. Second, the new Deep Extraction of Re-parameterization (DER) backbone network is designed, which adopts the re-parameterization structure to reduce the number of parameters and computation of the model. Additionally, each DER-block fuses different scales of features to enhance the accuracy of the model’s detection capabilities. Finally, Rep-YOLO is optimized using a slim-neck structure, which reduces the complexity of the Rep-YOLO while maintaining sufficient accuracy. The results showed that the Rep-YOLO model proposed in this paper achieved an accuracy of \(87.5\%\), a recall rate of \(77.2\%\), an Average Precision (AP) of \(83.1\%\), and a Frame Per Second (FPS) of 71.9. Compared to eight different models, the recall, AP50, and FPS of the Rep-YOLO model were improved. The research shows that the Rep-YOLO model can provide a real-time and efficient method for mine personnel detection. Source code is released in https://github.com/DrLSB/Rep-YOLO.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14

Similar content being viewed by others

Data Availability

The datasets generated during and/or analyzed during the current study are not publicly available due [their containing information that could compromise the privacy of research participants] but are available from the corresponding author on reasonable request.

References

  1. Wang, Q., Ge, S.: Uncovering the effects of external demand on China’s coal consumption: A global input-output analysis. J. Clean. Prod. 245, 118877 (2020)

    Article  Google Scholar 

  2. Qi, Rui, Liu, Tongyi, Jia, Qingxuan, Sun, Li., Liu, Jiangyi: Simulating the sustainable effect of green mining construction policies on coal mining industry of China 226, 392–406 (2019)

  3. Ge, Xiaosan, Shuai, Su, Haiyang, Yu, Chen, Gang, Xiaoping, Lu: Smart mine construction based on knowledge engineering and internet of things 14, 1060 (2018)

  4. Zhu, Qiang, Yeh, Mei-Chen, Cheng, Kwang-Ting, Avidan, Shai: Fast human detection using a cascade of histograms of oriented gradients. In: 2006 IEEE computer society conference on computer vision and pattern recognition (CVPR), vol. 2, pp. 1491–1498 (2006)

  5. Suykens, J.A.K., Vandewalle, J.: Least squares support vector machine classifiers. Neural Process. Lett. 9, 293–300 (1999)

    Article  Google Scholar 

  6. Girshick, Ross, Donahue, Jeff, Darrell, Trevor, Malik, Jitendra: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 580–587 (2014)

  7. Ren, Shaoqing, He, Kaiming, Girshick, Ross, Sun, Jian: Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems 28 (2015)

  8. Cai, Zhaowei, Vasconcelos, Nuno: Cascade r-cnn: Delving into high quality object detection. IN: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 6154–6162 (2018)

  9. Liu, Wei, Anguelov, Dragomir, Erhan, Dumitru, Szegedy, Christian, Reed, Scott, Fu, Cheng-Yang, Berg, Alexander C.: Ssd: Single shot multibox detector. In: European conference on computer vision, pp. 21–37 (2016)

  10. Redmon, Joseph, Divvala, Santosh, Girshick, Ross, Farhadi, Ali: You only look once: Unified, real-time object detection. IN: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 779–788 (2016)

  11. Redmon, Joseph, Farhadi, Ali: YOLO9000: better, faster, stronger. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7263–7271 (2017)

  12. Redmon, Joseph, Farhadi, Ali: Yolov3: An incremental improvement (2018). arXiv preprint arXiv:1804.02767

  13. Bochkovskiy, Alexey, Wang, Chien-Yao, Liao, Hong-Yuan Mark: Yolov4: Optimal speed and accuracy of object detection (2020). arXiv preprint arXiv:2004.10934

  14. Xu, Shangliang, Wang, Xinxin, Lv, Wenyu, Chang, Qinyao, Cui, Cheng, Deng, Kaipeng, Wang, Guanzhong, Dang, Qingqing, Wei, Shengyu, Du, Yuning, others: PP-YOLOE: An evolved version of YOLO (2022). arXiv preprint arXiv:2203.16250

  15. Wang, Chien-Yao, Yeh, I-Hau, Liao, Hong-Yuan Mark: You only learn one representation: Unified network for multiple tasks (2021). arXiv preprint arXiv:2105.04206

  16. Rauf, Rabia, Shahid, Ahmad R., Ziauddin, Sheikh, Safi, Asad Ali: Pedestrian detection using HOG, LUV and optical flow as features with AdaBoost as classifier. In: 2016 Sixth International Conference on Image Processing Theory, Tools and Applications (IPTA), pp. 1–4 (2016)

  17. Hong, G.-S., Kim, B.-G., Hwang, Y.-S., Kwon, K.-K.: Fast multi-feature pedestrian detection algorithm based on histogram of oriented gradient using discrete wavelet transform. Multimedia Tools and Applications 75, 15229–15245 (2016)

    Article  Google Scholar 

  18. Zhi, N., Mao, S.J., Li, M.: Enhancement algorithm based on illumination adjustment for non-uniform illuminance video images in coal mine. J. China Coal Soc. 42, 2190–2197 (2017)

    Google Scholar 

  19. Zhu, L.Y.: Research on mine image enhancement and underground personnel detection, China University of Mining and Technology (2019)

  20. Li, X., Wang, S., Liu, B., Chen, W., Fan, W., Tian, Z.: Improved YOLOv4 network using infrared images for personnel detection in coal mines. J. Electron. Imaging 31, 013017–013017 (2022)

    Google Scholar 

  21. Wei, X., Zhang, H., Liu, S., Lu, Y.: Pedestrian detection in underground mines via parallel feature transfer network. Pattern Recogn. 103, 107195 (2020)

    Article  Google Scholar 

  22. Kou, F., Xiao, W., He, H., Chen, R.: Research on Target Detection in Underground Coal Mines Based on Improved YOLOv5. Journal of Electronics & Information Technology 45, 1–8 (2022)

    Google Scholar 

  23. Zhang, M.Z.: Underground pedestrian detection model based on Dense-YOLO network. Journal of Mine Automation 48 (2022)

  24. Knausgård, Kristian Muri, Wiklund, Arne, Sørdalen, Tonje Knutsen, Halvorsen, Kim Tallaksen, Kleiven, Alf Ring, Jiao, Lei, Goodwin, Morten: Temperate fish detection and classification: a deep learning based approach. In: Applied Intelligence, pp. 1–14 (2022)

  25. Sun, B., Wang, X., Li, H., Dong, F., Wang, Y.: Small-target ship detection in SAR images based on densely connected deep neural network with attention in complex scenes. Appl. Intell. 53, 4162–4179 (2023)

    Article  Google Scholar 

  26. Hu, Jie, Shen, Li, Sun, Gang: Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7132–7141 (2018)

  27. Woo, Sanghyun, Park, Jongchan, Lee, Joon-Young, Kweon: In So, Cbam: Convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV), pp. 3–19 (2018)

  28. Hou, Qibin, Zhou, Daquan, Feng, Jiashi: Coordinate attention for efficient mobile network design. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 13713–13722 (2021)

  29. Wang, Qilong, Wu, Banggu, Zhu, Pengfei, Li, Peihua, Zuo, Wangmeng, Hu, Qinghua: ECA-Net: Efficient channel attention for deep convolutional neural networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 11534–11542 (2020)

  30. Sheng, W., Yu, X., Lin, J., Chen, X.: Faster RCNN Target Detection Algorithm Integrating CBAM and FPN. Appl. Sci. 13, 6913 (2023)

    Article  Google Scholar 

  31. Tang, Jun, Gong, Sihang, Wang, Yanjiang, Liu, Baodi, Du, Chunyu, Gu, Boyang: Beyond coordinate attention: spatial-temporal recalibration and channel scaling for skeleton-based action recognition. In: Signal, Image and Video Processing, pp. 1–8 (2023)

  32. Niu, JieYi, Xie, ZhiHua, Li, Yi, Cheng, SiJia, Fan, JiaWei: Scale fusion light CNN for hyperspectral face recognition with knowledge distillation and attention mechanism. In: Signal, Applied Intelligence, pp. 1–15 (2022)

  33. Huang, Zilong, Wang, Xinggang, Huang, Lichao, Huang, Chang, Wei, Yunchao, Liu, Wenyu: Ccnet: Criss-cross attention for semantic segmentation. In: Proceedings of the IEEE/CVF international conference on computer vision, pp. 603–612 (2019)

  34. Howard, Andrew G., Zhu, Menglong, Chen, Bo, Kalenichenko, Dmitry, Wang, Weijun, Weyand, Tobias, Andreetto, Marco, Adam, Hartwig: Mobilenets: Efficient convolutional neural networks for mobile vision applications (2017). arXiv preprint arXiv:1704.04861

  35. Ioffe, Sergey, Szegedy, Christian: Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning, pp. 448–456 (2015)

  36. Sandler, Mark, Howard, Andrew, Zhu, Menglong, Zhmoginov, Andrey, Chen, Liang-Chieh: Mobilenetv2: Inverted residuals and linear bottlenecks, Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4510–4520 (2018)

  37. Ding, Xiaohan, Zhang, Xiangyu, Ma, Ningning, Han, Jungong, Ding, Guiguang, Sun, Jian: Repvgg: Making vgg-style convents great again. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 13733–13742 (2021)

  38. Vasu, Pavan Kumar Anasosalu, Gabriel, James, Zhu, Jeff, Tuzel, Oncel, Ranjan, Anurag: MobileOne: An Improved One Millisecond Mobile Backbone. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7907–7917 (2023)

  39. Elfwing, S., Uchibe, E., Doya, K.: Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 107, 3–11 (2018)

    Article  Google Scholar 

  40. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research 15, 1929–1958 (2014)

    MathSciNet  Google Scholar 

  41. Li, Hulin, Li, Jun, Wei, Hanbing, Liu, Zheng, Zhan, Zhenfei, Ren, Qiliang: Slim-neck by GSConv: A better design paradigm of detector architectures for autonomous vehicles (2022). arXiv preprint arXiv:2206.02424

  42. Lee, Youngwan, Hwang, Joong-won, Lee, Sangrok, Bae, Yuseok, Park, Jongyoul: An energy and GPU-computation efficient backbone network for real-time object detection, Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, pp. 752–760 (2019)

  43. Wang, Chien-Yao, Liao, Hong-Yuan Mark, Wu, Yueh-Hua, Chen, Ping-Yang, Hsieh, Jun-Wei, Yeh, I-Hau: CSPNet: A new backbone that can enhance learning capability of CNN. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, pp. 390–391 (2020)

  44. Zheng, Z., Wang, P., Liu, W., Li, J., Ye, R., Ren, D.: Distance-IoU loss: Faster and better learning for bounding box regression. Proceedings of the AAAI conference on artificial intelligence 34, 12993–13000 (2020)

    Article  Google Scholar 

  45. Tan, Mingxing, Pang, Ruoming, Le, Quoc V.: Efficientdet: Scalable and efficient object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10781–10790 (2020)

  46. Glenn, J.: yolov5. Git code (2020). https://github.com/ultralytics/yolov5

  47. Ge, Zheng, Liu, Songtao, Wang, Feng, Li, Zeming, Sun, Jian: Yolox: Exceeding yolo series in 2021 (2021). arXiv preprint arXiv:2107.08430

  48. Li, Chuyi, Li, Lulu, Jiang, Hongliang, Weng, Kaiheng, Geng, Yifei, Li, Liang, Ke, Zaidan, Li, Qingyuan, Cheng, Meng, Nie, Weiqiang, others: YOLOv6: A single-stage object detection framework for industrial applications (2022). arXiv preprint arXiv:2209.02976

  49. Glenn, J.: yolov8. Git code (2023). https://github.com/ultralytics/ultralytics/tree/main/ultralytics/models/v8

  50. Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. IEEE computer society conference on computer vision and pattern recognition 1, 886–893 (2005)

    Google Scholar 

Download references

Acknowledgements

This work was supported by Grants from the National Natural Science Foundation of China (52174198).

Author information

Authors and Affiliations

Authors

Contributions

XS contributed to conceptualization, methodology, and writing—original draft preparation. SL was involved in conceptualization, methodology, writing—original draft preparation, and writing—reviewing and editing. XL contributed to conceptualization and data curation. ZL contributed to conceptualization and writing-reviewing. HL contributed to conceptualization and methodology.

Corresponding author

Correspondence to Shibo Liu.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Ethics approval

Written informed consent for publication of this paper was obtained from the Xi’an University of Science and Technology, Xi’an Key Laboratory of Electrical Equipment Condition Monitoring and Power Supply Security, and all authors.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Shao, X., Liu, S., Li, X. et al. Rep-YOLO: an efficient detection method for mine personnel. J Real-Time Image Proc 21, 28 (2024). https://doi.org/10.1007/s11554-023-01407-3

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11554-023-01407-3

Keywords

Navigation