Skip to main content
Log in

Yolov4-based hybrid feature enhancement network with robust object detection under adverse weather conditions

  • Original Paper
  • Published:
Signal, Image and Video Processing Aims and scope Submit manuscript

Abstract

Investigations into the behaviour of pedestrians and autonomous driving both frequently employ object detection. It has always been a popular area for research in computer vision and artificial intelligence. Because of the advancement of deep learning, object detectors are improving in accuracy and speed. However, the majority of them struggle to reconcile speed and precision. Because of this, the object detection model employed in this work, which is based on a better You Only Look Once Version 4 (YOLOv4) algorithm, considers both detection accuracy and efficiency. Several different sensor types, including cameras and mmWave (millimetre Wave) radar, are employed to observe the environment. This study employs a YOLOv4-based hybrid feature enhancement network for robust object detection under adverse weather conditions (AWC), utilizing the CRUW dataset. A camera–radar-fused approach that detects things that are cross-supervised by a deep radar object detection network, such as the YOLOv4 algorithm, reduces processing complexity. To correct category imbalance and further boost the robustness of the suggested model, the class-balanced sampling strategy (CBSS) is employed. Python software was employed to analyse the suggested method. In terms of mAP and accuracy, the suggested method is contrasted with the current OD ensemble framework and faster R-CNN methodologies. The suggested method’s accuracy and mAP are higher than those of the other existing methods; as a result, the proposed method’s mAP produces 40% and its accuracy produces 98.7%, respectively. Additionally, the suggested method performs better when classifying the object and detecting it, but it also strengthens the system’s stability and reduces category imbalance.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13

Similar content being viewed by others

Availability of data and materials

Data sharing is not relevant to this article since there were no datasets generated or analysed in the course of the present study.

References

  1. Sharma, T., Debaque, B., Duclos, N., Chehri, A., Kinder, B., Fortier, P.: Deep learning-based object detection and scene perception under bad weather conditions. Electronics 11(4), 563 (2022)

    Article  Google Scholar 

  2. Li, Y.J., Park, J., O'Toole, M., Kitani, K.: Modality-agnostic learning for radar-lidar fusion in vehicle detection. In proceedings of the IEEE/CVF conference on computer vision and pattern recognition pp. 918–927 (2022)

  3. Hahner, M., Sakaridis, C., Bijelic, M., Heide, F., Yu, F., Dai, D., Van Gool, L.: Lidar snowfall simulation for robust 3d object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp 16364–16374) (2022)

  4. Vs, V., Poster, D., You, S., Hu, S. Patel, V.M.: Meta-UDA: Unsupervised domain adaptive thermal object detection using meta-learning. In proceedings of the IEEE/CVF winter conference on applications of computer vision (pp 1412–1423) (2022)

  5. Shuai, X., Shen, Y., Tang, Y., Shi, S., Ji, L., Xing, G., millieye: a lightweight mmwave radar and camera fusion system for robust object detection. In proceedings of the international conference on internet-of-things design and implementation (pp. 145–157) (2021)

  6. Kim, Y., Hwang, H., Shin, J.: Robust object detection under harsh autonomous-driving environments. IET Image Proc. 16(4), 958–971 (2022)

    Article  Google Scholar 

  7. Osipov, A., Pleshakova, E., Gataullin, S., Korchagin, S., Ivanov, M., Finogeev, A., Yadav, V.: Deep learning method for recognition and classification of images from video recorders in difficult weather conditions. Sustainability 14(4), 2420 (2022)

    Article  Google Scholar 

  8. Cygert, S. and Czyżewski, A.: Robust object detection with multi-input multi-output faster r-cnn. In international conference on image analysis and processing (pp. 572–583) Springer, Cham. (2022)

  9. Hahner, M., Sakaridis, C., Dai, D. and Van Gool, L.: Fog simulation on real LiDAR point clouds for 3D object detection in adverse weather. In proceedings of the ieee/cvf international conference on computer vision (pp. 15283–15292) (2021)

  10. Boppana, U.M., Mustapha, A., Jacob, K., Deivanayagampillai, N.: Comparative analysis of single-stage yolo algorithms for vehicle detection under extreme weather conditions. In: IOT with smart systems, pp. 637–645. Springer, Singapore (2022)

    Chapter  Google Scholar 

  11. Zimmer, W., Grabler, M. and Knoll, A.: (2022) Real-time and robust 3D object detection within road-side LiDARs using domain adaptation. arXiv preprint arXiv:2204.00132

  12. Wang, Y., Jiang, Z., Li, Y., Hwang, J.N., Xing, G., Liu, H.: RODNet: a real-time radar object detection network cross-supervised by camera-radar fused object 3D localization. IEEE J. Sel. Top. Signal Process. 15(4), 954–967 (2021)

    Article  ADS  Google Scholar 

  13. Han, J., Liang, X., Xu, H., Chen, K., Hong, L., Ye, C., Zhang, W., Li, Z., Liang, X., Xu, C.: Soda10m: towards large-scale object detection benchmark for autonomous driving. arXiv e-prints, pp. arXiv-2106 (2021)

  14. Wang, W., Zhang, J., Zhai, W., Cao, Y., Tao, D.: Robust object detection via adversarial novel style exploration. IEEE Trans. Image Process. 31, 1949–1962 (2022)

    Article  ADS  CAS  PubMed  Google Scholar 

  15. Liu, Z., Cai, Y., Wang, H., Chen, L., Gao, H., Jia, Y., Li, Y.: Robust target recognition and tracking of self-driving cars with radar and camera information fusion under severe weather conditions. IEEE Trans. Intell. Trans. Sys. 23(7), 6640–6653 (2021)

    Article  Google Scholar 

  16. Qian, K., Zhu, S., Zhang, X. and Li, L.E.: Robust multimodal vehicle detection in foggy weather using complementary lidar and radar signals. In proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 444–453) (2021)

  17. Han, J., Liang, X., Xu, H., Chen, K., Hong, L., Mao, J., Ye, C., Zhang, W., Li, Z., Liang, X., Xu, C.: SODA10M: a Large-scale 2D Self/Semi-supervised object detection dataset for autonomous driving. arXiv preprint arXiv:2106.11118 (2021)

  18. Wang, Y., Jiang, Z., Gao, X., Hwang, J.N., Xing, G. and Liu, H.: Rodnet: radar object detection using cross-modal supervision. In proceedings of the IEEE/CVF winter conference on applications of computer vision (pp. 504–513) (2021)

  19. Nguyen, V., Tran, D., Tran, M., Nguyen, N., Nguyen, V.: Robust vehicle detection under adverse weather conditions using auto-encoder feature. Int. J. Mach. Learn. Comput 10(4), 549–555 (2020)

    Article  Google Scholar 

  20. Krišto, M., Ivasic-Kos, M., Pobar, M.: Thermal object detection in difficult weather conditions using YOLO. IEEE access 8, 125459–125476 (2020)

    Article  Google Scholar 

  21. Rothmeier, T., Huber, W.: Performance evaluation of object detection algorithms under adverse weather conditions. In international conference on intelligent transport systems (pp. 211–222). Springer, Cham. (2020)

  22. Bijelic, M., Gruber, T., Mannan, F., Kraus, F., Ritter, W., Dietmayer, K., Heide, F.: Seeing through fog without seeing fog: deep multimodal sensor fusion in unseen adverse weather. In proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 11682–11692) (2020)

  23. Yusuf, M.M., Karim, T., Saif, A.S.: A robust method for lane detection under adverse weather and illumination conditions using convolutional neural network. In proceedings of the international conference on computing advancements (pp. 1–8). (2020)

  24. Sheeny, M., Wallace, A., Wang, S.: 300 GHz radar object recognition based on deep neural networks and transfer learning. IET Radar Sonar Navig. 14(10), 1483–1493 (2020)

    Article  Google Scholar 

  25. Sindagi, V.A., Oza, P., Yasarla, R., Patel, V.M.: Prior-based domain adaptive object detection for hazy and rainy conditions. In european conference on computer vision (pp. 763–780). Springer, Cham, (2020)

Download references

Funding

Not applicable.

Author information

Authors and Affiliations

Authors

Contributions

All authors contributed to the design and implementation of the research, to the analysis of the results and the script of the manuscript.

Corresponding author

Correspondence to Shankar M. Patil.

Ethics declarations

Competing interests

The authors declare no competing interests.

Ethics approval

Not applicable.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

Variables

Description

\(\left( {\rho_{c} , \theta_{c} } \right)\)

Projected location in range-azimuth coordinates

\(\left( {x^{c} - x_{{{\text{or}}}} } \right)\)

Place of the object in camera BEV coordinates

\(\left( {x_{{{\text{or}}}} ,z_{{{\text{or}}}} } \right)\)

Place of the radar origin in camera BEV coordinates

\({\text{cls}}\)

Object class

\(d_{i}\)

Depth of the object

\(c_{i}\)

Depth of confidence

\(\delta \left( {{\text{cls}}} \right)\)

Typical azimuth error for camera localization

\(N\left( \cdot \right)\)

Object’s probability map

\(\epsilon \left( \cdot \right)\)

Azimuth resolution of the radar

\(\delta_{j}^{r}\)

Range resolution

\(P_{{\left( {{\text{cls}}} \right)}}^{{{\text{CRF}}}}\)

Fused probability map

\(P_{{{\text{local}}}}\)

Point-wise features

\(P\)

Point cloud

\(V\)

Voxel features

\(D\)

Dimension

\(R^{3}\)

Voxel grids number

\(B\)

Fixed number

\(h_{\theta }\)

MLP’s shared learnable parameters

\(W\)

Width

\(H\)

Height

\(C\)

Channel count

\(F\)

Feature map

\(\overline{F}^{w,h}\)

Local surrounding region

\(\overline{W}\)

Width of local region

\(\overline{H}\)

Height of local region

\(\left( {w, h} \right)\)

Weights associated with each pixel

\(P_{{{\text{local}}}}^{w,h}\)

Attentive contextual feature

\(\overline{f}_{i}^{w,h}\)

Feature at the \(i^{th}\) location

\(\hat{\phi }_{j}\)

Balanced softmax output

\(\hat{\phi }_{y}\)

Conditional probability of the correct class

\(x^{ + }\)

Positive sample

\(x^{ - }\)

Negative sample

\(j\)

Class

\(D_{{{\text{train}}}}\)

Training set

\(1/n_{j}\)

Inverse loss weight

\(1/n_{j}^{2}\)

Over-balance problem

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Patil, S.M., Pawar, S.D., Mhatre, S.N. et al. Yolov4-based hybrid feature enhancement network with robust object detection under adverse weather conditions. SIViP (2024). https://doi.org/10.1007/s11760-024-03068-6

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11760-024-03068-6

Keywords

Navigation