Abstract
Investigations into the behaviour of pedestrians and autonomous driving both frequently employ object detection. It has always been a popular area for research in computer vision and artificial intelligence. Because of the advancement of deep learning, object detectors are improving in accuracy and speed. However, the majority of them struggle to reconcile speed and precision. Because of this, the object detection model employed in this work, which is based on a better You Only Look Once Version 4 (YOLOv4) algorithm, considers both detection accuracy and efficiency. Several different sensor types, including cameras and mmWave (millimetre Wave) radar, are employed to observe the environment. This study employs a YOLOv4-based hybrid feature enhancement network for robust object detection under adverse weather conditions (AWC), utilizing the CRUW dataset. A camera–radar-fused approach that detects things that are cross-supervised by a deep radar object detection network, such as the YOLOv4 algorithm, reduces processing complexity. To correct category imbalance and further boost the robustness of the suggested model, the class-balanced sampling strategy (CBSS) is employed. Python software was employed to analyse the suggested method. In terms of mAP and accuracy, the suggested method is contrasted with the current OD ensemble framework and faster R-CNN methodologies. The suggested method’s accuracy and mAP are higher than those of the other existing methods; as a result, the proposed method’s mAP produces 40% and its accuracy produces 98.7%, respectively. Additionally, the suggested method performs better when classifying the object and detecting it, but it also strengthens the system’s stability and reduces category imbalance.
Similar content being viewed by others
Availability of data and materials
Data sharing is not relevant to this article since there were no datasets generated or analysed in the course of the present study.
References
Sharma, T., Debaque, B., Duclos, N., Chehri, A., Kinder, B., Fortier, P.: Deep learning-based object detection and scene perception under bad weather conditions. Electronics 11(4), 563 (2022)
Li, Y.J., Park, J., O'Toole, M., Kitani, K.: Modality-agnostic learning for radar-lidar fusion in vehicle detection. In proceedings of the IEEE/CVF conference on computer vision and pattern recognition pp. 918–927 (2022)
Hahner, M., Sakaridis, C., Bijelic, M., Heide, F., Yu, F., Dai, D., Van Gool, L.: Lidar snowfall simulation for robust 3d object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp 16364–16374) (2022)
Vs, V., Poster, D., You, S., Hu, S. Patel, V.M.: Meta-UDA: Unsupervised domain adaptive thermal object detection using meta-learning. In proceedings of the IEEE/CVF winter conference on applications of computer vision (pp 1412–1423) (2022)
Shuai, X., Shen, Y., Tang, Y., Shi, S., Ji, L., Xing, G., millieye: a lightweight mmwave radar and camera fusion system for robust object detection. In proceedings of the international conference on internet-of-things design and implementation (pp. 145–157) (2021)
Kim, Y., Hwang, H., Shin, J.: Robust object detection under harsh autonomous-driving environments. IET Image Proc. 16(4), 958–971 (2022)
Osipov, A., Pleshakova, E., Gataullin, S., Korchagin, S., Ivanov, M., Finogeev, A., Yadav, V.: Deep learning method for recognition and classification of images from video recorders in difficult weather conditions. Sustainability 14(4), 2420 (2022)
Cygert, S. and Czyżewski, A.: Robust object detection with multi-input multi-output faster r-cnn. In international conference on image analysis and processing (pp. 572–583) Springer, Cham. (2022)
Hahner, M., Sakaridis, C., Dai, D. and Van Gool, L.: Fog simulation on real LiDAR point clouds for 3D object detection in adverse weather. In proceedings of the ieee/cvf international conference on computer vision (pp. 15283–15292) (2021)
Boppana, U.M., Mustapha, A., Jacob, K., Deivanayagampillai, N.: Comparative analysis of single-stage yolo algorithms for vehicle detection under extreme weather conditions. In: IOT with smart systems, pp. 637–645. Springer, Singapore (2022)
Zimmer, W., Grabler, M. and Knoll, A.: (2022) Real-time and robust 3D object detection within road-side LiDARs using domain adaptation. arXiv preprint arXiv:2204.00132
Wang, Y., Jiang, Z., Li, Y., Hwang, J.N., Xing, G., Liu, H.: RODNet: a real-time radar object detection network cross-supervised by camera-radar fused object 3D localization. IEEE J. Sel. Top. Signal Process. 15(4), 954–967 (2021)
Han, J., Liang, X., Xu, H., Chen, K., Hong, L., Ye, C., Zhang, W., Li, Z., Liang, X., Xu, C.: Soda10m: towards large-scale object detection benchmark for autonomous driving. arXiv e-prints, pp. arXiv-2106 (2021)
Wang, W., Zhang, J., Zhai, W., Cao, Y., Tao, D.: Robust object detection via adversarial novel style exploration. IEEE Trans. Image Process. 31, 1949–1962 (2022)
Liu, Z., Cai, Y., Wang, H., Chen, L., Gao, H., Jia, Y., Li, Y.: Robust target recognition and tracking of self-driving cars with radar and camera information fusion under severe weather conditions. IEEE Trans. Intell. Trans. Sys. 23(7), 6640–6653 (2021)
Qian, K., Zhu, S., Zhang, X. and Li, L.E.: Robust multimodal vehicle detection in foggy weather using complementary lidar and radar signals. In proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 444–453) (2021)
Han, J., Liang, X., Xu, H., Chen, K., Hong, L., Mao, J., Ye, C., Zhang, W., Li, Z., Liang, X., Xu, C.: SODA10M: a Large-scale 2D Self/Semi-supervised object detection dataset for autonomous driving. arXiv preprint arXiv:2106.11118 (2021)
Wang, Y., Jiang, Z., Gao, X., Hwang, J.N., Xing, G. and Liu, H.: Rodnet: radar object detection using cross-modal supervision. In proceedings of the IEEE/CVF winter conference on applications of computer vision (pp. 504–513) (2021)
Nguyen, V., Tran, D., Tran, M., Nguyen, N., Nguyen, V.: Robust vehicle detection under adverse weather conditions using auto-encoder feature. Int. J. Mach. Learn. Comput 10(4), 549–555 (2020)
Krišto, M., Ivasic-Kos, M., Pobar, M.: Thermal object detection in difficult weather conditions using YOLO. IEEE access 8, 125459–125476 (2020)
Rothmeier, T., Huber, W.: Performance evaluation of object detection algorithms under adverse weather conditions. In international conference on intelligent transport systems (pp. 211–222). Springer, Cham. (2020)
Bijelic, M., Gruber, T., Mannan, F., Kraus, F., Ritter, W., Dietmayer, K., Heide, F.: Seeing through fog without seeing fog: deep multimodal sensor fusion in unseen adverse weather. In proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 11682–11692) (2020)
Yusuf, M.M., Karim, T., Saif, A.S.: A robust method for lane detection under adverse weather and illumination conditions using convolutional neural network. In proceedings of the international conference on computing advancements (pp. 1–8). (2020)
Sheeny, M., Wallace, A., Wang, S.: 300 GHz radar object recognition based on deep neural networks and transfer learning. IET Radar Sonar Navig. 14(10), 1483–1493 (2020)
Sindagi, V.A., Oza, P., Yasarla, R., Patel, V.M.: Prior-based domain adaptive object detection for hazy and rainy conditions. In european conference on computer vision (pp. 763–780). Springer, Cham, (2020)
Funding
Not applicable.
Author information
Authors and Affiliations
Contributions
All authors contributed to the design and implementation of the research, to the analysis of the results and the script of the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Ethics approval
Not applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
Variables | Description |
---|---|
\(\left( {\rho_{c} , \theta_{c} } \right)\) | Projected location in range-azimuth coordinates |
\(\left( {x^{c} - x_{{{\text{or}}}} } \right)\) | Place of the object in camera BEV coordinates |
\(\left( {x_{{{\text{or}}}} ,z_{{{\text{or}}}} } \right)\) | Place of the radar origin in camera BEV coordinates |
\({\text{cls}}\) | Object class |
\(d_{i}\) | Depth of the object |
\(c_{i}\) | Depth of confidence |
\(\delta \left( {{\text{cls}}} \right)\) | Typical azimuth error for camera localization |
\(N\left( \cdot \right)\) | Object’s probability map |
\(\epsilon \left( \cdot \right)\) | Azimuth resolution of the radar |
\(\delta_{j}^{r}\) | Range resolution |
\(P_{{\left( {{\text{cls}}} \right)}}^{{{\text{CRF}}}}\) | Fused probability map |
\(P_{{{\text{local}}}}\) | Point-wise features |
\(P\) | Point cloud |
\(V\) | Voxel features |
\(D\) | Dimension |
\(R^{3}\) | Voxel grids number |
\(B\) | Fixed number |
\(h_{\theta }\) | MLP’s shared learnable parameters |
\(W\) | Width |
\(H\) | Height |
\(C\) | Channel count |
\(F\) | Feature map |
\(\overline{F}^{w,h}\) | Local surrounding region |
\(\overline{W}\) | Width of local region |
\(\overline{H}\) | Height of local region |
\(\left( {w, h} \right)\) | Weights associated with each pixel |
\(P_{{{\text{local}}}}^{w,h}\) | Attentive contextual feature |
\(\overline{f}_{i}^{w,h}\) | Feature at the \(i^{th}\) location |
\(\hat{\phi }_{j}\) | Balanced softmax output |
\(\hat{\phi }_{y}\) | Conditional probability of the correct class |
\(x^{ + }\) | Positive sample |
\(x^{ - }\) | Negative sample |
\(j\) | Class |
\(D_{{{\text{train}}}}\) | Training set |
\(1/n_{j}\) | Inverse loss weight |
\(1/n_{j}^{2}\) | Over-balance problem |
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Patil, S.M., Pawar, S.D., Mhatre, S.N. et al. Yolov4-based hybrid feature enhancement network with robust object detection under adverse weather conditions. SIViP (2024). https://doi.org/10.1007/s11760-024-03068-6
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11760-024-03068-6