Multi-Input All-Weather Streetlight to Reduce Carbon Footprint of Illuminating Road Infrastructure
DOI:
https://doi.org/10.47611/jsrhs.v13i4.8337Keywords:
computer vision, vehicle detection, convolutional neural network, machine learning, infrared sensorAbstract
Street lighting systems account for 15-40% of the total energy spent in cities worldwide. An ideal energy efficient street lighting system should ensure driver safety during the night with good visibility while reducing energy uses and thus costs. This project proposes a solution using a vision-based vehicle detection to respond to real-time road and traffic conditions. The system comprises of a vehicular detection model comprised of a computer vision object detection model that can be integrated into existing surveillance systems, such as a closed-circuit television (CCTV) camera, which allows for the automatic control and dimming of LED streetlights. The computer vision object detection model uses the YOLOv8 architecture, allowing for massively decreased computational requirements and model complexity while allowing for improved model accuracy. The developed model outperforms other models similarly trained in vehicular detection, such as a model from a 2023 paper based on YOLOX. The developed model has an 89% decrease in computational requirements, requiring 14.3 GFLOPS compared to 131.0 GFLOPS, and it has a 79.85% decrease in model complexity, with 11.14 MParams compared to 55.31 MParams. The developed model also performs better with a 4.12% increase in accuracy, with a mAP of 0.961 compared to 0.923. Decreased computation requirements and model parameter count allows the model to run on more affordable hardware, allowing a more widespread adoption in communities. More installations of the developed solution means that the energy consumption of streetlights can be rapidly decreased, thus providing long lasting environmental benefits.
Downloads
References or Bibliography
Boshoff, D. W. (2019, February 18). Case Study: LED vs HPS Street lights : Provident Procurement. ProvProcure. https://provprocure.com/case-study-led-vs-hps-street-lights/
He, Q., Xu, A., Ye, Z., Wen, Z., & Cai, T. (2023). Object detection based on lightweight YOLOX for autonomous driving. Sensors, 23(17), 7596. https://doi.org/10.3390/s23177596
Hong, S. G., Kim, N. S., & Kim, W. (2013). Reduction of false alarm signals for PIR sensor in realistic outdoor surveillance. Etri Journal, 35(1), 80–88. https://doi.org/10.4218/etrij.13.0112.0219
Iparraguirre, O., Amundarain, A., Brazález, A., & Borro, D. (2021). Sensors on the move: onboard Camera-Based Real-Time traffic alerts paving the way for cooperative roads. Sensors, 21(4), 1254. https://doi.org/10.3390/s21041254
Jin, J., Fatemi, A., Lira, W., Yu, F., Leng, B., Ma, R., Mahdavi‐Amiri, A., & Zhang, H. (2021). RaidaR: A Rich Annotated Image Dataset of Rainy Street Scenes. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2104.04606
Redmon, J., Santosh, D. H. H., Ross, G., & Farhadi, A. (2015). You Only Look Once: Unified, Real-Time Object Detection. arXiv (Cornell University). https://doi.org/10.48550/arxiv.1506.02640
Sang, J., Wu, Z., Guo, P., Hu, H., Xiang, H., Zhang, Q., & Cai, B. (2018). An improved YOLOV2 for vehicle detection. Sensors, 18(12), 4272. https://doi.org/10.3390/s18124272
Shende, A., Gatpade, P., Marekar, S., & Bhaisare, S. (2022). Review of Various Technology Automatic Dimming Control of LED Street Light. International Journal for Research in Applied Science and Engineering Technology, 10(4), 2016–2018. https://doi.org/10.22214/ijraset.2022.41642
Singh, R. (2020). Day time and night time road images. Kaggle. https://www.kaggle.com/datasets/raman77768/day-time-and-night-time-road-images/data
Subramani, C., Surya, S. M., Gowtham, J., Chari, R., Srinivasan, S., Siddharth, J. P., & Shrimali, H. (2019). Energy efficiency and pay-back calculation on street lighting systems. AIP Conference Proceedings. https://doi.org/10.1063/1.5112267
Terven, J. R., & Cordova-Esparza, D. (2023). A comprehensive review of YOLO: from YOLOV1 and beyond. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2304.00501
Yang, G., Song, X., Huang, C., Deng, Z., Shi, J., & Zhou, B. (2019). DrivingStereo: A Large-Scale Dataset for Stereo Matching in Autonomous Driving Scenarios. The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.org/10.1109/cvpr.2019.00099
Yu, F., Chen, H., Wang, X., Xian, W., Chen, Y., Liu, F., Madhavan, V., & Darrell, T. (2020). BDD100K: A Diverse Driving Dataset for Heterogeneous Multitask Learning. The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.org/10.1109/cvpr42600.2020.00271
Zhang, Y., Guo, Z., Wu, J., Tian, Y., Tang, H., & Guo, X. (2022). Real-Time Vehicle Detection Based on Improved YOLO v5. Sustainability, 14(19), 12274. https://doi.org/10.3390/su141912274
Published
How to Cite
Issue
Section
Copyright (c) 2024 Cody Chen; Weile Shen

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
Copyright holder(s) granted JSR a perpetual, non-exclusive license to distriute & display this article.


