Multi-Input All-Weather Streetlight to Reduce Carbon Footprint of Illuminating Road Infrastructure

Authors

  • Cody Chen Los Gatos High School
  • Weile Shen

DOI:

https://doi.org/10.47611/jsrhs.v13i4.8337

Keywords:

computer vision, vehicle detection, convolutional neural network, machine learning, infrared sensor

Abstract

Street lighting systems account for 15-40% of the total energy spent in cities worldwide. An ideal energy efficient street lighting system should ensure driver safety during the night with good visibility while reducing energy uses and thus costs. This project proposes a solution using a vision-based vehicle detection to respond to real-time road and traffic conditions. The system comprises of a vehicular detection model comprised of a computer vision object detection model that can be integrated into existing surveillance systems, such as a closed-circuit television (CCTV) camera, which allows for the automatic control and dimming of LED streetlights. The computer vision object detection model uses the YOLOv8 architecture, allowing for massively decreased computational requirements and model complexity while allowing for improved model accuracy. The developed model outperforms other models similarly trained in vehicular detection, such as a model from a 2023 paper based on YOLOX. The developed model has an 89% decrease in computational requirements, requiring 14.3 GFLOPS compared to 131.0 GFLOPS, and it has a 79.85% decrease in model complexity, with 11.14 MParams compared to 55.31 MParams. The developed model also performs better with a 4.12% increase in accuracy, with a mAP of 0.961 compared to 0.923. Decreased computation requirements and model parameter count allows the model to run on more affordable hardware, allowing a more widespread adoption in communities. More installations of the developed solution means that the energy consumption of streetlights can be rapidly decreased, thus providing long lasting environmental benefits.

Downloads

Download data is not yet available.

References or Bibliography

Boshoff, D. W. (2019, February 18). Case Study: LED vs HPS Street lights : Provident Procurement. ProvProcure. https://provprocure.com/case-study-led-vs-hps-street-lights/

He, Q., Xu, A., Ye, Z., Wen, Z., & Cai, T. (2023). Object detection based on lightweight YOLOX for autonomous driving. Sensors, 23(17), 7596. https://doi.org/10.3390/s23177596

Hong, S. G., Kim, N. S., & Kim, W. (2013). Reduction of false alarm signals for PIR sensor in realistic outdoor surveillance. Etri Journal, 35(1), 80–88. https://doi.org/10.4218/etrij.13.0112.0219

Iparraguirre, O., Amundarain, A., Brazález, A., & Borro, D. (2021). Sensors on the move: onboard Camera-Based Real-Time traffic alerts paving the way for cooperative roads. Sensors, 21(4), 1254. https://doi.org/10.3390/s21041254

Jin, J., Fatemi, A., Lira, W., Yu, F., Leng, B., Ma, R., Mahdavi‐Amiri, A., & Zhang, H. (2021). RaidaR: A Rich Annotated Image Dataset of Rainy Street Scenes. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2104.04606

Redmon, J., Santosh, D. H. H., Ross, G., & Farhadi, A. (2015). You Only Look Once: Unified, Real-Time Object Detection. arXiv (Cornell University). https://doi.org/10.48550/arxiv.1506.02640

Sang, J., Wu, Z., Guo, P., Hu, H., Xiang, H., Zhang, Q., & Cai, B. (2018). An improved YOLOV2 for vehicle detection. Sensors, 18(12), 4272. https://doi.org/10.3390/s18124272

Shende, A., Gatpade, P., Marekar, S., & Bhaisare, S. (2022). Review of Various Technology Automatic Dimming Control of LED Street Light. International Journal for Research in Applied Science and Engineering Technology, 10(4), 2016–2018. https://doi.org/10.22214/ijraset.2022.41642

Singh, R. (2020). Day time and night time road images. Kaggle. https://www.kaggle.com/datasets/raman77768/day-time-and-night-time-road-images/data

Subramani, C., Surya, S. M., Gowtham, J., Chari, R., Srinivasan, S., Siddharth, J. P., & Shrimali, H. (2019). Energy efficiency and pay-back calculation on street lighting systems. AIP Conference Proceedings. https://doi.org/10.1063/1.5112267

Terven, J. R., & Cordova-Esparza, D. (2023). A comprehensive review of YOLO: from YOLOV1 and beyond. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2304.00501

Yang, G., Song, X., Huang, C., Deng, Z., Shi, J., & Zhou, B. (2019). DrivingStereo: A Large-Scale Dataset for Stereo Matching in Autonomous Driving Scenarios. The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.org/10.1109/cvpr.2019.00099

Yu, F., Chen, H., Wang, X., Xian, W., Chen, Y., Liu, F., Madhavan, V., & Darrell, T. (2020). BDD100K: A Diverse Driving Dataset for Heterogeneous Multitask Learning. The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.org/10.1109/cvpr42600.2020.00271

Zhang, Y., Guo, Z., Wu, J., Tian, Y., Tang, H., & Guo, X. (2022). Real-Time Vehicle Detection Based on Improved YOLO v5. Sustainability, 14(19), 12274. https://doi.org/10.3390/su141912274

Published

11-30-2024

How to Cite

Chen, C., & Shen, W. (2024). Multi-Input All-Weather Streetlight to Reduce Carbon Footprint of Illuminating Road Infrastructure. Journal of Student Research, 13(4). https://doi.org/10.47611/jsrhs.v13i4.8337

Issue

Section

HS Research Projects