Academic literature on the topic 'Deep learning, computer vision, safety, road scene understanding'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Deep learning, computer vision, safety, road scene understanding.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Deep learning, computer vision, safety, road scene understanding"

1

Trabelsi, Rim, Redouane Khemmar, Benoit Decoux, Jean-Yves Ertaud, and Rémi Butteau. "Recent Advances in Vision-Based On-Road Behaviors Understanding: A Critical Survey." Sensors 22, no. 7 (2022): 2654. http://dx.doi.org/10.3390/s22072654.

Full text
Abstract:
On-road behavior analysis is a crucial and challenging problem in the autonomous driving vision-based area. Several endeavors have been proposed to deal with different related tasks and it has gained wide attention recently. Much of the excitement about on-road behavior understanding has been the labor of advancement witnessed in the fields of computer vision, machine, and deep learning. Remarkable achievements have been made in the Road Behavior Understanding area over the last years. This paper reviews 100+ papers of on-road behavior analysis related work in the light of the milestones achieved, spanning over the last 2 decades. This review paper provides the first attempt to draw smart mobility researchers’ attention to the road behavior understanding field and its potential impact on road safety to the whole road agents such as: drivers, pedestrians, stuffs, etc. To push for an holistic understanding, we investigate the complementary relationships between different elementary tasks that we define as the main components of road behavior understanding to achieve a comprehensive understanding of approaches and techniques. For this, five related topics have been covered in this review, including situational awareness, driver-road interaction, road scene understanding, trajectories forecast, driving activities, and status analysis. This paper also reviews the contribution of deep learning approaches and makes an in-depth analysis of recent benchmarks as well, with a specific taxonomy that can help stakeholders in selecting their best-fit architecture. We also finally provide a comprehensive discussion leading us to identify novel research directions some of which have been implemented and validated in our current smart mobility research work. This paper presents the first survey of road behavior understanding-related work without overlap with existing reviews.
APA, Harvard, Vancouver, ISO, and other styles
2

TS, Prof Nishchitha. "Real Time Object Detection in Autonomous Vehicle Using Yolo V8." INTERNATIONAL JOURNAL OF SCIENTIFIC RESEARCH IN ENGINEERING AND MANAGEMENT 09, no. 05 (2025): 1–9. https://doi.org/10.55041/ijsrem48914.

Full text
Abstract:
Abstract Autonomous vehicles rely heavily on real-time object detection to ensure safe and efficient navigation in dynamic environments. This paper explores the implementation of YOLOv8 (You Only Look Once, version 8), a state-of-the-art deep learning model for object detection, within autonomous driving systems. YOLOv8 offers enhanced speed, accuracy, and lightweight deployment capabilities compared to its predecessors, making it highly suitable for real-time applications. The model is trained and evaluated on datasets such as KITTI and COCO to detect and classify various objects including pedestrians, vehicles, traffic signs, and lane markings. The integration of YOLOv8 with on-board vehicle sensors and edge computing units enables rapid inference and low-latency decision-making. Experimental results demonstrate that YOLOv8 achieves high mean average precision (mAP) with low computational overhead, affirming its potential for deployment in real-world autonomous driving scenarios. This work highlights the advantages of YOLOv8 in improving the perception module of self-driving cars and addresses challenges related to detection in complex, real-time traffic conditions. Keywords Real-Time Object Detection Autonomous Vehicles YOLOv8 Deep Learning Computer Vision Convolutional Neural Networks (CNNs) Traffic Scene Understanding Edge Computing Mean Average Precision (mAP) Self-Driving Cars Sensor Fusion Road Safety
APA, Harvard, Vancouver, ISO, and other styles
3

Prasoona, Samala. "FogNet: An Enhanced Object Detection Model for Vehicle and Human Recognition in Foggy Conditions." INTERNATIONAL JOURNAL OF SCIENTIFIC RESEARCH IN ENGINEERING AND MANAGEMENT 09, no. 06 (2025): 1–9. https://doi.org/10.55041/ijsrem49854.

Full text
Abstract:
Abstract : Foggy weather makes it really hard for vehicle detection systems to work properly. The visibility drops, and objects on the road become hard to recognize. To tackle this problem, we developed a smart and lightweight detection method based on an improved version of the YOLOv10 model. This system doesn't just rely on raw images-it first applies a series of advanced preprocessing techniques. These include data transformations, Dehaze Formers, and dark channel methods that help clean up the foggy images and bring out the important details. By doing this, we reduce the effect of haze and make the key features more visible. We also added a special attention module to the model. This helps the system focus better on the important parts of the image by understanding both the surroundings and finer details. It's especially useful for spotting small or partially hidden vehicles and people in foggy scenes. On top of that, we improved the feature extraction process using a lightweight yet powerful module. This makes the system faster and more efficient, without compromising on accuracy. Overall, our approach offers a solid and reliable solution for detecting vehicles and humans even in tough foggy conditions, making roads safer and detection systems more dependable. Keywords: Foggy Weather, Vehicle Detection, Human Detection, YOLOv10, Lightweight Model, Image Preprocessing, Fog Removal, Dehaze Techniques, Dark Channel Method, Attention Module, Feature Extraction, Object Detection, Low Visibility, Real-Time Detection, Deep Learning, Computer Vision, Road Safety, Smart Detection System, Adverse Weather Detection, Autonomous Driving.
APA, Harvard, Vancouver, ISO, and other styles
4

Deepa Mane, Et al. "A Review on Cross Weather Traffic Scene Understanding Using Transfer Learning for Intelligent Transport System." International Journal on Recent and Innovation Trends in Computing and Communication 11, no. 10 (2023): 2027–38. http://dx.doi.org/10.17762/ijritcc.v11i10.8886.

Full text
Abstract:
Intelligent transport systems (ITS) have revolutionized the transportation industry by integrating cutting-edge technologies to enhance road safety, reduce traffic congestion and optimize the transportation network. Scene understanding is a critical component of ITS that enables real-time decision-making by interpreting the environment's contextual information. However, achieving accurate scene understanding requires vast amounts of labeled data, which can be costly and time-consuming. It is quite challenging to Understand traffic scene captured from vehicle mounted cameras. In recent times, the combination of road scene-graph representations and graph learning techniques has demonstrated superior performance compared to cutting-edge deep learning methods across various tasks such as action classification, risk assessment, and collision prediction. It's a grueling problem due to large variations under different illumination conditions. Transfer learning is a promising approach to address this challenge. Transfer learning involves leveraging pre-trained deep learning models on large-scale datasets to develop efficient models for new tasks with limited data. In the context of ITS, transfer learning can enable accurate scene understanding with less data by reusing learned features from other domains.
 This paper presents a comprehensive overview of the application of transfer learning for scene understanding in cross domain. It highlights the benefits of transfer learning for ITS and presents various transfer learning techniques used for scene understanding. This survey paper provides systematic review on cross domain outdoor scene understanding and transfer learning approaches from different perspective, presents information on current state of art and significant methods in choosing the right transfer learning model for specific scene understanding applications.
APA, Harvard, Vancouver, ISO, and other styles
5

Jose, Mekha, Joshy Avin, R. Paleri Abishek, Mohan Athul, and Jasim R. M. Ali. "A Review on Contribution and Influence of Artificial Intelligence in Road Safety and Optimal Routing." International Journal on Emerging Research Areas (IJERA) 04, no. 02 (2025): 56–60. https://doi.org/10.5281/zenodo.14669339.

Full text
Abstract:
Pothole detection is crucial for road safety and maintenance, driving research towards automated and efficient detection systems. Traditional methods present limitations: public reporting, while cost-effective, relies on citizen participation and lacks real-time information; vibration-based methods, using accelerometers to detect vehicle vibrations, require driving over potholes. Image/video processing techniques offer a proactive approach by analysing visual data to identify potholes. These methods often leverage computer vision algorithms, 3D scene reconstruction, and machine learning strategies for enhanced accuracy. Researchers are exploring deep learning models like Convolutional Neural Networks (CNNs) and YOLOv2 to im- prove real-time pothole detection accuracy and efficiency. These advancements, including stereo vision-based systems with high detection rates and pixel-level accuracy, contribute to timely pothole detection and repair, ultimately improving road safety
APA, Harvard, Vancouver, ISO, and other styles
6

Samo, Madiha, Jimiama Mosima Mafeni Mase, and Grazziela Figueredo. "Deep Learning with Attention Mechanisms for Road Weather Detection." Sensors 23, no. 2 (2023): 798. http://dx.doi.org/10.3390/s23020798.

Full text
Abstract:
There is great interest in automatically detecting road weather and understanding its impacts on the overall safety of the transport network. This can, for example, support road condition-based maintenance or even serve as detection systems that assist safe driving during adverse climate conditions. In computer vision, previous work has demonstrated the effectiveness of deep learning in predicting weather conditions from outdoor images. However, training deep learning models to accurately predict weather conditions using real-world road-facing images is difficult due to: (1) the simultaneous occurrence of multiple weather conditions; (2) imbalanced occurrence of weather conditions throughout the year; and (3) road idiosyncrasies, such as road layouts, illumination, and road objects, etc. In this paper, we explore the use of a focal loss function to force the learning process to focus on weather instances that are hard to learn with the objective of helping address data imbalances. In addition, we explore the attention mechanism for pixel-based dynamic weight adjustment to handle road idiosyncrasies using state-of-the-art vision transformer models. Experiments with a novel multi-label road weather dataset show that focal loss significantly increases the accuracy of computer vision approaches for imbalanced weather conditions. Furthermore, vision transformers outperform current state-of-the-art convolutional neural networks in predicting weather conditions with a validation accuracy of 92% and an F1-score of 81.22%, which is impressive considering the imbalanced nature of the dataset.
APA, Harvard, Vancouver, ISO, and other styles
7

Pavel, Monirul Islam, Siok Yee Tan, and Azizi Abdullah. "Vision-Based Autonomous Vehicle Systems Based on Deep Learning: A Systematic Literature Review." Applied Sciences 12, no. 14 (2022): 6831. http://dx.doi.org/10.3390/app12146831.

Full text
Abstract:
In the past decade, autonomous vehicle systems (AVS) have advanced at an exponential rate, particularly due to improvements in artificial intelligence, which have had a significant impact on social as well as road safety and the future of transportation systems. However, the AVS is still far away from mass production because of the high cost of sensor fusion and a lack of combination of top-tier solutions to tackle uncertainty on roads. To reduce sensor dependency and to increase manufacturing along with enhancing research, deep learning-based approaches could be the best alternative for developing practical AVS. With this vision, in this systematic review paper, we broadly discussed the literature of deep learning for AVS from the past decade for real-life implementation in core fields. The systematic review on AVS implementing deep learning is categorized into several modules that cover activities including perception analysis (vehicle detection, traffic signs and light identification, pedestrian detection, lane and curve detection, road object localization, traffic scene analysis), decision making, end-to-end controlling and prediction, path and motion planning and augmented reality-based HUD, analyzing research works from 2011 to 2021 that focus on RGB camera vision. The literature is also analyzed for final representative outcomes as visualization in augmented reality-based head-up display (AR-HUD) with categories such as early warning, road markings for improved navigation and enhanced safety with overlapping on vehicles and pedestrians in extreme visual conditions to reduce collisions. The contribution of the literature review includes detailed analysis of current state-of-the-art deep learning methods that only rely on RGB camera vision rather than complex sensor fusion. It is expected to offer a pathway for the rapid development of cost-efficient and more secure practical autonomous vehicle systems.
APA, Harvard, Vancouver, ISO, and other styles
8

Hindarto, Djarot. "Enhancing Road Safety with Convolutional Neural Network Traffic Sign Classification." sinkron 8, no. 4 (2023): 2810–18. http://dx.doi.org/10.33395/sinkron.v8i4.13124.

Full text
Abstract:
Recent computer vision and deep learning breakthroughs have improved road safety by automatically classifying traffic signs. This research uses CNNs to classify traffic signs to improve road safety. Autonomous vehicles and intelligent driver assistance systems require accurate traffic sign detection and classification. Using deep learning, we created a CNN model that can recognize and classify road traffic signs. This research uses a massive dataset of labeled traffic sign photos for training and validation. These CNN algorithms evaluate images and produce real-time predictions to assist drivers and driverless cars in understanding traffic signs. Advanced driver assistance systems, navigation systems, and driverless vehicles can use this technology to give drivers more precise information, improving their decision-making and road safety. Researcher optimized CNN model design, training, and evaluation metrics during development. The model was rigorously tested and validated for robustness and classification accuracy. The research also solves real-world driving obstacles like illumination, weather, and traffic signal obstructions. This research shows deep learning-based traffic sign classification can dramatically improve road safety. This technology can prevent accidents and enhance traffic management by accurately recognizing and interpreting traffic signs. It is also a potential step toward a safer, more efficient transportation system with several automotive and intelligent transportation applications. Road safety is a global issue, and CNN-based traffic sign classification can reduce accidents and improve driving. On filter 3, Convolutional Neural Network training accuracy reached 98.9%, while validation accuracy reached 88.23%.
APA, Harvard, Vancouver, ISO, and other styles
9

Xu, Zhaosheng, Zhongming Liao, Xiaoyong Xiao, Suzana Ahmad, Norizan Mat Diah, and Azlan Ismail. "Target image detection algorithm of complex road scene based on improved multi-scale adaptive feature fusion technology." International Journal for Simulation and Multidisciplinary Design Optimization 16 (2025): 6. https://doi.org/10.1051/smdo/2025004.

Full text
Abstract:
Understanding road scenes is crucial to the safe driving of autonomous vehicles, and object detection in road scenes is necessary to develop driving assistance systems. Current object detection algorithms are not very good at handling complex road scenes, and public datasets do not always adequately represent city traffic. Using Improved Multi-Scale Adaptive Feature Fusion Technology (IMSAFFT), this work suggests a real-time traffic information identification method to fix the issues of low detection accuracy of road scenes and high false detection rates in panoramic video images. In addition, a semantic recognition algorithm for a road scene based on image data is suggested. This study introduces computer vision-based approaches, including colour and texture recognition, object detection, and scene context understanding using Deep Neural Networks (DNN). An increasing number of deeper stacked layers allows the deep neural network to learn more complicated high-level semantic features, and the features' quality improves with time. A learning rate adaptive adjustment technique has been utilized to make training more efficient. After that, this improved detector is used to identify vehicles in original road environments. The suggested technique surpassed traditional detectors in the experiments with a high accuracy rate and processing speed. It worked well in real-world traffic situations for detecting overlapping, multiple, distant, and small objects. The simulation outcomes illustrate that the recommended IMSAFFT model increases the accuracy ratio of 98.4%, target image detection ratio of 97.4%, traffic prediction rate of 96.5%, processing speed rate of 10.4% and F1-score ratio of 95.4% compared to other existing models.
APA, Harvard, Vancouver, ISO, and other styles
10

Kherraki, Amine, Shahzaib Saqib Warraich, Muaz Maqbool, and Rajae El Ouazzani. "Residual balanced attention network for real-time traffic scene semantic segmentation." International Journal of Electrical and Computer Engineering (IJECE) 13, no. 3 (2023): 3281. http://dx.doi.org/10.11591/ijece.v13i3.pp3281-3289.

Full text
Abstract:
<span lang="EN-US">Intelligent transportation systems (ITS) are among the most focused research in this century. Actually, autonomous driving provides very advanced tasks in terms of road safety monitoring which include identifying dangers on the road and protecting pedestrians. In the last few years, deep learning (DL) approaches and especially convolutional neural networks (CNNs) have been extensively used to solve ITS problems such as traffic scene semantic segmentation and traffic signs classification. Semantic segmentation is an important task that has been addressed in computer vision (CV). Indeed, traffic scene semantic segmentation using CNNs requires high precision with few computational resources to perceive and segment the scene in real-time. However, we often find related work focusing only on one aspect, the precision, or the number of computational parameters. In this regard, we propose RBANet, a robust and lightweight CNN which uses a new proposed balanced attention module, and a new proposed residual module. Afterward, we have simulated our proposed RBANet using three loss functions to get the best combination using only 0.74M parameters. The RBANet has been evaluated on CamVid, the most used dataset in semantic segmentation, and it has performed well in terms of parameters’ requirements and precision compared to related work.</span>
APA, Harvard, Vancouver, ISO, and other styles
More sources

Dissertations / Theses on the topic "Deep learning, computer vision, safety, road scene understanding"

1

Schoen, Fabio. "Deep learning methods for safety-critical driving events analysis." Doctoral thesis, 2022. http://hdl.handle.net/2158/1260238.

Full text
Abstract:
In this thesis, we propose to study the data of crash and near-crash events, collectively called safety-critical driving events. Such data include a footage of the event, acquired from a camera mounted inside the vehicle, and the data from a GPS/IMU module, i.e., speed, acceleration and angular velocity. First, we introduce a novel problem, that we call unsafe maneuver classification, that aims at classifying safety-critical driving events based on the maneuver that leads to the unsafe situation and we propose a two-stream neural architecture based on Convolutional Neural Networks that performs sensor fusion and address the classification task. Then, we propose to integrate the output of an object detector in the classification task, to provide the network explicit knowledge of the entities in the scene. We design a specific architecture that leverages a tracking algorithm to extract information of a single real-world object over time, and then uses attention to ground the prediction on a single (or a few) objects, i.e., the dangerous or in danger ones, leveraging a solution that we called Spatio-Temporal Attention Selector (STAS). Finally, we propose to address video captioning of safety-critical events, with the goal of providing a description of the dangerous situation in a human-understandable form.
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Deep learning, computer vision, safety, road scene understanding"

1

Manfio Barbosa, Felipe, and Fernando Santos Osório. "3D Perception for Autonomous Mobile Robots Navigation Using Deep Learning for Safe Zones Detection: A Comparative Study." In Computer on the Beach. Universidade do Vale do Itajaí, 2021. http://dx.doi.org/10.14210/cotb.v12.p072-079.

Full text
Abstract:
Computer vision plays an important role in intelligent systems, particularly for autonomous mobile robots and intelligent vehicles. It is essential to the correct operation of such systems, increasing safety for users/passengers and also for other people in the environment. One of its many levels of analysis is semantic segmentation, which provides powerful insights in scene understanding, a task of utmost importance in autonomous navigation. Recent developments have shown the power of deep learning models applied to semantic segmentation. Besides, 3D data shows up as a richer representation of the world. Although there are many studies comparing the performances of several semantic segmentation models, they mostly consider the task over 2D images and none of them include the recent GAN models in the analysis. In this paper, we carry out the study, implementation and comparison of recent deep learning models for 3D semantic image segmentation. We consider the FCN, SegNet and Pix2Pix models. The 3D images are captured indoors and gathered in a dataset created for the scope of this project. Our main objective is to evaluate and compare the models’ performances and efficiency in detecting obstacles, safe and unsafe zones for autonomous mobile robots navigation. Considering as metrics the mean IoU values, number of parameters and inference time, our experiments show that Pix2Pix, a recent Conditional Generative Adversarial Network, outperforms the FCN and SegNet models in the
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!