Jidong J. Yang’s research while affiliated with University of Georgia and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (53)


Figure 1. The overall architecture of our model. A pretrained VAE extracts clear features, while the DerainAE and DepthNet modules handle rainy images. Latent space comparison between clear and rainy features improves depth estimation and deraining images prediction.
Figure 4. Vehicle detection results using YOLOv11 on the RainKITTI2015 dataset. Red bounding boxes denote the detected vehicles.
Comparsion results on the RainCityscapes testing dataset. We report the average, minimum and maximum of PSNR and SSIM mertrics on PReNet, DID-MDN and our model.
Ablation settings (A-E). Compared to our full model, we conduct an ablation study by removing component(s) to evaluate their respective contributions.
PSNR and SSIM results of the model trained without depth latent constraint (WO depth latent) on three outdoor datasets: RainCityScapes, RainKITTI2012, and RainKITTI2015.

+3

Leveraging Scene Geometry and Depth Information for Robust Image Deraining
  • Article
  • Full-text available

January 2025

·

25 Reads

Ningning Xu

·

Jidong J. Yang

Image deraining holds great potential for enhancing the vision of autonomous vehicles in rainy conditions, contributing to safer driving. Previous works have primarily focused on employing a single network architecture to generate derained images. However, they often fail to fully exploit the rich prior knowledge embedded in the scenes. Particularly, most methods overlook the depth information that can provide valuable context about scene geometry and guide more robust deraining. In this work, we introduce a novel learning framework that integrates multiple networks: an AutoEncoder for deraining, an auxiliary network to incorporate depth information, and two supervision networks to enforce feature consistency between rainy and clear scenes. This multi-network design enables our model to effectively capture the underlying scene structure, producing clearer and more accurately derained images, leading to improved object detection for autonomous vehicles. Extensive experiments on three widely used datasets demonstrated the effectiveness of our proposed method.

Download

Enhancing nighttime vehicle detection with day-to-night style transfer and labeling-free augmentation

January 2025

·

23 Reads

Applied Computing and Intelligence

Deep learning-based object detection models perform well under daytime conditions but face significant challenges at night, primarily because they are predominantly trained on daytime images. Additionally, training with nighttime images presents another challenge: Even human annotators struggle to accurately label objects in low-light conditions. This issue is particularly pronounced in transportation applications, such as detecting vehicles and other objects of interest on rural roads at night, where street lighting is often absent, and headlights may introduce undesirable glare. In this study, we addressed these challenges by introducing a novel framework for labeling-free data augmentation, leveraging synthetic data generated by the Car Learning to Act (CARLA) simulator for day-to-night image style transfer. Specifically, the framework incorporated the efficient attention Generative Adversarial Network for realistic day-to-night style transfer and used CARLA-generated synthetic nighttime images to help the model learn the vehicle headlight effect. To evaluate the efficacy of the proposed framework, we fine-tuned the state-of-the-art object detection model with an augmented dataset curated for rural nighttime environments, achieving significant improvements in nighttime vehicle detection. This novel approach was simple yet effective, offering a scalable solution to enhance deep learning-based detection systems in low-visibility environments and extended the applicability of object detection models to broader real-world contexts.



Enhancing autonomous vehicle safety in rain: a data-centric approach for clear vision

December 2024

·

4 Reads

Autonomous vehicles face significant challenges in navigating adverse weather, particularly rain, due to the visual impairment of camera-based systems. In this study, we leveraged contemporary deep learning techniques to mitigate these challenges, aiming to develop a vision model that processes live vehicle camera feeds to eliminate rain-induced visual hindrances, yielding visuals closely resembling clear, rain-free scenes. Using the Car Learning to Act (CARLA) simulation environment, we generated a comprehensive dataset of clear and rainy images for model training and testing. In our model, we employed a classic encoder-decoder architecture with skip connections and concatenation operations. It was trained using novel batching schemes designed to effectively distinguish high-frequency rain patterns from low-frequency scene features across successive image frames. To evaluate the model performance, we integrated it with a steering module that processes front-view images as input. The results demonstrated notable improvements in steering accuracy, underscoring the model's potential to enhance navigation safety and reliability in rainy weather conditions.


Leveraging Scene Geometry and Depth Information for Robust Image Deraining

December 2024

·

2 Reads

Image deraining holds great potential for enhancing the vision of autonomous vehicles in rainy conditions, contributing to safer driving. Previous works have primarily focused on employing a single network architecture to generate derained images. However, they often fail to fully exploit the rich prior knowledge embedded in the scenes. Particularly, most methods overlook the depth information that can provide valuable context about scene geometry and guide more robust deraining. In this work, we introduce a novel learning framework that integrates multiple networks: an AutoEncoder for deraining, an auxiliary network to incorporate depth information, and two supervision networks to enforce feature consistency between rainy and clear scenes. This multi-network design enables our model to effectively capture the underlying scene structure, producing clearer and more accurately derained images, leading to improved object detection for autonomous vehicles. Extensive experiments on three widely-used datasets demonstrated the effectiveness of our proposed method.


Enhancing Nighttime Vehicle Detection with Day-to-Night Style Transfer and Labeling-Free Augmentation

December 2024

·

20 Reads

Existing deep learning-based object detection models perform well under daytime conditions but face significant challenges at night, primarily because they are predominantly trained on daytime images. Additionally, training with nighttime images presents another challenge: even human annotators struggle to accurately label objects in low-light conditions. This issue is particularly pronounced in transportation applications, such as detecting vehicles and other objects of interest on rural roads at night, where street lighting is often absent, and headlights may introduce undesirable glare. This study addresses these challenges by introducing a novel framework for labeling-free data augmentation, leveraging CARLA-generated synthetic data for day-to-night image style transfer. Specifically, the framework incorporates the Efficient Attention Generative Adversarial Network for realistic day-to-night style transfer and uses CARLA-generated synthetic nighttime images to help the model learn vehicle headlight effects. To evaluate the efficacy of the proposed framework, we fine-tuned the YOLO11 model with an augmented dataset specifically curated for rural nighttime environments, achieving significant improvements in nighttime vehicle detection. This novel approach is simple yet effective, offering a scalable solution to enhance AI-based detection systems in low-visibility environments and extend the applicability of object detection models to broader real-world contexts.


Feature Group Tabular Transformer: A Novel Approach to Traffic Crash Modeling and Causality Analysis

December 2024

·

4 Reads

Reliable and interpretable traffic crash modeling is essential for understanding causality and improving road safety. This study introduces a novel approach to predicting collision types by utilizing a comprehensive dataset fused from multiple sources, including weather data, crash reports, high-resolution traffic information, pavement geometry, and facility characteristics. Central to our approach is the development of a Feature Group Tabular Transformer (FGTT) model, which organizes disparate data into meaningful feature groups, represented as tokens. These group-based tokens serve as rich semantic components, enabling effective identification of collision patterns and interpretation of causal mechanisms. The FGTT model is benchmarked against widely used tree ensemble models, including Random Forest, XGBoost, and CatBoost, demonstrating superior predictive performance. Furthermore, model interpretation reveals key influential factors, providing fresh insights into the underlying causality of distinct crash types.


An Information Gradient Approach to Optimizing Traffic Sensor Placement in Statewide Networks

October 2024

·

14 Reads

·

1 Citation

Traffic sensors are vital to the development and operation of Intelligent Transportation Systems, providing essential data for traffic monitoring, management, and transportation infrastructure planning. However, optimizing the placement of these sensors, particularly across large and complex statewide highway networks, remains a challenging task. In this research, we presented a novel search algorithm designed to address this challenge by leveraging information gradients from K-nearest neighbors within an embedding space. Our method enabled more informed and strategic sensor placement under budget and resource constraints, enhancing overall network coverage and data quality. Additionally, we incorporated spatial kriging analysis, harnessing spatial correlations of existing sensors to refine and reduce the search space. Our proposed approach was tested against the widely used Genetic Algorithm, demonstrating superior efficiency in terms of convergence time and producing more effective solutions with reduced information loss.



Analyzing the importance of network topology in AADT estimation: insights from travel demand models using graph neural networks

September 2024

·

56 Reads

Transportation

Metropolitan traffic networks are becoming increasingly complex due to the growing population and diverse range of travel modes. However, the limited installation of continuous count stations leads to partially observable networks, posing a significant challenge for effective highway planning and design practices at various scales. Travel demand models have been developed and calibrated using sparse traffic counts at the metropolitan level. Nevertheless, these models are cumbersome to recalibrate and rerun whenever network changes occur. To overcome this challenge, we propose a flexible learning-based approach that extracts embedded knowledge from large-scale activity-based travel demand models to estimate Annual Average Daily Traffic (AADT). The approach offers two primary advantages: (1) directly learning network flow patterns based on segment attributes and network topology that can be transferred across regions, and (2) enabling efficient and reliable AADT estimation for projects of various scales. Our study explores a wide range of machine learning techniques, including novel graph neural networks that explicitly account for network topology, as well as modern and traditional regression and regression kriging models, which either disregard or implicitly consider network topology. We conducted extensive experiments using the loaded network data from the activity-based travel demand model for the Atlanta metropolitan area. Our findings underscore the importance of network topology in AADT estimation, with the diffusion graph convolutional network model demonstrating the best performance in both transductive and inductive settings. Additionally, modern tree ensemble models such as random forest regressor and CatBoost, despite their ignorance of network topology, show the second-best inductive performance with relatively lightweight structures.


Citations (29)


... Despite their widespread use and demonstrated performance improvements across various tasks [6,1,2,3,7,8], fixed pseudo-labeling methods have an inherent limitation: they depend heavily on a predetermined threshold to filter low-quality pseudolabels. In this approach, any unlabeled data point with confidence higher than the threshold is utilized for training, while the rest is discarded, regardless of the category. ...

Reference:

DyConfidMatch: Dynamic Thresholding and Re-sampling for 3D Semi-supervised Learning
An Information Gradient Approach to Optimizing Traffic Sensor Placement in Statewide Networks

... Lu et al. leveraged the encoder-decoder architecture of autoencoders, effectively correcting the detected vehicle positions and speeds in traffic flow data. Huang et al. [38] combined clustering algorithms with autoencoders using diffusion graph convolution networks to repair road network sensor data, achieving an experimental accuracy of over 99%. Boquet et al. [39] highlighted the potential of variational autoencoders (VAEs) in traffic data imputation and provided a detailed explanation of their principles. ...

Cluster-guided denoising graph auto-encoder for enhanced traffic data imputation and fault detection
  • Citing Article
  • October 2024

Expert Systems with Applications

... Measures to control flow speeds along streams and rivers are, therefore, essential to keep the flow speed below these structures at acceptable levels. The soil support conditions surrounding and supporting the foundation elements of a bridge can also be monitored by specialized instrumentation that can support the maintenance and preservation efforts of a historic bridge (46,47). ...

A review of bridge scour monitoring techniques and developments in vibration based scour monitoring for bridge foundations

Advances in Bridge Engineering

Alan Kazemian

·

Tien Yee

·

Metin Oguzmert

·

[...]

·

Dale Goff

... High-strength steels and composites not Lako & Barko that can withstand the extreme conditions that occur during intense hydrodynamic processes. P. Deng et al. (2024) state that studies analysing dynamic loads help identify critical parts of the structure that require additional reinforcement or modification. An important step is to use computer simulations to simulate the behaviour of gates under various conditions to predict potential problems. ...

Deep Learning-Based Flood Detection for Bridge Monitoring Using Accelerometer Data

... Reference [12] enhanced fault detection in time-series traffic sensor data through symmetric contrastive learning. Their approach has improved the accuracy of sensor data analysis, which is vital for the real-time processing capabilities required for autonomous navigation [13]. ...

Symmetric contrastive learning for robust fault detection in time-series traffic sensor data

International Journal of Data Science and Analytics

... In [58], the authors proposed a dynamic phasing sequence in acyclic signal control with Q-Learning. The authors created three models with different state depictions to study the optimum state model for various traffic scenarios. ...

Enhancing the Robustness of Traffic Signal Control with StageLight: A Multiscale Learning Approach

... The TSLP in this study can be stated as conditional upon the existing sensor locations, choosing the locations of new sensors to optimize the network coverage by maximizing network-level information gain. Optimality is sought in an embedding space that considers network topology [13]. Maximizing information gain is equivalent to minimizing the Kullback-Leibler (KL) divergence between the data distribution, P(x), and the model distribution, Q(x). ...

Strategic Sensor Placement in Expansive Highway Networks: A Novel Framework for Maximizing Information Gain

... The paper in Ref. [23] introduces a novel approach to leveraging features learned from both supervised and self-supervised paradigms, to improve image classification tasks, specifically for vehicle classification. Two state-of-the-art self-supervised learning methods, DINO and data2vec, were evaluated and compared by the authors for their representation learning of vehicle images. ...

Image-Based Vehicle Classification by Synergizing Features from Supervised and Self-Supervised Learning Paradigms

... The model architecture devised in this study draws inspiration from two influential architectures: The Deep Convolutional Generative Adversarial Network (DCGAN) [2] and the U-Net [3]. DCGAN, renowned for its applications in computer vision such as image generation [4,5,6], style transfer [7,8], and data augmentation [9,10,11], serves as a foundational pillar in our approach. Moreover, U-Net, initially developed for biomedical image segmentation, is important in modern diffusion models for iterative image denoising [12,13,14]. ...

Co-supervised learning paradigm with conditional generative adversarial networks for sample-efficient classification

Applied Computing and Intelligence

... Forget Gate (n t ): This gate is essential in culling information considered as non-essential from the cell state [20]. It operates through the following equation: ...

Deep-Learning-Based Temporal Prediction for Mitigating Dynamic Inconsistency in Vehicular Live Loads on Roads and Bridges