SynthAI

Synthetic material for AI training

Using a game engine to generate synthetic training data offers significant advantages for AI training in image classification, object detection, and animation classification. Synthetic data, created in controlled virtual environments, allows for the generation of large, diverse, and perfectly labeled datasets. This contrasts with human-labeled material, which is often expensive, time-consuming, and prone to errors. Synthetic data can be tailored to specific needs, ensuring comprehensive coverage of various scenarios and edge cases that might be underrepresented in real-world data. Additionally, synthetic environments facilitate the manipulation of variables, such as lighting and angles, providing a robust training ground for AI models. Overall, the use of synthetic material enhances the efficiency and accuracy of AI training compared to traditional human-labeled datasets.

Example

The Roof AI Training project focuses on training an AI system using Unreal Engine 5 (UE5) to generate synthetic data for machine learning applications, specifically for detecting damaged asphalt shingles on roofs. It highlights the need for massive labeled datasets to train AI effectively and discusses the role of Convolutional Neural Networks (CNNs) in image recognition tasks. UE5 is utilized to create procedurally generated roofing textures and environments, providing a cost-effective and scalable solution for data generation. The research project involves generating high-quality synthetic images of roofs, including damaged areas, to train and evaluate a CNN model. The AI aspect includes using Simplex Noise and 3-tab pattern algorithms for realistic texture generation, followed by training a CNN using PyTorch on the generated data. Preliminary results indicate that the model accurately detects most missing shingles, though further improvements are needed in bounding box precision and detection accuracy. Future goals include enhancing the procedural generation algorithm, optimizing the CNN architecture for better performance, and iteratively refining the AI through continuous training and evaluation using both synthetic and real-world data.

    

Protected: Book: Design in the Age of Extended Reality

This content is password protected. To view it please enter your password below:

Eye-Tracking for Drivers’ Visual Behavior

Impacts of Work Zone Traffic Signage Devices and Environment Complexity on Drivers’ Visual Behavior and Workers Safety.

Ph.D student: Adebisi, Adekunle. CEAS – Civil & Arch Eng & Const Mgmt

Undergraduate student: Nathan Deininger, 

Faculty. Ming Tang

The objective of this study is to investigate the safety of roadway workers under varying environmental and work zone conditions. To achieve the objectives, a driving simulator-based experiment is proposed to evaluate drivers’ visual attention under various work zone scenarios using eye-tracking technologies.

Grant.

  • Using Eye- Tracking to Study the Effectiveness of Visual Communication. UHP Discovery funding. University Honor Program. UC. $5,000. Faculty advisor. 2021.
  • Adekunle Adebisi  (Ph.D student at the College of Engineering and Applied Science) applied and received a $3,200 Emerging Fellowship Award By Academic Advisory Council for Signage Research and Education (AACSRE).

 

Tableau test

Power BI test

test PowerBI