CVG HOLO

CVG-HOLO – WAYFINDING HOLOGRAM PROJECT

XR-Lab is working with Cincinnati/Northern Kentucky International Airport (CVG), in collaboration with UC Center for Simulations & Virtual Environments Research, to

  1. Develop and demonstrate a wayfinding hologram.
  2. Evaluate the hologram signage’s performance to augment passengers’ wayfinding experience.
  3. Develop concepts of Concourse-B store renovation, integrating emerging digital technologies related to Extended Reality
  4. Develop a digital twin model of the CVG Concourse-B store area.

The project will apply various methods, including eye-tracking, motion capture, motion tracking, and computer vision.

Hologram. Reference Image from SVG news. 10.2023

Project Client: Josh Edwards, Sr. Manager, Innovation Cincinnati/Northern Kentucky International Airport

UC Team:

  • eXtended Reality Lab: Ming Tang, Director eXtended Reality Lab Digital Futures tangmg@ucmail.uc.edu
  • UCSIM Project Lead: Chris M. Collins.  Director. Center for Simulations & Virtual Environments Research
  • ARCH 7014 students. Fall. 2023

concept of hologram in CVG. by students in ARCH 7014. Fall 2023, UC. 

Thanks for the support from the UHP Discovery Summer program. 

Check out more on the student projects and eye-tracking analysis on CVG renovation.  or way-finding research projects and publications at XR-Lab. 

Wayfinding through VR

Use VR walkthrough for wayfinding research. Players’ routes, and walking behavior, such as head movement, are captured and evaluated.

Credit: restaurant designed by Eian Bennett.
More info on the wayfinding and Egress at the simulated DAAP building can be found here.

 

Cloud-based Digital Twin

Clients are one click away from interacting with a Digital Twin model on their personal devices. No installation is required.

The XR-Lab’s project showcases a cloud-based Digital Twin (DT) model, designed for accessibility and interaction via mobile devices. This advanced DT model allows multiple users to engage with its complex features directly through touch screens, eliminating the need for app installations. Clients can effortlessly access the content using a simple URL in a web browser on their personal iOS or Android mobile devices and tablets. The project is distinguished by its photorealistic renderings, which are streamed to clients at high frame rates, ensuring a visually rich and seamless experience. Furthermore, our DT model is an integration of various cutting-edge technologies, including Building Information Modeling (BIM), Metadata, IoT sensor data, 360-degree images/videos, and web3D content, creating a comprehensive and interactive digital environment.

 

More information on Future of Work: Understanding the interrelationships between humans and technology to improve the quality of work-life in smart buildings.

Kao Metaverse

The University of Cincinnati, through its Digital Futures complex, will work collaboratively with the UC Center for Simulations & Virtual Environments Research, Lindner College of Business, UC DAAP XR-Lab, and Kao to develop concepts and a minimum viable product for a Jergens virtual tanning experience called the ‘Glowverse.’

KAO-STP-Glowverse VR retail project. GLOWVERSE VIRTUAL SPA (VR) development. UC-SIM + XR-Lab +College of Business. Funded by KAO.  PI: Chris Collins. Co-PI: Ming Tang, Noah Van. $52,122. Period: 2.2023-11.2023.

  • UCSIM Project Lead: Chris M. Collins, Center for Simulations & Virtual Environments
    Research
    UCSIM Technical Lead: Ryan Gorsuch, Center for Simulations & Virtual Environments
    Research
  • UC DAAP Design Lead: Ming Tang, Director of XR-Lab. Registered Architect, RA, NCARB, LEED AP (BD+C),
    and Associate Professor at the School of Architecture and Interior Design, College of Design,
    Architecture, Art, and Planning
  • UC LCOB Marketing Lead: Noah Van Bergen, Asst. Professor, Marketing, Linder Business
    Honors Faculty, Carl H. Lindner College of Business

Solar System in VR

Extended Reality 3D Model Application in Space Exploration and Planetary Habitation

 

Developed by Mikhail NikolaenkoStudent, XR-Lab fellow, University of Cincinnati.

Summary:

This research project incorporates the use of data science, astronomy, and VR to create a visually interactive learning tool for students, academics, enthusiasts, and professionals alike to learn about areas of space exploration that will be easily accessible to anyone with a VR device such as an Oculus Quest 2. The application will include an accurate mapping of different celestial bodies, such as planets and stars, and the model will be fully interactable through functions such as scaling, time manipulation, and highlighting. The uses of this proposed application range from basic elementary applications (e.g., learning about our solar system in astronomy courses) to astronomical data research (e.g., viewing spectra of celestial objects found by Gaia)

PDFs:  Final Report; Poster

Acknowledgments: 

Project Advisor: Ming Tang

Research Funding Source: University of Cincinnati Space Research Institute for Discovery and Exploration 

External Celestial Data: This work has made use of data from the European Space Agency (ESA) mission Gaia (https://www.cosmos.esa.int/gaia), processed by the Gaia Data Processing and Analysis Consortium (DPAC,https://www.cosmos.esa.int/web/gaia/dpac/consortium). Funding for the DPAC has been provided by national institutions, in particular, the institutions participating in the Gaia Multilateral Agreement. 

Internal Solar System Data: Acton, C.H.; “Ancillary Data Services of NASA’s Navigation and Ancillary Information Facility;” Planetary and Space Science, Vol. 44, No. 1, pp. 65-70, 1996. DOI 10.1016/0032-0633(95)00107-7 

Charles Acton, Nathaniel Bachman, Boris Semenov, Edward Wright; A look toward the future in the handling of space science mission geometry; Planetary and Space Science (2017); DOI 10.1016/j.pss.2017.02.013 

References & Sources: 

1. NASA. (2018, December 14). NASA’s eyes: Eyes on exoplanets. NASA. Retrieved October 11, 2022, from https://eyes.nasa.gov/eyes-on-exoplanets.html


Demo at the 2023 Undergraduate Scholarly Showcase. 04.20.2023