UCSD Computer Scientists Develop Ubiquitous Video Application for 3D Environments

June 7, 2005 / By Doug Ramsey

McCurdy Video 1
Grad student Neil McCurdy explains how RealityFlythrough traces its roots to tele-presence, tele-reality and ubiquitous computing.
Length: 3:26

San Diego, CA, June 7, 2005 --  At MobiSys 2005 in Seattle yesterday, computer scientists affiliated with Calit2 at UCSD introduced a new technique for mixing images and video feeds from mobile cameras to provide remote viewers with a virtual window into a physical environment. Dubbed 'RealityFlythrough,' the application constructs a 3D virtual environment dynamically out of the live video streams.

"Instead of watching all the feeds simultaneously on a bank of monitors, the viewer can navigate an integrated, interactive environment as if it were a video game," said UCSD computer science and engineering professor Bill Griswold, who is working on the project with Ph.D. candidate Neil McCurdy. "RealityFlythrough creates the illusion of complete live camera coverage in a physical space. It's a new form of situational awareness, and we designed a system that can work in unforgiving environments with intermittent network connectivity."

McCurdy
Professor Bill Griswold (left) and Neil McCurdy test the Reality Flythrough system during the May 12 MMST disaster drill.

The researchers at UCSD's Jacobs School of Engineering have already begun testing the software for homeland security and emergency response, but they say that the technology has other potential consumer uses as well. "With virtual tourism, for instance, you could walk down the streets of Bangkok to see what it will be like before getting there," said McCurdy. "Another really cool application is pre-drive driving instructions. Imagine going to your favorite mapping website, where currently you get a set of instructions to turn left here or right there, and instead, you can 'fly' through the drive before doing it."  Other commercial uses could include virtual shopping, virtual mobility for the disabled, any-angle stadium views or 'my day' diaries.

At MobiSys 2005, McCurdy presented a joint paper* with Griswold about RealityFlythrough and a "systems architecture for ubiquitous video." The third international conference on mobile systems, applications and services brings together academic and industry researchers in the area of mobile and wireless systems.

Griswold and McCurdy are testing their new system as part of the WIISARD (Wireless Internet Information System for Medical Response in Disasters) project, which is funded by NIH's National Library of Medicine. During a May 12 disaster drill organized by San Diego's Metropolitan Medical Strike Team, the researchers shadowed a hazmat team responding to a simulated terrorist attack. They wore cameras mounted on their hardhats, tilt sensors with magnetic compasses, and global positioning (GPS) devices. Walking through the simulated disaster scene at the city's Cruise Ship Terminal, McCurdy and Griswold captured continuous video to be fed over an ad hoc wireless network to a makeshift command post nearby.

McCurdy Video 2
Neil McCurdy outlines potential commercial applications of the new software, e.g. virtual tourism.
Length: 1:49

The RealityFlythrough software automatically stitches the feeds together, by integrating the visual data with the camera's location and direction it is pointing. "Our system works in ubiquitous and dynamic environments, and the cameras themselves are moving and shifting," said McCurdy, who expects to finish his Ph.D. in 2006. "RealityFlythrough situates still photographs or live video in a three-dimensional environment, making the transition between two cameras while projecting the images onto the screen. We're cheating and flattening space into two dimensions and then re-projecting the images in 3D space."

The UCSD researchers say the biggest research challenge was to overcome the limitation of incomplete coverage of live video streams. "Every square meter of a space cannot be viewed from every angle with a live video stream at any given moment," said Griswold, a Calit2 participant and head of the ActiveCampus project. "We had to find a way to fill in the empty space that would give the user a sense of how the video streams relate to one another spatially."

Flythrough Software
 The software dynamically stitches together still images and live video feeds to create a 3D environment.

Their solution: RealityFlythrough fills in the gaps in coverage with the most recent still images captured during camera pans. The software then blends the imagery with smooth transitions that simulate the sensation of a human performing a walking camera pan - even when one of the images is a still frame. If older images are not desirable (e.g. in some security applications), the fill-in images can be omitted, or shown in sepia, or include an icon displaying how old the photo is.

The fundamental research finding to date, according to McCurdy, is that some of the processing can be offloaded to the human. "We take advantage of a principle called closure, which allows our brains to make sense of incomplete information.  The visual cortex does this all the time when it corrects for blind spots in our vision, for example," explained the graduate student. "RealityFlythrough supplies as much information as possible to the human operator, and the operator can easily fill in the blanks.that operator must fill in the blanks."

Human input is especially important indoors, where GPS cannot provide adequate location information. McCurdy carried a 'dead reckoning' device on his back during the May 12 disaster drill. The device uses gyros and other components to track body movement directions and footsteps from the moment the user enters an indoor area.

McCurdy Video 3
 'Tele-Reality in the Wild' is a research video produced by Neil McCurdy in 2004 about an earlier version of the technology.
Length: 6:53

Since dead-reckoning systems lose accuracy over time, the researchers implemented a system that allows the camera operators to periodically correct their locations."We created a Wizard-of-Oz approach to correcting inadequate location information," explained McCurdy. "Since we're combining this self-reporting technology with GPS or dead reckoning, it only has to be done occasionally. From all the footage we got from the May 12 drill, I only had to put in four corrections, and that was sufficient to give us pretty good accuracy indoors."

McCurdy will work on refining the system for his dissertation. And if consumers start to show interest in RealityFlythrough, McCurdy holds open the possibility of starting up a company to commercialize the technology -- but only after first finishing his Ph.D.

"A Systems Architecture for Ubiquitous Video," Neil J. McCurdy and William G. Griswold. MobiSys 2005.  

Related Links
RealityFlythrough Website
May 12 Disaster Drill
WIISARD
Calit2
MobiSys 2005

Related Articles
UCSD Researchers Test Wireless Technologies in Simulated Medical Disaster Response Drill

Media Contacts Media Contact: Doug Ramsey, (858) 822-5825, dramsey@ucsd.edu