AsiaIPEX is a one-stop-shop for players in the IP industry, facilitating IP trade and connection to the IP world. Whether you are a patent owner interested in selling your IP, or a manufacturer looking to buy technologies to upgrade your operation, you will find the portal a useful resource.

Robust Visual-Inertial Sensor Fusion For Navigation, Localization, Mapping, And 3D Reconstruction

Technology Benefits
Uses integrated inertial and vision measurementsImproved robustness and performanceFocuses on handling outliers
Technology Application
Augmented and virtual realityRoboticsAutonomous vehicles and flying robotsIndoor localization in GPS-denied areasEgo-motion estimation
Detailed Technology Description
Researchers led by Professor Stefano Soatto have developed a novel sensor fusion system that integrates inertial and vision measurements to estimate 3D positon and orientation, along with a point-cloud model of the 3D world surrounding it. This invention has better robustness and performance that other performing VINS schemes, such as Google Tango, and with the same computational footprint. This unique technology addresses the problem of inferring ego-motion of a sensor platform from visual and inertial measurements, focusing on handling outliers.
Application No.
20160140729
Others

Background

Vision-augmented navigation or VINS is central to augmented and virtual reality, robotics, autonomous vehicles, and navigation in mobile phones. The future growth of these applications depends on reliable navigation in dynamic environments, thus improvement to these systems is of importance. Current methods rely upon low-level processing of visual data for 3D motion estimation. However, the processing is substantially useless and easily 60 – 90% of sparse features selected and tracked across frames are inconsistent with a single rigid motion due to illumination effects, occlusions, and independently moving objects. These effects are global to the scene, while low-level processing is local to the image, so it is not realistic to expect significant improvements in the vision front-end. Instead, it is critical for algorithms utilizing vision to leverage other sensory modalities, such as inertial.


Related Materials

E. S. Jones and S. Soatto. Visual-Inertial Navigation, Mapping and Localization: A Scalable Real-Time Causal Approach. The International Journal of Robotics Research. 2011.
K. Tsotsos, A. Chiuso, and S. Soatto. Robust Inference for Visual-Inertial Sensor Fusion. 2015 IEEE International Conference on Robotics and Automation. 2015.


Additional Technologies by these Inventors


Tech ID/UC Case

27401/2015-346-0


Related Cases

2015-346-0

*Abstract
UCLA researchers in the Computer Science Department have invented a novel model for a visual-inertial system (VINS) for navigation, localization, mapping, and 3D reconstruction applications.
*IP Issue Date
May 19, 2016
*Principal Investigator

Name: Stefano Soatto

Department:


Name: Konstantine Tsotsos

Department:

Country/Region
USA

For more information, please click Here
Mobile Device