Intrinsic 3D vision system

Generate precise 3D point cloud data with a novel deep learning stereo architecture, unlocking new automation possibilities.

Read the paper

“The system's robustness and accuracy are validated on a challenging industrial dataset, showcasing its potential for industrial applications.”

Kartik Venkataraman, R&D Director, Intrinsic
Advancement #1

Deep multi-modal fusion

A first-of-its-kind deep learning algorithm that fuses RGB, IR, and polarization sensor data from multiple viewpoints into a single cost volume, enabling high-resolution, highly accurate 3D reconstructions.

Advancement #2

Multi-modal sim2real

We demonstrate a purely synthetic training framework for our multimodal sensor fusion network to prevent over-fitting. Collecting and passing polarization images (both angle & strength) along with RGB and IR, through a neural network to compute the cost volume, improving depth maps for transparent and reflective objects.

Advancement #3

Multi-unit auto calibration

An automatic calibration pipeline that precisely aligns multiple cameras (or sensors) to each other and corrects “drift”, allowing for flexible camera placement and high quality 3D without the need for manual calibration.

the research

Download the papers

More accurate 3D reconstruction enhances robotic performance

This paper has been published to Siggraph Asia

Download Papers

2 files, 20.6 mb