Organizers

Sean Fanello, Julien Valentin, Jonathan Taylor, Christoph Rhemann, Adarsh Kowdle, Jürgen Sturm,

Christine Kaeser-Chen, Pavel Pidlypenskyi, Rohit Pandey, Andrea Tagliasacchi, Sameh Khamis,

David Kim, Mingsong Dou, Kaiwen Guo, Danhang Tang, Shahram Izadi

Augmented Perception - Google

ECCV 2018 Tutorial

The website for the ECCV 2018 tutorial is: http://augmentedperception.github.io/eccv18/

Description

The emergence of virtual and augmented reality has increased the demand of robust systems for 3D capture, reconstruction and understanding. Designing such systems involves developing high quality sensors and efficient algorithms that can leverage new and existing technologies. To this end, we design depth sensors with two properties that drastically simplify the problem of fusing incomplete sensor data. First, an UltraFast depth stream that significantly reduces the frame to frame motion. Second, by allowing multiple sensors to be easily combined (without interference) we can eliminate occlusions. As a result, we have developed a wide range of efficient algorithms for reconstruction, tracking and understanding that are designed to work with this technology. In this tutorial we will walk the audience through the ins and outs of building such a stack from the ground up. From building such a sensor to applications for mixed reality with particular emphasis on high speed 3D capture systems.

When and Where

Friday June 22nd 2018
Ballroom A - Calvin L. Rampton Salt Palace Convention Center, Salt Lake City

Program

                               
Time
Title
Speaker
9:00 - 9:10
Introduction
Shahram Izadi
AP - Google
Session 1:
Depth Sensors & 3D Capture
9:10 - 9:40
Depth Sensors and Algorithms: What, When, Where
Adarsh Kowdle
AP - Google
9:40 - 10:10
Triangulation Methods: From Theory to Mobile
Christoph Rhemann
AP - Google
10:10 - 10:30
UltraFast Machine Learning for Triangulation Methods
Sean Fanello
AP - Google
10:30 - 11:00
Coffee Break
11:00 - 11:15
StereoNet: Guided Hierarchical Refinement for Real-Time Edge-Aware Depth Prediction
Sameh Khamis
AP - Google
11:15 - 11:30
Active Stereo Net: Self-Supervised End-to-End Training for Active Stereo Systems
Yinda Zhang
Princeton University / AP - Google
Session 2:
Tracking & Relocalization
11:30 - 12:00
Non-Linear Optimization Methods
Jonathan Taylor
AP - Google
12:00 - 13:00
Lunch Break
13:00 - 13:45
Localization and Mapping - ARCore
Konstantine Tsotsos
ARCore - Google
13:45 - 14:00
Machine Learning for Camera Relocalization
Julien Valentin
AP - Google
Session 3:
World Understanding
14:00 - 14:15
3D Scene Understanding
Juergen Sturm
AP - Google
14:15 - 14:30
Semantic Labeling
Thomas Funkhouser
Princeton University / AP - Google
Session 4:
Non Rigid Tracking & Reconstruction
14:30 - 15:00
Parametric Tracking
Anastasia Tkach
Sofien Bouaziz
AP - Google
15:00 - 15:30
Non Rigid Reconstruction of Humans
Mingsong Dou
Kaiwen Guo
AP - Google
15:30 - 16:00
Coffee Break
Session 5:
Machine Learning and Interactive Applications
16:00 - 16:15
LookinGood - Enhancing Performance Capture with Real-time Neural Re-rendering
Rohit Pandey
AP - Google
16:15 - 16:30
Fast Machine Learning on Embedded Devices
Christine Kaeser-Chen
Mobile Vision - Google
16:30 - 16:45
Digital AR
David Kim
AP - Google

Please contact Shahram Izadi or Sean Fanello if you have any questions or to request the slides.