Skip to main content
Oral defences & examinations, Thesis defences

Masters Thesis Defense: Soroush Saryazdi


Date & time
Monday, April 26, 2021
11 a.m. – 1 p.m.
Speaker(s)

Soroush Saryazdi

Cost

This event is free

Where

Online

Candidate: Soroush Saryazdi
Thesis Title: End-to-end Representation Learning for 3D Reconstruction
Date & Time: April 26th, 2021 @ 11:00 am – 13:00 pm

Online (Zoom)
Examining Committee:

Dr. Andre Delong - Chair
Dr. Mudur and Dr. Mendhurwar- Supervisors

Dr. Thomas Fevens - Examiner
Dr. Andre Delong - Examiner

 

Abstract:

Physically based rendering requires the digital representation of a scene to include both 3D geometry and material appearance properties of objects in the scene. Reconstructing such 3D representations from images of real world environments has been a long standing goal in the fields of computer vision, computer graphics, robotics, augmented and virtual reality, etc. Recently, representation learning based approaches have transformed the landscape of several domains such as image recognition and semantic segmentation. However, the prevalent way of using learning-based approaches for 3D reconstruction is still to replace one of the components in the 3D reconstruction pipeline with a pre-trained model. In this thesis, we propose approaches for using neural networks in conjunction with the 3D reconstruction pipeline such that they can be trained end-to-end based on a single end objective (e.g., to reconstruct an accurate 3D representation). 

Our main contributions include the following: 

- A fully differentiable dense visual SLAM system for reconstructing the 3D 

geometry of a scene from a sequence of RGB-D images, called ∇SLAM. This work, carried out in collaboration with the Robotics and Embodied AI Lab (REAL) at MILA, resulted in the release of the first open-source library for differentiable SLAM. 

- We propose the disentangled rendering loss for training neural networks to estimate material appearance parameters from image(s) of a near-flat surface. The disentangled rendering loss allows the network to weigh the importance of each material appearance parameter based on its effect on the final appearance of the material, while also having desirable mathematical properties for gradient-based training. 

- We describe work towards an end-to-end trainable model that can simultaneously reconstruct the 3D geometry and predict the material appearance properties of a scene. A publicly available dataset for training such a model is not currently available. So, we have created a dataset of material appearance properties for complex scenes which we intend to release publicly. 

Our approach enjoys many of the benefits of classical 3D reconstruction approaches such as interpretability (due to the modular nature) and the ability to use well-understood components from the reconstruction pipeline. Further, this approach also enjoys representation learning benefits such as the capability of solving challenging tasks which have been difficult to solve by designing explicit algorithms (e.g., material appearance property estimation for complex scenes), and their strong performance on end-to-end training tasks.

Back to top

© Concordia University