Research Overview

Large-Scale Scene 3D Reconstruction and Understanding

3D Reconstruction of UCAS

We focus on 3D computer vision research, especially focus on large-scale scene 3D reconstruction and understanding. Here is a brief introduction to our research work.

A complete large-scale scene 3D reconstruction and understanding system includes autonomous data acquisition, joint pose calibration, 3D geometric reconstruction, 3D semantic segmentation, 3D vectorized representation, long-term global localization, and incremental map updating. These modules form a closed-loop system in a progressive manner as shown in Figure 1, which supports digital cities, unmanned systems, digital twin and other fields. Below, we briefly describe the main research content of each key module in the 3D reconstruction and understanding system.

Large-scale scene 3D reconstruction and understanding system
Large-scale scene 3D reconstruction and understanding system

Autonomous Data Acquisition: Data acquisition is the first step in 3D reconstruction and understanding system, which directly affects the computational accuracy and stability of subsequent steps. The purpose of autonomous scene data acquisition is to use robots to autonomously collect massive data of large-scale complex scenes to achieve a complete coverage of the scene, while avoiding the subjectivity, experience and irreproducibility of manual data collection. This research includes robot global route dynamic planning, local aera discovery and exploration, optimal acquisition viewpoint selection, scene data coverage evaluation, etc.

Robot data acquisition (left: data acquisition robot, middle: 360-degree panoramic camera, right: 16-line LiDAR)
Robot data acquisition (left: data acquisition robot, middle: 360-degree panoramic camera, right: 16-line LiDAR)

Joint Pose Calibration: After acquiring massive scene data, the purpose of joint pose calculation is to calculate the precise spatial pose of the sensor at each moment in a unified coordinate system, so as to achieve spatio-temporal alignment of massive data. Especially when we collect data from multiple platforms (UAVs, unmanned vehicles, robots, etc.) and multiple modes (images, lasers, millimeter waves), accurate sensor pose calibration is critical for subsequent multi-mode integrating reconstruction and understanding. This research includes cross-modal data matching, robust scene graph construction, incremental and global scene reconstruction, online SLAM and offline SfM fusion, device-cloud collaborative reconstruction, etc.

Vehicle image pose calibration (25,000 images, 5,000 square meters) (left: front view image, right: camera trajectory and 3D sparse map)
Vehicle image pose calibration (25,000 images, 5,000 square meters) (left: front view image, right: camera trajectory and 3D sparse map)

3D Geometric Reconstruction: After obtaining the accurate spatial pose of the sensor at each moment, the purpose of 3D geometric reconstruction is to calculate a complete and accurate 3D geometric model of the scene. The geometric model here may be a dense 3D point cloud, a dense triangular mesh, or a dense 3D textured mesh. The geometric model is a 3D digitization of the real world, which is a reconstruction of the overall structure and local details of the scene. This research includes multi-view dense matching, multi-mode point cloud fusion, point cloud meshing and texture mapping, etc.

3D geometric map reconstruction from aerial images (13,000 images, 10 square kilometers)
3D geometric map reconstruction from aerial images (13,000 images, 10 square kilometers)

3D Semantic Segmentation: After obtaining the complete 3D geometric model of the scene, the purpose of 3D semantic segmentation is to obtain the semantic attributes of each geometric primitive (point/facet) in the 3D model, and parse the scene into semantic components with different attributes (buildings, roads, bridges, vegetation, vehicles, etc.) to achieve 3D understanding of large-scale complex scenes. This research mainly explores 3D segmentation networks used for 3D data, segmentation methods based on fusion of 2D images and 3D geometry, cross-domain and cross-scene adaptation of pre-trained segmentation models, etc.

3D semantic segmentation of urban scenes (left: 3D geometric model, right: 3D semantic model)
3D semantic segmentation of urban scenes (left: 3D geometric model, right: 3D semantic model)

3D Vectorized Representation: After decomposing the scene into different semantic parts, the purpose of 3D vectorized representation is to transform a specific semantic component (such as a single building, a single road, a single floor, etc.) into a highly structured, highly compact, and standard-compliant vectorized representation, such as CAD model or BIM model. This research includes the structural decomposition of semantic components, the global regularization of the main structure, the global consistent assembly of structural components, the inference of component topological relations, etc.

3D vectorized modeling of a large garage (left: vehicle images, middle: camera trajectory and 3D sparse map, right: 3D vectorized map)
3D vectorized modeling of a large garage (left: vehicle images, middle: camera trajectory and 3D sparse map, right: 3D vectorized map)

Long-term Global Localization: The reconstructed 3D geometric/semantic/vectorized 3D model could be further used as a global map to provide the ability of global localization for robots entering the scene. The purpose of long-term global localization is to compute the 6-DoF pose of a robot based to the correspondence between the robot real-time data and the 3D map. According to the different sensors used by the robot, the localization algorithm is also different, and the most challenging one is the visual localization problem based on a single image. This research includes all-weather and all-time visual localization, semantic visual localization, multi-source fusion localization, etc.

Monocular visual localization in indoor scene (left: monocular phone image, right: visual localization trajectory)
Monocular visual localization in indoor scene (left: monocular phone image, right: visual localization trajectory)

Incremental Map Updating: The constructed 3D map may change over time, and the purpose of incremental map updating is to discover and update the changed areas in the scene. Since the cost of updating a large-scale 3D map as a whole is very high, it is usually only necessary to perform local incremental updates for the changed areas. This research mainly explores the accurate registration of local and global maps, the change detection method of 3D maps, the reliability measurement of changed regions, and the map incremental update method. The updated 3D map could be returned to the first module again to start a new round of scene data acquisition.

Large-scale scene 3D change detection (left: urban scene, right: open-pit mine)
Large-scale scene 3D change detection (left: urban scene, right: open-pit mine)

The above modules are connected in series to form a complete large-scale 3D scene reconstruction and understanding system, supporting practical applications in digital cities, unmanned systems, digital twin and other fields, such as 3D modeling of urban scenes, 3D HD-map for unmanned vehicles, 3D modeling and visual localization for indoor and outdoor scenes, industrial 3D digital twins, etc. Figure 9 shows several typical applications that we have carried out.

Typical application cases
Typical application cases

During the construction of this system, we continue to improve the capability of each module so that it can accept more challenging inputs and produce more ideal outputs. At the same time, exploring new ideas and strategies, and making substantial changes to one or more of these modules is also our constant pursuit.


Selected Publications:

McSfM: Multi-camera Based Incremental Structure-from-Motion. IEEE Transactions on Image Processing, 32: 6441-6456, 2023.

DOI

PanoVLM: Low-Cost and Accurate Panoramic Vision and LiDAR Fused Mapping. ISPRS Journal of Photogrammetry and Remote Sensing, 206: 149-167, 2023.

Code DOI

Incremental Rotation Averaging. International Journal of Computer Vision, 129: 1202–1216, 2021.

DOI

Dynamic Graph Cuts in Parallel. IEEE Transactions on Image Processing, 26(8): 3775-3788, 2017.

DOI

Dynamic Parallel and Distributed Graph Cuts. IEEE Transactions on Image Processing, 25(12): 5511-5525, 2016.

DOI