• Skip navigation
  • Skip to navigation
  • Skip to the bottom
Simulate organization breadcrumb open Simulate organization breadcrumb close
Chair of Visual Computing
  • FAUTo the central FAU website
  1. Friedrich-Alexander-Universität
  2. Technische Fakultät
  3. Department Informatik
Suche öffnen
  • Campo
  • StudOn
  • FAUdir
  • Jobs
  • Map
  • Help
  1. Friedrich-Alexander-Universität
  2. Technische Fakultät
  3. Department Informatik

Chair of Visual Computing

Navigation Navigation close
  • Research
    • Rendering and Visualization
    • Geometric Modeling and 3D Reconstruction
    • Virtual, Mixed, and Augmented Reality
    • Visual Computing for Digital Humanities and Social Sciences
    • Visual Healthcare Computing
    Research
  • Publications
  • Teaching
    • Vertiefungsrichtung Visual Computing
    • Summer Term 2025
    • Winter Term 2024/25
    • Theses
    Teaching
  • Staff
  • Arrival and Contact
  1. Home
  2. Publications
  3. VolumeDeform: Real-time Volumetric Non-rigid Reconstruction

VolumeDeform: Real-time Volumetric Non-rigid Reconstruction

In page navigation: Publications
  • Adaptive stray-light compensation in dynamic multi-projection mapping
  • Adaptive Temporal Sampling for Volumetric Path Tracing of Medical Data
  • Analytic Displacement Mapping using Hardware Tessellation
  • Anisotropic Surface Based Deformation
  • Auto-Calibration for Dynamic Multi-Projection Mapping on Arbitrary Surfaces
  • Automated Heart Localization in Cardiac Cine MR Data
  • Demo of Face2Face: Real-time Face Capture and Reenactment of RGB Videos
  • Enhanced Sphere Tracing
  • Evaluating the Usability of Recent Consumer-Grade 3D Input Devices
  • Face2Face: Real-time Face Capture and Reenactment of RGB Videos
  • FaceForge: Markerless Non-Rigid Face Multi-Projection Mapping
  • FaceInCar: Real-time Dense Monocular Face Tracking of a Driver
  • FaceVR: Real-Time Facial Reenactment and Eye Gaze Control in Virtual Reality
  • GroPBS: Fast Solver for Implicit Electrostatics of Biomolecules
  • Grundsätzliche Überlegungen zur Edition des Bestandes an Münzen der FAU als frei zugängliche Datenbank im WWW
  • HeadOn: Real-time Reenactment of Human Portrait Videos
  • Hierarchical Multi-Layer Screen-Space Ray Tracing
  • Hybrid Mono-Stereo Rendering in Virtual Reality
  • Interactive Model-based Reconstruction of the Human Head using an RGB-D Sensor
  • Interactive Painting and Lighting in Dynamic Multi-Projection Mapping
  • Learning Real-Time Ambient Occlusion from Distance Representations
  • Low-Cost Real-Time 3D Reconstruction of Large-Scale Excavation Sites using an RGB-D Camera
  • Multi-Layer Depth of Field Rendering with Tiled Splatting
  • Multi-Resolution Attributes for Hardware Tessellated Objects
  • Real-time 3D Reconstruction at Scale using Voxel Hashing
  • Real-time Collision Detection for Dynamic Hardware Tessellated Objects
  • Real-time Expression Transfer for Facial Reenactment
  • Real-time Local Displacement using Dynamic GPU Memory Management
  • Real-Time Pixel Luminance Optimization for Dynamic Multi-Projection Mapping
  • Reality Forge: Interactive Dynamic Multi-Projection Mapping
  • Robust Blending and Occlusion Compensation in Dynamic Multi-Projection Mapping
  • Shape Adaptive Cut Lines
  • Spherical Fibonacci Mapping
  • State of the Art Report on Real-time Rendering with Hardware Tessellation
  • Stray-Light Compensation in Dynamic Projection Mapping
  • Visualization and Deformation Techniques for Entertainment and Training in Cultural Heritage
  • VolumeDeform: Real-time Volumetric Non-rigid Reconstruction

VolumeDeform: Real-time Volumetric Non-rigid Reconstruction

Matthias Innmann

Dr.-Ing. Matthias Innmann

Department of Computer Science
Chair of Computer Science 9 (Computer Graphics)

Cauerstraße 11
91058 Erlangen
  • Email: matthias.innmann@fau.de
  • Website: http://lgdv.cs.fau.de/people/card/matthias/innmann/

Paper

  • Innmann M., Zollhöfer M., Nießner M., Theobald C., Stamminger M.:
    VolumeDeform: Real-Time Volumetric Non-rigid Reconstruction
    14th European Conference on Computer Vision (ECCV) (Amsterdam, October 8, 2016 - October 16, 2016)
    In: Springer International Publishing (ed.): ECCV 2016 - Proceedings of the European Conference on Computer Vision, Cham: 2016
    DOI: 10.1007/978-3-319-46484-8_22
    URL: https://www.lgdv.tf.fau.de/?p=425
    BibTeX: Download

Abstract

We present a novel approach for the reconstruction of dynamic geometric shapes using a single hand-held consumer-grade RGB-D sensor at real-time rates. Our method builds up the scene model from scratch during the scanning process, thus it does not require a pre-defined shape template to start with. Geometry and motion are parameterized in a unified manner by a volumetric representation that encodes a distance field of the surface geometry as well as the non-rigid space deformation. Motion tracking is based on a set of extracted sparse color features in combination with a dense depth constraint. This enables accurate tracking and drastically reduces drift inherent to standard model-to-depth alignment. We cast finding the optimal deformation of space as a non-linear regularized variational optimization problem by enforcing local smoothness and proximity to the input constraints. The problem is tackled in real-time at the camera’s capture rate using a data-parallel flip-flop optimization strategy. Our results demonstrate robust tracking even for fast motion and scenes that lack geometric features.

Video

Display external content

At this point content of an external provider (source: YouTube) is integrated. When displaying, data may be transferred to third parties or cookies may be stored, therefore your consent is required.

You can find more information and the possibility to revoke your consent in our privacy policy.

I agree

Downloads

Supplemental

Dataset

We provide a dataset containing RGB-D data of a variety of objects, for the purpose of real-time non-rigid reconstruction. The RGB-D data contains sequences taken from a PrimeSense sensor (color and depth images). Additionally, we provide meshes extracted for several frame (live reconstruction and canonical model).
Please refer to this publication when using the dataset.

Format

For each scene, we provide a zip file containing a sequence of RGB-D camera frames (X_data.zip). Each sequence contains:

Color frames (frame-XXXXXX.color.png): RGB, 24-bit, PNG
Depth frames (frame-XXXXXX.depth.png): depth (mm), 16-bit, PNG (invalid depth is set to 0)

Camera Calibration:

The color and depth camera intrinsics for each sequence are provided in colorIntrinsics.txt and depthIntrinsics.txt. Note that these are the default values provided and we did not perform any calibration.

Meshes:

The extracted meshes are contained in X_canonical.zip (frame-XXXXXX.canonical.ply) and X_reconstruction.zip (frame-XXXXXX.mesh.ply) for every 100th frame and the last frame. The transformations from worldspace to cameraspace for these frames are also provided in X_reconstruction.zip (frame-XXXXXX.world-to-camera.txt).

License

The data has been released under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 License.

Chair of Visual Computing
(Lehrstuhl für Graphische Datenverarbeitung)

Cauerstraße 11
91058 Erlangen
Deutschland
  • Imprint
  • Privacy
  • Facebook
  • RSS Feed
  • Xing
Up