liu.seSearch for publications in DiVA
Change search
ReferencesLink to record
Permanent link

Direct link
Depth Data Processing and 3D Reconstruction Using the Kinect v2
Linköping University, Department of Electrical Engineering, Computer Vision. Linköping University, Faculty of Science & Engineering. (ISY)
2015 (English)Independent thesis Advanced level (degree of Master (Two Years)), 20 credits / 30 HE creditsStudent thesis
Abstract [en]

The Kinect v2 is a RGB-D sensor manufactured as a gesture interaction tool for the entertainment console XBOX One. In this thesis we will use it to perform 3D reconstruction and investigate its ability to measure depth. In order to sense both color and depth the Kinect v2 has two cameras: one RGB camera and one infrared camera used to produce depth and near infrared images. These cameras need to be calibrated if we want to use them for 3D reconstruction. We present a calibration procedure for simultaneously calibrating the cameras and extracting their relative pose. This enables us to construct colored meshes of the environment. When we know the camera parameters of the infrared camera, the depth images could be used to perform the Kinectfusion algorithm. This produces well-formed meshes of the environment by combining many depth frames taken from several camera poses.The Kinect v2 uses a time-of-flight technology were the phase shifts are extracted from amplitude modulated infrared light signals produced by an emitter. The extracted phase shifts are then converted to depth values. However, the extraction of phase shifts includes a phase unwrapping procedure, which is sensitive to noise and can result in large depth errors.By utilizing the ability to access the raw phase measurements from the device we managed to modify the phase unwrapping procedure. This new procedure includes an extraction of several hypotheses for the unwrapped phase and a spatial propagation to select amongst them. This proposed method has been compared with the available drivers in the open source library libfreenect2 and the Microsoft Kinect SDK v2. Our experiments show that the depth images of the two available drivers have similar quality and our proposed method improves over libfreenect2. The calculations in the proposed method are more expensive than those in libfreenect2 but it still runs at 2.5× real time. However, contrary to libfreenect2 the proposed method lacks a filter that removes outliers from the depth images. It turned out that this is an important feature when performing Kinect fusion and future work should thus be focused on adding an outlier filter.

Place, publisher, year, edition, pages
2015. , 71 p.
Keyword [en]
RGB-D, time-of-flight, 3D reconstruction, Kinect v2
National Category
Engineering and Technology Signal Processing
URN: urn:nbn:se:liu:diva-121401ISRN: LiTH-ISY-EX–15/4884–SEOAI: diva2:854680
Subject / course
Computer Vision Laboratory
Available from: 2015-09-17 Created: 2015-09-17 Last updated: 2015-09-18Bibliographically approved

Open Access in DiVA

fulltext(42746 kB)427 downloads
File information
File name FULLTEXT01.pdfFile size 42746 kBChecksum SHA-512
Type fulltextMimetype application/pdf

Search in DiVA

By author/editor
Järemo Lawin, Felix
By organisation
Computer VisionFaculty of Science & Engineering
Engineering and TechnologySignal Processing

Search outside of DiVA

GoogleGoogle Scholar
Total: 427 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

Total: 1069 hits
ReferencesLink to record
Permanent link

Direct link