Change search
ReferencesLink to record
Permanent link

Direct link
Depth Sensing for 3DTV: A Survey
Mid Sweden University, Faculty of Science, Technology and Media, Department of Information and Communication systems. (Realistic3D)ORCID iD: 0000-0002-2578-7896
Mid Sweden University, Faculty of Science, Technology and Media, Department of Information and Communication systems. (Realistic3D)
Mid Sweden University, Faculty of Science, Technology and Media, Department of Information and Communication systems. (Realistic3D)
2013 (English)In: IEEE Multimedia, ISSN 1070-986X, E-ISSN 1941-0166, Vol. 20, no 4, 10-17 p.Article in journal (Refereed) Published
Abstract [en]

In the context of 3D video systems, depth information could be used to render a scene from additional viewpoints. Although there have been many recent advances in this area, including the introduction of the Microsoft Kinect sensor, the robust acquisition of such information continues to be a challenge. This article reviews three depth-sensing approaches for 3DTV. The authors discuss several approaches for acquiring depth information and provides a comparative analysis of their characteristics.

Place, publisher, year, edition, pages
IEEE Computer Society, 2013. Vol. 20, no 4, 10-17 p.
Keyword [en]
3D video, scene acquisition, capture, depth sensing, stereo analysis, structured lighting, time-of-flight, sensor fusion
National Category
Signal Processing
URN: urn:nbn:se:miun:diva-20416DOI: 10.1109/MMUL.2013.53ISI: 000327723900007ScopusID: 2-s2.0-84890069117Local ID: STCOAI: diva2:669203
Knowledge Foundation, 2009/0264

This work has been supportedby grant 2009/0264 of the KK Foundation, Sweden; grant 00156702 of the EU European Regional Development Fund,Mellersta Norrland, Sweden; and grant 00155148 ofLänsstyrelsenVästernorrland, Sweden.

Available from: 2013-12-03 Created: 2013-12-03 Last updated: 2016-10-20Bibliographically approved
In thesis
1. Gaining Depth: Time-of-Flight Sensor Fusion for Three-Dimensional Video Content Creation
Open this publication in new window or tab >>Gaining Depth: Time-of-Flight Sensor Fusion for Three-Dimensional Video Content Creation
2014 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

The successful revival of three-dimensional (3D) cinema has generated a great deal of interest in 3D video. However, contemporary eyewear-assisted displaying technologies are not well suited for the less restricted scenarios outside movie theaters. The next generation of 3D displays, autostereoscopic multiview displays, overcome the restrictions of traditional stereoscopic 3D and can provide an important boost for 3D television (3DTV). Then again, such displays require scene depth information in order to reduce the amount of necessary input data. Acquiring this information is quite complex and challenging, thus restricting content creators and limiting the amount of available 3D video content. Nonetheless, without broad and innovative 3D television programs, even next-generation 3DTV will lack customer appeal. Therefore simplified 3D video content generation is essential for the medium's success.

This dissertation surveys the advantages and limitations of contemporary 3D video acquisition. Based on these findings, a combination of dedicated depth sensors, so-called Time-of-Flight (ToF) cameras, and video cameras, is investigated with the aim of simplifying 3D video content generation. The concept of Time-of-Flight sensor fusion is analyzed in order to identify suitable courses of action for high quality 3D video acquisition. In order to overcome the main drawback of current Time-of-Flight technology, namely the high sensor noise and low spatial resolution, a weighted optimization approach for Time-of-Flight super-resolution is proposed. This approach incorporates video texture, measurement noise and temporal information for high quality 3D video acquisition from a single video plus Time-of-Flight camera combination. Objective evaluations show benefits with respect to state-of-the-art depth upsampling solutions. Subjective visual quality assessment confirms the objective results, with a significant increase in viewer preference by a factor of four. Furthermore, the presented super-resolution approach can be applied to other applications, such as depth video compression, providing bit rate savings of approximately 10 percent compared to competing depth upsampling solutions. The work presented in this dissertation has been published in two scientific journals and five peer-reviewed conference proceedings. 

In conclusion, Time-of-Flight sensor fusion can help to simplify 3D video content generation, consequently supporting a larger variety of available content. Thus, this dissertation provides important inputs towards broad and innovative 3D video content, hopefully contributing to the future success of next-generation 3DTV.

Place, publisher, year, edition, pages
Sundsvall: Mittuniversitetet, 2014. 228 p.
Mid Sweden University doctoral thesis, ISSN 1652-893X ; 185
3D video, Time-of-Flight, depth map acquisition, optimization, 3DTV, ToF, upsampling, super-resolution, sensor fusion
National Category
Computer Systems
urn:nbn:se:miun:diva-21938 (URN)978-91-87557-49-1 (ISBN)
Public defence
2014-06-04, L111, Holmgatan 10, Sundsvall, 10:00 (English)
Knowledge Foundation, 2009/0264
Available from: 2014-05-16 Created: 2014-05-14 Last updated: 2014-05-16Bibliographically approved

Open Access in DiVA

fulltext(1597 kB)168 downloads
File information
File name FULLTEXT02.pdfFile size 1597 kBChecksum SHA-512
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Search in DiVA

By author/editor
Schwarz, SebastianOlsson, RogerSjöström, Mårten
By organisation
Department of Information and Communication systems
In the same journal
IEEE Multimedia
Signal Processing

Search outside of DiVA

GoogleGoogle Scholar
Total: 168 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

Altmetric score

Total: 781 hits
ReferencesLink to record
Permanent link

Direct link