In this paper we address the estimation of scene range for multiview 3DTV applications. Based on their similar setup
requirements, we propose a depth camera system combining the strengths of Structured Light and Active Stereo
techniques to obtain a fast depth map recovery for a larger variety of scenes and from a wider view with less occlusion.
The proposed system is composed by a stereo rig and a projector. A stripe-based one shot pattern is projected into the
target scene. Using an efficient decoding technique, reliable correspondences can be found between each camera and the
projected pattern. Decoded areas common to both cameras are triangulated and used to calibrate the projector on the fly
in order to estimate depth in decoded areas only visible in one camera. Correspondences in undecoded areas are
estimated via a stereo matching procedure. The different estimated data is then combined in a single depth map. We
introduce a 3 projection method that improves the structured light identification in strong illumination conditions and
more robust to color objects surfaces of the scene. We demonstrate the efficacy of the integration method experimental
results with special emphasis on its performance in the context of the development and contents creation for depth
image-based representation (DIBR) 3DTV.
In this paper, we present the depth value accuracy requirements for novel view synthesis in 3DTV and our approach for
depth camera based on hybrid method of structured light and active stereo matching. We first review the background on
the use of multiview video plus depth(MVD) representation for future 3DTV system, depth acquisition methods and
tools for real-time and non-real-time dense depth map acquisition in terms of performance and limitations. Then we use a
simple analysis model for novel view synthesis based on DIBR to draw depth value accuracy requirements. Finally we
present our approach on depth camera development along with future directions and huddles in developing real-time
depth camera usable for 3DTV content production.
Display adaptation is one of the key factors for the success of visual services. For two-dimensional (2D) imaging, the
display adaptation is generally accomplished by 2D image re-sampling (i.e., up-/down-sampling). However, when it
comes to stereoscopic three-dimensional (S3D) images, 2D re-sampling methods are inadequate because additional
consideration on the third dimension of depth is not incorporated. In this paper, we analyze S3D image resizing from two
aspects: geometrical deformation and frequency-domain aliasing. A number of S3D displays are available in the market
and they have various screen dimensions. As we have more varieties of the displays, efficient S3D image resizing is
becoming more emphasized. We present the conditions for equi-proportional S3D image resizing using the model of the
binocular vision; and propose an anti-aliasing filter for stereoscopic 3D image up/down-sampling.
KEYWORDS: Visualization, Panoramic photography, 3D visualizations, Video, Binary data, 3D image processing, 3D modeling, 3D displays, 3D metrology, Cameras
This paper describes issues and consideration on authoring of 3D visual content based on MPEG-4 Systems. The issues include types of 3D visual content; functionalities for user-interaction; 3D scene composition for rendering; and the 3D visual content file format. MPEG-4 includes several extensions compared to MPEG-1 and -2 especially in terms of object-based interactivity. In addition to advanced A/V compression methods (MPEG-4 Part2, Part3 and Part 10), MPEG-4 Part1, which is named "Systems", contains additional tools such as binary/textual file format, multiplexing, synchronization scheme, scene description, etc. The use of MPEG-4 can resolve the problem of format diversity while providing high interactivity to users. There has been little investigation on 3D visual content authoring. Throughout the paper, we will present which issues need to be determined and how currently available tools can be effectively utilized for 3D visual content creation.
A new algorithm for the rate control of videos considering the sensitivity of the human visual system (HVS) is presented. The method adopts the three-step structure of MPEG-2 Test Model 5 (TM5) rate control, while a new measure for the macroblock (MB) activity based on spatiotemporal sensitivity is introduced. Experimental results show that the proposed activity measure outperforms the spatial activity of TM5 in picture quality.
In this work, we present a novel method for bit allocation problem that aims to minimize overall distortion subject to bit rate constraint. It has been proved that optimal solution can be found by a method using the Lagrangian method with dynamic programming. However, the optimal bit allocation for block-based interframe coding is practically unattainable because of interframe dependency of macroblocks caused by motion compensation. In order to reduce the computational burden maintaining the result close to optimal, we propose an alternative method. We derive a partitioned form of the bit allocation problem: a frame-level problem and one-frame macroblock-level problems. Then we use a two-phase optimization technique with an interframe dependency model and a rate-distortion model.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.