An efficient stereo matching algorithm for computing stable disparity map sequence from video footage is presented.
The algorithm is based on both the spatial and temporal consistency in the stereo sequences, and high quality disparity
maps are achieved. Weber local descriptors (WLD) are extracted for each color channel from current stereo pairs, and
the raw matching costs between the images are initialized by WLD. Orthogonal integral image (OII) technique along
with minimum spanning tree (MST) is used to aggregate the similar pixels and preserve disparity edges adaptively. MST
takes place of the process of voting support regions in OII technique and provides a specific support region for each pixel.
The nodes of MST are all the image pixels, and the weight of edges are absolute difference between the nearest
neighboring pixels. It’s a global method, and can achieve more accurate disparity maps than traditional OII technique.
Three-frame subtraction is used to determine the temporal consistency between adjacent frames. The motion region is
extracted and the disparity map of motive region is renewed. The disparity of current frame with the renewed disparity
and the one for last frame is confirmed. The proposed approach has been tested on the real stereo sequences, and the
results are satisfactory.
An arbitrary view synthesis method from 2D-Plus-Depth image for real-time auto-stereoscopic display is presented.
Traditional methods use depth image based rendering (DIBR) technology, which is a process of synthesizing “virtual”
views of a scene from still or moving images and associated per-pixel depth information. All the virtual view images are
generated and then the ultimate stereo-image is synthesized. DIBR can greatly decrease the number of reference images
and is flexible and efficient as the depth images are used. However it causes some problems such as the appearance of
holes in the rendered image, and the occurrence of depth discontinuity on the surface of the object at virtual image plane.
Here, reversed disparity shift pixel rendering is used to generate the stereo-image directly, and the target image won’t
generate holes. To avoid duplication of calculation and also to be able to match with any specific three-dimensional
display, a selecting table is designed to pick up appropriate virtual viewpoints for auto-stereoscopic display. According to
the selecting table, only sub-pixels of the appropriate virtual viewpoints are calculated, so calculation amount is
independent of the number of virtual viewpoints. In addition, 3D image warping technology is used to translate depth
information to parallax between virtual viewpoints and parallax, and the viewer can adjust the
zero-parallax-setting-plane (ZPS) and change parallax conveniently to suit his/her personal preferences. The proposed
method is implemented with OPENGL and demonstrated on a laptop computer with a 2.3 GHz Intel Core i5 CPU and
NVIDA GeForce GT540m GPU. We got a frame rate 30 frames per second with 4096×2340 video. High synthesis
efficiency and good stereoscopic sense can be obtained. The presented method can meet the requirements of
real-time ultra-HD super multi-view auto-stereoscopic display.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.