Muutke küpsiste eelistusi

E-raamat: Emerging Technologies for 3D Video - Creation, Coding, Transmission and Rendering: Creation, Coding, Transmission and Rendering [Wiley Online]

  • Formaat: 520 pages
  • Ilmumisaeg: 17-May-2013
  • Kirjastus: John Wiley & Sons Inc
  • ISBN-10: 1118583590
  • ISBN-13: 9781118583593
Teised raamatud teemal:
  • Wiley Online
  • Hind: 153,31 €*
  • * hind, mis tagab piiramatu üheaegsete kasutajate arvuga ligipääsu piiramatuks ajaks
  • Formaat: 520 pages
  • Ilmumisaeg: 17-May-2013
  • Kirjastus: John Wiley & Sons Inc
  • ISBN-10: 1118583590
  • ISBN-13: 9781118583593
Teised raamatud teemal:

With the expectation of greatly enhanced user experience, 3D video is widely perceived as the next major advancement in video technology. In order to fulfil the expectation of enhanced user experience, 3D video calls for new technologies addressing efficient content creation, representation/coding, transmission and display.

Emerging Technologies for 3D Video will deal with all aspects involved in 3D video systems and services, including content acquisition and creation, data representation and coding, transmission, view synthesis, rendering, display technologies, human perception of depth and quality assessment.

Key features:

  • Offers an overview of key existing technologies for 3D video
  • Provides a discussion of advanced research topics and future technologies
  • Reviews relevant standardization efforts
  • Addresses applications and implementation issues
  • Includes contributions from leading researchers

The book is a comprehensive guide to 3D video systems and services suitable for all those involved in this field, including engineers, practitioners, researchers as well as professors, graduate and undergraduate students, and managers making technological decisions about 3D video.

Preface xvii
List of Contributors
xxi
Acknowledgements xxv
PART I CONTENT CREATION
1 Consumer Depth Cameras and Applications
3(14)
Seungkyu Lee
1.1 Introduction
3(1)
1.2 Time-of-Flight Depth Camera
3(8)
1.2.1 Principle
4(2)
1.2.2 Quality of the Measured Distance
6(5)
1.3 Structured Light Depth Camera
11(1)
1.3.1 Principle
11(1)
1.4 Specular and Transparent Depth
12(3)
1.5 Depth Camera Applications
15(2)
1.5.1 Interaction
15(1)
1.5.2 Three-Dimensional Reconstruction
15(1)
References
16(1)
2 SFTI: Space-from-Time Imaging
17(20)
Ahmed Kirmani
Andrea Colaco
Vivek K. Goyal
2.1 Introduction
17(1)
2.2 Background and Related Work
18(3)
2.2.1 Light Fields, Reflectance Distribution Functions, and Optical Image Formation
18(2)
2.2.2 Time-of-Flight Methods for Estimating Scene Structure
20(1)
2.2.3 Synthetic Aperture Radar for Estimating Scene Reflectance
20(1)
2.3 Sampled Response of One Source-Sensor Pair
21(3)
2.3.1 Scene, Illumination, and Sensor Abstractions
21(1)
2.3.2 Scene Response Derivation
22(2)
2.3.3 Inversion
24(1)
2.4 Diffuse Imaging: SFTI for Estimating Scene Reflectance
24(6)
2.4.1 Response Modeling
24(4)
2.4.2 Image Recovery using Linear Backprojection
28(2)
2.5 Compressive Depth Acquisition: SFTI for Estimating Scene Structure
30(4)
2.5.1 Single-Plane Response to Omnidirectional Illumination
30(2)
2.5.2 Spatially-Patterned Measurement
32(1)
2.5.3 Algorithms for Depth Map Reconstruction
33(1)
2.6 Discussion and Future Work
34(3)
Acknowledgments
35(1)
References
35(2)
3 2D-to-3D Video Conversion: Overview and Perspectives
37(25)
Carlos Vazquez
Liang Zhang
Filippo Speranza
Nils Plath
Sebastian Knorr
3.1 Introduction
37(1)
3.2 The 2D-to-3D Conversion Problem
38(3)
3.2.1 General Conversion Approach
38(1)
3.2.2 Depth Cues in Monoscopic Video
39(2)
3.3 Definition of Depth Structure of the Scene
41(7)
3.3.1 Depth Creation Methods
42(2)
3.3.2 Depth Recovery Methods
44(4)
3.4 Generation of the Second Video Stream
48(8)
3.4.1 Depth to Disparity Mapping
48(1)
3.4.2 View Synthesis and Rendering Techniques
49(4)
3.4.3 Post-Processing for Hole-Filling
53(3)
3.5 Quality of Experience of 2D-to-3D Conversion
56(1)
3.6 Conclusions
57(5)
References
58(4)
4 Spatial Plasticity: Dual-Camera Configurations and Variable Interaxial
62(19)
Ray Zone
4.1 Stereoscopic Capture
62(1)
4.2 Dual-Camera Arrangements in the 1950s
63(2)
4.3 Classic "Beam-Splitter" Technology
65(1)
4.4 The Dual-Camera Form Factor and Camera Mobility
66(2)
4.5 Reduced 3D Form Factor of the Digital CCD Sensor
68(3)
4.6 Handheld Shooting with Variable Interaxial
71(2)
4.7 Single-Body Camera Solutions for Stereoscopic Cinematography
73(3)
4.8 A Modular 3D Rig
76(1)
4.9 Human Factors of Variable Interaxial
76(5)
References
78(3)
PART II REPRESENTATION, CODING AND TRANSMISSION
5 Disparity Estimation Techniques
81(21)
Mounir Kaaniche
Raffaele Gaetano
Marco Cagnazzo
Beatrice Pesquet-Popescu
5.1 Introduction
81(1)
5.2 Geometrical Models for Stereoscopic Imaging
82(6)
5.2.1 The Pinhole Camera Model
82(3)
5.2.2 Stereoscopic Imaging Systems
85(3)
5.3 Stereo Matching Process
88(3)
5.3.1 Disparity Information
88(1)
5.3.2 Difficulties in the Stereo Matching Process
88(1)
5.3.3 Stereo Matching Constraints
89(1)
5.3.4 Fundamental Steps Involved in Stereo Matching Algorithms
89(2)
5.4 Overview of Disparity Estimation Methods
91(7)
5.4.1 Local Methods
91(2)
5.4.2 Global Methods
93(5)
5.5 Conclusion
98(4)
References
98(4)
6 3D Video Representation and Formats
102(19)
Marco Cagnazzo
Beatrice Pesquet-Popescu
Frederic Dufaux
6.1 Introduction
102(1)
6.2 Three-Dimensional Video Representation
103(6)
6.2.1 Stereoscopic 3D (S3D) Video
103(1)
6.2.2 Multiview Video (MVV)
104(1)
6.2.3 Video-Plus-Depth
105(2)
6.2.4 Multiview Video-Plus-Depth (MVD)
107(1)
6.2.5 Layered Depth Video (LDV)
107(2)
6.3 Three-Dimensional Video Formats
109(9)
6.3.1 Simulcast
109(1)
6.3.2 Frame-Compatible Stereo Interleaving
110(3)
6.3.3 MPEG-4 Multiple Auxiliary Components (MAC)
113(1)
6.3.4 MPEG-C Part 3
113(1)
6.3.5 MPEG-2 Multiview Profile (MVP)
113(1)
6.3.6 Multiview Video Coding (MVC)
114(4)
6.4 Perspectives
118(3)
Acknowledgments
118(1)
References
119(2)
7 Depth Video Coding Technologies
121(18)
Elie Gabriel Mora
Giuseppe Valenzise
Joel Jung
Beatrice Pesquet-Popescu
Marco Cagnazzo
Frederic Dufaux
7.1 Introduction
121(1)
7.2 Depth Map Analysis and Characteristics
122(1)
7.3 Depth Map Coding Tools
123(9)
7.3.1 Tools that Exploit the Inherent Characteristics of Depth Maps
123(4)
7.3.2 Tools that Exploit the Correlations with the Associated Texture
127(2)
7.3.3 Tools that Optimize Depth Map Coding for the Quality of the Synthesis
129(3)
7.4 Application Example: Depth Map Coding Using "Don't Care" Regions
132(4)
7.4.1 Derivation of "Don't Care" Regions
133(1)
7.4.2 Transform Domain Sparsification Using "Don't Care" Regions
134(1)
7.4.3 Using "Don't Care" Regions in a Hybrid Video Codec
135(1)
7.5 Concluding Remarks
136(3)
Acknowledgments
137(1)
References
137(2)
8 Depth-Based 3D Video Formats and Coding Technology
139(23)
Anthony Vetro
Karsten Muller
8.1 Introduction
139(2)
8.1.1 Existing Stereo/Multiview Formats
140(1)
8.1.2 Requirements for Depth-Based Format
140(1)
8.1.3
Chapter Organization
141(1)
8.2 Depth Representation and Rendering
141(3)
8.2.1 Depth Format and Representation
142(1)
8.2.2 Depth-Image-Based Rendering
143(1)
8.3 Coding Architectures
144(3)
8.3.1 AVC-Based Architecture
144(2)
8.3.2 HEVC-Based Architecture
146(1)
8.3.3 Hybrid
146(1)
8.4 Compression Technology
147(6)
8.4.1 Inter-View Prediction
148(1)
8.4.2 View Synthesis Prediction
148(1)
8.4.3 Depth Resampling and Filtering
149(1)
8.4.4 Inter-Component Parameter Prediction
150(1)
8.4.5 Depth Modelling
151(1)
8.4.6 Bit Allocation
152(1)
8.5 Experimental Evaluation
153(5)
8.5.1 Evaluation Framework
153(2)
8.5.2 AVC-Based 3DV Coding Results
155(1)
8.5.3 HEVC-Based 3DV Coding Results
156(2)
8.5.4 General Observations
158(1)
8.6 Concluding Remarks
158(4)
References
159(3)
9 Coding for Interactive Navigation in High-Dimensional Media Data
162(25)
Ngai-Man Cheung
Gene Cheung
9.1 Introduction
162(1)
9.2 Challenges and Approaches of Interactive Media Streaming
163(3)
9.2.1 Challenges: Coding Efficiency and Navigation Flexibility
163(2)
9.2.2 Approaches to Interactive Media Streaming
165(1)
9.3 Example Solutions
166(6)
9.3.1 Region-of-Interest (RoI) Image Browsing
166(1)
9.3.2 Light-Field Streaming
167(1)
9.3.3 Volumetric Image Random Access
168(1)
9.3.4 Video Browsing
168(1)
9.3.5 Reversible Video Playback
169(1)
9.3.6 Region-of-Interest (RoI) Video Streaming
169(3)
9.4 Interactive Multiview Video Streaming
172(12)
9.4.1 Interactive Multiview Video Streaming (IMVS)
172(7)
9.4.2 IMVS with Free Viewpoint Navigation
179(2)
9.4.3 IMVS with Fixed Round-Trip Delay
181(3)
9.5 Conclusion
184(3)
References
184(3)
10 Adaptive Streaming of Multiview Video Over P2P Networks
187(20)
C. Goktug Gurler
A. Murat Tekalp
10.1 Introduction
187(1)
10.2 P2P Overlay Networks
188(4)
10.2.1 Overlay Topology
188(1)
10.2.2 Sender-Driven versus Receiver-Driven P2P Video Streaming
189(1)
10.2.3 Layered versus Cross-Layer Architecture
190(1)
10.2.4 When P2P is Useful: Regions of Operation
191(1)
10.2.5 BitTorrent: A Platform for File Sharing
191(1)
10.3 Monocular Video Streaming Over P2P Networks
192(5)
10.3.1 Video Coding
193(1)
10.3.2 Variable-Size Chunk Generation
193(1)
10.3.3 Time-Sensitive Chunk Scheduling Using Windowing
194(1)
10.3.4 Buffer-Driven Rate Adaptation
195(1)
10.3.5 Adaptive Window Size and Scheduling Restrictions
195(1)
10.3.6 Multiple Requests from Multiple Peers of a Single Chunk
196(1)
10.4 Stereoscopic Video Streaming over P2P Networks
197(4)
10.4.1 Stereoscopic Video over Digital TV
197(1)
10.4.2 Rate Adaptation in Stereo Streaming: Asymmetric Coding
197(3)
10.4.3 Use Cases: Stereoscopic Video Streaming over P2P Network
200(1)
10.5 MVV Streaming over P2P Networks
201(6)
10.5.1 MVV Streaming over IP
201(1)
10.5.2 Rate Adaptation for MVV: View Scaling
201(1)
10.5.3 Use Cases: MVV Streaming over P2P Network
202(1)
References
203(4)
PART III RENDERING AND SYNTHESIS
11 Image Domain Warping for Stereoscopic 3D Applications
207(24)
Oliver Wang
Manuel Lang
Nikolce Stefanoski
Alexander Sorkine-Hornung
Olga Sorkine-Hornung
Aljoscha Smolic
Markus Gross
11.1 Introduction
207(1)
11.2 Background
208(1)
11.3 Image Domain Warping
209(1)
11.4 Stereo Mapping
210(3)
11.4.1 Problems in Stereoscopic Viewing
210(1)
11.4.2 Disparity Range
210(1)
11.4.3 Disparity Sensitivity
211(1)
11.4.4 Disparity Velocity
211(1)
11.4.5 Summary
212(1)
11.4.6 Disparity Mapping Operators
212(1)
11.4.7 Linear Operator
212(1)
11.4.8 Nonlinear Operator
212(1)
11.4.9 Temporal Operator
213(1)
11.5 Warp-Based Disparity Mapping
213(5)
11.5.1 Data Extraction
213(1)
11.5.2 Warp Calculation
214(2)
11.5.3 Applications
216(2)
11.6 Automatic Stereo to Multiview Conversion
218(3)
11.6.1 Automatic Stereo to Multiview Conversion
218(1)
11.6.2 Position Constraints
219(1)
11.6.3 Warp Interpolation and Extrapolation
219(1)
11.6.4 Three-Dimensional Video Transmission Systems for Multiview Displays
220(1)
11.7 IDW for User-Driven 2D-3D Conversion
221(4)
11.7.1 Technical Challenges of 2D-3D Conversion
222(3)
11.8 Multi-Perspective Stereoscopy from Light Fields
225(3)
11.9 Conclusions and Outlook
228(3)
Acknowledgments
229(1)
References
229(2)
12 Image-Based Rendering and the Sampling of the Plenoptic Function
231(18)
Christopher Gilliam
Mike Brookes
Pier Luigi Dragotti
12.1 Introduction
231(1)
12.2 Parameterization of the Plenoptic Function
232(3)
12.2.1 Light Field and Surface Light Field Parameterization
232(2)
12.2.2 Epipolar Plane Image
234(1)
12.3 Uniform Sampling in a Fourier Framework
235(7)
12.3.1 Spectral Analysis of the Plenoptic Function
236(3)
12.3.2 The Plenoptic Spectrum under Realistic Conditions
239(3)
12.4 Adaptive Plenoptic Sampling
242(4)
12.4.1 Adaptive Sampling Based on Plenoptic Spectral Analysis
244(2)
12.5 Summary
246(3)
12.5.1 Outlook
246(1)
References
247(2)
13 A Framework for Image-Based Stereoscopic View Synthesis from Asynchronous Multiview Data
249(26)
Felix Klose
Christian Lipski
Marcus Magnor
13.1 The Virtual Video Camera
249(9)
13.1.1 Navigation Space Embedding
251(1)
13.1.2 Space-Time Tetrahedralization
252(3)
13.1.3 Processing Pipeline
255(1)
13.1.4 Rendering
256(1)
13.1.5 Application
257(1)
13.1.6 Limitations
258(1)
13.2 Estimating Dense Image Correspondences
258(6)
13.2.1 Belief Propagation for Image Correspondences
259(1)
13.2.2 A Symmetric Extension
260(1)
13.2.3 SIFT Descriptor Downsampling
261(1)
13.2.4 Construction of Message-Passing Graph
261(1)
13.2.5 Data Term Compression
262(1)
13.2.6 Occlusion Removal
263(1)
13.2.7 Upsampling and Refinement
263(1)
13.2.8 Limitations
263(1)
13.3 High-Quality Correspondence Edit
264(1)
13.3.1 Editing Operations
264(1)
13.3.2 Applications
265(1)
13.4 Extending to the Third Dimension
265(10)
13.4.1 Direct Stereoscopic Virtual View Synthesis
266(1)
13.4.2 Depth-Image-Based Rendering
267(1)
13.4.3 Comparison
267(1)
13.4.4 Concluding with the "Who Cares?" Post-Production Pipeline
268(2)
References
270(5)
PART IV DISPLAY TECHNOLOGIES
14 Signal Processing for 3D Displays
275(20)
Janusz Konrad
14.1 Introduction
275(1)
14.2 3D Content Generation
276(11)
14.2.1 Automatic 2D-to-3D Image Conversion
276(4)
14.2.2 Real-Time Intermediate View Interpolation
280(6)
14.2.3 Brightness and Color Balancing in Stereopairs
286(1)
14.3 Dealing with 3D Display Hardware
287(5)
14.3.1 Ghosting Suppression for Polarized and Shuttered Stereoscopic 3D Displays
287(2)
14.3.2 Aliasing Suppression for Multiview Eyewear-Free 3D Displays
289(3)
14.4 Conclusions
292(3)
Acknowledgments
293(1)
References
293(2)
15 3D Display Technologies
295(18)
Thierry Borel
Didier Doyen
15.1 Introduction
295(1)
15.2 Three-Dimensional Display Technologies in Cinemas
295(6)
15.2.1 Three-Dimensional Cinema Projectors Based on Light Polarization
296(3)
15.2.2 Three-Dimensional Cinema Projectors Based on Shutters
299(1)
15.2.3 Three-Dimensional Cinema Projectors Based on Interference Filters
300(1)
15.3 Large 3D Display Technologies in the Home
301(8)
15.3.1 Based on Anaglyph Glasses
301(1)
15.3.2 Based on Shutter Glasses
302(2)
15.3.3 Based on Polarized Glasses
304(2)
15.3.4 Without Glasses
306(3)
15.4 Mobile 3D Display Technologies
309(2)
15.4.1 Based on Parallax Barriers
310(1)
15.4.2 Based on Lighting Switch
310(1)
15.5 Long-Term Perspectives
311(1)
15.6 Conclusion
312(1)
References
312(1)
16 Integral Imaging
313(23)
Jun Arai
16.1 Introduction
313(1)
16.2 Integral Photography
314(5)
16.2.1 Principle
314(1)
16.2.2 Integral Photography with a Concave Lens Array
315(2)
16.2.3 Holocoder Hologram
317(1)
16.2.4 IP using a Retrodirective Screen
318(1)
16.2.5 Avoiding Pseudoscopic Images
318(1)
16.3 Real-Time System
319(6)
16.3.1 Orthoscopic Conversion Optics
319(3)
16.3.2 Applications of the Ultra-High-Resolution Video System
322(3)
16.4 Properties of the Reconstructed Image
325(5)
16.4.1 Geometrical Relationship of Subject and Spatial Image
325(1)
16.4.2 Resolution
326(3)
16.4.3 Viewing Area
329(1)
16.5 Research and Development Trends
330(4)
16.5.1 Acquiring and Displaying Spatial Information
330(1)
16.5.2 Elemental Image Generation from 3D Object Information
331(1)
16.5.3 Three-Dimensional Measurement
332(1)
16.5.4 Hologram Conversion
333(1)
16.6 Conclusion
334(2)
References
334(2)
17 3D Light-Field Display Technologies
336(13)
Peter Tamas Kovacs
Tibor Balogh
17.1 Introduction
336(1)
17.2 Fundamentals of 3D Displaying
337(2)
17.3 The Holo Vizio Light-Field Display System
339(3)
17.3.1 Design Principles and System Parameters
340(1)
17.3.2 Image Organization
341(1)
17.4 Holo Vizio Displays and Applications
342(3)
17.4.1 Desktop Displays
342(1)
17.4.2 Large-Scale Displays
343(1)
17.4.3 Cinema Display
343(1)
17.4.4 Software and Content Creation
344(1)
17.4.5 Applications
344(1)
17.5 The Perfect 3D Display
345(1)
17.6 Conclusions
345(4)
References
345(4)
PART V HUMAN VISUAL SYSTEM AND QUALITY ASSESSMENT
18 3D Media and the Human Visual System
349(28)
Simon J. Watt
Kevin J. MacKenzie
18.1 Overview
349(1)
18.2 Natural Viewing and S3D Viewing
349(1)
18.3 Perceiving 3D Structure
350(4)
18.3.1 Perceiving Depth from Binocular Disparity
352(2)
18.4 `Technical' Issues in S3D Viewing
354(3)
18.4.1 Cross-Talk
355(1)
18.4.2 Low Image Luminance and Contrast
355(1)
18.4.3 Photometric Differences Between Left-and Right-Eye Images
355(1)
18.4.4 Camera Misalignments and Differences in Camera Optics
356(1)
18.4.5 Window Violations
356(1)
18.4.6 Incorrect Specular Highlights
356(1)
18.5 Fundamental Issues in S3D Viewing
357(1)
18.6 Motion Artefacts from Field-Sequential Stereoscopic Presentation
357(4)
18.6.1 Perception of Flicker
359(1)
18.6.2 Perception of Unsmooth or Juddering Motion
359(1)
18.6.3 Distortions in Perceived Depth from Binocular Disparity
360(1)
18.6.4 Conclusions
360(1)
18.7 Viewing Stereoscopic Images from the `Wrong' Place
361(5)
18.7.1 Capture Parameters
361(3)
18.7.2 Display Parameters and Viewer Parameters
364(1)
18.7.3 Are Problems of Incorrect Geometry Unique to S3D?
364(2)
18.7.4 Conclusions
366(1)
18.8 Fixating and Focusing on Stereoscopic Images
366(6)
18.8.1 Accommodation, Vergence and Viewing Distance
367(1)
18.8.2 Accommodation and Vergence in the Real World and in S3D
367(1)
18.8.3 Correcting Focus Cues in S3D
368(1)
18.8.4 The Stereoscopic Zone of Comfort
369(1)
18.8.5 Specifying the Zone of Comfort for Cinematography
370(1)
18.8.6 Conclusions
371(1)
18.9 Concluding Remarks
372(5)
Acknowledgments
372(1)
References
372(5)
19 3D Video Quality Assessment
377(18)
Philippe Hanhart
Francesca De Simone
Martin Rerabek
Touradj Ebrahimi
19.1 Introduction
377(1)
19.2 Stereoscopic Artifacts
378(1)
19.3 Subjective Quality Assessment
379(5)
19.3.1 Psycho-perceptual (or Psychophysical) Experiments
380(2)
19.3.2 Descriptive (or Explorative) Approaches
382(1)
19.3.3 Hybrid Approaches
382(1)
19.3.4 Open Issues
383(1)
19.3.5 Future Directions
384(1)
19.4 Objective Quality Assessment
384(11)
19.4.1 Objective Quality Metrics
384(1)
19.4.2 From 2D to 3D
385(1)
19.4.3 Including Depth Information
386(1)
19.4.4 Beyond Image Quality
387(1)
19.4.5 Open Issues
388(1)
19.4.6 Future Directions
389(1)
References
389(6)
PART VI APPLICATIONS AND IMPLEMENTATION
20 Interactive Omnidirectional Indoor Tour
395(21)
Jean-Charles Bazin
Olivier Saurer
Friedrich Fraundorfer
Marc Pollefeys
20.1 Introduction
395(1)
20.2 Related Work
396(1)
20.3 System Overview
397(1)
20.4 Acquisition and Preprocessing
398(3)
20.4.1 Camera Model
398(2)
20.4.2 Data Acquisition
400(1)
20.4.3 Feature Extraction
401(1)
20.4.4 Key-Frame Selection
401(1)
20.5 SfM Using the Ladybug Camera
401(1)
20.6 Loop and Junction Detection
401(1)
20.7 Interactive Alignment to Floor Plan
402(3)
20.7.1 Notation
402(1)
20.7.2 Fusing SfM with Ground Control Points
403(2)
20.8 Visualization and Navigation
405(3)
20.8.1 Authoring
405(1)
20.8.2 Viewer
405(3)
20.9 Vertical Rectification
408(2)
20.9.1 Existing Studies
408(1)
20.9.2 Procedure Applied
408(1)
20.9.3 Line Extraction
408(1)
20.9.4 Line Clustering and VP Estimation
409(1)
20.10 Experiments
410(4)
20.10.1 Vertical Rectification
410(1)
20.10.2 Trajectory Estimation and Mapping
411(3)
20.11 Conclusions
414(2)
Acknowledgments
414(1)
References
414(2)
21 View Selection
416(16)
Fahad Daniyal
Andrea Cavallaro
21.1 Introduction
416(1)
21.2 Content Analysis
417(4)
21.2.1 Pose
417(2)
21.2.2 Occlusions
419(1)
21.2.3 Position
419(2)
21.2.4 Size
421(1)
21.2.5 Events
421(1)
21.3 Content Ranking
421(3)
21.3.1 Object-Centric Quality of View
422(1)
21.3.2 View-Centric Quality of View
423(1)
21.4 View Selection
424(2)
21.4.1 View Selection as a Scheduling Problem
425(1)
21.4.2 View Selection as an Optimization Problem
425(1)
21.5 Comparative Summary and Outlook
426(6)
References
429(3)
22 3D Video on Mobile Devices
432(18)
Arnaud Bourge
Alain Bellon
22.1 Mobile Ecosystem, Architecture, and Requirements
432(1)
22.2 Stereoscopic Applications on Mobile Devices
433(3)
22.2.1 3D Video Camcorder
434(1)
22.2.2 3D Video Player
434(1)
22.2.3 3D Viewing Modalities
434(1)
22.2.4 3D Graphics Applications
435(1)
22.2.5 Interactive Video Applications
435(1)
22.2.6 Monoscopic 3D
435(1)
22.3 Stereoscopic Capture on Mobile Devices
436(6)
22.3.1 Stereo-Camera Design
436(1)
22.3.2 Stereo Imaging
437(1)
22.3.3 Stereo Rectification, Lens Distortion, and Camera Calibration
438(2)
22.3.4 Digital Zoom and Video Stabilization
440(2)
22.3.5 Stereo Codecs
442(1)
22.4 Display Rendering on Mobile Devices
442(3)
22.4.1 Local Auto-Stereoscopic Display
442(1)
22.4.2 Remote HD Display
443(1)
22.4.3 Stereoscopic Rendering
443(2)
22.5 Depth and Disparity
445(3)
22.5.1 View Synthesis
445(1)
22.5.2 Depth Map Representation and Compression Standards
446(1)
22.5.3 Other Usages
447(1)
22.6 Conclusions
448(2)
Acknowledgments
448(1)
References
448(2)
23 Graphics Composition for Multiview Displays
450(18)
Jean Le Feuvre
Yves Mathieu
23.1 An Interactive Composition System for 3D Displays
450(1)
23.2 Multimedia for Multiview Displays
451(3)
23.2.1 Media Formats
451(1)
23.2.2 Multimedia Languages
452(1)
23.2.3 Multiview Displays
453(1)
23.3 GPU Graphics Synthesis for Multiview Displays
454(4)
23.3.1 3D Synthesis
454(1)
23.3.2 View Interleaving
455(2)
23.3.3 3D Media Rendering
457(1)
23.4 DIBR Graphics Synthesis for Multiview Displays
458(8)
23.4.1 Quick Overview
458(1)
23.4.2 DIBR Synthesis
459(1)
23.4.3 Hardware Compositor
460(2)
23.4.4 DIBR Pre- and Post-Processing
462(2)
23.4.5 Hardware Platform
464(2)
23.5 Conclusion
466(2)
Acknowledgments
466(1)
References
466(2)
24 Real-Time Disparity Estimation Engine for High-Definition 3DTV Applications
468(19)
Yu-Cheng Tseng
Tian-Sheuan Chang
24.1 Introduction
468(1)
24.2 Review of Disparity Estimation Algorithms and Implementations
469(2)
24.2.1 DP-Based Algorithms and Implementations
469(1)
24.2.2 GC-Based Algorithms and Implementations
470(1)
24.2.3 BP-Based Algorithms and Implementations
470(1)
24.3 Proposed Hardware-Efficient Algorithm
471(5)
24.3.1 Downsampled Matching Cost for Full Disparity Range
472(1)
24.3.2 Hardware-Efficient Cost Diffusion Method
472(1)
24.3.3 Upsampling Disparity Maps
473(1)
24.3.4 Temporal Consistency Enhancement Methods
474(1)
24.3.5 Occlusion Handling
475(1)
24.4 Proposed Architecture
476(3)
24.4.1 Overview of Architecture
476(1)
24.4.2 Computational Modules
477(1)
24.4.3 External Memory Access
478(1)
24.5 Experimental Results
479(4)
24.5.1 Comparison of Disparity Quality
479(1)
24.5.2 Analysis of Sampling Factor
480(1)
24.5.3 Implementation Result
481(2)
24.6 Conclusion
483(4)
References
483(4)
Index 487
Frédéric Dufaux, Télécom Paris Tech, CNRS, France

Béatrice Pesquet-Popescu, Télécom Paris Tech, France

Marco Cagnazzo, Télécom Paris Tech, France