No.
|
Title
|
TBP
|
Available
|
|
23001-4 Codec Configuration Representation
|
|
|
9584
|
Study Text of ISO/IEC CD 23001-4 Codec Configuration Representation
|
N
|
08/03/17
|
|
23002-4 Video Tool Library
|
|
|
9585
|
Reconfigurable Video Coding Requirements V 4.0
|
N
|
08/01/18
|
9586
|
Overview of Reconfigurable Video Coding (RVC)
|
Y
|
08/02/02
|
9587
|
Study Text of CD ISO/IEC 23002-4 Video Tool Library
|
N
|
08/03/17
|
9588
|
Extensions of Video Tool Library under consideration
|
N
|
08/02/04
|
9589
|
Description of Core Experiments in RVC
|
N
|
08/02/04
|
9590
|
RVC Simulation Model (RSM) V7.0
|
N
|
08/02/04
|
9591
|
RVC Work Plan and FU Development Status
|
N
|
08/01/18
|
9592
|
RVC Conformance Testing Working Draft V4.0
|
N
|
08/02/04
|
9593
|
Description of Exploration Experiments in RVC
|
N
|
08/01/18
|
9594
|
Methodologies for Video Toolbox Extension V2.0
|
N
|
08/03/24
|
6Explorations – Free Viewpoint Video/Television
The exploratory work on free-viewpoint video has its roots in the “3DAV” exploration, which was originally started in December 2001, and later led in a first CfP on multiview video compression technology (current MVC development in JVT). As discussed in the previous meeting, FTV can be defined as a compressed representation and associated technologies which enable generating a large number of different views from a sparse view set. This most probably (from technologies currently known) requires implementation of depth/disparity map estimation (non-normative), definition of depth/disparity map representation/compression and interpolation/rendering method (not clear yet whether the latter should be non-normative or normative). All of these elements rely on each other, such that proper technology selection will most probably not be simple. Furthermore, higher distortion may be expected than for MVC (or at least quality may not be measurable in terms of pixel fidelity, geometric distortions may appear that might only be noticeable under certain observation conditions). The amount of distortion most probably would also depend on compactness (density of views) and complexity of the methods. Depending on concrete application, the view number to be generated may range from two for simple stereoscopic up to "many" for almost-free walk-through a scene.
In Antalya, more clarification was achieved about the focus of the next phase of the FTV work. Realistic market needs appear to be existing in supporting standardized formats for upcoming 3D (M-view) displays where the number of views M as locally generated influences the quality of visual perception. While currently numbers of approximately M=9 are used in prototypes, while for the future M of up to 40-50 could be expected. Even then, the view angle support will be relatively narrow (max. 20 degrees left-to-right), which is a clear (and implementation-wise realistic) limitation as compared to the “really free” FTV scenario. One additional advantage could be that with such narrow view alteration ranges, co-planar camera setups could still be useful. After extensive discussion, the group agreed that the name of "3D Video" is very well reflecting this subset scenario of FTV (namely, enabling technology for 3D video displays).
Related to the last meeting’s call for test sequences and depth/disparity estimation algorithms the following input contributions were brought:
-
2 contributions announcing generation of new test materials (15089, 15102), both will use dense camera arrays. First proposal for 80 cameras with 5 cm baseline (convergent); the other 15 cameras with 7 cm baseline (linear/co-planar).
-
4 contributions on depth estimation (15090, 15119, 15175, 15191)
-
2 contributions on view generation (15090, 15120)
In addition, M15101 reports corrected camera parameters for an existing sequence, M15047 and M15088 relate to more generic applications and requirements of depth map technology and FTV (no need detected to revise the apps & reqs document based on these contributions). M15088 indicated possibilities to perform skipped view encoding when higher-quality depth information is available.
Following the focus as described above, it was concluded that co-planar camera setup would be the optimum case for this kind of 3D Video applications, and test sequences should be captured according to this. Nevertheless, slight rectification would most likely be necessary even in the parallel setup, due to slight variation in camera properties and the impossibility for perfect mechanical adjustment. Nevertheless, the original shots should be as close to co-planar as possible to keep rectification artifacts to a minimum. Sequences should be provided in rectified, illumination- and color-compensated version.
Following this, a new call for test sequences was produced, which also includes a high-level description of 3DTV and FTV to make the purposes clear for which the new materials should be useful. Again, it is called to provide depth maps, depth estimation and interpolation software packages. To get more evidence about the possible elements of the 3D video chain, an Exploration Experiment was started to find out how depth estimation and interpolation inter-relate, based on the proposals that were brought (and for which software must be made available in this context). The results of this should bring evidence by the next meeting about how we can find out about
-
whether sufficiently good depth estimation algorithms are available
-
which level of quality can be achieved in view synthesis, and whether e.g. PSNR comparison against original views is useful
-
suitability of test sequences we have (and will have after next meeting) for purposes of upcoming CfP
Documents reviewed
m15047
|
Consideration of Depth Format
|
Taka Senoh, yamamoto@nict.go.jp, oi@nict.go.jp, mishina@nict.go.jp, m-okui@nict.go.jp
Reports about various versions of depth, e.g. absolute z-depth, disparity. Various versions of normalization of depth and defining depth ranges. Proponents should check relationship with definitions in 23002-3.
|
m15088
|
Available Technologies for FTV
|
Masayuki Tanimoto, Toshiaki Fujii, Kazuyoshi Suzuki
Presented.
|
m15089
|
Contribution of Nagoya University on FTV Test Material
|
Masayuki Tanimoto, Toshiaki Fujii, Kazuyoshi Suzuki, Norishige Fukushima
|
m15090
|
Improvement of Depth Map Estimation and View Synthesis
|
Masayuki Tanimoto, Toshiaki Fujii, Kazuyoshi Suzuki
|
m15098
|
Inter-View Skip Mode for FTV using Depth Information
|
Gang Zhu, Xiaozhong Xu, Ping Yang, Yun He
|
m15101
|
Corrected Camera Parameters for N9468; Call for Contributions on FTV Test Material?
|
Aljoscha Smolic, Heribert Brust, Karsten Mueller, Marcus Mueller, Thomas Wiegand
|
m15102
|
Progress Report on 3DTV Video Acquisition
|
Ingo Feldmann, Marcus Mueller, Frederik Zilly, Ralf Tanger, Karsten Mueller, Aljoscha Smolic, Peter Kauff, Thomas Wiegand
|
m15119
|
Segment-based Multi-view Depth Map Estimation for FTV
|
Sang-Beom Lee, Kwan-Jung Oh
|
m15120
|
Virtual View Synthesis for FTV
|
Sang-Tae Na, Kwan-Jung Oh
|
m15175
|
Depth Map Estimation Software
|
Olgierd Stankiewicz, Krzysztof Wegner.
|
m15191
|
Segment-based Disparity Estimation using Foreground Sepration
|
Gi-Miun Um, Taeone Kim, Namho Hur, Jinwoong Kim
|
Output documents:
No.
|
Title
|
TBP
|
Available
|
|
Exploration – Free Viewpoint TV Coding
|
|
|
9595
|
Call for Contributions on 3D Video Test Material (Update)
|
Y
|
08/01/18
|
9596
|
Description of Exploration Experiments in 3D Video
|
N
|
08/01/18
|
-
– JVT report
Source
|
JVT Management Team (Gary J. Sullivan, Jens-Rainer Ohm, Thomas Wiegand, and Ajay Luthra)
|
Dostları ilə paylaş: |