Video Compression

14,000,000 Leading Edge Experts on the ideXlab platform

Scan Science and Technology

Contact Leading Edge Experts & Companies

Scan Science and Technology

Contact Leading Edge Experts & Companies

The Experts below are selected from a list of 34650 Experts worldwide ranked by ideXlab platform

B. Bai - One of the best experts on this subject based on the ideXlab platform.

  • Multiview Video Compression
    2009
    Co-Authors: B. Bai
    Abstract:

    With the progress of computer graphics and computer vision technologies, 3D/multiview Video applications such as 3D-TV and tele-immersive conference become more and more popular and are very likely to emerge as a prime application in the near future. A successful 3D/multiview Video system needs synergistic integration of various technologies such as 3D/multiview Video acquisition, Compression, transmission and rendering. In this thesis, we focus on addressing the challenges for multiview Video Compression. In particular, we have made 5 major contributions: (1) We propose a novel neighbor-based multiview Video Compression system which helps remove the inter-view redundancies among multiple Video streams and improve the performance. An optimal stream encoding order algorithm is designed to enable the encoder to automatically decide the stream encoding order and find the best reference streams. (2) A novel multiview Video transcoder is designed and implemented. The proposed multiview Video transcoder can be used to encode multiple compressed Video streams and reduce the cost of multiview Video acquisition system. (3) A learning-based multiview Video Compression scheme is invented. The novel multiview Video Compression algorithms are built on the recent advances on semi-supervised learning algorithms and achieve Compression by finding a sparse representation of images. (4) Two novel distributed source coding algorithms, EETG and SNS-SWC, are put forward. Both EETG and SNS-SWC are capable to achieve the whole Slepian-Wolf rate region and are syndrome-based schemes. EETG simplifies the code construction algorithm for distributed source coding schemes using extended Tanner graph and is able to handle mismatched bits at the encoder. SNS-SWC has two independent decoders and thus can simplify the decoding process. (5) We propose a novel distributed multiview Video coding scheme which allows flelxible rate allocation between two distributed multiview Video encoders. SNS-SWC is used as the underlying Slepian-Wolf coding scheme. It is the first work to realize simultaneous Slepian-Wolf coding of stereo Videos with the help of a distributed source code that achieves the whole Slepian-Wolf rate region. The proposed scheme has a better rate-distortion performance than the separate H.264 coding scheme in the high-rate case.

  • ICME - An Efficient Multiview Video Compression Scheme
    2005 IEEE International Conference on Multimedia and Expo, 1
    Co-Authors: B. Bai, Pierre Boulanger, Janelle Harms
    Abstract:

    Multiview Video Compression is important to the image-based 3D Video applications. In this paper, we proposes a novel neighbor-based multiview Video Compression scheme. It is essentially a MPEG2-like block-based scheme. In particular, a method to decide the stream encoding order is presented. The resulting stream encoding order can better decorrelate spatial redundancies among multiple Video streams than the center approach. Experimental results confirm the superiority of the proposed neighbor approach over the center approach and MPEG2 for multiview Video Compression

Janelle Harms - One of the best experts on this subject based on the ideXlab platform.

  • ICME - An Efficient Multiview Video Compression Scheme
    2005 IEEE International Conference on Multimedia and Expo, 1
    Co-Authors: B. Bai, Pierre Boulanger, Janelle Harms
    Abstract:

    Multiview Video Compression is important to the image-based 3D Video applications. In this paper, we proposes a novel neighbor-based multiview Video Compression scheme. It is essentially a MPEG2-like block-based scheme. In particular, a method to decide the stream encoding order is presented. The resulting stream encoding order can better decorrelate spatial redundancies among multiple Video streams than the center approach. Experimental results confirm the superiority of the proposed neighbor approach over the center approach and MPEG2 for multiview Video Compression

Lubomir Bourdev - One of the best experts on this subject based on the ideXlab platform.

  • Learned Video Compression
    2019 IEEE CVF International Conference on Computer Vision (ICCV), 2019
    Co-Authors: Oren Rippel, Sanjay Nair, Steve Branson, Alexander Anderson, Lubomir Bourdev
    Abstract:

    We present a new algorithm for Video coding, learned end-to-end for the low-latency mode. In this setting, our approach outperforms all existing Video codecs across nearly the entire bitrate range. To our knowledge, this is the first ML-based method to do so. We evaluate our approach on standard Video Compression test sets of varying resolutions, and benchmark against all mainstream commercial codecs in the low-latency mode. On standard-definition Videos, HEVC/H.265, AVC/H.264 and VP9 typically produce codes up to 60% larger than our algorithm. On high-definition 1080p Videos, H.265 and VP9 typically produce codes up to 20% larger, and H.264 up to 35% larger. Furthermore, our approach does not suffer from blocking artifacts and pixelation, and thus produces Videos that are more visually pleasing. We propose two main contributions. The first is a novel architecture for Video Compression, which (1) generalizes motion estimation to perform any learned compensation beyond simple translations, (2) rather than strictly relying on previously transmitted reference frames, maintains a state of arbitrary information learned by the model, and (3) enables jointly compressing all transmitted signals (such as optical flow and residual). Secondly, we present a framework for ML-based spatial rate control - a mechanism for assigning variable bitrates across space for each frame. This is a critical component for Video coding, which to our knowledge had not been developed within a machine learning setting.

  • learned Video Compression
    arXiv: Image and Video Processing, 2018
    Co-Authors: Oren Rippel, Sanjay Nair, Steve Branson, Carissa Lew, Alexander G Anderson, Lubomir Bourdev
    Abstract:

    We present a new algorithm for Video coding, learned end-to-end for the low-latency mode. In this setting, our approach outperforms all existing Video codecs across nearly the entire bitrate range. To our knowledge, this is the first ML-based method to do so. We evaluate our approach on standard Video Compression test sets of varying resolutions, and benchmark against all mainstream commercial codecs, in the low-latency mode. On standard-definition Videos, relative to our algorithm, HEVC/H.265, AVC/H.264 and VP9 typically produce codes up to 60% larger. On high-definition 1080p Videos, H.265 and VP9 typically produce codes up to 20% larger, and H.264 up to 35% larger. Furthermore, our approach does not suffer from blocking artifacts and pixelation, and thus produces Videos that are more visually pleasing. We propose two main contributions. The first is a novel architecture for Video Compression, which (1) generalizes motion estimation to perform any learned compensation beyond simple translations, (2) rather than strictly relying on previously transmitted reference frames, maintains a state of arbitrary information learned by the model, and (3) enables jointly compressing all transmitted signals (such as optical flow and residual). Secondly, we present a framework for ML-based spatial rate control: namely, a mechanism for assigning variable bitrates across space for each frame. This is a critical component for Video coding, which to our knowledge had not been developed within a machine learning setting.

Xin Jin - One of the best experts on this subject based on the ideXlab platform.

  • learning for Video Compression
    IEEE Transactions on Circuits and Systems for Video Technology, 2020
    Co-Authors: Zhibo Chen, Xin Jin
    Abstract:

    One key challenge to learning-based Video Compression is that motion predictive coding, a very effective tool for Video Compression, can hardly be trained into a neural network. In this paper, we propose the concept of PixelMotionCNN (PMCNN) which includes motion extension and hybrid prediction networks. PMCNN can model spatiotemporal coherence to effectively perform predictive coding inside the learning network. On the basis of PMCNN, we further explore a learning-based framework for Video Compression with additional components of iterative analysis/synthesis and binarization. The experimental results demonstrate the effectiveness of the proposed scheme. Although entropy coding and complex configurations are not employed in this paper, we still demonstrate superior performance compared with MPEG-2 and achieve comparable results with H.264 codec. The proposed learning-based scheme provides a possible new direction to further improve Compression efficiency and functionalities of future Video coding.

Didier Le J Gall - One of the best experts on this subject based on the ideXlab platform.

  • the mpeg Video Compression algorithm
    Signal Processing-image Communication, 1992
    Co-Authors: Didier Le J Gall
    Abstract:

    Abstract The Video Compression technique developed by MPEG covers many applications from interactive systems on CD-ROM to delivery of Video information over telecommunications networks. The MPEG Video Compression algorithm relies on two basic techniques: block based motion compensation for the reduction of the temporal redundancy and transform domain based Compression for the reduction of spatial redundancy. Motion compensation techniques are applied with both predictive and interpolative techniques. The prediction error signal is further compressed with spatial redundancy reduction (DCT). The quality of the compressed Video with the MPEG algorithm at about 1.5 Mbit/s has been compared to that of consumer grade VCR's.