T. Wiegand

Technische Universität Berlin, Berlín, Berlin, Germany

Are you T. Wiegand?

Claim your profile

Publications (279)137.34 Total impact

  • [Show abstract] [Hide abstract]
    ABSTRACT: High Efficiency Video Coding (HEVC) is the most recent jointly developed video coding standard of ITU-T Visual Coding Experts Group (VCEG) and ISO/IEC Moving Picture Experts Group (MPEG). Although its basic architecture is built along the conventional hybrid block-based approach of combining prediction with transform coding, HEVC includes a number of coding tools with greatly enhanced coding-efficiency capabilities relative to those of prior video coding standards. Among these tools are new transform coding techniques that include the support for dyadically increasing transform block sizes ranging from 4 × 4 to 32 × 32, the partitioning of residual blocks into variable block-size transforms by using a quadtree-based partitioning dubbed as residual quadtree (RQT) as well as some properly designed entropy coding techniques for quantized transform coefficients of variable transform block sizes. In this paper, we describe these HEVC techniques for transform coding with a particular focus on the RQT structure and the entropy coding stage and demonstrate their benefit in terms of improved coding efficiency by experimental results.
    IEEE Journal of Selected Topics in Signal Processing 12/2013; 7(6):978-989. · 3.30 Impact Factor
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: This paper proposes an image interpolation algorithm exploiting sparse representation for natural images. It involves three main steps: (a) obtaining an initial estimate of the high resolution image using linear methods like FIR filtering, (b) promoting sparsity in a selected dictionary through iterative thresholding, and (c) extracting high frequency information from the approximation to refine the initial estimate. For the sparse modeling, a shearlet dictionary is chosen to yield a multiscale directional representation. The proposed algorithm is compared to several state-of-the-art methods to assess its objective as well as subjective performance. Compared to the cubic spline interpolation method, an average PSNR gain of around 0.8 dB is observed over a dataset of 200 images.
    08/2013;
  • Source
  • [Show abstract] [Hide abstract]
    ABSTRACT: The paper describes an extension of the high efficiency video coding (HEVC) standard for coding of multi-view video and depth data. In addition to the known concept of disparity-compensated prediction, inter-view motion parameter and inter-view residual prediction for coding of the dependent video views have been developed and integrated. Furthermore, for depth coding, new intra coding modes, a modified motion compensation and motion vector coding as well as the concept of motion parameter inheritance are part of the HEVC extension. A novel encoder control uses view synthesis optimization, which guarantees that high quality intermediate views can be generated based on the decoded data. The bitstream format supports the extraction of partial bitstreams, so that conventional 2D video, stereo video and the full multi-view video plus depth (MVD) format can be decoded from a single bitstream. Objective and subjective results are presented, demonstrating that the proposed approach provides about 50% bit rate savings in comparison to HEVC simulcast and about 20% in comparison to a straightforward multi-view extension of HEVC without the newly developed coding tools.
    IEEE Transactions on Image Processing 05/2013; · 3.20 Impact Factor
  • [Show abstract] [Hide abstract]
    ABSTRACT: Intra prediction is a fundamental tool in video coding with hybrid block-based architecture. Recent investigations have shown that one of the most beneficial elements for a higher compression performance in high-resolution videos is the incorporation of larger block structures. Thus in this work, we investigate the performance of novel intra prediction modes based on different image completion techniques in a new video coding scheme with large block structures. Image completion methods exploit the fact that high frequency image regions yield high coding costs when using classical H.264/AVC prediction modes. This problem is tackled by investigating the incorporation of several intra predictors using the concept of Laplace partial differential equation (PDE), Least Square (LS) based linear prediction and the Auto Regressive model. A major aspect of this article is the evaluation of the coding performance in a qualitative (i.e. coding efficiency) manner. Experimental results show significant improvements in compression (up to 7.41 %) by integrating the LS-based linear intra prediction.
    Proc SPIE 02/2013;
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: This paper describes an extension of the upcoming High Efficiency Video Coding (HEVC) standard for supporting spatial and quality scalable video coding. Besides scalable coding tools known from scalable profiles of prior video coding standards such as H.262/MPEG-2 Video and H.264/MPEG-4 AVC, the proposed scalable HEVC extension includes new coding tools that further improve the coding efficiency of the enhancement layer. In particular, new coding modes by which base and enhancement layer signals are combined for forming an improved enhancement layer prediction signal have been added. All scalable coding tools have been integrated in a way that the low-level syntax and decoding process of HEVC remain unchanged to a large extent. Simulation results for typical application scenarios demonstrate the effectiveness of the proposed design. For spatial and quality scalable coding with two layers, bit-rate savings of about 20-30% have been measured relative to simulcasting the layers, which corresponds to a bit-rate overhead of about 5-15% relative to single-layer coding of the enhancement layer.
    Proc SPIE 02/2013;
  • [Show abstract] [Hide abstract]
    ABSTRACT: The paper describes a scalable video coding extension of the upcoming HEVC video coding standard for spatial and quality scalable coding. Besides coding tools known from scalable profiles of prior video coding standards, it includes new coding tools that further improve the enhancement layer coding efficiency. The effectiveness of the proposed scalable HEVC extension is demonstrated by comparing the coding efficiency to simulcast and single-layer coding for several test sequences and coding conditions.
    Data Compression Conference (DCC), 2013; 01/2013
  • P. Merkle, K. Muller, T. Wiegand
    [Show abstract] [Hide abstract]
    ABSTRACT: This paper presents a new approach for the depth coding part of a 3D video coding extension based on the Multiview Video plus Depth (MVD) representation. Our approach targets a higher coding efficiency for the depth component and is motivated by the fact that depth signals have specific characteristics that differ from video. For this purpose we apply the method of wedgelet segmentation with residual adaptation for depth blocks by implementing a new set of coding and prediction modes and by optimizing the algorithms for efficient processing and signaling. The results show that a bit rate reduction of up to 6% is achieved for the depth component, using a 3D video codec based on the high-efficiency video coding (HEVC) technology. Apart from the depth coding gains, wedgelets lead to a considerably better rendered view quality.
    Multimedia and Expo (ICME), 2013 IEEE International Conference on; 01/2013
  • [Show abstract] [Hide abstract]
    ABSTRACT: The high bit rates of high-definition or 3D-services require a huge share of the valuable terrestrial spectrum, especially when targeting wide coverage areas. This article describes how to provide future services with the state-of-the-art digital terrestrial TV technology DVB-T2 in a flexible and cost-efficient way. The combination of layered media such as the scalable and 3D extension of the H.264/AVC or emerging H.265/HEVC format with the physical layer pipes feature of DVB-T2 enables flexible broadcast of services with differentiated protection of the quality layers. This opens up new ways of service provisioning such as graceful degradation for mobile or fixed reception. This article shows how existing DVB-T2 and MPEG-2 transport stream mechanisms need to be configured for offering such services over DVB-T2. A detailed description of the setup of such services and the involved components is given.
    IEEE Communications Magazine 01/2013; 51(10):76-82. · 3.66 Impact Factor
  • Source
    ICIP'13 (Melbourne, Australia, 2013), Proc., to appear.; 01/2013
  • H. Lakshman, H. Schwarz, T. Wiegand
    [Show abstract] [Hide abstract]
    ABSTRACT: Fractional sample interpolation with finite impulse response (FIR) filters is commonly used for motion-compensated prediction (MCP). The FIR filtering can be viewed as a signal decomposition using restricted basis functions. The concept of generalized interpolation provides a greater degree of freedom for selecting basis functions. We developed a generalized interpolation framework for MCP using fixed-point infinite impulse response and FIR filters. An efficient multiplication-free design of the algorithm that is suited for hardware implementation is shown. A detailed analysis of average and worst case complexities compared to FIR filter-based interpolation techniques is provided. Average bitrate savings of around 2.0% compared to an 8-tap FIR filter are observed over the high-efficiency video coding dataset at a similar worst case complexity.
    IEEE Transactions on Circuits and Systems for Video Technology 01/2013; 23(3):455-466. · 1.82 Impact Factor
  • Source
  • [Show abstract] [Hide abstract]
    ABSTRACT: With the prospective High Effciency Video Coding (HEVC) standard as jointly developed by ITU-T VCEG and ISO/IEC MPEG, a new step in video compression capability is achieved. Technically, HEVC is a hybrid video-coding approach using quadtree-based block partitioning together with motion-compensated prediction. Even though a high degree of adaptability is achieved by quadtree-based block partitioning, this approach is intrinsically tied to certain drawbacks which may result in redundant sets of motion parameters to be transmitted. In order to remove those redundancies, a block-merging algorithm for HEVC is proposed. This algorithm generates a single motion-parameter set for a whole region of contiguous motion-compensated blocks. Simulation results show that the proposed merging technique works more effciently than a conceptually similar direct mode.
    Proc SPIE 10/2012;
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: The most recent video compression technology is High Efficiency Video Coding (HEVC). This soon to be completed standard is a joint development by Video Coding Experts Group (VCEG) of ITU-T and Moving Picture Experts Group (MPEG) of ISO/IEC. As one of its major technical novelties, HEVC supports variable prediction and transform block sizes using the quadtree approach for block partitioning. In terms of entropy coding, the Draft International Standard (DIS) of HEVC specifies context-based adaptive binary arithmetic coding (CABAC) as the single mode of operation. In this paper, a description of the specific CABAC-based entropy coding part in HEVC is given that is related to block structures and transform coefficient levels. In addition, experimental results are presented that indicate the benefit of the transform-coefficient level coding design in HEVC in terms of improved coding performance and reduced complexity.
    Proc SPIE 10/2012;
  • [Show abstract] [Hide abstract]
    ABSTRACT: The state-of-the-art hybrid video coding standard H.264/AVC defines two entropy-coding schemes with different complexity-performance trade-offs. Supporting these two schemes within a single standard raises several problems ranging from higher efforts for product development to increased silicon costs for hardware implementations. To overcome these issues, this work proposes a unified and complexity-scalable entropy-coding framework that is based on PIPE/V2V. The proposed framework uses a single set of tools for all configurations and achieves the same complexity-performance trade-offs as the existing entropy-coding schemes through scalability.
    Proceedings / ICIP ... International Conference on Image Processing 09/2012;
  • [Show abstract] [Hide abstract]
    ABSTRACT: Entropy coding schemes like context-based adaptive binary arithmetic coding (CABAC) as used in the H.264/AVC video coding standard are established in the market. To increase throughput and possibilities for parallelization, probability interval partitioning entropy coding (PIPE) was developed. It replaces the arithmetic entropy coding step of the CABAC scheme with a set of bin coders, each operating at a fixed probability. In this paper, the use of variable-to-variable length (v2v) codes for the bin coders is discussed and a concept for increasing the throughput of table-based v2v-coder implementations is developed. It is based on code word concatenation and thus virtually neutral in terms of coding efficiency. Using this technique in the emerging HEVC video coding standard show an overall decoding time reduction of up to 5% when compared to the unchanged CABAC.
    Consumer Electronics (ICCE 2012), Berlin, Germany. IEEE International Conference on. 09/2012;
  • [Show abstract] [Hide abstract]
    ABSTRACT: Perception-oriented video coding based on image analysis and completion (IAC) has received increasing attention over recent years. Several approaches have been proposed that follow diverging strategies: from full compatibility to hybrid block transform coding to alternative codec designs. In this paper, we review the most significant issues in IAC coding and their relevance for future codec designs. The most promising pathways will be analyzed along with limitations and open issues. The challenges in IAC-based video coding will be summarized using methods developed by the authors and others for block-based and region-based IAC coding.
    IEEE Journal of Selected Topics in Signal Processing 07/2012; 27(6):579–594. · 3.30 Impact Factor
  • [Show abstract] [Hide abstract]
    ABSTRACT: In hybrid video coding, an entropy coding scheme transmits the quantized transform coefficients, resulting from block-based transformation and quantization of the difference between the prediction signal and the original signal, and additional side information. The state-of-the-art hybrid video coding standard H.264/AVC defines two different entropy coding schemes with different complexity-performance tradeoff. As a result, the support for two different entropy coding schemes has to be maintained and introduces several problems. To overcome these issues, a unified solution is proposed, which is based on the PIPE/V2V coding concept. It achieves the same complexity-performance trade-offs as the existing entropy coding schemes by scalability. The advantage of the proposed scheme over the existing concept is the usage of the same set of tools for all configurations. Simulation results and complexity analysis on hardware show the efficiency of the proposed scheme.
    Picture Coding Symposium (PCS 2012), Kraków, Poland. 05/2012;
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: HTTP-based video streaming has been gaining popularity within the recent years. There are multiple benefits of relying on HTTP/TCP connections, such as the usage of the widely deployed network caches to relieve video servers from sending the same content to a high number of users and the avoidance of traversal issues with firewalls and NATs typical for RTP/UDP-based solutions. Therefore, many service providers resort to adopt HTTP streaming as the basis for their services. In this paper, the benefits of using the Scalable Video Coding (SVC) for a HTTP streaming service are shown, and the SVC based approach is compared to the AVC based approach. We show that network resources are more efficiently used and how the benefits of the traditional techniques can even be heightened by adopting the Scalable Video Coding (SVC) as the video codec for adaptive low delay streaming over HTTP. For the latter small playout-buffers are considered hence allowing low media access latency in the delivery chain and it is shown that adaptation is more effectively performed with the SVC based approach.
    Signal Processing-image Communication - SIGNAL PROCESS-IMAGE COMMUN. 04/2012;
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: An approach to the direct measurement of perception of video quality change using electroencephalography (EEG) is presented. Subjects viewed 8-s video clips while their brain activity was registered using EEG. The video signal was either uncompressed at full length or changed from uncompressed to a lower quality level at a random time point. The distortions were introduced by a hybrid video codec. Subjects had to indicate whether they had perceived a quality change. In response to a quality change, a positive voltage change in EEG (the so-called P3 component) was observed at latency of about 400-600 ms for all subjects. The voltage change positively correlated with the magnitude of the video quality change, substantiating the P3 component as a graded neural index of the perception of video quality change within the presented paradigm. By applying machine learning techniques, we could classify on a single-trial basis whether a subject perceived a quality change. Interestingly, some video clips wherein changes were missed (i.e., not reported) by the subject were classified as quality changes, suggesting that the brain detected a change, although the subject did not press a button. In conclusion, abrupt changes of video quality give rise to specific components in the EEG that can be detected on a single-trial basis. Potentially, a neurotechnological approach to video assessment could lead to a more objective quantification of quality change detection, overcoming the limitations of subjective approaches (such as subjective bias and the requirement of an overt response). Furthermore, it allows for real-time applications wherein the brain response to a video clip is monitored while it is being viewed.
    IEEE Transactions on Image Processing 02/2012; 21(5):2619-29. · 3.20 Impact Factor

Publication Stats

12k Citations
137.34 Total Impact Points

Institutions

  • 2009–2012
    • Technische Universität Berlin
      • Department of Telecommunication Systems
      Berlín, Berlin, Germany
  • 2001–2012
    • Fraunhofer Heinrich-Hertz-Institute HHI
      • Department of Image Processing (IP)
      Berlín, Berlin, Germany
  • 2011
    • Cisco Systems, Inc
      San Jose, California, United States
  • 2008
    • European Broadcasting Union
      Genève, Geneva, Switzerland
  • 2003–2008
    • Technische Universität München
      • Lehrstuhl für Nachrichtentechnik LNT
      München, Bavaria, Germany
    • University of Zagreb
      • Faculty of Electrical Engineering and Computing (FER)
      Zagrabia, Grad Zagreb, Croatia
  • 2005
    • Microsoft
      Washington, West Virginia, United States
  • 2002
    • Stanford University
      • Information Systems Laboratory
      Stanford, CA, United States
    • University of British Columbia - Vancouver
      Vancouver, British Columbia, Canada
  • 1996–1999
    • Universitätsklinikum Erlangen
      Erlangen, Bavaria, Germany
  • 1995
    • University of California, Santa Barbara
      Santa Barbara, California, United States