CA2796778C - Methods and devices for encoding and decoding transform domain filters - Google Patents

Methods and devices for encoding and decoding transform domain filters Download PDF

Info

Publication number
CA2796778C
CA2796778C CA2796778A CA2796778A CA2796778C CA 2796778 C CA2796778 C CA 2796778C CA 2796778 A CA2796778 A CA 2796778A CA 2796778 A CA2796778 A CA 2796778A CA 2796778 C CA2796778 C CA 2796778C
Authority
CA
Canada
Prior art keywords
block
transformed
samples
filter
prediction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CA2796778A
Other languages
French (fr)
Other versions
CA2796778A1 (en
Inventor
Jing Wang
Xiang Yu
Dake He
Steven Wayne Campbell
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BlackBerry Ltd
Original Assignee
BlackBerry Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by BlackBerry Ltd filed Critical BlackBerry Ltd
Publication of CA2796778A1 publication Critical patent/CA2796778A1/en
Application granted granted Critical
Publication of CA2796778C publication Critical patent/CA2796778C/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/004Predictors, e.g. intraframe, interframe coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/189Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
    • H04N19/19Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding using optimisation based on Lagrange multipliers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/625Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using discrete cosine transform [DCT]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • H04N19/82Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop

Abstract

Methods and devices for encoding and decoding data using transform domain filtering are described. The encoder determines a set of transform domain filter coefficients to be applied to a transform domain prediction. The filtering may, in some cases, also apply to transform domain reconstructions. Rate-distortion optimization may be used to determine the optimal filter coefficients on a frame-basis, coding-unit- basis, or other basis. Multiple filters may be developed and communicated from the encoder to the decoder for different combinations of transform block size, coding mode, prediction mode, and texture type. In other cases, the filtering is applied in the pixel-domain to a pixel-domain prediction or a pixel-domain reconstruction of a block of samples.

Description

METHODS AND DEVICES FOR ENCODING AND
DECODING TRANSFORM DOMAIN FILTERS
FIELD
[0002] The present application generally relates to data compression and, in particular, to methods and devices for encoding and decoding transform domain filters.
BACKGROUND
[0003] Data compression occurs in a number of contexts. It is very commonly used in communications and computer networking to store, transmit, and reproduce information efficiently. It finds particular application in the encoding of images, audio and video. Video presents a significant challenge to data compression because of the large amount of data required for each video frame and the speed with which encoding and decoding often needs to occur. The current state-of-the-art for video encoding is the ITU-T H.264/AVC
video coding standard. It defines a number of different profiles for different applications, including the Main profile, Baseline profile and others. A next-generation video encoding standard is currently under development through a joint initiative of MPEG-ITU: High Efficiency Video Coding (HEVC).
[0004] There are a number of standards for encoding/decoding images and videos, including H.264, that use block-based coding processes. In these processes, the image or Our 101-0092CAPI RIM 43609-CA-PAT

frame is divided into blocks, typically 4x4 or 8x8, and the blocks are spectrally transformed into coefficients, quantized, and entropy encoded. In many cases, the data being transformed is not the actual pixel data, but is residual data following a prediction operation. Predictions can be intra-frame, i.e. block-to-block within the frame/image, or inter-frame, i.e. between frames (also called motion prediction). It is expected that HEVC will also have these features.
[0005] When spectrally transforming residual data, many of these standards prescribe the use of a discrete cosine transform (DCT) or some variant thereon. The resulting DCT
coefficients are then quantized using a quantizer that employs a uniform quantization step size.
[0006] Quantization is lossy. In other words, it introduces distortion that shows up as noise in the reconstructed images or videos. Accordingly, many existing compression schemes utilize some form of post-processing, i.e. filtering, to try to remove quantization noise from reconstructed pixels. Examples include deblocking filters, de-noising filters, or other pixel-domain filters.
[0007] Work in lossy compression, e.g., audio/voice coding, video coding, image coding, etc., tends to focus on improving rate-distortion performance. That is, the objective of most encoding and decoding schemes is to find an optimal balance between distortion and coding rate. A rate-distortion optimization expression of the type J = D + AR
is typically used, wherein the Lagrangian multiplier A, represents the desired trade-off between coding rate and distortion.
BRIEF SUMMARY
100081 The present application describes methods and encoders/decoders for achieving rate-distortion improvements in lossy data compression. In some embodiments, the data compression process or devices described herein may be applied to the encoding and decoding of audio data, image data, and/or video data. In particular, the present application describes a method and process of data compression that uses transform domain filtering.
100091 In one aspect, the present application describes a method of decoding filters from a bitstream of encoded data for use in a decoder for reconstructing blocks of samples for Our 101-0092CAP1 RIM 43609-CA-PAT

an image. The method includes reading filter-type flags from the bitstream, wherein the filter-type flags identify one or more filter types; and for each of the filter types identified, decoding a set of filter coefficients from the bitstream.
[0010] In another aspect, the present application describes a method of reconstructing a block of samples from a bitstream of encoded data in a decoder, wherein the decoder has a spectrally transformed prediction block of samples associated with the block of samples. The method includes reading filter-type flags from the bitstream, wherein the filter-type flags identify one or more filter types; for each of the filter types identified, decoding a set of filter coefficients from the bitstream; and calculating a filtered transform domain reconstruction by applying, to at least a portion of the transformed prediction block, one of the sets of filter coefficients based on its filter type matching characteristics associated with the block of samples.
[0011] In a further aspect, the present application describes encoders and decoders configured to implement such methods of encoding and decoding.
[0012] In yet a further aspect, the present application describes non-transitory computer-readable media storing computer-executable program instructions which, when executed, configured a processor to perform the described methods of encoding and/or decoding.
[0013] Other aspects and features of the present application will be understood by those of ordinary skill in the art from a review of the following description of examples in conjunction with the accompanying figures.
BRIEF DESCRIPTION OF THE DRAWINGS
[0014] Reference will now be made, by way of example, to the accompanying drawings which show example embodiments of the present application, and in which:
[0015] Figure 1 shows, in block diagram form, an encoder for encoding video;
[0016] Figure 2 shows, in block diagram form, a decoder for decoding video;
Our 10 1-0092CA P I RIM 43609-CA-PAT

100171 Figure 3 shows a block diagram of an example encoder with transform domain filtering;
[0018] Figure 4 shows a flowchart illustrating an example method of encoding;
[0019] Figure 5 shows a block diagram of another example encoder with transform domain filtering;
[0020] Figure 6 shows a flowchart illustrating another example method of encoding;
[0021] Figure 7 shows a flowchart illustrating a first example method of decoding with transform domain filtering;
[0022] Figure 8 shows a flowchart illustrating a second example method of decoding with transform domain filtering;
[0023] Figure 9 shows a simplified block diagram of an example embodiment of an encoder;
[0024] Figure 10 shows a simplified block diagram of an example embodiment of a decoder;
[0025] Figure 11 shows an example grouping of coefficient positions for filtering in a 4x4 block;
[0026] Figure 12 shows another example grouping of coefficient positions for filtering in a 4x4 block; and [0027] Figure 13 illustrates the various angles-of-prediction defined for intra-coding.
100281 Similar reference numerals may have been used in different figures to denote similar components.
DESCRIPTION OF EXAMPLE EMBODIMENTS
[0029] In the description that follows, some example embodiments are described with reference to the H.264 standard for video coding. Those ordinarily skilled in the art will understand that the present application is not limited to H.264 but may be applicable to other video coding/decoding standards, including possible future standards, such as HEVC. It will also be appreciated that the present application is not necessarily limited to video Our 101-0092CAP1 RIM 43609-CA-PAT

coding/decoding and may be applicable to audio coding/decoding, image coding/decoding, or the lossy coding/decoding of any other data. The present application is broadly applicable to any lossy data compression process that employs a prediction operation, a spectral transform of a residual following the prediction operation, and quantization of the transformed residual, irrespective of the type of data being coded/decoded.
[0030] In the description that follows, when referring to video or images the terms frame, slice, tile and rectangular slice group may be used somewhat interchangeably. Those of skill in the art will appreciate that, in the case of the H.264 standard, a frame may contain one or more slices. It will also be appreciated that certain encoding/decoding operations are performed on a frame-by-frame basis, some are performed on a slice-by-slice basis, some tile-by-tile, and some by rectangular slice group, depending on the particular requirements of the applicable image or video coding standard. In any particular embodiment, the applicable image or video coding standard may determine whether the operations described below are performed in connection with frames and/or slices and/or tiles and/or rectangular slice groups, as the case may be. Accordingly, those ordinarily skilled in the art will understand, in light of the present disclosure, whether particular operations or processes described herein and particular references to frames, slices, tiles, rectangular slice groups are applicable to frames, slices, tiles, rectangular slice groups, or some or all of those for a given embodiment. This also applies to coding units, groups of coding units, etc., as will become apparent in light of the description below.
[0031] To the extent that the processes or methods described below are applied to images and/or video they may be applied to a portion of a video or image, such as a frame, a slice, a Group-of-Pictures (GOP), or on any other basis, such as to a coding unit, or group of coding units. To the extent that the process or methods described herein are applied to audio, such as music or voice data, they may be applied to a grouping or sequence of data points, e.g.
an audio sample. It should be noted that the terms picture and frame could be used interchangeably.
[0032] The description herein refers to filters and filter coefficients for use in the transform domain. It will be understood from the following description that the transform domain filter is a set of scalars or multipliers, each selected to be applied to the transform domain coefficient in a particular transform domain coefficient position. The transform Our I01-0092CAP I RIM 43609-CA-PAT

domain filter may be applied to a reconstruction in the transform domain or to a prediction in the transform domain, as will be explained below. The term "filter coefficient" may be used interchangeably herein with "scalar", "adaptive prediction scalar", or "transform domain filter"(TDF).
[0033] Although many of the examples below relate to the determination of filter coefficients for transform-domain filtering, it will be appreciated that the filter coefficient transmission processes described herein may be applied to filter coefficients for filtering in the pixel-domain instead. In other words, the processes for encoding and decoding filter information described herein may be applied to transform-domain filter information or pixel-domain filter information.
[0034] Reference is now made to Figure 1, which shows, in block diagram form, an encoder 10 for encoding video. Reference is also made to Figure 2, which shows a block diagram of a decoder 50 for decoding video. It will be appreciated that the encoder 10 and decoder 50 described herein may each be implemented on an application-specific or general purpose computing device, containing one or more processing elements and memory. The operations performed by the encoder 10 or decoder 50, as the case may be, may be implemented by way of application-specific integrated circuit, for example, or by way of stored program instructions executable by a general purpose processor. The device may include additional software, including, for example, an operating system for controlling basic device functions. The range of devices and platforms within which the encoder 10 or decoder 50 may be implemented will be appreciated by those ordinarily skilled in the art having regard to the following description.
[0035] The encoder 10 receives a video source 12 and produces an encoded bitstream 14. The decoder 50 receives the encoded bitstream 14 and outputs a decoded video frame 16.
The encoder 10 and decoder 50 may be configured to operate in conformance with a number of video compression standards. For example, the encoder 10 and decoder 50 may be H.264/AVC compliant. In other embodiments, the encoder 10 and decoder 50 may conform to other video compression standards, including evolutions of the H.264/AVC
standard, like HEVC.
[0036] The encoder 10 includes a spatial predictor 21, a coding mode selector 20, transform processor 22, quantizer 24, and entropy encoder 26. As will be appreciated by Our 101-0092CAP1 RIM 43609-CA-PAT

those ordinarily skilled in the art, the coding mode selector 20 determines the appropriate coding mode for the video source, for example whether the subject frame/slice is of!, P, or B
type, and whether particular coding units (e.g. macroblocks) within the frame/slice are inter or intra coded. The transform processor 22 performs a transform upon the spatial domain data.
In particular, the transform processor 22 applies a block-based transform to convert spatial domain data to spectral components. For example, in many embodiments a discrete cosine transform (DCT) is used. Other transforms, such as a discrete sine transform or others may be used in some instances. The block-based transform is performed on a macroblock or sub-block basis, depending on the size of the macroblocks. In the H.264 standard, for example, a typical 16x16 macroblock contains sixteen 4x4 transform blocks and the DCT
process is performed on the 4x4 blocks. In some cases, the transform blocks may be 8x8, meaning there are four transform blocks per macroblock. In yet other cases, the transform blocks may be other sizes. In some cases, a 16x16 macroblock may include a non-overlapping combination of 4x4 and 8x8 transform blocks.
[0037] Applying the block-based transform to a block of pixel data results in a set of transform domain coefficients. A "set" in this context is an ordered set in which the coefficients have coefficient positions. In some instances the set of transform domain coefficients may be considered a "block" or matrix of coefficients. In the description herein the phrases a "set of transform domain coefficients" or a "block of transform domain coefficients" are used interchangeably and are meant to indicate an ordered set of transform domain coefficients.
[0038] The set of transform domain coefficients is quantized by the quantizer 24. The quantized coefficients and associated information are then encoded by the entropy encoder 26.
[0039] Intra-coded frames/slices (L e. type I) are encoded without reference to other frames/slices. In other words, they do not employ temporal prediction. However intra-coded frames do rely upon spatial prediction within the frame/slice, as illustrated in Figure 1 by the spatial predictor 21. That is, when encoding a particular block the data in the block may be compared to the data of nearby pixels within blocks already encoded for that frame/slice.
Using a prediction algorithm, the source data of the block may be converted to residual data.
The transform processor 22 then encodes the residual data. H.264, for example, prescribes nine spatial prediction modes for 4x4 transform blocks. In some embodiments, each of the Our 101-0092CAP I RIM 43609-CA-PAT
- 8 -nine modes may be used to independently process a block, and then rate-distortion optimization is used to select the best mode.
[00401 The H.264 standard also prescribes the use of motion prediction/compensation to take advantage of temporal prediction. Accordingly, the encoder 10 has a feedback loop that includes a de-quantizer 28, inverse transform processor 30, and deblocking processor 32.
These elements mirror the decoding process implemented by the decoder 50 to reproduce the frame/slice. A frame store 34 is used to store the reproduced frames. In this manner, the motion prediction is based on what will be the reconstructed frames at the decoder 50 and not on the original frames, which may differ from the reconstructed frames due to the lossy compression involved in encoding/decoding. A motion predictor 36 uses the frames/slices stored in the frame store 34 as source frames/slices for comparison to a current frame for the purpose of identifying similar blocks. Accordingly, for macroblocks to which motion prediction is applied, the "source data" which the transform processor 22 encodes is the residual data that comes out of the motion prediction process. For example, it may include information regarding the reference frame, a spatial displacement or "motion vector", and residual pixel data that represents the differences (if any) between the reference block and the current block. Information regarding the reference frame and/or motion vector may not be processed by the transform processor 22 and/or quantizer 24, but instead may be supplied to the entropy encoder 26 for encoding as part of the bitstream along with the quantized coefficients.
[0041] Those ordinarily skilled in the art will appreciate the details and possible variations for implementing H.264 encoders.
[0042] The decoder 50 includes an entropy decoder 52, dequantizer 54, inverse transform processor 56, spatial compensator 57, and deblocking processor 60. A
frame buffer 58 supplies reconstructed frames for use by a motion compensator 62 in applying motion compensation. The spatial compensator 57 represents the operation of recovering the video data for a particular intra-coded block from a previously decoded block.
[0043] The bitstream 14 is received and decoded by the entropy decoder 52 to recover the quantized coefficients. Side information may also be recovered during the entropy decoding process, some of which may be supplied to the motion compensation loop for use in Our 101-0092CAP1 RIM 43609-CA-PAT
- 9 -motion compensation, if applicable. For example, the entropy decoder 52 may recover motion vectors and/or reference frame information for inter-coded macroblocks.
[0044] The quantized coefficients are then dequantized by the dequantizer 54 to produce the transform domain coefficients, which are then subjected to an inverse transform by the inverse transform processor 56 to recreate the "video data". It will be appreciated that, in some cases, such as with an intra-coded macroblock, the recreated "video data" is the residual data for use in spatial compensation relative to a previously decoded block within the frame. The spatial compensator 57 generates the video data from the residual data and pixel data from a previously decoded block. In other cases, such as inter-coded macroblocks, the recreated "video data" from the inverse transform processor 56 is the residual data for use in motion compensation relative to a reference block from a different frame. Both spatial and motion compensation may be referred to herein as "prediction operations".
[0045] The motion compensator 62 locates a reference block within the frame buffer 58 specified for a particular inter-coded macroblock. It does so based on the reference frame information and motion vector specified for the inter-coded macroblock. It then supplies the reference block pixel data for combination with the residual data to arrive at the reconstructed video data for that macroblock.
[0046] A deblocking process may then be applied to a reconstructed frame/slice, as indicated by the deblocking processor 60. After deblocking, the frame/slice is output as the decoded video frame 16, for example for display on a display device. It will be understood that the video playback machine, such as a computer, set-top box, DVD or Blu-Ray player, and/or mobile handheld device, may buffer decoded frames in a memory prior to display on an output device.
[0047] It is expected that HEVC-compliant encoders and decoders will have many of these same or similar features.
[0048] In the description of example embodiments that follows, the following symbol definitions are used:
x = the original frame/block of pixels p = the predictive block of pixels cõ = the transform of x Our 10I-0092CAP I RIM 43609-CA-PAT
- 10 -Cp= the transform of p z= the pixel domain residual = x ¨ p cz= the transform of z, i.e. the transform domain coefficients u = the quantized transform domain coefficients = Q(c,), where Q() denotes the quantization function, and Q-1(=) denotes the inverse quantization function.
For example, for a uniform scalar quantization function, the output of Q(c) at position (i,j) is given by j)=sign(c,rfloor(abs(c,,, j)/q, j), where sign() denotes the sign function, floor() denotes the floor function, abs() denotes the absolute function, and q, j denotes the quantization step size at position (i,j).
= the reconstructed block of pixels = p +1-1(Q-1(u)).
[0049] The distortion in a reconstructed block of pixels is the difference between the original block of pixels, x, and its reconstruction at the decoder. Filtering is sometimes applied to the reconstruction in order to reduce distortion. From a signal processing point of view, this may be categorized as low pass filtering. Such filtering generally involves a convolution operation in the pixel domain. It will be appreciated that convolution operations are computationally demanding. Because of the feedback loop at the encoder this computational complexity is present in both the encoding and decoding operations.
[0050] In accordance with one aspect of the present application, the quantization distortion is filtered in the transform domain instead of in the pixel domain.
Based on having a spectral transform, such as DCT, with unitary property, the filtering normally done to minimize the difference in the pixel domain between the original block of pixels and the reconstructed block of pixels may be carried out in the transform domain before the reconstruction is inverse transformed to create the reconstructed block of pixels.
[0051] In the transform domain, the quantization distortion may be expressed as:
\ 12 CX ¨ CP ¨ (2-1(U)12 Ei I Czi ¨ Cmij ¨ (2-1 j) I (1) [0052] In the case where scalar quantization is used, Q1(u) = u * q, where *
represents element-wise multiplication between vectors/matrices.
Our 101-0092CAP I RIM 43609-CA-PAT

- I -[0053] Because of the unitary property of the transform operation, e.g. DCT, the quality loss in the pixel domain is equivalent to the distortion in the transform domain. In other words:
I cx cp (u)I2 = P 1-1(Q (u)) 12 (2) [0054] where T represents the inverse transform operation.
[0055] Convolution in the pixel domain is the equivalent of scaling/multiplication in the transform domain, meaning that a filter realized in the transform domain will be more computationally efficient to implement. Moreover, u*q is already available in the transform domain, thus further reducing the computational complexity.
[0056] The aim of a transform domain filter is to reduce quantization distortion. In order for a decoder to apply such a filter, the encoder will transmit the filter coefficients in the bitstream of encoded data. Accordingly, design of the filter at the encoder may be considered from a rate-distortion point of view. A rate-distortion expression for finding a desired filter for N blocks xl, x2, ..., xN is as follows:

mint. EZ=1 lexk ¨1* (cpk+ Q'(uk))I + A = r(f) (3-a) [0057] Equation (3-a) reflects a Lagrangian cost function for finding an optimal filter from a rate-distortion point of view. The first term represents the distortion. It will be noted that the distortion in this term is the difference between the transformed original block of samples and the filtered transform domain reconstruction. The cost of sending the filter coefficients is represented by the rate function r(f). k is the Lagrange multiplier. An alternative rate-distortion expression (the justification for which is set out below) is:
[0058] minf EZ=1 I cxk f * (cp,) ¨ Q-1(uk)I2 + A = r(f) (3-1)) [0059] The symbol fin Equation (3-a) represents the filter; more specifically, a set or block of filter coefficients in a filter matrix used to scale the transform domain reconstruction.
The filter coefficients may also be applied to the prediction, which may also lead to a rate distortion coding performance boost, as shown in (3-b). The filter coefficients may be referred to herein as transform-domain filter(TDF). Equations (3-a) and (3-b) provides some basis for selecting the filter coefficients. It will be recalled that the quantization vector q is a scaling Our 10I-0092CAP I RIM 43609-CA-PAT

factor. In some embodiments, the quantization vector q may be a vector/matrix specifying quantization values for each coefficient position in the block/matrix. An example is described in U.S. Patent Application No. 12/830,548, filed July 6, 2010, and entitled "Methods and Devices for Data Compression Using Adaptive Reconstruction Levels", owned in common -- herewith. In such a situation, in which the quantization scaling factor for each coefficient position will be signaled to the decoder, Equation (3) may be reformulated as:
minf Elivc=ilcxk ¨ f * (cm + uk * q)I2 + A = r(f) (4) [0060] Furthermore, it will be recalled that the quantization operation to calculate the quantized transform domain coefficients, uk, occurs after the prediction operation that -- produces the prediction Pk and, thus, the transformed prediction block cpk.
It has also been observed by the present inventors that some elements of cpk are negatively correlated with the elements of czk. Accordingly, Equation (4) may be simplified as:
minf EZ=licxk ¨ f * cPk 12 2 = r(f) (5) [0061] Equation (5) realizes a filter that reduces the residual energy. The residual in -- this case is the difference between the original block of samples x and the filtered predictive block of samples.
[0062] Equations (3) and (5) lead to two possible processes for generating a transform domain filter/. The process according to Equation (3) results in a filter for filtering the reconstruction in the transform domain. The process according to Equation (5) results in a -- filter for filtering the prediction in the transform domain. They will be described in turn below.
[0063] If the filter coefficients are calculated and signaled sufficiently infrequently, the rate aspect of Equation (3) may be negligible. Using this simplification, the expression for finding a desired filter may be restated as follows:
\ 12 mint. EL, Iczk ¨ f * (cpk + (2-1(uk))1 (6) [0064] The foregoing discussion is based on finding a filter for use in filtering a prediction block in the transform domain. The filter may be designed based upon the Our I01-0092CAP I RIM 43609-CA-PAT

statistics of a previous slice or the statistics of the current slice. In some cases, the filter may be designed based upon the statistics of more than one slice, picture, frame, GoP or other collection of data samples. It has been noted that the optimal filter varies depending upon a number of factors because of variation in the data statistics amongst transform units of various types. Relevant factors include the transform unit size, such as 4x4, 8x8, 16x16, 32x32, 64x64, etc.; the transform unit type, e.g. DST or DCT; the texture type, e.g.
luma or chroma;
and, the coding mode, e.g. inter-coding or intra-coding. In addition, data statistics can vary or be dependent upon motion vectors, prediction modes, etc. Accordingly, in some embodiments different filters may be developed for different permutations of one or more of these factors. For example, separate filters may be developed for 4x4 DST
blocks, 4x4 DCT
blocks, 8x8 DCT blocks, 16x16 DCT blocks, and 32x32 DCT blocks. For each of those groups or categories of filter, multiple filters may be developed based on permutations of intra or inter coding mode and luma and chroma texture types. Yet further granularity in filter design may be found if within those groups multiple filters are developed based on groups of motion vectors and/or groups of intra-prediction modes. It will be appreciated that in such embodiments, a number of filters may be developed and communicated to the decoder, meaning that the rate aspect of the rate-distortion calculation in filter design is non-negligible.
[0065] With a set of inter-coded blocks, the transform coefficients may be grouped based upon motion prediction mode, motion vector magnitude, motion vector direction, or any combination thereof.
[0066] Within blocks, rather than developing filter coefficients for each individual coefficient position, some coefficient positions may be grouped based on statistical similarities. Those coefficient positions in the same group would share the same filter coefficient. Reference may be made to Figure 11, which illustrates one example embodiment of a grouping of coefficients for filtering a 4x4 block 700. In this example, the filter includes seven filter coefficients (indicated by numbers 0 through 6 within the block 700). The coefficient groups are indicated by the heavier lines, and the filter coefficient applicable to each position is indicated. Figure 12 shows another example embodiment of a grouping of coefficients for filtering a 4x4 block 702. In this example, the filter includes four filter coefficients. The grouping of coefficient positions to use a common filter coefficient may be Our 101-0092CAP1 RIM 43609-CA-PAT

applied to larger blocks. The specific pattern used to group coefficients may depend on statistical analysis of the coefficients.
[0067] In some cases, the pattern is similar to the patterns shown in Figures 11 or 12 but scaled up to the larger block size. For example, the pattern and the right and bottom edges may be extended to the larger size of the block. As an example, the 4x4 pattern shown in Figure 11 may be extended to 8x8 by including the filter coefficients in column 0, rows 4 to 7, in group number 5. The filter coefficients in row 0, columns 4-7, are in group 4. All remaining added filter coefficients are in group 6.
[0068] It will be appreciated that other patterns may also be used and that the selection of a suitable pattern for grouping coefficient positions in the filter design may be based upon empirical data statistics.
[0069] With intra-coding, filters may be developed that apply to specified groupings of direction prediction modes. For example, in the current HEVC development it is proposed that there will be 33 directional prediction modes plus a DC prediction for intra-coded blocks.
Figure 13 illustrates the intra-prediction angle definitions for HEVC.
[0070] Data may be grouped on the basis of angle-of-prediction direction. For example, the following intra-coding mode filter types may be defined:
Filter type = 0: angle of prediction direction: 45 deg ¨67.5 deg Filter type = 1: angle of prediction direction: 67.5 deg ¨ 112.5 deg Filter type = 2: angle of prediction direction: 112.5 deg¨ 157.5 deg Filter type = 3: angle of prediction direction: 157.5 deg ¨202.5 deg Filter type = 4: angle of prediction direction: 202.5 deg ¨225 deg Filter type = 5: DC prediction [0071] Other groupings on the basis of angle-of-prediction may also be used in other embodiments.
[0072] Reference is now made to Figure 3, which shows a simplified block diagram of an encoder 100 with adaptive transform domain filtering. The encoder 100 includes a prediction operator 102 for selecting or generating the predictive block of samples/pixels p.
The prediction operator 102 receives the original block of samples/pixels x, and may have access to a frame store 104 in which previously encoded and reconstructed frames are stored.
The prediction operator 102 may apply temporal or spatial prediction. In other words, the Our 101-0092CAP I RIM 43609-CA-PAT

prediction operator 102 may perform intra spatial prediction, or may apply inter motion estimation, depending on the frame type and coding mode, etc.
[0073] The original block of samples x and the predictive block of samples p are input to a spectral transform operator 106, such as a DCT. The spectral transform operator 106 produces the transformed original block of samples cx and the transformed predictive block of samples cp. The encoder 100 includes a residue calculator 108, which finds the transform domain residual cz as the difference between the transformed original block of samples cx and the transformed predictive block of samples cp. A quantizer 110 quantizes the transform domain residual cz to generate the quantized transform domain coefficients u.
[0074] The encoder 100 further includes a filter generator 112. The filter generator 112 receives the transformed original block of samples cx, the transformed predictive block of samples cp, the quantized transform domain coefficients u, and the quantization matrix/vector or scaling factor q. From this data the filter generator 112 produces the filter coefficients f, for example using Equation (3) or (6). It will be appreciated that in an embodiment in which multiple filters are developed for application to different categories of transform units (e.g.
based on permutations of size, texture, coding mode, etc.), the filter generator 112 categorizes the filter (filter coefficientsf) based upon these characteristics of the original block of samples cx.
[0075] The filter coefficients fare included in the bitstream output from the encoder.
In many embodiments, the filter coefficients may be entropy encoded and the encoded coefficients are inserted in the bitstream. In yet further embodiments, the filter coefficients may be quantized prior to entropy encoding. It will be understood that in that latter situation the decoder will need to decode the quantized coefficients and then dequantize them to realize reconstructed filter coefficients. It will also be appreciated that the quantization may introduce distortion and that a rate-distortion optimization process may be used in the quantization process to select the quantization step size for the filter coefficients.
[0076] The entropy encoding of the filter coefficients may rely upon any suitable encoding scheme. Examples include variable length coding (VLC), arithmetic coding, or variable-length-to-variable-length (V2V) coding.
[0077] The entropy encoding of the filter coefficients may be included in the beginning of the stream for a slice, e.g., in the slice header, or at the end of a slice, depending Our 101-0092CAP I RIM 43609-CA-PAT

on applications. When it is transmitted in the slice header, the coefficients are usually calculated based on some previously-coded slices. In this case, the filtering process may be applied during the mode selection. One drawback of doing this might be the data mismatch between previously-coded slices and the current slice. When the coefficients are transmitted at the end of a slice, they can be calculated based on the current slice. Yet, the filtering process cannot be included in the mode selection process, unless multi-pass coding is allowed.
[0078] One example implementation is now described with reference to Figure 4, which shows, in flowchart form, one example embodiment of the process 200 for data compression with adaptive transform domain filtering. The process 200 begins in operation 202 with receipt of the original block of samples/pixels x. A prediction operation is then used to find the predictive block of samples p in operation 204.
[0079] In operation 206, the encoder calculates the transform domain original cx, the transform domain prediction cp, and the transform domain residual coefficients cz. As noted above the residual may be found in the pixel domain and then transformed, or may be calculated as the difference between cx and cp.
[0080] The quantized transform domain coefficients u are generated in operation 208.
[0081] In operation 210, the encoder assesses whether it has reached the end of the coding unit and/or frame and/or picture and/or group-of-pictures (depending upon whether the filters are developed on a coding unit, frame, picture, or GOP basis). If not, then it returns to operation 202 to encode the next block of original samples. If so, then in operation 212 the encoder calculates the filter coefficients for the coding unit/frame/picture/etc. Operation 212 may include calculating multiple filters if the encoder is configured to create filters specific to predefined filter types. The coefficient positions are indicated using the indices i and j. Each filter coefficientA= may be calculated as a ratio of sums taken over the number of blocks n, where n is the number of blocks associated with that filter type in the coding unit/frame/picture/etc., as the case may be. In one embodiment, the numerator is the sum of values from the same ijth coefficient position in the transform domain original block cx. The denominator may be the sum of values from the same it coefficient position of the transform domain reconstruction, which is the transform domain predictive block cp plus the reconstructed transform domain residual given by Q-1(ul. The filter coefficients may, thus, be calculated in accordance with the expression:
Our 101-0092CAPI RIM 43609-CA-PAT

= Ci (7) EL(cpk,u+Q-1(uk,u)) [0082] In an alternative embodiment, the following expression may be used. The following expression is still an approximate solution to Equation (4), and is based on the assumption that the Lagrangian multiplier is zero, which may be acceptable if r(f) is small:
EZ=1 cxof(cPkji+Q-1(ukii)) (8) EZ.,1(cpk,u+Q-1(uk,u))2 [0083] The encoder may be configured to use a default filter coefficient in the case that the denominator sums to zero, so as to avoid a divide-by-zero problem.
The default filter coefficient may, in some embodiments, be zero. It will also be appreciated that filter coefficients are not necessarily calculated for all coefficient positions ij;
in some cases, the calculated filter coefficients will only include coefficients for a predetermined number of the coefficient positions, and the remaining positions use a default filter coefficient, for example, set to I. As an example, in the case of a 4x4 block, the filter coefficients may only be calculated for the first n coefficient positions in the scan order, where n may be preselected to be any suitable integer less than 16. Remaining positions may be assumed to be a preselected filter value or null filter value. It will be understood that the preselected or null filter values are not transmitted in the bitstream.
[0084] Although operation 212 is illustrated as occurring after the end of the coding unit/frame/picture/etc. has been reached for ease of illustration, it will be appreciated that the encoder may accumulate values throughout the encoding process of operations 202 to 210 for building the sums in the numerator and denominator, meaning that operation 212 is embedded in the looped process of operations 202 through 210. Moreover, the encoder may categorize the values and accumulate separate numerator/denominator values for the various filter types dependent upon whether it is processing a block of samples having those characteristics.
[0085] In operation 212, the encoder may evaluate whether a sufficient number of values for a particular filter type have been accumulated to justify designing a filter for that filter type. For example, the encoder may find that a frame/picture contains mostly 4x4 DCT
Our 101-0092CAP1 RIM 43609-CA-PAT

inter-coded and 8x8 DCT inter-coded blocks, thereby justifying the development of filters for these filter types, but has only one or a few 16x16 or 32x32 DCT inter-coded blocks. Using preset threshold values, the encoder may determine for which filter types it will create a filter.
The encoder may also use a rate-distortion analysis to determine whether to create a filter for a particular filter type.
[0086] In operation 214, the encoder, using an entropy encoding scheme, entropy encodes the generated quantized transform domain coefficients u and the filter coefficients!
to create an output bitstream of encoded data. In the case of multiple filters, it also signals to the decoder which filters are encoded in the bitstream. It will be understood that other data is also entropy encoded in the bitstream in certain embodiments, such as motion vectors, coding mode indices, quantization scaling matrices, etc. As noted above, the coding of the filter coefficients may use any suitable entropy encoding scheme, including VLC, V2V, or arithmetic coding. Also as noted above, the filter coefficients may, in some embodiments, be quantized prior to encoding. In some instances, rather than encoding the filter coefficients themselves, a prediction and residual are encoded. For example, the prediction may be based on the immediately preceding filter coefficients with the residual being the difference between those coefficients and the currently calculated filter coefficients. In some cases, the prediction operation may involve indexing a predetermined or previously calculated set of filter coefficients that are best match to the currently calculated filter coefficients, with the residual being the difference between the indexed coefficients and the current coefficients. More generally, the prediction is some function f(x), where x are the coefficients of the previous unit and y are the coefficients of the current unit. In such a case, the residual encoded is y-f(x). Example encoding algorithms for transmitting filter data are set out below.
[0087] The example encoding and decoding processes described below are applied to transform domain filter data in the examples. It will be appreciated that these encoding and decoding processes may alternatively be applied to pixel-domain filter data in some embodiments.
[0088] Reference will now be made to Figure 5, which shows another example embodiment of an encoder 300 configured for transform domain filtering. As with the encoder 100 of Figure 3, the encoder 300 includes a prediction operator 302, a frame store Our 101-0092CAP I RIM 43609-CA-PAT

304 and a spectral transform operator 306. The spectral transform operator 306 outputs the blocks of transform coefficients cx and the blocks of prediction transform coefficients cp.
[0089] The encoder 300 includes a filter generator 308 that receives the blocks of transform coefficients cx and the blocks of prediction transform coefficients cp. The filter generator 308 produces the filter coefficients in accordance with Equation (5). That is, the filter generator 308 finds filter coefficients that tend to minimize the difference between the blocks of transform coefficients c and the blocks of prediction transform coefficients cp.
100901 The encoder 300 further includes a residual calculator 310. In this example embodiment, the residuals are calculated based upon the filtered prediction.
That is, the residual calculator 310 finds the quantized transform domain coefficients u on the basis of:
u = Q (c, ¨(f* cp)) (9) [0091] One example implementation is now described with reference to Figure 6, which shows, in flowchart form, a further example embodiment of the process 400 for data compression with adaptive transform domain filtering. The process 400 begins in operation 402 with receipt of the original blocks of samples x. A prediction operation is then used to find the predictive block of samples p in operation 404.
[0092] In operation 406, the encoder calculates the blocks of transform domain coefficients c, and the blocks of prediction transform coefficients cp.
Operation 408 is the calculation of the filter coefficients. It will be appreciated that operations 402 to 406 may be repeated for all blocks of a coding unit or frame, as the case may be. It will also be appreciated that the calculation in operation 408 may be embedded with the loop of operations 402 to 406, such that the encoder sums values in the numerator and denominator as the values are calculated in operation 406.
[0093] Each it filter coefficient fy may be calculated as a ratio of sums taken over the number of blocks n, where n is the number of blocks in the coding unit or frame, as the case may be. In one embodiment, the numerator is the sum of values from the same 11th coefficient position in the blocks of transform coefficients cx. The denominator may be the sum of values from the same ijth coefficient position of the blocks of prediction transform coefficients cp.
Note that this calculation does not include the reconstructed residuals as part of determining Our 101 -0092CAP I RIM 43609-CA-PAT

the filter coefficients. The filter coefficients may, thus, be calculated in accordance with the expression:
= - cx k, (10IL]) Eicr=i(cPk,Ei) [0094] An alternative expression for finding the coefficients is set out below:
E'=1 cxk,t1=(cpk,,j) = __________________________________ ,2 (I 1) 41=1(CP0j) [0095] The encoder may be configured to use a default filter coefficient in the case that the denominator sums to zero, so as to avoid a divide-by-zero problem.
The default filter coefficient may, in some embodiments, be zero.
[0096] After the filter coefficientsf have been determined for the coding unit/frame, then in operation 410 the transform domain residue is determined. In particular the encoder calculates the quantized transform domain coefficients, u, using Equation (8), for example.
[0097] In operation 412, the quantized transform domain coefficients u, the filter coefficients f and other data is then entropy encoded to produce a bitstream of encoded data.
[0098] From the foregoing detailed discussion of examples, it will be understood that the determination and application of a filterfmay be on a frame basis, on a coding unit basis, or on the basis of any other collection of blocks, such as a picture, group of pictures, or slice.
[0099] To enable the decoder to start decoding as quickly as possible, the encoder may delay outputting the encoded bitstream by one frame (or coding unit, etc., as the case may be) so that the filter coefficientsf may be specified at the beginning of the encoded data for the frame/coding unit.
Transmission offilters [00100] In one example implementation, the filters may be transmitted or communicated from the encoder to the decoder in the slice header or in another header. The Our 101-0092CAP1 RIM 43609-CA-PAT

header may include a set of flags that signal which filter types are encoded in the bitstream.
The ordering of the flags may be predefined or may be specified by the encoder to the decoder in side information for the video. After the flags, the filter coefficients for the various filters are encoded in the bitstream. The filter coefficients may specify transform-domain filters or pixel-domain filters. Some of the examples below specifically reference transform-domain filtering, but may be applied to the transmission of filter data for pixel-domain filtering in other embodiments.
1001011 In one example embodiment, let TDF[qP][DataType][i] denote the filter coefficient for a given qP, given DataType, and a frequency group index i. The frequency group index i refers to the group of coefficient positions, for example as illustrated in Figures
11 and 12 for 4x4 blocks. TDF[qP][DataType][i] may be represented with 7-bit arithmetic precision. In a pixel-domain example, the group index i may refer to another grouping, such as by pixel position within a block.
[00102] An example process for transmitting TDF[qP][DataType][i] is as follows:
1001031 TDF[qP][DataType][i] is clipped to be within a range of [64, 1911 ([0.5, 1.5]
in the floating-point representation). TDF[qP][DataType][i] predicted with a fixed prediction value of 128. The prediction error is then defined as TDFd[qP][DataType][i] =
TDF[qP][DataType][i] ¨ 128, which is within a range of [-64, 63]. Experiments show that TDF[qP][DataType][i] is normally close to 1 and correspondingly 0 for TDFd [e][DataType][i]. TDFd[e][DataType][i] may be directly entropy-coded and transmitted as part of the slice header in the bitstream. Examples of entropy coding schemes for encoding TDFd include a fixed-length code, a Golomb-Rice code based on the probability distribution of the indices, and a Huffman code designed based on the probability distribution of the indices.
[00104] One example syntax for the bitstream is illustrated in the following pseudo-code for a function TDF_param() implemented at the decoder to obtain the filter coefficients:
TDF_param() ( Descriptor TDF_slice_flag u(1) If(TDF_slice_flag) for(i=0; i <=1µIUM_TDF_DATA_TYPES; i++) u(1) Our 101-0092CAP1 RIM 43609-CA-PAT

TDF_data_type _flag u(1) read_TDF_data(i) read_TDF_data (TDF_data type_id){
for(i=0: i<=6; i++){
TDFd[slice_qp][ TDF_data_type_id][i] se(v) [00105] In the above syntax, se(v) is the signed integer Exp-Golumb code. The value NUM_TDF_DATA_TYPES specifies the total number of prediction filter types, i.e.
the total number of filter-type flags (TDF_data_type_flag) to be read from the bitstream.
[00106] In another example embodiment, the filter coefficients may be encoded and decoded using the following scheme.
[00107] The filter coefficients are first predicted as TDFp[qP][DataType][i]. The prediction error is given by TDFd[qP][DataType][i] = TDF[qP][DataType][i] ¨
TDFp[qP][DataType][i]. The prediction error may be converted to an index value through quantization: TDF_delta_idx[qP][DataType][i] = Q(TDFd[qP][DataType][i] ). The syntax element TDF_delta_idx[qP][DataType][i] specifies the corresponding quantization index, which is entropy-coded and transmitted in the bitstream. Q(-) denotes the forward quantization function.
[00108] At the decoder, TDF is reconstructed from the corresponding received TDF_delta_idx as follows: TDF'[qP][DataType][i] = invQ( TDF_delta_idx[e][DataType][i] ) + TDFp[qP][DataType][i] , where TDF' denotes the reconstruction of TDF parameter, and invQ(.) denotes the inverse quantization function.
[00109] The prediction TDFp may be found in a number of possible ways.
One is to use a fixed default prediction value. For example, a default prediction value may be selected to be TDFp[qP][DataType][i] = 128. Note that different default prediction values may be used based on quantization parameter, data type, index, temporal layer of slice, etc. Another possibility is to base the prediction on previously-coded TDF values. For example, the Our 101-0092CAP I RIM 43609-CA-PAT

prediction of the filter coefficient for one frequency position based on the reconstructed filter coefficients of its neighboring frequency positions.
[00110] The quantization of TDFd may also be implemented in a number of ways. For example, in one case uniform quantization may be used. In uniform quantization, given a quantization step size q, the forward and inverse quantization of the prediction error TDFd are given by:
Forward quantization: TDF_delta_idx = Q(TDFd) = sign(TDFd) = round ( abs(TDFd) /q +f), where 1>150 is a rounding offset Inverse quantization: TDFd' = invQ(TDF_delta_idx) = TDF_delta_idx = q [00111] Non-uniform quantizers may also be designed for quantizing the prediction error TDFd based on its statistics.
[00112] Another example syntax for the bitstream is illustrated in the following pseudo-code for the function TDF_param() implemented at the decoder to obtain the filter coefficients based on encoding of the quantized prediction error:
TDF_param() ( Descriptor TDF_slice_flag u(1) If(TDF_slice_flag) for(i=0; i <=NUM_TDF_DATA_TYPES; i++) u(1) TDF_data_type_flag u(1) read_TDF_data(i) read_TDF_data (TDF_data_type_id)( for(i=0; i<=6; i++){
TDF_delta_idx[slice_qp][ TDF_data_type_id][i] vlc(v) Our 101-0092CAP I RIM 43609-CA-PAT

[00113] Reference is now made to Figure 7, which shows, in flowchart form, a process 500 for decoding an encoded bitstream of data using transform domain filtering. The process 500 includes receiving the bitstream in operation 502 and entropy decoding the bitstream in operation 504 to recover the quantized transform domain coefficients u, the filter coefficients f, and other data, such as data that enables the decoder to determine the prediction (for example, motion vector(s) and reference frame indices). The bitstream may be based upon one of the syntaxes discussed above. The entropy decoding of the bitstream may include reading the flags indicating which filter types are present and then, based on the flag values, entropy decoding the filter coefficients for each filter type that is present.
[00114] In operation 506, the decoder calculates the filtered transform domain reconstruction, ci , by calculating the transform domain reconstruction as the transform domain prediction block plus the dequantized transform domain residual, and then applying the filter f The calculation can be expressed as:
= f * (cp + (2-1 (u)) (12) [00115] It will be appreciated that operation 506 may include the decoder identifying or generating the predictive block of samples based on the decoded prediction data. For example, the decoder may identify the reference block from a reference frame based on motion vector information. Operation 506 may further include spectrally transforming the predictive block of samples to obtain the transform domain prediction block cp. Equation (12) may then be used to realize the transform domain reconstructed block.
[00116] The decoder then, in operation 508, inverse transforms the transform domain reconstruction to produce the reconstructed pixel domain block .
[00117] Another example decoding process 600 will now be described with reference to Figure 8, which shows the process 600 in flowchart form. The process 600 in Figure 8 is applicable to embodiments in which the residual was calculated from the filtered prediction at the encoder, as was described above in connection with Figures 5 and 6. The process 600 may also be applicable to cases in which the residual was calculated from the unfiltered prediction, like in Figures 3 and 4, but where the encoder transmits a quantization matrix q that specifies coefficient-position based quantization step sizes. In this case, the scaling Our 101-0092CAP I RIM 43609-CA-PAT

represented by the filter matrix f may be incorporated in the quantization matrix q, meaning that it need not be applied to the reconstructed transform domain residuals at the decoder.
[00118] In operations 602 and 604 the decoder receives the bitstream and entropy decodes it to recover the quantized transform domain coefficients u, the quantization step size (or quantization vector or matrix q, in some embodiments), the filter coefficientsf, and prediction data (such as motion vector(s), etc.). In operation 606, the decoder calculates the filtered transform domain reconstruction, ci , by calculating the sum of the filtered transform domain prediction block and the dequantized transform domain residual. The calculation may be expressed as:
f * cp + (2-1(u) (13) [00119] As detailed above in connection with operation 506, operation 606 may include the decoder generating or identifying the predictive block of samples and spectrally transforming the predictive block of samples to obtain the transform domain prediction block 17*
[00120] In one embodiment, the filter or TDF in Equations (12) or (13) may include a single scalar applicable to a group of transform coefficients, i.e for linearly scaling each of the coefficients in the group. The group of coefficients may be formed based on various coding parameters such as transform sizes, transform types, frequency positions, prediction modes, and data textures.
[00121] The decoder then inverse transform the filtered transform domain reconstruction to obtain the reconstructed block of pixels .X .
[00122] Although the processed described above and illustrated in Figures 8 and 9 relate to the transmission and use of transform-domain filter data, it will be appreciated that similar processes may be used to transmit and use pixel-domain filter data.
For example, the decoder may receive and decode the bitstream to obtain indices u, filter coefficients f, and prediction data. The decoder may then reconstruct pixel-domain residual blocks by dequantizing the indices and inverse transforming the dequantized transform coefficients. In the pixel-domain, the decoder may then filter the prediction blocks using a filter specified by the filter coefficientsf or it may reconstruct the pixel data by combining the residual block with the predicted block and then filter the resulting reconstructed pixel block. In either case, Our I 0 I -0092CAP I RIM 43609-CA-PAT

it will be understood that the filter coefficients fmay be specified in the bitstream using indices, predictions, prediction error, quantization, or combinations thereof, as described above.
[00123] Reference is now made to Figure 9, which shows a simplified block diagram of an example embodiment of an encoder 900. The encoder 900 includes a processor 902, memory 904, and an encoding application 906. The encoding application 906 may include a computer program or application stored in memory 904 and containing instructions for configuring the processor 902 to perform steps or operations such as those described herein.
For example, the encoding application 906 may encode and output bitstreams encoded in accordance with the processes described herein. The input data points may relate to audio, images, video, or other data that may be subject of a lossy data compression scheme. The encoding application 906 may include a filter generation module 908 configured to determine filter coefficients for a frame, coding unit, or other collection of blocks.
The encoding application 906 may include an entropy encoder configured to entropy encode the filter coefficients/ and other data as part of the bitstream. It will be understood that the encoding application 906 may be stored in on a computer readable medium, such as a compact disc, flash memory device, random access memory, hard drive, etc.
[00124] Reference is now also made to Figure 10, which shows a simplified block diagram of an example embodiment of a decoder 1000. The decoder 1000 includes a processor 1002, a memory 1004, and a decoding application 1006. The decoding application 1006 may include a computer program or application stored in memory 1004 and containing instructions for configuring the processor 1002 to perform steps or operations such as those described herein. The decoding application 1006 may include an entropy decoder 1008 and a reconstruction module 1010 configured to obtain filter coefficients/ and to use them to calculate a filtered transform domain reconstruction, which the decoding application 1006 is configured to inverse transform to produce the reconstructed block of pixels.
It will be understood that the decoding application 1006 may be stored in on a computer readable medium, such as a compact disc, flash memory device, random access memory, hard drive, etc.
[00125] It will be appreciated that the decoder and/or encoder according to the present application may be implemented in a number of computing devices, including, without Our 101-0092CAP I RIM 43609-CA-PAT

limitation, servers, suitably programmed general purpose computers, audio/video encoding and playback devices, set-top television boxes, television broadcast equipment, and mobile devices. The decoder or encoder may be implemented by way of software containing instructions for configuring a processor to carry out the functions described herein. The software instructions may be stored on any suitable computer-readable memory, including CDs, RAM, ROM, Flash memory, etc.
[00126] It will be understood that the encoder described herein and the module, routine, process, thread, or other software component implementing the described method/process for configuring the encoder may be realized using standard computer programming techniques and languages. The present application is not limited to particular processors, computer languages, computer programming conventions, data structures, other such implementation details. Those skilled in the art will recognize that the described processes may be implemented as a part of computer-executable code stored in volatile or non-volatile memory, as part of an application-specific integrated chip (ASIC), etc.
[00127] Certain adaptations and modifications of the described embodiments can be made. Therefore, the above discussed embodiments are considered to be illustrative and not restrictive.
Our I01-0092CAP I RIM 43609-CA-PAT

Claims (32)

WHAT IS CLAIMED IS:
1. A method of reconstructing a block of samples from a bitstream of encoded data in a decoder, the method comprising:
reconstructing a spectrally transformed prediction block of samples associated with the block of samples;
reading filter-type flags from the bitstream, wherein the filter-type flags identify one or more filter types;
for each of the filter types identified, decoding a set of filter coefficients from the bitstream; and calculating a filtered transform domain reconstruction by applying, to at least a portion of the transformed prediction block, one of the sets of filter coefficients based on its filter type matching characteristics associated with the block of samples.
2. The method claimed in claim 1, wherein calculating includes determining the characteristics of the block of samples and selecting the one of the sets of filter coefficients based on its filter type matching the characteristics.
3. The method claimed in claim 1 or claim 2, wherein the filter types are based on at least one of transform size, coding mode, prediction mode, and texture type.
4. The method claimed in any one of claims 1 to 3, wherein the characteristics include motion vector characteristics, and wherein the filter types include two or more filter types applicable to respective groups of inter-coded blocks, and wherein each respective group of inter-coded blocks is associated with a respective group of motion vector characteristics.
5. The method claimed in any one of claims 1 to 3, wherein the characteristics include directional prediction modes, and wherein the filter types include two or more filter types applicable to respective groups of intra-coded blocks, and wherein each respective group of intra-coded blocks is associated with a group of directional prediction modes.
6. The method claimed in any one of claims 1 to 5, wherein the set of filter coefficients defines a filter, wherein the set of filter coefficients includes a coefficient for each coefficient group in the filter, and wherein at least one coefficient group includes two or more coefficient positions.
7. The method claimed in any one of claims 1 to 6, wherein calculating the filtered transform domain reconstruction comprises calculating a filtered prediction by scaling at least a portion of the transformed prediction block by the filter coefficients and dequantizing a block of quantized transform domain coefficients, and wherein the filtered transform domain reconstruction comprises the sum of the filtered prediction and the dequantized block of transform domain coefficients.
8. The method claimed in any one of claims 1 to 6, wherein calculating the filtered transform domain reconstruction comprises:
dequantizing a block of quantized transform domain coefficients;
calculating a transform domain reconstruction block as the transformed prediction block plus the dequantized block of transform domain coefficients; and scaling the transform domain reconstruction block using the filter coefficients.
9. A method of reconstructing a block of samples from a bitstream of encoded data in a decoder, the method comprising:
reconstructing a prediction block of samples associated with the block of samples;
reading filter-type flags from the bitstream, wherein the filter-type flags identify one or more filter types;
for each of the filter types identified, decoding a set of filter coefficients from the bitstream; and calculating a filtered pixel-domain reconstruction by applying, to at least a portion of the prediction block, one of the sets of filter coefficients based on its filter type matching characteristics associated with the block of samples.
10. A decoder, comprising:
a processor;
a memory; and a decoding application stored in memory and containing instructions for configuring the processor to perform the method claimed in any one of claims 1 to 9.
11. A non-transitory computer-readable medium storing computer-executable instructions which, when executed by a processor, configure the processor to perform the method claimed in any one of claims 1 to 9.
12. A method of encoding a block of samples in an encoder, the method comprising:
determining a prediction block for predicting the block of samples;
spectrally transforming the block of samples to generate a transformed original block of samples;
spectrally transforming the prediction block to generate a transformed prediction block;
determining a set of filter coefficients based the transformed prediction block and the transformed original block of samples; and encoding the set of filter coefficients and a quantized transformed residual block in a bitstream of encoded data, whereby the set of filter coefficients are used for transform domain filtering during reconstruction of the block of samples.
13. The method claimed in claim 12, wherein the set of filter coefficients defines a filter, wherein the set of filter coefficients includes a coefficient for each coefficient group in the filter, and wherein at least one coefficient group includes two or more coefficient positions.
14. The method claimed in claim 12, wherein determining a set of filter coefficients includes determining the set of filter coefficients based upon a sum of transformed prediction blocks and a sum of transformed original blocks of samples from a plurality of blocks of samples.
15. The method claimed in claim 14, wherein the plurality of blocks of samples form one of a coding unit, a frame, a picture, or a group of pictures.
16. The method claimed in claim 14, wherein the set of filter coefficients is determined as the minimum of a sum of the absolute value squared of the difference between the transformed original blocks of samples and a product, wherein the product comprises the filter coefficients times the sum of the transformed prediction blocks and dequantized quantized transformed residual blocks.
17. The method claimed in claim 14, wherein the set of filter coefficients is determined as the ratio of the sum of the transformed original blocks of samples to the sum of the transformed prediction blocks.
18. The method claimed in claim 14, wherein the set of filter coefficients is determined as the ratio of:
the sum of the transformed original blocks of samples to the sum of the transformed prediction blocks and dequantized quantized transformed residual blocks.
19. The method claimed in claim 12, further comprising:
determining a residual block as the difference between the block of samples and the prediction block;
transforming the residual block to generate a transformed residual block; and quantizing the transformed residual block to create the quantized transformed residual block.
20. The method claimed in claim 12, further comprising:
determining a transformed residual block as the difference between the transformed original block of samples and the transformed prediction block; and quantizing the transformed residual block to create the quantized transformed residual block.
21. The method claimed in claim 12, further comprising filtering the transformed prediction bock using the set of filter coefficients to create a filtered transformed prediction block, determining a transformed residual block based on the transformed original block and the filtered transformed prediction block, and quantizing the transformed residual block to generate the quantized transformed residual block.
22. An encoder, comprising:
memory storing a block of samples to be encoded;
a processor;
an encoding application containing processor-executable instructions which, when executed by the processor cause the processor to:
determine a prediction block for predicting the block of samples, spectrally transform the block of samples to generate a transformed original block of samples, spectrally transform the prediction block to generate a transformed prediction block, determine a set of filter coefficients based the transformed prediction block and the transformed original block of samples, and encode the set of filter coefficients and a quantized transformed residual block in a bitstream of encoded data, whereby the set of filter coefficients are used for transform domain filtering during reconstruction of the block of samples.
23. The encoder claimed in claim 22, wherein the set of filter coefficients defines a filter, wherein the set of filter coefficients includes a coefficient for each coefficient group in the filter, and wherein at least one coefficient group includes two or more coefficient positions.
24. The encoder claimed in claim 22, wherein the processor determines a set of filter coefficients by determining the set of filter coefficients based upon a sum of transformed prediction blocks and a sum of transformed original blocks of samples from a plurality of blocks of samples.
25. The encoder claimed in claim 24, wherein the plurality of blocks of samples form one of a coding unit, a frame, a picture, or a group of pictures.
26. The encoder claimed in claim 24, wherein the set of filter coefficients is determined as the minimum of a sum of the absolute value squared of the difference between the transformed original blocks of samples and a product, wherein the product comprises the filter coefficients times the sum of the transformed prediction blocks and dequantized quantized transformed residual blocks.
27. The encoder claimed in claim 24, wherein the set of filter coefficients is determined as the ratio of the sum of the transformed original blocks of samples to the sum of the transformed prediction blocks.
28. The encoder claimed in claim 24, wherein the set of filter coefficients is determined as the ratio of:
the sum of the transformed original blocks of samples to the sum of the transformed prediction blocks and dequantized quantized transformed residual blocks.
29. The encoder claimed in claim 24, wherein the instructions, when executed further cause the processor to:
determine a residual block as the difference between the block of samples and the prediction block;
transform the residual block to generate a transformed residual block; and quantize the transformed residual block to create the quantized transformed residual block.
30. The encoder claimed in claim 24, wherein the instructions, when executed, further cause the processor to:
determine a transformed residual block as the difference between the transformed original block of samples and the transformed prediction block; and quantize the transformed residual block to create the quantized transformed residual block.
31. The encoder claimed in claim 24, wherein the instructions, when executed, further cause the processor to:
filter the transformed prediction bock using the set of filter coefficients to create a filtered transformed prediction block;
determine a transformed residual block based on the transformed original block and the filtered transformed prediction block; and quantize the transformed residual block to generate the quantized transformed residual block.
32. A non-transitory computer-readable medium storing computer-executable instructions for encoding a block of samples, wherein the instructions, when executed by a processor, cause the processor to:
determine a prediction block for predicting the block of samples, spectrally transform the block of samples to generate a transformed original block of samples, spectrally transform the prediction block to generate a transformed prediction block, determine a set of filter coefficients based the transformed prediction block and the transformed original block of samples, and encode the set of filter coefficients and a quantized transformed residual block in a bitstream of encoded data, whereby the set of filter coefficients are used for transform domain filtering during reconstruction of the block of samples.
CA2796778A 2011-11-21 2012-11-20 Methods and devices for encoding and decoding transform domain filters Active CA2796778C (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201161562294P 2011-11-21 2011-11-21
US61/562,294 2011-11-21

Publications (2)

Publication Number Publication Date
CA2796778A1 CA2796778A1 (en) 2013-05-21
CA2796778C true CA2796778C (en) 2016-12-13

Family

ID=47630077

Family Applications (1)

Application Number Title Priority Date Filing Date
CA2796778A Active CA2796778C (en) 2011-11-21 2012-11-20 Methods and devices for encoding and decoding transform domain filters

Country Status (4)

Country Link
US (2) US8942495B2 (en)
EP (1) EP2595382B1 (en)
KR (1) KR101506372B1 (en)
CA (1) CA2796778C (en)

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2595382B1 (en) * 2011-11-21 2019-01-09 BlackBerry Limited Methods and devices for encoding and decoding transform domain filters
KR102061201B1 (en) * 2012-04-12 2019-12-31 주식회사 골드피크이노베이션즈 Methods of transformation based on block information and appratuses using the same
US9756346B2 (en) 2012-10-08 2017-09-05 Google Inc. Edge-selective intra coding
US9407915B2 (en) * 2012-10-08 2016-08-02 Google Inc. Lossless video coding with sub-frame level optimal quantization values
WO2015015436A2 (en) * 2013-07-30 2015-02-05 Riversilica Technologies Pvt Ltd System and method for performing optimal temporal predictive mode decision in h.264 video coding
US10116933B2 (en) * 2013-10-14 2018-10-30 Mediatek Inc. Method of lossless mode signaling for video system with lossless and lossy coding
JP6730191B2 (en) * 2013-12-22 2020-07-29 エルジー エレクトロニクス インコーポレイティド Method and apparatus for predicting video signal using predicted signal and transform-coded signal
US20150365703A1 (en) * 2014-06-13 2015-12-17 Atul Puri System and method for highly content adaptive quality restoration filtering for video coding
CN107079171B (en) * 2014-10-01 2021-01-29 Lg 电子株式会社 Method and apparatus for encoding and decoding video signal using improved prediction filter
CN105812950A (en) * 2014-12-31 2016-07-27 鸿富锦精密工业(深圳)有限公司 Multimedia device and video communication method
US11076153B2 (en) * 2015-07-31 2021-07-27 Stc.Unm System and methods for joint and adaptive control of rate, quality, and computational complexity for video coding and video delivery
CN108353175B (en) * 2015-09-21 2021-03-12 Lg 电子株式会社 Method and apparatus for processing video signal using coefficient-induced prediction
KR20180068334A (en) 2015-11-12 2018-06-21 엘지전자 주식회사 Method and apparatus for intra prediction in a video coding system
FR3049084B1 (en) * 2016-03-15 2022-11-11 Fraunhofer Ges Forschung CODING DEVICE FOR PROCESSING AN INPUT SIGNAL AND DECODING DEVICE FOR PROCESSING A CODED SIGNAL
KR101930389B1 (en) * 2016-06-15 2019-01-02 주식회사 에벤에셀케이 Video File Compression Method, Device and Computer Program Thereof
US10742979B2 (en) * 2016-12-21 2020-08-11 Arris Enterprises Llc Nonlinear local activity for adaptive quantization
WO2020055288A2 (en) * 2018-09-16 2020-03-19 Huawei Technologies Co., Ltd. Apparatus and method for filtering in video coding with look-up table selected based on bitstream information
US10769819B2 (en) * 2018-09-24 2020-09-08 Qualcomm Incorporated Floating-point data compression
CN114287135A (en) 2019-08-23 2022-04-05 北京字节跳动网络技术有限公司 Cropping in reference picture resampling
KR20220082839A (en) * 2019-10-23 2022-06-17 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 Calculations for multiple coding tools
KR20220080107A (en) 2019-10-23 2022-06-14 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 Signaling for reference picture resampling

Family Cites Families (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3298915B2 (en) 1991-12-28 2002-07-08 ソニー株式会社 Encoding device
JP3781194B2 (en) 1995-10-20 2006-05-31 ノキア コーポレイション Motion vector field coding
DE69623342T2 (en) 1996-04-19 2003-04-17 Nokia Corp VIDEO ENCODER AND DECODER WITH MOTION-BASED IMAGE SEGMENTATION AND IMAGE ASSEMBLY
GB2317525B (en) 1996-09-20 2000-11-08 Nokia Mobile Phones Ltd A video coding system
FI106071B (en) 1997-03-13 2000-11-15 Nokia Mobile Phones Ltd Adaptive filter
GB2343319B (en) 1998-10-27 2003-02-26 Nokia Mobile Phones Ltd Video coding
US6735249B1 (en) 1999-08-11 2004-05-11 Nokia Corporation Apparatus, and associated method, for forming a compressed motion vector field utilizing predictive motion coding
JP4596718B2 (en) 1999-08-11 2010-12-15 ノキア コーポレイション Adaptive motion vector field coding
US6738423B1 (en) 2000-01-21 2004-05-18 Nokia Mobile Phones Ltd. Method for encoding and decoding video information, a motion compensated video encoder and a corresponding decoder
FI116819B (en) 2000-01-21 2006-02-28 Nokia Corp Procedure for transferring images and an image encoder
EP1279293A1 (en) 2000-01-21 2003-01-29 Nokia Corporation A motion estimation method and a system for a video coder
FI117534B (en) 2000-01-21 2006-11-15 Nokia Corp A method for filtering digital images and a filter
US6741648B2 (en) 2000-11-10 2004-05-25 Nokia Corporation Apparatus, and associated method, for selecting an encoding rate by which to encode video frames of a video sequence
US20020122491A1 (en) 2001-01-03 2002-09-05 Marta Karczewicz Video decoder architecture and method for using same
US6920175B2 (en) 2001-01-03 2005-07-19 Nokia Corporation Video coding architecture and methods for using same
US6765963B2 (en) 2001-01-03 2004-07-20 Nokia Corporation Video decoder architecture and method for using same
US6696993B2 (en) 2001-03-23 2004-02-24 Nokia Corporation Variable length coding
US6856701B2 (en) 2001-09-14 2005-02-15 Nokia Corporation Method and system for context-based adaptive binary arithmetic coding
US6950469B2 (en) 2001-09-17 2005-09-27 Nokia Corporation Method for sub-pixel value interpolation
US6690307B2 (en) 2002-01-22 2004-02-10 Nokia Corporation Adaptive variable length coding of digital video
US7289674B2 (en) 2002-06-11 2007-10-30 Nokia Corporation Spatial prediction based intra coding
US6795584B2 (en) 2002-10-03 2004-09-21 Nokia Corporation Context-based adaptive variable length coding for adaptive block transforms
US7664176B2 (en) 2004-07-09 2010-02-16 Nokia Corporation Method and system for entropy decoding for scalable video bit stream
US7570827B2 (en) 2004-07-14 2009-08-04 Slipstream Data Inc. Method, system and computer program product for optimization of data compression with cost function
US7336837B2 (en) 2005-01-11 2008-02-26 Nokia Corporation Method and system for coding/decoding of a video bit stream for fine granularity scalability
JP4146444B2 (en) 2005-03-16 2008-09-10 株式会社東芝 Video encoding method and apparatus
US7756206B2 (en) 2005-04-13 2010-07-13 Nokia Corporation FGS identification in scalable video coding
US20070147497A1 (en) 2005-07-21 2007-06-28 Nokia Corporation System and method for progressive quantization for scalable image and video coding
JP4756949B2 (en) 2005-08-08 2011-08-24 キヤノン株式会社 Image decoding apparatus, control method therefor, computer program, and computer-readable storage medium
US8005140B2 (en) 2006-03-17 2011-08-23 Research In Motion Limited Soft decision and iterative video coding for MPEG and H.264
EP2082580A4 (en) * 2006-11-17 2010-01-13 Lg Electronics Inc Method and apparatus for decoding/encoding a video signal
JP2010525658A (en) 2007-04-19 2010-07-22 トムソン ライセンシング Adaptive reference image data generation for intra prediction
EP2048886A1 (en) 2007-10-11 2009-04-15 Panasonic Corporation Coding of adaptive interpolation filter coefficients
US8576906B2 (en) * 2008-01-08 2013-11-05 Telefonaktiebolaget L M Ericsson (Publ) Adaptive filtering
US9967590B2 (en) 2008-04-10 2018-05-08 Qualcomm Incorporated Rate-distortion defined interpolation for video coding based on fixed filter or adaptive filter
US10123050B2 (en) 2008-07-11 2018-11-06 Qualcomm Incorporated Filtering video data using a plurality of filters
KR101518237B1 (en) 2008-09-01 2015-05-15 삼성전자주식회사 Method and apparatus for inverse quantization, and method and apparatus for decoding of image
EP2192786A1 (en) * 2008-11-27 2010-06-02 Panasonic Corporation Frequency domain filters for video coding
US8577159B2 (en) * 2010-09-14 2013-11-05 Blackberry Limited Methods and devices for data compression with adaptive filtering in the transform domain
EP2595382B1 (en) * 2011-11-21 2019-01-09 BlackBerry Limited Methods and devices for encoding and decoding transform domain filters

Also Published As

Publication number Publication date
CA2796778A1 (en) 2013-05-21
KR20130056190A (en) 2013-05-29
US8942495B2 (en) 2015-01-27
KR101506372B1 (en) 2015-03-26
US20150110408A1 (en) 2015-04-23
EP2595382B1 (en) 2019-01-09
US9264722B2 (en) 2016-02-16
EP2595382A2 (en) 2013-05-22
US20130129241A1 (en) 2013-05-23
EP2595382A3 (en) 2014-08-20

Similar Documents

Publication Publication Date Title
CA2796778C (en) Methods and devices for encoding and decoding transform domain filters
US9282329B2 (en) Methods and devices for data compression using offset-based adaptive reconstruction levels
CA2808451C (en) Methods and devices for data compression with adaptive filtering in the transform domain
EP2396770B1 (en) Adaptive quantization with balanced pixel-domain distortion distribution in image processing
US10178391B2 (en) Methods and devices for data compression using a non-uniform reconstruction space
CN107396116B (en) Video encoding and decoding apparatus and non-transitory computer-readable storage medium
EP2252059B1 (en) Image encoding and decoding method and device
US8582639B2 (en) Methods and devices for data compression using adaptive reconstruction levels
EP2575364A1 (en) Methods and devices for data compression using a non-uniform reconstruction space
EP3151560B1 (en) Intra-coding mode-dependent quantization tuning
CA2803202C (en) Methods and devices for data compression using adaptive reconstruction levels
JP6497100B2 (en) Video encoding apparatus, method and program