[go: up one dir, main page]

WO2011145328A1 - Procédé de décodage d'images, procédé de codage d'images, dispositif de décodage d'images, et dispositif de codage d'images - Google Patents

Procédé de décodage d'images, procédé de codage d'images, dispositif de décodage d'images, et dispositif de codage d'images Download PDF

Info

Publication number
WO2011145328A1
WO2011145328A1 PCT/JP2011/002723 JP2011002723W WO2011145328A1 WO 2011145328 A1 WO2011145328 A1 WO 2011145328A1 JP 2011002723 W JP2011002723 W JP 2011002723W WO 2011145328 A1 WO2011145328 A1 WO 2011145328A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
angle
edge vector
unit
transformation matrix
Prior art date
Application number
PCT/JP2011/002723
Other languages
English (en)
Japanese (ja)
Inventor
陽司 柴原
西 孝啓
寿郎 笹井
ステファン ウィットマン
Original Assignee
パナソニック株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by パナソニック株式会社 filed Critical パナソニック株式会社
Publication of WO2011145328A1 publication Critical patent/WO2011145328A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/14Coding unit complexity, e.g. amount of activity or edge presence estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/12Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
    • H04N19/122Selection of transform size, e.g. 8x8 or 2x4x8 DCT; Selection of sub-band transforms of varying structure or type
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

Definitions

  • the present invention relates to an image decoding method and an image decoding apparatus that involve inverse transformation from an image frequency domain to a spatial domain, and an image coding method and an image coding apparatus that involve an image transformation from a spatial domain to a frequency domain.
  • H.264 ITU-T As an example of the video coding standard, H.264 ITU-T standard called 26x and ISO / IEC standard called MPEG-x.
  • MPEG-x As an example of the video coding standard, H.264 ITU-T standard called 26x and ISO / IEC standard called MPEG-x.
  • the latest video coding standard is H.264. H.264 / MPEG-4AVC (see, for example, Non-Patent Document 1).
  • Such an image encoding device conforming to the moving image encoding standard includes a conversion unit, a quantization unit, and an entropy encoding unit in order to encode image data at a low bit rate.
  • the conversion unit outputs a plurality of frequency coefficients with reduced correlation by converting the image data from the spatial domain to the frequency domain.
  • the quantization unit outputs a plurality of quantization coefficients having a small total data amount by quantizing the plurality of frequency coefficients output from the conversion unit.
  • the entropy encoding unit outputs an encoded signal (encoded image) obtained by compressing the image data by encoding the plurality of quantization coefficients output from the quantization unit using an entropy encoding algorithm.
  • the conversion unit obtains a conversion input vector x n which is a vector (N-dimensional signal) having N points of elements as image data to be converted. Then, the conversion unit, as shown in (Equation 1), by performing a transformation T with respect to its inverting input vectors x n, and outputs the converted output (Transform Output) vector y n.
  • the transformation T is expressed by a matrix product of a transformation matrix A of an N ⁇ N matrix and a transformation input vector x n as shown in (Formula 2). Therefore, the elements yi of converting the output vector y n, using the elements a ik of the transformation matrix A, is expressed as shown in (Equation 3).
  • the transformation matrix A is designed so that the correlation of the transformation input vector (input signal) is reduced, and energy is concentrated on the lower dimension side in the transformation output vector (output signal).
  • KLT Kerhunen Loeve Transform
  • KLT is a method for deriving an optimum transform coefficient based on the statistical properties of an input signal, or a transform method using the derived optimum transform coefficient.
  • KLT the basis (base of conversion coefficient A) is designed based on statistical properties. Also, such KLT is known as a conversion that greatly eliminates the correlation of input signals and can concentrate energy to the low frequency (low dimension) side very efficiently. Conversion efficiency (conversion performance, objective performance) Or objective image quality) is high. Therefore, encoding efficiency can be improved by using this KLT.
  • the prediction direction may not match the direction in which the correlation is generated. In such a case, it is not optimal to switch the transformation matrix according to the prediction mode of in-plane prediction.
  • An object is to provide a decoding method, an image decoding device, an image encoding method, and an image encoding device.
  • an image decoding method for decoding an encoded image for each block, and includes a decoded block spatially adjacent to a processing target block; An edge vector detection step for detecting an edge vector in at least one of the decoded blocks that are temporally adjacent to the processing target block, and based on the detected angle of the edge vector, the processing target block An inverse transform matrix determining step for determining an inverse transform matrix of the block, an inverse quantization step for inversely quantizing an encoded difference block corresponding to the block to be processed, and an inverse quantization using the determined inverse transform matrix An inverse transform step of generating a difference block by inversely transforming the encoded difference block, and the processing target Comprising a prediction step of generating a prediction block by predicting the lock, and an adding step of generating a decoded block by adding said prediction block and the difference block.
  • the inverse transformation matrix can be determined based on the angle of the edge vector of the decoded block that is close in time or space to the target block. That is, the inverse transformation can be performed using an inverse transformation matrix determined based on the angle of the edge vector that can be detected in smaller units. Therefore, in an encoded image decoded by such a decoding apparatus, conversion can be performed using a conversion matrix appropriately determined so as to reduce the correlation occurring in the edge direction, and the encoding efficiency can be improved. Can be improved.
  • the processing target It is preferable to determine the inverse matrix of the block.
  • an inverse transformation matrix of the processing target block is determined based on an angle of the edge vector, When the intensity value is less than the first threshold, it is preferable to determine an inverse transformation matrix of the processing target block based on the use design angle.
  • the inverse transformation matrix can be determined based on the angle of the edge vector, and when the intensity value of the edge vector is small, the inverse transformation matrix is determined based on the design angle used. be able to.
  • the greater the intensity value of the edge vector the higher the possibility that the direction of the edge vector matches the direction of the edge of the processing target block. That is, the inverse transformation matrix can be appropriately determined according to the reliability of the edge vector detected in the block spatially adjacent to the processing target block.
  • the edge vector is selected from a plurality of inverse transformation matrices held in association with a plurality of design angles. It is preferable that an inverse transformation matrix corresponding to a design angle closest to the angle is selected as the inverse transformation matrix of the processing target block.
  • the use design angle and the use angle are selected from the plurality of inverse transformation matrices.
  • the inverse transformation matrix corresponding to the design angle closest to the average with the angle of the edge vector is selected as the inverse transformation matrix of the processing target block, and the intensity value is less than the second threshold. It is preferable that an inverse transformation matrix corresponding to a design angle closest to the design angle used is selected as an inverse transformation matrix of the processing target block from among the inverse transformation matrices.
  • the inverse transformation matrix is determined based on both the edge vector angle and the used design angle, or the inverse transformation matrix is determined based only on the used design angle. It becomes possible to do. Therefore, it is possible to determine the inverse transformation matrix more appropriately according to the reliability of the edge vector.
  • the prediction step based on a motion vector obtained from an encoded stream including the encoded image, a decoded block that is temporally close to the processing target block is obtained as a motion compensation block.
  • a block is generated, and the edge vector detection step includes a first edge vector detection step for detecting a first edge vector in the motion compensation block, and a second in a decoded block spatially adjacent to the processing target block.
  • the angle of the edge of the target block can be predicted with high accuracy based on the edge vectors detected in each of the block spatially adjacent to the target block and the motion compensation block, and more appropriately reversed. It becomes possible to determine the transformation matrix.
  • the method further includes a correction step of correcting an angle of the detected edge vector using a correction angle obtained from the encoded stream including the encoded image, and the correction is performed in the inverse transformation matrix determination step. It is preferable to determine an inverse transformation matrix of the processing target block based on the angle of the edge vector.
  • the inverse transformation step includes a division step of dividing the inversely quantized encoded difference block into a second part and a second transformation output, and the second transformation using the determined inverse transformation matrix.
  • the first conversion output is generated by integrating the second inverse conversion step for generating the first part by integrating the first part and the second part by inversely converting the conversion output of It is preferable to include an integration step and a first inverse conversion step of generating the difference block by inversely converting the first converted output.
  • the inverse transformation matrix can be appropriately determined even in the two-stage transformation, and the coding efficiency can be improved.
  • an image encoding method is an image encoding method for encoding an image for each block, and is an encoded block spatially close to a processing target block
  • a transformation matrix determining step for determining a transformation matrix of the processing target block based on the detected angle of the edge vector; a prediction step for generating a prediction block by predicting the processing target block; and the processing target block
  • a subtraction step for generating a difference block by subtracting the prediction block from It was using the transformation matrix, including a conversion step of converting the difference block generated, and a quantization step of quantizing the transformed the difference block.
  • the transformation matrix can be determined based on the angle of the edge vector of the decoded block that is close to the target block temporally or spatially. That is, the transformation matrix can be determined based on the angle of the edge vector that can be detected in smaller units. In other words, it is possible to appropriately determine the transformation matrix so as to reduce the correlation occurring in the edge direction, and it is possible to improve the coding efficiency by performing transformation using such a transformation matrix. It becomes.
  • the present invention can be realized not only as such an image decoding method or an image encoding method, but also as a configuration unit that executes processing of each step included in such an image decoding method or an image encoding method. It can also be realized as an image decoding device or an image encoding device provided. Such an image decoding apparatus or image encoding apparatus can also be configured as an integrated circuit.
  • the present invention can also be realized as a program that causes a computer to execute each step included in the image decoding method or the image encoding method.
  • a program can be distributed via a non-temporary recording medium such as a CD-ROM (Compact Disc Only Memory) or a transmission medium such as the Internet.
  • FIG. 1A is a block diagram illustrating a functional configuration of an image encoding device according to an aspect of the present invention.
  • FIG. 1B is a flowchart illustrating a processing operation of the image encoding device according to an aspect of the present invention.
  • FIG. 2A is a block diagram illustrating a functional configuration of the image decoding device according to one aspect of the present invention.
  • FIG. 2B is a flowchart illustrating a processing operation of the image decoding device according to one aspect of the present invention.
  • FIG. 3 is a block diagram showing a functional configuration of the coding apparatus according to Embodiment 1 of the present invention.
  • FIG. 4 is a flowchart showing the processing operation of the coding apparatus according to Embodiment 1 of the present invention.
  • FIG. 1A is a block diagram illustrating a functional configuration of an image encoding device according to an aspect of the present invention.
  • FIG. 1B is a flowchart illustrating a processing operation of the image encoding device according to
  • FIG. 5A is a conceptual diagram for explaining the directionality of the correlation of the prediction error signal of a block including an edge.
  • FIG. 5B is a conceptual diagram for explaining the directionality of the correlation of the prediction error signal of the block including the edge.
  • FIG. 5C is a conceptual diagram for explaining the directionality of the correlation of the prediction error signal of the block including the edge.
  • FIG. 6A is a diagram for explaining an example of processing for determining a transformation matrix based on the angle of an edge vector.
  • FIG. 6B is a diagram for explaining an example of processing for determining a transformation matrix based on the angle of an edge vector.
  • FIG. 7A is a diagram for explaining another example of the process of determining the transformation matrix based on the angle of the edge vector.
  • FIG. 7B is a diagram for explaining another example of the process of determining the transformation matrix based on the angle of the edge vector.
  • FIG. 8 is a block diagram showing a functional configuration of the decoding apparatus according to Embodiment 2 of the present invention.
  • FIG. 9 is a flowchart showing processing operations of the decoding apparatus according to Embodiment 2 of the present invention.
  • FIG. 10 is a flowchart showing processing operations of the decoding apparatus according to Embodiment 3 of the present invention.
  • FIG. 11 is a conceptual diagram for explaining processing for determining an inverse transformation matrix based on an angle of an edge vector.
  • FIG. 12A is a flowchart showing a processing operation of the decoding apparatus according to Embodiment 4 of the present invention.
  • FIG. 12B is a conceptual diagram for explaining an example of processing for determining an inverse transformation matrix based on an angle of an edge vector.
  • FIG. 13 is a diagram for explaining an example of a reference angle correction method.
  • FIG. 14 is a conceptual diagram showing an outline of processing in the fifth embodiment of the present invention.
  • FIG. 15 is a block diagram showing a functional configuration of the decoding apparatus according to Embodiment 5 of the present invention.
  • FIG. 16 is a flowchart showing processing operations of the decoding apparatus according to Embodiment 5 of the present invention.
  • FIG. 17 is a flowchart illustrating a method for determining an angle with respect to a neighboring block.
  • FIG. 18A is a diagram for describing processing for calculating an average of angles.
  • FIG. 18B is a diagram for describing processing for calculating an average of angles.
  • FIG. 19A is a diagram for explaining the starting point of an angle.
  • FIG. 19B is a diagram for explaining the angle circulation processing.
  • FIG. 20A is a block diagram showing a functional configuration related to forward conversion of two-stage conversion in the coding apparatus according to Embodiment 6 of the present invention.
  • FIG. 20B is a flowchart showing the flow of the forward conversion process of the two-stage conversion in the encoding apparatus according to Embodiment 6 of the present invention.
  • FIG. 20C is a block diagram showing a functional configuration related to inverse transformation of two-stage transformation in the decoding apparatus according to Embodiment 6 of the present invention.
  • FIG. 20A is a block diagram showing a functional configuration related to forward conversion of two-stage conversion in the coding apparatus according to Embodiment 6 of the present invention.
  • FIG. 20B is a flowchart showing the flow of the forward conversion process of the two-stage conversion in
  • FIG. 20D is a flowchart showing the flow of the inverse transformation process of the two-stage transformation in the decoding apparatus according to Embodiment 6 of the present invention.
  • FIG. 21A is a diagram illustrating an example of a first conversion output of a 4 ⁇ 4 block.
  • FIG. 21B is a conceptual diagram for explaining a method of selecting one piece of division integration information from a plurality of division integration information.
  • FIG. 22 is a diagram for explaining an example of eight pieces of division integration information corresponding to eight design angles on a one-to-one basis.
  • FIG. 23 is an overall configuration diagram of a content supply system that implements a content distribution service.
  • FIG. 24 is an overall configuration diagram of a digital broadcasting system.
  • FIG. 25 is a block diagram illustrating a configuration example of a television.
  • FIG. 26 is a block diagram illustrating a configuration example of an information reproducing / recording unit that reads and writes information from and on a recording medium that is an optical disk.
  • FIG. 27 is a diagram illustrating a structure example of a recording medium that is an optical disk.
  • FIG. 28A shows an example of a mobile phone.
  • FIG. 28B is a block diagram illustrating a configuration example of a mobile phone.
  • FIG. 29 is a diagram showing a structure of multiplexed data.
  • FIG. 30 is a diagram schematically showing how each stream is multiplexed in the multiplexed data.
  • FIG. 31 is a diagram showing in more detail how the video stream is stored in the PES packet sequence.
  • FIG. 32 is a diagram showing the structure of TS packets and source packets in multiplexed data.
  • FIG. 33 is a diagram illustrating a data structure of the PMT.
  • FIG. 34 shows the internal structure of multiplexed data information.
  • FIG. 35 shows the internal structure of stream attribute information.
  • FIG. 36 is a diagram showing steps for identifying video data.
  • FIG. 37 is a block diagram illustrating a configuration example of an integrated circuit that realizes the moving image encoding method and the moving image decoding method according to each embodiment.
  • FIG. 38 is a diagram showing a configuration for switching the drive frequency.
  • FIG. 39 is a diagram illustrating steps for identifying video data and switching between driving frequencies.
  • FIG. 40 is a diagram illustrating an example of a look-up table in which video data standards are associated with drive frequencies.
  • FIG. 41A is a diagram illustrating an example of a configuration for sharing a module of the signal processing unit.
  • FIG. 41B is a diagram illustrating another example of
  • FIG. 1A is a block diagram illustrating a functional configuration of an image encoding device 10 according to an aspect of the present invention.
  • FIG. 1B is a flowchart illustrating a processing operation of the image encoding device 10 according to an aspect of the present invention.
  • the image encoding device 10 includes an edge vector detection unit 11, a transformation matrix determination unit 12, a prediction unit 13, a subtraction unit 14, a transformation unit 15, and a quantization unit 16. .
  • the image encoding device 10 encodes an image for each block.
  • the operation of each component of the image encoding device 10 illustrated in FIG. 1A will be described with reference to FIG. 1B.
  • the edge vector detection unit 11 encodes a block in which an encoded block that is spatially close to a processing target block (hereinafter, simply referred to as “target block”) is reconfigured and a temporally close proximity to the target block.
  • An edge vector is detected in at least one of the reconstructed blocks (S11).
  • the block spatially adjacent to the target block is a block near the target block (hereinafter also referred to as “neighboring block”).
  • the block spatially adjacent to the target block includes an adjacent block spatially adjacent to the target block and a block adjacent to the adjacent block.
  • a block that is temporally close to the target block is a block included in a reference picture that has been encoded and reconstructed before the picture including the target block.
  • the transformation matrix determination unit 12 determines the transformation matrix of the target block based on the detected angle of the edge vector (S12).
  • a transformation matrix is a matrix used for transformation from the spatial domain to the frequency domain. Examples of the transform from the spatial domain to the frequency domain include discrete cosine transform (DCT: Discrete Cosine Transform), discrete sine transform (DST: Discrete Sine Transform), or discrete Hadamard transform (DHT).
  • DCT discrete cosine transform
  • DST Discrete Sine Transform
  • DHT discrete Hadamard transform
  • the prediction unit 13 generates a prediction block by predicting the target block (S13). For example, the prediction unit 13 generates a prediction block by performing in-plane prediction based on the detected edge vector direction. For example, the prediction unit 13 may generate a prediction block by performing motion compensation prediction.
  • the subtraction unit 14 generates a difference block by subtracting the prediction block from the target block (S14).
  • the conversion unit 15 converts the generated difference block using the determined conversion matrix (S15).
  • transformation means transformation from a spatial domain to a frequency domain (so-called orthogonal transformation).
  • the quantization unit 16 quantizes the converted difference block (S16).
  • the transformation matrix can be determined based on the angle of the edge vector of the decoded block that is temporally or spatially close to the target block. That is, the transformation matrix can be determined based on the angle of the edge vector that can be detected in smaller units. Therefore, it is possible to appropriately determine the transformation matrix so as to reduce the correlation occurring in the edge direction, and it is possible to improve the coding efficiency by performing transformation using such a transformation matrix. It becomes.
  • FIG. 2A is a block diagram illustrating a functional configuration of the image decoding device 20 according to an aspect of the present invention.
  • FIG. 2B is a flowchart showing the processing operation of the image decoding device 20 according to an aspect of the present invention.
  • the image decoding device 20 includes an edge vector detection unit 21, an inverse transformation matrix determination unit 22, an inverse quantization unit 23, an inverse transformation unit 24, a prediction unit 25, and an addition unit 26. Is provided.
  • the image decoding device 20 decodes the encoded image for each block.
  • the operation of each component of the image decoding device 20 illustrated in FIG. 2A will be described with reference to FIG. 2B.
  • the edge vector detection unit 21 detects an edge vector in at least one of a decoded block spatially close to the target block and a decoded block temporally close to the target block (S21). ).
  • the inverse transformation matrix determination unit 22 determines the inverse transformation matrix of the target block based on the detected angle of the edge vector (S22).
  • the inverse quantization unit 23 inversely quantizes the encoded difference block corresponding to the target block (S23).
  • the inverse transform unit 24 generates a difference block by inversely transforming the inversely quantized encoded difference block using the determined inverse transform matrix (S24).
  • the prediction unit 25 generates a prediction block by predicting the target block (S25).
  • the adding unit 26 generates a decoded block by adding the prediction block and the difference block (S26).
  • the inverse transformation matrix can be determined based on the angle of the edge vector of the decoded block that is close to the target block temporally or spatially. That is, the inverse transformation can be performed using the inverse transformation matrix determined based on the angle of the edge vector that can be detected in smaller units. Therefore, in an encoded image decoded by such a decoding device, the conversion can be performed using a conversion matrix appropriately determined so as to reduce the correlation occurring in the edge direction, and the encoding efficiency Can be improved.
  • FIG. 3 is a block diagram showing a functional configuration of coding apparatus 100 according to Embodiment 1 of the present invention.
  • the encoding apparatus 100 encodes the input video signal Vin for each block and outputs an encoded signal Str.
  • the encoding apparatus 100 includes a subtraction unit 144, a transform quantization unit 142, an encoding unit 143, an inverse quantization inverse transform unit 123, an adder 129, a frame memory 124, a candidate vector derivation unit 125, An edge vector deriving unit 126, a prediction unit 127, a prediction conversion control unit 130, and an encoding control unit 145 are provided.
  • the encoding device 100 corresponds to the image encoding device 10 described above.
  • a group including the candidate vector deriving unit 125 and the edge vector deriving unit 126 of the encoding device 100 corresponds to the edge vector detecting unit 11 of the image encoding device 10.
  • the prediction conversion control unit 130, the prediction unit 127, and the subtraction unit 144 of the encoding device 100 correspond to the transformation matrix determination unit 12, the prediction unit 13, and the subtraction unit 14 of the image encoding device 10, respectively.
  • the transform quantization unit 142 of the encoding device 100 corresponds to a group including the transform unit 15 and the quantization unit 16 of the image encoding device 10.
  • the subtraction unit 144 generates a prediction error signal PredErr that is a difference signal between the video signal Vin and the prediction signal Pred.
  • the prediction error signal PredErr is output to the transform quantization unit 142.
  • the transform quantization unit 142 transforms the prediction error signal PredErr, scans, and quantizes it to generate a coding coefficient Coeff.
  • the coding coefficient Coeff is output to the coding unit 143 and the inverse quantization and inverse transform unit 123.
  • the encoding unit 143 generates an encoded signal Str by entropy compressing the encoding coefficient Coeff.
  • the inverse quantization inverse transform unit 123 inversely quantizes the coding coefficient Coeff, performs inverse scan, and inverse transform, thereby generating a quantized prediction error signal QPredErr.
  • the quantized prediction error signal QPredErr is output to the adder 129.
  • the addition unit 129 generates a decoded signal Recon by adding the quantized prediction error signal QPredErr and the prediction signal Pred. Then, the addition unit 129 stores the decoded signal Recon in the frame memory 124.
  • the candidate vector deriving unit 125 acquires the image signal BlkNgh of the decoded block near the target block from the frame memory 124. That is, the candidate vector deriving unit 125 determines at least one of the reconstructed block of the encoded block spatially close to the target block and the reconstructed block of the encoded block temporally close to the target block as the frame memory 124. Get from. Then, the candidate vector deriving unit 125 derives a plurality of candidate vectors Vec by performing edge detection filter processing or the like on the image signal BlkNgh.
  • the edge vector deriving unit 126 detects the edge vector EdgeVec based on the plurality of candidate vectors Vec. In other words, the edge vector deriving unit 126 performs edge vectors in at least one of the reconstructed block of the encoded block spatially close to the target block and the reconstructed block of the encoded block temporally close to the target block. Is detected.
  • the edge vector deriving unit 126 selects a candidate vector having the maximum vector size (norm) from among a plurality of candidate vectors as the edge vector EdgeVec. For example, the edge vector deriving unit 126 calculates the edge vector EdgeVec by performing a statistical averaging process on the plurality of candidate vectors.
  • the edge vector derivation unit 126 may detect the edge vector EdgeVec by excluding the candidate vector Vec that does not indicate the target block. Further, the edge vector deriving unit 126 may select the candidate vector Vec having the longest length of the portion intersecting the target block among the straight lines obtained by extending the candidate vector Vec as the edge vector EdgeVec.
  • the encoding control unit 145 determines mode information Mode that is prediction or conversion type information, quantization accuracy, and the like so that encoding efficiency is increased.
  • the prediction conversion control unit 130 determines a conversion matrix based on the edge vector EdgeVec and the mode information Mode. Then, the predictive transformation control unit 130 sends the transformation control signal CtrlTQ, which is a control signal designating the decided transformation matrix and the inverse transformation matrix corresponding to the transformation matrix, to the transformation quantization unit 142 and the inverse quantization inverse transformation unit 123. Output. Further, the predictive conversion control unit 130 outputs a predictive control signal CtrlP, which is a control signal designating a prediction method, to the predicting unit 127.
  • the prediction unit 127 generates a prediction signal Pred by acquiring a reference image signal BlkRef, which is a decoded image signal at a position required for prediction, from the frame memory 124 according to the prediction control signal CtrlP.
  • the transform quantization unit 142 performs transform using the transform matrix specified by the transform control signal CtrlTQ.
  • the inverse quantization inverse transform unit 123 performs inverse transform using an inverse transform matrix specified by the transform control signal CtrlTQ.
  • FIG. 4 is a flowchart showing the processing operation of the encoding apparatus 100 according to Embodiment 1 of the present invention.
  • the input image included in the video signal Vin is divided into blocks composed of at least one pixel and encoded in units of blocks (S101, S106).
  • the edge vector derivation unit 126 reconstructs a block (a neighboring block) in which an encoded block spatially close to the target block is reconstructed and a coded block that is temporally close to the target block.
  • An edge vector is detected in at least one of the blocks (motion compensation block) (S102).
  • the edge vector deriving unit 126 detects an edge vector in the neighboring block.
  • the edge vector deriving unit 126 uses the edge vector in the motion compensation block at the position indicated by the motion vector in the reference picture stored in the frame memory 124. Is detected.
  • the edge vector deriving unit 126 When the motion vector has decimal precision information and the prediction block is generated by applying the motion compensation interpolation filter to the motion compensation block, the edge vector deriving unit 126 performs the process in the motion compensation block before the interpolation filter application. What is necessary is just to detect an edge vector. In this case, since the interpolation filter process and the edge vector detection process can be executed in parallel, the processing speed can be improved. Note that the edge vector deriving unit 126 may detect an edge vector in a motion compensation block (prediction block) after the interpolation filter is applied.
  • the prediction conversion control unit 130 determines a prediction mode based on the detected angle and intensity of the edge vector (S103).
  • the prediction unit 127 generates a prediction block (prediction signal Pred) according to the determined prediction mode (S104).
  • the subtraction part 144 produces
  • the prediction conversion control unit 130 determines the conversion matrix of the target block based on the angle of the edge vector (S113).
  • the transform quantization unit 142 transforms the difference block (difference signal PredErr) using the determined transform matrix, scans, and quantizes the transform block to generate the coding coefficient Coeff of the target block.
  • the inverse quantization inverse transform unit 123 inversely quantizes the coding coefficient Coeff, performs inverse scan, and performs inverse transform using the inverse transform matrix corresponding to the determined transform matrix, thereby reconstructing the difference block (quantized prediction).
  • An error signal QPredErr is generated (S114).
  • the adding unit 129 generates a reconstructed block (decoded signal Recon) by adding the reconstructed difference block and the prediction block, and stores the reconstructed block in the frame memory 124 (S105).
  • 5A to 5C are conceptual diagrams for explaining the directionality of the correlation of the prediction error signal of the block including the edge.
  • the target block has an edge Edge of some angle as shown in FIG. 5A.
  • the block is further subdivided or an appropriate prediction method is selected.
  • block division information or prediction mode information will increase. Therefore, in order to suppress the code amount, the predicted edge signal Pred often cannot be completely matched with the actual edge.
  • the prediction error signal PredErr tends to have a large energy in the vicinity of the edge as shown by the hatched area in FIG. 5B.
  • the prediction error signal PredErr tends to have a high correlation in the direction of the edge and a weak correlation in the direction orthogonal to the edge.
  • the transformation matrix is determined so that the correlation in the edge direction, which is the direction of strong correlation, is reduced to the maximum. Therefore, it is possible to improve the compression performance in the compression encoding method that combines transformation and quantization.
  • FIG. 6A and 6B are diagrams for explaining an example of processing for determining a transformation matrix based on the angle of an edge vector.
  • the predictive conversion control unit 130 specifies a design angle closest to the angle EAng of the edge vector, and selects a conversion matrix corresponding to the design angle.
  • the prediction conversion control unit 130 specifies the design angle T1 as the conversion angle Tsel corresponding to the angle EANG of the edge vector.
  • the predictive conversion control unit 130 selects a conversion matrix corresponding to the conversion angle Tsel.
  • the angle EAng of the edge vector is defined as shown in FIG. 6B.
  • FIG. 7A and 7B are diagrams for explaining another example of the process of determining the transformation matrix based on the angle of the edge vector.
  • a transformation matrix is held in association with a predefined angle range (R0 to R3).
  • R0 to R3 a predefined angle range
  • the predictive conversion control unit 130 performs vertical and horizontal conversion from a plurality of conversion matrices shown in FIG. 7B.
  • a transformation matrix that is both DST is selected.
  • the transformation matrix is determined such that the direction closer to the edge vector direction is DST and the direction away from the edge vector direction is DCT, so that the correlation can be further reduced. Become. This is because DST can reduce the correlation more than DCT even at a position away from the adjacent block where the difference in direction between the edge vector detected in the adjacent block and the edge of the target block becomes large. .
  • the transformation matrix is determined based on the angle of the edge vector detected in the decoded block that is temporally or spatially close to the target block. be able to. That is, the transformation matrix can be determined based on the angle of the edge vector that can be detected in smaller units. In other words, it is possible to appropriately determine the transformation matrix so as to reduce the correlation occurring in the edge direction, and it is possible to improve the coding efficiency by performing transformation using such a transformation matrix. It becomes.
  • FIG. 8 is a block diagram showing a functional configuration of the decoding apparatus 200 according to Embodiment 2 of the present invention.
  • the same components as those in FIG. 3 are denoted by the same reference numerals, and description thereof will be omitted as appropriate.
  • the decoding apparatus 200 decodes the encoded signal Str encoded by the encoding apparatus 100 according to the first embodiment for each block, and outputs a decoded screen signal Vout.
  • the decoding device 200 includes a decoding unit 122, an inverse quantization inverse transformation unit 123, an addition unit 129, a frame memory 124, a candidate vector derivation unit 125, an edge vector derivation unit 126, a prediction unit 127, and a prediction transformation. And a control unit 130.
  • the decoding device 200 corresponds to the image decoding device 20 described above.
  • a group including the candidate vector deriving unit 125 and the edge vector deriving unit 126 of the decoding device 200 corresponds to the edge vector detecting unit 21 of the image decoding device 20.
  • the prediction conversion control unit 130, the prediction unit 127, and the addition unit 129 of the decoding device 200 correspond to the inverse transformation matrix determination unit 22, the prediction unit 25, and the addition unit 26 of the image decoding device 20, respectively.
  • the inverse quantization inverse transform unit 123 of the decoding device 200 corresponds to a group including the inverse quantization unit 23 and the inverse transform unit 24 of the image decoding device 20.
  • the decoding unit 122 entropy-decodes the encoded stream Str, and acquires the encoded coefficient Coeff and mode information Mode.
  • the candidate vector deriving unit 125 acquires the image signal BlkNgh of the decoded block near the target block from the frame memory 124 as in the first embodiment. That is, the candidate vector deriving unit 125 acquires at least one of the decoded block spatially close to the target block and the decoded block temporally close to the target block from the frame memory 124. Then, the candidate vector deriving unit 125 derives a plurality of candidate vectors Vec by performing edge detection filter processing or the like on the image signal BlkNgh.
  • the edge vector deriving unit 126 detects the edge vector EdgeVec based on the plurality of candidate vectors Vec. That is, the edge vector deriving unit 126 detects edge vectors in at least one of a decoded block spatially close to the target block and a decoded block temporally close to the target block.
  • the prediction conversion control unit 130 determines an inverse conversion matrix based on the angle of the edge vector EdgeVec. Then, the predictive transform control unit 130 outputs a transform control signal CtrlTQ, which is a control signal designating the determined inverse transform matrix, to the inverse quantization inverse transform unit 123. Further, the predictive conversion control unit 130 outputs a predictive control signal CtrlP, which is a control signal designating a prediction method, to the predicting unit 127.
  • the inverse quantization inverse transform unit 123 inversely quantizes the coding coefficient Coeff and generates a prediction error signal PredErr by performing inverse transform from the frequency domain to the spatial domain using an inverse transform matrix specified by the transform control signal CtrlTQ. .
  • the addition unit 129 generates a decoded signal Recon by adding the quantized prediction error signal QPredErr and the prediction signal Pred. Then, the addition unit 129 stores the decoded signal Recon in the frame memory 124.
  • the decoded signal Recon stored in the frame memory 124 is output as a decoded video signal Vout.
  • FIG. 9 is a flowchart showing the processing operation of the decoding apparatus 200 according to Embodiment 2 of the present invention.
  • steps that perform the same processing as in FIG. 4 are given the same symbols, and descriptions thereof are omitted as appropriate.
  • the decoding apparatus 200 performs processing in units of blocks, similar to the encoding apparatus 100 according to Embodiment 1 (S101, S106). Specifically, the decoding device 200 decodes the encoded image included in the encoded stream Str for each block. The processing from step S102 to S104 is the same as that in FIG.
  • the predictive transformation control unit 130 determines an inverse transformation matrix of the target block based on the detected angle of the edge vector (S113). Note that the method for determining the inverse transformation matrix is the same as the method for determining the transformation matrix in the first embodiment, and thus detailed description thereof is omitted.
  • the inverse quantization inverse transform unit 123 inverse quantizes the coding coefficient Coeff and performs inverse transform using the determined inverse transform matrix to generate a difference block (quantized prediction error signal QPredErr). (S114).
  • the adding unit 129 generates a decoded block (decoded signal Recon) by adding the prediction block and the difference block (S105).
  • the decoded block generated in this way is stored in the frame memory 124 and output to the outside.
  • an inverse transformation matrix can be determined based on the angle of an edge vector of a decoded block that is temporally or spatially close to the target block. . That is, the inverse transformation can be performed using an inverse transformation matrix determined based on the angle of the edge vector that can be detected in smaller units. Therefore, in an encoded image decoded by such a decoding apparatus, conversion can be performed using a conversion matrix appropriately determined so as to reduce the correlation occurring in the edge direction, and the encoding efficiency can be improved. Can be improved.
  • the decoding device according to Embodiment 3 of the present invention is a modification of the decoding device of Embodiment 2.
  • the angle of the edge vector is corrected using the difference angle obtained from the encoded stream, and the inverse transformation matrix is determined based on the angle of the corrected edge vector.
  • the decoding apparatus according to the present embodiment has the same functional configuration as that of the decoding apparatus according to the second embodiment shown in FIG.
  • FIG. 10 is a flowchart showing the processing operation of the decoding apparatus 200 according to Embodiment 3 of the present invention.
  • the edge vector deriving unit 126 detects an edge vector as in the second embodiment (S102).
  • the predictive conversion control unit 130 corrects the angle of the edge vector using the difference angle (S133). Specifically, the predictive conversion control unit 130 corrects the edge vector angle EAng by adding the difference angle DeltaAng to the edge vector angle EAng. On the other hand, when the difference angle is not included in the encoded stream (NO in S132), the predictive conversion control unit 130 does not correct the angle EAng of the edge vector.
  • FIG. 11 is a conceptual diagram for explaining an example of processing for determining an inverse transformation matrix based on the angle of an edge vector.
  • a case where the detected angle EAng of the edge vector (angle before correction) is closest to the design angle T1 will be described as an example.
  • an inverse transformation matrix corresponding to the design angle T1 closest to the edge vector angle EAng is selected.
  • the inverse transformation matrix corresponding to the design angle T3 closest to the angle of the edge vector after the difference angle is added is selected. It is assumed that the increment amount of the difference angle can be controlled in units of streams or frames by the encoding device.
  • the difference angle can be notified to the decoding apparatus as necessary. Therefore, even when edge vector detection fails due to noise or the like, the inverse transformation matrix can be notified correctly, so that the compression performance can be improved.
  • the angle of the edge vector can be corrected using the correction angle. Therefore, even if the edge vector close to the direction of the edge of the target block cannot be detected in the block adjacent to the target block, the inverse transformation matrix can be appropriately determined using the corrected angle. As a result, it is possible to reduce the correlation that occurs in the direction of the edge and improve the coding efficiency.
  • the difference angle does not necessarily need to be an angle for adding to the angle Eang of the edge vector.
  • the difference angle may be an angle for multiplying the angle EANG of the edge vector.
  • the difference angle may be an angle for replacing the edge vector angle EAng. That is, the difference angle may be a correction angle for correcting the angle of the edge vector.
  • encoding apparatus 100 may also determine a transformation matrix, similarly to decoding apparatus 200 according to the present embodiment.
  • the prediction conversion control unit 130 of the encoding device 100 may correct the edge vector angle EAng by adding the difference angle DeltaAng to the edge vector angle EAng.
  • the difference angle may be a difference between the angle of the edge vector of the neighboring block detected by the edge vector deriving unit 126 and the angle of the edge vector of the target block, for example. More specifically, for example, when the difference between the angle of the edge vector of the neighboring block and the angle of the edge vector of the target block is greater than or equal to the threshold value, the prediction conversion control unit 130 determines the difference as the difference angle. That's fine.
  • the decoding device according to Embodiment 4 of the present invention is a modification of the decoding device according to Embodiments 2 and 3.
  • the inverse transformation matrix is determined based on the reference angle acquired from the encoded stream and the angle of the edge vector. Note that the decoding apparatus according to the present embodiment has the same functional configuration as that of the decoding apparatus according to Embodiment 2 (FIG. 8), and therefore description of the block diagram is omitted.
  • FIG. 12A is a flowchart showing the processing operation of the decoding apparatus 200 according to Embodiment 4 of the present invention.
  • the edge vector deriving unit 126 detects an edge vector as in the second and third embodiments (S102).
  • the predictive conversion control unit 130 acquires the reference angle ExplctAng of the target block from the encoded stream (S141).
  • the reference angle ExplctAng is an angle for determining the inverse transformation matrix of the target block.
  • the reference angle ExplctAng is an angle in the prediction direction of in-plane prediction.
  • the reference angle ExplctAng may be a design angle corresponding to the inverse transformation matrix.
  • the reference angle ExplctAng is preferably expressed with a very coarse particle size. Thereby, the code amount of the reference angle can be suppressed.
  • the reference angle ExplctAng may be encoded in the encoded stream as a flag indicating 0 degree when “0” and 90 degrees when “1”. Further, for example, the reference angle ExplctAng may be encoded in the encoded stream as a flag indicating one of three angles of 0 degree, 45 degrees, and 90 degrees.
  • the predictive conversion control unit 130 acquires the reference angle explicitly notified from the encoding device in this way.
  • the predictive conversion control unit 130 determines whether or not to correct the reference angle ExplctAng using the edge vector angle EAng (S142). For example, the predictive conversion control unit 130 determines whether or not to correct the reference angle ExplctAng based on the difference between the reference angle ExplctAng and the edge vector angle EAng.
  • the predictive conversion control unit 130 corrects the reference angle ExplAngAng using the angle of the edge vector (S143).
  • the predictive conversion control unit 130 does not correct the reference angle ExplctAng.
  • the prediction conversion control unit 130 determines an inverse conversion matrix based on the reference angle ExplctAng (S144). That is, the predictive conversion control unit 130 determines the inverse transform matrix using the reference angle ExplctAng instead of the edge vector angle EAnng in the second embodiment. Other steps are the same as those in the second embodiment, and thus the description thereof is omitted.
  • FIG. 12B is a conceptual diagram for explaining an example of processing for determining an inverse transformation matrix based on an angle of an edge vector.
  • the predictive conversion control unit 130 corrects the reference angle ExplcAng obtained from the encoded stream using the angle Eang of the edge vector so that the corrected reference angle falls within the correction range UpdateRange. . Then, the prediction conversion control unit 130 selects an inverse conversion matrix corresponding to the design angle T3 closest to the reference angle corrected as described above.
  • a conversion angle BlkAng is an angle for determining an inverse conversion matrix. That is, the predictive conversion control unit 130 determines an inverse conversion matrix based on the conversion angle BldAng. Specifically, the conversion angle BlkAng is a corrected reference angle when the reference angle is corrected, and is the reference angle itself when the reference angle is not corrected.
  • the predictive conversion control unit 130 uses the angle EANG of the edge vector as the target block. Conversion angle BlkAng. That is, when the absolute value of the difference value between the angle EANG of the edge vector and the reference angle ExplctAng is less than the first angle, the predictive conversion control unit 130 determines that the reference angle ExplctAng matches the angle Eang of the edge vector. The angle ExplctAng is corrected.
  • the predictive conversion control unit 130 determines the angle EAng of the edge vector as the conversion angle BlkAng of the target block.
  • the predictive conversion control unit 130 determines the conversion angle BlkAng by correcting the reference angle ExplctAng using the angle EAng of the edge vector. Specifically, when the absolute value of the difference value between the edge vector angle EAng and the reference angle ExplctAng is greater than or equal to the first angle and less than the second angle, the predictive conversion control unit 130 uses the edge vector angle EAng as a reference. The angle ExplctAng is corrected.
  • the predictive conversion control unit 130 corrects the reference angle ExplctAng so that the corrected reference angle ExplctAng falls within a predetermined update range UpdateRange centered on the reference angle ExplctAng.
  • the predictive conversion control unit 130 corrects the reference angle ExplctAng so as to approach the angle Eang of the edge vector by a predetermined angle from the reference angle ExplctAng. This predetermined angle may be determined according to the increment of the reference angle, for example.
  • the edge vector angle EAng when the edge vector angle EAng is moderately separated from the reference angle ExplctAng, the accuracy of the edge vector angle EAng may be low. It is not preferable that ExplctAng is corrected. Therefore, the reference angle ExplctAng is corrected so that the corrected reference angle ExplctAng falls within a narrow range (Range 1b) centered on the reference angle ExplctAng.
  • the prediction conversion control unit 130 determines the reference angle ExplctAng as the target block conversion angle BlkAng without considering the angle of the edge vector. That is, the predictive conversion control unit 130 does not correct the reference angle ExplctAng when the absolute value of the difference value between the angle EANG of the edge vector and the reference angle ExplctAng is equal to or larger than the second angle.
  • the edge vector deriving unit 126 determines the magnitude of the difference between the reference angle ExplctAng and the angle of the candidate vector, in addition to the size of the candidate vector or the length of the portion where the straight line obtained by extending the candidate vector intersects the target block. Considering this, the edge vector may be detected. That is, the edge vector deriving unit 126 may detect an edge vector by excluding candidate vectors whose difference from the reference angle ExplctAng is larger than a predetermined threshold.
  • an explicit reference angle having a large angle increment but high reliability, and an angle of a high-definition edge vector that may be low in reliability, Can be combined to determine the inverse transformation matrix more appropriately, and the compression performance can be improved.
  • encoding apparatus 100 may also determine a transformation matrix, similarly to decoding apparatus 200 according to the present embodiment. Specifically, the prediction conversion control unit 130 of the encoding device 100 may determine a conversion matrix in the same manner as described above based on the reference angle and the angle of the edge vector.
  • the decoding device is a modification of the decoding device according to Embodiments 2 to 4.
  • the inverse transformation matrix of the target block is determined based on the design angle (usage design angle) of the inverse transformation matrix used for the inverse transformation of the block spatially adjacent to the target block.
  • FIG. 14 is a conceptual diagram showing an outline of processing in the fifth embodiment of the present invention.
  • An inverse transformation matrix is determined based on the angle of the edge vector EVecU of the upper adjacent block BlkUp adjacent above the target block BlkCur and the angle of the edge vector EVecL of the left adjacent block BlkLeft adjacent to the left of the target block BlkCur.
  • the target block is an inter-frame prediction block
  • the inverse transformation matrix is determined in consideration of the angle of the edge vector EVecT of the block (motion compensation block) BlkRef at the position in the reference picture pointed to by the motion vector MV. It is determined.
  • the design angle (use design angle) of the inverse transform matrix used for the inverse transform of the upper adjacent block BlkUp and the left adjacent block BlkLeft is also considered.
  • an inverse transformation matrix is determined.
  • FIG. 15 is a block diagram showing a functional configuration of the decoding apparatus 300 according to Embodiment 5 of the present invention.
  • the same components as those described in FIG. 8 are given the same reference numerals, description thereof will be omitted as appropriate, and differences will be mainly described.
  • the decoding device 300 includes a selector 150 and a transformation matrix history memory 151 in addition to the components shown in FIG.
  • the selector 150 When the prediction of the target block is inter-frame prediction, the selector 150 outputs a prediction signal Pred that is an image signal of a prediction block (motion compensation block) generated by the prediction unit 127 to the candidate vector derivation unit 125.
  • a prediction signal Pred that is an image signal of a prediction block (motion compensation block) generated by the prediction unit 127 to the candidate vector derivation unit 125.
  • the selector 150 obtains the image signal BlkNgh of the decoded block (hereinafter referred to as “neighboring block”) in the vicinity of the target block from the frame memory 124 and derives a candidate vector. To the unit 125.
  • the prediction unit 127 When the target block prediction is inter-frame prediction and a motion compensation interpolation filter is applied, the prediction unit 127 outputs an image signal before the interpolation filter is applied to the selector 150. Also good. In this case, the motion compensation interpolation filter application process and the edge detection process may be executed in parallel.
  • the selector 150 may also output the image signal BlkNgh to the candidate vector deriving unit 125 in addition to the prediction signal Pred when the prediction of the target block is inter-frame prediction.
  • the edge vector deriving unit 126 detects edge vectors in a plurality of neighboring blocks and motion compensation blocks, and outputs the detected plurality of edge vectors EdgeVec to the prediction conversion control unit 130.
  • the transformation matrix history memory 151 is a memory in which the design angle (use design angle) of the inverse transformation matrix used for the inverse transformation of the decoded block or the type of the transformation matrix is stored.
  • the transformation matrix history memory 151 stores the design angle (or the type of transformation matrix) indicated by the transformation control signal CtrlTQ output from the prediction transformation control unit 130.
  • the prediction conversion control unit 130 acquires the history conversion control signal StrCtrlTQ indicating the design angle (or the type of the conversion matrix) of the spatially adjacent block of the target block from the conversion matrix history memory 151. Further, the predictive conversion control unit 130 determines an inverse transform matrix based on the plurality of edge vectors and the history transform control signal StrCtrlTQ, and a transform control signal CtrlTQ indicating the determined inverse transform matrix and the design angle of the inverse transform matrix. Is output to the inverse quantization inverse transform unit 123 and the transform matrix history memory 151.
  • FIG. 16 is a flowchart showing the processing operation of the decoding apparatus 300 according to Embodiment 5 of the present invention. Note that steps that perform the same processing as the steps shown in FIG. 10 are given the same reference numerals, description thereof will be omitted as appropriate, and differences will be mainly described.
  • the edge vector deriving unit 126 when the prediction of the target block is inter-frame prediction, the edge vector deriving unit 126 performs edge processing in both the prediction signal Pred that is the image signal of the motion compensation block and the image signal BlkNgh of the neighboring block. Detect vectors.
  • the edge vector deriving unit 126 obtains a motion compensation block BlkRef that is a block at a position in the reference picture indicated by the motion vector MV. Then, the edge vector deriving unit 126 detects the edge vector EVecT in the motion compensation block BlkRef (S151). That is, the edge vector deriving unit 126 detects the first edge vector in the motion compensation block.
  • the edge vector deriving unit 126 may detect the edge vector EVecT in the motion compensation block BlkRef after the motion compensation interpolation filter is applied. . Thereby, the angle Temporal Ang of the edge vector EVecT of the motion compensation block is obtained. In the case of in-plane prediction, step S151 is not executed.
  • the edge vector derivation unit 126 detects the edge vector EVecU and the edge vector EVecL in the upper adjacent block BlkUp and the left adjacent block BlkL among the neighboring blocks (S152). That is, the edge vector derivation unit 126 detects the second edge vector in the decoded block that is spatially close to the target block.
  • edge vector deriving unit 126 may detect an edge vector in the upper left adjacent block or the upper right adjacent block in addition to the above two adjacent blocks. In the case of inter-frame prediction, step S152 may not be executed.
  • the predictive transformation control unit 130 determines an angle with respect to the neighboring block based on the angle of the edge vector of the neighboring block and the design angle of the inverse transformation matrix used for the inverse transformation of the neighboring block (S153). Specifically, first, the predictive transformation control unit 130 determines an angle UAng for the upper adjacent block BlkUp based on the angle of the edge vector EVecU and the design angle of the inverse transformation matrix used in the upper adjacent block BlkUp. . Furthermore, the predictive transformation control unit 130 determines an angle LAng for the left adjacent block BlkLeft based on the angle of the edge vector EVecL and the design angle of the inverse transformation matrix used in the left adjacent block BlkLeft. Details of the processing in step S153 will be described later with reference to FIG.
  • the predictive conversion control unit 130 performs an average process or a median process using the angle Temporal Ang of the edge vector EVecT of the motion compensation block and the angles (UAng, LAng) with respect to a plurality of adjacent blocks, thereby performing the target block. Is calculated (S154).
  • FIG. 17 is a flowchart showing an angle determination method (S153) with respect to neighboring blocks.
  • the larger the intensity value of the edge vector the higher the possibility that the direction of the edge vector matches the direction of the edge of the target block.
  • the predictive conversion control unit 130 determines the angle of the edge vector of the adjacent block as the angle with respect to the adjacent block (S162). That is, when the strength value indicating the strength of the edge vector of the adjacent block is equal to or greater than the first threshold, the predictive conversion control unit 130 determines the angle of the edge vector as the angle with respect to the adjacent block.
  • the predictive transformation control unit 130 determines the angle of the edge vector of the adjacent block and the design angle of the inverse transformation matrix used in the inverse transformation of the adjacent block.
  • the average with (use design angle) is determined as an angle with respect to the adjacent block (S163). That is, when the strength value indicating the strength of the edge vector of the adjacent block is less than the first threshold and greater than or equal to the second threshold, the predictive conversion control unit 130 Is determined as the angle with respect to the adjacent block. Note that the second threshold value is smaller than the first threshold value.
  • the predictive conversion control unit 130 performs the inverse conversion of the adjacent block.
  • the design angle of the used inverse transformation matrix is determined as an angle with respect to the adjacent block (S164). That is, when the strength value indicating the strength of the edge vector of the adjacent block is less than the second threshold, the predictive conversion control unit 130 determines the design angle to be used as the angle with respect to the adjacent block.
  • the inverse transformation matrix is determined based on the angle with respect to the adjacent block determined in this way. That is, when the intensity value of the edge vector is greater than or equal to the first threshold value, the predictive transformation control unit 130 determines an inverse transformation matrix of the target block based on the angle of the edge vector, and the intensity value of the edge vector is the first threshold value. In the case of less than, the inverse transformation matrix of the target block is determined based on at least the design angle used.
  • the predictive transformation control unit 130 targets the inverse transformation matrix corresponding to the design angle closest to the angle of the edge vector from among a plurality of inverse transformation matrices. Select as the inverse matrix of the block. Also, for example, when the intensity value of the edge vector is less than the first threshold value and greater than or equal to the second threshold value, the predictive transformation control unit 130 uses the design angle and edge vector used from a plurality of inverse transformation matrices. The inverse transformation matrix corresponding to the design angle closest to the average with the angle is selected as the inverse transformation matrix of the target block.
  • the predictive conversion control unit 130 selects an inverse transform matrix corresponding to the design angle closest to the design angle to be used from among a plurality of inverse transform matrices. , Select as the inverse transformation matrix of the target block.
  • step S153 or S154 described above the prediction conversion control unit 130 does not calculate the average of the angles using the angle expressed in radians in the local coordinate system, but instead of calculating the X component of the vector Vec in the XY coordinate system.
  • An average of a plurality of angles may be calculated using Vx and the Y component Vy.
  • Vx and Vy are the X component and the Y component of the edge vector (see FIG. 18A). Note that Vx and Vy may be weighted using the length of the line segment that cuts the target block.
  • Vx and Vy are calculated as follows using the vector magnitude K defined for the design angle ⁇ .
  • Vx cos ( ⁇ ) * K
  • Vy sin ( ⁇ ) * K
  • the vector size K is notified from the encoding device to the decoding device (see FIG. 18B).
  • K is encoded in the encoded stream Str.
  • the predictive conversion control unit 130 calculates the angle of the vector constituted by the average of Vx and Vy obtained in this way as the average of a plurality of angles.
  • the predictive conversion control unit 130 can calculate an angle with higher reliability by calculating the average of a plurality of angles in consideration of the size of the edge vector.
  • the predictive conversion control unit 130 may calculate a median value of a plurality of angles as an average of the plurality of angles. In such a case, the predictive conversion control unit 130 may calculate the angle of the vector formed from the median values of Vx and Vy as an average of a plurality of angles.
  • the predictive conversion control unit 130 when calculating the average of a plurality of angles, the predictive conversion control unit 130 preferably performs an angle circulation process so that a difference between the plurality of angles becomes small.
  • the circulation processing of such an angle will be described with reference to FIGS. 19A and 19B.
  • the edge vector angle can range from 0 to 2 ⁇ .
  • the design angle of the orthogonal transformation matrix circulates with a ⁇ period. That is, the transformation matrix corresponding to ⁇ [rad] and the transformation matrix corresponding to ⁇ + ⁇ [rad] are the same. Therefore, for example, a transformation matrix may be prepared in association with an angle in the range of 0 to ⁇ .
  • the predictive conversion control unit 130 may add or subtract ⁇ from the angle.
  • the process of adding or subtracting the circulation period (here, ⁇ ) is referred to as an angle circulation process.
  • the above-mentioned circulation processing requires special processing in the averaging processing. For example, when calculating the average of ⁇ 1 [rad] and ⁇ 2 [rad], if the circulation processing of ⁇ 2 is required, the average calculation results of ⁇ 1 and ⁇ 2 + ⁇ and the average of ⁇ 1 + ⁇ 2- ⁇ are different. In such a case, it is desirable for the predictive conversion control unit 130 to determine whether to add or subtract ⁇ from ⁇ 2 so that the difference between ⁇ 1 and ⁇ 2 is minimized. For example, in the case of FIG. 19B, ⁇ 2 ⁇ is closer to ⁇ 1 than ⁇ 2 + ⁇ . Therefore, the predictive conversion control unit 130 calculates the average of ⁇ 1 and ⁇ 2 as ( ⁇ 1 + ⁇ 2- ⁇ ) / 2. Such a cyclic process applies not only to the averaging process but also to the median process.
  • the predictive conversion control unit 130 first determines a reference angle among the angles of the three or more vectors. For example, when the averaging process of three angles ⁇ 1, ⁇ 2, ⁇ 3 ⁇ is performed, the predictive conversion control unit 130 determines, for example, ⁇ 1 as a reference angle, and ⁇ 2 so that the difference between ⁇ 1 and ⁇ 2 becomes small. Is circulated so that the difference between ⁇ 1 and ⁇ 3 is reduced. For example, the prediction conversion control unit 130 may determine the reference angle as the design angle of the conversion matrix of the left adjacent block.
  • the starting point of the angle is defined as shown in FIG. 19A. That is, it is preferable that the upper right direction is the starting point of the angle.
  • encoding is generally performed by raster scanning by dividing a frame into a plurality of blocks. Assuming such a raster scan, the frequency of the prediction direction increases in the right direction and downward direction (or diagonally downward direction). That is, by defining the upper right direction as the base point of the angle so that an angle with a high frequency as the prediction direction is included in the range from 0 to ⁇ , the frequency of the cyclic processing can be suppressed.
  • edge vector deriving unit 126 (FIGS. 3 and 8), and the edge vector deriving step S102 (FIGS. 4 and 9). This process can also be applied and the same effect can be obtained.
  • the inverse transformation matrix can be determined based on the angle of the edge vector, and the intensity value of the edge vector is When it is small, the inverse transformation matrix can be determined based on the design angle used. That is, the inverse transformation matrix can be appropriately determined according to the reliability of the edge vector detected in the neighboring block.
  • an inverse transformation matrix is determined based on both the angle of the edge vector and the design angle of use according to the magnitude of the intensity value of the edge vector. It is possible to determine an inverse transformation matrix based only on the design angle. Therefore, it is possible to determine the inverse transformation matrix more appropriately according to the reliability of the edge vector.
  • the angle of the edge of the target block is determined with high accuracy based on the edge vector detected in each of the block spatially close to the target block and the motion compensation block.
  • the inverse transformation matrix can be determined more appropriately.
  • the decoding apparatus 300 Since it is possible to perform inverse transform using the appropriately determined inverse transform matrix, the decoding apparatus 300 according to the present embodiment can improve encoding efficiency.
  • the decoding apparatus 300 determines the inverse transformation matrix based on the angle of the edge vector and the use design angle, but it is not always necessary to determine the inverse transformation matrix in this way. That is, the predictive conversion control unit 130 may determine the inverse transform matrix of the target block based only on the use design angle among the angle of the edge vector and the use design angle. Specifically, step S164 may be executed regardless of the strength of the edge vector in FIG. Thereby, the decoding apparatus 300 does not need to detect an edge vector, and can reduce the processing load.
  • the angle of the edge vector (first edge vector) detected in the motion compensation block the angle of the edge vector (second edge vector) detected in the neighboring block, and the use design angle
  • the inverse transformation matrix is determined based on the above, it is not necessarily determined in this way.
  • the predictive conversion control unit 130 calculates the average of the angle of the first edge vector and the angle of the second edge vector as the angle of the edge vector, and reverses the target block based on the calculated angle of the edge vector.
  • a transformation matrix may be determined.
  • the conversion matrix history memory 151 for storing the use design angle is not necessary, and the configuration of the decoding device 300 can be simplified.
  • encoding apparatus 100 may also determine a transformation matrix, similarly to decoding apparatus 300 according to the present embodiment.
  • the encoding apparatus 100 may also include a selector 150 and a transformation matrix history memory 151 in addition to the components shown in FIG.
  • the predictive transform control unit 130 of the encoding device 100 determines the transform matrix of the target block based on the angle of the edge vector, and the strength of the edge vector
  • the transformation matrix of the target block may be determined based on at least the design angle used. That is, the prediction conversion control unit 130 of the encoding device 100 may determine the conversion matrix by the same processing as in FIGS.
  • the encoding method or decoding method described in the first to fifth embodiments can be applied to two-stage conversion.
  • a case will be described in which the encoding method or decoding method of the first to fifth embodiments is applied to two-stage conversion.
  • FIG. 20A is a block diagram showing a functional configuration related to forward conversion of two-stage conversion in the encoding apparatus according to Embodiment 6 of the present invention. Specifically, FIG. 20A shows components included in the transform quantization unit 142.
  • the encoding apparatus according to the present embodiment includes a first conversion unit 400, a division unit 410, a second conversion unit 420, and an integration unit 430.
  • the first conversion unit 400 generates a first conversion output Tout1 by converting a conversion input signal (prediction error signal PredErr in FIG. 3). For example, when Tin is an NxN block, Tout1 is also an NxN block.
  • the dividing unit 410 divides this Tout1 into a first part Tpart1 and a second part Tpart2 according to the division integration information CtrlTsel. For example, the dividing unit 410 divides Tout1 into a low frequency portion and a remaining high frequency portion.
  • the division integration information is information indicating a block division method and an integration method.
  • the second conversion unit 420 generates the second conversion output Tout2 by performing the second conversion of the first part Tpart1.
  • the conversion matrix used for the second conversion is indicated by the second conversion control signal CtrlT2. That is, the second conversion unit 420 generates the second conversion output Tout2 by converting the first partial Tpart1 using the conversion matrix determined by the prediction conversion control unit 130.
  • the integration unit 430 generates a two-stage conversion output Tout by integrating the second part Tpart2 and the second conversion output Tout2 according to the divided integration information CtrlTsel.
  • the transform quantization unit 142 quantizes the two-stage transform output Tout to generate a coding coefficient Coeff.
  • FIG. 20B is a flowchart showing the flow of the forward conversion process of the two-stage conversion in the encoding apparatus according to Embodiment 6 of the present invention.
  • the first conversion unit 400 generates a first conversion output by converting the difference block (prediction error signal PredErr) (S171).
  • the dividing unit 410 divides the first converted output into a first part and a second part (S172).
  • the second conversion unit 420 generates a second conversion output by converting the first part using the conversion matrix determined by the prediction conversion control unit 130 (S173).
  • the integration unit 430 generates a two-stage conversion output by integrating the second part and the second conversion output (S174).
  • the transform quantization unit 142 quantizes the two-stage transform output (transformed difference block) (S175).
  • FIG. 20C is a block diagram showing a functional configuration related to inverse transformation of two-stage transformation in the decoding apparatus according to Embodiment 6 of the present invention. Specifically, FIG. 20C shows components included in the inverse quantization inverse transform unit 123.
  • the decoding apparatus according to the present embodiment includes a dividing unit 410, a second inverse transform unit 460, an integration unit 430, and a first inverse transform unit 450.
  • the dividing unit 410 divides the input RTout of the two-stage inverse transform into the decoded second part RTPart2 (second part) and the decoded second converted output RTout2 (second converted output) according to the divided integration information CtrlTsel. .
  • the second inverse transform unit 460 generates a decoded first part RTpart1 (first part) by inversely transforming RTout2 using the inverse transform matrix indicated by the second transform control signal CtrlT2. That is, the second inverse transform unit 460 generates the decoded first partial RTpart1 by inversely transforming the decoded second transform output RTout2 using the inverse transform matrix determined by the prediction transform control unit 130.
  • the integration unit 430 generates an inverse transformation output RTin (quantized prediction error signal QPredErr in FIG. 8) of two-stage transformation by integrating RTPart2 and RTout2 in accordance with the division integration information CtrlTsel.
  • RTin quantized prediction error signal QPredErr in FIG. 8
  • FIG. 20D is a flowchart showing a flow of the inverse transformation process of the two-stage transformation in the decoding apparatus according to Embodiment 6 of the present invention.
  • the inverse quantization inverse transform unit 123 inversely quantizes the encoded difference block (encoded coefficient Coeff) (S181).
  • the dividing unit 410 divides the dequantized encoded difference block into the second part and the second converted output (S182).
  • the second inverse transform unit 460 generates the first part by performing inverse transform on the second transform output using the inverse transform matrix determined by the predictive transform control unit 130 (S183).
  • the integration unit 430 generates the first conversion output by integrating the first part and the second part (S184).
  • the first inverse transform unit 450 generates a difference block by inversely transforming the first transform output (S185).
  • the transformation matrix used by the first transformation unit 400 and the inverse transformation matrix used by the first inverse transformation unit 450 are also determined according to the transformation matrix or inverse transformation matrix determination method described in the first to fifth embodiments. May be.
  • division integration information CtrlTsel may be information indicating the division method or the integration method optimized for the design angle of the transformation matrix or the inverse transformation matrix described in the first to fifth embodiments. The concept of such division integration information will be described with reference to FIGS. 21A and 21B.
  • FIG. 21A is a diagram illustrating an example of a first conversion output of a 4 ⁇ 4 block.
  • the upper left is the low range.
  • the first portion is further compressed by the second conversion, with the low-frequency side where energy tends to concentrate as the first portion and the other high-frequency portion as the second portion.
  • Gij ⁇ 0, 1 ⁇ is used to indicate to which part the N ⁇ N elements of the first conversion output belong.
  • i indicates a position in the vertical downward direction
  • FIG. 21B is a conceptual diagram for explaining a method for selecting one piece of division integration information from a plurality of division integration information.
  • the division integration information also has a range from 0 to ⁇ [rad] with the upper right direction as the base point of the angle. Is pre-defined in a design angle unit that divides. In the example of FIG. 21B, four pieces of division integration information are defined.
  • the prediction conversion control unit 130 determines the division integration information based on the detected angle of the edge vector. Specifically, the predictive conversion control unit 130 selects the division integration information corresponding to the design angle closest to the angle of the edge vector from the plurality of division integration information held in association with the plurality of design angles. To do.
  • FIG. 22 is a diagram for explaining an example of eight pieces of division integration information corresponding to eight design angles on a one-to-one basis.
  • the first part is composed of four elements included in a 4 ⁇ 4 block.
  • the division integration information can be defined so that the design angle and the position of the first part have a relationship. That is, the dividing unit 410 can determine the position of the first portion with respect to an arbitrary angle by performing the dividing process according to such divided integration information.
  • the division integration information is designed for a specific angle. Therefore, when determining the angle with respect to the adjacent block, the predictive conversion control unit 130 determines the angle with respect to the adjacent block based on the design angle of the division integration information in addition to the angle of the edge vector and the design angle of the conversion matrix. May be.
  • the encoding device or the decoding device According to the encoding device or the decoding device according to the present embodiment, it is possible to appropriately determine a transformation matrix or an inverse transformation matrix even in the two-stage transformation, and to improve the coding efficiency. It becomes possible.
  • the processing described in each of the above embodiments is performed by recording a program for realizing the configuration of the image (moving image) encoding method or the image (moving image) decoding method described in each of the above embodiments on a storage medium.
  • the storage medium may be any medium that can record a program, such as a magnetic disk, an optical disk, a magneto-optical disk, an IC card, and a semiconductor memory.
  • FIG. 23 is a diagram showing an overall configuration of a content supply system ex100 that realizes a content distribution service.
  • the communication service providing area is divided into desired sizes, and base stations ex106, ex107, ex108, ex109, and ex110, which are fixed wireless stations, are installed in each cell.
  • the content supply system ex100 includes a computer ex111, a PDA (Personal Digital Assistant) ex112, a camera ex113, a mobile phone ex114, a game machine ex115 via the Internet ex101, the Internet service provider ex102, the telephone network ex104, and the base stations ex106 to ex110. Etc. are connected.
  • PDA Personal Digital Assistant
  • each device may be directly connected to the telephone network ex104 without going through the base stations ex106 to ex110 which are fixed wireless stations.
  • the devices may be directly connected to each other via short-range wireless or the like.
  • the camera ex113 is a device that can shoot moving images such as a digital video camera
  • the camera ex116 is a device that can shoot still images and movies such as a digital camera.
  • the mobile phone ex114 is a GSM (Global System for Mobile Communications) method, a CDMA (Code Division Multiple Access) method, a W-CDMA (Wideband-Code Division Multiple Access L (Semiconductor Access) method, a W-CDMA (Wideband-Code Division Multiple Access L method, or a high access rate).
  • GSM Global System for Mobile Communications
  • CDMA Code Division Multiple Access
  • W-CDMA Wideband-Code Division Multiple Access L (Semiconductor Access) method
  • W-CDMA Wideband-Code Division Multiple Access L method
  • a high access rate A High Speed Packet Access
  • PHS Personal Handyphone System
  • the camera ex113 and the like are connected to the streaming server ex103 through the base station ex109 and the telephone network ex104, thereby enabling live distribution and the like.
  • the content for example, music live video
  • the streaming server ex103 streams the content data transmitted to the requested client.
  • the client include a computer ex111, a PDA ex112, a camera ex113, a mobile phone ex114, a game machine ex115, and the like that can decode the encoded data.
  • Each device that has received the distributed data decodes and reproduces the received data.
  • the encoded processing of the captured data may be performed by the camera ex113, the streaming server ex103 that performs the data transmission processing, or may be performed in a shared manner.
  • the decryption processing of the distributed data may be performed by the client, the streaming server ex103, or may be performed in a shared manner.
  • still images and / or moving image data captured by the camera ex116 may be transmitted to the streaming server ex103 via the computer ex111.
  • the encoding process in this case may be performed by any of the camera ex116, the computer ex111, and the streaming server ex103, or may be performed in a shared manner.
  • encoding / decoding processes are generally performed by the computer ex111 and the LSI ex500 included in each device.
  • LSIex500 may be even configured of a single chip or a plurality of chips.
  • moving image encoding / decoding software may be incorporated in any recording medium (CD-ROM, flexible disk, hard disk, etc.) that can be read by the computer ex111 and the like, and encoding / decoding processing may be performed using the software. Good.
  • moving image data acquired by the camera may be transmitted. The moving image data at this time is data encoded by the LSI ex500 included in the mobile phone ex114.
  • the streaming server ex103 may be a plurality of servers or a plurality of computers, and may process, record, and distribute data in a distributed manner.
  • the encoded data can be received and reproduced by the client.
  • the information transmitted by the user can be received, decrypted and reproduced in real time by the client, and even a user who does not have special rights or facilities can realize personal broadcasting.
  • At least one of the video encoding device and the video decoding device of each of the above embodiments is incorporated in the digital broadcasting system ex200.
  • the broadcast station ex201 multiplexed data obtained by multiplexing music data and the like on video data is transmitted to a communication or satellite ex202 via radio waves.
  • This video data is data encoded by the moving image encoding method described in the above embodiments.
  • the broadcasting satellite ex202 transmits a radio wave for broadcasting, and the home antenna ex204 capable of receiving the satellite broadcast receives the radio wave.
  • the received multiplexed data is decoded and reproduced by a device such as the television (receiver) ex300 or the set top box (STB) ex217.
  • the reader / recorder ex218 also reads and decodes multiplexed data recorded on a recording medium ex215 such as a DVD or BD, or encodes a video signal on the recording medium ex215 and, in some cases, multiplexes and writes it with a music signal. It is possible to implement the moving picture decoding apparatus or moving picture encoding apparatus shown in each embodiment. In this case, the reproduced video signal is displayed on the monitor ex219, and the video signal can be reproduced in another device or system by the recording medium ex215 on which the multiplexed data is recorded.
  • a moving picture decoding device may be mounted in a set-top box ex217 connected to a cable ex203 for cable television or an antenna ex204 for satellite / terrestrial broadcasting, and this may be displayed on a monitor ex219 of the television.
  • the moving picture decoding apparatus may be incorporated in the television instead of the set top box.
  • FIG. 25 is a diagram illustrating a television (receiver) ex300 that uses the moving picture decoding method and the moving picture encoding method described in the above embodiments.
  • the television ex300 obtains or outputs multiplexed data in which audio data is multiplexed with video data via the antenna ex204 or the cable ex203 that receives the broadcast, and demodulates the received multiplexed data.
  • the modulation / demodulation unit ex302 that modulates multiplexed data to be transmitted to the outside, and the demodulated multiplexed data is separated into video data and audio data, or the video data and audio data encoded by the signal processing unit ex306 Is provided with a multiplexing / separating unit ex303.
  • the television ex300 decodes each of the audio data and the video data, or encodes each information, an audio signal processing unit ex304, a signal processing unit ex306 including the video signal processing unit ex305, and outputs the decoded audio signal.
  • the television ex300 includes an interface unit ex317 including an operation input unit ex312 that receives an input of a user operation.
  • the television ex300 includes a control unit ex310 that controls each unit in an integrated manner, and a power supply circuit unit ex311 that supplies power to each unit.
  • the interface unit ex317 includes a bridge ex313 connected to an external device such as a reader / recorder ex218, a recording unit ex216 such as an SD card, and an external recording such as a hard disk.
  • a driver ex315 for connecting to a medium, a modem ex316 for connecting to a telephone network, and the like may be included.
  • the recording medium ex216 is capable of electrically recording information by using a nonvolatile / volatile semiconductor memory element to be stored.
  • Each part of the television ex300 is connected to each other via a synchronous bus.
  • the television ex300 receives a user operation from the remote controller ex220 or the like, and demultiplexes the multiplexed data demodulated by the modulation / demodulation unit ex302 by the multiplexing / demultiplexing unit ex303 based on the control of the control unit ex310 having a CPU or the like. Furthermore, in the television ex300, the separated audio data is decoded by the audio signal processing unit ex304, and the separated video data is decoded by the video signal processing unit ex305 using the decoding method described in the above embodiments.
  • the decoded audio signal and video signal are output to the outside from the output unit ex309.
  • these signals may be temporarily stored in the buffers ex318, ex319, etc. so that the audio signal and the video signal are reproduced in synchronization.
  • the television ex300 may read multiplexed data from recording media ex215 and ex216 such as a magnetic / optical disk and an SD card, not from broadcasting.
  • recording media ex215 and ex216 such as a magnetic / optical disk and an SD card
  • the television ex300 receives a user operation from the remote controller ex220 or the like, and encodes an audio signal with the audio signal processing unit ex304 based on the control of the control unit ex310, and converts the video signal with the video signal processing unit ex305. Encoding is performed using the encoding method described in (1).
  • the encoded audio signal and video signal are multiplexed by the multiplexing / demultiplexing unit ex303 and output to the outside. When multiplexing, these signals may be temporarily stored in the buffers ex320 and ex321 so that the audio signal and the video signal are synchronized.
  • a plurality of buffers ex318, ex319, ex320, and ex321 may be provided as illustrated, or one or more buffers may be shared. Further, in addition to the illustrated example, data may be stored in the buffer as a buffer material that prevents system overflow and underflow, for example, between the modulation / demodulation unit ex302 and the multiplexing / demultiplexing unit ex303.
  • the television ex300 has a configuration for receiving AV input of a microphone and a camera, and performs encoding processing on the data acquired from them. Also good.
  • the television ex300 has been described as a configuration capable of the above-described encoding processing, multiplexing, and external output. However, these processing cannot be performed, and only the above-described reception, decoding processing, and external output are possible. It may be.
  • the decoding process or the encoding process may be performed by either the television ex300 or the reader / recorder ex218, or the television ex300 and the reader / Recorder ex218 may share each other.
  • FIG. 26 shows a configuration of the information reproducing / recording unit ex400 when data is read from or written to the optical disk.
  • the information reproducing / recording unit ex400 includes elements ex401, ex402, ex403, ex404, ex405, ex406, and ex407 described below.
  • the optical head ex401 irradiates a laser spot on the recording surface of the recording medium ex215 that is an optical disc to write information, and detects information reflected from the recording surface of the recording medium ex215 to read the information.
  • the modulation recording unit ex402 electrically drives a semiconductor laser built in the optical head ex401 and modulates the laser beam according to the recording data.
  • the reproduction demodulator ex403 amplifies the reproduction signal obtained by electrically detecting the reflected light from the recording surface by the photodetector built in the optical head ex401, separates and demodulates the signal component recorded on the recording medium ex215, and is necessary. To play back information.
  • the buffer ex404 temporarily holds information to be recorded on the recording medium ex215 and information reproduced from the recording medium ex215.
  • the disk motor ex405 rotates the recording medium ex215.
  • the servo control unit ex406 moves the optical head ex401 to a predetermined information track while controlling the rotational drive of the disk motor ex405, and performs a laser spot tracking process.
  • the system control unit ex407 controls the entire information reproduction / recording unit ex400.
  • the system control unit ex407 uses various kinds of information held in the buffer ex404, and generates and adds new information as necessary, as well as the modulation recording unit ex402, the reproduction demodulation unit This is realized by recording / reproducing information through the optical head ex401 while operating the ex403 and the servo control unit ex406 in a coordinated manner.
  • the system control unit ex407 is composed of, for example, a microprocessor, and executes these processes by executing a read / write program.
  • the optical head ex401 has been described as irradiating a laser spot, but it may be configured to perform higher-density recording using near-field light.
  • FIG. 27 shows a schematic diagram of a recording medium ex215 that is an optical disk.
  • Guide grooves grooves
  • address information indicating the absolute position on the disc is recorded in advance on the information track ex230 by changing the shape of the groove.
  • This address information includes information for specifying the position of the recording block ex231 that is a unit for recording data, and the recording block is specified by reproducing the information track ex230 and reading the address information in a recording or reproducing apparatus.
  • the recording medium ex215 includes a data recording area ex233, an inner peripheral area ex232, and an outer peripheral area ex234.
  • the area used for recording the user data is the data recording area ex233, and the inner circumference area ex232 and the outer circumference area ex234 arranged on the inner circumference or outer circumference of the data recording area ex233 are used for specific purposes other than user data recording. Used.
  • the information reproducing / recording unit ex400 reads / writes encoded audio data, video data, or multiplexed data obtained by multiplexing these data with respect to the data recording area ex233 of the recording medium ex215.
  • an optical disk such as a single-layer DVD or BD has been described as an example.
  • the present invention is not limited to these, and an optical disk having a multilayer structure and capable of recording other than the surface may be used.
  • an optical disc with a multi-dimensional recording / reproducing structure such as recording information using light of different wavelengths in the same place on the disc, or recording different layers of information from various angles. It may be.
  • the car ex210 having the antenna ex205 can receive data from the satellite ex202 and the like, and the moving image can be reproduced on a display device such as the car navigation ex211 that the car ex210 has.
  • the configuration of the car navigation ex211 may include a configuration in which a GPS receiving unit is added to the configuration illustrated in FIG. 25, and the same may be applied to the computer ex111, the mobile phone ex114, and the like.
  • FIG. 28 (a) is a diagram showing a mobile phone ex114 using the moving picture decoding method and the moving picture encoding method described in the above embodiment.
  • the mobile phone ex114 includes an antenna ex350 for transmitting and receiving radio waves to and from the base station ex110, a camera unit ex365 capable of taking video and still images, a video captured by the camera unit ex365, a video received by the antenna ex350, and the like Is provided with a display unit ex358 such as a liquid crystal display for displaying the decoded data.
  • the mobile phone ex114 further includes a main body unit having an operation key unit ex366, an audio output unit ex357 such as a speaker for outputting audio, an audio input unit ex356 such as a microphone for inputting audio, It is an interface unit with a memory unit ex367 that stores encoded data or decoded data such as still images, recorded audio, received video, still images, and mails, or a recording medium that similarly stores data.
  • a slot portion ex364 is provided.
  • the cellular phone ex114 has a power supply circuit ex361, an operation input control unit ex362, and a video signal processing unit ex355 for a main control unit ex360 that comprehensively controls each part of the main body including the display unit ex358 and the operation key unit ex366.
  • a camera interface unit ex363, an LCD (Liquid Crystal Display) control unit ex359, a modulation / demodulation unit ex352, a multiplexing / demultiplexing unit ex353, an audio signal processing unit ex354, a slot unit ex364, and a memory unit ex367 are connected to each other via a bus ex370. ing.
  • the power supply circuit unit ex361 starts up the mobile phone ex114 in an operable state by supplying power from the battery pack to each unit.
  • the mobile phone ex114 converts the audio signal collected by the audio input unit ex356 in the voice call mode into a digital audio signal by the audio signal processing unit ex354 based on the control of the main control unit ex360 having a CPU, a ROM, a RAM, and the like. This is subjected to spectrum spread processing by the modulation / demodulation unit ex352, digital-analog conversion processing and frequency conversion processing by the transmission / reception unit ex351, and then transmitted via the antenna ex350.
  • the mobile phone ex114 amplifies the received data received through the antenna ex350 in the voice call mode, performs frequency conversion processing and analog-digital conversion processing, performs spectrum despreading processing in the modulation / demodulation unit ex352, and performs voice signal processing unit After converting to an analog audio signal at ex354, this is output from the audio output unit ex357.
  • the text data of the e-mail input by operating the operation key unit ex366 of the main unit is sent to the main control unit ex360 via the operation input control unit ex362.
  • the main control unit ex360 performs spread spectrum processing on the text data in the modulation / demodulation unit ex352, performs digital analog conversion processing and frequency conversion processing in the transmission / reception unit ex351, and then transmits the text data to the base station ex110 via the antenna ex350.
  • almost the reverse process is performed on the received data and output to the display unit ex358.
  • the video signal processing unit ex355 compresses the video signal supplied from the camera unit ex365 by the moving image encoding method described in the above embodiments.
  • the encoded video data is sent to the multiplexing / demultiplexing unit ex353.
  • the audio signal processing unit ex354 encodes the audio signal picked up by the audio input unit ex356 while the camera unit ex365 images a video, a still image, and the like, and sends the encoded audio data to the multiplexing / demultiplexing unit ex353. To do.
  • the multiplexing / demultiplexing unit ex353 multiplexes the encoded video data supplied from the video signal processing unit ex355 and the encoded audio data supplied from the audio signal processing unit ex354 by a predetermined method, and is obtained as a result.
  • the multiplexed data is subjected to spread spectrum processing by the modulation / demodulation unit (modulation / demodulation circuit unit) ex352, digital-analog conversion processing and frequency conversion processing by the transmission / reception unit ex351, and then transmitted through the antenna ex350.
  • the multiplexing / separating unit ex353 separates the multiplexed data into a video data bit stream and an audio data bit stream, and the video data encoded via the synchronization bus ex370 is a video signal processing unit.
  • the encoded audio data is supplied to the audio signal processing unit ex354.
  • the video signal processing unit ex355 decodes the video signal by decoding using a video decoding method corresponding to the video encoding method shown in each of the above embodiments, and from the display unit ex358 via the LCD control unit ex359, For example, video and still images included in a moving image file linked to a home page are displayed.
  • the audio signal processing unit ex354 decodes the audio signal, and the audio output unit ex357 outputs the audio.
  • the transmission / reception terminal having both an encoder and a decoder
  • there are three types of terminals such as the mobile phone ex114, such as a transmitting terminal having only an encoder and a receiving terminal having only a decoder.
  • the implementation form of can be considered.
  • multiplexed data in which music data is multiplexed with video data is received and transmitted.
  • character data related to video is multiplexed. It may be converted data, or may be video data itself instead of multiplexed data.
  • the moving picture encoding method or the moving picture decoding method described in each of the above embodiments can be used in any of the above-described devices / systems, and as a result, described in each of the above embodiments. Effects can be obtained.
  • multiplexed data obtained by multiplexing audio data or the like with video data is configured to include identification information indicating which standard the video data conforms to.
  • identification information indicating which standard the video data conforms to.
  • FIG. 29 is a diagram showing a structure of multiplexed data.
  • multiplexed data is obtained by multiplexing one or more of a video stream, an audio stream, a presentation graphics stream (PG), and an interactive graphics stream.
  • the video stream indicates the main video and sub-video of the movie
  • the audio stream (IG) indicates the main audio portion of the movie and the sub-audio mixed with the main audio
  • the presentation graphics stream indicates the subtitles of the movie.
  • the main video indicates a normal video displayed on the screen
  • the sub-video is a video displayed on a small screen in the main video.
  • the interactive graphics stream indicates an interactive screen created by arranging GUI components on the screen.
  • the video stream is encoded by the moving image encoding method or apparatus shown in the above embodiments, or the moving image encoding method or apparatus conforming to the conventional standards such as MPEG-2, MPEG4-AVC, and VC-1. ing.
  • the audio stream is encoded by a method such as Dolby AC-3, Dolby Digital Plus, MLP, DTS, DTS-HD, or linear PCM.
  • Each stream included in the multiplexed data is identified by PID. For example, 0x1011 for video streams used for movie images, 0x1100 to 0x111F for audio streams, 0x1200 to 0x121F for presentation graphics, 0x1400 to 0x141F for interactive graphics streams, 0x1B00 to 0x1B1F are assigned to video streams used for sub-pictures, and 0x1A00 to 0x1A1F are assigned to audio streams used for sub-audio mixed with the main audio.
  • FIG. 30 is a diagram schematically showing how multiplexed data is multiplexed.
  • a video stream ex235 composed of a plurality of video frames and an audio stream ex238 composed of a plurality of audio frames are converted into PES packet sequences ex236 and ex239, respectively, and converted into TS packets ex237 and ex240.
  • the data of the presentation graphics stream ex241 and interactive graphics ex244 are converted into PES packet sequences ex242 and ex245, respectively, and further converted into TS packets ex243 and ex246.
  • the multiplexed data ex247 is configured by multiplexing these TS packets into one stream.
  • FIG. 31 shows in more detail how the video stream is stored in the PES packet sequence.
  • the first row in FIG. 31 shows a video frame sequence of the video stream.
  • the second level shows a PES packet sequence.
  • a plurality of Video Presentation Units in a video stream are divided into pictures, stored in the payload of the PES packet.
  • Each PES packet has a PES header, and a PTS (Presentation Time-Stamp) that is a display time of a picture and a DTS (Decoding Time-Stamp) that is a decoding time of a picture are stored in the PES header.
  • PTS Presentation Time-Stamp
  • DTS Decoding Time-Stamp
  • FIG. 32 shows the format of TS packets that are finally written in the multiplexed data.
  • the TS packet is a 188-byte fixed-length packet composed of a 4-byte TS header having information such as a PID for identifying a stream and a 184-byte TS payload for storing data.
  • the PES packet is divided and stored in the TS payload.
  • a 4-byte TP_Extra_Header is added to a TS packet, forms a 192-byte source packet, and is written in multiplexed data.
  • TP_Extra_Header information such as ATS (Arrival_Time_Stamp) is described.
  • ATS indicates the transfer start time of the TS packet to the PID filter of the decoder.
  • Source packets are arranged in the multiplexed data as shown in the lower part of FIG. 32, and the number incremented from the head of the multiplexed data is called SPN (source packet number).
  • TS packets included in the multiplexed data include PAT (Program Association Table), PMT (Program Map Table), PCR (Program Clock Reference), and the like in addition to each stream such as video / audio / caption.
  • PAT indicates what the PID of the PMT used in the multiplexed data is, and the PID of the PAT itself is registered as 0.
  • the PMT has the PID of each stream such as video / audio / subtitles included in the multiplexed data and the attribute information of the stream corresponding to each PID, and has various descriptors related to the multiplexed data.
  • the descriptor includes copy control information for instructing permission / non-permission of copying of multiplexed data.
  • the PCR corresponds to the ATS in which the PCR packet is transferred to the decoder. Contains STC time information.
  • FIG. 33 is a diagram for explaining the data structure of the PMT in detail.
  • a PMT header describing the length of data included in the PMT is arranged at the head of the PMT. After that, a plurality of descriptors related to multiplexed data are arranged. The copy control information and the like are described as descriptors. After the descriptor, a plurality of pieces of stream information regarding each stream included in the multiplexed data are arranged.
  • the stream information is composed of a stream descriptor describing a stream type, a stream PID, and stream attribute information (frame rate, aspect ratio, etc.) for identifying a compression codec of the stream. There are as many stream descriptors as the number of streams existing in the multiplexed data.
  • the multiplexed data is recorded together with the multiplexed data information file.
  • the multiplexed data information file is management information of multiplexed data, has a one-to-one correspondence with the multiplexed data, and includes multiplexed data information, stream attribute information, and an entry map.
  • the multiplexed data information is composed of a system rate, a reproduction start time, and a reproduction end time.
  • the system rate indicates a maximum transfer rate of multiplexed data to a PID filter of a system target decoder described later.
  • the ATS interval included in the multiplexed data is set to be equal to or less than the system rate.
  • the playback start time is the PTS of the first video frame of the multiplexed data
  • the playback end time is set by adding the playback interval for one frame to the PTS of the video frame at the end of the multiplexed data.
  • attribute information about each stream included in the multiplexed data is registered for each PID.
  • the attribute information has different information for each video stream, audio stream, presentation graphics stream, and interactive graphics stream.
  • the video stream attribute information includes the compression codec used to compress the video stream, the resolution of the individual picture data constituting the video stream, the aspect ratio, and the frame rate. It has information such as how much it is.
  • the audio stream attribute information includes the compression codec used to compress the audio stream, the number of channels included in the audio stream, the language supported, and the sampling frequency. With information. These pieces of information are used for initialization of the decoder before the player reproduces it.
  • the stream type included in the PMT is used.
  • video stream attribute information included in the multiplexed data information is used.
  • the video encoding shown in each of the above embodiments for the stream type or video stream attribute information included in the PMT.
  • FIG. 36 shows the steps of the moving picture decoding method according to the present embodiment.
  • step exS100 the stream type included in the PMT or the video stream attribute information included in the multiplexed data information is acquired from the multiplexed data.
  • step exS101 it is determined whether or not the stream type or the video stream attribute information indicates multiplexed data generated by the moving picture encoding method or apparatus described in the above embodiments. To do.
  • step exS102 each of the above embodiments.
  • Decoding is performed by the moving picture decoding method shown in the form.
  • the conventional information Decoding is performed by a moving image decoding method compliant with the standard.
  • the moving picture encoding method or apparatus or the moving picture decoding method or apparatus described in this embodiment can be used in any of the above-described devices and systems.
  • FIG. 37 shows the configuration of an LSI ex500 that is made into one chip.
  • the LSI ex500 includes elements ex501, ex502, ex503, ex504, ex505, ex506, ex507, ex508, and ex509 described below, and each element is connected via a bus ex510.
  • the power supply circuit unit ex505 starts up to an operable state by supplying power to each unit when the power supply is in an on state.
  • the LSI ex500 uses the AV I / O ex509 to perform the microphone ex117 and the camera ex113 based on the control of the control unit ex501 including the CPU ex502, the memory controller ex503, the stream controller ex504, the drive frequency control unit ex512, and the like.
  • the AV signal is input from the above.
  • the input AV signal is temporarily stored in an external memory ex511 such as SDRAM.
  • the accumulated data is divided into a plurality of times as appropriate according to the processing amount and the processing speed and sent to the signal processing unit ex507, and the signal processing unit ex507 encodes an audio signal and / or video. Signal encoding is performed.
  • the encoding process of the video signal is the encoding process described in the above embodiments.
  • the signal processing unit ex507 further performs processing such as multiplexing the encoded audio data and the encoded video data according to circumstances, and outputs the result from the stream I / Oex 506 to the outside.
  • the output multiplexed data is transmitted to the base station ex107 or written to the recording medium ex215. It should be noted that data should be temporarily stored in the buffer ex508 so as to be synchronized when multiplexing.
  • the memory ex511 has been described as an external configuration of the LSI ex500.
  • a configuration included in the LSI ex500 may be used.
  • the number of buffers ex508 is not limited to one, and a plurality of buffers may be provided.
  • the LSI ex500 may be made into one chip or a plurality of chips.
  • control unit ex501 includes the CPU ex502, the memory controller ex503, the stream controller ex504, the drive frequency control unit ex512, and the like, but the configuration of the control unit ex501 is not limited to this configuration.
  • the signal processing unit ex507 may further include a CPU.
  • the CPU ex502 may be configured to include a signal processing unit ex507 or, for example, an audio signal processing unit that is a part of the signal processing unit ex507.
  • the control unit ex501 is configured to include a signal processing unit ex507 or a CPU ex502 having a part thereof.
  • LSI LSI
  • IC system LSI
  • super LSI ultra LSI depending on the degree of integration
  • the method of circuit integration is not limited to LSI, and implementation with a dedicated circuit or a general-purpose processor is also possible.
  • An FPGA Field Programmable Gate Array
  • a reconfigurable processor that can reconfigure the connection and setting of circuit cells inside the LSI may be used.
  • FIG. 38 shows a configuration ex800 in the present embodiment.
  • the drive frequency switching unit ex803 sets the drive frequency high when the video data is generated by the moving image encoding method or apparatus described in the above embodiments. Then, it instructs the decoding processing unit ex801 that executes the moving picture decoding method described in each of the above embodiments to decode the video data.
  • the video data is video data compliant with the conventional standard, compared to the case where the video data is generated by the moving picture encoding method or apparatus shown in the above embodiments, Set the drive frequency low. Then, it instructs the decoding processing unit ex802 compliant with the conventional standard to decode the video data.
  • the drive frequency switching unit ex803 includes the CPU ex502 and the drive frequency control unit ex512 in FIG.
  • the decoding processing unit ex801 that executes the moving picture decoding method shown in each of the above embodiments and the decoding processing unit ex802 that complies with the conventional standard correspond to the signal processing unit ex507 in FIG.
  • the CPU ex502 identifies which standard the video data conforms to. Then, based on the signal from the CPU ex502, the drive frequency control unit ex512 sets the drive frequency. Further, based on the signal from the CPU ex502, the signal processing unit ex507 decodes the video data.
  • the identification of the video data for example, it is conceivable to use the identification information described in the eighth embodiment.
  • the identification information is not limited to that described in Embodiment 8, and any information that can identify which standard the video data conforms to may be used. For example, it is possible to identify which standard the video data conforms to based on an external signal that identifies whether the video data is used for a television or a disk. In some cases, identification may be performed based on such an external signal. Further, the selection of the drive frequency in the CPU ex502 may be performed based on, for example, a look-up table in which video data standards and drive frequencies are associated with each other as shown in FIG. The look-up table is stored in the buffer ex508 or the internal memory of the LSI, and the CPU ex502 can select the drive frequency by referring to this look-up table.
  • FIG. 39 shows steps for executing the method of the present embodiment.
  • the signal processing unit ex507 acquires identification information from the multiplexed data.
  • the CPU ex502 identifies whether the video data is generated by the encoding method or apparatus described in each of the above embodiments based on the identification information.
  • the CPU ex502 sends a signal for setting the drive frequency high to the drive frequency control unit ex512. Then, the drive frequency control unit ex512 sets a high drive frequency.
  • step exS203 the CPU ex502 drives a signal for setting the drive frequency low. This is sent to the frequency control unit ex512. Then, in the drive frequency control unit ex512, the drive frequency is set to be lower than that in the case where the video data is generated by the encoding method or apparatus described in the above embodiments.
  • the power saving effect can be further enhanced by changing the voltage applied to the LSI ex500 or the device including the LSI ex500 in conjunction with the switching of the driving frequency.
  • the drive frequency is set to be low, it is conceivable that the voltage applied to the LSI ex500 or the device including the LSI ex500 is set low as compared with the case where the drive frequency is set high.
  • the setting method of the driving frequency may be set to a high driving frequency when the processing amount at the time of decoding is large, and to a low driving frequency when the processing amount at the time of decoding is small. It is not limited to the method.
  • the amount of processing for decoding video data compliant with the MPEG4-AVC standard is larger than the amount of processing for decoding video data generated by the moving picture encoding method or apparatus described in the above embodiments. It is conceivable that the setting of the driving frequency is reversed to that in the case described above.
  • the method for setting the drive frequency is not limited to the configuration in which the drive frequency is lowered.
  • the voltage applied to the LSI ex500 or the apparatus including the LSI ex500 is set high.
  • the video data conforms to the conventional standards such as MPEG-2, MPEG4-AVC, VC-1, etc.
  • the identification information indicates that the video data is generated by the moving image encoding method or apparatus described in each of the above embodiments, the driving of the CPU ex502 is stopped.
  • the CPU ex502 is temporarily stopped because there is enough processing. Is also possible. Even when the identification information indicates that the video data is generated by the moving image encoding method or apparatus described in each of the above embodiments, if there is enough processing, the CPU ex502 is temporarily driven. It can also be stopped. In this case, it is conceivable to set the stop time shorter than in the case where the video data conforms to the conventional standards such as MPEG-2, MPEG4-AVC, and VC-1.
  • a plurality of video data that conforms to different standards may be input to the above-described devices and systems such as a television and a mobile phone.
  • the signal processing unit ex507 of the LSI ex500 needs to support a plurality of standards in order to be able to decode even when a plurality of video data complying with different standards is input.
  • the signal processing unit ex507 corresponding to each standard is individually used, there is a problem that the circuit scale of the LSI ex500 increases and the cost increases.
  • a decoding processing unit for executing the moving picture decoding method shown in each of the above embodiments and a decoding conforming to a standard such as MPEG-2, MPEG4-AVC, or VC-1
  • the processing unit is partly shared.
  • An example of this configuration is shown as ex900 in FIG.
  • the moving picture decoding method shown in each of the above embodiments and the moving picture decoding method compliant with the MPEG4-AVC standard are processed in processes such as entropy coding, inverse quantization, deblocking filter, and motion compensation. Some contents are common.
  • the decoding processing unit ex902 corresponding to the MPEG4-AVC standard is shared, and for other processing contents unique to the present invention that do not correspond to the MPEG4-AVC standard, the dedicated decoding processing unit ex901 is used.
  • Configuration is conceivable.
  • a dedicated decoding processing unit ex901 is used for transformation or inverse transformation, and other entropy coding, deblocking filter, motion, etc. It is conceivable to share the decoding processing unit for any or all of the compensation.
  • the decoding processing unit for executing the moving picture decoding method described in each of the above embodiments is shared, and the processing content unique to the MPEG4-AVC standard May be configured to use a dedicated decoding processing unit.
  • ex1000 in FIG. 41 (b) shows another example in which processing is partially shared.
  • a dedicated decoding processing unit ex1001 corresponding to processing content specific to the present invention
  • a dedicated decoding processing unit ex1002 corresponding to processing content specific to other conventional standards
  • a moving picture decoding method of the present invention A common decoding processing unit ex1003 corresponding to processing contents common to other conventional video decoding methods is used.
  • the dedicated decoding processing units ex1001 and ex1002 are not necessarily specialized in the processing content specific to the present invention or other conventional standards, and may be capable of executing other general-purpose processing.
  • the configuration of the present embodiment can be implemented by LSI ex500.
  • the circuit scale of the LSI can be reduced and the cost can be reduced. It is possible to reduce.
  • the encoding efficiency can be improved and it can be used for various purposes such as storage, transmission, and communication.
  • it can be used for high-resolution information display devices and imaging devices such as televisions, digital video recorders, car navigation systems, mobile phones, digital cameras, and digital video cameras, and has high utility value.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Discrete Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne un procédé de décodage d'images qui décode une image codée bloc par bloc, le procédé comprenant les étapes consistant à détecter un vecteur de bord au sein d'au moins un bloc décodé qui est spatialement proche d'un bloc à traiter, ou au sein d'un bloc décodé qui est temporellement proche d'un bloc à traiter (S21) ; déterminer une matrice de transformée inverse du bloc à traiter sur la base de l'angle du vecteur de bord détecté (S22) ; quantifier de manière inverse un bloc de différence codé qui correspond au bloc à traiter (S23) ; générer un bloc de différence en utilisant la matrice de transformée inverse déterminée afin de transformer de manière inverse le bloc de différence codé et quantifié de manière inverse (S24) ; générer un bloc de prédiction en prédisant le bloc à traiter (S25) ; et générer le bloc décodé en ajoutant le bloc de prédiction et le bloc de différence (S26).
PCT/JP2011/002723 2010-05-18 2011-05-17 Procédé de décodage d'images, procédé de codage d'images, dispositif de décodage d'images, et dispositif de codage d'images WO2011145328A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US34570310P 2010-05-18 2010-05-18
US61/345,703 2010-05-18

Publications (1)

Publication Number Publication Date
WO2011145328A1 true WO2011145328A1 (fr) 2011-11-24

Family

ID=44991446

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2011/002723 WO2011145328A1 (fr) 2010-05-18 2011-05-17 Procédé de décodage d'images, procédé de codage d'images, dispositif de décodage d'images, et dispositif de codage d'images

Country Status (1)

Country Link
WO (1) WO2011145328A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013105367A1 (fr) * 2012-01-13 2013-07-18 ルネサスエレクトロニクス株式会社 Procédé de décodage de données d'image, dispositif pour décoder des données d'image, procédé de codage de données d'image, dispositif pour coder des données d'image
CN109993762A (zh) * 2019-02-21 2019-07-09 华南理工大学 一种基于非对称逆布局模式的快速边缘检测方法

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009147843A (ja) * 2007-12-18 2009-07-02 Sony Corp 符号化装置及び方法、復号装置及び方法、並びにプログラム
WO2009090884A1 (fr) * 2008-01-18 2009-07-23 Panasonic Corporation Procédé de codage d'image et procédé de décodage d'image

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009147843A (ja) * 2007-12-18 2009-07-02 Sony Corp 符号化装置及び方法、復号装置及び方法、並びにプログラム
WO2009090884A1 (fr) * 2008-01-18 2009-07-23 Panasonic Corporation Procédé de codage d'image et procédé de décodage d'image

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ALI TABATABAI ET AL.: "AVC based Intra Prediction for Improved Visual Quality", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/ SC29/WG11 1ST MEETING, 15 April 2010 (2010-04-15), DRESDEN, DE, pages 1 - 4 *
I. AMOUOU ET AL.: "Description of video coding technology proposal by France Telecom", NTT, 15 April 2010 (2010-04-15), DRESDEN, DE, pages 1 - 42 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013105367A1 (fr) * 2012-01-13 2013-07-18 ルネサスエレクトロニクス株式会社 Procédé de décodage de données d'image, dispositif pour décoder des données d'image, procédé de codage de données d'image, dispositif pour coder des données d'image
CN109993762A (zh) * 2019-02-21 2019-07-09 华南理工大学 一种基于非对称逆布局模式的快速边缘检测方法
CN109993762B (zh) * 2019-02-21 2022-12-16 华南理工大学 一种基于非对称逆布局模式的快速边缘检测方法

Similar Documents

Publication Publication Date Title
JP6643665B2 (ja) 動画像符号化方法、動画像復号化方法、動画像符号化装置および動画像復号化装置
JP6112320B2 (ja) 動画像復号化方法および動画像符号化方法
JP6064235B2 (ja) 符号化方法、プログラムおよび符号化装置
JP6090625B2 (ja) 画像復号方法及び画像復号装置
WO2012117728A1 (fr) Procédé de codage d'images vidéo, procédé de décodage d'images vidéo, dispositif de codage d'images vidéo, dispositif de décodage d'images vidéo et dispositif de codage/décodage d'images vidéo
JP6210375B2 (ja) 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置及び画像符号化復号装置
WO2012023281A1 (fr) Procédé de décodage d'image vidéo, procédé de codage d'image vidéo, appareil de décodage d'image vidéo, appareil de codage d'image vidéo
JP6414712B2 (ja) 多数の参照ピクチャを用いる動画像符号化方法、動画像復号方法、動画像符号化装置、および動画像復号方法
WO2012120840A1 (fr) Procédé de décodage d'image, procédé de codage d'image, dispositif de décodage d'image et dispositif de codage d'image
JPWO2011129090A1 (ja) 符号化歪み除去方法、符号化方法、復号化方法、符号化歪み除去装置、符号化装置、および復号化装置
WO2012039136A1 (fr) Procédé de codage d'image, procédé de décodage d'image, appareil de codage d'image et appareil de décodage d'image
WO2013001813A1 (fr) Procédé de codage d'image, procédé de décodage d'image, dispositif de codage d'image et dispositif de décodage d'image
JP6399433B2 (ja) 画像符号化方法、画像復号方法、画像符号化装置及び画像復号装置
JP6187887B2 (ja) 符号化方法および符号化装置
WO2011135841A1 (fr) Procédé de codage d'image, procédé de décodage d'image, appareil de codage d'image et appareil de décodage d'image
WO2011145328A1 (fr) Procédé de décodage d'images, procédé de codage d'images, dispositif de décodage d'images, et dispositif de codage d'images
WO2012042810A1 (fr) Procédé de codage d'image, procédé de décodage d'image, dispositif de codage d'image, dispositif de décodage d'image et système de traitement d'image
WO2012096157A1 (fr) Procédé de codage d'image, procédé de décodage d'image, dispositif de codage d'image et dispositif de décodage d'image
WO2012077347A1 (fr) Procédé de décodage
WO2012086166A1 (fr) Procédé de codage d'image et procédé de décodage d'image

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11783265

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11783265

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP