WO2006072992A1 - 圧縮符号化装置、伸張復号化装置 - Google Patents
圧縮符号化装置、伸張復号化装置 Download PDFInfo
- Publication number
- WO2006072992A1 WO2006072992A1 PCT/JP2005/000122 JP2005000122W WO2006072992A1 WO 2006072992 A1 WO2006072992 A1 WO 2006072992A1 JP 2005000122 W JP2005000122 W JP 2005000122W WO 2006072992 A1 WO2006072992 A1 WO 2006072992A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- frame
- reference camera
- captured
- imaged
- camera
- Prior art date
Links
- 230000006835 compression Effects 0.000 title claims description 65
- 238000007906 compression Methods 0.000 title claims description 65
- 230000006837 decompression Effects 0.000 title claims description 25
- 230000033001 locomotion Effects 0.000 claims abstract description 165
- 238000000034 method Methods 0.000 claims abstract description 36
- 239000013598 vector Substances 0.000 claims description 51
- 230000010365 information processing Effects 0.000 claims description 13
- 239000000284 extract Substances 0.000 claims description 4
- 230000015572 biosynthetic process Effects 0.000 claims description 3
- 238000003384 imaging method Methods 0.000 claims description 3
- 238000003786 synthesis reaction Methods 0.000 claims description 3
- 230000002194 synthesizing effect Effects 0.000 claims description 3
- 239000000872 buffer Substances 0.000 description 38
- 238000013139 quantization Methods 0.000 description 30
- 238000004458 analytical method Methods 0.000 description 21
- 238000012545 processing Methods 0.000 description 20
- 230000008569 process Effects 0.000 description 15
- 238000010586 diagram Methods 0.000 description 13
- 238000005516 engineering process Methods 0.000 description 5
- 102100037812 Medium-wave-sensitive opsin 1 Human genes 0.000 description 3
- 230000002457 bidirectional effect Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 238000004091 panning Methods 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000002427 irreversible effect Effects 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/105—Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
- H04N19/137—Motion inside a coding unit, e.g. average field, frame or block difference
- H04N19/139—Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/172—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
Definitions
- Compression encoding device decompression decoding device
- the present invention relates to a technique that is effective when applied to an apparatus or method for encoding / decoding an image captured from multiple viewpoints.
- the coding efficiency improves as the prediction efficiency improves.
- the correlation between moving images taken at each viewpoint means that a part of the object or background reflected in a certain viewpoint camera is also reflected in the camera of another viewpoint.
- the frame at camera m and the camera n Compared to the same time frame, the same object and the same background may be captured when both cameras are capturing images in close directions at close positions. For this reason, it is possible to perform a prediction code using a motion vector by regarding a frame captured at the same time by such different cameras as a frame captured by the same camera.
- the frame by the other camera is used. It is possible to increase the efficiency of code prediction when using motion prediction code
- the motion vector corresponds to the parallax between the two cameras.
- Patent Document 1-17 there is Patent Document 1-17.
- Patent Document 1 JP 2001-186516 A
- Patent Document 2 Japanese Translation of Special Publication 2002-523943
- Patent Document 3 Japanese Patent Laid-Open No. 2002-300607
- Patent Document 4 Japanese Patent No. 3426668
- Patent Document 5 Japanese Patent Laid-Open No. 06-98312
- Patent Document 6 Japanese Patent Laid-Open No. 10-191394
- Patent Document 7 JP 2000-23918
- the reference camera refers to a camera that captures a frame used for prediction between cameras. Other cameras are called reference cameras. Conventionally, only the reference camera is set based on the arrangement information of each viewpoint camera, and no effective setting standard has been proposed. For this reason, the improvement of the efficiency of the sign key has not been realized sufficiently.
- FIGS. 9 and 10 are diagrams showing problems of the conventional technology.
- the problems of the conventional technology are explained using Figs.
- the three triangles indicate cameras CI, C2, and C3, respectively.
- three ellipses arranged in the direction of camera movement indicate the subject.
- Cameras CI, C2, and C3 capture the imaging direction while moving in the right direction.
- Figure 10 shows each camera CI, C2, C3 at each time T (nl), T (n), ⁇ ( ⁇ + 1). It is a figure which shows the example of the flame
- the present invention aims to solve these problems and provide an apparatus, a method, and the like that can improve prediction efficiency and encoding efficiency by selecting an optimal frame used for prediction.
- a frame captured by itself is compressed and encoded by motion prediction using only a frame captured by itself, and a frame captured by itself is captured by itself.
- a compression encoding device includes a determination unit, a compression encoding unit, a prediction information generation unit, and a synthesis unit.
- the determination unit determines the subject in the captured frame before the captured frame is captured. Based on the motion! /, Determine the frame of the other power mela that is used for motion prediction.
- the compression encoding unit compresses the frame captured by the reference camera by motion prediction using the frame of the other power mer determined by the determination unit and the other frame captured by the reference camera. Do sign y.
- the compression coding means performs compression coding based on intra-frame prediction or motion prediction using only other frames imaged by the reference camera for the frames imaged by the reference force camera. In other words, the compression encoding means uses the same method as the conventional method for predicting motion for each frame imaged by the reference camera using only the frame imaged by the same power memera. Do the trap.
- the prediction information creating means includes information indicating whether each frame is a frame imaged by the reference camera or a frame imaged by the reference camera, and for the frame imaged by the reference camera, the frame and motion Prediction information including information for associating with other frames used for prediction is generated.
- the synthesizing unit includes a single piece of moving image data including a plurality of frames after the code captured by the reference camera, a plurality of frames after the code captured by the reference camera, and motion prediction information. Is generated.
- the frame force used for motion prediction when the determination unit performs compression coding on the frame imaged by the reference camera is previously captured. It is determined based on the movement of the subject in the determined frame. Then, this frame captured by the reference camera is compressed and encoded by motion prediction using the frame determined by the determining means. For this reason, the optimum frame used for motion prediction is determined based on the motion of the subject in the past frame, and the prediction efficiency and the encoding efficiency are improved.
- the determining means provides a direction of a motion vector acquired at the time of motion prediction performed on a frame captured before the frame to be processed is captured. According to the above, the movement of the subject may be judged and the frame of the other power mela used for the motion prediction may be determined.
- the determining means in the first aspect of the present invention is that the frame to be processed is It is installed in the direction opposite to the direction of the motion vector acquired in the motion prediction performed in the captured frame before the target frame is captured with respect to the captured reference camera. It may be configured to determine a frame captured by another camera as a frame used for motion prediction. With this configuration, it is possible to perform motion prediction using a frame having a higher correlation with the processing target frame. Therefore, prediction efficiency and encoding efficiency can be improved.
- the first aspect of the present invention may be configured to further include reference camera determination means for determining which of the plurality of cameras should be the reference camera at a predetermined cycle. good.
- the reference camera is accurately selected, the optimal frame used for motion prediction is determined, and prediction efficiency and coding efficiency are improved.
- the reference camera determination unit determines the reference camera based on the movement of the subject in the captured frame before the frame to be processed is captured. It can be configured as
- the reference camera determination unit provides a motion vector acquired at the time of motion prediction performed on a frame captured before a frame to be processed is captured.
- the reference camera may be determined by judging the movement of the subject according to the orientation of the camera.
- the motion prediction information creating unit does not perform motion prediction using a frame imaged by another camera on a frame imaged by the reference camera.
- the compression-coded time information may be further included in the motion prediction information.
- a second aspect of the present invention is a decompression decoding apparatus that decompresses and decodes moving image data created by the compression coding apparatus according to the first aspect of the present invention, comprising a judging means and Includes decompression decoding means.
- the determining means extracts moving image data force motion prediction information, and determines for each frame whether the frame was captured by the reference camera or the reference camera. Then, the decompression decoding means is based on only the other frames imaged by the same camera for the frames determined by the determining means as the frames imaged by the reference camera. For the frame that is determined to be a frame captured by the reference camera by the determination means, the frame that has been captured by the same camera and the frame that is captured by the other camera are used. Performs extended decoding by motion estimation based on this.
- the first aspect and the second aspect may be realized by executing a program by the information processing apparatus. That is, the present invention specifies the processing executed by each means in the first aspect and the second aspect described above as a program for causing the information processing apparatus to execute or a recording medium on which the program is recorded. be able to. Further, the present invention may be specified by a method in which the information processing apparatus executes the processing executed by each of the above-described means.
- an optimal frame to be used for motion prediction is determined, and prediction efficiency and encoding efficiency can be improved.
- FIG. 1 is a diagram showing a prediction code encoding method for each frame imaged by a reference camera and a prediction code encoding method for each frame imaged by a reference camera.
- FIG. 2 is a diagram illustrating a configuration example of moving image data.
- FIG. 3 is a diagram illustrating an example of functional blocks of a compression code generator.
- FIG. 4 is a flowchart showing a processing example of a reference camera determination process and a reference destination determination process.
- FIG. 5 is a diagram showing an arrangement example of cameras.
- FIG. 6 A table showing the strength and power of each camera serving as a reference camera, and which reference camera is a reference frame when it is a reference camera.
- FIG. 7 is a diagram illustrating an example of functional blocks of the decompression decoding apparatus.
- FIG. 8 is a flowchart showing a processing example of decoding camera determination processing.
- FIG. 9 is a diagram showing problems with the conventional technology.
- FIG. 10 is a diagram showing problems with the conventional technology.
- FIG. 1 is a diagram illustrating a prediction code encoding method for each frame imaged by the reference camera and a prediction code encoding method for each frame imaged by the reference camera.
- the frame positioned at the root of the arrow is predicted by referring to the frame positioned at the tip of the arrow.
- the I frame indicates a frame on which intraframe coding is performed.
- I frame other frame Are encoded without any reference to.
- the P frame indicates a forward temporal prediction coding frame.
- a P frame is a frame captured by the same camera, and is predicted by referring to an I frame or another P frame located immediately before in time.
- B frame indicates a bidirectional temporal prediction encoded frame.
- the B frame is a frame imaged by the same camera, and refers to an I frame or P frame positioned immediately before in time and an I frame or P frame positioned immediately after in time to predict a prediction code. Is done.
- the concepts of I-frame, P-frame, and B-frame are the same as MPEG-1, 2, 4, and so on.
- the I 'frame is predicted by referring to only the I frame of the same time imaged by the reference camera.
- Such predictive encoding by referring to a frame imaged by another camera is hereinafter referred to as “parallax predictive encoding”.
- the P 'frame is obtained by referring to the P frame at the same time captured by the reference camera, and the ⁇ frame or other P' frame that is captured by the same camera and positioned immediately before in time.
- B 'frame is a B frame at the same time taken by the reference camera, a frame taken by the same camera, and I, frame or P, which is located immediately in time, and taken by the same camera.
- Predictive coding is performed by referring to a ⁇ frame or P ′ frame that is immediately after the frame.
- FIG. 2 is a diagram illustrating a configuration example of moving image data.
- the moving image data includes S EQH and GOP.
- SEQH is inserted into the full power unit of GOP (Group of pictures).
- SEQH (n) indicates the GOP force nth GOP that follows.
- SEQH includes the total number of cameras, the parallax prediction type of each camera (that is, the force that the camera that captured each frame is either the reference camera or the reference camera, and an identifier that indicates the reference destination camera in each reference camera.
- GOP (m, n) indicates the nth GOP of camera m, which is a group of frames arranged along the time axis.
- GOPH includes GOPH and a plurality of frames.
- GOPH GOP header Information GOPH includes an identifier indicating the camera that captured the frame, and parallax prediction, and includes time information (frame number relative value from the GOP head).
- Frame is encoded frame data.
- One Frame includes a Frame Header and a plurality of MBs (Macro Blocks).
- Frame Header is frame header information.
- the Frame Header contains the prediction type of the frame (I, P, B, ⁇ ', ⁇ ', ⁇ ⁇ ,).
- MB indicates macroblock information.
- Each MB includes MBType, MV, and DCTCoeff.
- MBType includes the prediction type (Intra, Inter, Bi-Direction) and quantization coefficient of each macroblock.
- the prediction type further includes an identifier of a frame referred to in unidirectional prediction (Inter) or bidirectional prediction (Bi- Direction).
- the prediction type includes one identifier for unidirectional prediction and two identifiers for bidirectional prediction.
- MV is vector information. In the following, this beta information is described separately for motion vectors and disparity vectors.
- the motion vector indicates vector information between frames captured by the same power camera, and indicates vector information between frames captured by a camera different from the disparity vector.
- DCTCoeff is the quantization DCT coefficient information of the prediction error.
- the compression encoding device 1 includes a CPU (central processing unit), a main storage device (RAM), an auxiliary storage device, and the like connected via a bus in terms of hardware.
- the auxiliary storage device is configured using a nonvolatile storage device.
- Non-volatile memory devices mentioned here include so-called ROM (Read-Only Memory: EPROM (Erasable Programmable Read-Only Memory), EEPROM (Electrically Erasable Programmable Read-Only Memory), mask ROM, etc.), FRAM ( Ferroelectric RAM) and hard disk.
- FIG. 3 is a diagram illustrating an example of functional blocks of the compression encoding device 1.
- the compression encoding device 1 mainly includes various programs (OS, applications, etc.) stored in the auxiliary storage device.
- the input frame buffer 101, the subtractor 102, the DCT quantization unit 103, the IDCT inverse quantization unit 104, the calorie calculator 105, the frame buffer 106, and the motion 'disparity vector compensation are loaded into the storage device and executed by the CPU. It functions as a device including a unit 107, a variable length coding unit 108, a header adding unit 109, a control unit 110, and the like.
- DCT quantization section 103 IDCT inverse quantization section 104, motion and disparity vector compensation section 107, variable length coding section 108, header addition section 109, and control section 110 are realized by the program being executed by the CPU. . Further, the DCT quantization unit 103, the IDCT inverse quantization unit 104, the motion / disparity vector compensation unit 107, the variable length coding unit 108, the header addition unit 109, and the control unit 110 may be configured as dedicated chips. Next, each functional unit included in the compression coding device 1 will be described.
- the input frame buffer 101 buffers a moving image input to the compression encoding device 1.
- the compression coding device 1 receives a moving image captured by the basic camera and a moving image captured by the reference camera. Therefore, the input frame buffer 101 buffers the moving image captured by the basic camera and the moving image captured by the reference camera.
- the input frame buffer 101 outputs frame data in units of code key processing of each camera according to an instruction from the control unit 110.
- the encoding processing unit may be one frame or a plurality of frames such as 1 GOP (Group of Pictures).
- a frame output by the input frame buffer 101 that is, a frame to be subjected to compression coding processing is referred to as an input frame.
- the subtractor 102 calculates the difference between the input frame and prediction information by motion compensation or parallax compensation, and outputs the result as prediction difference information.
- the DCT quantization unit 103 performs DCT (Discrete Cosine Transform) operation and quantization operation.
- the DCT quantization unit 103 performs DCT operation on the prediction difference information calculated by the subtractor 102 in units of blocks, quantizes the DCT coefficient, and outputs the resulting quantized DCT coefficient.
- the IDCT inverse quantization unit 104 performs IDCT (Inverse Discrete Cosine Transform) operation (hereinafter also referred to as “inverse DCT operation”) and inverse quantization operation.
- IDCT Inverse Discrete Cosine Transform
- inverse DCT operation Inverse Discrete Cosine Transform
- the IDCT inverse quantization unit 104 performs inverse quantization and inverse DCT operation on the quantized DCT coefficient to obtain an inverse DCT operation result.
- the adder 105 adds the inverse DCT calculation result and the prediction result by motion compensation or disparity compensation to generate a local decoded moving image.
- the frame buffer 106 stores local decoded moving images.
- the frame buffer 106 outputs a designated frame of a designated camera moving image in accordance with an instruction from the control unit 110.
- a frame output by the frame buffer 106 that is, a frame used for motion prediction or disparity prediction is referred to as a prediction source frame.
- the motion 'parallax vector compensation unit 107 performs prediction based on block matching using the input frame and the prediction source frame in accordance with an instruction from the control unit 110.
- the motion / disparity vector compensation unit 107 outputs the motion vector information and the disparity vector information to the variable length coding unit 108.
- the motion / disparity vector compensation unit 107 outputs the prediction information to the subtractor 102.
- Motion vector information, disparity vector information, and prediction information are used to minimize prediction errors.
- the motion 'parallax vector compensation unit 107 outputs the motion vector information for the entire frame to the control unit 110.
- the motion vector information is, for example, the average and variance of motion vector information in all blocks.
- variable-length code unit 108 generates frame-compressed frame data by variable-length coding the quantization result. Also, the motion vector information used for motion compensation and the disparity vector information used for disparity compensation are passed to the header adding unit 109.
- the header adding unit 109 adds information such as a camera number, camera type (base camera or reference camera), other cameras to be referred to, etc. to each frame or a plurality of frames after compression encoding. Generate data. ⁇ Control unit>
- the control unit 110 controls the sign of a moving image captured by each camera, determines a reference camera (reference camera determination process), and determines a frame to be referenced for a predicted sign of a frame captured by a reference camera. (Reference destination decision processing) is performed. Further, the control unit 110 is connected so as to be able to instruct each function unit. The controller 110 also receives parameters of each camera (placement information for each camera) and external force. The arrangement information may be an absolute position of each camera or a relative position.
- the reference camera determination process and the reference destination determination process will be described.
- FIG. 4 is a flowchart showing an operation example of the control unit 110 in the base camera determination process and the reference destination determination process. An operation example of the control unit 110 will be described with reference to FIG. The following processing is executed for each set of GOPs. That is, the process shown in FIG. 4 is executed in units of a plurality of GOPs collected by one SEQH.
- the control unit 110 determines whether or not a panning has occurred in the last frame of the immediately preceding GOP (SOI). For example, the control unit 110 determines the occurrence of the vanning based on the motion vector information in the last frame of the previous GOP (for example, the average and variance of the motion vector information in all blocks). In other words, the occurrence of panning is determined based on the movement of the subject imaged in this last frame. In this case, the control unit 110 determines whether the average value of the motion vectors in the horizontal direction is greater than or equal to a threshold and whether the variance is equal to or less than the threshold. When these two conditions are satisfied, the control unit 110 determines that the vanning has occurred. On the other hand, if one of these two conditions is not satisfied, control unit 110 determines that nosing has not occurred.
- SOI the last frame of the immediately preceding GOP
- the control unit 110 calculates C (NZ2 nK) and sets the camera corresponding to the calculation result as the reference camera (S03).
- C (m) is an identifier indicating the m-th camera, and it is assumed that numbers are assigned in the order in which the cameras are arranged in one direction.
- N indicates the total number of cameras. N represents an integer of 0 or more.
- the value of K is a positive value and is set appropriately by the designer in accordance with the distance between the cameras and the distance between the camera and the subject. In this case, a camera in which the center of the camera row and the central force are both in the left and right directions at equal intervals (K) is set as the reference camera.
- FIG. 5 is a diagram illustrating an arrangement example of cameras. Further, FIG. 6 is a table showing the strength and power of each camera serving as a reference camera, and which reference camera is a reference frame when it is a reference camera.
- seven cameras C (l) and C (7) are arranged in numerical order on the X axis. In FIG. 5, the cameras are arranged at equal intervals or at arbitrary intervals in the direction perpendicular to the shooting direction (Z-axis direction).
- the control unit 110 determines in which direction the vanning is occurring (S02). This direction can be determined from the motion vector information used when determining the occurrence of the vanning. That is, the direction of occurrence of vanning can be determined from the direction of this vector.
- C l + nK
- the control unit 110 determines the reference frame for each frame imaged by the reference camera as the nearest reference frame on the right side (S07). In other words, the control unit 110 sets the reference frame for each frame imaged by the reference camera as the nearest reference camera installed in the direction opposite to the direction in which the Banning occurs.
- control unit 110 determines that the vanning has occurred on the right (S02 right), C (N nK) is set as the reference camera (S05). Then, the control unit 110 determines the reference frame for each frame imaged by the reference camera as the nearest reference frame on the left side (SO 8).
- the decompression decoding device 2 includes a CPU (Central Processing Unit), a main storage device (RAM), an auxiliary storage device and the like connected via a bus in hardware.
- the auxiliary storage device is configured using a nonvolatile storage device.
- Non-volatile memory devices mentioned here include so-called ROM (Read-Only Memory: EPROM (Erasable Programmable Read-Only Memory), EEPROM (Electrically Erasable Programmable Read-Only Memory), mask ROM, etc.), FRAM ( Ferroelectric RAM) and hard disk.
- FIG. 7 is a diagram illustrating an example of functional blocks of the decompression decoding apparatus 2.
- the decompression decoding device 2 loads various programs (OS, applications, etc.) stored in the auxiliary storage device into the main storage device and executes them by the CPU, whereby the header analysis unit 201, the encoded video buffer 202, a variable length decoding unit 203, an IDCT inverse quantization unit 204, an adder 205, a display moving image buffer 206, a frame buffer 207, a motion 'disparity vector compensation unit 208, a control unit 209, and the like.
- OS operating system
- applications etc.
- the header analysis unit 201, variable length decoding unit 203, IDCT inverse quantization unit 204, motion / disparity vector compensation unit 208, and control unit 209 are realized by the program being executed by the CPU. Further, the header 201, the variable length decoding unit 203, the IDCT inverse quantization unit 204, the motion / disparity vector compensation unit 208, and the control unit 209 may be configured as dedicated chips. Next, each functional unit included in the decompression decoding apparatus 2 will be described.
- the header analysis unit 201 extracts header information from the input moving image data (moving image data created by the compression encoding device 1). Specifically, the header information includes the identifier of the camera that captured each frame, its type (base camera or reference camera), the identifier of the camera that captured the reference frame, and the like. Also, the header analysis unit 201 receives from the control unit 209 the identifier of the camera that has captured the frame to be decoded. Based on this identifier, the header analysis unit 201 determines whether or not each input frame should be decoded, and only the frame that is determined to be decoded is encoded video buffer 202. To pass.
- the header analysis unit 201 determines a GOP corresponding to a camera to be decoded from among GOPs including frames captured by each camera (decoding camera determination process). By executing this processing, it becomes possible to omit the decoding processing for unnecessary frames, and to increase the processing speed. At the same time, it is possible to reduce the amount of data to be notched, and for example, it is possible to reduce the scale of the display moving image buffer 206 and the frame notifier 207. Further, the header analysis unit 201 notifies the control unit 209 of the identifier of the frame passed to the code motion image buffer 202.
- FIG. 8 is a flowchart showing an operation example of the header analysis unit 201 in the decoding camera determination process. The processing shown in this flowchart is executed every time SEQH is detected.
- the header analysis unit 201 acquires an identifier indicating a camera from the control unit 209 (Sl l).
- the header analysis unit 201 extracts and analyzes SEQ H from the input moving image data (S12). Through this analysis, the header analysis unit 201 acquires the error prediction type of each camera, the identifier of the camera to be referred to, and the like.
- the header analysis unit 201 searches for GOPH. At this time, if the next SEQH is detected, that is, if there is no frame imaged by the next camera, this process is terminated (S13—No). On the other hand, when GOPH is detected (S13—Yes), the header analysis unit 201 analyzes this GOPH and obtains an identifier of the camera. Then, the header analysis unit 201 determines whether or not the identifier included in GOPH matches the identifier passed from the control unit 209. If they match (S14—Yes), it is determined that decoding should be performed for each frame included in this GOP (S19). Then, the processes after S13 are executed again.
- the header analysis unit 201 determines whether the identifier is a camera corresponding to the camera reference destination of the identifier passed from the control unit 209. If it is not the reference destination (S15—No), the processing after S13 is executed. On the other hand, when it is a reference destination (S15—Yes), the header analysis unit 201 analyzes GOPH and analyzes the time during which disparity prediction is not performed (S16). And if the parallax prediction is performed all the time (S 17—Yes), this It is determined that the decoding key for each frame included in the GOP should be executed (S19).
- the header analysis unit 201 determines a frame for which disparity prediction is performed, and determines that only that frame should be decoded ( S18). Then, the processes after S13 are executed again.
- the code motion picture buffer 202 buffers each frame determined to be decoded by the header analysis unit 201.
- the encoded video buffer 202 outputs a frame in units of decoding key processing in accordance with an instruction from the control unit 209.
- the decryption processing unit may be one frame or multiple frames like GOP.
- variable length decoding unit 203 performs variable length decoding on the quantized DCT coefficient that has been subjected to variable length coding, and passes the result to the IDCT inverse quantization unit 204.
- the variable length decoding unit 203 also performs variable length decoding on the motion vector information and the disparity vector information and passes them to the motion / disparity vector compensation unit 208.
- the IDCT inverse quantization unit 204 performs IDCT operation and inverse quantization operation.
- the IDCT inverse quantization unit 204 performs inverse quantization and inverse DCT operation on the quantized DCT coefficient to obtain an inverse DCT operation result.
- the adder 205 adds the inverse DCT calculation result and the prediction result by motion compensation or disparity compensation to generate a decoded moving image.
- the display moving image buffer 206 buffers the decoded moving image data generated by the adder 205. At this time, the display moving image buffer 208 buffers the data of the decoded moving image corresponding to the camera designated to be displayed from the outside. Then, the display moving image buffer 206 sequentially outputs the data being noffered.
- the frame buffer 207 stores the decoded moving image data. Buffer the data. Further, the frame buffer 207 also accumulates frames captured by other cameras that are necessary for decoding the frame to be processed regardless of designation from the outside.
- the motion 'disparity vector compensation unit 208 reads out, from the frame notifier 207, a frame used for motion prediction and parallax prediction necessary for decoding a frame to be processed. Then, the motion 'disparity vector compensation unit 208 acquires motion vector information ⁇ disparity vector information from the variable length decoding unit 203. Then, the prediction result is acquired, and the prediction result is passed to the adder 205.
- the control unit 209 performs decoding key control on the input moving image data.
- the control unit 209 is connected so as to be able to instruct each functional unit.
- an identifier indicating a camera that has captured a moving image to be output to the outside from the display moving image buffer 208 is input to the control unit 209 from the outside.
- One or more identifiers may be used. Then, the control unit 209 passes this input identifier to the header analysis unit 201.
- Each frame (I, frame, P, frame, B, frame) captured by the reference camera is not limited to a frame captured by the reference camera at the same time, but refers to a frame captured by another reference camera.
- Predictive encoding may be performed, or prediction encoding may be performed with reference to frames captured at different times by a standard camera or other reference power cameras.
- the cameras may be installed in any way, such as wavy lines, circles, crosses, squares, etc.
- the present invention can obtain an effect by being applied to an apparatus that encodes / decodes images picked up from multiple viewpoints.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
Claims
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2005/000122 WO2006072992A1 (ja) | 2005-01-07 | 2005-01-07 | 圧縮符号化装置、伸張復号化装置 |
CN2005800461596A CN101099392B (zh) | 2005-01-07 | 2005-01-07 | 压缩编码装置、解压缩解码装置 |
JP2006550569A JP4893311B2 (ja) | 2005-01-07 | 2005-01-07 | 圧縮符号化装置、伸張復号化装置 |
EP05703387A EP1843597A4 (en) | 2005-01-07 | 2005-01-07 | COMPRESSION / CODING DEVICE, DECOMPRESSING / DECODING DEVICE |
US11/774,264 US20070253482A1 (en) | 2005-01-07 | 2007-07-06 | Compression-coding device and decompression-decoding device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2005/000122 WO2006072992A1 (ja) | 2005-01-07 | 2005-01-07 | 圧縮符号化装置、伸張復号化装置 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/774,264 Continuation US20070253482A1 (en) | 2005-01-07 | 2007-07-06 | Compression-coding device and decompression-decoding device |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2006072992A1 true WO2006072992A1 (ja) | 2006-07-13 |
Family
ID=36647480
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2005/000122 WO2006072992A1 (ja) | 2005-01-07 | 2005-01-07 | 圧縮符号化装置、伸張復号化装置 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20070253482A1 (ja) |
EP (1) | EP1843597A4 (ja) |
JP (1) | JP4893311B2 (ja) |
CN (1) | CN101099392B (ja) |
WO (1) | WO2006072992A1 (ja) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010098134A1 (ja) * | 2009-02-27 | 2010-09-02 | パナソニック株式会社 | 記録媒体、再生装置、及び集積回路 |
WO2010100875A1 (ja) * | 2009-03-02 | 2010-09-10 | パナソニック株式会社 | 記録媒体、再生装置、及び集積回路 |
WO2011074189A1 (ja) * | 2009-12-15 | 2011-06-23 | パナソニック株式会社 | 画像符号化方法及び画像符号化装置 |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2014027448A (ja) * | 2012-07-26 | 2014-02-06 | Sony Corp | 情報処理装置、情報処理方法、及びプログラム |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0698312A (ja) | 1992-09-16 | 1994-04-08 | Fujitsu Ltd | 画像高能率符号化方式 |
JPH06113338A (ja) * | 1992-09-30 | 1994-04-22 | Fujitsu Ltd | 多眼式立体映像の動き補償予測方法 |
JPH07143494A (ja) * | 1993-11-19 | 1995-06-02 | Sanyo Electric Co Ltd | 動画像符号化方法 |
JPH09261653A (ja) * | 1996-03-18 | 1997-10-03 | Sharp Corp | 多視点画像符号化装置 |
JPH10191394A (ja) | 1996-12-24 | 1998-07-21 | Sharp Corp | 多視点画像符号化装置 |
JPH1169381A (ja) * | 1997-08-25 | 1999-03-09 | Sony Corp | 立体画像用ビデオ信号生成装置、立体画像用ビデオ信号送出装置、立体画像用ビデオ信号受信装置および立体画像用ビデオ信号送出切り替え装置 |
JP2000023918A (ja) | 1998-07-10 | 2000-01-25 | Canon Inc | 眼科撮影装置 |
JP2002016945A (ja) * | 2000-06-29 | 2002-01-18 | Toppan Printing Co Ltd | 画像軽量化手法を用いた三次元画像表現システム |
JP2002300607A (ja) | 2001-03-29 | 2002-10-11 | Nippon Hoso Kyokai <Nhk> | 符号化装置及び復号化装置 |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
NL9301358A (nl) * | 1993-08-04 | 1995-03-01 | Nederland Ptt | Transcodeerinrichting. |
US5703961A (en) * | 1994-12-29 | 1997-12-30 | Worldscape L.L.C. | Image transformation and synthesis methods |
US6137912A (en) * | 1998-08-19 | 2000-10-24 | Physical Optics Corporation | Method of multichannel data compression |
JP3592168B2 (ja) * | 1999-12-22 | 2004-11-24 | キヤノン株式会社 | 画像データの符号化復号化方法及び装置 |
US20020009137A1 (en) * | 2000-02-01 | 2002-01-24 | Nelson John E. | Three-dimensional video broadcasting system |
CN101304536B (zh) * | 2002-11-20 | 2011-11-02 | 松下电器产业株式会社 | 动态图像编码方法及动态图像编码装置 |
-
2005
- 2005-01-07 WO PCT/JP2005/000122 patent/WO2006072992A1/ja active Application Filing
- 2005-01-07 JP JP2006550569A patent/JP4893311B2/ja not_active Expired - Fee Related
- 2005-01-07 EP EP05703387A patent/EP1843597A4/en not_active Ceased
- 2005-01-07 CN CN2005800461596A patent/CN101099392B/zh not_active Expired - Fee Related
-
2007
- 2007-07-06 US US11/774,264 patent/US20070253482A1/en not_active Abandoned
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0698312A (ja) | 1992-09-16 | 1994-04-08 | Fujitsu Ltd | 画像高能率符号化方式 |
JPH06113338A (ja) * | 1992-09-30 | 1994-04-22 | Fujitsu Ltd | 多眼式立体映像の動き補償予測方法 |
JPH07143494A (ja) * | 1993-11-19 | 1995-06-02 | Sanyo Electric Co Ltd | 動画像符号化方法 |
JP3426668B2 (ja) | 1993-11-19 | 2003-07-14 | 三洋電機株式会社 | 動画像符号化方法 |
JPH09261653A (ja) * | 1996-03-18 | 1997-10-03 | Sharp Corp | 多視点画像符号化装置 |
JPH10191394A (ja) | 1996-12-24 | 1998-07-21 | Sharp Corp | 多視点画像符号化装置 |
JPH1169381A (ja) * | 1997-08-25 | 1999-03-09 | Sony Corp | 立体画像用ビデオ信号生成装置、立体画像用ビデオ信号送出装置、立体画像用ビデオ信号受信装置および立体画像用ビデオ信号送出切り替え装置 |
JP2000023918A (ja) | 1998-07-10 | 2000-01-25 | Canon Inc | 眼科撮影装置 |
JP2002016945A (ja) * | 2000-06-29 | 2002-01-18 | Toppan Printing Co Ltd | 画像軽量化手法を用いた三次元画像表現システム |
JP2002300607A (ja) | 2001-03-29 | 2002-10-11 | Nippon Hoso Kyokai <Nhk> | 符号化装置及び復号化装置 |
Non-Patent Citations (4)
Title |
---|
HIKEAKI KIMATA ET AL.: "System Design of Free Viewpoint Video Communication", COMPUTER AND INFORMATION TECHNOLOGY, 2004. CIT '04. THE FOURTH INTERNATIONAL CONFERENCE ON, IEEE, LOS ALAMITOS, CA, USA, 14 September 2004 (2004-09-14), pages 52 - 59, XP003012676 |
KHAN S ET AL.: "Consistent labeling of tracked objects in multiple cameras with overlapping fields of view", IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, IEEE SERVICE CENTER, vol. 25, no. 10, 1 October 2003 (2003-10-01), pages 1355 - 1360, XP011101862, DOI: doi:10.1109/TPAMI.2003.1233912 |
LIM J ET AL.: "A multiview sequence CODEC with view scalability" SIGNAL PROCESSING. IMAGE COMMUNICATOIN", vol. 19, 1 March 2004, ELSEVIER SCIENCE PUSLISHERS, pages: 239 - 256 |
See also references of EP1843597A4 |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010098134A1 (ja) * | 2009-02-27 | 2010-09-02 | パナソニック株式会社 | 記録媒体、再生装置、及び集積回路 |
JPWO2010098134A1 (ja) * | 2009-02-27 | 2012-08-30 | パナソニック株式会社 | 記録媒体、再生装置、及び集積回路 |
WO2010100875A1 (ja) * | 2009-03-02 | 2010-09-10 | パナソニック株式会社 | 記録媒体、再生装置、及び集積回路 |
US8861940B2 (en) | 2009-03-02 | 2014-10-14 | Panasonic Corporation | Recording medium, playback device and integrated circuit |
WO2011074189A1 (ja) * | 2009-12-15 | 2011-06-23 | パナソニック株式会社 | 画像符号化方法及び画像符号化装置 |
Also Published As
Publication number | Publication date |
---|---|
CN101099392A (zh) | 2008-01-02 |
US20070253482A1 (en) | 2007-11-01 |
JPWO2006072992A1 (ja) | 2008-06-12 |
EP1843597A4 (en) | 2010-07-14 |
CN101099392B (zh) | 2010-10-13 |
EP1843597A1 (en) | 2007-10-10 |
JP4893311B2 (ja) | 2012-03-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR100774296B1 (ko) | 움직임 벡터 부호화 방법, 복호화 방법 및 그 장치 | |
WO2006073116A1 (ja) | 映像符号化方法及び装置、映像復号方法及び装置、それらのプログラムおよびそれらプログラムを記録した記録媒体 | |
JPH09294266A (ja) | 動きベクトル探索方法 | |
US6271774B1 (en) | Picture data processor, picture data decoder and picture data encoder, and methods thereof | |
US8606024B2 (en) | Compression-coding device and decompression-decoding device | |
WO2006072992A1 (ja) | 圧縮符号化装置、伸張復号化装置 | |
JP5402953B2 (ja) | 伸張復号化装置、伸張復号化方法、及び、プログラム | |
US8897368B2 (en) | Image coding device, image coding method, image coding integrated circuit and image coding program | |
JP5402960B2 (ja) | 伸張復号化装置、及び、伸張復号化方法 | |
KR100926607B1 (ko) | 압축 부호화 장치, 신장 복호화 장치 | |
JP3690259B2 (ja) | 動画像高速符号化装置及び動画像高速符号化方法 | |
KR100926608B1 (ko) | 압축 부호화 장치, 신장 복호화 장치 | |
WO2006038568A1 (ja) | 映像符号化方法及び装置、映像復号方法及び装置、それらのプログラムおよびそれらプログラムを記録した記録媒体 | |
JP2007517438A (ja) | 参照フレームの数を固定する符号化方式で画像の参照ブロックを取得する方法 | |
JP2011130499A (ja) | 伸張復号化装置、及び、伸張復号化方法 | |
JP2012049753A (ja) | 映像圧縮装置、映像圧縮方法およびプログラム | |
JP3188081B2 (ja) | 画像符号化方法および画像符号化装置 | |
JP2008245237A (ja) | 画像処理方法、画像処理装置およびそれを用いた撮像装置 | |
JP5274181B2 (ja) | 動画像符号化装置および動画像符号化方法 | |
JP2006165801A (ja) | 符号化モード判定装置及び方法 | |
JP2000295567A (ja) | 符号化データ編集装置 | |
JP2007325230A (ja) | 動きベクトル復号化方法および復号化装置 | |
JP2003309803A (ja) | ビデオストリーム編集装置 | |
JP2010068219A (ja) | 動画像符号化装置 | |
JPH0662388A (ja) | 動画像圧縮装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2006550569 Country of ref document: JP Ref document number: 11774264 Country of ref document: US Ref document number: 200580046159.6 Country of ref document: CN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2005703387 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020077018034 Country of ref document: KR |
|
WWP | Wipo information: published in national office |
Ref document number: 2005703387 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 11774264 Country of ref document: US |