CN100481963C - Visual difference calculating method for video image in multi-vision point system - Google Patents
Visual difference calculating method for video image in multi-vision point system Download PDFInfo
- Publication number
- CN100481963C CN100481963C CNB2005100774732A CN200510077473A CN100481963C CN 100481963 C CN100481963 C CN 100481963C CN B2005100774732 A CNB2005100774732 A CN B2005100774732A CN 200510077473 A CN200510077473 A CN 200510077473A CN 100481963 C CN100481963 C CN 100481963C
- Authority
- CN
- China
- Prior art keywords
- video image
- image
- target video
- frame
- search
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Landscapes
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Image Processing (AREA)
Abstract
Description
技术领域 technical field
本发明设计一种多视点系统中视频图像的视差计算方法,属于视频图像处理技术领域。The invention designs a parallax calculation method of video images in a multi-viewpoint system, and belongs to the technical field of video image processing.
背景技术 Background technique
用于多视点视频系统的块匹配法是最常用的两个视频图像之间的视差计算方法,其基本思想是将目标图像分割为固定尺寸的块,并搜索每个块在参考图像中的最佳匹配。经典的搜索技术是在搜索窗内的完全搜索,其基本算法如下(设块尺寸为N乘N):The block matching method used in multi-view video systems is the most commonly used disparity calculation method between two video images. best match. The classic search technique is a complete search within the search window, and its basic algorithm is as follows (set the block size to be N times N):
若
其中in
T为图像块的空间位移量;T’是目标图像和参考图像中两个对应图像块之间相关性达到最大时的空间位移量;m、n分别是空间位移量T的横坐标和纵坐标,T is the spatial displacement of the image block; T' is the spatial displacement when the correlation between the two corresponding image blocks in the target image and the reference image reaches the maximum; m and n are the abscissa and ordinate of the spatial displacement T respectively ,
fl、fr是目标图像和参考图像中两个对应图像块的像素灰度值。f l , f r are the pixel gray values of two corresponding image blocks in the target image and the reference image.
传统的块匹配视差估计算法由于算法简单、计算量小而被普遍采用。然而,对整个目标图像进行完全搜索时,其计算量很大。The traditional block-matching disparity estimation algorithm is widely used because of its simple algorithm and small amount of calculation. However, it is computationally expensive to perform a full search on the entire target image.
发明内容 Contents of the invention
本发明的目的是提出一种多视点系统中视频图像的视差计算方法,在传统的块匹配基础上提出了基于立体摄像几何及视频特性的自适应算法,提高了视差计算的效率。The purpose of the present invention is to propose a disparity calculation method for video images in a multi-viewpoint system. On the basis of traditional block matching, an adaptive algorithm based on stereo camera geometry and video characteristics is proposed, which improves the efficiency of disparity calculation.
本发明提出的多视点系统中视频图像的视差计算方法,包括以下步骤:The parallax calculation method of the video image in the multi-viewpoint system that the present invention proposes, comprises the following steps:
(1)利用块匹配法计算目标视频图像中第一帧视频图像的完整视差;(1) Utilize the block matching method to calculate the complete disparity of the first frame video image in the target video image;
(2)以上一帧目标视频图像中的相应图像块的视差值为当前位置,以LW为搜索半径,对参考视频图像进行有限搜索;(2) The disparity value of the corresponding image block in the above frame target video image is the current position, and the reference video image is limitedly searched with LW as the search radius;
(3)若在参考视频图像中搜索到与目标视频图像中相对应的图像块,则进行步骤(8),若搜索不到与目标视频图像中相对应的图像块,则进行步骤(4);(3) if searching for the image block corresponding to the target video image in the reference video image, then proceed to step (8), if not searching for the corresponding image block in the target video image, then proceed to step (4) ;
(4)以上一帧目标视频图像中的相应图像块的视差值为当前位置,向左对参考视频图像进行方向搜索,搜索范围为LW+1~LW+DW1;(4) The disparity value of the corresponding image block in the above frame target video image is the current position, and the direction search is carried out to the left for the reference video image, and the search range is LW+1~LW+DW1;
(5)若在参考视频图像中搜索到与目标视频图像中相对应的图像块,则进行步骤(8),若搜索不到与目标视频图像中相对应的图像块,则进行步骤(6);(5) if searching for the image block corresponding to the target video image in the reference video image, then proceed to step (8), if not searching for the corresponding image block in the target video image, then proceed to step (6) ;
(6)以上一帧目标视频图像中的相应图像块的视差值为当前位置,向右对参考视频图像进行方向搜索,搜索范围为LW+1~LW+DW2;(6) The parallax value of the corresponding image block in the above frame target video image is the current position, and the direction search is carried out to the right for the reference video image, and the search range is LW+1~LW+DW2;
(7)若在参考视频图像中搜索到与目标视频图像中相对应的图像块,则进行步骤(8),若搜索不到与目标视频图像中相对应的图像块,则取目标视频图像中平均象素灰度差最小的图像块作为对应的图像块,进行步骤(8);(7) If in the reference video image, search for the image block corresponding to the target video image, then proceed to step (8), if not find the corresponding image block in the target video image, then get the image block corresponding to the target video image The image block with the smallest average pixel gray level difference is used as the corresponding image block, and then step (8);
(8)根据目标视频图像中图像块的位置和搜索到的参考图像块的位置,得到目标视频图像中的图像块与参考视频图像中相应图像块之间的视差。(8) Obtain the disparity between the image block in the target video image and the corresponding image block in the reference video image according to the position of the image block in the target video image and the position of the searched reference image block.
本发明提出的多视点系统中视频图像的视差计算方法,其优点是提高了算法的计算效率,只需比传统的块匹配算法更少的计算量,就可以得到相同的视差计算结果的准确性。The disparity calculation method of video images in the multi-viewpoint system proposed by the present invention has the advantage of improving the calculation efficiency of the algorithm, and only needs less calculation than the traditional block matching algorithm to obtain the same accuracy of the disparity calculation results .
具体实施方式 Detailed ways
本发明提出的多视点系统中视频图像的视差计算方法,首先利用块匹配法计算目标视频图像中第一帧视频图像的完整视差;以上一帧目标视频图像中的相应图像块的视差值为当前位置,以LW为搜索半径,对参考视频图像进行有限搜索;若在参考视频图像中搜索到与目标视频图像中相对应的图像块,则根据目标视频图像中图像块的位置和搜索到的参考图像块的位置,得到目标视频图像中的图像块与参考视频图像中相应图像块之间的视差,若搜索不到与目标视频图像中相对应的图像块,则以上一帧目标视频图像中的相应图像块的视差值为当前位置,向左对参考视频图像进行方向搜索,搜索范围为LW+1~LW+DW1;若在参考视频图像中搜索到与目标视频图像中相对应的图像块,则根据目标视频图像中图像块的位置和搜索到的参考图像块的位置,得到目标视频图像中的图像块与参考视频图像中相应图像块之间的视差,若搜索不到与目标视频图像中相对应的图像块,则以上一帧目标视频图像中的相应图像块的视差值为当前位置,向右对参考视频图像进行方向搜索,搜索范围为LW+1~LW+DW2;若在参考视频图像中搜索到与目标视频图像中相对应的图像块,则根据目标视频图像中图像块的位置和搜索到的参考图像块的位置,得到目标视频图像中的图像块与参考视频图像中相应图像块之间的视差,若搜索不到与目标视频图像中相对应的图像块,则取目标视频图像中平均像素灰度差最小的图像块作为对应的图像块,根据目标视频图像中图像块的位置和搜索到的参考图像块的位置,得到目标视频图像中的图像块与参考视频图像中相应图像块之间的视差。The method for calculating the disparity of video images in the multi-viewpoint system proposed by the present invention first utilizes the block matching method to calculate the complete disparity of the first frame video image in the target video image; the disparity value of the corresponding image block in the above frame of target video image is At the current position, with LW as the search radius, a limited search is performed on the reference video image; if the image block corresponding to the target video image is found in the reference video image, according to the position of the image block in the target video image and the searched Refer to the position of the image block to obtain the disparity between the image block in the target video image and the corresponding image block in the reference video image. If the image block corresponding to the target video image cannot be searched, then The disparity value of the corresponding image block is the current position, search the direction of the reference video image to the left, and the search range is LW+1~LW+DW1; if the image corresponding to the target video image is found in the reference video image block, then according to the position of the image block in the target video image and the position of the searched reference image block, the disparity between the image block in the target video image and the corresponding image block in the reference video image is obtained. The corresponding image block in the image, then the disparity value of the corresponding image block in the target video image of the previous frame is the current position, and the direction search is performed on the reference video image to the right, and the search range is LW+1~LW+DW2; if Search for the image block corresponding to the target video image in the reference video image, then according to the position of the image block in the target video image and the position of the searched reference image block, obtain the image block in the target video image and the reference video image The disparity between the corresponding image blocks in the target video image, if the corresponding image block in the target video image cannot be searched, the image block with the smallest average pixel gray level difference in the target video image is taken as the corresponding image block, according to the target video image The position of the image block and the searched position of the reference image block are used to obtain the disparity between the image block in the target video image and the corresponding image block in the reference video image.
在多视点摄像系统中,方向性与均匀性是与视差有关的两个性质,在视差计算时若能充分利用这些几何特性,可以简化视差匹配的算法,实现快速匹配:In a multi-viewpoint camera system, directionality and uniformity are two properties related to parallax. If these geometric characteristics can be fully utilized in parallax calculation, the algorithm of parallax matching can be simplified and fast matching can be achieved:
1、方向性:1. Direction:
对于同一景物,其透视投影左图像相对于右图像局部地向水平方向的反方向移动。也就是说,如果正水平方向是向右,则同一空间点在左图像中的投影位置比右图像偏左。这个特性为本算法的方向性搜索提供了依据,比如在搜索左图像各块在右图像中的匹配块时,可优先向右方搜索。For the same scene, the left image of its perspective projection moves to the opposite direction of the horizontal direction locally relative to the right image. That is to say, if the positive horizontal direction is to the right, the projected position of the same spatial point in the left image is to the left of the right image. This feature provides a basis for the directional search of this algorithm, for example, when searching for the matching blocks of the blocks in the left image in the right image, you can search to the right first.
2、均匀性:2. Uniformity:
偏振线上相邻图像子块间的视差差异实际上是很小的,因此,在进行视差匹配时,以邻近块已求出的视差矢量作为参考,在小范围内搜索当前块的视差矢量,能迅速准确地搜索出当前视差这一特性,称之为有限搜索。The disparity difference between adjacent image sub-blocks on the polarization line is actually very small. Therefore, when disparity matching is performed, the disparity vector of the adjacent block is used as a reference to search for the disparity vector of the current block in a small range. The property that can quickly and accurately search out the current disparity is called limited search.
视差计算的目的是应用于视频的编、解码,因此可以结合视频的特性,即相邻视频帧间的视差相关度非常高,进一步提高块匹配的效率;另外,由于算法是应用于视频,可以利用前面视频帧的搜索结果来修改本帧的参数值,从而使得参数本身可以自动适应视频的特性。The purpose of disparity calculation is to apply to video encoding and decoding, so it can combine the characteristics of video, that is, the disparity correlation between adjacent video frames is very high, and further improve the efficiency of block matching; in addition, because the algorithm is applied to video, it can Use the search results of the previous video frame to modify the parameter value of this frame, so that the parameter itself can automatically adapt to the characteristics of the video.
对于视频的第一帧,采用传统搜索算法得到其完整视差;从第二帧开始,每一帧都在前一帧的视差结果基础上进行视差计算。For the first frame of the video, the traditional search algorithm is used to obtain its complete disparity; starting from the second frame, each frame is calculated based on the disparity result of the previous frame.
本发明方法中的搜索半径LW、DW1和DW2,必须根据对目标图像进行搜索的结果不断调整,以提高搜索效率。三个搜索半径参数的调整方法完全相同,以下以LW为例进行说明:The search radii LW, DW1 and DW2 in the method of the present invention must be constantly adjusted according to the search results of the target image, so as to improve the search efficiency. The adjustment methods of the three search radius parameters are exactly the same, and the following uses LW as an example to illustrate:
为每一帧图像设定一个调整参数ALW,则第n+1帧图像的LW=(第n帧的ALW+第n—1帧的ALW+第n—2帧的ALW)/3,其中第n帧的ALW根据本帧在LW范围内搜索后的结果是否满足匹配准则而定,若满足匹配准则,则第n帧的ALW为搜索到的视差值与上一帧视差值差的绝对值;若不满足匹配准则,则第n帧的ALW为第n—1帧的ALW加9。An adjustment parameter ALW is set for each frame image, then the LW=(ALW of the nth frame+the ALW of the n-1 frame+the ALW of the n-2 frame)/3 of the n+1th frame image, wherein the nth frame The ALW of the frame depends on whether the search result of this frame in the LW range satisfies the matching criterion. If the matching criterion is satisfied, the ALW of the nth frame is the absolute value of the difference between the searched disparity value and the previous frame disparity value; If the matching criterion is not met, the ALW of the nth frame is the ALW of the n-1th frame plus 9.
上述匹配准则可采用平均绝对差准则(MAD)或其他的相似度准则,这里定义一种离散度准则,性能优于一般的MAD准则。设fl(i,j)、fr(i,j)为参考视频图像和目标视频图像块的像素灰度,块的尺寸为N乘以N,在满足立体平行摄像系统条件下,定义离散度函数为:The above-mentioned matching criterion may adopt the mean absolute difference criterion (MAD) or other similarity criterion. Here, a dispersion criterion is defined, and its performance is better than that of the general MAD criterion. Let f l (i, j), f r (i, j) be the pixel grayscale of the reference video image and the target video image block, the size of the block is N times N, and under the condition of satisfying the stereo parallel camera system, define the discrete The degree function is:
当B(d)为最小时,两个图像块达到匹配,此时的d即为两图像块之间的视差。When B(d) is the minimum, the two image blocks are matched, and d at this time is the disparity between the two image blocks.
Claims (1)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CNB2005100774732A CN100481963C (en) | 2005-06-24 | 2005-06-24 | Visual difference calculating method for video image in multi-vision point system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CNB2005100774732A CN100481963C (en) | 2005-06-24 | 2005-06-24 | Visual difference calculating method for video image in multi-vision point system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN1688168A CN1688168A (en) | 2005-10-26 |
CN100481963C true CN100481963C (en) | 2009-04-22 |
Family
ID=35306208
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CNB2005100774732A Expired - Fee Related CN100481963C (en) | 2005-06-24 | 2005-06-24 | Visual difference calculating method for video image in multi-vision point system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN100481963C (en) |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101627634B (en) * | 2006-10-16 | 2014-01-01 | 诺基亚公司 | Systems and methods for using parallel decodable slices for multiview video coding |
CN101174331B (en) * | 2006-11-01 | 2011-07-27 | 深圳市蓝韵实业有限公司 | Maximum density projection generating method for medical image |
EP2116063B1 (en) * | 2007-01-04 | 2017-03-08 | Thomson Licensing | Methods and apparatus for multi-view information conveyed in high level syntax |
CN102256129B (en) * | 2007-10-15 | 2013-01-02 | 华为技术有限公司 | Method and system for determining corresponding macro block |
CN101415116B (en) * | 2007-10-15 | 2011-08-03 | 华为技术有限公司 | A method and system for determining corresponding macroblocks |
CN101600123B (en) * | 2009-07-06 | 2010-08-25 | 四川大学 | Method for shifting parallel vidicon array for obtaining parallactic image |
CN101754038B (en) * | 2009-12-09 | 2012-05-30 | 青岛海信网络科技股份有限公司 | Parallax extraction method |
CN102075779B (en) * | 2011-02-21 | 2013-05-08 | 北京航空航天大学 | Intermediate view synthesizing method based on block matching disparity estimation |
CN102215417A (en) * | 2011-05-04 | 2011-10-12 | 山东大学 | Parallax prediction method capable of establishing mathematical model based on block matching |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1332430A (en) * | 2001-07-27 | 2002-01-23 | 南开大学 | 3D tracking and measurement method of moving objects by 2D code |
CN1534328A (en) * | 2003-03-21 | 2004-10-06 | ������������ʽ���� | Parallactic screen and multi-view display |
GB2405517A (en) * | 2003-08-30 | 2005-03-02 | Sharp Kk | Multiple view display |
CN1591089A (en) * | 2003-07-28 | 2005-03-09 | 三星电子株式会社 | Image displaying unit of a 3D image system having multi-viewpoints |
CN1607840A (en) * | 2003-08-05 | 2005-04-20 | 三星电子株式会社 | Apparatus and method for generating 3D image signal using space-division method |
JP2005110113A (en) * | 2003-10-01 | 2005-04-21 | Nec Corp | Multi-visual point image transmission system and method, multi-visual point image compression apparatus and method, and multi-visual point image decompression apparatus, method and program |
-
2005
- 2005-06-24 CN CNB2005100774732A patent/CN100481963C/en not_active Expired - Fee Related
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1332430A (en) * | 2001-07-27 | 2002-01-23 | 南开大学 | 3D tracking and measurement method of moving objects by 2D code |
CN1534328A (en) * | 2003-03-21 | 2004-10-06 | ������������ʽ���� | Parallactic screen and multi-view display |
CN1591089A (en) * | 2003-07-28 | 2005-03-09 | 三星电子株式会社 | Image displaying unit of a 3D image system having multi-viewpoints |
CN1607840A (en) * | 2003-08-05 | 2005-04-20 | 三星电子株式会社 | Apparatus and method for generating 3D image signal using space-division method |
GB2405517A (en) * | 2003-08-30 | 2005-03-02 | Sharp Kk | Multiple view display |
JP2005110113A (en) * | 2003-10-01 | 2005-04-21 | Nec Corp | Multi-visual point image transmission system and method, multi-visual point image compression apparatus and method, and multi-visual point image decompression apparatus, method and program |
Non-Patent Citations (2)
Title |
---|
基于MPEG4多重辅助部件的多视点视频编码方案. 张虎,孙立峰,钟玉琢.计算机工程与应用. 2005 |
基于MPEG4多重辅助部件的多视点视频编码方案. 张虎,孙立峰,钟玉琢.计算机工程与应用. 2005 * |
Also Published As
Publication number | Publication date |
---|---|
CN1688168A (en) | 2005-10-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP1661384B1 (en) | Semantics-based motion estimation for multi-view video coding | |
CN101193302B (en) | Illumination Compensation Method and Device and Video Codec Method and Device | |
CN102254348B (en) | Virtual viewpoint mapping method based o adaptive disparity estimation | |
US9137518B2 (en) | Method and system for converting 2D image data to stereoscopic image data | |
US9361699B2 (en) | Imaging system and method | |
US20120320986A1 (en) | Motion vector estimation method, multiview video encoding method, multiview video decoding method, motion vector estimation apparatus, multiview video encoding apparatus, multiview video decoding apparatus, motion vector estimation program, multiview video encoding program, and multiview video decoding program | |
EP1932366A2 (en) | Adaptive motion search range | |
WO2009097714A1 (en) | Depth searching method and depth estimating method for multi-viewing angle video image | |
CN108460792B (en) | Efficient focusing stereo matching method based on image segmentation | |
US20150172715A1 (en) | Picture encoding method, picture decoding method, picture encoding apparatus, picture decoding apparatus, picture encoding program, picture decoding program, and recording media | |
JP2007036800A (en) | Video encoding method, video decoding method, video encoding program, video decoding program, and computer-readable recording medium on which these programs are recorded | |
CN100481963C (en) | Visual difference calculating method for video image in multi-vision point system | |
JP4605716B2 (en) | Multi-view image compression encoding method, apparatus, and program | |
JP4944046B2 (en) | Video encoding method, decoding method, encoding device, decoding device, program thereof, and computer-readable recording medium | |
CN104202606B (en) | One kind determines method based on HEVC estimation starting points | |
CN103139569B (en) | The coding of multi-view point video, coding/decoding method, device and codec | |
Cheung et al. | On modeling the rendering error in 3D video | |
CN102378009B (en) | Fast algorithm for estimating optical parallax | |
An et al. | Theory and experiment analysis of disparity for stereoscopic image pairs | |
CN113610055A (en) | All-optical video sequence intra-frame prediction method based on gradient information | |
CN105187813B (en) | A kind of time domain uniformity deep video method of estimation based on adaptive weight | |
CN102215417A (en) | Parallax prediction method capable of establishing mathematical model based on block matching | |
CN114332188B (en) | Parallax image generation method, device and medium for binocular vision device | |
Zhang et al. | An improved depth map estimation for coding and view synthesis | |
KR101138921B1 (en) | Apparatus and method for encoding multi-view video |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20090422 Termination date: 20140624 |
|
EXPY | Termination of patent right or utility model |