CN102075779A - Intermediate view synthesizing method based on block matching disparity estimation - Google Patents
Intermediate view synthesizing method based on block matching disparity estimation Download PDFInfo
- Publication number
- CN102075779A CN102075779A CN2011100415383A CN201110041538A CN102075779A CN 102075779 A CN102075779 A CN 102075779A CN 2011100415383 A CN2011100415383 A CN 2011100415383A CN 201110041538 A CN201110041538 A CN 201110041538A CN 102075779 A CN102075779 A CN 102075779A
- Authority
- CN
- China
- Prior art keywords
- view
- piece
- parallax
- target image
- right view
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Landscapes
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
技术领域technical field
本发明涉及一种双目数字图像的处理方法,特别涉及一种双目数字图像中间视点合成方法。The invention relates to a processing method of a binocular digital image, in particular to a method for synthesizing intermediate viewpoints of a binocular digital image.
背景技术Background technique
随着多媒体技术的不断发展,图像和视频技术也由二维向三维发展,交互性将成为未来多媒体技术的一个主要特征。交互式三维视频系统的关键技术之一就是虚拟视点合成技术。虚拟视点绘合成是所有立体显示系统终端不可缺少的模块,在远程视频会议、自由视点立体电视等高端多媒体领域中也具有至关重要的作用。为了使用户可以在场景中漫游,实现“连续的环视”,在多视点视频采集的过程中,摄像机的数量应尽可能多,但由于放置无限个摄像机以实现视点无缝切换的不现实性,为了显示任意视点的视图,必须在客户端进行虚拟视点的合成,通过对已有视点的分析,合成用户所要观察的视点。因此,虚拟视点合成技术是多媒体领域一项非常重要的新兴技术。With the continuous development of multimedia technology, image and video technology is also developing from two-dimensional to three-dimensional, and interactivity will become a main feature of future multimedia technology. One of the key technologies of interactive 3D video system is virtual view synthesis technology. Virtual viewpoint rendering synthesis is an indispensable module for all stereoscopic display system terminals, and it also plays a vital role in high-end multimedia fields such as remote video conferencing and free viewpoint stereoscopic TV. In order to enable users to roam in the scene and realize "continuous look around", the number of cameras should be as many as possible during the multi-view video collection process, but due to the impracticality of placing infinite cameras to achieve seamless switching of viewpoints, In order to display the view of any viewpoint, the virtual viewpoint must be synthesized on the client side, and the viewpoint that the user wants to observe must be synthesized by analyzing the existing viewpoint. Therefore, virtual view synthesis technology is a very important emerging technology in the field of multimedia.
虚拟视点合成技术,在原理上分为基于模型的绘制方法(GBR)和基于图像的绘制方法(IBR)两种,GBR的优点在于几何模型表现全面且能满足灵活多变的视角变化需求,作为传统算法已经发展得比较成熟,交互性好。但该方法有两个主要缺点,即创建高精度的三维物体及场景模型计算量庞大,几乎不可能完成;以及计算费时,实时性差,绘制时间取决于物体的复杂性。另外,如果最终目的仅是合成中间视,那么建立三维模型可能是不需要的。鉴于此,可以直接利用真实场景作为参考视点的图像合成新视图的IBR方法逐渐得到关注和发展。IBR方法生成的视图真实感较强,与场景的融合较好,而且由于避免了重建过程,使其更便于进行复杂场景的实时绘制,具有运算复杂度不随场景变化而变化,无需对场景进行几何建模等特点。基于视差的中间视合成方法正是IBR的分支。Virtual viewpoint synthesis technology is divided into model-based rendering (GBR) and image-based rendering (IBR) in principle. The advantage of GBR is that the geometric model is comprehensive and can meet the needs of flexible and changeable viewing angles. The traditional algorithm has been developed relatively mature and has good interactivity. However, this method has two main disadvantages, that is, creating a high-precision 3D object and scene model is computationally intensive and almost impossible; and the calculation is time-consuming, with poor real-time performance, and the rendering time depends on the complexity of the object. Also, building a 3D model may not be necessary if the ultimate goal is only to synthesize mesoscopic vision. In view of this, the IBR method, which can directly use the real scene as the reference viewpoint image to synthesize a new view, has gradually received attention and development. The view generated by the IBR method has a strong sense of reality and is better integrated with the scene, and because it avoids the reconstruction process, it is more convenient for real-time rendering of complex scenes, and the computational complexity does not change with the scene. modeling features. The method of intermediate view synthesis based on parallax is just a branch of IBR.
在各种IBR方法当中,基于深度图像的虚拟绘制(Depth-Image Based Rendering,DIBR)发展得比较成熟,这类方法通过利用深度摄像机拍摄或利用各原始视图之间的几何约束和视差信息来获取每个彩色像素对应的深度信息,并根据深度信息绘制三维场景的虚拟图像。例如,蒋刚毅等人提出的利用参考视点深度图的平滑度将彩色图像分割成大小不同的块,然后进行三维图像变换将每个块投影到虚拟视点,以完成对该视点虚拟视图的绘制的方法(参见蒋刚毅,朱波,郁梅.一种基于可分级块的虚拟视点图像绘制方法[P].中国发明专利:200910153324.8,2009-10-14.)。又如骆凯等人提出首先对深度图像进行形态学处理,然后利用视点变换方程生成目标视点,最后对含有空洞的目标视点采用图像修复算法进行后处理的方法(参见骆凯,李东晓,冯雅美,张明.基于DIBR和图像修复的任意视点绘制[J].中国图象图形学报,2010,15(3):443-449.)。DIBR方法将场景的深度信息引入到虚拟视点场景当中,减少了虚拟视点绘制所需的参考视点数目,但深度图像的获取是非常困难的,深度摄像机可以直接获取场景深度信息,但造价昂贵,多目摄像机组标定困难;而直接从彩色图像中提取出每个像素的深度信息是非常困难的,由于深度信息的抽密度和准确度极大影响着虚拟视点的合成精度,从原始视图中提取深度信息往往是不可靠的。Among the various IBR methods, Depth-Image Based Rendering (DIBR) has developed relatively maturely. This type of method is obtained by using the depth camera or using the geometric constraints and disparity information between the original views. Each color pixel corresponds to the depth information, and draws a virtual image of the three-dimensional scene according to the depth information. For example, Jiang Gangyi et al proposed to use the smoothness of the reference viewpoint depth map to divide the color image into blocks of different sizes, and then perform three-dimensional image transformation to project each block to the virtual viewpoint to complete the drawing of the virtual view of the viewpoint. Method (see Jiang Gangyi, Zhu Bo, Yu Mei. A virtual viewpoint image rendering method based on scalable blocks [P]. Chinese invention patent: 200910153324.8, 2009-10-14.). Another example is that Luo Kai et al. proposed a method of morphologically processing the depth image first, then using the viewpoint transformation equation to generate the target viewpoint, and finally using an image inpainting algorithm to post-process the target viewpoint containing holes (see Luo Kai, Li Dongxiao, Feng Yamei , Zhang Ming. Arbitrary viewpoint rendering based on DIBR and image inpainting [J]. Chinese Journal of Image and Graphics, 2010, 15(3): 443-449.). The DIBR method introduces the depth information of the scene into the virtual viewpoint scene, which reduces the number of reference viewpoints required for virtual viewpoint rendering, but it is very difficult to obtain the depth image. The depth camera can directly obtain the scene depth information, but it is expensive and many It is difficult to calibrate the target camera group; it is very difficult to directly extract the depth information of each pixel from the color image, because the pumping density and accuracy of the depth information greatly affect the synthesis accuracy of the virtual viewpoint, and it is difficult to extract the depth information from the original view. Information is often unreliable.
采用几何方法获取深度信息的过程中,有一个步骤就是求视差场,因此,直接使用视差来进行虚拟视点绘制是方便的。在这类方法当中,视差估计的准确性是获得好的中间视图的关键,因此,目前的算法主要集中于如何获得准确的视差场。但是精准的视差场需要很大的计算量,这在实际应用当中并不实用。经典的视差估计算法总是无法在精确度和快速性两个方面取得平衡,这些算法在搜索视差时,为了达到较高的准确度,往往采用复杂的约束条件和搜索策略,但对精度的提高作用并不明显,反而使得算法运行时间加长;采用在局部区域搜索像素的最佳匹配点可以提高视差估计的效率,但代价是视差估计精度较全局搜索降低。In the process of acquiring depth information using geometric methods, one step is to calculate the parallax field. Therefore, it is convenient to directly use the parallax to draw the virtual viewpoint. Among such methods, the accuracy of disparity estimation is the key to obtaining a good intermediate view, so current algorithms mainly focus on how to obtain accurate disparity fields. However, accurate parallax fields require a large amount of calculation, which is not practical in practical applications. Classical disparity estimation algorithms are always unable to strike a balance between accuracy and rapidity. These algorithms often use complex constraints and search strategies in order to achieve higher accuracy when searching for disparity, but the improvement of accuracy The effect is not obvious, but it makes the running time of the algorithm longer; the efficiency of disparity estimation can be improved by searching for the best matching point of pixels in the local area, but the cost is that the accuracy of disparity estimation is lower than that of global search.
在视差估计环节当中,吕朝辉等人在视差估计中提出了一种基于自适应权值的方法,并在匹配的代价函数中引入视差平滑性约束项,这种方法获得的视差场是较为准确的,但由于约束条件复杂,计算量大大增加(参见吕朝晖,袁惇.基于视差估计的中间视合成[J].光电子·激光.2007,18(7):855-858)。骆艳等人提出在视差估计中加入图像分割算法,在获得视差场之后,对输入图像进行基于灰度值的图像分割,将图像分为具有灰度值不同的若干区域,区域内部的点则具有灰度相似性。在求出视差场之后,根据同一分割区域的像素点具有相同或相似的视差值这一假设,对视差场进行平滑(参见骆艳,安平,张兆扬.基于视差场校正和区域分割的中间视图像生成与内插方法[J].通信学报.2004.25(10):127-133.)。这样做可以纠正一部分错误匹配的视差,在一定程度上达到视差场平滑的目的,但会造成另一部分原来正确的视差被赋予错误的视差值,而且,图像分割模块的增加,使整个算法的计算量增加。In the process of disparity estimation, Lu Chaohui et al. proposed a method based on adaptive weights in disparity estimation, and introduced disparity smoothness constraints into the matching cost function. The disparity field obtained by this method is more accurate. , but due to complex constraint conditions, the amount of calculation is greatly increased (see Lu Zhaohui, Yuan Dun. Mesopic synthesis based on disparity estimation [J]. Optoelectronics·Laser. 2007, 18(7): 855-858). Luo Yan et al. proposed to add an image segmentation algorithm to the disparity estimation. After obtaining the disparity field, the input image is segmented based on the gray value, and the image is divided into several regions with different gray values. The points inside the region are have a grayscale similarity. After the disparity field is obtained, the disparity field is smoothed according to the assumption that pixels in the same segmented region have the same or similar disparity value (see Luo Yan, Anping, Zhang Zhaoyang. Intermediate view based on disparity field correction and region segmentation Image generation and interpolation method [J]. Journal of Communications. 2004.25(10): 127-133.). Doing so can correct a part of the wrongly matched disparity, and achieve the purpose of smoothing the disparity field to a certain extent, but it will cause another part of the original correct disparity to be given a wrong disparity value. Moreover, the increase of the image segmentation module makes the whole algorithm The amount of calculation increases.
发明内容Contents of the invention
本发明要解决的技术问题是:为克服现有技术的不足,本发明提供一种基于块匹配视差估计的中间视图合成方法,在不影响合成视图质量的同时大大降低了运算复杂度,节省了算法运行时间。The technical problem to be solved by the present invention is: in order to overcome the deficiencies of the prior art, the present invention provides an intermediate view synthesis method based on block matching disparity estimation, which greatly reduces the computational complexity and saves Algorithm runtime.
本发明解决其技术问题所采用的技术方案包括:一种基于块匹配视差估计的中间视图合成方法,其特征在于包括以下步骤:The technical solution adopted by the present invention to solve the technical problems includes: a method for synthesizing intermediate views based on block matching parallax estimation, characterized in that it includes the following steps:
(1)输入拍摄自同一场景,同一时刻,摄像机位于同一水平高度的左右两幅图像,要求这两幅图像仅在拍摄视角上存在差异;(1) Input two left and right images taken from the same scene at the same moment, and the camera is at the same horizontal height. It is required that the two images only have differences in shooting angles;
(2)若两幅输入图像为彩色图像,则分别将其转化为灰度图像;若两幅输入图像为灰度图像,则执行步骤(3);(2) If the two input images are color images, then convert them into grayscale images respectively; if the two input images are grayscale images, then perform step (3);
(3)判断两幅输入图像尺寸是否相同,若不同,提示错误并跳出;若相同,执行步骤(4);(3) Determine whether the size of the two input images is the same, if not, prompt an error and jump out; if they are the same, perform step (4);
(4)为一种基于块匹配的视差估计方法,具体是以右视图为目标图像,左视图为参考图像,将目标视图分为固定大小的块,针对目标图像中每个块,在参考图像中逐个搜索与之最相近的块,计算出每个目标图像块与参考图像中匹配块之间的位移矢量dLR,即为左视图到右视图中的块视差;再以左视图为目标图像,右视图为参考图像,重复步骤(4)求出右视图到左视图中的块视差dRL;(4) is a disparity estimation method based on block matching. Specifically, the right view is used as the target image, and the left view is used as the reference image. The target view is divided into blocks of fixed size. For each block in the target image, in the reference image Search for the closest blocks one by one in , and calculate the displacement vector d LR between each target image block and the matching block in the reference image, which is the block disparity from the left view to the right view; then take the left view as the target image , the right view is a reference image, repeat step (4) to obtain the block parallax d RL from the right view to the left view;
(5)根据步骤(4)中求出的左视图到右视图的块视差dLR,利用双目视觉视差原理,可以求出经过初步视点合成后基于左视图到右视图块视差的中间视图IM L;(其中上标M代表“中间视图”;下标L代表“左视图到右视图”);根据步骤(4)中求出的右视图到左视图的块视差dRL,同理可以求出经过初步视点合成后基于右视图到左视图块视差的中间视图IM R(其中下标R代表“右视图到左视图”)理论上,dLR与dRL、IM L与IM R应相等,但由于视差块匹配的误差是无法完全消除的,所以需要进一步对初步视点合成后的两个中间视图进行处理;(5) According to the block parallax d LR from the left view to the right view obtained in step (4), using the principle of binocular vision parallax, the middle view I based on the block parallax from the left view to the right view after preliminary viewpoint synthesis can be obtained M L ; (wherein the superscript M represents "middle view"; the subscript L represents "left view to right view"); according to the block parallax d RL from the right view to the left view obtained in step (4), in the same way Calculate the middle view I M R based on the block disparity from the right view to the left view after preliminary view synthesis (the subscript R stands for "right view to left view") Theoretically, d LR and d RL , I M L and I M R should be equal, but since the error of disparity block matching cannot be completely eliminated, it is necessary to further process the two intermediate views after the initial viewpoint synthesis;
(6)为一种在左视图和右视图中分别寻找待合成视图最佳匹配点的方法,具体是对于步骤(5)中求出的IM L与IM R这两个初步合成的中间视图,针对其中每一个像素,在左视图和右视图中逐个寻找IM L与IM R中每个像素的最佳匹配点IL(xL,y)和IR(xR,y)(其中,IL和IR代表左视图和右视图中的像素点灰度值;(x,y)代表像素点的坐标;xL和xR分别代表在左视图和右视图中找到的最佳匹配点的横坐标值,由于双目摄像机处于同一水平高度,所以在左视图和右视图中找到的最佳匹配点的纵坐标值相等,用y表示),并依据最佳匹配点的加权对IM L与IM R进行空洞填充(即使用一定的灰度值填补IM L与IM R中没有映射到得空白区域),最终得到中间视图合成的结果IM。(6) is a method for finding the best matching point of the view to be synthesized in the left view and the right view respectively, specifically for the intermediate points of the two preliminary synthesis of I M L and I M R obtained in step (5). View, for each pixel, find the best matching points I L (x L , y) and I R (x R , y) for each pixel in I M L and I M R in the left view and right view one by one (Wherein, I L and I R represent the pixel gray value in the left view and the right view; (x, y) represent the coordinates of the pixel; x L and x R represent the maximum value found in the left view and the right view respectively The abscissa value of the best matching point, since the binocular cameras are at the same horizontal height, the ordinate values of the best matching point found in the left view and the right view are equal, denoted by y), and based on the weighting of the best matching point Hole filling is performed on I ML and I MR (that is, a certain gray value is used to fill the blank area that is not mapped to I ML and I MR ) , and finally the result of intermediate view synthesis I M is obtained.
上述所述步骤(4)中的基于块匹配的视差估计方法,采用以下步骤实现:The parallax estimation method based on block matching in the above-mentioned step (4) adopts the following steps to realize:
(i)对参考图像进行扩边处理,在参考图像的左侧和右侧分别增加k个像素单位(k须满足包含所有双目视图中只存在于一幅视图的场景内容),并令这些像素的灰度值为0,将增补的边缘置为黑色;(i) Carry out edge expansion processing on the reference image, add k pixel units on the left and right sides of the reference image respectively (k must satisfy the scene content that only exists in one view in all binocular views), and make these The grayscale value of the pixel is 0, and the supplementary edge is set to black;
(ii)将目标图像分为M×N的块,其中M为每块宽度,N为长度;(ii) the target image is divided into M * N blocks, where M is the width of each block, and N is the length;
(iii)求出目标图像与参考图像对应位置块的SAD,作为搜索时用于比较的初始值,其中块的大小为M×N,左上角坐标为(m,n)的目标图像中的块与左上角坐标为(p,q)的参考图像块之间的绝对值误差和最小绝对差SAD为:(iii) Find the SAD of the block corresponding to the target image and the reference image, as the initial value for comparison during the search, where the size of the block is M×N, and the coordinates of the upper left corner are (m, n) of the block in the target image The absolute value error and the minimum absolute difference SAD between the reference image block whose coordinates are (p, q) in the upper left corner are:
其中,(m,n)为目标图像中块左上角的像素点坐标;(p,q)为参考图像中左上角的像素点坐标;I1、I2分别为目标图像和参考图像在某一坐标点的灰度值;Among them, (m, n) is the pixel coordinates of the upper left corner of the block in the target image; ( p , q) is the pixel coordinates of the upper left corner of the reference image; The gray value of the coordinate point;
(iv)在参考图像中,将搜索起点设置为目标图像块的左上角坐标(m,n),在横坐标区间为(m-60,m+60)范围内进行匹配,求出每个匹配位置与待匹配目标图像块间的SAD,使SAD值取得最小的参考图像对应位置即为最佳匹配块,并保留这个SAD最小值;(iv) In the reference image, set the search starting point as the coordinates (m, n) of the upper left corner of the target image block, perform matching within the range of (m-60, m+60) in the abscissa interval, and find out each matching The SAD between the position and the target image block to be matched, so that the corresponding position of the reference image with the smallest SAD value is the best matching block, and the minimum SAD value is retained;
(v)将搜索到的最佳匹配块位置记录下来,并求出目标块与最佳匹配块之间的位移矢量d,其中d(i,j)=(m-p,n-q),即视差;(v) record the best matching block position searched, and find the displacement vector d between the target block and the best matching block, where d (i, j)=(m-p, n-q), i.e. parallax;
(vi)如果接受匹配的是目标图像中位于右下角的块,不包括增补的边缘,即左上角坐标为(X+60-M,Y+60-N),则结束匹配;否则,找到下一个待匹配的目标图像块,返回步骤(iii);(vi) If the block in the lower right corner of the target image is accepted for matching, excluding the supplementary edge, that is, the coordinates of the upper left corner are (X+60-M, Y+60-N), then end the matching; otherwise, find the next A target image block to be matched, return to step (iii);
(v)将以块为单位的视差扩展至以像素为单位,即d(i×M+m,j×N+n)=d(i,j);其中m∈[0,M-1],n∈[0,N-1]。以左视图为参考图像,右视图为目标图像,可以计算出左视图到右视图的块视差dLR;再以右视图为参考图像,左视图为目标图像,可以计算出右视图到左视图的块视差dRL。本发明采用这种方法求出dLR和dRL,用于步骤(5)和步骤(6)等。求出这两个视差,才能够继续进行中间视图合成。(v) Extend the block-based disparity to pixel-based units, ie d(i×M+m, j×N+n)=d(i, j); where m∈[0,M-1] , n ∈ [0, N-1]. Taking the left view as the reference image and the right view as the target image, the block disparity d LR from the left view to the right view can be calculated; then using the right view as the reference image and the left view as the target image, the distance from the right view to the left view can be calculated Block parallax d RL . The present invention adopts this method to obtain d LR and d RL for use in step (5) and step (6). Only by calculating these two disparities can the intermediate view synthesis be continued.
上述所述步骤(5)中,利用双目视觉视差原理求出初步视点合成得到的中间视图的方法如下:In above-mentioned step (5), utilize binocular visual parallax principle to obtain the method for the intermediate view that preliminary viewpoint synthesis obtains as follows:
位于同一水平高度的两个摄像机,同时对场景进行拍摄,根据双目视差原理,有:Two cameras at the same level shoot the scene at the same time. According to the principle of binocular parallax, there are:
xR=xL+dLR(x,y)x R = x L + d LR (x, y)
IR(x,y)=IL(x+dLR(x,y),y)I R (x, y) = I L (x+d LR (x, y), y)
其中,xL为左视图一点在摄像机坐标系中的横坐标,以像素为单位;XR为该点在右视图中对应点的横坐标;dLR为以左视图为目标图像,右视图为参考图像求出的视差值,Among them, x L is the abscissa of a point on the left view in the camera coordinate system, in pixels; X R is the abscissa of the corresponding point in the right view; d LR is the target image of the left view, and the right view is The disparity value obtained from the reference image,
虚拟视角的合成用如下亮度加权获得:The composition of the virtual perspective is obtained by weighting the brightness as follows:
IM(xL,y)=(1-α)IL(xL,y)+αIR(xR,y)I M (x L , y) = (1-α) I L (x L , y) + αI R (x R , y)
设α为位置参数,令α=0为左视图所在位置,α=1为右视图所在位置,则α在(0,1)区间内可以代表原始视图间直线(基线)上任意位置,而待合成的中间视与左右原始视图之间关系表示为:Let α be the position parameter, let α=0 be the position of the left view, and α=1 be the position of the right view, then α can represent any position on the straight line (baseline) between the original views in the interval (0, 1), and to be The relationship between the synthesized intermediate view and the left and right original views is expressed as:
xM=xL+αdLR(x,y)=xR+(1-α)dRL(x,y)x M =x L +αd LR (x,y)=x R +(1-α)d RL (x,y)
将以上关系代入加权公式可以分别得到用dLR和dRL求出的IM L与IM R如下:Substituting the above relationship into the weighting formula can obtain the I M L and I M R calculated by d LR and d RL respectively as follows:
IM L(xL+αdLR,y)=(1-α)IL(xL,y)+αIR(xL+dLR,y)I M L (x L +αd LR ,y)=(1-α)I L (x L ,y)+αI R (x L +d LR ,y)
IM R(xR+(1-α)dRL,y)=(1-α)IL(xR+dRL,y)+αIR(xR,y)I M R (x R +(1-α)d RL ,y)=(1-α)I L (x R +d RL ,y)+αI R (x R ,y)
其中,IM L和IM R为初步视点合成得到的中间视图;IL和IR分别为左右原始视图;xL和xR均可遍历整幅图像。Among them, I M L and I M R are the intermediate views obtained by preliminary viewpoint synthesis; I L and I R are the left and right original views respectively; x L and x R can traverse the entire image.
上述所述步骤(6)中,在左视图和右视图中分别寻找待合成视图最佳匹配点的方法实现为:In the above-mentioned step (6), the method of finding the best matching point of the view to be synthesized in the left view and the right view respectively is realized as:
(i)求出左视图到右视图的视差图dLR中的最大值dLR max;(i) Find the maximum value d LR max in the disparity map d LR from the left view to the right view;
(ii)求出函数gL(x)=|xI-(x+αdRL(x,y))|在x∈[xI-αdRL max,xI+αdRL max]区间内的最小值min(gL(x))=gL(xL),并保存使gL(x)取得最小的x值,记为xL,其中,xI为中间视图中点的横坐标,取值为xI∈[0,图像宽度-1];(ii) Calculate the minimum value of the function g L (x)=|x I -(x+αd RL (x, y))| in the interval of x∈[x I -αd RL max , x I +αd RL max ] Value min(g L (x))=g L (x L ), and save the x value that makes g L (x) obtain the minimum value, denoted as x L , wherein, x I is the abscissa of the middle point of the middle view, take The value is x I ∈ [0, image width - 1];
(iii)求出右视图到左视图的视差图dRL中的最大值dRL max;(iii) find the maximum value d RL max in the disparity map d RL from the right view to the left view;
(iv)求出判定函数gR(x)=|xI-(x+αdRL(x,y))|在x∈[xI-αdRL,xI+αdRL]区间内的最小值min(gR(x))=gR(xR),并保存xR;(iv) Find the minimum value of the judgment function g R (x)=|x I -(x+αd RL (x, y))| in the interval of x∈[x I -αd RL , x I +αd RL ] min(g R (x)) = g R (x R ), and save x R ;
(v)按照亮度加权公式IM(x,y)=(1-α)IL(xL,y)+αIR(xR,y),最终得到中间视图合成的结果IM。(v) According to the brightness weighting formula I M (x, y) = (1-α) I L (x L , y) + α I R (x R , y), the result I M of intermediate view synthesis is finally obtained.
本发明与现有技术相比所具有的优点:Compared with the prior art, the present invention has the following advantages:
(1)本发明在视差估计环节采用在较大范围内的小步长搜索,搜索精度基本等同于穷举型的全局搜索,搜索精度高于经典视差估计算法的局部搜索方案;(1) The present invention adopts a small-step search in a larger range in the parallax estimation link, and the search accuracy is basically equal to the exhaustive global search, and the search accuracy is higher than the local search scheme of the classic parallax estimation algorithm;
(2)本发明在视差估计中根据用于合成中间视的原始视图,是依据双目视差原理拍摄自同一水平高度的两幅图像,理论上不存在垂直方向的视差,故仅对水平方向进行匹配搜索,大大减少了运算量,减少了算法运行时间;(2) In the parallax estimation, according to the original view for synthesizing intermediate vision, the present invention is based on the principle of binocular parallax to shoot two images from the same horizontal height. In theory, there is no parallax in the vertical direction, so only the horizontal direction is performed. Matching search greatly reduces the amount of computation and the running time of the algorithm;
(3)实验表明,本发明采用的方法能够合成具有良好视觉效果的中间视图,特别是前景距离摄像机较远或场景的灰度变化比较平缓时;并且在合成视图精度基本不变的情况下对视图合成速度有有效改善。(3) experiments show that the method adopted in the present invention can synthesize an intermediate view with good visual effect, especially when the foreground is far away from the camera or the grayscale change of the scene is relatively gentle; View composition speed has been significantly improved.
附图说明Description of drawings
图1是本发明基于块匹配视差估计的中间视图合成算法流程图;Fig. 1 is the flow chart of the intermediate view synthesis algorithm based on block matching parallax estimation in the present invention;
图2是本发明中视差估计部分流程图;Fig. 2 is a flow chart of the parallax estimation part in the present invention;
图3是本发明中中间视图合成部分流程图;Fig. 3 is a flow chart of the middle view synthesis part in the present invention;
图4是原始视图中的左视图;Figure 4 is the left view in the original view;
图5是原始视图中的右视图;Figure 5 is a right view in the original view;
图6是本发明中初步合成结果中的左视图;Fig. 6 is the left side view in the preliminary synthesis result among the present invention;
图7是本发明中初步合成结果中的右视图;Fig. 7 is the right side view in the preliminary synthesis result among the present invention;
图8是本发明中最终的中间视图合成结果,位置参数为α=0.25;Fig. 8 is the final intermediate view synthesis result in the present invention, and the position parameter is α=0.25;
图9是本发明中最终的中间视图合成结果,位置参数为α=0.5;Fig. 9 is the final intermediate view synthesis result in the present invention, and the position parameter is α=0.5;
图10是本发明中最终的中间视图合成结果,位置参数为α=0.75。Fig. 10 is the final synthesis result of the middle view in the present invention, and the position parameter is α=0.75.
具体实施方式Detailed ways
本发明的基于块匹配视差估计的中间视图合成方法,使用基于块匹配的视差估计分别得到从左视图到右视图以及从右视图到左视图的视差场,并分别根据这两个视差场进行中间视合成,最后利用视差场和初步合成的视图计算出待合成视图的每一个像素点在左右视图中的最佳匹配点,根据对应位置最佳匹配点的亮度加权为待合成视图中的每一点赋亮度值,以完成两视图连线上任意一个位置的视图合成。The intermediate view synthesis method based on block matching disparity estimation of the present invention uses the disparity estimation based on block matching to respectively obtain the disparity fields from the left view to the right view and from the right view to the left view, and perform intermediate view according to these two disparity fields respectively. View synthesis, and finally use the parallax field and the initially synthesized view to calculate the best matching point of each pixel of the view to be synthesized in the left and right views, and weight each point in the view to be synthesized according to the brightness of the best matching point at the corresponding position Assign a brightness value to complete the view synthesis at any position on the line connecting the two views.
本实施例中所有“视图”、“图像”均指数字位图,横坐标为从左至右,纵坐标为从上至下,均从0开始计数,像素点表示形式为(X,Y)。图1所示为本发明的基于块匹配视差估计的中间视图合成算法流程图;图2所示为本发明中视差估计部分的流程;图3所示为中间视图合成部分流程图,具体步骤如下:All "views" and "images" in this embodiment refer to digital bitmaps, the abscissa is from left to right, and the ordinate is from top to bottom, all counting from 0, and the pixel point representation form is (X, Y) . Fig. 1 shows the flow chart of the intermediate view synthesis algorithm based on block matching parallax estimation in the present invention; Fig. 2 shows the process flow of the disparity estimation part in the present invention; Fig. 3 shows the flow chart of the intermediate view synthesis part, and the specific steps are as follows :
(1)输入拍摄自同一场景,同一时刻,摄像机位于同一水平高度的两幅图像,要求这两幅图像仅在拍摄视角上存在差异。对一组输入图像的以上要求满足双目原理,分别相当于用“左右眼”对场景进行观察和记录。而本发明所要完成的工作即是合成并输出“左右眼”之间任意一点处对场景进行观察所得到的结果;(1) Input two images taken from the same scene at the same moment, and the camera is at the same horizontal height. It is required that the two images only have differences in shooting angles. The above requirements for a set of input images meet the binocular principle, which is equivalent to observing and recording the scene with "left and right eyes". And the work to be completed by the present invention is to synthesize and output the result obtained by observing the scene at any point between the "left and right eyes";
(2)若两幅输入图像为彩色图像,则分别将其转化为灰度图像;若两幅输入图像为灰度图像,则执行步骤(3)。在彩色图像转化为灰度图时,按照三个颜色分量RGB比重分别为0.11、0.59、0.30进行加权,所得结果为对应像素灰度值;(2) If the two input images are color images, convert them into grayscale images respectively; if the two input images are grayscale images, perform step (3). When the color image is converted into a grayscale image, weighting is performed according to the proportions of the three color components RGB being 0.11, 0.59, and 0.30 respectively, and the obtained result is the grayscale value of the corresponding pixel;
(3)判断两幅输入图像尺寸是否相同,若不同,提示错误并跳出程序;若相同,执行步骤(4)。如果用于合成中间视图的左右图像尺寸不同,则无法将两幅图的对应像素进行视图插值,从而得到最终结果;(3) Determine whether the sizes of the two input images are the same, if not, prompt an error and jump out of the program; if they are the same, execute step (4). If the left and right images used to synthesize the middle view have different sizes, the corresponding pixels of the two images cannot be interpolated to obtain the final result;
(4)以右视图为目标图像,左视图为参考图像,将目标视图分为固定大小的块,在参考图像中搜索与之最相近的块,计算出每个目标图像块与参考图像中匹配块之间的位移矢量,即左视图到右视图中的块视差dLR;再以左视图为目标图像,右视图为参考图像,重复步骤(4)求出右视图到左视图中的块视差dRL,具体流程如图2所示;(4) Take the right view as the target image and the left view as the reference image, divide the target view into blocks of fixed size, search for the closest block in the reference image, and calculate the matching between each target image block and the reference image The displacement vector between the blocks, that is, the block parallax d LR from the left view to the right view; then take the left view as the target image, and the right view as the reference image, and repeat step (4) to find the block parallax from the right view to the left view d RL , the specific process is shown in Figure 2;
(5)根据步骤(4)中求出的dLR,根据双目视觉视差原理求出经过初步视点合成后基于左视图到右视图块视差的中间视图IM L;根据步骤(4)中求出的dRL,求出经过初步视点合成后基于右视图到左视图块视差的中间视图IM R。由于IM L和IM R的求解过程利用的视差信息不完全,求IM L时仅用到dLR,而求IM R时仅用到dRL,并且加权公式采用前向映射,故IM L和IM R中会产生空洞,即没有填充的区域,需要进行进一步处理;(5) According to the d LR obtained in step (4), obtain the intermediate view I M L based on the block parallax from the left view to the right view after preliminary viewpoint synthesis according to the principle of binocular visual parallax; The obtained d RL is used to obtain the intermediate view I MR based on the block disparity between the right view and the left view after preliminary view synthesis. Because the disparity information used in the process of solving I ML and I MR is incomplete, only d LR is used when calculating I ML , and only d RL is used when calculating I MR , and the weighting formula adopts forward mapping , so Voids are generated in IML and IMR , that is, areas without filling , which require further processing;
(6)对于求出的IM L与IM R这两个初步合成的中间视图,针对其中每一个像素,按照图3所示流程,分别在左视图和右视图中逐个寻找IM L与IM R中每个像素的最佳匹配点IL(xL,y)和IR(xR,y),并依据最佳匹配点的灰度值加权对初步合成的视图进行空洞填充,最终得到中间视图合成的结果IM。(6) For the two initially synthesized intermediate views of I M L and I M R obtained, for each pixel, according to the process shown in Figure 3, search for I M L and I M R in the left view and right view one by one. The best matching points I L (x L , y) and I R (x R , y) of each pixel in I M R are weighted according to the gray value of the best matching point to fill holes in the initially synthesized view, Finally, the result I M of intermediate view synthesis is obtained.
其中,上述所述步骤(4)中的基于块匹配的视差估计方法,采用以下步骤实现:Wherein, the parallax estimation method based on block matching in the above-mentioned step (4) is realized by the following steps:
(i)对参考图像进行扩边处理,在参考图像的左侧和右侧分别增加k个像素单位(k须满足包含所有双目视图中只存在于一幅视图的场景内容,经实验,本发明取k=60),并令其为0,将增补的部分置为黑色。这些增补的黑色区域将用于搜索靠近目标图像左右边缘的块时,使参考图像的边缘不超过搜索范围;(i) Perform edge expansion processing on the reference image, and add k pixel units on the left and right sides of the reference image respectively (k must satisfy the scene content that only exists in one view in all binocular views. After experiments, this The invention takes k=60), and let it be 0, and set the added part to black. These supplementary black areas will be used to search for blocks close to the left and right edges of the target image, so that the edges of the reference image do not exceed the search range;
(ii)将目标图像分为M×N的块。一般处理图像时为了表述简便,通常使M=N,在本发明说明书附图中的范例当中,采用大小为16×16的块进行匹配;(ii) Divide the target image into M×N blocks. In general, for the sake of simplicity of expression when processing images, M=N is usually used. In the example in the accompanying drawings of the specification of the present invention, a block with a size of 16×16 is used for matching;
(iii)求出目标图像与参考图像对应位置块的SAD,作为搜索时用于比较的初始值。(iii) Obtain the SAD of the block corresponding to the target image and the reference image, and use it as an initial value for comparison during searching.
SAD的定义为两组元素数目相同的数组对应元素相减的绝对值之和。在本发明中,块的大小为M×N,左上角坐标为(m,n)的目标图像中的块与左上角坐标为(p,q)的参考图像块之间的绝对值误差和SAD为:SAD is defined as the sum of the absolute values of the subtraction of the corresponding elements of two arrays with the same number of elements. In the present invention, the size of the block is M×N, and the absolute value error and SAD for:
其中,(m,n)为目标图像中块左上角的像素点坐标;(p,q)为参考图像中左上角的像素点坐标;I1、I2分别为目标图像和参考图像在某一坐标点的灰度值;Among them, (m, n) is the pixel coordinates of the upper left corner of the block in the target image; ( p , q) is the pixel coordinates of the upper left corner of the reference image; The gray value of the coordinate point;
该定义对下文中所有SAD均适用;This definition applies to all SADs hereinafter;
(iv)在参考图像中,将搜索起点设置为目标图像块的左上角坐标(m,n),在横坐标区间为(m-60,m+60)范围内进行匹配,求出每个匹配位置与待匹配目标图像块间的SAD,使SAD值取得最小的参考图像对应位置即为最佳匹配块,设这个最佳匹配块的坐标为(p,q),并保留这个SAD最小值;(iv) In the reference image, set the search starting point as the coordinates (m, n) of the upper left corner of the target image block, perform matching within the range of (m-60, m+60) in the abscissa interval, and find out each matching The SAD between the position and the target image block to be matched, so that the corresponding position of the reference image with the smallest SAD value is the best matching block, the coordinates of this best matching block are (p, q), and the minimum SAD value is reserved;
(v)将搜索到的最佳匹配块位置(p,q)记录下来,并求出目标块与最佳匹配块之间的位移矢量d,以像素为单位,其中d(i,j)=(m-p,n-q),即视差矢量。在本发明中,由于所取输入图像为在同一水平高度上拍摄,因此视差矢量的垂直方向上的分量为零,仅存在水平方向上的分量;(v) Record the searched best matching block position (p, q), and find the displacement vector d between the target block and the best matching block, in pixels, where d(i, j) = (m-p, n-q), which is the disparity vector. In the present invention, since the input images taken are taken at the same horizontal height, the vertical component of the parallax vector is zero, and only the horizontal component exists;
(vi)如果接受匹配的是目标图像中位于最后一行有意义的最后一块,也就是目标图像右下角(不包括增补的边缘)的块,其左上角坐标为(X+60-M,Y+60-N),则结束匹配;否则,找到下一个待匹配的目标图像块,返回步骤(iii)。采用这种方法对分成若干块的目标图像进行遍历,最终将得到目标图像中所有块的视差值;(vi) If the accepted match is the last meaningful block in the last line of the target image, that is, the block in the lower right corner of the target image (excluding the supplementary edge), the coordinates of the upper left corner are (X+60-M, Y+ 60-N), then end the matching; otherwise, find the next target image block to be matched, and return to step (iii). Using this method to traverse the target image divided into several blocks, the disparity values of all blocks in the target image will be obtained finally;
(v)将以块为单位的视差扩展至以像素为单位,即d(i×M+m,j×N+n)=d(i,j);其中m∈[0,M-1],n∈[0,N-1]。以左视图为参考图像,右视图为目标图像,可以计算出左视图到右视图的块视差dLR;再以右视图为参考图像,左视图为目标图像,可以计算出右视图到左视图的块视差dRL。(v) Extend the block-based disparity to pixel-based units, ie d(i×M+m, j×N+n)=d(i, j); where m∈[0,M-1] , n ∈ [0, N-1]. Taking the left view as the reference image and the right view as the target image, the block disparity d LR from the left view to the right view can be calculated; then using the right view as the reference image and the left view as the target image, the distance from the right view to the left view can be calculated Block parallax d RL .
上述所述步骤(5)中,根据双目视觉视差原理求出初步中间视图的方法具体如下:In the above-mentioned step (5), the method for obtaining the preliminary intermediate view according to the principle of binocular visual parallax is specifically as follows:
使用位于同一水平高度的两个摄像机,同时对场景进行拍摄,根据双目视差原理,拍摄得到的两视图具有如下几何关系:Use two cameras at the same level to shoot the scene at the same time. According to the principle of binocular parallax, the two views obtained by shooting have the following geometric relationship:
xR=xL+dLR(x,y)x R = x L + d LR (x, y)
IR(x,y)=IL(x+dLR(x,y),y)I R (x, y) = I L (x+d LR (x, y), y)
其中,xL为左视图一点在摄像机坐标系中的横坐标,以像素为单位;xR为该点在右视图中对应点的横坐标;dLR为以左视图为目标图像,右视图为参考图像求出的视差值。Among them, x L is the abscissa of a point on the left view in the camera coordinate system, in pixels; x R is the abscissa of the corresponding point in the right view; d LR is the target image of the left view, and the right view is The disparity value obtained from the reference image.
这两摄像机连线上任意位置虚拟视角的合成,可以使用如下亮度加权公式获得:The composition of the virtual viewing angle at any position on the connection between the two cameras can be obtained using the following brightness weighting formula:
IM(xL,y)=(1-α)IL(xL,y)+αIR(xR,y)I M (x L , y) = (1-α) I L (x L , y) + αI R (x R , y)
其中,IM为要合成的中间视图,α为位置参数,令α=0为左视图所在位置,α=1为右视图所在位置,则α在(0,1)区间内可以代表原始视图间直线(基线)上任意位置。而待合成的中间视与左右原始视图之间可用如下关系表示:Among them, I M is the middle view to be synthesized, α is the position parameter, let α=0 be the position of the left view, α=1 is the position of the right view, then α can represent the original view in the (0,1) interval Any position on the straight line (baseline). The relationship between the intermediate view to be synthesized and the left and right original views can be expressed as follows:
xM=xL+αdLR(x,y)=xR+(1-α)dRL(x,y)x M =x L +αd LR (x,y)=x R +(1-α)d RL (x,y)
将以上关系代入加权公式可以分别得到用dLR和dRL求出的IM L与IM R如下:Substituting the above relationship into the weighting formula can obtain the I M L and I M R calculated by d LR and d RL respectively as follows:
IM L(xL+αdLR,y)=(1-α)IL(xL,y)+αIR(xL+dLR,y)I M L (x L +αd LR ,y)=(1-α)I L (x L ,y)+αI R (x L +d LR ,y)
IM R(xR+(1-α)dRL,y)=(1-α)IL(xR+dRL,y)+αIR(xR,y)I M R (x R +(1-α)d RL ,y)=(1-α)I L (x R +d RL ,y)+αI R (x R ,y)
其中,IM L和IM R为初步视点合成得到的中间视图;IL和IR分别为左右原始视图;xL和xR均可遍历整幅图像。Among them, I M L and I M R are the intermediate views obtained by preliminary viewpoint synthesis; I L and I R are the left and right original views respectively; x L and x R can traverse the entire image.
采用这两个公式对图4、图5进行中间视图合成,合成的视图如图6、图7所示。由于每幅合成的视图只利用了一个视差场,即以左视图为目标图像合成的视图只用到了dRL,而以右视图为目标图像合成的视图只用到了dLR,视差信息的缺乏使得合成的结果不够准确,存在着较多误差;更为明显的是,这一步骤中采用的合成公式使用前向映射方法,即等式左边待赋值的IM,坐标并不能遍历到整幅图像,因此导致合成的视图当中存在很多空白的没有映射到的区域,非常影响观赏,不能够被当作最终合成结果。Use these two formulas to synthesize the intermediate views of Figures 4 and 5, and the synthesized views are shown in Figures 6 and 7. Since each synthesized view uses only one disparity field, that is, the view synthesized with the left view as the target image only uses d RL , while the view synthesized with the right view as the target image only uses d LR , the lack of disparity information makes The result of the synthesis is not accurate enough, and there are many errors; more obviously, the synthesis formula used in this step uses the forward mapping method, that is, the coordinates of the I M to be assigned on the left side of the equation cannot traverse the entire image , so there are many blank unmapped areas in the synthesized view, which greatly affects viewing and cannot be regarded as the final synthesized result.
上述所述步骤(6)中,在左视图和右视图中分别寻找待合成视图最佳匹配点,是由于在上述所述步骤(5)中求出的中间视图缺乏视差信息,并且合成公式采用前向映射方法,导致合成的视图当中存在误差和很多空白区域,非常影响观赏,不能够被当作最终合成结果,所以必须进一步进行处理,对未映射到得空洞区域进行填充,方法参见图3的流程图,具体采用如下步骤实现:In the above-mentioned step (6), the best matching points of the views to be synthesized are respectively searched for in the left view and the right view, because the middle view obtained in the above-mentioned step (5) lacks disparity information, and the synthesis formula adopts The forward mapping method leads to errors and many blank areas in the synthesized view, which greatly affects viewing and cannot be regarded as the final synthesis result. Therefore, further processing is required to fill the unmapped empty areas. The method is shown in Figure 3. The flow chart is implemented in the following steps:
(i)遍历左视图到右视图的视差图dLR,求出dLR中的最大值dLR max;(i) traverse the disparity map d LR from the left view to the right view, and find the maximum value d LR max in d LR ;
(ii)在右图像中,求出判定函数gL(x)=|xI-(x+αdRL(x,y))|在x∈[xI-αdRL max,xI+αdRL max]区间内的最小值min(gL(x))=gL(xL),并保存使gL(x)取得最小的x值,记为xL,xL即为该点的最佳匹配点。其中,xI为中间视图中点的横坐标,取值为xI∈[0,图像宽度-1];遍历整幅参考图像,即可得到左视图在右视图中的所有最佳匹配点;(ii) In the right image, obtain the decision function g L (x)=|x I -(x+αd RL (x,y))|at x∈[x I -αd RL max , x I +αd RL max ] the minimum value min(g L (x))=g L (x L ) in the interval, and save the x value that makes g L (x) get the minimum value, denoted as x L , x L is the minimum value of the point best match point. Among them, x I is the abscissa of the middle point in the middle view, and the value is x I ∈ [0, image width - 1]; traversing the entire reference image, you can get all the best matching points of the left view in the right view;
(iii)求出右视图到左视图的视差图dRL中的最大值dRL max;(iii) find the maximum value d RL max in the disparity map d RL from the right view to the left view;
(iv)在左图像中,求出判定函数gR(x)=|xI-(x+αdRL(x,y))|在x∈[xI-αdRL,xI+αdRL]区间内的最小值min(gR(x))=gR(xR),并保存使gR(x)取得最小的x值,记为xR,xR即为该点的最佳匹配点;(iv) In the left image, determine the decision function g R (x)=|x I -(x+αd RL (x,y))| at x∈[x I -αd RL , x I +αd RL ] The minimum value min(g R (x)) in the interval = g R (x R ), and save the value of x that makes g R (x) the smallest, denoted as x R , and x R is the best match at this point point;
(v)按照亮度加权公式IM(x,y)=(1-α)IL(xL,y)+αIR(xR,y),最终得到中间视图合成的结果IM,如图8、图9、图10,其位置参数分别为α=0.25、α=0.5、α=0.75。该公式充分利用了左视图、右视图的几何信息,以及左视图到右视图,和右视图到左视图的两个视差场,因此得到的结果是比较准确的。从附图实例证明,本发明确实得到了很好的视觉效果。(v) According to the brightness weighting formula I M (x, y) = (1-α) I L (x L , y) + α I R (x R , y), finally get the result I M of the intermediate view synthesis, as shown in the figure 8. As shown in Fig. 9 and Fig. 10, the position parameters are respectively α=0.25, α=0.5, and α=0.75. This formula makes full use of the geometric information of the left view and the right view, and the two disparity fields from the left view to the right view, and from the right view to the left view, so the obtained result is relatively accurate. Prove from accompanying drawing example, the present invention has really obtained good visual effect.
本发明说明书中未作详细描述的内容属于本领域专业技术人员公知的现有技术。The contents not described in detail in the description of the present invention belong to the prior art known to those skilled in the art.
Claims (4)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 201110041538 CN102075779B (en) | 2011-02-21 | 2011-02-21 | Intermediate view synthesizing method based on block matching disparity estimation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 201110041538 CN102075779B (en) | 2011-02-21 | 2011-02-21 | Intermediate view synthesizing method based on block matching disparity estimation |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102075779A true CN102075779A (en) | 2011-05-25 |
CN102075779B CN102075779B (en) | 2013-05-08 |
Family
ID=44034099
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN 201110041538 Expired - Fee Related CN102075779B (en) | 2011-02-21 | 2011-02-21 | Intermediate view synthesizing method based on block matching disparity estimation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102075779B (en) |
Cited By (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102254348A (en) * | 2011-07-25 | 2011-11-23 | 北京航空航天大学 | Block matching parallax estimation-based middle view synthesizing method |
CN102263957A (en) * | 2011-07-25 | 2011-11-30 | 北京航空航天大学 | A Disparity Estimation Method Based on Search Window Adaptation |
CN103108200A (en) * | 2011-11-09 | 2013-05-15 | 晨星软件研发(深圳)有限公司 | Processing method of stereoscopic images |
CN103384337A (en) * | 2012-05-02 | 2013-11-06 | 索尼公司 | Image processing apparatus and image processing method |
CN103458257A (en) * | 2012-05-31 | 2013-12-18 | 财团法人工业技术研究院 | Hole filling method for multi-view disparity map |
CN104065954A (en) * | 2014-07-03 | 2014-09-24 | 中国传媒大学 | A fast detection method for disparity range of high-definition stereoscopic video |
US9076249B2 (en) | 2012-05-31 | 2015-07-07 | Industrial Technology Research Institute | Hole filling method for multi-view disparity maps |
CN105007478A (en) * | 2015-07-08 | 2015-10-28 | 南京信息工程大学 | View synthesis method based on mean shift stereo matching |
CN105657401A (en) * | 2016-01-13 | 2016-06-08 | 深圳创维-Rgb电子有限公司 | Naked eye 3D display method and system and naked eye 3D display device |
CN103843336B (en) * | 2011-10-13 | 2016-08-24 | 索尼公司 | Image processing apparatus, image processing method |
CN107392965A (en) * | 2017-07-24 | 2017-11-24 | 王华锋 | A kind of distance-finding method being combined based on deep learning and binocular stereo vision |
CN107430782A (en) * | 2015-04-23 | 2017-12-01 | 奥斯坦多科技公司 | Method for being synthesized using the full parallax squeezed light field of depth information |
CN107493465A (en) * | 2017-09-18 | 2017-12-19 | 郑州轻工业学院 | A kind of virtual multi-view point video generation method |
WO2018014324A1 (en) * | 2016-07-22 | 2018-01-25 | 北京大学深圳研究生院 | Method and device for synthesizing virtual viewpoints in real time |
CN110177216A (en) * | 2019-06-28 | 2019-08-27 | Oppo广东移动通信有限公司 | Image processing method, image processing device, mobile terminal and storage medium |
CN110473247A (en) * | 2019-07-30 | 2019-11-19 | 中国科学院空间应用工程与技术中心 | Solid matching method, device and storage medium |
CN111464804A (en) * | 2020-04-08 | 2020-07-28 | 北京小米松果电子有限公司 | Omnidirectional parallax view synthesis method and device and storage medium |
TWI701639B (en) * | 2018-12-11 | 2020-08-11 | 緯創資通股份有限公司 | Method of identifying foreground object in image and electronic device using the same |
CN112365400A (en) * | 2020-10-27 | 2021-02-12 | 深圳大学 | Rapid super-resolution reconstruction method for light field angle |
CN113132706A (en) * | 2021-03-05 | 2021-07-16 | 北京邮电大学 | Controllable position virtual viewpoint generation method and device based on reverse mapping |
CN114612566A (en) * | 2020-12-04 | 2022-06-10 | 深圳超多维科技有限公司 | Image processing method, image processor and terahertz imaging system |
CN114881841A (en) * | 2022-04-19 | 2022-08-09 | 维沃移动通信有限公司 | Image generation method and device |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1688168A (en) * | 2005-06-24 | 2005-10-26 | 清华大学 | Visual difference calculating method for video image in multi-vision point system |
CN1767655A (en) * | 2005-10-18 | 2006-05-03 | 宁波大学 | A method for disparity estimation of multi-viewpoint video images |
US20070159476A1 (en) * | 2003-09-15 | 2007-07-12 | Armin Grasnick | Method for creating a stereoscopic image master for imaging methods with three-dimensional depth rendition and device for displaying a steroscopic image master |
CN101312540A (en) * | 2008-07-03 | 2008-11-26 | 浙江大学 | A Virtual Viewpoint Synthesis Method Based on Depth and Occlusion Information |
CN101873508A (en) * | 2010-05-25 | 2010-10-27 | 东南大学 | Method of mesoscopic synthesis based on improved BP algorithm |
-
2011
- 2011-02-21 CN CN 201110041538 patent/CN102075779B/en not_active Expired - Fee Related
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070159476A1 (en) * | 2003-09-15 | 2007-07-12 | Armin Grasnick | Method for creating a stereoscopic image master for imaging methods with three-dimensional depth rendition and device for displaying a steroscopic image master |
CN1688168A (en) * | 2005-06-24 | 2005-10-26 | 清华大学 | Visual difference calculating method for video image in multi-vision point system |
CN1767655A (en) * | 2005-10-18 | 2006-05-03 | 宁波大学 | A method for disparity estimation of multi-viewpoint video images |
CN101312540A (en) * | 2008-07-03 | 2008-11-26 | 浙江大学 | A Virtual Viewpoint Synthesis Method Based on Depth and Occlusion Information |
CN101873508A (en) * | 2010-05-25 | 2010-10-27 | 东南大学 | Method of mesoscopic synthesis based on improved BP algorithm |
Non-Patent Citations (3)
Title |
---|
吕朝辉等: "基于图像的中间视合成方法综述", 《电视技术》 * |
杨海涛等: "应用于多视点视频编码的基于深度特征的图像区域分割与区域视差估计", 《光学学报》 * |
梁广民: "立体视频会议中基于对象的中间视生成", 《计算机工程与应用》 * |
Cited By (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102263957A (en) * | 2011-07-25 | 2011-11-30 | 北京航空航天大学 | A Disparity Estimation Method Based on Search Window Adaptation |
CN102263957B (en) * | 2011-07-25 | 2013-07-03 | 北京航空航天大学 | Search-window adaptive parallax estimation method |
CN102254348A (en) * | 2011-07-25 | 2011-11-23 | 北京航空航天大学 | Block matching parallax estimation-based middle view synthesizing method |
CN103843336B (en) * | 2011-10-13 | 2016-08-24 | 索尼公司 | Image processing apparatus, image processing method |
CN103108200B (en) * | 2011-11-09 | 2016-03-02 | 晨星软件研发(深圳)有限公司 | To the processing method of stereopsis |
CN103108200A (en) * | 2011-11-09 | 2013-05-15 | 晨星软件研发(深圳)有限公司 | Processing method of stereoscopic images |
CN103384337A (en) * | 2012-05-02 | 2013-11-06 | 索尼公司 | Image processing apparatus and image processing method |
CN103384337B (en) * | 2012-05-02 | 2016-08-31 | 索尼公司 | Image processing apparatus and image processing method |
US9076249B2 (en) | 2012-05-31 | 2015-07-07 | Industrial Technology Research Institute | Hole filling method for multi-view disparity maps |
CN103458257B (en) * | 2012-05-31 | 2015-05-27 | 财团法人工业技术研究院 | Hole filling method for multi-view disparity map |
CN103458257A (en) * | 2012-05-31 | 2013-12-18 | 财团法人工业技术研究院 | Hole filling method for multi-view disparity map |
CN104065954B (en) * | 2014-07-03 | 2016-05-18 | 中国传媒大学 | A kind of disparity range method for quick of high definition three-dimensional video-frequency |
CN104065954A (en) * | 2014-07-03 | 2014-09-24 | 中国传媒大学 | A fast detection method for disparity range of high-definition stereoscopic video |
CN107430782A (en) * | 2015-04-23 | 2017-12-01 | 奥斯坦多科技公司 | Method for being synthesized using the full parallax squeezed light field of depth information |
CN105007478A (en) * | 2015-07-08 | 2015-10-28 | 南京信息工程大学 | View synthesis method based on mean shift stereo matching |
CN105657401A (en) * | 2016-01-13 | 2016-06-08 | 深圳创维-Rgb电子有限公司 | Naked eye 3D display method and system and naked eye 3D display device |
CN105657401B (en) * | 2016-01-13 | 2017-10-24 | 深圳创维-Rgb电子有限公司 | A kind of bore hole 3D display methods, system and bore hole 3D display device |
WO2018014324A1 (en) * | 2016-07-22 | 2018-01-25 | 北京大学深圳研究生院 | Method and device for synthesizing virtual viewpoints in real time |
CN107392965A (en) * | 2017-07-24 | 2017-11-24 | 王华锋 | A kind of distance-finding method being combined based on deep learning and binocular stereo vision |
CN107493465B (en) * | 2017-09-18 | 2019-06-07 | 郑州轻工业学院 | A kind of virtual multi-view point video generation method |
CN107493465A (en) * | 2017-09-18 | 2017-12-19 | 郑州轻工业学院 | A kind of virtual multi-view point video generation method |
US10964050B2 (en) * | 2018-12-11 | 2021-03-30 | Wistron Corporation | Method for identifying foreground object in image and electronic device |
TWI701639B (en) * | 2018-12-11 | 2020-08-11 | 緯創資通股份有限公司 | Method of identifying foreground object in image and electronic device using the same |
CN110177216A (en) * | 2019-06-28 | 2019-08-27 | Oppo广东移动通信有限公司 | Image processing method, image processing device, mobile terminal and storage medium |
CN110473247A (en) * | 2019-07-30 | 2019-11-19 | 中国科学院空间应用工程与技术中心 | Solid matching method, device and storage medium |
US11245886B2 (en) | 2020-04-08 | 2022-02-08 | Beijing Xiaomi Pinecone Electronics Co., Ltd. | Method and apparatus for synthesizing omni-directional parallax view, and storage medium |
CN111464804A (en) * | 2020-04-08 | 2020-07-28 | 北京小米松果电子有限公司 | Omnidirectional parallax view synthesis method and device and storage medium |
CN112365400A (en) * | 2020-10-27 | 2021-02-12 | 深圳大学 | Rapid super-resolution reconstruction method for light field angle |
CN112365400B (en) * | 2020-10-27 | 2024-05-28 | 深圳大学 | Rapid light field angle super-resolution reconstruction method |
CN114612566A (en) * | 2020-12-04 | 2022-06-10 | 深圳超多维科技有限公司 | Image processing method, image processor and terahertz imaging system |
CN113132706A (en) * | 2021-03-05 | 2021-07-16 | 北京邮电大学 | Controllable position virtual viewpoint generation method and device based on reverse mapping |
CN114881841A (en) * | 2022-04-19 | 2022-08-09 | 维沃移动通信有限公司 | Image generation method and device |
Also Published As
Publication number | Publication date |
---|---|
CN102075779B (en) | 2013-05-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102075779B (en) | Intermediate view synthesizing method based on block matching disparity estimation | |
CN102254348B (en) | Virtual viewpoint mapping method based o adaptive disparity estimation | |
CN104756489B (en) | A kind of virtual visual point synthesizing method and system | |
KR101415147B1 (en) | A Boundary Noise Removal and Hole Filling Method for Virtual Viewpoint Image Generation | |
TWI469088B (en) | Depth map generation module for foreground object and the method thereof | |
Cheng et al. | Spatio-temporally consistent novel view synthesis algorithm from video-plus-depth sequences for autostereoscopic displays | |
CN103248909B (en) | Method and system of converting monocular video into stereoscopic video | |
CN110853151A (en) | Three-dimensional point set recovery method based on video | |
WO2009097714A1 (en) | Depth searching method and depth estimating method for multi-viewing angle video image | |
CN104065946B (en) | Hole Filling Method Based on Image Sequence | |
US7505623B2 (en) | Image processing | |
CN102263957A (en) | A Disparity Estimation Method Based on Search Window Adaptation | |
CN106791774A (en) | Virtual visual point image generating method based on depth map | |
CN101873508A (en) | Method of mesoscopic synthesis based on improved BP algorithm | |
CN103679739A (en) | Virtual view generating method based on shielding region detection | |
Bleyer et al. | Temporally consistent disparity maps from uncalibrated stereo videos | |
Lie et al. | Key-frame-based background sprite generation for hole filling in depth image-based rendering | |
JP2013005025A (en) | Stereoscopic image generating device, stereoscopic image generating method, program, and recording medium | |
Lee et al. | High-Resolution Depth Map Generation by Applying Stereo Matching Based on Initial Depth Informaton | |
Lee et al. | Segment-based multi-view depth map estimation using belief propagation from dense multi-view video | |
Knorr et al. | From 2D-to stereo-to multi-view video | |
Zhu et al. | Virtual view synthesis using stereo vision based on the sum of absolute difference | |
KR20220071935A (en) | Method and Apparatus for Deriving High-Resolution Depth Video Using Optical Flow | |
CN109379577B (en) | Video generation method, device and equipment of virtual viewpoint | |
CN107610070A (en) | Free stereo matching process based on three shooting collections |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
C17 | Cessation of patent right | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20130508 Termination date: 20140221 |