[go: up one dir, main page]

CN103493482B - The method and apparatus of a kind of extraction and optimized image depth map - Google Patents

The method and apparatus of a kind of extraction and optimized image depth map Download PDF

Info

Publication number
CN103493482B
CN103493482B CN201280004184.8A CN201280004184A CN103493482B CN 103493482 B CN103493482 B CN 103493482B CN 201280004184 A CN201280004184 A CN 201280004184A CN 103493482 B CN103493482 B CN 103493482B
Authority
CN
China
Prior art keywords
pixel
value
correlation
depth
sampling
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201280004184.8A
Other languages
Chinese (zh)
Other versions
CN103493482A (en
Inventor
赵兴朋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hisense Visual Technology Co Ltd
Original Assignee
Qingdao Hisense Xinxin Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Qingdao Hisense Xinxin Technology Co Ltd filed Critical Qingdao Hisense Xinxin Technology Co Ltd
Publication of CN103493482A publication Critical patent/CN103493482A/en
Application granted granted Critical
Publication of CN103493482B publication Critical patent/CN103493482B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/59Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

The embodiment of the present invention relates to the method and apparatus of a kind of extraction and optimized image depth map.Described method comprises: the scene degree of correlation obtaining each pixel in current source picture and current source picture; To the continuous down-sampling of current source picture, obtain the scene degree of correlation of each pixel in current each down-sampling source images; The pixel that each pixel in current down-sampling source images is corresponding with down-sampling source images before carries out block matching motion Vector operation, obtains the motion vector value of each pixel in current down-sampling source images; In cumulative current down-sampling source images, the motion vector value of each pixel, extracts the ID value of each pixel from motion vector cumulative sum, and ID value forms ID figure; Utilize the scene degree of correlation of each pixel in the scene degree of correlation of each pixel in source images and each down-sampling source images to surpass the disposal of gentle filter and up-sampling process continuously to pixel each in ID figure, obtain source images depth map.

Description

The method and apparatus of a kind of extraction and optimized image depth map
Technical field
The present invention relates to technical field of image processing, especially relate to the method and apparatus of a kind of extraction and optimized image depth map.
Background technology
Three-dimensional stereo display technique high speed development recent years, three-dimensional terminal presentation facility (ThreeDimensions, 3D) as, 3D TV, the rapid emergence of 3D game machine etc. has become the inevitable outcome of development in science and technology.Due to 3D film source resource scarcity, involve great expense, so the 2D converting 3D sequence of frames of video to by common two dimension (TwoDimensions, 2D) sequence of frames of video turns 3D technology become hot fields in three-dimensional stereo display technique.
Turn in 3D technology at 2D, Focal point and difficult point is the extraction of 2D video image depth map.The physical significance of depth map is: in 2D sequence of frames of video, different pictures content is from the how far of beholder, is to form the topmost information source of 3D anaglyph.The ways and means of current extraction depth map is various, mainly comprises the depth information based on contour of object extraction object, the color segmentation based on image extracts depth map, extracts depth map based on Virtual Space intersection point, based on the motion vector extraction depth map of object and based on key frame semi-automatic extraction depth map etc.But the technology of these numerous extraction depth maps all exists serious defect mostly, the too large or Human disturbance factor of clear amount of calculation that depth map is smudgy is too many etc., is so just difficult to the display demand reaching 3D terminal presentation facility.
Prior art provides a kind of method generating two dimensional video sequence depth map, as shown in Figure 1, the key frame of the method first in selecting video frame sequence, and the artificial depth map generating key frame, then coupling estimates the moving displacement between video successive frame characteristic point, draws the depth map of present frame according to key frame depth maps and moving displacement.This method of document introduction can extract present frame depth map to a certain extent, but this method needs the depth map manually choosing and calculate key frame, is unfavorable for that the full-automation of depth map generates, and then is difficult to promote at industrial circle; Be a bit carrying out being easy to cause matching error time coupling is estimated in addition, and then depth map information also can cause matching error, and the depth map of extraction so often soft edge, depth information is concavo-convex unbalanced.
Summary of the invention
The need that the object of the invention is to solve prior art are manually chosen and are extracted the depth map of key frame, cause the depth map precision of extraction low, occur the problem of error, provide the method and apparatus of a kind of extraction and optimized image depth map.
In first aspect, embodiments provide a kind of method of extraction and optimized image depth map, described method comprises: the scene degree of correlation obtaining each pixel in current source picture and described current source picture, and described current source picture is current video successive frame sequence;
To the continuous down-sampling of described current source picture, obtain the scene degree of correlation of each pixel in current each down-sampling source images;
The pixel that each pixel in described current down-sampling source images is corresponding with down-sampling source images before carries out block matching motion Vector operation, obtains the motion vector value of each pixel in described current down-sampling source images;
The motion vector value of each pixel in cumulative described current down-sampling source images respectively, from motion vector cumulative sum, extract the ID value of each pixel, described ID value forms source images ID figure;
Utilize the scene degree of correlation of each pixel in the scene degree of correlation of each pixel in described source images and described each down-sampling source images to surpass the disposal of gentle filter and up-sampling process continuously to each pixel in described ID figure, obtain described source images depth map.
In second aspect, embodiments provide the device of a kind of extraction and optimized image depth map, described device comprises: the first acquiring unit, for obtaining the scene degree of correlation of each pixel in current source picture and described current source picture, described current source picture is current video successive frame sequence;
Second acquisition unit, for the continuous down-sampling of described current source picture, obtains the scene degree of correlation of each pixel in current each down-sampling source images;
3rd acquiring unit, carries out block matching motion Vector operation for the pixel that each pixel in described current down-sampling source images is corresponding with down-sampling source images before, obtains the motion vector value of each pixel in described current down-sampling source images;
Computing unit, for the motion vector value of each pixel in the described current down-sampling source images that adds up respectively, extract the ID value of each pixel from motion vector cumulative sum, described ID value forms source images ID figure;
First processing unit, for utilizing the scene degree of correlation of each pixel in the scene degree of correlation of each pixel in described source images and described each down-sampling source images to surpass the disposal of gentle filter and up-sampling process continuously to each pixel in described ID figure, obtain described source images depth map.
By application embodiment of the present invention the methods and apparatus disclosed, the corresponding scene degree of correlation is obtained in the different down-sampling stage, ID figure is extracted by motion vector cumulative sum, utilize the scene degree of correlation of different sample phase to carry out iteration to ID figure and surpass smoothing processing, final generation source images depth map, improve the picture quality of depth map, make depth map profile more clear, this method also allows the computing cost of whole process remain within zone of reasonableness simultaneously.
Accompanying drawing explanation
Fig. 1 is that prior art extracts depth map flow chart;
Fig. 2 method flow diagram that depth map extracts and optimizes disclosed in the embodiment of the present invention;
Fig. 3 is embodiment of the present invention current source picture schematic diagram;
Fig. 4 is the schematic diagram that the embodiment of the present invention calculates arbitrary pixel scene degree of correlation;
Fig. 5 is the schematic diagram of arbitrary pixel scene degree of correlation on embodiment of the present invention computation bound;
Fig. 6 is the ID figure of embodiment of the present invention source images;
Fig. 7 to be the embodiment of the present invention be arbitrary pixel assigns weight coefficient schematic diagram;
Fig. 8 to be the embodiment of the present invention be on border arbitrary pixel assigns weight coefficient schematic diagram;
Fig. 9 is the structure chart that the embodiment of the present invention surpasses smoothing filter;
Figure 10 is depth map after the optimization of embodiment of the present invention source images;
Figure 11 is the installation drawing of the disclosed extraction of the embodiment of the present invention and optimized image depth map.
Embodiment
For making the object of the embodiment of the present invention, technical scheme and advantage clearly, below in conjunction with the accompanying drawing in the embodiment of the present invention, technical scheme in the embodiment of the present invention is clearly and completely described, obviously, described embodiment is the present invention's part embodiment, instead of whole embodiments.Based on the embodiment in the present invention, those of ordinary skill in the art, not making the every other embodiment obtained under creative work prerequisite, belong to the scope of protection of the invention.
For ease of the understanding to the embodiment of the present invention, be further explained explanation below in conjunction with accompanying drawing with specific embodiment, embodiment does not form the restriction to the embodiment of the present invention.
The method of image procossing disclosed in the embodiment of the present invention is described in detail, Fig. 2 method flow diagram that depth map extracts and optimizes disclosed in the embodiment of the present invention for Fig. 2.
As shown in Figure 2, in embodiments of the present invention, first continuous current source picture is obtained, described current source picture is two-dimentional successive frame sequence, extract the scene degree of correlation of each pixel in current source picture, to current source picture respectively through horizontal and vertical twice 1/2 down-samplings operations, and after twice 1/2 down-samplings, carry out the scene degree of correlation extracting each pixel accordingly, obtain each pixel in each stage resolution source image the scene degree of correlation.
Utilize after twice 1/2 down-samplings, resolution is that the pixel corresponding with 1/16 resolution source image before of the pixel in the source images of 1/16 carries out block matching motion Vector operation, draw the motion vector value of pixel in current 1/16 resolution source image, the motion vector value of pixel in cumulative current 1/16 resolution source image, extract the ID value of pixel in current 1/16 resolution source image based on motion vector cumulative sum, the ID value of each pixel forms ID figure; But the resolution of described ID figure is 1/16 of source images, and therefore soft edge is unintelligible, also strict optimization process to be carried out to described ID figure.
Super the disposal of gentle filter based on the 1/16 resolution scene degree of correlation is carried out to 1/16 resolution ID figure, and carries out four iteration and surpass smothing filtering; 1/16 complete depth of resolution figure of the disposal of gentle filter is surpassed to iteration and carries out horizontal and vertical twice up-sampling respectively, obtain 1/4 resolution ID figure, super the disposal of gentle filter based on the 1/4 resolution scene degree of correlation is carried out to 1/4 depth of resolution figure, and carries out twice iteration and surpass smothing filtering; Further the depth map that twice up-sampling obtains original resolution size is carried out to the 1/4 depth of resolution figure processed, an iteration of carrying out again based on the scene degree of correlation surpasses the disposal of gentle filter, finally get the depth map after optimization process, the step of specific implementation as described below:
Step 210, acquisition two dimensional source image, and from source images, extract the scene degree of correlation;
Particularly, first obtain continuous current source picture, described current source picture is two-dimentional successive frame sequence, and as shown in Figure 3, Fig. 3 is embodiment of the present invention current source picture schematic diagram;
Meanwhile, from current source picture, the scene degree of correlation is extracted;
In embodiments of the present invention, the scene degree of correlation is arbitrary pixel (appoint get pixel centered by a pixel) and the degree of correlation of neighbor pixel around it in a two field picture.Utilize central pixel point place R (red), G (green), B (indigo plant) value subtract each other with the value of around neighbor pixel place R (red), G (green), B (indigo plant) successively, and get the absolute value of difference.If the absolute value of the difference of the neighbor on some directions is less than the relevance threshold preset, so, just the degree of correlation flag bit in the party upwards degree of correlation mark groove is arranged 1, otherwise arrange 0.Described degree of correlation mark groove buffer [] is the buffer of 8 bit wides, and 8 degree of correlation flag bits from lowest order to highest order deposit the degree of correlation information that central pixel point 8 neighbors nearest with it press clock-wise order rotation successively.Centered by storing 1, pixel is relevant to neighbor pixel, and centered by storing 0, pixel is uncorrelated with neighbor pixel.Any one pixel in a degree of correlation mark groove map source image.
As shown in Figure 4, Fig. 4 is the schematic diagram that the embodiment of the present invention calculates arbitrary pixel scene degree of correlation; Wherein, red pixel point is the central pixel point chosen, the coordinate of central pixel point is (x, y), the pixel adjacent with described central pixel point is neighbor pixel, in the diagram, there are 8 neighbor pixels, and 8 neighbors are pressed clockwise numbering, as No. 0-7, described numbering correspond to the lowest order of scene degree of correlation mark groove respectively to highest order;
According to preceding method, judge the degree of correlation of each neighbor pixel and central pixel point, if the neighbor pixel of numbering 1 and the relevant of central pixel point, then in degree of correlation mark groove buffer [1], store 1; Therefore, the formula of degree of correlation mark groove buffer [m] is:
buffer [ m ] = 1 , | f ( x , y ) - f ( x ± k , y ± k ) | R + | f ( x , y ) - f ( x ± k , y ± k ) | G + | f ( x , y ) - f ( x ± k , y ± k ) | B ≤ A ; 0 , | f ( x , y ) - f ( x ± k , y ± k ) | R + | f ( x , y ) - f ( x ± k , y ± k ) | G + | f ( x , y ) - f ( x ± k , y ± k ) | B > A ;
(formula 1)
Wherein, k=0,1; 0≤m≤7; M ∈ Z; F (x, y), f (x ± k, y ± k) are pixel R (red), the value of G (green), B (indigo plant) component; A is the threshold value of the scene degree of correlation; The m position that buffer [m] is the degree of correlation; M is neighbor piont mark.
At this, illustratively, the situation of the pixel degree of correlation on computation bound, as shown in Figure 5, Fig. 5 is the schematic diagram of arbitrary pixel scene degree of correlation on embodiment of the present invention computation bound, wherein, red pixel point is the central pixel point chosen, the coordinate of central pixel point is (x, y), the pixel adjacent with described central pixel point is neighbor pixel, in Figure 5, the coordinate of central pixel point gets x=0 and y=0, namely central pixel point is positioned at the first row of current source picture, first row, now, the mode of numbering as hereinbefore, only need calculate and be numbered 2 in dotted line frame, 3, the degree of correlation of 4 pixels, be numbered 0, 1, 5, 6, 7, pixel does not exist, therefore, directly by degree of correlation mark groove buffer [0], buffer [1], buffer [5], buffer [6], buffer [7] assignment is 0, pixel processing method on other boundary position is identical with it, therefore, repeat no more.
It should be noted that, the calculating degree of correlation of foregoing description is described for 1 pixel, in Practical Calculation, all to calculate each pixel, in source images each pixel as calculated after, all there is 1 degree of correlation mark groove, all degree of correlation mark grooves of all pixels constitute the scene degree of correlation of current source picture.
Step 220, horizontal and vertical 1/2 down-sampling is carried out to source images, obtain 1/4 resolution source image, and extract the scene degree of correlation from 1/4 resolution source image;
Particularly, to the current source picture obtained respectively through 1/2 horizontal and vertical down-sampling operation, and after 1/2 down-sampling, from each pixel of 1/4 resolution source image, extract the scene degree of correlation, obtain the scene degree of correlation of each pixel in 1/4 resolution source image and 1/4 resolution source image.
In the sample phase of current 1/4 resolution, calculate each pixel scene degree of correlation in 1/4 resolution source image by the method that step 210 describes.
It should be noted that, prior art is treated to the down-sampling of source images, does not repeat them here.
Step 230, horizontal and vertical 1/2 down-sampling is carried out again to 1/4 resolution source image, obtain 1/16 resolution source image, and extract the scene degree of correlation from 1/16 resolution source image;
Particularly, to 1/4 resolution source image obtained respectively through 1/2 horizontal and vertical down-sampling operation, and after 1/2 down-sampling, from 1/16 resolution source image, extract the scene degree of correlation, obtain the scene degree of correlation of each pixel in 1/16 resolution source image and 1/16 resolution source image.
In the sample phase of current 1/16 resolution, calculate each pixel scene degree of correlation in 1/16 resolution source image by the method that step 210 describes.
It should be noted that, prior art is treated to the down-sampling of 1/4 resolution source image, does not repeat them here.
Step 240,1/16 resolution source image and before 1/16 resolution source image carry out block matching motion Vector operation;
Particularly, utilize in step 230,1/16 resolution source image after down-sampling carries out block matching motion Vector operation with 1/16 resolution source image before, obtain the motion vector value of each pixel in current 1/16 resolution source image, and the motion vector value of each pixel in current 1/16 resolution source image obtained is added up.
It should be noted that, the motion vector based on Block-matching is cumulative is also prior art, does not repeat them here.
Extract ID value in step 250, motion vector cumulative sum, form ID figure;
Particularly, according to the description of step 240, the motion vector value obtaining each pixel in current 1/16 resolution source image adds up, the ID value of each pixel is extracted from described motion vector cumulative sum, all extract ID value to pixel whole in 1/16 resolution source image, the ID value of whole pixel forms the ID figure of 1/16 resolution source image;
At this, illustratively, based in motion vector cumulative sum, ID value is extracted to each pixel;
Assuming that in continuous two source images, the maximum offset of moving object is 3.5% width of current source picture, and now, the gray value that representative motion vector value is corresponding is 255, the following formula of gray value size so representated by unit picture element displacement:
I = 255 W * 3.5 % (formula 2)
Wherein, W is the width of image;
If the image block motion vector mould calculated is the value shown in table 1, at this, illustrate by way of example, in table 1, calculate 9 motion vector moulds.
Table 1 motion vector mould
a b c
d e f
g h k
So, the gray value that above-mentioned 9 motion vector moulds are corresponding is:
Table 2 motion vector gray value
a*I b*I c*I
d*I e*I f*I
g*I h*I k*I
Illustrate that the source images of acquisition is two-dimensional video frame sequence aforementioned, the depth information of arbitrary pixel is extracted from described two-dimensional frames sequence, in order to keep pixel after stop motion, its depth value still exists, so that obtain its movable information at any time, by the depth value of arbitrary pixel stored in degree of depth buffer.Otherwise once pixel stopped motion in current source picture, so, in source images, the motion vector of this pixel will be zero afterwards, now, if directly will be obtained the result of mistake by current pixel point motion vector value compute depth information.Therefore, the accumulated value for the depth information before pixel that degree of depth buffer is deposited, because degree of depth buffer exists maximum, so the maximum of controlling depth buffer is D total.
Owing to having obtained the gray value of arbitrary pixel in the preamble, the cumulative sum of the whole gray value of current depth figure has been D new, before in degree of depth buffer, whole gray value cumulative sums of all depth maps are D accif, always D newsimply be added to D accin, finally can exceed the maximum D of degree of depth buffer cumulative sum totaland cause overflowing, cause pixel depth information to run off; Therefore, in embodiments of the present invention,
If D new+ D acc< D totaltime, then D acc_depth(x, y)=D acc_depth(x, y)+D new_depth(x, y) (formula 3)
If D new+ D acc> D totaltime, then D acc _ depth ( x , y ) = D acc _ depth ( x , y ) * D total - D new D acc (formula 4)
If D total - D new D acc < 0 , Then D total - D new D acc = 0 ; (formula 5)
If D total - D new D acc > 1 , Then D total - D new D acc = 1 . (formula 6)
Wherein 0≤x≤h-1,0≤y≤w-1.H, w are respectively height and the width of down-sampling 1/16 resolution source image.D acc_depth(x, y) be expressed as each pixel before the cumulative sum of motion vector gray value, D new_depth(x, y) is expressed as each pixel current motion vector gray value.
It should be noted that, the ID value of extracting from motion vector cumulative sum of foregoing description is described for partial pixel point, in Practical Calculation, all to extract each pixel, extract each pixel ID value in 1/16 resolution source image after, form the ID figure of 1/16 resolution source image, as shown in Figure 6, Fig. 6 is the ID figure of embodiment of the present invention source images.
Step 260, super the disposal of gentle filter is carried out to 1/16 resolution ID figure, and up-sampling process;
Particularly, from the ID figure of Fig. 6, due in step 250, in block matching motion Vector operation process, there is comparatively big error, cause extracted ID map contour smudgy clear, therefore, in this step strict optimization process is carried out to 1/16 resolution ID figure.
According to the description of step 230, obtain the scene degree of correlation of 1/16 resolution source image in step 230, therefore, the scene degree of correlation according to 1/16 resolution source image obtained carries out super the disposal of gentle filter to 1/16 resolution ID figure, and carries out four iterative filterings; Then, horizontal and vertical twice up-sampling is carried out respectively to the 1/16 resolution ID figure that iteration surpasses the disposal of gentle filter complete, obtains 1/4 resolution ID figure.
In embodiments of the present invention, according to the description of step 230, obtained the scene degree of correlation of 1/16 resolution source image pixel, the scene degree of correlation according to 1/16 resolution source image pixel is optimized process to 1/16 resolution ID figure; The pixel adjacent with central pixel point is defined in the calculating degree of correlation, in super the disposal of gentle filter in this step, different weight coefficients is distributed again by each neighbor pixel, as shown in Figure 7, Fig. 7 to be the embodiment of the present invention be arbitrary pixel assigns weight coefficient schematic diagram;
The weight coefficient of each neighbor pixel is respectively as the filter taps coefficient of super smoothing filter.Described super smoothing filter is low pass filter, because this filter coefficient factor is 8, and to be regularly distributed on 8 directions of central pixel point, so, filtering performance is high, can effectively concavo-convex sharp keen high-frequency noise and high fdrequency component in level and smooth ID figure.
When the central pixel point chosen is not on the border of ID figure, according to the degree of correlation mark groove at central pixel point place, obtain the correlation circumstance of central pixel point and 8 neighbor pixels.If a certain neighbor pixel and the central pixel point degree of correlation are 1, the weight coefficient of this neighbor pixel is then multiplied by with the gray value of neighbor pixel, if a certain neighbor pixel and the central pixel point degree of correlation are 0, the gray value of central pixel point is then multiplied by with the weight coefficient of neighbor pixel, finally, the result that 8 neighbor pixels are multiplied has been added up be used as the result after to 1/16 resolution ID figure smothing filtering.
To be the embodiment of the present invention to assign weight coefficient schematic diagram to a border pixel of taking up an official post Fig. 8, as shown in Figure 8, wherein, red pixel point is the central pixel point chosen, the coordinate of central pixel point is (x, y), the pixel adjacent with described central pixel point is neighbor pixel, in fig. 8, the coordinate of central pixel point gets x=0 and y=0, namely central pixel point is positioned at the first row of current source picture, first row, now, the mode of numbering as hereinbefore, only need according to being numbered 2 in dotted line frame, 3, the degree of correlation of 4 pixels carries out filtering, be numbered 0, 1, 5, 6, 7, pixel does not exist, therefore, degree of correlation mark groove buffer [0], buffer [1], buffer [5], buffer [6], buffer [7] value is 0, pixel processing method on other boundary position is identical with it, therefore, repeat no more.
Fig. 9 is the structure chart that the embodiment of the present invention surpasses smoothing filter, and as shown in Figure 9, the coef0-coef7 in Fig. 9 is the weight coefficient of neighbor pixel, is also the tap coefficient of super smoothing filter; In embodiments of the present invention, coef0=coef2=coef4=coef6=1/6 is set; Coef1=coef3=coef5=coef7=1/12; When setting weight coefficient, the weight coefficient of 8 neighbor pixels be met and be 1, be i.e. coef0+coef1+coef2+coef3+coef4+coef5+coef6+coef7=1.
Therefore, to the formula that arbitrary pixel carries out filtering be:
f ( x , y ) = [ 1 6 * &Sigma; n = 0 3 [ ~ buffer [ 2 n ] ] + 1 12 * &Sigma; n = 0 3 [ ~ buffer [ 2 n + 1 ] ] ] * f ( x , y ) + 1 6 * &Sigma; n = 0 3 [ buffer [ 2 n ] * f ( 2 n ) ] + 1 12 * &Sigma; n = 0 3 [ buffer [ 2 n + 1 ] * f ( 2 n + 1 ) ] (formula 7)
Wherein, n ∈ Z, n=0,1,2,3; The scene degree of correlation that buffer [] is neighbor pixel; The negate of the scene degree of correlation that ~ buffer [] is neighbor pixel; The gray value at pixel (x, y) place centered by f (x, y).
Carrying out four iteration to described 1/16 resolution ID figure surpasses after the disposal of gentle filter completes, each pixel in 1/16 complete resolution ID figure of the disposal of gentle filter is surpassed to iteration and carries out horizontal and vertical twice up-sampling respectively, obtain the depth value of each 1/4 definition pixel point, the depth value of each 1/4 definition pixel point forms 1/4 depth of resolution figure, obtains 1/4 depth of resolution figure.
It should be noted that being described for 1 pixel ID figure filtering according to the degree of correlation of foregoing description, in Practical Calculation, all will carry out filtering to each pixel.
Up-sampling iteration being surpassed to 1/16 complete resolution ID figure of the disposal of gentle filter is treated to prior art, does not repeat them here.
Step 270, super the disposal of gentle filter is carried out to 1/4 depth of resolution figure, and up-sampling process;
Particularly, 1/4 depth of resolution figure is obtained according to step 260, according to the description of step 220, obtain the scene degree of correlation of 1/4 resolution source image in a step 220, therefore, the scene degree of correlation according to 1/4 resolution source image obtained carries out super the disposal of gentle filter to 1/4 depth of resolution figure, and carries out twice iterative filtering; Then, horizontal and vertical twice up-sampling is carried out respectively to the 1/4 resolution ID figure that iteration surpasses the disposal of gentle filter complete, obtain the depth value of each original resolution pixel, the depth value of each original resolution pixel forms original resolution depth map, obtains the ID figure of original resolution;
In filtering in this step, by the method that step 260 describes, super smothing filtering is carried out to 1/4 depth of resolution figure.
Depth map after step 280, acquisition optimization process;
Particularly, obtain original resolution depth map according to step 270, according to the description of step 210, obtain the scene degree of correlation of source images in step 210, therefore, according to the scene degree of correlation of the source images obtained, an iteration is carried out to original resolution depth map and surpass the disposal of gentle filter; Finally, source images depth map is obtained.
As shown in Figure 10, Figure 10 is depth map after source images optimization process; Compared with the ID figure of Fig. 6 source images, ID figure resolution is lower, pixel is less, image outline is unintelligible, carrying out after comparatively successive ignition surpasses smothing filtering and up-sampling, ID figure resolution raises, and pixel becomes many, depth map profile is more clear, improves the picture quality of depth map.
By method disclosed in the application embodiment of the present invention, the corresponding scene degree of correlation is obtained in the different down-sampling stage, ID figure is extracted by motion vector cumulative sum, utilize the scene degree of correlation in different down-sampling stage to carry out iteration to ID figure and surpass smoothing processing, carry out up-sampling process, final generation source images depth map, improves the picture quality of depth map simultaneously, make depth map profile more clear, this method also allows the computing cost of whole process remain within zone of reasonableness simultaneously.
Correspondingly, above-described embodiment describes the method for extraction and optimized image depth map, and correspondingly, the device of also usable image process realizes.As shown in figure 11, Figure 11 is the installation drawing of the disclosed extraction of the embodiment of the present invention and optimized image depth map.The device of described extraction and optimized image depth map comprises: the first acquiring unit 1110, and for obtaining the scene degree of correlation of each pixel in current source picture and described current source picture, described current source picture is current video successive frame sequence;
Second acquisition unit 1120, for the continuous down-sampling of described current source picture, obtains the scene degree of correlation of each pixel in current each down-sampling source images;
3rd acquiring unit 1130, carry out block matching motion Vector operation for the pixel that each pixel in described current down-sampling source images is corresponding with down-sampling source images before, obtain the motion vector value of each pixel in described current down-sampling source images;
Computing unit 1140, for the motion vector value of each pixel in the described current down-sampling source images that adds up respectively, extract the ID value of each pixel from motion vector cumulative sum, described ID value forms source images ID figure;
First processing unit 1150, for utilizing the scene degree of correlation of each pixel in the scene degree of correlation of each pixel in described source images and described each down-sampling source images to surpass the disposal of gentle filter and up-sampling process continuously to each pixel in described ID figure, obtain described source images depth map.
In described device, the first acquiring unit 1110 specifically for: select arbitrary pixel as central pixel point, will with described central pixel point neighbor piont mark;
Obtain the difference of the red R of described central pixel point, green G, blue B component value and the red R of each described neighbor pixel, green G, blue B component value, and described difference is taken absolute value;
By described absolute value compared with scene relevance threshold, if described absolute value is less than scene relevance threshold, then the scene degree of correlation of described neighbor pixel is set to 1 and is correlated with, otherwise, be set to 0 uncorrelated;
The scene degree of correlation of described neighbor pixel stored in a buffer, described buffer is specially
buffer [ m ] = 1 , | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | R + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | G + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | B &le; A ; 0 , | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | R + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | G + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | B > A ;
Wherein, k=0,1; 0≤m≤7; M ∈ Z; F (x, y), f (x ± k, y ± k) are the red R of pixel, green G, the value of blue B component; A is scene relevance threshold; The m position that buffer [m] is the degree of correlation; M is neighbor piont mark.
In described device, second acquisition unit 1120 specifically for: 1/2 horizontal and vertical down-sampling is carried out to described current source picture, obtains the scene degree of correlation of each pixel in current 1/4 resolution source image and current 1/4 resolution source image;
1/2 horizontal and vertical down-sampling is carried out to described current 1/4 resolution source image, obtains the scene degree of correlation of each pixel in current 1/16 resolution source image and current 1/16 resolution source image.
In described device, 3rd acquiring unit 1130 specifically for: the pixel that each pixel in described current 1/16 resolution source image is corresponding with described 1/16 resolution source image before carries out block matching motion Vector operation, obtains the motion vector value of each pixel in described current 1/16 resolution source image;
The motion vector value of each pixel in cumulative described current 1/16 resolution source image respectively.
In described device, computing unit 1140 specifically for: the motion vector mould and the unit picture element displacement gray value that obtain each pixel displacement, described unit picture element displacement gray value is w is the width of image;
Described motion vector mould is multiplied with described unit picture element displacement gray value, obtains the motion vector gray value of described each pixel;
The cumulative sum of the motion vector gray value of described each pixel is stored in degree of depth buffer;
If D new+ D acc< D total; Then D acc_depth(x, y)=D acc_depth(x, y)+D new_depth(x, y);
If D new+ D acc> D total; Then D acc _ depth ( x , y ) = D acc _ depth ( x , y ) * D total - D new D acc ;
If D total - D new D acc < 0 , Then D total - D new D acc = 0 ;
If D total - D new D acc > 1 , Then D total - D new D acc = 1 ;
Wherein, described (x, y) is each pixel coordinate; D acc_depth(x, y) is the cumulative sum of motion vector gray value before each pixel; D new_depth(x, y) is each pixel current motion vector gray value; D totalfor the maximum of degree of depth buffer cumulative sum, D newfor the cumulative sum of the whole gray value of current depth figure; D accfor whole gray value cumulative sums of all depth maps before in degree of depth buffer.
In described device, the first processing unit 1150 specifically for: for described each neighbor pixel assigns weight coefficient, described weight coefficient is the tap coefficient of super smothing filtering;
Call in described buffer the scene degree of correlation of the described neighbor pixel stored;
If the scene degree of correlation of described neighbor pixel be 1 be correlated with, then the weight coefficient that neighbor pixel gray value and neighbor pixel distribute is multiplied;
If the scene degree of correlation of described neighbor pixel is 0 uncorrelated, then the weight coefficient that central pixel point gray value and neighbor pixel distribute is multiplied;
The value that cumulative described neighbor pixel gray value is multiplied with the weight coefficient that neighbor pixel distributes, and the value that described central pixel point gray value is multiplied with the weight coefficient that neighbor pixel distributes;
f ( x , y ) = [ 1 6 * &Sigma; n = 0 3 [ ~ buffer [ 2 n ] ] + 1 12 * &Sigma; n = 0 3 [ ~ buffer [ 2 n + 1 ] ] ] * f ( x , y ) + 1 6 * &Sigma; n = 0 3 [ buffer [ 2 n ] * f ( 2 n ) ] + 1 12 * &Sigma; n = 0 3 [ buffer [ 2 n + 1 ] * f ( 2 n + 1 ) ]
Wherein, n ∈ Z, n=0,1,2,3; The scene degree of correlation that buffer [] is neighbor pixel; The negate of the scene degree of correlation that ~ buffer [] is neighbor pixel; The gray value at pixel (x, y) place centered by f (x, y).
In described device, the first processing unit 1150 further specifically for: four iteration are carried out to described ID figure and surpass the disposal of gentle filter, surpass the depth map after the disposal of gentle filter through four iteration carry out horizontal and vertical twice up-sampling to described;
Obtain the depth value of each 1/4 definition pixel point, the depth value of described each 1/4 definition pixel point forms 1/4 depth of resolution figure;
Twice iteration is carried out to described 1/4 depth of resolution figure and surpasses the disposal of gentle filter, surpass the depth map after the disposal of gentle filter through twice iteration carry out horizontal and vertical twice up-sampling to described;
Obtain the depth value of each original resolution pixel, the depth value of described each original resolution pixel forms original resolution depth map;
An iteration is carried out to described original resolution depth map and surpasses the disposal of gentle filter, obtain source images depth map.
In described device, to be describedly specially for described each neighbor pixel coefficient that assigns weight: the weight coefficient of cumulative neighbor pixel and be 1.
By device disclosed in the application embodiment of the present invention, the corresponding scene degree of correlation is obtained in the different down-sampling stage, ID figure is extracted by motion vector cumulative sum, utilize the scene degree of correlation in different down-sampling stage to carry out iteration to ID figure and surpass smoothing processing, carry out up-sampling process, final generation source images depth map, improves the picture quality of depth map simultaneously, make depth map profile more clear, this method also allows the computing cost of whole process remain within zone of reasonableness simultaneously.
Professional should recognize further, in conjunction with unit and the algorithm steps of each example of embodiment disclosed herein description, can realize with electronic hardware, computer software or the combination of the two, in order to the interchangeability of hardware and software is clearly described, generally describe composition and the step of each example in the above description according to function.These functions perform with hardware or software mode actually, depend on application-specific and the design constraint of technical scheme.Professional and technical personnel can use distinct methods to realize described function to each specifically should being used for, but this realization should not think the scope exceeding the embodiment of the present invention.
The software module that the method described in conjunction with embodiment disclosed herein or the step of algorithm can use hardware, processor to perform, or the combination of the two is implemented.Software module can be placed in the storage medium of other form any known in random asccess memory (RAM), internal memory, read-only memory (ROM), electrically programmable ROM, electrically erasable ROM, register, hard disk, moveable magnetic disc, CD-ROM or technical field.
Above-described embodiment; the object of the embodiment of the present invention, technical scheme and beneficial effect are further described; be understood that; the foregoing is only the embodiment of the embodiment of the present invention; be not intended to limit the present invention the protection range of embodiment; within all spirit in the embodiment of the present invention and principle, any amendment made, equivalent replacement, improvement etc., within the protection range that all should be included in the embodiment of the present invention.

Claims (16)

1. a method for extraction and optimized image depth map, it is characterized in that, described method comprises:
Obtain the scene degree of correlation of each pixel in current source picture and described current source picture, described current source picture is current video successive frame sequence;
To the continuous down-sampling of described current source picture, obtain the scene degree of correlation of each pixel in current each down-sampling source images;
The pixel that each pixel in described current down-sampling source images is corresponding with down-sampling source images before carries out block matching motion Vector operation, obtains the motion vector value of each pixel in described current down-sampling source images;
The motion vector value of each pixel in cumulative described current down-sampling source images respectively, from motion vector cumulative sum, extract the ID value of each pixel, described ID value forms source images ID figure;
Utilize the scene degree of correlation of each pixel in the scene degree of correlation of each pixel in described source images and described each down-sampling source images to surpass the disposal of gentle filter and up-sampling process continuously to each pixel in described ID figure, obtain described source images depth map.
2. the method for extraction according to claim 1 and optimized image depth map, is characterized in that, in described acquisition current source picture and described current source picture, the scene degree of correlation of each pixel is specially:
Select arbitrary pixel as central pixel point, for described central pixel point neighbor piont mark;
Obtain the difference of the red R of described central pixel point, green G, blue B value and the red R of described neighbor pixel, green G, blue B component value, and described difference is taken absolute value;
By described absolute value compared with scene relevance threshold, if described absolute value is less than scene relevance threshold, then the scene degree of correlation of described neighbor pixel is set to 1 and is correlated with, otherwise, be set to 0 uncorrelated;
The scene degree of correlation of described neighbor pixel stored in a buffer, described buffer is specially
b u f f e r &lsqb; m &rsqb; = 1 , | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | R + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | G + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | B &le; A ; 0 , | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | R + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | G + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | B > A ; Wherein, k=0,1; 0≤m≤7; M ∈ Z; F (x, y), f (x ± k, y ± k) are the red R of pixel, green G, the value of blue B component; A is scene relevance threshold; The m position that buffer [m] is the degree of correlation; M is neighbor piont mark.
3. the method for extraction according to claim 1 and optimized image depth map, is characterized in that, described to the continuous down-sampling of described current source picture, and the scene degree of correlation obtaining each pixel in current each down-sampling source images is specially:
1/2 horizontal and vertical down-sampling is carried out to described current source picture, obtains the scene degree of correlation of each pixel in current 1/4 resolution source image and current 1/4 resolution source image;
1/2 horizontal and vertical down-sampling is carried out to described current 1/4 resolution source image, obtains the scene degree of correlation of each pixel in current 1/16 resolution source image and current 1/16 resolution source image.
4. the method for extraction according to claim 3 and optimized image depth map, it is characterized in that, the described pixel that each pixel in described current down-sampling source images is corresponding with down-sampling source images before carries out block matching motion Vector operation, and the motion vector value obtaining each pixel in described current down-sampling source images is specially:
The pixel that each pixel in described current 1/16 resolution source image is corresponding with described 1/16 resolution source image before carries out block matching motion Vector operation, obtains the motion vector value of each pixel in described current 1/16 resolution source image;
The motion vector value of each pixel in cumulative described current 1/16 resolution source image respectively.
5. the method for extraction according to claim 1 and optimized image depth map, is characterized in that, the described ID value extracting each pixel from motion vector cumulative sum is specially:
Obtain motion vector mould and the unit picture element displacement gray value of each pixel displacement, described unit picture element displacement gray value is w is the width of image;
Described motion vector mould is multiplied with described unit picture element displacement gray value, obtains the motion vector gray value of described each pixel;
The cumulative sum of the motion vector gray value of described each pixel is stored in degree of depth buffer;
If D new+ D acc< D total; Then D acc_depth(x, y)=D acc_depth(x, y)+D new_depth(x, y);
If D new+ D acc> D total; Then D a c c _ d e p t h ( x , y ) = D a c c _ d e p t h ( x , y ) * D t o t a l - D n e w D a c c ;
If D t o t a l - D n e w D a c c < 0 , Then D t o t a l - D n e w D a c c = 0 ;
If D t o t a l - D n e w D a c c > 1 , Then D t o t a l - D n e w D a c c = 1 ;
Wherein, described (x, y) is each pixel coordinate; D acc_depth(x, y) is the cumulative sum of motion vector gray value before each pixel; D new_depth(x, y) is each pixel current motion vector gray value; D totalfor the maximum of degree of depth buffer cumulative sum, D newfor the cumulative sum of the whole gray value of current depth figure; D accfor whole gray value cumulative sums of all depth maps before in degree of depth buffer.
6. the method for extraction according to claim 2 and optimized image depth map, it is characterized in that, describedly utilize the scene degree of correlation of each pixel in the scene degree of correlation of each pixel in described source images and described each down-sampling source images to surpass the disposal of gentle filter continuously to each pixel in described ID figure to be specially:
For described each neighbor pixel assigns weight coefficient, described weight coefficient is the tap coefficient of super smothing filtering;
Call in described buffer the scene degree of correlation of the described neighbor pixel stored;
If the scene degree of correlation of described neighbor pixel be 1 be correlated with, then the weight coefficient that neighbor pixel gray value and neighbor pixel distribute is multiplied;
If the scene degree of correlation of described neighbor pixel is 0 uncorrelated, then the weight coefficient that central pixel point gray value and neighbor pixel distribute is multiplied;
The value that cumulative described neighbor pixel gray value is multiplied with the weight coefficient that neighbor pixel distributes, and the value that described central pixel point gray value is multiplied with the weight coefficient that neighbor pixel distributes;
f ( x , y ) = &lsqb; 1 6 * &Sigma; n = 0 3 &lsqb; ~ b u f f e r &lsqb; 2 n &rsqb; &rsqb; + 1 12 * &Sigma; n = 0 3 &lsqb; ~ b u f f e r &lsqb; 2 n + 1 &rsqb; &rsqb; &rsqb; * f ( x , y ) + 1 6 * &Sigma; n = 0 3 * &lsqb; b u f f e r &lsqb; 2 n &rsqb; * f ( 2 n ) &rsqb; + 1 12 * &Sigma; n = 0 3 &lsqb; b u f f e r &lsqb; 2 n + 1 &rsqb; * f ( 2 n + 1 ) &rsqb;
Wherein, n ∈ Z, n=0,1,2,3; The scene degree of correlation that buffer [] is neighbor pixel; The negate of the scene degree of correlation that ~ buffer [] is neighbor pixel; The gray value at pixel (x, y) place centered by f (x, y).
7. the method for extraction according to claim 6 and optimized image depth map, is characterized in that, in described ID figure, each pixel carries out continuous up-sampling and is specially:
Four iteration being carried out to described ID figure and surpasses the disposal of gentle filter, carrying out horizontal and vertical twice up-sampling to surpassing the depth map after the disposal of gentle filter through described four iteration;
Obtain the depth value of each 1/4 definition pixel point, the depth value of described each 1/4 definition pixel point forms 1/4 depth of resolution figure;
Twice iteration being carried out to described 1/4 depth of resolution figure and surpasses the disposal of gentle filter, carrying out horizontal and vertical twice up-sampling to surpassing the depth map after the disposal of gentle filter through described twice iteration;
Obtain the depth value of each original resolution pixel, the depth value of described each original resolution pixel forms original resolution depth map;
An iteration is carried out to described original resolution depth map and surpasses the disposal of gentle filter, obtain source images depth map.
8. the method for extraction according to claim 6 and optimized image depth map, is characterized in that, is describedly specially for described each neighbor pixel coefficient that assigns weight: the weight coefficient of cumulative neighbor pixel and be 1.
9. a device for extraction and optimized image depth map, it is characterized in that, described device comprises:
First acquiring unit, for obtaining the scene degree of correlation of each pixel in current source picture and described current source picture, described current source picture is current video successive frame sequence;
Second acquisition unit, for the continuous down-sampling of described current source picture, obtains the scene degree of correlation of each pixel in current each down-sampling source images;
3rd acquiring unit, carries out block matching motion Vector operation for the pixel that each pixel in described current down-sampling source images is corresponding with down-sampling source images before, obtains the motion vector value of each pixel in described current down-sampling source images;
Computing unit, for the motion vector value of each pixel in the described current down-sampling source images that adds up respectively, extract the ID value of each pixel from motion vector cumulative sum, described ID value forms source images ID figure;
First processing unit, for utilizing the scene degree of correlation of each pixel in the scene degree of correlation of each pixel in described source images and described each down-sampling source images to surpass the disposal of gentle filter and up-sampling process continuously to each pixel in described ID figure, obtain described source images depth map.
10. the device of extraction according to claim 9 and optimized image depth map, is characterized in that, described first acquiring unit specifically for:
Select arbitrary pixel as central pixel point, will with described central pixel point neighbor piont mark;
Obtain the difference of the red R of described central pixel point, green G, blue B component value and the red R of each described neighbor pixel, green G, blue B component value, and described difference is taken absolute value;
By described absolute value compared with scene relevance threshold, if described absolute value is less than scene relevance threshold, then the scene degree of correlation of described neighbor pixel is set to 1 and is correlated with, otherwise, be set to 0 uncorrelated;
The scene degree of correlation of described neighbor pixel stored in a buffer, described buffer is specially
b u f f e r &lsqb; m &rsqb; = 1 , | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | R + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | G + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | B &le; A ; 0 , | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | R + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | G + | f ( x , y ) - f ( x &PlusMinus; k , y &PlusMinus; k ) | B > A ;
Wherein, k=0,1; 0≤m≤7; M ∈ Z; F (x, y), f (x ± k, y ± k) are the red R of pixel, green G, the value of blue B component; A is scene relevance threshold; The m position that buffer [m] is the degree of correlation; M is neighbor piont mark.
The device of 11. extractions according to claim 9 and optimized image depth map, is characterized in that, described second acquisition unit specifically for:
1/2 horizontal and vertical down-sampling is carried out to described current source picture, obtains the scene degree of correlation of each pixel in current 1/4 resolution source image and current 1/4 resolution source image;
1/2 horizontal and vertical down-sampling is carried out to described current 1/4 resolution source image, obtains the scene degree of correlation of each pixel in current 1/16 resolution source image and current 1/16 resolution source image.
The device of 12. extractions according to claim 11 and optimized image depth map, is characterized in that, described 3rd acquiring unit specifically for:
The pixel that each pixel in described current 1/16 resolution source image is corresponding with described 1/16 resolution source image before carries out block matching motion Vector operation, obtains the motion vector value of each pixel in described current 1/16 resolution source image;
The motion vector value of each pixel in cumulative described current 1/16 resolution source image respectively.
The device of 13. extractions according to claim 9 and optimized image depth map, is characterized in that, described computing unit specifically for:
Obtain motion vector mould and the unit picture element displacement gray value of each pixel displacement, described unit picture element displacement gray value is w is the width of image;
Described motion vector mould is multiplied with described unit picture element displacement gray value, obtains the motion vector gray value of described each pixel;
The cumulative sum of the motion vector gray value of described each pixel is stored in degree of depth buffer;
If D new+ D acc< D total; Then D acc_depth(x, y)=D acc_depth(x, y)+D new_depth(x, y);
If D new+ D acc> D total; Then D a c c _ d e p t h ( x , y ) = D a c c _ d e p t h ( x , y ) * D t o t a l - D n e w D a c c ;
If D t o t a l - D n e w D a c c < 0 , Then D t o t a l - D n e w D a c c = 0 ;
If D t o t a l - D n e w D a c c > 1 , Then D t o t a l - D n e w D a c c = 1 ;
Wherein, described (x, y) is each pixel coordinate; D acc_depth(x, y) is the cumulative sum of motion vector gray value before each pixel; D new_depth(x, y) is each pixel current motion vector gray value; D totalfor the maximum of degree of depth buffer cumulative sum, D newfor the cumulative sum of the whole gray value of current depth figure; D accfor whole gray value cumulative sums of all depth maps before in degree of depth buffer.
The device of 14. extractions according to claim 10 and optimized image depth map, is characterized in that, described first processing unit specifically for:
For described each neighbor pixel assigns weight coefficient, described weight coefficient is the tap coefficient of super smothing filtering;
Call in described buffer the scene degree of correlation of the described neighbor pixel stored;
If the scene degree of correlation of described neighbor pixel be 1 be correlated with, then the weight coefficient that neighbor pixel gray value and neighbor pixel distribute is multiplied;
If the scene degree of correlation of described neighbor pixel is 0 uncorrelated, then the weight coefficient that central pixel point gray value and neighbor pixel distribute is multiplied;
The value that cumulative described neighbor pixel gray value is multiplied with the weight coefficient that neighbor pixel distributes, and the value that described central pixel point gray value is multiplied with the weight coefficient that neighbor pixel distributes;
f ( x , y ) = &lsqb; 1 6 * &Sigma; n = 0 3 &lsqb; ~ b u f f e r &lsqb; 2 n &rsqb; &rsqb; + 1 12 * &Sigma; n = 0 3 &lsqb; ~ b u f f e r &lsqb; 2 n + 1 &rsqb; &rsqb; &rsqb; * f ( x , y ) + 1 6 * &Sigma; n = 0 3 * &lsqb; b u f f e r &lsqb; 2 n &rsqb; * f ( 2 n ) &rsqb; + 1 12 * &Sigma; n = 0 3 &lsqb; b u f f e r &lsqb; 2 n + 1 &rsqb; * f ( 2 n + 1 ) &rsqb;
Wherein, n ∈ Z, n=0,1,2,3; The scene degree of correlation that buffer [] is neighbor pixel; The negate of the scene degree of correlation that ~ buffer [] is neighbor pixel; The gray value at pixel (x, y) place centered by f (x, y).
The device of 15. extractions according to claim 14 and optimized image depth map, is characterized in that, described first processing unit further specifically for:
Four iteration being carried out to described ID figure and surpasses the disposal of gentle filter, carrying out horizontal and vertical twice up-sampling to surpassing the depth map after the disposal of gentle filter through described four iteration;
Obtain the depth value of each 1/4 definition pixel point, the depth value of described each 1/4 definition pixel point forms 1/4 depth of resolution figure;
Twice iteration being carried out to described 1/4 depth of resolution figure and surpasses the disposal of gentle filter, carrying out horizontal and vertical twice up-sampling to surpassing the depth map after the disposal of gentle filter through described twice iteration;
Obtain the depth value of each original resolution pixel, the depth value of described each original resolution pixel forms original resolution depth map;
An iteration is carried out to described original resolution depth map and surpasses the disposal of gentle filter, obtain source images depth map.
16. the device of extraction according to claim 15 and optimized image depth map, is characterized in that, to be describedly specially for described each neighbor pixel coefficient that assigns weight: the weight coefficient of cumulative neighbor pixel and be 1.
CN201280004184.8A 2012-05-08 2012-05-08 The method and apparatus of a kind of extraction and optimized image depth map Active CN103493482B (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2012/075187 WO2013166656A1 (en) 2012-05-08 2012-05-08 Method and device for extracting and optimizing depth map of image

Publications (2)

Publication Number Publication Date
CN103493482A CN103493482A (en) 2014-01-01
CN103493482B true CN103493482B (en) 2016-01-20

Family

ID=49550062

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201280004184.8A Active CN103493482B (en) 2012-05-08 2012-05-08 The method and apparatus of a kind of extraction and optimized image depth map

Country Status (2)

Country Link
CN (1) CN103493482B (en)
WO (1) WO2013166656A1 (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104978554B (en) * 2014-04-08 2019-02-05 联想(北京)有限公司 The processing method and electronic equipment of information
CN104318586B (en) * 2014-09-26 2017-04-26 燕山大学 Adaptive morphological filtering-based motion blur direction estimation method and device
CN104394399B (en) * 2014-10-31 2016-08-24 天津大学 Three limit filtering methods of deep video coding
CN105721852B (en) * 2014-11-24 2018-12-14 奥多比公司 For determining the method, storage equipment and system of the capture instruction of depth refined image
TWI672677B (en) * 2017-03-31 2019-09-21 鈺立微電子股份有限公司 Depth map generation device for merging multiple depth maps
CN107204011A (en) * 2017-06-23 2017-09-26 万维云视(上海)数码科技有限公司 A kind of depth drawing generating method and device
CN110049242B (en) * 2019-04-18 2021-08-24 腾讯科技(深圳)有限公司 Image processing method and device
CN114943793B (en) * 2021-02-10 2025-03-18 北京字跳网络技术有限公司 Fluid rendering method, device, electronic device and storage medium
CN114240751A (en) * 2021-12-16 2022-03-25 海宁奕斯伟集成电路设计有限公司 Image processing apparatus, method, and program

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101141648A (en) * 2007-09-20 2008-03-12 上海广电(集团)有限公司中央研究院 Column diagram based weight predicting method
CN101582171A (en) * 2009-06-10 2009-11-18 清华大学 A method and device for creating a depth map
CN101754040A (en) * 2008-12-04 2010-06-23 三星电子株式会社 Method and appratus for estimating depth, and method and apparatus for converting 2d video to 3d video
WO2010123909A1 (en) * 2009-04-20 2010-10-28 Dolby Laboratories Licensing Corporation Directed interpolation and data post-processing
CN101945288A (en) * 2010-10-19 2011-01-12 浙江理工大学 H.264 compressed domain-based image depth map generation method
CN101951511A (en) * 2010-08-19 2011-01-19 深圳市亮信科技有限公司 Method for layering video scenes by analyzing depth
CN101969564A (en) * 2010-10-29 2011-02-09 清华大学 Upsampling method for depth video compression of three-dimensional television
CN102098527A (en) * 2011-01-28 2011-06-15 清华大学 Method and device for transforming two dimensions into three dimensions based on motion analysis

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101141648A (en) * 2007-09-20 2008-03-12 上海广电(集团)有限公司中央研究院 Column diagram based weight predicting method
CN101754040A (en) * 2008-12-04 2010-06-23 三星电子株式会社 Method and appratus for estimating depth, and method and apparatus for converting 2d video to 3d video
WO2010123909A1 (en) * 2009-04-20 2010-10-28 Dolby Laboratories Licensing Corporation Directed interpolation and data post-processing
CN101582171A (en) * 2009-06-10 2009-11-18 清华大学 A method and device for creating a depth map
CN101951511A (en) * 2010-08-19 2011-01-19 深圳市亮信科技有限公司 Method for layering video scenes by analyzing depth
CN101945288A (en) * 2010-10-19 2011-01-12 浙江理工大学 H.264 compressed domain-based image depth map generation method
CN101969564A (en) * 2010-10-29 2011-02-09 清华大学 Upsampling method for depth video compression of three-dimensional television
CN102098527A (en) * 2011-01-28 2011-06-15 清华大学 Method and device for transforming two dimensions into three dimensions based on motion analysis

Also Published As

Publication number Publication date
WO2013166656A1 (en) 2013-11-14
CN103493482A (en) 2014-01-01

Similar Documents

Publication Publication Date Title
CN103493482B (en) The method and apparatus of a kind of extraction and optimized image depth map
CN101287143B (en) Method of converting planar video to stereoscopic video based on real-time man-machine dialogue
CN101635859B (en) A method and device for converting flat video to stereoscopic video
CN102254348B (en) Virtual viewpoint mapping method based o adaptive disparity estimation
CN102625127B (en) Optimization method suitable for virtual viewpoint generation of 3D television
CN111325693B (en) A Large-scale Panoramic Viewpoint Synthesis Method Based on Single Viewpoint RGB-D Image
CN102665086B (en) Method for obtaining parallax by using region-based local stereo matching
US12283029B2 (en) Image inpainting method and electronic device
CN103248906B (en) Method and system for acquiring depth map of binocular stereo video sequence
Pearson et al. Plenoptic layer-based modeling for image based rendering
CN104954780A (en) DIBR (depth image-based rendering) virtual image restoration method applicable to high-definition 2D/3D (two-dimensional/three-dimensional) conversion
CN109509163B (en) A method and system for multi-focus image fusion based on FGF
Eder et al. Mapped convolutions
CN107958481A (en) A kind of three-dimensional rebuilding method and device
CN101557534B (en) Method for generating disparity map from video close frames
CN109118434A (en) A kind of image pre-processing method
CN106127691A (en) Panoramic picture mapping method
CN102957936B (en) Virtual viewpoint generation method from video single viewpoint to multiple viewpoints
CN106028020B (en) A kind of virtual perspective image cavity complementing method based on multi-direction prediction
CN102761764B (en) Upper sampling method used for depth picture of three-dimensional stereo video
CN102075777B (en) Method for converting planar video image into three-dimensional video image based on moving object
CN103559701A (en) Two-dimensional single-view image depth estimation method based on DCT coefficient entropy
CN102447932B (en) Reconstruction method of view point of free view point video
CN102750681A (en) Processing device and method for sharpening edge of image
CN108198140A (en) Three-dimensional collaboration filtering and noise reduction method based on NCSR models

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
C41 Transfer of patent application or patent right or utility model
TR01 Transfer of patent right

Effective date of registration: 20170222

Address after: 266100 Zhuzhou Road, Laoshan District, Shandong, No. 151, No.

Patentee after: QINGDAO HISENSE ELECTRONICS Co.,Ltd.

Address before: 266100 Zhuzhou Road, Laoshan District, Shandong, No. 151, No.

Patentee before: HISENSE HIVIEW TECH Co.,Ltd.

CP01 Change in the name or title of a patent holder
CP01 Change in the name or title of a patent holder

Address after: 266100 No. 151, Zhuzhou Road, Laoshan District, Shandong, Qingdao

Patentee after: Hisense Visual Technology Co., Ltd.

Address before: 266100 No. 151, Zhuzhou Road, Laoshan District, Shandong, Qingdao

Patentee before: QINGDAO HISENSE ELECTRONICS Co.,Ltd.