[go: up one dir, main page]

CN109978766A - Image magnification method and image amplifying device - Google Patents

Image magnification method and image amplifying device Download PDF

Info

Publication number
CN109978766A
CN109978766A CN201910185936.9A CN201910185936A CN109978766A CN 109978766 A CN109978766 A CN 109978766A CN 201910185936 A CN201910185936 A CN 201910185936A CN 109978766 A CN109978766 A CN 109978766A
Authority
CN
China
Prior art keywords
image
pixel
training
resolution
interpolation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910185936.9A
Other languages
Chinese (zh)
Other versions
CN109978766B (en
Inventor
朱江
赵斌
周明忠
吴宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
TCL China Star Optoelectronics Technology Co Ltd
Original Assignee
Shenzhen China Star Optoelectronics Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen China Star Optoelectronics Technology Co Ltd filed Critical Shenzhen China Star Optoelectronics Technology Co Ltd
Priority to CN201910185936.9A priority Critical patent/CN109978766B/en
Priority to PCT/CN2019/085764 priority patent/WO2020181641A1/en
Publication of CN109978766A publication Critical patent/CN109978766A/en
Application granted granted Critical
Publication of CN109978766B publication Critical patent/CN109978766B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4007Scaling of whole images or parts thereof, e.g. expanding or contracting based on interpolation, e.g. bilinear interpolation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/403Edge-driven scaling; Edge-based scaling

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)
  • Editing Of Facsimile Originals (AREA)

Abstract

The present invention provides a kind of image magnification method and image amplifying device.The image magnification method includes the following steps: to obtain the original image with first resolution;Interpolation amplification is carried out to original image by preset first interpolation algorithm, obtains the First Transition image with second resolution, second resolution is greater than first resolution;Interpolation amplification is carried out to original image by preset second interpolation algorithm, and the image after interpolation amplification is smoothed, obtains second transfer image acquisition with second resolution;Edge detection is carried out to original image, obtains the marginal information of original image;Weight output model is established, and the marginal information of original image is inputted into weight output model, generates the blending weight of target image;According to blending weight and preset fusion formula fusion First Transition image and the second transfer image acquisition, the target image with second resolution is obtained, image border smooth transition is can be realized, promotes image amplification effect, reduce image amplification cost.

Description

Image magnification method and image amplifying device
Technical field
The present invention relates to field of display technology more particularly to a kind of image magnification methods and image amplifying device.
Background technique
With computer technology, the high speed development of modern communications technology enters today of information age, people in human society Demand to image information it is also more more and more urgent.Current image digitazation input equipment is all by micro- on sampled images Zonule generates corresponding pixel, to form a dot matrixed image data, i.e., for fixed image input condition For fixed image, retrievable data volume is relatively-stationary.
And with the development of technology with the variation of the market demand, requirement of the consumer to the display quality of display device be more next Higher, correspondingly, the resolution ratio of display device is also higher and higher, the video and demand signals of high-res are also increasing.But It is still to have many video file and signal source resolution ratio lower at present, the video file and signal source of these low resolution When high resolution display part is shown, need by enhanced processing.Since the performance of image amplification directly determines view The quality of frequency display, therefore video display system improves the vision body of user there is an urgent need to the image magnification method of high quality It tests.
Currently, common image magnification method is generally interpolation amplification, typical interpolation amplification method includes closest inserts Value, bilinear interpolation, bicubic interpolation and polynomial interopolation etc., wherein closest interpolation algorithm is most simple, but closest interpolation It is discontinuous that algorithm is also most also easy to produce pixel value, and so as to cause blocking artifact, it is fuzzy in turn result in image, picture quality effect after amplification It is general not ideal enough.Bilinear interpolation algorithm is complex, and bilinear interpolation algorithm is not in the discontinuous situation of pixel value, Amplified picture quality is higher, but the edge contour of each theme and detail section in image can be made to become to a certain extent It must obscure, and bicubic interpolation and the algorithm of polynomial interopolation are more complicated.
Further, in the prior art, usually the flat site of image and fringe region are carried out using identical algorithm Interpolation amplification often makes in the image finally presented that there are apparent sawtooth when selecting better simply algorithm to carry out operation Or distortion;When selecting more complex algorithm to carry out operation, although being avoided that the distortion of image, entire calculating process is consumed The time taken is longer, and the requirement to hardware is higher, cannot be considered in terms of image amplification effect and amplification cost.
Summary of the invention
The purpose of the present invention is to provide a kind of image magnification methods, can be realized image border smooth transition, promote figure As amplification effect, image amplification cost is reduced.
The object of the invention is also to provide a kind of image amplifying devices, can be realized image border smooth transition, are promoted Image amplification effect reduces image amplification cost.
To achieve the above object, the present invention provides a kind of image magnification method, include the following steps:
Step S1, the original image with first resolution is obtained;
Step S2, interpolation amplification is carried out to the original image by preset first interpolation algorithm, obtains that there is second point The First Transition image of resolution, the second resolution are greater than first resolution;
Step S3, by preset second interpolation algorithm to the original image carry out interpolation amplification, and to interpolation amplification after Image be smoothed, obtain second transfer image acquisition with second resolution;
Step S4, edge detection is carried out to the original image, obtains the marginal information of the original image;
Step S5, weight output model is established, and the marginal information of original image is inputted into weight output model, generates target The blending weight of image;
Step S6, the First Transition image and the second transition figure are merged according to blending weight and preset fusion formula Picture obtains the target image with second resolution.
First interpolation algorithm is closest interpolation, bilinear interpolation, bicubic interpolation or polynomial interopolation algorithm, institute Stating the second interpolation algorithm is closest interpolation algorithm;
In the step S3 mode of smoothing processing be using preset smoothing operator to interpolation amplification in step S3 after Image carries out convolution;
Wherein, the smoothing operator is any of matrix 1 to matrix 5:
The original image includes multiple original pixel of array arrangement, and the First Transition image includes the multiple of array arrangement First pixel, second transfer image acquisition include multiple second pixels of array arrangement, and the target image includes array arrangement Multiple object pixels;
In the step S4, the marginal information of the original image includes the edge letter of each original pixel in the original image Breath;
In the step S5, the corresponding marginal information of each original pixel is inputted into weight output model, is generated and the preimage The blending weight of the corresponding object pixel in position of element;
In the step S6, the preset fusion formula are as follows:
Vp=(1- λ) × Vcb+ λ × Vs;
Wherein, the Vp is the gray value of object pixel, and Vcb is the first pixel corresponding with the position of the object pixel Gray value, Vs be the second pixel corresponding with the position of the object pixel gray value, λ be the object pixel fusion weigh Value, 0≤λ≤1.
The step of establishing weight output model in the step S5 specifically includes: obtaining a plurality of training data, and according to institute It states a plurality of training data and the weight output model is generated by machine learning training;
Wherein, the method for obtaining a plurality of training data are as follows:
The training image with first resolution is provided, the training image includes multiple trained pixels of array arrangement;
Edge detection is carried out to the training image, obtains the marginal information of each trained pixel;
Interpolation amplification is carried out to the training image by preset first interpolation algorithm, obtains that there is second resolution First Transition training image;
Interpolation amplification is carried out to the training image by preset second interpolation algorithm, and to the image after interpolation amplification It is smoothed, obtains the second lead-in training image with second resolution;
Multiple and different blending weights is chosen, merges institute according to the fusion formula and the plurality of different blending weight First Transition training image and the second lead-in training image are stated, generates multiple training objective images with second resolution, often One training objective image includes multiple training objective pixels of array arrangement;
There is provided the training image the corresponding standard target image with second resolution, the standard target image includes Multiple standard target pixels of array arrangement;
Determine that the sum of the grayscale values of each standard target pixel is in each training of same position with the standard target pixel The smallest training objective pixel of the difference of the gray value of object pixel;
To generate the blending weight of the smallest training objective pixel of the difference as the standard target pixel pair with the position The standard fusion weight answered;
The a plurality of training data for respectively corresponding each standard target pixel is formed, each training data includes a standard target The marginal information of the corresponding standard fusion weight of pixel and trained pixel corresponding with the standard target pixel.
The step S5 further include: the target image is divided into multiple regions, calculates each mesh in each region Mark pixel blending weight mean value, and using the mean value as the region in each object pixel blending weight.
The present invention also provides a kind of image amplifying devices, comprising: acquiring unit, first to be connected with the acquiring unit are put Big unit, the second amplifying unit being connected with the acquiring unit, the edge detection unit being connected with the acquiring unit and institute The connected weight of edge detection unit is stated to generate unit and generate with first amplifying unit, the second amplifying unit and weight The connected integrated unit of unit;
The acquiring unit is used to obtain the original image with first resolution;
First amplifying unit is used to carry out interpolation amplification to the original image by preset first interpolation algorithm, obtains To the First Transition image with second resolution, the second resolution is greater than first resolution;
Second amplifying unit is used to carry out interpolation amplification to the original image by preset second interpolation algorithm, and Image after interpolation amplification is smoothed, second transfer image acquisition with second resolution is obtained;
The edge detection unit is used to carry out edge detection to the original image, generates the edge letter of the original image Breath;
The weight generates unit and exports for establishing weight output model, and by the marginal information of original image input weight Model generates the blending weight of target image;
The integrated unit is used to merge the First Transition image and the according to blending weight and preset fusion formula Two transfer image acquisitions obtain the target image with second resolution.
First interpolation algorithm is closest interpolation, bilinear interpolation, bicubic interpolation or polynomial interopolation, described the Two interpolation algorithms are closest interpolation;
The mode that second amplifying unit is smoothed is using preset smoothing operator to by the second amplification Image after unit interpolation amplification carries out convolution;
Wherein, the smoothing operator is any of matrix 1 to matrix 5:
The original image includes multiple original pixel of array arrangement, and the First Transition image includes the multiple of array arrangement First pixel, second transfer image acquisition include multiple second pixels of array arrangement, and the target image includes array arrangement Multiple object pixels;
The marginal information for the original image that the edge detection unit generates specifically includes each original in the original image The marginal information of pixel;
The weight generates unit and the corresponding marginal information of each original pixel is inputted weight output model, generates and the original The blending weight of the corresponding object pixel in the position of pixel;
Preset fusion formula in the integrated unit are as follows:
Vp=(1- λ) × Vcb+ λ × Vs;
Wherein, the Vp is the gray value of object pixel, and Vcb is the first pixel corresponding with the position of the object pixel Gray value, Vs be the second pixel corresponding with the position of the object pixel gray value, λ be the object pixel fusion weigh Value, 0≤λ≤1.
The weight generates unit by obtaining a plurality of training data, and passes through engineering according to a plurality of training data It practises training and generates the weight output model;
Wherein, a plurality of training data of acquisition specifically includes:
The training image with first resolution is provided, the training image includes multiple trained pixels of array arrangement;
Edge detection is carried out to the training image, obtains the marginal information of each trained pixel;
Interpolation amplification is carried out to the training image by preset first interpolation algorithm, obtains that there is second resolution First Transition training image;
Interpolation amplification is carried out to the training image by preset second interpolation algorithm, and to the image after interpolation amplification It is smoothed, obtains the second lead-in training image with second resolution;
Multiple and different blending weights is chosen, merges institute according to the fusion formula and the plurality of different blending weight First Transition training image and the second lead-in training image are stated, generates multiple training objective images with second resolution, often One training objective image includes multiple training objective pixels of array arrangement;
There is provided the training image the corresponding standard target image with second resolution, the standard target image includes Multiple standard target pixels of array arrangement;
Determine that the sum of the grayscale values of each standard target pixel is in each training of same position with the standard target pixel The smallest training objective pixel of the difference of the gray value of object pixel;
To generate the blending weight of the smallest training objective pixel of the difference as the standard target pixel pair with the position The standard fusion weight answered;
The a plurality of training data for respectively corresponding each standard target pixel is formed, each training data includes a standard target The marginal information of the corresponding standard fusion weight of pixel and trained pixel corresponding with the standard target pixel.
The weight generates unit and is also used to the target image being divided into multiple regions, and calculates in each region The mean value of the blending weight of each object pixel, and using the mean value as the region in each object pixel blending weight.
Beneficial effects of the present invention: the present invention provides a kind of image magnification method.Described image amplification method includes as follows Step: the original image with first resolution is obtained;Interpolation is carried out to the original image by preset first interpolation algorithm to put Greatly, the First Transition image with second resolution is obtained, the second resolution is greater than first resolution;Pass through preset Two interpolation algorithms carry out interpolation amplification to the original image, and are smoothed to the image after interpolation amplification, are had Second transfer image acquisition of second resolution;Edge detection is carried out to the original image, obtains the marginal information of the original image;It builds Write value output model, and the marginal information of original image is inputted into weight output model, generate the blending weight of target image;Root The First Transition image and the second transfer image acquisition are merged according to blending weight and preset fusion formula, obtains having second to differentiate The target image of rate can be realized image border smooth transition, promote image amplification effect, reduces image amplification cost.This hair It is bright that a kind of image amplifying device is also provided, it can be realized image border smooth transition, promote image amplification effect, reduce image and put Big cost.
Detailed description of the invention
For further understanding of the features and technical contents of the present invention, it please refers to below in connection with of the invention detailed Illustrate and attached drawing, however, the drawings only provide reference and explanation, is not intended to limit the present invention.
In attached drawing,
Fig. 1 is the flow chart of image magnification method of the invention;
Fig. 2 is the schematic diagram of image amplifying device of the invention.
Specific embodiment
Further to illustrate technological means and its effect adopted by the present invention, below in conjunction with preferred implementation of the invention Example and its attached drawing are described in detail.
Referring to Fig. 1, the present invention provides a kind of image magnification method, include the following steps:
Step S1, the original image with first resolution is obtained.
Specifically, the original image includes multiple original pixel of array arrangement.
Step S2, interpolation amplification is carried out to the original image by preset first interpolation algorithm, obtains that there is second point The First Transition image of resolution, the second resolution are greater than first resolution.
Specifically, first interpolation algorithm is closest interpolation, bilinear interpolation, bicubic interpolation or polynomial interopolation Algorithm.
Further, the First Transition image includes multiple first pixels of array arrangement.
Step S3, by preset second interpolation algorithm to the original image carry out interpolation amplification, and to interpolation amplification after Image be smoothed, obtain second transfer image acquisition with second resolution.
Specifically, second transfer image acquisition includes multiple second pixels of array arrangement.
Further, second interpolation algorithm is closest interpolation algorithm.
Specifically, the mode of the smoothing processing is using preset smoothing operator to the figure after interpolation amplification in step S3 As carrying out convolution;
Wherein, the smoothing operator is any of matrix 1 to matrix 5:
Step S4, edge detection is carried out to the original image, obtains the marginal information of the original image.
Specifically, edge detection is carried out to the original image by Sobel (Sobel) operator in the step S4.
Specifically, in the step S4, the marginal information of the original image includes each original pixel in the original image Marginal information.
Step S5, weight output model is established, and the marginal information of original image is inputted into weight output model, generates target The blending weight of image.
Specifically, the target image includes multiple object pixels of array arrangement.
Specifically, weight output model is established by machine learning in the step S5.
Further, the step of establishing weight output model in the step S5 specifically includes: a plurality of training data is obtained, And the weight output model is generated by machine learning training according to a plurality of training data, the training data can be anti- Being associated with for the marginal information of original image and the blending weight of target image is reflected, described in generating by machine learning training Weight output model.
Specifically, the method for obtaining a plurality of training data are as follows:
The training image with first resolution is provided, the training image includes multiple trained pixels of array arrangement;
Edge detection is carried out to the training image, obtains the marginal information of each trained pixel;
Interpolation amplification is carried out to the training image by preset first interpolation algorithm, obtains that there is second resolution First Transition training image;
Interpolation amplification is carried out to the training image by preset second interpolation algorithm, and to the image after interpolation amplification It is smoothed, obtains the second lead-in training image with second resolution;
Multiple and different blending weights is chosen, merges institute according to the fusion formula and the plurality of different blending weight First Transition training image and the second lead-in training image are stated, generates multiple training objective images with second resolution, often One training objective image includes multiple training objective pixels of array arrangement;
There is provided the training image the corresponding standard target image with second resolution, the standard target image includes Multiple standard target pixels of array arrangement;
Determine that the sum of the grayscale values of each standard target pixel is in each training of same position with the standard target pixel The smallest training objective pixel of the difference of the gray value of object pixel;
To generate the blending weight of the smallest training objective pixel of the difference as the standard target pixel pair with the position The standard fusion weight answered;
The a plurality of training data for respectively corresponding each standard target pixel is formed, each training data includes a standard target The marginal information of the corresponding standard fusion weight of pixel and trained pixel corresponding with the standard target pixel.
Specifically, in the step S5, the corresponding marginal information of each original pixel is inputted into weight output model, generate with The blending weight of the corresponding object pixel in the position of the original pixel.
Step S6, the First Transition image and the second transition figure are merged according to blending weight and preset fusion formula Picture obtains the target image with second resolution.
Specifically, in the step S6, the preset fusion formula are as follows:
Vp=(1- λ) × Vcb+ λ × Vs;
Wherein, the Vp is the gray value of object pixel, and Vcb is the first pixel corresponding with the position of the object pixel Gray value, Vs be the second pixel corresponding with the position of the object pixel gray value, λ be the object pixel fusion weigh Value, 0≤λ≤1.
Specifically, the original pixel, the first pixel, the second pixel and object pixel include red component, green component And blue component, respectively to the ash of the red component of original pixel, blue component and green component in the step S2 and step S3 Angle value is handled to obtain the gray value of first pixel and the second pixel red component, blue component and green component, and In the step S6, the gray scale of first pixel and the second pixel red component, blue component and green component is merged respectively Value obtains the gray value of the red component of object pixel, green component and blue component.
Further, image transition is gentler, in the step S5 the blending weight for obtaining each object pixel it Afterwards, include the steps that asking equal, the target image is specially divided into multiple regions, calculate each in each region The mean value of the blending weight of object pixel, and using the mean value as the region in each object pixel blending weight, preferably Ground, each region include 3 × 3 pixels.
It should be noted that of the invention amplifies original image generation First Transition image and second by two methods respectively Transfer image acquisition, wherein First Transition image is more preferable for flat site amplification effect, and the second transfer image acquisition is directed to marginal zone The amplification effect in domain is more preferable, then establishes a weight output model by machine learning algorithm, and output one and the edge of original image are believed Relevant blending weight is ceased, the First Transition image and the second transfer image acquisition are merged, when object pixel is biased to fringe region, The accounting of the second transfer image acquisition is larger when then First Transition image is merged with the second transfer image acquisition, when object pixel is biased to flat region When domain, then the accounting of First Transition image is larger when First Transition image is merged with the second transfer image acquisition, can be realized image side Edge smooth transition, promotes image amplification effect, reduces image amplification cost, conceptual design is simple, is easy to be fabricated to corresponding core Piece, and cost is relatively low, the blending weight accuracy exported by the weight output model that machine learning is established is higher.
Referring to Fig. 2, the present invention also provides a kind of image amplifying devices, comprising: acquiring unit 10 and the acquiring unit 10 the first connected amplifying units 20, the second amplifying unit 30 being connected with the acquiring unit 10 and the acquiring unit 10 Connected edge detection unit 40, the weight being connected with the edge detection unit 40 generate unit 50 and put with described first Big unit 20, the second amplifying unit 30 and weight generate the connected integrated unit 60 of unit 50;
The acquiring unit 10 is used to obtain the original image with first resolution;
First amplifying unit 20 is used to carry out interpolation amplification to the original image by preset first interpolation algorithm, The First Transition image with second resolution is obtained, the second resolution is greater than first resolution;
Second amplifying unit 30 is used to carry out interpolation amplification to the original image by preset second interpolation algorithm, And the image after interpolation amplification is smoothed, obtain second transfer image acquisition with second resolution;
The edge detection unit 40 is used to carry out edge detection to the original image, generates the edge letter of the original image Breath;
The weight generates unit 50 for establishing weight output model, and the marginal information of original image input weight is defeated Model out generates the blending weight of target image;
The integrated unit 60 be used to be merged according to blending weight and preset fusion formula the First Transition image and Second transfer image acquisition obtains the target image with second resolution.
Specifically, first interpolation algorithm is that closest interpolation, bilinear interpolation, bicubic interpolation or multinomial are inserted Value, second interpolation algorithm are closest interpolation;
The mode that second amplifying unit 30 is smoothed is to be put using preset smoothing operator to by second Image after 30 interpolation amplification of big unit carries out convolution;
Wherein, the smoothing operator is any of matrix 1 to matrix 5:
Specifically, the edge detection unit 40 carries out edge inspection to the original image by Sobel (Sobel) operator It surveys.
Specifically, the original image includes multiple original pixel of array arrangement, and the First Transition image includes array row Multiple first pixels of cloth, second transfer image acquisition include multiple second pixels of array arrangement, and the target image includes Multiple object pixels of array arrangement;
The marginal information for the original image that the edge detection unit 40 generates specifically includes each in the original image The marginal information of original pixel;
The weight generates unit 50 and the corresponding marginal information of each original pixel is inputted weight output model, generates and is somebody's turn to do The blending weight of the corresponding object pixel in the position of original pixel;
Preset fusion formula in the integrated unit 60 are as follows:
Vp=1- λ × Vcb+ λ × Vs;
Wherein, the Vp is the gray value of object pixel, and Vcb is the first pixel corresponding with the position of the object pixel Gray value, Vs be the second pixel corresponding with the position of the object pixel gray value, λ be the object pixel fusion weigh Value, 0≤λ≤1.
Specifically, the weight generates unit 50 by obtaining a plurality of training data, and according to a plurality of training data The weight output model is generated by machine learning training;
Wherein, a plurality of training data of acquisition specifically includes:
The training image with first resolution is provided, the training image includes multiple trained pixels of array arrangement;
Edge detection is carried out to the training image, obtains the marginal information of each trained pixel;
Interpolation amplification is carried out to the training image by preset first interpolation algorithm, obtains that there is second resolution First Transition training image;
Interpolation amplification is carried out to the training image by preset second interpolation algorithm, and to the image after interpolation amplification It is smoothed, obtains the second lead-in training image with second resolution;
Multiple and different blending weights is chosen, merges institute according to the fusion formula and the plurality of different blending weight First Transition training image and the second lead-in training image are stated, generates multiple training objective images with second resolution, often One training objective image includes multiple training objective pixels of array arrangement;
There is provided the training image the corresponding standard target image with second resolution, the standard target image includes Multiple standard target pixels of array arrangement;
Determine that the sum of the grayscale values of each standard target pixel is in each training of same position with the standard target pixel The smallest training objective pixel of the difference of the gray value of object pixel;
To generate the blending weight of the smallest training objective pixel of the difference as the standard target pixel pair with the position The standard fusion weight answered;
The a plurality of training data for respectively corresponding each standard target pixel is formed, each training data includes a standard target The marginal information of the corresponding standard fusion weight of pixel and trained pixel corresponding with the standard target pixel.
Specifically, the original pixel, the first pixel, the second pixel and object pixel include red component, green component And blue component, respectively to the red component of original pixel, blue point in first amplifying unit 20 and the second amplifying unit 30 The gray value of amount and green component is handled to obtain first pixel and the second pixel red component, blue component and green The gray value of component, and the integrated unit 60 merges first pixel and the second pixel red component, blue component respectively And the gray value of green component obtains the gray value of the red component of object pixel, green component and blue component.
Further, image transition is gentler, in the step S5 the blending weight for obtaining each object pixel it Afterwards, include the steps that asking equal, the target image is specially divided into multiple regions, calculate each in each region The mean value of the blending weight of object pixel, and using the mean value as the region in each object pixel blending weight, preferably Ground, each region include 3 × 3 pixels.
It should be noted that of the invention amplifies original image generation First Transition image and second by two methods respectively Transfer image acquisition, wherein First Transition image is more preferable for flat site amplification effect, and the second transfer image acquisition is directed to marginal zone The amplification effect in domain is more preferable, then establishes a weight output model by machine learning algorithm, and output one and the edge of original image are believed Relevant blending weight is ceased, the First Transition image and the second transfer image acquisition are merged, when object pixel is biased to fringe region, The accounting of the second transfer image acquisition is larger when then First Transition image is merged with the second transfer image acquisition, when object pixel is biased to flat region When domain, then the accounting of First Transition image is larger when First Transition image is merged with the second transfer image acquisition, can be realized image side Edge smooth transition, promotes image amplification effect, reduces image amplification cost, conceptual design is simple, is easy to be fabricated to corresponding core Piece, and cost is relatively low, the blending weight accuracy exported by the weight output model that machine learning is established is higher.
In conclusion the present invention provides a kind of image magnification method.Described image amplification method includes the following steps: to obtain Original image with first resolution;Interpolation amplification is carried out to the original image by preset first interpolation algorithm, is had There is the First Transition image of second resolution, the second resolution is greater than first resolution;It is calculated by preset second interpolation Method carries out interpolation amplification to the original image, and is smoothed to the image after interpolation amplification, obtains having second to differentiate Second transfer image acquisition of rate;Edge detection is carried out to the original image, obtains the marginal information of the original image;It is defeated to establish weight Model out, and the marginal information of original image is inputted into weight output model, generate the blending weight of target image;It is weighed according to fusion Value and preset fusion formula merge the First Transition image and the second transfer image acquisition, obtain the target with second resolution Image can be realized image border smooth transition, promote image amplification effect, reduces image amplification cost.The present invention also provides A kind of image amplifying device can be realized image border smooth transition, promote image amplification effect, reduces image amplification cost.
The above for those of ordinary skill in the art can according to the technique and scheme of the present invention and technology Other various corresponding changes and modifications are made in design, and all these change and modification all should belong to the claims in the present invention Protection scope.

Claims (10)

1. a kind of image magnification method, which comprises the steps of:
Step S1, the original image with first resolution is obtained;
Step S2, interpolation amplification is carried out to the original image by preset first interpolation algorithm, obtained with second resolution First Transition image, the second resolution be greater than first resolution;
Step S3, interpolation amplification is carried out to the original image by preset second interpolation algorithm, and to the figure after interpolation amplification As being smoothed, second transfer image acquisition with second resolution is obtained;
Step S4, edge detection is carried out to the original image, obtains the marginal information of the original image;
Step S5, weight output model is established, and the marginal information of original image is inputted into weight output model, generates target image Blending weight;
Step S6, the First Transition image and the second transfer image acquisition are merged according to blending weight and preset fusion formula, obtained To the target image with second resolution.
2. image magnification method as described in claim 1, which is characterized in that first interpolation algorithm be closest interpolation, Bilinear interpolation, bicubic interpolation or polynomial interopolation algorithm, second interpolation algorithm are closest interpolation algorithm;
The mode of smoothing processing is using preset smoothing operator to the image after interpolation amplification in step S3 in the step S3 Carry out convolution;
Wherein, the smoothing operator is any of matrix 1 to matrix 5:
3. image magnification method as described in claim 1, which is characterized in that the original image includes multiple originals of array arrangement Pixel, the First Transition image include multiple first pixels of array arrangement, and second transfer image acquisition includes array arrangement Multiple second pixels, the target image includes multiple object pixels of array arrangement;
In the step S4, the marginal information of the original image includes the marginal information of each original pixel in the original image;
In the step S5, the corresponding marginal information of each original pixel is inputted into weight output model, is generated and the original pixel The blending weight of the corresponding object pixel in position;
In the step S6, the preset fusion formula are as follows:
Vp=(1- λ) × Vcb+ λ × Vs;
Wherein, the Vp is the gray value of object pixel, and Vcb is the ash of the first pixel corresponding with the position of the object pixel Angle value, Vs be the second pixel corresponding with the position of the object pixel gray value, λ be the object pixel blending weight, 0 ≤λ≤1。
4. image magnification method as claimed in claim 3, which is characterized in that establish weight output model in the step S5 Step specifically includes: obtaining a plurality of training data, and is passed through described in machine learning training generation according to a plurality of training data Weight output model;
Wherein, the method for obtaining a plurality of training data are as follows:
The training image with first resolution is provided, the training image includes multiple trained pixels of array arrangement;
Edge detection is carried out to the training image, obtains the marginal information of each trained pixel;
Interpolation amplification is carried out to the training image by preset first interpolation algorithm, obtains first with second resolution Lead-in training image;
Interpolation amplification is carried out to the training image by preset second interpolation algorithm, and the image after interpolation amplification is carried out Smoothing processing obtains the second lead-in training image with second resolution;
Multiple and different blending weights is chosen, according to the fusion formula and the plurality of different blending weight fusion described the One lead-in training image and the second lead-in training image generate multiple training objective images with second resolution, Mei Yixun Practice multiple training objective pixels that target image includes array arrangement;
There is provided the training image the corresponding standard target image with second resolution, the standard target image includes array Multiple standard target pixels of arrangement;
Determine that the sum of the grayscale values of each standard target pixel is in each training objective of same position with the standard target pixel The smallest training objective pixel of the difference of the gray value of pixel;
To generate the blending weight of the smallest training objective pixel of the difference as corresponding with the standard target pixel of the position Standard fusion weight;
The a plurality of training data for respectively corresponding each standard target pixel is formed, each training data includes a standard target pixel Corresponding standard fusion weight and trained pixel corresponding with the standard target pixel marginal information.
5. image magnification method as claimed in claim 3, which is characterized in that the step S5 further include: by the target figure As being divided into multiple regions, calculate the mean value of the blending weight of each object pixel in each region, and using the mean value as The blending weight of each object pixel in the region.
6. a kind of image amplifying device characterized by comprising acquiring unit (10) is connected with the acquiring unit (10) First amplifying unit (20), the second amplifying unit (30) being connected with the acquiring unit (10), with the acquiring unit (10) Connected edge detection unit (40), the weight being connected with the edge detection unit (40) generate unit (50) and with it is described First amplifying unit (20), the second amplifying unit (30) and weight generate unit (50) connected integrated unit (60);
The acquiring unit (10) is used to obtain the original image with first resolution;
First amplifying unit (20) is used to carry out interpolation amplification to the original image by preset first interpolation algorithm, obtains To the First Transition image with second resolution, the second resolution is greater than first resolution;
Second amplifying unit (30) is used to carry out interpolation amplification to the original image by preset second interpolation algorithm, and Image after interpolation amplification is smoothed, second transfer image acquisition with second resolution is obtained;
The edge detection unit (40) is used to carry out edge detection to the original image, generates the edge letter of the original image Breath;
The weight generates unit (50) and exports for establishing weight output model, and by the marginal information of original image input weight Model generates the blending weight of target image;
The integrated unit (60) is used to merge the First Transition image and the according to blending weight and preset fusion formula Two transfer image acquisitions obtain the target image with second resolution.
7. image amplifying device as claimed in claim 6, which is characterized in that first interpolation algorithm be closest interpolation, Bilinear interpolation, bicubic interpolation or polynomial interopolation, second interpolation algorithm are closest interpolation;
The mode that second amplifying unit (30) is smoothed is using preset smoothing operator to by the second amplification Image after unit (30) interpolation amplification carries out convolution;
Wherein, the smoothing operator is any of matrix 1 to matrix 5:
8. image amplifying device as claimed in claim 6, which is characterized in that the original image includes multiple originals of array arrangement Pixel, the First Transition image include multiple first pixels of array arrangement, and second transfer image acquisition includes array arrangement Multiple second pixels, the target image includes multiple object pixels of array arrangement;
The marginal information for the original image that the edge detection unit (40) generates specifically includes each original in the original image The marginal information of pixel;
The weight generates unit (50) and the corresponding marginal information of each original pixel is inputted weight output model, generates and the original The blending weight of the corresponding object pixel in the position of pixel;
Preset fusion formula in the integrated unit (60) are as follows:
Vp=(1- λ) × Vcb+ λ × Vs;
Wherein, the Vp is the gray value of object pixel, and Vcb is the ash of the first pixel corresponding with the position of the object pixel Angle value, Vs be the second pixel corresponding with the position of the object pixel gray value, λ be the object pixel blending weight, 0 ≤λ≤1。
9. image amplifying device as claimed in claim 8, which is characterized in that it is more by obtaining that the weight generates unit (50) Training data, and machine learning training is passed through according to a plurality of training data and generates the weight output model;
Wherein, a plurality of training data of acquisition specifically includes:
The training image with first resolution is provided, the training image includes multiple trained pixels of array arrangement;
Edge detection is carried out to the training image, obtains the marginal information of each trained pixel;
Interpolation amplification is carried out to the training image by preset first interpolation algorithm, obtains first with second resolution Lead-in training image;
Interpolation amplification is carried out to the training image by preset second interpolation algorithm, and the image after interpolation amplification is carried out Smoothing processing obtains the second lead-in training image with second resolution;
Multiple and different blending weights is chosen, according to the fusion formula and the plurality of different blending weight fusion described the One lead-in training image and the second lead-in training image generate multiple training objective images with second resolution, Mei Yixun Practice multiple training objective pixels that target image includes array arrangement;
There is provided the training image the corresponding standard target image with second resolution, the standard target image includes array Multiple standard target pixels of arrangement;
Determine that the sum of the grayscale values of each standard target pixel is in each training objective of same position with the standard target pixel The smallest training objective pixel of the difference of the gray value of pixel;
To generate the blending weight of the smallest training objective pixel of the difference as corresponding with the standard target pixel of the position Standard fusion weight;
The a plurality of training data for respectively corresponding each standard target pixel is formed, each training data includes a standard target pixel Corresponding standard fusion weight and trained pixel corresponding with the standard target pixel marginal information.
10. image amplifying device as claimed in claim 8, which is characterized in that the weight generate unit (50) be also used to by The target image is divided into multiple regions, and calculates the mean value of the blending weight of each object pixel in each region, and The blending weight of each object pixel in using the mean value as the region.
CN201910185936.9A 2019-03-12 2019-03-12 Image amplifying method and image amplifying device Active CN109978766B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201910185936.9A CN109978766B (en) 2019-03-12 2019-03-12 Image amplifying method and image amplifying device
PCT/CN2019/085764 WO2020181641A1 (en) 2019-03-12 2019-05-07 Image magnification method and image magnification device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910185936.9A CN109978766B (en) 2019-03-12 2019-03-12 Image amplifying method and image amplifying device

Publications (2)

Publication Number Publication Date
CN109978766A true CN109978766A (en) 2019-07-05
CN109978766B CN109978766B (en) 2020-10-16

Family

ID=67078601

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910185936.9A Active CN109978766B (en) 2019-03-12 2019-03-12 Image amplifying method and image amplifying device

Country Status (2)

Country Link
CN (1) CN109978766B (en)
WO (1) WO2020181641A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112308808A (en) * 2019-07-31 2021-02-02 北京金山云网络技术有限公司 Image processing method and device and electronic equipment
CN112381714A (en) * 2020-10-30 2021-02-19 南阳柯丽尔科技有限公司 Image processing method, device, storage medium and equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050078883A1 (en) * 2003-10-08 2005-04-14 Yi Jong-Hyon Digital image processing device and method
CN101911112A (en) * 2007-12-25 2010-12-08 日本电气株式会社 Image processing device, image processing method, image decompressing device, image compressing device, image transmission system, and storage medium
CN102800069A (en) * 2012-05-22 2012-11-28 湖南大学 Image super-resolution method for combining soft decision self-adaptation interpolation and bicubic interpolation
CN102842111A (en) * 2012-07-09 2012-12-26 许丹 Enlarged image compensation method and device
EP1947603A3 (en) * 2007-01-22 2013-08-07 Sharp Kabushiki Kaisha Image upsampling technique
CN104299185A (en) * 2014-09-26 2015-01-21 京东方科技集团股份有限公司 Image magnification method, image magnification device and display device

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8180180B2 (en) * 2008-12-29 2012-05-15 Arcsoft Hangzhou Co., Ltd. Method for magnifying images and videos
CN106204454B (en) * 2016-01-26 2019-06-21 西北工业大学 High-precision and fast image interpolation method based on texture edge adaptive data fusion
CN106709875B (en) * 2016-12-30 2020-02-18 北京工业大学 A Compressed Low-Resolution Image Restoration Method Based on Joint Deep Network

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050078883A1 (en) * 2003-10-08 2005-04-14 Yi Jong-Hyon Digital image processing device and method
EP1947603A3 (en) * 2007-01-22 2013-08-07 Sharp Kabushiki Kaisha Image upsampling technique
CN101911112A (en) * 2007-12-25 2010-12-08 日本电气株式会社 Image processing device, image processing method, image decompressing device, image compressing device, image transmission system, and storage medium
CN102800069A (en) * 2012-05-22 2012-11-28 湖南大学 Image super-resolution method for combining soft decision self-adaptation interpolation and bicubic interpolation
CN102842111A (en) * 2012-07-09 2012-12-26 许丹 Enlarged image compensation method and device
CN104299185A (en) * 2014-09-26 2015-01-21 京东方科技集团股份有限公司 Image magnification method, image magnification device and display device

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
王会鹏等: "《一种基于区域的双三次图像插值算法》", 《计算机工程》 *
赵旦峰等: "《基于阈值判断的区域指导插值算法》", 《系统工程与电子技术》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112308808A (en) * 2019-07-31 2021-02-02 北京金山云网络技术有限公司 Image processing method and device and electronic equipment
CN112308808B (en) * 2019-07-31 2025-01-10 北京金山云网络技术有限公司 Image processing method, device and electronic equipment
CN112381714A (en) * 2020-10-30 2021-02-19 南阳柯丽尔科技有限公司 Image processing method, device, storage medium and equipment

Also Published As

Publication number Publication date
CN109978766B (en) 2020-10-16
WO2020181641A1 (en) 2020-09-17

Similar Documents

Publication Publication Date Title
US7791652B2 (en) Image processing apparatus, image capture apparatus, image output apparatus, and method and program for these apparatus
EP3076364B1 (en) Image filtering based on image gradients
US7792384B2 (en) Image processing apparatus, image processing method, program, and recording medium therefor
CN110264426B (en) Image distortion correction method and device
US10614551B2 (en) Image interpolation methods and related image interpolation devices thereof
US20130094781A1 (en) Image processing apparatus
US6813041B1 (en) Method and apparatus for performing local color correction
US10255665B2 (en) Image processing device and method, image capturing device, program, and record medium
JP4498361B2 (en) How to speed up Retinex-type algorithms
CN103942755A (en) Image brightness adjusting method and device
JP2006013558A (en) Image processing apparatus and image processing program
KR102337835B1 (en) Directional Scaling Systems and Methods
JP2004165840A (en) Image processing program
US7751642B1 (en) Methods and devices for image processing, image capturing and image downscaling
US20090022402A1 (en) Image-resolution-improvement apparatus and method
CN109978766A (en) Image magnification method and image amplifying device
TW201536029A (en) Image downsampling apparatus and method
US9928577B2 (en) Image correction apparatus and image correction method
JP4369030B2 (en) Image correction method and apparatus, and computer-readable recording medium storing image correction program
JP4992379B2 (en) Image gradation conversion apparatus, program, electronic camera, and method thereof
JP4147155B2 (en) Image processing apparatus and method
KR101585187B1 (en) Image Processing Method and Apparatus for Integrated Multi-scale Retinex Based on CIELAB Color Space for Preserving Color
TWI413019B (en) Image adjusting circuit and image adjusting method
JP4032200B2 (en) Image data interpolation method, image data interpolation device, and computer readable recording medium recording image data interpolation program
CN102447817B (en) Image processing device and space image noise eliminating method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CP01 Change in the name or title of a patent holder

Address after: 9-2 Tangming Avenue, Guangming New District, Shenzhen City, Guangdong Province

Patentee after: TCL China Star Optoelectronics Technology Co.,Ltd.

Address before: 9-2 Tangming Avenue, Guangming New District, Shenzhen City, Guangdong Province

Patentee before: Shenzhen China Star Optoelectronics Technology Co.,Ltd.

CP01 Change in the name or title of a patent holder