CN110177282A - A kind of inter-frame prediction method based on SRCNN - Google Patents
A kind of inter-frame prediction method based on SRCNN Download PDFInfo
- Publication number
- CN110177282A CN110177282A CN201910388829.6A CN201910388829A CN110177282A CN 110177282 A CN110177282 A CN 110177282A CN 201910388829 A CN201910388829 A CN 201910388829A CN 110177282 A CN110177282 A CN 110177282A
- Authority
- CN
- China
- Prior art keywords
- image
- frame
- resolution
- super
- image sequence
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 14
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 19
- 238000012549 training Methods 0.000 claims abstract description 17
- 238000012360 testing method Methods 0.000 claims description 10
- 238000012795 verification Methods 0.000 claims description 8
- 230000000694 effects Effects 0.000 claims description 5
- 238000013139 quantization Methods 0.000 claims description 5
- 238000013507 mapping Methods 0.000 claims description 4
- 238000005070 sampling Methods 0.000 claims 1
- 238000013528 artificial neural network Methods 0.000 abstract description 2
- 238000013135 deep learning Methods 0.000 abstract description 2
- 238000000605 extraction Methods 0.000 abstract description 2
- 230000005540 biological transmission Effects 0.000 description 7
- 230000006835 compression Effects 0.000 description 6
- 238000007906 compression Methods 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000007429 general method Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000005192 partition Methods 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/124—Quantisation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/42—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
技术领域technical field
本发明属于视频编码领域中的帧间预测,主要为了提高视频传输效率,具体涉及一种基于SRCNN的帧间预测方法。The invention belongs to inter-frame prediction in the field of video coding, mainly for improving video transmission efficiency, and specifically relates to an SRCNN-based inter-frame prediction method.
背景技术Background technique
超分辨率(Super-Resolution)意味着将一幅低分辨率(Low Resolution)图像转变成一幅高分辨率(HighResolution)图像,通常可以提高图像质量、清晰度。超分辨率卷积神经网络(Super-ResolutionConvolutionalNeuralNetwork,SRCNN)是一个应用于图像超分辨率重建的卷积神经网络,通过提取图像块的特征,对特征进行非线性映射后,重建出高分辨率的图像。此卷积神经网络自提出后,被广泛使用,准确性和可靠性得到了很好的验证。Super-resolution (Super-Resolution) means converting a low-resolution (Low Resolution) image into a high-resolution (High Resolution) image, usually improving image quality and clarity. Super-Resolution Convolutional Neural Network (SRCNN) is a convolutional neural network applied to image super-resolution reconstruction. By extracting the features of image blocks and nonlinearly mapping the features, a high-resolution image can be reconstructed. image. Since this convolutional neural network was proposed, it has been widely used, and its accuracy and reliability have been well verified.
在当今这个信息时代里,从科学家们的研究和统计的数据表明,人类获取的来自外界的信息中,大概75%的是通过眼睛获得的,眼睛获得的信息通过视觉系统转换成图像并传输到大脑。随着当今生活水平的快速提高,人们对图像视频质量要求越来越高。而图像、视频的分辨率不断提高也为信息传输带来巨大的挑战。越清晰的图像、视频意味着更大的数据量和需要更高的传输速率。为了保证人们的观感舒适度,现今电影等视频的帧率一般要高于24帧每秒,如果将每一帧的图像保存下来,再逐帧播放,不仅对硬盘容量要求特别高,并且对播放设备的传输、显示速率都有着巨大的挑战。如果以此方式播放视频,因为传输速率的限制,那么将不会存在2K、4K等高清视频。视频编码技术极大程度上消除了图像序列之间的冗余,使得视频的数据量大大压缩,配合现有硬件技术,让超高清视频走进人们的生活中,极大程度上满足了人们的观感需求。In today's information age, the research and statistical data of scientists show that about 75% of the information obtained by humans from the outside world is obtained through the eyes, and the information obtained by the eyes is converted into images through the visual system and transmitted to brain. With the rapid improvement of today's living standards, people have higher and higher requirements for image and video quality. The continuous improvement of image and video resolution also brings huge challenges to information transmission. Clearer images and videos mean larger data volumes and higher transmission rates. In order to ensure people's viewing comfort, the frame rate of movies and other videos is generally higher than 24 frames per second. If the image of each frame is saved and played frame by frame, not only the hard disk capacity is particularly high, but also the playback There are huge challenges in the transmission and display rate of the device. If the video is played in this way, due to the limitation of the transmission rate, there will be no high-definition videos such as 2K and 4K. Video coding technology has largely eliminated the redundancy between image sequences, greatly compressing the data volume of video, combined with existing hardware technology, allowing ultra-high-definition video to enter people's lives, to a great extent satisfying people's needs perception needs.
帧间预测是视频编码中最主要的一环,是利用视频图像帧间的相关性,即时间相关性,来达到图像压缩的目的,广泛用于普通电视、会议电视、视频电话、高清晰度电视的压缩编码。在图像传输技术中,活动图像特别是电视图像是关注的主要对象。活动图像是由时间上以帧周期为间隔的连续图像帧组成的时间图像序列,它在时间上比在空间上具有更大的相关性。大多数电视图像相邻帧间细节变化是很小的,即视频图像帧间具有很强的相关性,利用帧所具有的相关性的特点进行帧间编码,可获得比帧内编码高得多的压缩比。Inter-frame prediction is the most important part of video coding. It uses the correlation between video image frames, that is, time correlation, to achieve the purpose of image compression. It is widely used in ordinary TV, conference TV, video phone, high-definition Compression coding for television. In image transmission technology, moving images, especially TV images, are the main objects of concern. A moving image is a temporal image sequence composed of consecutive image frames temporally spaced by a frame period, and it has greater correlation in time than in space. The detail changes between adjacent frames of most TV images are very small, that is, there is a strong correlation between video image frames, and inter-frame coding can be obtained by using the correlation characteristics of frames, which is much higher than intra-frame coding. the compression ratio.
在帧间预测编码中,由于活动图像邻近帧中的景物存在着一定的相关性。因此,可将活动图像分成若干块或宏块,并设法搜索出每个块或宏块在邻近帧图像中的位置,并得出两者之间的空间位置的相对偏移量,得到的相对偏移量就是通常所指的运动矢量,得到运动矢量的过程被称为运动估计。运动矢量和经过运动匹配后得到的预测误差共同发送到解码端,在解码端按照运动矢量指明的位置,从已经解码的邻近参考帧图像中找到相应的块或宏块,和预测误差相加后就得到了块或宏块在当前帧中的位置。通过运动估计可以去除帧间冗余度,使得视频传输的比特数大为减少,因此,运动估计是视频压缩处理系统中的一个重要组成部分。本节先从运动估计的一般方法入手,重点讨论了运动估计的三个关键问题:将运动场参数化、最优化匹配函数定义以及如何寻找到最优化匹配。In inter-frame predictive coding, there is a certain correlation between the scenes in the adjacent frames of the moving image. Therefore, the active image can be divided into several blocks or macroblocks, and the position of each block or macroblock in the adjacent frame image can be searched, and the relative offset of the spatial position between the two can be obtained. The offset is usually referred to as a motion vector, and the process of obtaining the motion vector is called motion estimation. The motion vector and the prediction error obtained after motion matching are sent to the decoding end together. At the decoding end, according to the position indicated by the motion vector, the corresponding block or macroblock is found from the decoded adjacent reference frame image, and the prediction error is added. The position of the block or macroblock in the current frame is obtained. The inter-frame redundancy can be removed through motion estimation, which greatly reduces the number of bits in video transmission. Therefore, motion estimation is an important part of the video compression processing system. This section starts with the general method of motion estimation, focusing on three key issues of motion estimation: parameterizing the motion field, defining the optimal matching function, and how to find the optimal matching.
发明内容Contents of the invention
本发明的目的是区别于主流的HEVC视频编码方式,提出一种基于SRCNN的帧间预测方法。本发明旨在使用超分辨率卷积神经网络对图像序列进行帧间预测。对图像序列做运动估计和运动补偿操作后,结合超分辨率卷积神经网络训练出特征模型。使用模型中的参数,可以在对图像进行超分辨率重建,同时对图像进行运动估计和运动补偿,得到与当前图像的下一帧图像基本一致的图像。The purpose of the present invention is to propose an inter-frame prediction method based on SRCNN, which is different from the mainstream HEVC video encoding method. The present invention aims at inter-prediction of image sequences using super-resolution convolutional neural networks. After performing motion estimation and motion compensation operations on the image sequence, a feature model is trained with a super-resolution convolutional neural network. Using the parameters in the model, the super-resolution reconstruction of the image can be performed, and the motion estimation and motion compensation can be performed on the image at the same time, so as to obtain an image that is basically consistent with the next frame of the current image.
本发明解决其技术问题所采用的技术方案包括如下步骤:The technical solution adopted by the present invention to solve its technical problems comprises the steps:
步骤1:收集大量不同场景的视频文件,按不同的量化参数(QP)对视频进行压缩;Step 1: collect a large amount of video files of different scenes, and compress the video according to different quantization parameters (QP);
步骤2:从视频中提取图像序列,提取图像序列时,前后两帧图像的时间间隔设置为t,t<0.1秒;Step 2: Extract the image sequence from the video. When extracting the image sequence, the time interval between the two frames of images before and after is set to t, and t<0.1 second;
步骤3:将图像序列中的部分划分为验证集。逐帧读取剩余图像,除了读取的图像序列的首帧外,每张图像使用当前帧和前一帧,计算两帧图像之间的残差,将前一帧图像和此残差结合,对其进行运动补偿,得到前一帧图像的预测帧。保存计算所得的预测帧序列,将预测帧图像序列进行划分,得到训练集和测试集,二者的比例为4:1。Step 3: Partition the image sequence into a validation set. Read the remaining images frame by frame. Except for the first frame of the read image sequence, each image uses the current frame and the previous frame, calculates the residual between the two frames, and combines the previous frame with the residual. Perform motion compensation on it to obtain the predicted frame of the previous frame image. Save the calculated predicted frame sequence, divide the predicted frame image sequence to obtain a training set and a test set, and the ratio of the two is 4:1.
步骤4:输入训练集和测试集,设置合适的超参数,使用超分辨率卷积神经网络(SRCNN)训练参数模型;Step 4: Input the training set and test set, set the appropriate hyperparameters, and use the super-resolution convolutional neural network (SRCNN) to train the parameter model;
步骤5:计算验证集内每一图像序列中的第i帧图像与第i+1帧的峰值信噪比(PSNR),记作PSRN1;读取参数模型中的参数对获取的图像序列内的第i帧图像进行处理,得到重建图像I;计算重建图像I与验证集内图像序列的第i帧图像之间的PSNR,记作PSNR2;Step 5: Calculate the peak signal-to-noise ratio (PSNR) of the i-th frame image and the i+1-th frame in each image sequence in the verification set, denoted as PSRN1; read the parameters in the parameter model to the obtained image sequence Process the i-th frame image to obtain a reconstructed image I; calculate the PSNR between the reconstructed image I and the i-th frame image of the image sequence in the verification set, denoted as PSNR2;
比较计算所得的两个PSNR值,若PSNR2≥PSNR1,则认为该模型有效;Compare the calculated two PSNR values, if PSNR2≥PSNR1, the model is considered valid;
若PSNR2<PSNR1,则认为模型效果不好;记ERR=PSNR1-PSNR2;若ERR<5,则认为训练超参数设置有问题,返回步骤4,调整学习率的超参数,然后重新训练参数模型;若ERR≥5,则认为数据集的划分策略又问题,返回步骤3,通过扩充数据集使数据集包含更多场景,重新划分训练集和测试集后进行训练和验证;If PSNR2<PSNR1, it is considered that the model effect is not good; record ERR=PSNR1-PSNR2; if ERR<5, it is considered that there is a problem with the training hyperparameter setting, return to step 4, adjust the hyperparameter of the learning rate, and then retrain the parameter model; If ERR ≥ 5, it is considered that the division strategy of the data set is problematic, return to step 3, expand the data set to include more scenes, re-divide the training set and test set, and then perform training and verification;
若两张图像差别较大,PSNR值超出最低预设阈值,则调整训练集、测试集;If the difference between the two images is large and the PSNR value exceeds the minimum preset threshold, adjust the training set and test set;
若两张图像差别较小,PSNR值在最佳预设阈值和最低预设阈值之间,则返回步骤4调整超分辨率卷积神经网络的参数,重新训练参数模型。If the difference between the two images is small and the PSNR value is between the optimal preset threshold and the lowest preset threshold, return to step 4 to adjust the parameters of the super-resolution convolutional neural network and retrain the parameter model.
所述的使用参数模型重建图像具体实现如下:The specific implementation of the image reconstruction using the parameter model is as follows:
1.将输入的低分辨率图像转到YCbCr色彩空间取灰度图,作为图像重建操作的输入i。对图像i进行下采样,下采样的步长设置为k,得到低维度的图像;1. Convert the input low-resolution image to the YCbCr color space to take the grayscale image as the input i of the image reconstruction operation. Downsample the image i, and set the downsampling step size to k to obtain a low-dimensional image;
2.对低维度的图像使用双三次插值,将其放大到目标大小,即输入的低分辨率图像大小;2. Use bicubic interpolation on the low-dimensional image to enlarge it to the target size, which is the size of the input low-resolution image;
3.读取参数模型中的参数,包括各个网络节点的权重和偏置。通过三层卷积网络对插值后的图像做非线性映射,得到的重建后的结果,图像I;3. Read the parameters in the parameter model, including the weight and bias of each network node. Perform nonlinear mapping on the interpolated image through a three-layer convolutional network, and obtain the reconstructed result, image I;
4.将图像I转回RGB彩色图,得到重建的高分辨率图像。4. Convert the image I back to an RGB color map to obtain a reconstructed high-resolution image.
本发明有益效果如下:The beneficial effects of the present invention are as follows:
本发明的创新性在于将深度学习应用到视频编码的帧间预测,使用卷积神经网络,对图像序列间的运动估计、运动补偿操作进行特征提取和训练学习。同时,使用超分辨率神经网络,在图像重建时,图像的画质会得到增强。The innovation of the present invention lies in the application of deep learning to the inter-frame prediction of video coding, and the use of convolutional neural networks to perform feature extraction and training learning on motion estimation and motion compensation operations between image sequences. At the same time, using the super-resolution neural network, the quality of the image will be enhanced when the image is reconstructed.
附图说明Description of drawings
图1是超分辨率卷积神经网络SRCNN的示意图;Figure 1 is a schematic diagram of a super-resolution convolutional neural network SRCNN;
图2是本发明实施的流程图。Figure 2 is a flowchart of the implementation of the present invention.
具体实施方式Detailed ways
本发明主要针对视频编码内的帧间预测方法进行算法创新,对于整个模型的训练流程进行了详细的介绍,以下结合附图,详细阐述本发明的具体实施步骤,本发明的目的和效果将变得更加明显。The present invention mainly focuses on the algorithm innovation of the inter-frame prediction method in video coding, and introduces the training process of the entire model in detail. The specific implementation steps of the present invention will be described in detail below in conjunction with the accompanying drawings. The purpose and effect of the present invention will change more obvious.
图1是超分辨率卷积神经网络SRCNN的示意图,从图中可以清楚地看到该卷积神经网络结构简单,通过非线性映射和图像重建,可以对图像的画质起到增强作用。运用该网络,可以在对图像序列进行帧间预测的同时,提高图像的分辨率。Figure 1 is a schematic diagram of the super-resolution convolutional neural network SRCNN. It can be clearly seen from the figure that the convolutional neural network has a simple structure and can enhance the image quality through nonlinear mapping and image reconstruction. Using this network, the resolution of the image can be improved while performing inter-frame prediction on the image sequence.
图2是本发明的实施流程图,其中具体操作包括:Fig. 2 is the implementation flowchart of the present invention, and wherein concrete operation comprises:
1.收集大量YUV格式的视频文件,包含各种不同的场景。1. Collect a large number of video files in YUV format, including various scenes.
2.使用不同的量化参数对视频文件进行压缩,量化参数越高,则压缩程度越高,主要关注量化参数在28至42之间的压缩比。2. Use different quantization parameters to compress video files. The higher the quantization parameter, the higher the degree of compression. The main focus is on the compression ratio of the quantization parameter between 28 and 42.
3.从视频文件中提取图像序列,根据不同时长的视频,提取不同数量的图像,来保证图像序列的间隔一致。为了保证前后两帧图像间的变化不大,提取图像的时间间隔要设置得很小,具体根据视频的长度来设置。3. Extract image sequences from video files, and extract different numbers of images according to videos of different durations to ensure consistent intervals between image sequences. In order to ensure that there is not much change between the two frames of images before and after, the time interval for extracting images should be set very small, specifically according to the length of the video.
4.对提取出来的每一张图像做运动估计、运动补偿,此操作具体为输入当前帧及下一帧图像,通过对比两帧图像,对当前帧做运动估计、运动补偿。4. Perform motion estimation and motion compensation for each extracted image. This operation is specifically to input the current frame and the next frame image, and perform motion estimation and motion compensation for the current frame by comparing the two frames of images.
5.使用处理好的图像序列,组织训练集和测试集。验证模型所需的验证集则需要用没有做过运动估计、运动补偿的图像序列。5. Using the processed image sequence, organize the training set and test set. The verification set required to verify the model requires an image sequence that has not been motion estimated or compensated.
6.输入训练集和测试集,设置合适的参数,使用超分辨率卷积神经网络SRCNN来训练模型。6. Input the training set and test set, set appropriate parameters, and use the super-resolution convolutional neural network SRCNN to train the model.
7.验证训练好的模型是否有效,通过对比原本提取出的下一帧图像和使用该模型参数重建出的图像,若两张图像几无差别,则可认为该模型有效。若两张图像有着明显的差别,还要根据不同情况作出调整。若两张图像的差别很大,则需要调整数据集,重新训练模型,如果两图像间差别不是很大,则需要在成像效果上做出改进,调整网络参数,重新训练出复合要求的模型。7. Verify whether the trained model is valid. By comparing the original extracted next frame image with the image reconstructed using the model parameters, if there is little difference between the two images, the model can be considered valid. If there are obvious differences between the two images, adjustments should be made according to different situations. If the difference between the two images is very large, you need to adjust the data set and retrain the model. If the difference between the two images is not very large, you need to improve the imaging effect, adjust the network parameters, and retrain the model with composite requirements.
在对比生成图像和原图像的下一帧图像时,需要结合视觉上的主观判断跟客观数值分析。主观上,通过肉眼观察两帧图像,如果两张图片几无差别,可以主观上认为模型有效。但由于原本前后帧图像的区别并不大,还需借助数学工具,来对两张图像进行比较。可以使用,即峰值信噪比即PSNR来对重建效果进行客观评价,PSNR是一种评价图像的客观标准,其公式如下:When comparing the generated image with the next frame of the original image, it is necessary to combine visual subjective judgment with objective numerical analysis. Subjectively, by observing two frames of images with the naked eye, if there is little difference between the two images, the model can be considered to be valid subjectively. However, since there is not much difference between the front and rear frame images, it is necessary to use mathematical tools to compare the two images. It can be used to objectively evaluate the reconstruction effect, that is, the peak signal-to-noise ratio (PSNR). PSNR is an objective standard for evaluating images. The formula is as follows:
其中,MSE为均方误差(Meansquarederror)。分别计算原图像与其下一帧图像、原图像和重建出的图像间的PSNR数值,若两者数值接近,则说明该模型效果很好,基本重建出了与原图像下一帧图像相同的图片。若后者的PSNR数值更高,则可以认为,程序在对图像进行帧间预测的同时,还提高了图像质量。Among them, MSE is mean square error (Meansquarederror). Calculate the PSNR values between the original image and its next frame image, the original image and the reconstructed image respectively. If the two values are close, it means that the model works well, and basically reconstructs the same picture as the next frame of the original image. . If the PSNR value of the latter is higher, it can be considered that the program improves the image quality while performing inter-frame prediction on the image.
借助PSNR,可以从客观上再次验证模型的准确性,以此来减少工作量,并保证该方案有效地实施。With the help of PSNR, the accuracy of the model can be verified again objectively, so as to reduce the workload and ensure the effective implementation of the scheme.
Claims (3)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910388829.6A CN110177282B (en) | 2019-05-10 | 2019-05-10 | An Inter-frame Prediction Method Based on SRCNN |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910388829.6A CN110177282B (en) | 2019-05-10 | 2019-05-10 | An Inter-frame Prediction Method Based on SRCNN |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110177282A true CN110177282A (en) | 2019-08-27 |
CN110177282B CN110177282B (en) | 2021-06-04 |
Family
ID=67690836
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910388829.6A Active CN110177282B (en) | 2019-05-10 | 2019-05-10 | An Inter-frame Prediction Method Based on SRCNN |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110177282B (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112155511A (en) * | 2020-09-30 | 2021-01-01 | 广东唯仁医疗科技有限公司 | Method for compensating human eye shake in OCT (optical coherence tomography) acquisition process based on deep learning |
CN112601095A (en) * | 2020-11-19 | 2021-04-02 | 北京影谱科技股份有限公司 | Method and system for creating fractional interpolation model of video brightness and chrominance |
CN113191945A (en) * | 2020-12-03 | 2021-07-30 | 陕西师范大学 | High-energy-efficiency image super-resolution system and method for heterogeneous platform |
CN113592719A (en) * | 2021-08-14 | 2021-11-02 | 北京达佳互联信息技术有限公司 | Training method of video super-resolution model, video processing method and corresponding equipment |
CN117313818A (en) * | 2023-09-28 | 2023-12-29 | 四川大学 | Method for training lightweight convolutional neural network and terminal equipment |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107133919A (en) * | 2017-05-16 | 2017-09-05 | 西安电子科技大学 | Time dimension video super-resolution method based on deep learning |
CN108012157A (en) * | 2017-11-27 | 2018-05-08 | 上海交通大学 | Construction method for the convolutional neural networks of Video coding fractional pixel interpolation |
CN108805808A (en) * | 2018-04-04 | 2018-11-13 | 东南大学 | A method of improving video resolution using convolutional neural networks |
CN109087243A (en) * | 2018-06-29 | 2018-12-25 | 中山大学 | A kind of video super-resolution generation method generating confrontation network based on depth convolution |
US20190139205A1 (en) * | 2017-11-09 | 2019-05-09 | Samsung Electronics Co., Ltd. | Method and apparatus for video super resolution using convolutional neural network with two-stage motion compensation |
-
2019
- 2019-05-10 CN CN201910388829.6A patent/CN110177282B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107133919A (en) * | 2017-05-16 | 2017-09-05 | 西安电子科技大学 | Time dimension video super-resolution method based on deep learning |
US20190139205A1 (en) * | 2017-11-09 | 2019-05-09 | Samsung Electronics Co., Ltd. | Method and apparatus for video super resolution using convolutional neural network with two-stage motion compensation |
CN108012157A (en) * | 2017-11-27 | 2018-05-08 | 上海交通大学 | Construction method for the convolutional neural networks of Video coding fractional pixel interpolation |
CN108805808A (en) * | 2018-04-04 | 2018-11-13 | 东南大学 | A method of improving video resolution using convolutional neural networks |
CN109087243A (en) * | 2018-06-29 | 2018-12-25 | 中山大学 | A kind of video super-resolution generation method generating confrontation network based on depth convolution |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112155511A (en) * | 2020-09-30 | 2021-01-01 | 广东唯仁医疗科技有限公司 | Method for compensating human eye shake in OCT (optical coherence tomography) acquisition process based on deep learning |
CN112601095A (en) * | 2020-11-19 | 2021-04-02 | 北京影谱科技股份有限公司 | Method and system for creating fractional interpolation model of video brightness and chrominance |
CN112601095B (en) * | 2020-11-19 | 2023-01-10 | 北京影谱科技股份有限公司 | Method and system for creating fractional interpolation model of video brightness and chrominance |
CN113191945A (en) * | 2020-12-03 | 2021-07-30 | 陕西师范大学 | High-energy-efficiency image super-resolution system and method for heterogeneous platform |
CN113191945B (en) * | 2020-12-03 | 2023-10-27 | 陕西师范大学 | Heterogeneous platform-oriented high-energy-efficiency image super-resolution system and method thereof |
CN113592719A (en) * | 2021-08-14 | 2021-11-02 | 北京达佳互联信息技术有限公司 | Training method of video super-resolution model, video processing method and corresponding equipment |
CN113592719B (en) * | 2021-08-14 | 2023-11-28 | 北京达佳互联信息技术有限公司 | Training method of video super-resolution model, video processing method and corresponding equipment |
CN117313818A (en) * | 2023-09-28 | 2023-12-29 | 四川大学 | Method for training lightweight convolutional neural network and terminal equipment |
Also Published As
Publication number | Publication date |
---|---|
CN110177282B (en) | 2021-06-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Hu et al. | FVC: A new framework towards deep video compression in feature space | |
CN110177282A (en) | A kind of inter-frame prediction method based on SRCNN | |
CN101668205B (en) | Self-adapting down-sampling stereo video compressed coding method based on residual error macro block | |
CN100463527C (en) | A method for disparity estimation of multi-viewpoint video images | |
CN107027029B (en) | High-performance video coding improvement method based on frame rate conversion | |
CN101729891B (en) | Method for encoding multi-view depth video | |
CN108924554B (en) | A Rate Distortion Optimization Method for Panoramic Video Coding Based on Spherical Weighted Structural Similarity | |
CN111083477B (en) | HEVC Optimization Algorithm Based on Visual Saliency | |
CN101710993A (en) | Block-based self-adaptive super-resolution video processing method and system | |
CN104602028B (en) | A kind of three-dimensional video-frequency B frames entire frame loss error concealing method | |
CN101600109A (en) | H.264 Downsizing Transcoding Method Based on Texture and Motion Features | |
CN101867816A (en) | Stereoscopic video asymmetric compression coding method based on human-eye visual characteristic | |
CN109905717A (en) | A H.264/AVC Coding Optimization Method Based on Space-Time Domain Downsampling and Reconstruction | |
CN106412572A (en) | Video stream encoding quality evaluation method based on motion characteristics | |
CN106961610A (en) | With reference to the ultra high-definition video new type of compression framework of super-resolution rebuilding | |
CN114222127A (en) | Video coding method, video decoding method and device | |
Xu et al. | Consistent visual quality control in video coding | |
CN102510496B (en) | Quick size reduction transcoding method based on region of interest | |
CN115002482B (en) | End-to-end video compression method and system using structural preserving motion estimation | |
Li et al. | Perceptual quality assessment of face video compression: A benchmark and an effective method | |
Amirpour et al. | A real-time video quality metric for HTTP adaptive streaming | |
Yang et al. | Content adaptive spatial–temporal rescaling for video coding optimization | |
CN115131254A (en) | Constant bit rate compressed video quality enhancement method based on two-domain learning | |
CN110401832B (en) | Panoramic video objective quality assessment method based on space-time pipeline modeling | |
Jenab et al. | Content-adaptive resolution control to improve video coding efficiency |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |