[go: up one dir, main page]

CN109788270B - 3D-360-degree panoramic image generation method and device - Google Patents

3D-360-degree panoramic image generation method and device Download PDF

Info

Publication number
CN109788270B
CN109788270B CN201811619904.7A CN201811619904A CN109788270B CN 109788270 B CN109788270 B CN 109788270B CN 201811619904 A CN201811619904 A CN 201811619904A CN 109788270 B CN109788270 B CN 109788270B
Authority
CN
China
Prior art keywords
image
right eye
layer
network model
eye views
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811619904.7A
Other languages
Chinese (zh)
Other versions
CN109788270A (en
Inventor
周强
高宏彬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Meilewei Electronic Technology Co ltd
Original Assignee
Nanjing Magewell Electronic Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Magewell Electronic Technology Co ltd filed Critical Nanjing Magewell Electronic Technology Co ltd
Priority to CN201811619904.7A priority Critical patent/CN109788270B/en
Publication of CN109788270A publication Critical patent/CN109788270A/en
Application granted granted Critical
Publication of CN109788270B publication Critical patent/CN109788270B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Processing (AREA)
  • Studio Devices (AREA)

Abstract

本发明涉及3D‑360度全景图像生成方法及装置,采用深度学习神经网络方法,利用环绕多相机图像画面直接生成左右眼视图,进而拼接合成3D‑360度全景画面,获取左右眼视图的整个过程由网络模型通过自动训练得到合适特征自动插值实现,不需要另外使用光流算法计算,大幅降低计算量,增加全景生成的鲁棒性的同时避免出现难以接受的伪影,提高了全景图像生成速度,同时图像生成过程可以应对光线复杂、含有近景目标和纹理不丰富的场景,图像质量高,高效高质量高稳定性的输出,可满足实时全景图像采集的需要。

Figure 201811619904

The present invention relates to a method and device for generating a 3D-360-degree panoramic image. The deep learning neural network method is adopted to directly generate left and right eye views by using surrounding multi-camera image pictures, and then splicing and synthesizing a 3D-360-degree panoramic picture to obtain the entire process of left and right eye views. It is realized by automatic interpolation of suitable features obtained by automatic training of the network model, without the need for additional calculation by optical flow algorithm, which greatly reduces the amount of calculation, increases the robustness of panorama generation, avoids unacceptable artifacts, and improves the speed of panorama image generation. At the same time, the image generation process can deal with scenes with complex light, close-range targets and not rich textures, high image quality, high-efficiency, high-quality and high-stability output, which can meet the needs of real-time panoramic image acquisition.

Figure 201811619904

Description

3D-360-degree panoramic image generation method and device
Technical Field
The invention relates to the technical field of camera shooting and image processing, in particular to a 3D-360-degree panoramic image generation method and device.
Background
With the increasing development of Virtual Reality (VR) technology, the production of VR content is becoming a short board for the development of the whole industry. 3D-360 degree panoramic video in VR content is an important direction of video industry, and the performance of the acquisition equipment plays a crucial role in video image quality.
The 3D-360 degree panoramic image is a shot image field of view reaching 360 degrees in the horizontal direction, simultaneously contains left and right eye panoramic views with horizontal parallax, and can be respectively displayed on left and right eye screens of VR glasses, so that a three-dimensional panoramic image display effect with reality is realized.
The current 3D-360 degree panoramic image splicing implementation method comprises the following steps: a plurality of images are shot by surrounding multi-camera hardware, then optical flows are calculated for two adjacent images in the plurality of images, the optical flows are that the positions of any pixel in overlapped pictures of the two images correspond to the position of the other image, the left eye view and the right eye view are interpolated by using the pixel level corresponding relation, and if a pair of virtual cameras simulating human eyes are positioned behind the two cameras, the imaging of the visual fields of the two virtual cameras between the two cameras can be completely obtained by interpolation of the two real cameras. To achieve 3D effect, two images of the left and right eyes of the same scene must be synthesized, and a 360-degree panorama is obtained by sequentially connecting a plurality of left and right eye images synthesized by two cameras.
In the above-described conventional 3D-360 panoramic image stitching process, the link directly determining the stitching quality is optical flow calculation of adjacent images, and the pixel level correspondence of the accurate adjacent views can generate a flawless and artifact-free interpolation result. The optical flow calculation is always a difficult problem in the computer vision industry, and is easy to fail in scenes with complex light, close objects, and poor texture, and the like, and the spliced images are directly mistaken at the time. Meanwhile, optical flow calculation is to obtain the pixel level position correspondence of any pixel in an image in another view, and the calculation amount is often large, so that the panoramic real-time stitching application which aims to achieve a high frame rate is difficult, the hardware calculation cost of the panoramic camera is increased, and the large-scale application of panoramic stitching is limited.
Disclosure of Invention
The invention aims to improve the existing 3D-360-degree panoramic image generation method and provides a 3D-360-degree panoramic image generation method and a device.
In order to achieve the technical purpose, the invention adopts the following technical scheme:
A3D-360 degree panoramic image generation method comprises the following steps:
acquiring images of cameras surrounding a plurality of cameras, one image for each camera;
preprocessing a plurality of acquired images into images meeting the input requirements of a network model; the network model is obtained by performing iterative training on a plurality of groups of image sample sets generated by the virtual camera and 3D left and right eye views corresponding to the image sample sets;
inputting the preprocessed multiple images into a network model, and calculating to obtain multiple left and right eye views;
carrying out post-processing on the obtained multiple left and right eye views, and recovering the size and the pixel value range of the original image;
splicing the plurality of left-eye views after the post-processing according to a sequence to obtain a left-eye panoramic view, splicing the plurality of right-eye views according to a sequence to obtain a right-eye panoramic view, and splicing the left-eye panoramic view and the right-eye panoramic view up and down to obtain the required panoramic image.
Further, the network model is obtained through convolutional neural network training, and the convolutional neural network comprises a plurality of first convolutional layers, a first activation function layer, a pooling layer, a plurality of anti-convolutional layers, a second convolutional layer and a second activation function layer which are sequentially connected.
Further, the training method of the network model comprises the following steps:
acquiring an image sample set generated by a virtual camera and a plurality of 3D left and right eye views corresponding to the image sample set;
preprocessing an image sample set;
inputting each group of preprocessed image sample set into a convolutional neural network, outputting a plurality of generated left and right eye views, calculating prediction errors according to the left and right eye views obtained by calculation of each group of image sample set and the left and right eye views obtained by a virtual camera, and performing iterative training on the convolutional neural network by adopting a supervised back propagation method to obtain a deep learning network model.
Further, the method for generating a plurality of left and right eye views by using the convolutional neural network comprises the following steps:
s1: performing convolution operation on the acquired image through the first convolution layer, performing nonlinear transformation on a convolution operation result through the first activation function layer, and performing pooling operation on a nonlinear transformation result through the pooling layer;
s2: repeating S1 to obtain a plurality of feature maps with descending scales;
s3: sampling the characteristic diagram result of the processing result of the front half part through the deconvolution layer to obtain a characteristic diagram with a plurality of scales rising continuously; then, connecting the first half part characteristic diagram and the second half part characteristic diagram of the network with the same scale in parallel, performing convolution operation on the processing result through a second convolution layer, and performing nonlinear transformation on the convolution operation result of the second convolution layer through a second activation function layer;
s4: the S3 is repeated to obtain the prediction results of the plurality of left and right eye views.
Further, the activation functions adopted in the first activation function layer and the second activation function layer are linear rectification functions; the pooling layer adopts a maximum pooling mode.
Further, the method for preprocessing the image or the image sample set comprises the following steps:
scaling the image to a standard size;
normalizing the scaled image pixel values so that all pixel values are between 0 and 1;
and averaging 0 of each pixel value in the normalized image.
Further, the image post-processing method comprises the following steps:
multiplying the pixel value of the image by a coefficient to restore the pixel value to an original pixel value range;
and enlarging the image restored by the numerical value range to the standard size.
Further, the method for acquiring the image sample set and the 3D left and right eye views corresponding to the image sample set by the virtual camera includes:
simulating a plurality of virtual cameras to be placed in a surrounding mode by using a VR graphic engine to form an annular virtual camera set;
setting a first group of virtual cameras to be completely horizontally arranged in equal proportion, wherein a virtual imaging scene comprises objects and textures of depth of field, and each virtual camera performs imaging independently to obtain a group of image sets;
placing a second group of two virtual cameras in the circular shape of the annular virtual camera set, wherein the second group of virtual cameras only record pixels in the vertical direction right in front of the optical center of the cameras, simulating the second group of two virtual cameras to rotate by taking the circle center of the annular virtual camera as the center, recording scanning imaging right in front of the optical center of the second group of two virtual cameras, recording a left eye view and a right eye view respectively when each scanning passes through the two peripheral virtual cameras, and recording left eye views and right eye views corresponding to any two adjacent peripheral virtual cameras when the second group of two virtual cameras are rotated by 360 degrees;
and repeating the steps to obtain a plurality of groups of image samples and left and right eye view data corresponding to the image samples.
The device of the invention is realized by the following technical scheme:
a 3D-360 degree panoramic image generation apparatus comprising:
the surrounding multi-camera image acquisition module is used for acquiring images shot by surrounding multi-cameras, and each camera acquires an image;
the network model training module is used for training a network model; the network model is obtained by performing iterative training on a plurality of groups of image sample sets generated by the virtual camera and 3D left and right eye views corresponding to the image sample sets;
the image preprocessing module is connected with the surrounding multi-camera image acquisition module and used for preprocessing the acquired image into an image meeting the input requirement of a network model;
the view prediction module is connected with the network model training module and the image preprocessing module and is used for inputting a plurality of preprocessed images into the network model generated by the network model training module to obtain a plurality of left and right eye views;
the view post-processing module is connected with the view prediction module and used for restoring the acquired multiple left and right eye views to the original image size and the pixel value range;
and the panoramic image splicing module is connected with the view post-processing module and used for splicing all the left eye views subjected to post-processing to obtain a left eye panoramic view, splicing all the right eye views subjected to post-processing to obtain a right eye panoramic view, and splicing the left eye panoramic view and the right eye panoramic view up and down to obtain the required panoramic image.
Further, the network model training module performs training based on a convolutional neural network, and comprises a plurality of first convolutional layers, a first activation function layer, a pooling layer, a plurality of anti-convolutional layers, a second convolutional layer and a second activation function layer which are connected in sequence. (ii) a
The first convolution layer performs convolution operation on the acquired image;
the first activation function layer carries out nonlinear transformation on the convolution operation result;
the pooling layer performs pooling operation on the nonlinear transformation result to obtain a characteristic diagram with a plurality of scales descending continuously;
the deconvolution layer samples the result of the characteristic diagram, and then the characteristic diagram with the same scale is connected in parallel with a new characteristic diagram obtained by sampling the characteristic diagram;
the second convolution layer carries out convolution operation on the processing result of the deconvolution layer;
and the second activation function layer performs nonlinear transformation on the convolution operation result of the second convolution layer to obtain a plurality of prediction results of left and right eye views.
The invention adopts a deep learning neural network method, utilizes surrounding multi-camera image frames to directly generate left and right eye views, and further splices and synthesizes a 3D-360-degree panoramic image, the whole process of acquiring the left and right eye views is realized by automatic interpolation of proper characteristics obtained by automatic training of a network model, optical flow algorithm calculation is not needed, the calculated amount is greatly reduced, the robustness of panoramic generation is increased, meanwhile, unacceptable artifacts are avoided, the panoramic image generation speed is improved, meanwhile, the image generation process can deal with scenes with complex light, close-range objects and insufficient textures, the image quality is high, high-efficiency, high-quality and high-stability output is realized, and the requirement of real-time panoramic image acquisition can be met.
Drawings
FIG. 1 is a view showing the structure of the apparatus of the present invention.
Detailed Description
The technical solution of the present invention will be further described with reference to the accompanying drawings and detailed description.
Example 1
This example specifically illustrates an implementation of the method of the present invention.
The 3D-360 degree panoramic image generation method comprises the following steps:
s100, acquiring images of cameras surrounding a plurality of cameras, wherein each camera has one image;
s200, preprocessing a plurality of acquired images into images meeting the input requirements of a network model; the method comprises the following steps:
scaling the image to a standard size;
normalizing the scaled image pixel values so that all pixel values are between 0 and 1;
and averaging 0 of each pixel value in the normalized image.
S300, inputting the preprocessed multiple images into a network model, and calculating to obtain multiple left and right eye views;
the network model is obtained through convolutional neural network training, the convolutional neural network comprises a plurality of first convolutional layers, a first activation function layer, a pooling layer, a plurality of anti-convolutional layers, a second convolutional layer and a second activation function layer which are sequentially connected, and the training method comprises the following steps:
s310, acquiring an image sample set generated by the virtual camera and a plurality of 3D left and right eye views corresponding to the image sample set;
the method for acquiring the image sample set and the 3D left and right eye views corresponding to the image sample set through the virtual camera comprises the following steps:
simulating a plurality of virtual cameras to be placed in a surrounding mode by using a VR graphic Engine (such as a non-regional Engine, Unity 3D, CryENGINE and the like) to form an annular virtual camera group;
the method comprises the steps that a first group of virtual cameras are arranged in a system and are arranged in a horizontal equal proportion completely, a virtual imaging scene comprises objects and textures of depth of field, the simulated scene comprises an indoor space and an outdoor space, the simulated objects comprise people, buildings, office supplies, trees, flowers and plants, large stadiums, parks, the sky, the sea bottom, tunnels and the like, real world textures can be attached to the virtual scene, and each virtual camera is used for imaging independently to obtain a group of image sets.
The second group of two virtual cameras are placed in the circular ring of the annular virtual camera set, the distance between the two virtual cameras is set to be 6.4cm, the two virtual cameras of the second group only record pixels in the vertical direction in front of the optical center of the cameras, the virtual cameras of the second group are simulated to rotate by taking the circle center of the annular virtual camera set as the center, scanning imaging in front of the optical center of the two virtual cameras is recorded, each scanning passes through the two peripheral virtual cameras, namely, a left eye view and a right eye view are respectively recorded, and the two virtual cameras of the second group rotate for 360 degrees, so that left eye views and right eye views corresponding to any two adjacent peripheral virtual cameras are recorded.
And repeating the process to obtain a large amount of imaging data of the surrounding multi-camera for training and left and right eye diagram data corresponding to the imaging data.
S320, preprocessing an image sample set;
the method comprises the following steps:
scaling the image to a standard size;
normalizing the scaled image pixel values so that all pixel values are between 0 and 1;
and averaging 0 of each pixel value in the normalized image.
S330, inputting each group of preprocessed image sample sets into a convolutional neural network, and outputting a plurality of generated left and right eye views, wherein the steps are as follows:
s331, performing convolution operation on the acquired image through the first convolution layer, performing nonlinear transformation on a convolution operation result through the first activation function layer, and performing pooling operation on a nonlinear transformation result through the pooling layer; the activation function adopted in the first activation function layer is a linear rectification function; the pooling layer adopts a maximum pooling mode;
s332, repeating S1 to obtain a plurality of feature maps with descending scales;
s333, sampling the feature map result of the processing result of the front half part through the deconvolution layer to obtain a feature map with a plurality of scales rising continuously; then, connecting the first half part characteristic diagram and the second half part characteristic diagram of the network with the same scale in parallel, performing convolution operation on the processing result through a second convolution layer, and performing nonlinear transformation on the convolution operation result of the second convolution layer through a second activation function layer; the activation function adopted in the second activation function layer is a linear rectification function;
s334 repeats S3 to obtain the prediction results of the left and right eye views.
And calculating prediction errors according to the left and right eye views obtained by calculating each group of image sample sets and the left and right eye views obtained by the virtual camera, and performing iterative training on the convolutional neural network by adopting a supervised back propagation method to obtain a deep learning network model.
S340, post-processing the left and right eye views to restore the size and the pixel value range of the original image;
the method comprises the following steps:
multiplying the pixel value of the image by a coefficient to restore the pixel value to an original pixel value range; for example, for an image with an original pixel value range of 0-255, multiplying the pixel value of the image by a coefficient of 255 to restore the image;
and enlarging the image restored by the numerical value range to the standard size.
S500, splicing the plurality of left-eye views after the post-processing in sequence to obtain a left-eye panoramic view, splicing the plurality of right-eye views in sequence to obtain a right-eye panoramic view, and splicing the left-eye panoramic view and the right-eye panoramic view up and down to obtain the required panoramic image.
Example 2
This example specifically illustrates an implementation of the apparatus of the present invention.
The 3D-360 degree panorama image generating apparatus as shown in fig. 1 includes:
the surrounding multi-camera image acquisition module is used for acquiring images shot by surrounding multi-cameras, and each camera acquires an image;
the network model training module is used for training a network model; the network model is obtained by performing iterative training on a plurality of groups of image sample sets generated by the virtual camera and 3D left and right eye views corresponding to the image sample sets;
the image preprocessing module is connected with the surrounding multi-camera image acquisition module and used for preprocessing the acquired image into an image meeting the input requirement of a network model;
the view prediction module is connected with the network model training module and the image preprocessing module and is used for inputting a plurality of preprocessed images into the network model generated by the network model training module to obtain a plurality of left and right eye views;
the view post-processing module is connected with the view prediction module and used for restoring the acquired multiple left and right eye views to the original image size and the pixel value range;
and the panoramic image splicing module is connected with the view post-processing module and used for splicing all the left eye views subjected to post-processing to obtain a left eye panoramic view, splicing all the right eye views subjected to post-processing to obtain a right eye panoramic view, and splicing the left eye panoramic view and the right eye panoramic view up and down to obtain the required panoramic image.
The network model training module is used for training based on a convolutional neural network and comprises a plurality of first convolution layers, a first activation function layer, a pooling layer, a plurality of deconvolution layers, a second convolution layer and a second activation function layer which are sequentially connected. (ii) a
The first convolution layer performs convolution operation on the acquired image;
the first activation function layer carries out nonlinear transformation on the convolution operation result;
the pooling layer performs pooling operation on the nonlinear transformation result to obtain a characteristic diagram with a plurality of scales descending continuously;
the deconvolution layer samples the result of the characteristic diagram, and then the characteristic diagram with the same scale is connected in parallel with a new characteristic diagram obtained by sampling the characteristic diagram;
the second convolution layer carries out convolution operation on the processing result of the deconvolution layer;
and the second activation function layer performs nonlinear transformation on the convolution operation result of the second convolution layer to obtain a plurality of prediction results of left and right eye views.
And calculating prediction errors according to the predicted left and right eye views and the left and right eye views obtained by the virtual camera corresponding to each group of images, and performing iterative training on the convolutional neural network and the deconvolution network by adopting a supervised back propagation method to obtain a deep learning network model.
The image preprocessing module comprises:
a scaling unit scaling the image to a standard size;
the normalization unit normalizes the zoomed image pixel values to ensure that all the pixel values are between 0 and 1;
and the normalization unit is used for carrying out 0 equalization processing on each pixel value in the normalized image.
The apparatus further comprises a virtual camera module; the system is used for acquiring an image sample set required by a network model training module and left and right eye view records corresponding to the images, and simulating the disturbance of the position and orientation of the camera during imaging through the parameter setting of the virtual camera.
The view post-processing module includes:
the value range recovery unit is used for recovering all the pixel value ranges of the images in the predicted left and right eye views to an original image value range;
and a scaling unit scaling the image to a standard size.

Claims (6)

1.一种3D-360度全景图像生成方法,其特征在于,包括如下步骤:1. a 3D-360 degree panoramic image generation method, is characterized in that, comprises the steps: 获取环绕多相机的各个相机的图像,每个相机一幅图像;Get images of each camera surrounding a multi-camera, one image per camera; 将获取的多幅图像预处理为符合网络模型输入要求的图像;其中,所述网络模型经对包括多组经过虚拟相机生成的图像样本集和与其对应的3D左右眼视图进行迭代训练获取;所述网络模型经卷积神经网络训练获取,所述卷积神经网络包括依次连接的多个第一卷积层、第一激活函数层、池化层、多个反卷积层、第二卷积层和第二激活函数层;Preprocessing the acquired images into images that meet the input requirements of the network model; wherein, the network model is obtained by iterative training including multiple groups of image sample sets generated by virtual cameras and their corresponding 3D left and right eye views; The network model is obtained through the training of a convolutional neural network, and the convolutional neural network includes a plurality of first convolutional layers, a first activation function layer, a pooling layer, a plurality of deconvolutional layers, and a second convolutional layer connected in sequence. layer and the second activation function layer; 所述网络模型的训练方法包括:The training method of the network model includes: 获取虚拟相机生成的图像样本集,及与其对应的多幅3D左右眼视图;Obtain the image sample set generated by the virtual camera and its corresponding multiple 3D left and right eye views; 对图像样本集进行预处理;Preprocess the image sample set; 将各组经过预处理的图像样本集输入卷积神经网络中,并输出生成的多幅左右眼视图,根据各组图像样本集计算得到的多幅左右眼视图和虚拟相机得到的多幅左右眼视图计算预测误差,并采用有监督的反向传播方法对卷积神经网络进行迭代训练,得到深度学习的网络模型;Input each group of preprocessed image sample sets into the convolutional neural network, and output the multiple left and right eye views generated. View to calculate the prediction error, and use the supervised backpropagation method to iteratively train the convolutional neural network to obtain a deep learning network model; 利用卷积神经网络生成多幅左右眼视图的方法包括:Methods for generating multiple left and right eye views using convolutional neural networks include: S1:通过第一卷积层对获取的图像进行卷积运算,通过第一激活函数层对卷积运算结果进行非线性变换,通过池化层对非线性变换结果进行池化操作;S1: Perform a convolution operation on the acquired image through the first convolution layer, perform nonlinear transformation on the result of the convolution operation through the first activation function layer, and perform a pooling operation on the nonlinear transformation result through the pooling layer; S2:重复S1获取多个尺度不断下降的特征图;S2: Repeat S1 to obtain multiple feature maps with decreasing scales; S3:通过反卷积层对前半部分处理结果的特征图结果采样,获取多个尺度不断上升的特征图;然后将相同尺度的网络前半部分特征图和后半部分特征图并联,再通过第二卷积层对处理结果进行卷积运算,通过第二激活函数层对第二卷积层的卷积运算结果进行非线性变换;S3: Sampling the feature map results of the first half of the processing results through the deconvolution layer to obtain multiple feature maps with increasing scales; The convolution layer performs a convolution operation on the processing result, and performs nonlinear transformation on the convolution operation result of the second convolution layer through the second activation function layer; S4:重复S3获取多幅左右眼视图的预测结果;S4: Repeat S3 to obtain prediction results of multiple left and right eye views; 将经过预处理的多幅图像输入网络模型中,计算获取多幅左右眼视图;Input the preprocessed multiple images into the network model, and calculate and obtain multiple left and right eye views; 获取的多幅左右眼视图进行后处理,恢复原图像大小及像素值值域;The acquired multiple left and right eye views are post-processed to restore the original image size and pixel value range; 将经后处理的多幅左眼视图按照顺序拼接得到左眼全景视图,多幅右眼视图按照顺序拼接得到右眼全景视图,将左眼全景视图和右眼全景视图上下拼接,即得到所需全景图像。The left-eye panoramic view is obtained by stitching the post-processed multiple left-eye views in sequence, the right-eye panoramic view is obtained by stitching multiple right-eye views in sequence, and the left-eye panoramic view and the right-eye panoramic view are stitched up and down to obtain the desired Panoramic image. 2.根据权利要求1所述的一种3D-360度全景图像生成方法,其特征在于,第一激活函数层、第二激活函数层中采用的激活函数为线性整流函数;池化层采用最大池化方式。2. The method for generating a 3D-360-degree panoramic image according to claim 1, wherein the activation function adopted in the first activation function layer and the second activation function layer is a linear rectification function; pooling method. 3.根据权利要求1所述的一种3D-360度全景图像生成方法,其特征在于,所述图像或图像样本集的预处理方法包括:3. The method for generating a 3D-360-degree panoramic image according to claim 1, wherein the preprocessing method for the image or image sample set comprises: 将图像缩放至标准尺寸;Scale the image to standard size; 将缩放后的图像像素数值归一化,使得所有像素数值位于0-1之间;Normalize the scaled image pixel values so that all pixel values lie between 0-1; 将归一化图像中每个像素数值进行0均值化。0-means each pixel value in the normalized image. 4.根据权利要求1所述的一种3D-360度全景图像生成方法,其特征在于,所述图像后处理方法包括:4. The method for generating a 3D-360-degree panoramic image according to claim 1, wherein the image post-processing method comprises: 将图像像素数值乘以系数,使其恢复至原图像像素值值域;Multiply the image pixel value by the coefficient to restore it to the original image pixel value range; 将数值值域恢复的图像放大至标准尺寸。Enlarges the image recovered by the numeric range to standard size. 5.根据权利要求1所述的一种3D-360度全景图像生成方法,其特征在于,通过虚拟相机获取图像样本集及与其对应的3D左右眼视图的方法包括:5. The method for generating a 3D-360-degree panoramic image according to claim 1, wherein the method for obtaining an image sample set and its corresponding 3D left and right eye views through a virtual camera comprises: 使用VR图形引擎模拟多个虚拟相机环绕摆放,形成环形虚拟相机组;Use the VR graphics engine to simulate the surrounding placement of multiple virtual cameras to form a ring-shaped virtual camera group; 设置第一组虚拟相机,使其完全水平等比例排布,虚拟成像场景中包含有景深的物体和纹理,每个虚拟相机单独成像,获取一组图像集;Set the first group of virtual cameras so that they are completely horizontally arranged in equal proportions. The virtual imaging scene contains objects and textures with depth of field. Each virtual camera is imaged separately to obtain a set of images; 在上述环形虚拟相机组圆内部放置第二组两个虚拟相机,第二组虚拟相机只记录相机光心正前方垂直方向像素,系统模拟该第二组两个虚拟相机以环形虚拟相机圆心为中心转动,记录第二组两个虚拟相机的光心正前方扫描成像,每扫描经过外围的两个虚拟相机,即分别记录一幅左眼视图和一幅右眼视图,将第二组两个虚拟相机转动360度,则记录了任意两个相邻的外围虚拟相机对应的左右眼视图;A second group of two virtual cameras is placed inside the above-mentioned annular virtual camera group circle. The second group of virtual cameras only records the pixels in the vertical direction directly in front of the optical center of the camera. The system simulates the second group of two virtual cameras with the center of the annular virtual camera as the center. Rotate and record the scanning imaging in front of the optical center of the second group of two virtual cameras. Each scan passes through the two peripheral virtual cameras, that is, a left-eye view and a right-eye view are recorded respectively, and the two virtual cameras in the second group are recorded. When the camera is rotated 360 degrees, the left and right eye views corresponding to any two adjacent peripheral virtual cameras are recorded; 重复上述步骤,获取多组图像样本和与其对应的左右眼视图数据。Repeat the above steps to obtain multiple sets of image samples and corresponding left and right eye view data. 6.一种3D-360度全景图像生成装置,其特征在于,包括:6. A 3D-360 degree panoramic image generation device, characterized in that, comprising: 环绕多相机图像获取模块,用于获取环绕多相机拍摄的图像,每个相机获取一幅图像;The surround multi-camera image acquisition module is used to acquire the images captured by the surround multi-cameras, and each camera acquires one image; 网络模型训练模块,用于训练网络模型;所述网络模型经对包括多组经过虚拟相机生成的图像样本集和与其对应的3D左右眼视图进行迭代训练获取;a network model training module, used for training a network model; the network model is obtained by iterative training including multiple groups of image sample sets generated by virtual cameras and their corresponding 3D left and right eye views; 所述网络模型训练模块基于卷积神经网络进行训练,包括依次连接的多个第一卷积层、第一激活函数层、池化层、多个反卷积层、第二卷积层和第二激活函数层;The network model training module is trained based on a convolutional neural network, including a plurality of first convolutional layers, a first activation function layer, a pooling layer, a plurality of deconvolutional layers, a second convolutional layer and a first convolutional layer connected in sequence. Two activation function layers; 所述第一卷积层对获取的图像进行卷积运算;The first convolution layer performs a convolution operation on the acquired image; 所述第一激活函数层对卷积运算结果进行非线性变换;The first activation function layer performs nonlinear transformation on the result of the convolution operation; 所述池化层对非线性变换结果进行池化操作,获取多个尺度不断下降的特征图;The pooling layer performs a pooling operation on the nonlinear transformation result to obtain a plurality of feature maps with decreasing scales; 所述反卷积层对特征图结果采样,然后将相同尺度的特征图和特征图采样获取的新的特征图并联;The deconvolution layer samples the feature map result, and then parallelizes the feature map of the same scale and the new feature map obtained by sampling the feature map; 所述第二卷积层对反卷积层处理结果进行卷积运算;The second convolution layer performs a convolution operation on the deconvolution layer processing result; 所述第二激活函数层对第二卷积层的卷积运算结果进行非线性变换,获取多幅左右眼视图的预测结果;The second activation function layer performs nonlinear transformation on the convolution operation result of the second convolution layer, and obtains prediction results of multiple left and right eye views; 图像预处理模块,连接环绕多相机图像获取模块,用于将获取的图像预处理为符合网络模型输入要求的图像;The image preprocessing module is connected to the surrounding multi-camera image acquisition module, and is used to preprocess the acquired image into an image that meets the input requirements of the network model; 视图预测模块,连接网络模型训练模块和图像预处理模块,用于将经过预处理的多幅图像输入网络模型训练模块生成的网络模型中,获取多幅左右眼视图;The view prediction module is connected to the network model training module and the image preprocessing module, and is used to input multiple preprocessed images into the network model generated by the network model training module to obtain multiple left and right eye views; 视图后处理模块,连接视图预测模块,用于将获取的多幅左右眼视图恢复至原图像大小及像素值值域;The view post-processing module is connected to the view prediction module, and is used to restore the acquired multiple left and right eye views to the original image size and pixel value range; 全景图像拼接模块,连接视图后处理模块,用于将经后处理的所有左眼视图拼接得到左眼全景视图,将将经后处理的所有右眼视图拼接得到右眼全景视图,将左眼全景视图和右眼全景视图上下拼接,即得到所需全景图像。The panoramic image stitching module is connected to the view post-processing module, which is used to stitch all the post-processed left-eye views to obtain the left-eye panoramic view, stitch all the post-processed right-eye views to obtain the right-eye panoramic view, and stitch the left-eye panoramic view. The view and the right eye panorama view are stitched up and down to obtain the desired panorama image.
CN201811619904.7A 2018-12-28 2018-12-28 3D-360-degree panoramic image generation method and device Active CN109788270B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811619904.7A CN109788270B (en) 2018-12-28 2018-12-28 3D-360-degree panoramic image generation method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811619904.7A CN109788270B (en) 2018-12-28 2018-12-28 3D-360-degree panoramic image generation method and device

Publications (2)

Publication Number Publication Date
CN109788270A CN109788270A (en) 2019-05-21
CN109788270B true CN109788270B (en) 2021-04-09

Family

ID=66498593

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811619904.7A Active CN109788270B (en) 2018-12-28 2018-12-28 3D-360-degree panoramic image generation method and device

Country Status (1)

Country Link
CN (1) CN109788270B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR3078564B1 (en) * 2018-03-01 2020-09-11 4D View Solutions THREE-DIMENSIONAL MODELING SYSTEM OF A SCENE BY MULTI-VIEW PHOTOGRAMMETRY
US12367546B2 (en) 2019-12-24 2025-07-22 Nvidia Corporation Panorama generation using one or more neural networks
CN111476759B (en) * 2020-03-13 2022-03-25 深圳市鑫信腾机器人科技有限公司 Screen surface detection method and device, terminal and storage medium
CN113313771B (en) * 2021-07-19 2021-10-12 山东捷瑞数字科技股份有限公司 Omnibearing measuring method for industrial complex equipment
CN114742703B (en) * 2022-03-11 2024-12-20 影石创新科技股份有限公司 Method, device, equipment and storage medium for generating binocular stereoscopic panoramic images

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011144793A1 (en) * 2010-05-18 2011-11-24 Teknologian Tutkimuskeskus Vtt Mobile device, server arrangement and method for augmented reality applications
CN107018400A (en) * 2017-04-07 2017-08-04 华中科技大学 It is a kind of by 2D Video Quality Metrics into 3D videos method
CN108616746A (en) * 2017-10-12 2018-10-02 叠境数字科技(上海)有限公司 The method that 2D panoramic pictures based on deep learning turn 3D panoramic pictures
CN108830890A (en) * 2018-04-24 2018-11-16 广州启辰电子科技有限公司 A method of scene geometric information being estimated from single image using production confrontation network

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108648161B (en) * 2018-05-16 2020-09-01 江苏科技大学 Binocular Vision Obstacle Detection System and Method Based on Asymmetric Kernel Convolutional Neural Network

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011144793A1 (en) * 2010-05-18 2011-11-24 Teknologian Tutkimuskeskus Vtt Mobile device, server arrangement and method for augmented reality applications
CN107018400A (en) * 2017-04-07 2017-08-04 华中科技大学 It is a kind of by 2D Video Quality Metrics into 3D videos method
CN108616746A (en) * 2017-10-12 2018-10-02 叠境数字科技(上海)有限公司 The method that 2D panoramic pictures based on deep learning turn 3D panoramic pictures
CN108830890A (en) * 2018-04-24 2018-11-16 广州启辰电子科技有限公司 A method of scene geometric information being estimated from single image using production confrontation network

Also Published As

Publication number Publication date
CN109788270A (en) 2019-05-21

Similar Documents

Publication Publication Date Title
CN109788270B (en) 3D-360-degree panoramic image generation method and device
KR101295471B1 (en) A system and method for 3D space-dimension based image processing
CN107862718B (en) 4D holographic video capture method
EP4036863A1 (en) Human body model reconstruction method and reconstruction system, and storage medium
CN109801234B (en) Image geometry correction method and device
WO2019041351A1 (en) Real-time aliasing rendering method for 3d vr video and virtual three-dimensional scene
CN106651808B (en) Fisheye diagram conversion method and device
CN112085659A (en) A panorama stitching fusion method, system and storage medium based on spherical screen camera
EP4091141A1 (en) Neural blending for novel view synthesis
Yan et al. Stereoscopic image stitching based on a hybrid warping model
CN117274501B (en) Drivable digital person modeling method, device, equipment and medium
CN107918948B (en) 4D video rendering method
CN109523635A (en) A kind of non-rigid reconstruction of 3D anthropometric scanning and measurement method and device
US11328436B2 (en) Using camera effect in the generation of custom synthetic data for use in training an artificial intelligence model to produce an image depth map
CN113298946B (en) House three-dimensional reconstruction and ground recognition method, device, equipment and storage medium
CN118196332A (en) Multi-view human body surface reconstruction method and system
JP6799468B2 (en) Image processing equipment, image processing methods and computer programs
CN116708862A (en) Virtual background generation method for live broadcasting room, computer equipment and storage medium
CN111629194B (en) Method and system for converting panoramic video into 6DOF video based on neural network
CN115035178A (en) Augmented reality display system, method and storage medium
Shilaskar et al. A scalable structure-from-motion framework for efficient 2D-to-3D reconstruction of historical artifacts
CN115314672A (en) A real-time three-dimensional panoramic video surveillance system
CN119172516A (en) A method for converting a two-dimensional image into a three-dimensional image
CN114529673A (en) Display method and display system for three-dimensional touring scene and display terminal
CN117527994A (en) Visual presentation method and system for space simulation shooting

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CP03 Change of name, title or address

Address after: No. 68-66 Suyuan Avenue, Jiangning District, Nanjing City, Jiangsu Province, 211100

Patentee after: Nanjing Meilewei Electronic Technology Co.,Ltd.

Country or region after: China

Address before: 211100 14th floor, building 3, No.89 Shengli Road, Jiangning Development Zone, Nanjing City, Jiangsu Province

Patentee before: NANJING MAGEWELL ELECTRONIC TECHNOLOGY Co.,Ltd.

Country or region before: China

CP03 Change of name, title or address