[go: up one dir, main page]

CN106878698A - Method and system for mobile naked-eye three-dimensional virtual reality based on optical path acquisition - Google Patents

Method and system for mobile naked-eye three-dimensional virtual reality based on optical path acquisition Download PDF

Info

Publication number
CN106878698A
CN106878698A CN201611202874.0A CN201611202874A CN106878698A CN 106878698 A CN106878698 A CN 106878698A CN 201611202874 A CN201611202874 A CN 201611202874A CN 106878698 A CN106878698 A CN 106878698A
Authority
CN
China
Prior art keywords
image
virtual reality
dimensional virtual
viewpoint
map
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201611202874.0A
Other languages
Chinese (zh)
Other versions
CN106878698B (en
Inventor
马魁
裴仁静
耿征
张梅
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN201611202874.0A priority Critical patent/CN106878698B/en
Publication of CN106878698A publication Critical patent/CN106878698A/en
Application granted granted Critical
Publication of CN106878698B publication Critical patent/CN106878698B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/261Image signal generators with monoscopic-to-stereoscopic image conversion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/302Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Processing Or Creating Images (AREA)

Abstract

本发明涉及本发明实施例提出一种基于光路采集的移动裸眼三维虚拟现实的方法和系统。其中,该方法可以包括:采集图像;根据采集到的图像获得两视点合成图索引图;根据加速度的位置信息和角速度的旋转信息,获得两视点图;基于图像的两视点合成图索引图,对两视点图渲染出合成图,得到三维虚拟现实显示图像。优选地,获得两视点图步骤具体可以包括:确定图像上子像素亮度影响权重,然后利用带有包围盒的最小二乘方法,确定两视点的合成图索引图;根据加速度的位置信息和角速度的旋转信息,调整所获取的图像的方位,获得两视点图。本发明实施例通过采用上述技术方案,解决了如何实现高质量的裸眼三维虚拟现实显示的技术问题。

The present invention relates to a method and system for moving naked-eye three-dimensional virtual reality based on optical path acquisition proposed by the embodiment of the present invention. Wherein, the method may include: collecting an image; obtaining a two-viewpoint synthetic map index map according to the collected image; obtaining a two-viewpoint map according to the position information of acceleration and the rotation information of angular velocity; based on the image-based two-viewpoint synthetic map index map, for The composite image is rendered by rendering the two-viewpoint image, and a three-dimensional virtual reality display image is obtained. Preferably, the step of obtaining the two-viewpoint map may specifically include: determining the influence weight of sub-pixel brightness on the image, and then using the least squares method with a bounding box to determine the composite map index map of the two viewpoints; according to the position information of the acceleration and the angular velocity Rotate the information, adjust the orientation of the acquired image, and obtain a two-viewpoint map. The embodiment of the present invention solves the technical problem of how to realize high-quality naked-eye three-dimensional virtual reality display by adopting the above technical solution.

Description

基于光路采集的移动裸眼三维虚拟现实的方法和系统Method and system for mobile naked-eye three-dimensional virtual reality based on optical path collection

技术领域technical field

本发明涉及三维显示方法,尤其涉及一种基于光路采集的移动裸眼三维虚拟现实的方法和系统。The invention relates to a three-dimensional display method, in particular to a method and system for moving naked-eye three-dimensional virtual reality based on optical path acquisition.

背景技术Background technique

目前,虚拟现实在全球掀起了浪潮,而三维虚拟现实则是一个较新的发展方向。虚拟现实技术是集合了跟踪系统、触觉系统、图像生成与显示系统和可视化显示设备的一个综合成果,但在实用商业化的路中仍存在很多有待解决的问题。虚拟现实大多是通过穿戴式设备头盔来集成相关技术,但是头带大小无法调节、头盔过重、长时间佩戴时设备的透气、散热性差,这些因素都会使得观看者的舒适体验度下降;由于虚拟现实的设备需要与电脑相连,进行信号传输,其常常是连接一根很长的线缆作为通讯保障。在观看者移动的情况下,注意力一直在显示设备上而忽略了脚下的线缆就存在被绊倒的危险;相对而言,性能中等偏上的虚拟现实设备,其价格也普遍偏高。作为一款娱乐工具,目前其显示资源十分有限,但售价超出普通大众的承受能力。三维虚拟现实相对于传统的虚拟现实,更加贴近人类实际感知,能够直观的给观看者带来深度效果。在国内外众多三维显示技术中,全息技术是通过干涉和衍射的光学现象真实的记录并还原出真实物体图像的记录和还原技术,但目前的全息影像装置采用激光进行影像采集与显示,设备造价极其昂贵。集成成像三维显示技术在图像视点串扰、视点数目等多方面存在问题。裸眼体视三维显示具有可观的商业前景,目前已经走出实验室阶段的体视三维显示装置,大多结构复杂、笨重并且需要多种材料设备和自行设计的控制装置,并不适合应用在移动性较强的虚拟现实技术中。At present, virtual reality has set off a wave in the world, and three-dimensional virtual reality is a relatively new development direction. Virtual reality technology is a comprehensive achievement that integrates tracking system, haptic system, image generation and display system and visual display equipment, but there are still many problems to be solved on the road to practical commercialization. Virtual reality mostly integrates related technologies through wearable device helmets, but the size of the headband cannot be adjusted, the helmet is too heavy, and the device has poor ventilation and heat dissipation when worn for a long time. These factors will reduce the comfort experience of the viewer; The actual equipment needs to be connected with the computer for signal transmission, which is often connected with a very long cable as a communication guarantee. When the viewer is moving, the attention is always on the display device and the cable under the feet is ignored, and there is a danger of tripping; relatively speaking, the price of virtual reality devices with medium-to-high performance is generally high. As an entertainment tool, its display resources are very limited at present, but the price is beyond the affordability of the general public. Compared with traditional virtual reality, 3D virtual reality is closer to the actual perception of human beings, and can intuitively bring depth effects to viewers. Among many three-dimensional display technologies at home and abroad, holographic technology is a recording and restoration technology that records and restores real object images through optical phenomena of interference and diffraction. However, current holographic imaging devices use lasers for image acquisition and display, and equipment costs extremely expensive. Integrated imaging 3D display technology has problems in many aspects such as image viewpoint crosstalk and the number of viewpoints. Naked-eye stereoscopic 3D display has considerable commercial prospects. Most of the stereoscopic 3D display devices that have gone out of the laboratory stage are complex in structure, heavy and require a variety of materials and equipment and self-designed control devices, which are not suitable for applications in mobile applications. Strong virtual reality technology.

移动终端由于其便携性和普及性,在三维虚拟现实技术中占据了先天性的优势。微透镜阵列裸眼三维显示技术基于覆盖在二维平面屏幕上的微透镜阵列,其实现相对简单,并且其三维图像重建的质量足以用于许多消费电子设备。用于显示技术的现有渲染方法,例如飞利浦的模算术或光线反投影方法,需要依赖于设备测量的精确校正标定以产生高质量三维显示效果。Due to its portability and popularity, mobile terminals have inherent advantages in 3D virtual reality technology. The microlens array naked-eye 3D display technology is based on a microlens array covered on a 2D flat screen, which is relatively simple to implement, and the quality of its 3D image reconstruction is sufficient for many consumer electronics devices. Existing rendering methods for display technologies, such as Philips' modular arithmetic or ray backprojection methods, require precise calibration calibrations that rely on device measurements to produce high-quality 3D displays.

有鉴于此,特提出本发明。In view of this, the present invention is proposed.

发明内容Contents of the invention

为了解决现有技术中的上述问题,即为了解决如何实现高质量的裸眼三维虚拟现实显示的技术问题而提出一种基于光路采集的移动裸眼三维虚拟现实的方法。此外,还提供一种基于光路采集的移动裸眼三维虚拟现实的系统。In order to solve the above-mentioned problems in the prior art, that is, to solve the technical problem of how to realize high-quality naked-eye three-dimensional virtual reality display, a method for moving naked-eye three-dimensional virtual reality based on optical path acquisition is proposed. In addition, a mobile naked-eye three-dimensional virtual reality system based on optical path collection is also provided.

为了实现上述目的,根据本发明的一个方面,提供以下技术方案:In order to achieve the above object, according to one aspect of the present invention, the following technical solutions are provided:

一种基于光路采集的移动裸眼三维虚拟现实的方法,该方法包括:A method for moving naked-eye three-dimensional virtual reality based on optical path collection, the method comprising:

采集图像;capture images;

根据采集到的图像获得两视点合成图索引图;Obtain a two-view composite map index map according to the collected images;

根据加速度的位置信息和角速度的旋转信息,获得两视点图;According to the position information of the acceleration and the rotation information of the angular velocity, a two-viewpoint map is obtained;

基于图像的两视点合成图索引图,对两视点图渲染出合成图,得到三维虚拟现实显示图像。Based on the image-based two-viewpoint synthetic image index image, the composite image is rendered for the two-viewpoint image to obtain a three-dimensional virtual reality display image.

进一步地,所述基于采集到的所述图像获得两视点合成图索引图具体包括:Further, the obtaining the two-view composite map index map based on the collected image specifically includes:

确定图像上子像素亮度影响权重;Determine the sub-pixel brightness influence weight on the image;

基于图像上子像素亮度影响权重,利用带有包围盒的最小二乘方法,确定两视点的合成图索引图。Based on the influence weight of sub-pixel brightness on the image, the index map of the composite image of two viewpoints is determined by using the method of least squares with bounding boxes.

进一步地,确定图像上子像素亮度影响权重具体可以包括:Further, determining the sub-pixel brightness influence weight on the image may specifically include:

根据下式确定图像上子像素亮度影响权重:Determine the sub-pixel brightness influence weight on the image according to the following formula:

其中,Imagec(t)(i,j,k)表示在第t个视点处采集到的图像上的子像素,t=0、1,i=1~H,j=1~W;H表示图像的高度;W表示图像的宽度;r表示安全窗口半径;m、n表示以安全窗口中心为原点的窗口坐标;q=1~3;Imagem(i+m,j+n,q)表示图像上的子像素;表示图像上子像素(i+m,j+n,q)点亮到最大亮度时对邻近像素(i,j,k)经过归一化后的亮度影响权重。Among them, Image c(t) (i, j, k) represents the sub-pixel on the image collected at the t-th viewpoint, t=0, 1, i=1~H, j=1~W; H represents The height of the image; W represents the width of the image; r represents the radius of the security window; m and n represent the window coordinates with the center of the security window as the origin; q=1~3; Image m (i+m,j+n,q) represents sub-pixels on the image; Indicates the normalized brightness influence weight of adjacent pixels (i, j, k) when the sub-pixel (i+m, j+n, q) on the image is lit to the maximum brightness.

进一步地,基于图像上子像素亮度影响权重,利用带有包围盒的最小二乘方法,确定两视点的合成图索引图,具体可以包括:Further, based on the influence weight of the sub-pixel brightness on the image, the least squares method with bounding boxes is used to determine the composite image index map of the two viewpoints, which may specifically include:

基于图像上子像素亮度影响权重,利用带有包围盒的最小二乘方法,根据下式确定两视点的合成图索引图:Based on the influence weight of the sub-pixel brightness on the image, using the least squares method with bounding boxes, the composite image index map of the two viewpoints is determined according to the following formula:

其中,表示两视点对应子像素亮度影响权重的差异;I表示两视点的合成图索引图,且被限定于包围盒BI={I∈Z1×H:L≤I≤U}中,L=0×I1×H且U=255×I1×H,I1×H表示H维的全1列向量;表示两视点采集图像的差异;H表示图像的高度。in, Indicates the difference in the weights of sub-pixel brightness influences corresponding to two viewpoints; I represents the composite image index map of two viewpoints, and is limited to the bounding box B I ={I∈Z 1×H :L≤I≤U}, L=0 ×I 1×H and U=255×I 1×H , I 1×H represents a full 1-column vector of H dimension; Indicates the difference between images collected from two viewpoints; H indicates the height of the image.

进一步地,在确定图像上子像素亮度影响权重步骤之前还可以包括:Further, before determining the sub-pixel brightness influence weight step on the image, it may also include:

对图像进行反畸变和兴趣区域提取操作处理。Anti-distortion and region-of-interest extraction operations are performed on the image.

进一步地,根据加速度的位置信息和角速度的旋转信息,获得两视点图,具体可以包括:Further, according to the position information of the acceleration and the rotation information of the angular velocity, a two-viewpoint map is obtained, which may specifically include:

根据加速度的位置信息和角速度的旋转信息,调整所获取的图像的方位,获得两视点图。According to the position information of the acceleration and the rotation information of the angular velocity, the orientation of the acquired image is adjusted to obtain a two-viewpoint image.

基于图像的两视点合成图索引图,对两视点图渲染出合成图,得到三维虚拟现实显示图像,具体包括:Based on the image-based two-viewpoint synthetic image index map, the two-viewpoint image is rendered to a composite image to obtain a three-dimensional virtual reality display image, specifically including:

根据下式渲染出合成图,得到三维虚拟现实显示图像:The composite image is rendered according to the following formula to obtain a 3D virtual reality display image:

Syn=I0×view0+I1×view1 Syn=I 0 ×view 0 +I 1 ×view 1

其中,Syn表示三维虚拟现实显示图像;I0表示第0视点的合成图索引图;I1表示第1视点的合成图索引图;view0表示第0视点图;view1表示第1视点图。Among them, Syn represents the 3D virtual reality display image; I 0 represents the synthetic map index map of the 0th viewpoint; I 1 represents the synthetic map index map of the 1st viewpoint; view 0 represents the 0th viewpoint map; view 1 represents the 1st viewpoint map.

为了实现上述目的,根据本发明的另一个方面,提供以下技术方案:In order to achieve the above object, according to another aspect of the present invention, the following technical solutions are provided:

一种基于光路采集的移动裸眼三维虚拟现实的系统,该系统可以包括:A mobile naked-eye three-dimensional virtual reality system based on optical path acquisition, the system may include:

图像获取单元,用于采集图像,并将图像发送至控制器;an image acquisition unit, configured to acquire images and send the images to the controller;

移动终端,包括屏幕,屏幕上设有微透镜阵列膜,用于将加速度的位置信息和角速度的旋转信息发送至控制器,并通过屏幕显示三维虚拟现实显示图像;The mobile terminal includes a screen, and a microlens array film is arranged on the screen, which is used to send the position information of the acceleration and the rotation information of the angular velocity to the controller, and display a three-dimensional virtual reality display image through the screen;

控制器,分别与图像获取单元和移动终端通信连接;用于对图像进行处理,得到两视点合成图索引图,且根据加速度的位置信息和角速度的旋转信息,获得两视点图,并基于图像的两视点合成图索引图,对两视点图渲染出合成图,得到三维虚拟现实显示图像,以及将三维虚拟现实显示图像发送至移动终端。The controller is connected to the image acquisition unit and the mobile terminal respectively; it is used to process the image to obtain the index map of the two-viewpoint composite map, and obtain the two-viewpoint map according to the position information of the acceleration and the rotation information of the angular velocity, and based on the image The two-viewpoint synthetic image index image is used to render a composite image for the two-viewpoint image to obtain a three-dimensional virtual reality display image, and to send the three-dimensional virtual reality display image to a mobile terminal.

进一步地,移动终端还可以包括Further, the mobile terminal can also include

加速度计,用于获取加速度的位置信息;Accelerometer, used to obtain the location information of the acceleration;

陀螺仪,用于获取角速度的旋转信息。Gyroscope, used to obtain rotation information of angular velocity.

进一步地图像获取单元为单目相机、单目摄像机、双目相机或双目摄像机。Further, the image acquisition unit is a monocular camera, a monocular camera, a binocular camera or a binocular camera.

进一步地,移动终端为手机或个人数字助理。Further, the mobile terminal is a mobile phone or a personal digital assistant.

进一步地,控制器为计算机、笔记本电脑、服务器或工控机。Further, the controller is a computer, a notebook computer, a server or an industrial computer.

本发明实施例提出一种基于光路采集的移动裸眼三维虚拟现实的方法和系统。其中,该方法可以包括:采集图像;基于采集到的图像获得两视点合成图索引图,根据加速度的位置信息和角速度的旋转信息,获得两视点图;基于图像的两视点合成图索引图,对两视点图渲染出合成图,得到三维虚拟现实显示图像。本发明实施例将三维显示技术与虚拟现实技术相结合,能够在微透镜阵列和移动终端屏幕的参数未知的情况下,呈现出高质量的三维显示效果。相比于传统虚拟现实设备减少了头盔的重量和舒适性欠佳的束缚感,同时增加了设备自身三维显示效果而非仅仅使用双目接收左右不同场景图片达到三维的效果。Embodiments of the present invention propose a method and system for mobile naked-eye three-dimensional virtual reality based on optical path acquisition. Wherein, the method may include: collecting an image; obtaining a two-viewpoint synthetic map index map based on the collected image, and obtaining a two-viewpoint map according to the position information of the acceleration and the rotation information of the angular velocity; based on the image-based two-viewpoint synthetic map index map, The composite image is rendered by rendering the two-viewpoint image, and a three-dimensional virtual reality display image is obtained. The embodiment of the present invention combines the three-dimensional display technology with the virtual reality technology, and can present a high-quality three-dimensional display effect under the condition that the parameters of the microlens array and the screen of the mobile terminal are unknown. Compared with traditional virtual reality equipment, it reduces the weight of the helmet and the sense of restraint due to poor comfort. At the same time, it increases the three-dimensional display effect of the equipment itself instead of just using binoculars to receive pictures of different scenes on the left and right to achieve a three-dimensional effect.

附图说明Description of drawings

图1是根据本发明实施例的基于光路采集的移动裸眼三维虚拟现实的系统的结构示意图;FIG. 1 is a schematic structural diagram of a mobile naked-eye three-dimensional virtual reality system based on optical path acquisition according to an embodiment of the present invention;

图2是根据本发明另一实施例的基于光路采集的移动裸眼三维虚拟现实的系统的结构示意图;FIG. 2 is a schematic structural diagram of a mobile naked-eye three-dimensional virtual reality system based on optical path acquisition according to another embodiment of the present invention;

图3是根据本发明实施例的用户手持手机在三个不同观察点的手机裸眼三维虚拟现实显示效果示意图;Fig. 3 is a schematic diagram of a naked-eye three-dimensional virtual reality display effect of a mobile phone held by a user at three different observation points according to an embodiment of the present invention;

图4是根据本发明实施例的基于光路采集的移动裸眼三维虚拟现实的方法的流程示意图。Fig. 4 is a schematic flowchart of a method for moving naked-eye three-dimensional virtual reality based on optical path acquisition according to an embodiment of the present invention.

具体实施方式detailed description

下面参照附图来描述本发明的优选实施方式。本领域技术人员应当理解的是,这些实施方式仅仅用于解释本发明的技术原理,并非旨在限制本发明的保护范围。Preferred embodiments of the present invention are described below with reference to the accompanying drawings. Those skilled in the art should understand that these embodiments are only used to explain the technical principles of the present invention, and are not intended to limit the protection scope of the present invention.

下面以一示例性实施例来说明本发明实施例的应用平台。其中,可以利用两个基准线为60mm的图像获取单元模拟人的双眼,将其放在距离移动终端屏幕约300mm处。利用一个密封的实验箱,为图像获取单元的光路采集提供一个与外界光线隔绝的密闭环境,并在图像获取单元的后面设置风扇,以进行散热。优选地,为了使捕获图像上的像素最大化地利用,可以给图像获取单元配备长焦镜头。测试图集在计算机上生成,并且通过局域网搭建依次传输给移动终端进行显示。The following uses an exemplary embodiment to illustrate the application platform of the embodiment of the present invention. Among them, two image acquisition units with a reference line of 60 mm can be used to simulate the eyes of a person, and they are placed at a distance of about 300 mm from the screen of the mobile terminal. A sealed experimental box is used to provide a closed environment isolated from external light for the optical path acquisition of the image acquisition unit, and a fan is installed behind the image acquisition unit for heat dissipation. Preferably, in order to maximize the use of pixels on the captured image, the image acquisition unit may be equipped with a telephoto lens. The test atlas is generated on the computer, and is sequentially transmitted to the mobile terminal for display through the construction of the local area network.

上述图像获取单元包括但不限于单目相机、双目相机、单目摄像机、双目摄像机、工业相机。其中,如果选用单目相机或单目摄像机,在实际应用中,则采用两个单目相机或单目摄像机。The above-mentioned image acquisition unit includes but is not limited to a monocular camera, a binocular camera, a monocular camera, a binocular camera, and an industrial camera. Wherein, if a monocular camera or a monocular camera is selected, in practical applications, two monocular cameras or monocular cameras are used.

本发明实施例提供一种基于光路采集的移动裸眼三维虚拟现实的系统。如图1所示,该系统10包括:图像获取单元12、移动终端14和控制器16。其中,图像获取单元12用于采集图像,并将图像发送至控制器。移动终端14包括屏幕,该屏幕上设有微透镜阵列膜,用于将加速度的位置信息和角速度的旋转信息发送至控制器,并通过屏幕显示三维虚拟现实显示图像。控制器16分别与图像获取单元和移动终端通信连接;用于对图像进行处理,得到两视点合成图索引图,且根据加速度的位置信息和角速度的旋转信息,获得两视点图,并基于该图像的两视点合成图索引图,对两视点图渲染出合成图,得到三维虚拟现实显示图像,以及将三维虚拟现实显示图像发送至移动终端。An embodiment of the present invention provides a mobile naked-eye three-dimensional virtual reality system based on optical path acquisition. As shown in FIG. 1 , the system 10 includes: an image acquisition unit 12 , a mobile terminal 14 and a controller 16 . Wherein, the image acquisition unit 12 is used to acquire images and send the images to the controller. The mobile terminal 14 includes a screen, on which a microlens array film is arranged, for sending the position information of the acceleration and the rotation information of the angular velocity to the controller, and displaying a three-dimensional virtual reality display image through the screen. The controller 16 is communicated with the image acquisition unit and the mobile terminal respectively; it is used to process the image to obtain a two-viewpoint synthetic map index map, and obtain a two-viewpoint map according to the position information of the acceleration and the rotation information of the angular velocity, and based on the image The two-viewpoint synthetic image index image is used to render the composite image for the two-viewpoint image to obtain a three-dimensional virtual reality display image, and to send the three-dimensional virtual reality display image to a mobile terminal.

该移动终端包括但不限于手机、个人数字助理。在本优选实施例中,将移动三维显示技术与虚拟现实技术相结合,利用移动终端可以实现可移动的裸眼三维虚拟现实显示。The mobile terminal includes, but is not limited to, a mobile phone and a personal digital assistant. In this preferred embodiment, the mobile three-dimensional display technology is combined with the virtual reality technology, and a mobile terminal can be used to realize a mobile naked-eye three-dimensional virtual reality display.

上述图像获取单元作为光路采集设备,其包括但不限于单目相机、单目摄像机、双目相机和双目摄像机。该图像获取单元可以与移动终端之间进行前期的校正和标定,之后进行三维虚拟现实的显示。The above-mentioned image acquisition unit, as an optical path acquisition device, includes but is not limited to a monocular camera, a monocular camera, a binocular camera and a binocular camera. The image acquisition unit can perform pre-calibration and calibration with the mobile terminal, and then perform three-dimensional virtual reality display.

上述控制器包括但不限于计算机、笔记本电脑、工控机、服务器。该服务器也可以为服务器集群。The aforementioned controllers include but are not limited to computers, notebook computers, industrial computers, and servers. The server can also be a server cluster.

在移动终端的屏幕上覆盖微透镜阵列膜,该微透镜阵列膜由成千上万个凸透镜组成。这样,在不同的方向上观看时,位于凸透镜下方的不同地方的子像素被放大到观察者眼中。A microlens array film is covered on the screen of the mobile terminal, and the microlens array film is composed of tens of thousands of convex lenses. In this way, when viewed in different directions, the sub-pixels located at different places under the convex lens are magnified into the viewer's eyes.

在本实施例中,控制器分别与图像获取单元和移动终端进行通信连接的方式包括但不限于WIFI网络、蓝牙、ZigBee、2G、3G、4G、5G。In this embodiment, the ways in which the controller communicates with the image acquisition unit and the mobile terminal respectively include but are not limited to WIFI network, Bluetooth, ZigBee, 2G, 3G, 4G, and 5G.

本发明实施例通过采用上述技术方案,能够在微透镜阵列和移动终端液晶屏幕的参数未知的情况下,实现复制性强、成本可控并且分辨率高的移动三维虚拟现实显示。By adopting the above-mentioned technical solution, the embodiment of the present invention can realize mobile three-dimensional virtual reality display with high reproducibility, controllable cost and high resolution under the condition that the parameters of the microlens array and the liquid crystal screen of the mobile terminal are unknown.

在一个可选的实施例中,移动终端可以包括陀螺仪和加速度计。其中,加速度计用于获取加速度的位置信息。陀螺仪用于获取角速度的旋转信息。In an optional embodiment, the mobile terminal may include a gyroscope and an accelerometer. Among them, the accelerometer is used to obtain the position information of the acceleration. Gyroscopes are used to obtain rotation information of angular velocity.

下面结合图2和图3以优选实施例的方式,来对基于光路采集的移动裸眼三维虚拟现实的系统的工作过程进行详细说明。本优选实施例以手机、相机、控制计算机分别作为移动终端、图像获取单元和控制器为例进行详细说明。其中,采用两个单目工业相机。手机屏幕上覆盖微透镜阵列膜。控制计算机通过路由器与手机和工业相机实现无线通信。The working process of the mobile naked-eye three-dimensional virtual reality system based on light path acquisition will be described in detail below in combination with FIG. 2 and FIG. 3 in the form of a preferred embodiment. In this preferred embodiment, a mobile phone, a camera, and a control computer are respectively used as a mobile terminal, an image acquisition unit, and a controller to describe in detail. Among them, two monocular industrial cameras are used. The mobile phone screen is covered with a microlens array film. The control computer realizes wireless communication with the mobile phone and the industrial camera through the router.

本优选实施例进行如下距离设置:测量观看使用者的双目距离、手持手机3时屏幕距离眼睛的距离,将两个相机6水平放置间距设置为用户7双目的间距,将手机屏幕与相机6的距离和用户7手持手机3的距离保持一致。This preferred embodiment carries out the following distance settings: measure the binocular distance of watching the user, the distance between the screen and the eyes when holding the mobile phone 3, two cameras 6 horizontal placement distances are set to the user 7 binocular distances, and the mobile phone screen and the camera 6 is consistent with the distance at which the user 7 holds the mobile phone 3 .

在光路采集实验盒4中,通过手机3经由光线5对工业相机6进行校正和标定。其中,标定方法包括但不限于主动视觉相机标定方法和相机自标定法。In the light path acquisition experiment box 4, the industrial camera 6 is calibrated and calibrated through the light 5 through the mobile phone 3. Wherein, the calibration method includes but not limited to an active vision camera calibration method and a camera self-calibration method.

将工业相机6连接到控制计算机1并进行图像采集,优化出两张初始合成图索引图,其效果是两张索引图在手机3中的显示效果是黑白颜色互补的。Connect the industrial camera 6 to the control computer 1 for image acquisition, and optimize the two initial synthetic image index images. The effect is that the display effect of the two index images in the mobile phone 3 is complementary in black, white and color.

通过手机3采集观看者的位移和旋转姿态,并通过WIFI网络与控制计算机1进行信号通讯,将观看者的位移和旋转姿态发送至控制计算机1。The mobile phone 3 collects the viewer's displacement and rotation posture, and communicates with the control computer 1 through the WIFI network, and sends the viewer's displacement and rotation posture to the control computer 1 .

控制计算机1分别与手机3和工业相机6通过路由器2进行信号通讯,控制手机点亮单个像素点,同时控制相机6采集图像;还对两个工业相机6采集到的图像进行优化处理生成两视点图;以及还接收手机3发来的观看者的位移和旋转姿态数据,且根据观看者的位移和旋转姿态数据,将要显示的两视点图和初始合成图索引图相互作用生成虚拟三维场景图像(三维虚拟现实图像),并将该虚拟三维场景图像反馈至手机3,进行裸眼三维显示。此时,观看者就能在不需要任何其他外界辅助设备的情况下观看到虚拟场景的三维效果。The control computer 1 performs signal communication with the mobile phone 3 and the industrial camera 6 respectively through the router 2, controls the mobile phone to light up a single pixel, and controls the camera 6 to collect images at the same time; also optimizes the images collected by the two industrial cameras 6 to generate two viewpoints and also receive the viewer's displacement and rotation attitude data sent by the mobile phone 3, and according to the viewer's displacement and rotation attitude data, the two-viewpoint figure to be displayed and the initial synthetic figure index figure interact to generate a virtual three-dimensional scene image ( 3D virtual reality image), and the virtual 3D scene image is fed back to the mobile phone 3 for naked-eye 3D display. At this time, the viewer can watch the three-dimensional effect of the virtual scene without any other external auxiliary equipment.

图3示例性地示出了用户手持手机在三个不同观察点的手机裸眼三维虚拟现实显示效果示意图。FIG. 3 exemplarily shows a schematic diagram of naked-eye three-dimensional virtual reality display effects of the mobile phone at three different viewing points when the user holds the mobile phone.

本优选实施例构建了一个复制性强、成本可控并且分辨率高、可移动的基于光路采集的裸眼三维虚拟现实的系统,其融入了虚拟现实技术,作为基于光路采集的自动校准和减少串扰的移动三维显示的方案,能够在微透镜阵列和手机液晶屏幕的参数未知的情况下,呈现出高质量、细腻的三维显示效果,能够针对个人手持手机的差异做个性化定制,相比于传统虚拟现实设备,减少了头盔的重量和舒适性欠佳的束缚感,同时增加了设备自身三维显示效果而非仅仅使用双目接收左右不同场景图片达到三维的效果。This preferred embodiment constructs a naked-eye three-dimensional virtual reality system based on optical path acquisition with strong reproducibility, controllable cost, high resolution, and mobility, which incorporates virtual reality technology as an automatic calibration and crosstalk reduction based on optical path acquisition. The mobile 3D display solution can present a high-quality, delicate 3D display effect when the parameters of the microlens array and the LCD screen of the mobile phone are unknown, and it can be customized according to the differences in personal mobile phones. Compared with the traditional The virtual reality device reduces the weight of the helmet and the sense of restraint caused by poor comfort, and at the same time increases the three-dimensional display effect of the device itself instead of just using binoculars to receive pictures of different scenes on the left and right to achieve a three-dimensional effect.

此外,本发明实施例提供一种基于光路采集的移动裸眼三维虚拟现实的方法。该方法可以应用于上述基于光路采集的移动裸眼三维虚拟现实的系统。如图4所示,该方法可以包括:In addition, an embodiment of the present invention provides a method for moving naked-eye three-dimensional virtual reality based on optical path acquisition. The method can be applied to the above-mentioned mobile naked-eye three-dimensional virtual reality system based on optical path acquisition. As shown in Figure 4, the method may include:

S400:采集图像。S400: Collect images.

S410:基于采集到的图像获得两视点的合成图索引图。S410: Obtain a composite map index map of two viewpoints based on the collected images.

具体地,本步骤可以通过步骤S414至步骤S418来实现。Specifically, this step may be implemented through steps S414 to S418.

S414:确定图像上子像素亮度影响权重。S414: Determine the influence weight of sub-pixel brightness on the image.

本步骤针对两视点的合成图索引图上的每一个子像素的值都考虑了周边子像素对其的影响,并通过采集到的图像得到不同子像素影响的权重值。In this step, the influence of surrounding subpixels is taken into account for the value of each subpixel on the composite image index map of two viewpoints, and the weight value of different subpixel influences is obtained through the collected images.

举例来说,当位于Imagem(i,j)窗口里的任何一个子像素Imagem(i,j,q)被点亮时,都会影响到捕获图像上、Imagec(t)(i,j)像素里的每一个子像素。For example, when any sub-pixel Image m (i, j, q) located in the Image m (i, j) window is lit, it will affect the captured image, Image c(t) (i, j ) for each sub-pixel in the pixel.

具体地,本步骤可以进一步包括:根据下式确定图像上子像素亮度影响权重:Specifically, this step may further include: determining the sub-pixel brightness influence weight on the image according to the following formula:

其中,Imagec(t)(i,j,k)表示在第t个视点处采集到的图像上的子像素,t=0、1,i=1~H,j=1~W;H表示图像的高度;W表示图像的宽度;r表示安全窗口半径;m、n表示以安全窗口中心为原点的窗口坐标;q=1~3;Imagem(i+m,j+n,q)表示图像上的子像素;表示图像上子像素(i+m,j+n,q)点亮到最大亮度时对邻近像素(i,j,k)经过归一化后的亮度影响权重。Among them, Image c(t) (i, j, k) represents the sub-pixel on the image collected at the t-th viewpoint, t=0, 1, i=1~H, j=1~W; H represents The height of the image; W represents the width of the image; r represents the radius of the security window; m and n represent the window coordinates with the center of the security window as the origin; q=1~3; Image m (i+m,j+n,q) represents sub-pixels on the image; Indicates the normalized brightness influence weight of adjacent pixels (i, j, k) when the sub-pixel (i+m, j+n, q) on the image is lit to the maximum brightness.

根据本步骤得到的子像素亮度影响权重可以索引出权重矩阵,以用于后续的处理。According to the sub-pixel brightness influence weight obtained in this step, a weight matrix can be indexed for subsequent processing.

S416:基于图像上子像素亮度影响权重,利用带有包围盒的最小二乘方法,确定两视点的合成图索引图。S416: Based on the influence weight of sub-pixel brightness on the image, using the method of least squares with bounding boxes, determine the composite image index map of the two viewpoints.

具体地,本步骤可以进一步包括:基于图像上子像素亮度影响权重,利用带有包围盒的最小二乘方法,根据下式确定两视点的合成图索引图:Specifically, this step may further include: based on the influence weight of sub-pixel brightness on the image, using the least squares method with bounding boxes, according to the following formula to determine the composite map index map of the two viewpoints:

其中,表示两视点对应子像素亮度影响权重的差异;I表示两视点的合成图索引图,且被限定于包围盒BI={I∈Z1×H:L≤I≤U}中,L=0×I1×H且U=255×I1×H,I1×H表示H维的全1列向量;表示两视点采集图像的差异;H表示图像的高度。in, Indicates the difference in the weights of sub-pixel brightness influences corresponding to two viewpoints; I represents the composite image index map of two viewpoints, and is limited to the bounding box B I ={I∈Z 1×H :L≤I≤U}, L=0 ×I 1×H and U=255×I 1×H , I 1×H represents a full 1-column vector of H dimension; Indicates the difference between images collected from two viewpoints; H indicates the height of the image.

作为示例,通过以下方式确定第0视点的合成图索引图I0As an example, the composite map index map I 0 of the 0th viewpoint is determined in the following manner:

步骤1:将I限定于包围盒BI={I∈Z1×H:L≤I≤U}中,其中,L=0×I1×H且U=255×I1×H,I1×H表示H维的全1列向量。Step 1: Confine I to the bounding box B I ={I∈Z 1×H :L≤I≤U}, where L=0×I 1×H and U=255×I 1×H , I 1 ×H represents an H-dimensional full 1-column vector.

步骤2:根据下式进行带有包围盒的最小二乘处理:Step 2: Perform least squares processing with bounding boxes according to the following formula:

其中,表示0视点和1视点捕获图像的差异, 表示两视点对应子像素亮度影响权重的差异,其为(H×W×3,H×W)×3大小的满秩稀疏矩阵,H表示图像的高度,W表示图像的宽度;I表示视点的合成图索引图。in, Indicates the difference between images captured at 0 viewpoint and 1 viewpoint, Indicates the difference in the weight of sub-pixel brightness influences corresponding to two viewpoints, which is a full-rank sparse matrix of size (H×W×3, H×W)×3, H represents the height of the image, W represents the width of the image; I represents the Composite graph index graph.

同理我们可以求得第1视点的合成图索引图。In the same way, we can obtain the composite map index map of the first viewpoint.

在一个优选的实施例中,本发明实施例在步骤S414之前还可以包括:In a preferred embodiment, before step S414, the embodiment of the present invention may further include:

S412:对图像进行反畸变和兴趣区域提取操作处理。S412: Perform anti-distortion and ROI extraction operations on the image.

举例来说,考虑到移动终端屏幕上距离为N(N取整数)个像素的两个像素之间的亮度不会相互产生影响,所以可以将移动终端屏幕被分成了若干个N×N大小的安全窗口。故,向移动终端传输N×N×3张测试图imagem即可进行实施。其中,测试图的分辨率为H×W。其中,H表示测试图的高度;W表示测试图的宽度。在具体实施过程中,利用位于两视点观测点的图像获取单元分别进行捕获,得到捕获图像,并对捕获图像进行反畸变、兴趣区域提取操作,从而得到分辨率为H×W的imagec(0)图像和分辨率为H×W的imagec(1)图像。其中,图像获取单元包括但不限于单目相机、单目摄像机、双目相机和双目摄像机、工业相机。For example, considering that the brightness of two pixels on the screen of the mobile terminal whose distance is N (where N is taken as an integer) pixels will not affect each other, the screen of the mobile terminal can be divided into several N×N sized security window. Therefore, it can be implemented by transmitting N×N×3 test images image m to the mobile terminal. Wherein, the resolution of the test chart is H×W. Among them, H represents the height of the test pattern; W represents the width of the test pattern. In the specific implementation process, the image acquisition units located at the observation points of the two viewpoints are used to capture separately to obtain the captured image, and the captured image is subjected to anti-distortion and region-of-interest extraction operations to obtain an image c(0 ) image and image c(1) image with resolution H×W. Wherein, the image acquisition unit includes but not limited to a monocular camera, a monocular camera, a binocular camera, a binocular camera, and an industrial camera.

S420:根据加速度的位置信息和角速度的旋转信息,获得两视点图。S420: Obtain a two-viewpoint map according to the position information of the acceleration and the rotation information of the angular velocity.

具体地,本步骤可以通过以下方式来实现:根据加速度的位置信息和角速度的旋转信息,调整采集图像的方位,获得两视点图。Specifically, this step may be implemented in the following manner: according to the position information of the acceleration and the rotation information of the angular velocity, the orientation of the collected image is adjusted to obtain a two-viewpoint image.

其中,加速度的位置信息可以通过移动终端的加速度计传感器进行采集而得到。角速度的旋转信息可以通过移动终端的陀螺仪传感器进行采集而得到。由此,可以采集到观看者的位移和旋转姿态,从而,利用该位移和旋转姿态实现移动裸眼三维虚拟现实的显示。Wherein, the location information of the acceleration can be obtained by collecting the accelerometer sensor of the mobile terminal. The rotation information of the angular velocity can be obtained through collection by the gyroscope sensor of the mobile terminal. Thus, the displacement and rotation posture of the viewer can be collected, so that the display of the mobile naked-eye three-dimensional virtual reality can be realized by using the displacement and rotation posture.

在实际应用中,可以通过由手机、相机和计算机搭建的平台,通过步骤S420在计算机端调整虚拟相机的位姿,获得两视点图。In practical applications, through the platform built by the mobile phone, the camera and the computer, the pose of the virtual camera can be adjusted on the computer side through step S420 to obtain a two-viewpoint map.

S430:基于图像的两视点合成图索引图,对两视点图渲染出合成图,得到三维虚拟现实显示图像。S430: Based on the image-based two-viewpoint synthetic image index image, render the composite image for the two-viewpoint image to obtain a three-dimensional virtual reality display image.

举例来说,根据下式渲染出合成图,得到三维虚拟现实显示图像:For example, the composite image is rendered according to the following formula to obtain a three-dimensional virtual reality display image:

Syn=I0×view0+I1×view1Syn=I 0 ×view 0 +I 1 ×view 1 ;

其中,Syn表示三维虚拟现实显示图像;I0表示第0视点的合成图索引图;I1表示第1视点的合成图索引图;view0表示第0视点图;view1表示第1视点图。Among them, Syn represents the 3D virtual reality display image; I 0 represents the synthetic map index map of the 0th viewpoint; I 1 represents the synthetic map index map of the 1st viewpoint; view 0 represents the 0th viewpoint map; view 1 represents the 1st viewpoint map.

在一个优选的实施例中,本步骤S430还可以通过以下方式来实现:将两视点图与两张显示效果为黑白互补的预定合成图索引图进行合成,得到三维虚拟现实显示图像。In a preferred embodiment, this step S430 can also be implemented in the following manner: combining the two-viewpoint images with two predetermined composite image index images whose display effects are black and white and complement each other to obtain a 3D virtual reality display image.

上述实施例中虽然将各个步骤按照上述先后次序的方式进行了描述,但是本领域技术人员可以理解,为了实现本实施例的效果,不同的步骤之间不必按照这样的次序执行,其可以同时(并行)执行或以颠倒的次序执行,这些简单的变化都在本发明的保护范围之内。In the above embodiment, although the various steps are described according to the above sequence, those skilled in the art can understand that in order to achieve the effect of this embodiment, different steps do not have to be executed in this order, and they can be performed at the same time ( Parallel) execution or execution in reversed order, these simple changes are all within the protection scope of the present invention.

需要说明的是,在对本发明各个实施例的描述过程中,出于简明的考虑,省略了相同的部分,本领域技术人员应能理解,在不出现冲突的情况下,对一个实施例的说明也可以应用于另一个实施例。It should be noted that, in the process of describing various embodiments of the present invention, for the sake of brevity, the same parts are omitted, and those skilled in the art should understand that, in the absence of conflict, the description of an embodiment It can also be applied to another embodiment.

还应当注意,本说明书中使用的语言主要是为了可读性和教导的目的,并不是为了解释或者限定本发明的保护范围。It should also be noted that the language used in this specification is mainly for the purpose of readability and teaching, and is not intended to explain or limit the protection scope of the present invention.

以上对本发明的示例实施例的详细描述是为了说明和描述的目的而提供。不是为了穷尽或将本发明限制为所描述的精确形式。显然,许多变型和改变对本领域技术人员而言是显而易见的。实施例的选择和描述是为了最佳地说明本发明的原理及其实际应用,从而使本领域其他技术人员能够理解本发明的各种实施例和适于特定使用预期的各种变型。本发明的实施例可以省略上述技术特征中的一些技术特征,仅解决现有技术中存在的部分技术问题。而且,所描述的技术特征可以进行任意组合。本发明的保护范围由所附权利要求及其等价物来限定,本领域技术其他人员可以对所附权利要求中所描述的技术方案进行各种变型或替换和组合,这些更改或替换之后的技术方案都将落入本发明的保护范围之内。The foregoing detailed description of example embodiments of the present invention has been presented for purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms described. Obviously, many modifications and changes will be apparent to those skilled in the art. The embodiment was chosen and described in order to best explain the principles of the invention and its practical application, thereby enabling others skilled in the art to understand the invention for various embodiments and with various modifications as are suited to the particular use contemplated. Embodiments of the present invention may omit some of the above-mentioned technical features, and only solve some technical problems existing in the prior art. Moreover, the described technical features can be combined arbitrarily. The scope of protection of the present invention is defined by the appended claims and their equivalents. Others skilled in the art can make various modifications or replacements and combinations of the technical solutions described in the appended claims. The technical solutions after these changes or replacements All will fall within the protection scope of the present invention.

Claims (12)

1. A method for moving naked eye three-dimensional virtual reality based on optical path acquisition is characterized by comprising the following steps:
collecting an image;
acquiring a two-viewpoint composite map index map based on the acquired image;
obtaining two viewpoint images according to the position information of the acceleration and the rotation information of the angular velocity;
rendering a synthetic graph to the two-viewpoint graph based on the two-viewpoint synthetic graph index graph of the image to obtain a three-dimensional virtual reality display image.
2. The method for moving naked eye three-dimensional virtual reality based on optical path acquisition according to claim 1, wherein the obtaining of the two-viewpoint synthetic image index map based on the acquired image specifically comprises:
determining a sub-pixel luminance impact weight on the image;
and determining the composite map index map of the two viewpoints by using a least square method with bounding boxes based on the sub-pixel brightness influence weight on the image.
3. The method for moving naked eye three-dimensional virtual reality based on optical path acquisition according to claim 2, wherein the determining the sub-pixel brightness influence weight on the image specifically comprises:
determining the sub-pixel luminance impact weight on the image according to:
Image c ( t ) ( i , j , k ) = Σ m = - r r Σ n = - r r Σ q = 1 3 w i + m , j + n , q ( t ) ( - m , - n , k ) Image m ( i + m , j + n , q )
wherein the Image isc(t)(i, j, k) represents a sub-pixel on the image acquired at the t-th viewpoint, t being 0, 1, i being 1 to H, and j being 1 to W; the H represents the height of the image; the W represents a width of the image; the r represents a security window radius; the m and the n represent window coordinates with the center of the safety window as an origin; the q is 1-3; the Imagem(i + m, j + n, q) represents a sub-pixel on the image; the above-mentionedRepresenting the normalized luminance impact weight for the neighboring pixel (i, j, k) when the sub-pixel (i + m, j + n, q) on the image is lit to maximum luminance.
4. The method for moving naked eye three-dimensional virtual reality based on optical path acquisition according to claim 2, wherein the determining the composite map index map of the two viewpoints by using a least square method with bounding boxes based on the sub-pixel brightness influence weight on the image specifically comprises:
determining the composite map index map for the two viewpoints using a least squares with bounding boxes based on sub-pixel luminance impact weights on the image according to the following equation:
▿ w × I = ▿ I c ;
wherein, theRepresenting the difference of the brightness influence weights of the sub-pixels corresponding to the two viewpoints; the I represents the composite map index map of the two viewpoints and is limited to a bounding box BI={I∈Z1×HL is less than or equal to I and less than or equal to U, L is 0 × I1×HAnd U255 × I1×HSaid I is1×HA full 1-column vector representing the H dimension; the above-mentionedRepresenting the difference of the two viewpoint collected images; the H represents the height of the image.
5. The method for moving naked eye three-dimensional virtual reality based on optical path acquisition according to claim 2, wherein before the step of determining the sub-pixel brightness influence weight on the image, the method further comprises:
and carrying out anti-distortion and interest region extraction operation processing on the image.
6. The method for moving naked eye three-dimensional virtual reality based on optical path acquisition according to claim 1, wherein the obtaining of the two-viewpoint images according to the position information of the acceleration and the rotation information of the angular velocity specifically comprises:
and adjusting the orientation of the acquired image according to the position information of the acceleration and the rotation information of the angular velocity to obtain the two-viewpoint image.
7. The method for moving naked eye three-dimensional virtual reality based on optical path acquisition according to claim 1, wherein rendering a composite map to obtain a three-dimensional virtual reality display image based on the two-viewpoint composite map index map of the image specifically comprises:
rendering a synthetic graph according to the following formula to obtain the three-dimensional virtual reality display image:
Syn=I0×view0+I1×view1
wherein the Syn represents the three-dimensional virtual reality display image; said I0A composite map index map indicating the 0 th view; said I1A composite map index map indicating the 1 st view; the view is0Represents a 0 th viewpoint map; the view is1The 1 st view is shown.
8. A system for moving naked eye three-dimensional virtual reality based on optical path acquisition is characterized by comprising:
the image acquisition unit is used for acquiring an image and sending the image to the controller;
the mobile terminal comprises a screen, wherein a micro-lens array film is arranged on the screen and used for sending position information of acceleration and rotation information of angular velocity to the controller and displaying a three-dimensional virtual reality display image through the screen;
the controller is in communication connection with the image acquisition unit and the mobile terminal respectively; the two-viewpoint synthetic graph index graph is obtained according to the position information of the acceleration and the rotation information of the angular velocity, a synthetic graph is rendered for the two-viewpoint synthetic graph based on the two-viewpoint synthetic graph index graph of the image, a three-dimensional virtual reality display image is obtained, and the three-dimensional virtual reality display image is sent to the mobile terminal.
9. The system for moving naked eye three-dimensional virtual reality based on optical path acquisition according to claim 8, wherein the mobile terminal comprises:
the accelerometer is used for acquiring position information of the acceleration;
and the gyroscope is used for acquiring the rotation information of the angular velocity.
10. The system for moving naked eye three-dimensional virtual reality based on optical path acquisition according to claim 8, wherein the image acquisition unit is a monocular camera, a monocular video camera, a binocular camera or a binocular video camera.
11. The system for mobile naked eye three-dimensional virtual reality based on optical path acquisition according to claim 8, wherein the mobile terminal is a mobile phone or a personal digital assistant.
12. The system for moving naked eye three-dimensional virtual reality based on light path acquisition according to claim 8, wherein the controller is a computer, a notebook computer, a server or an industrial personal computer.
CN201611202874.0A 2016-12-23 2016-12-23 The method and system of mobile naked eye three-dimensional virtual reality based on optical path acquisition Expired - Fee Related CN106878698B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611202874.0A CN106878698B (en) 2016-12-23 2016-12-23 The method and system of mobile naked eye three-dimensional virtual reality based on optical path acquisition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611202874.0A CN106878698B (en) 2016-12-23 2016-12-23 The method and system of mobile naked eye three-dimensional virtual reality based on optical path acquisition

Publications (2)

Publication Number Publication Date
CN106878698A true CN106878698A (en) 2017-06-20
CN106878698B CN106878698B (en) 2019-05-24

Family

ID=59164901

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611202874.0A Expired - Fee Related CN106878698B (en) 2016-12-23 2016-12-23 The method and system of mobile naked eye three-dimensional virtual reality based on optical path acquisition

Country Status (1)

Country Link
CN (1) CN106878698B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108391117A (en) * 2018-02-01 2018-08-10 周金润 A kind of mobile phone bore hole 3D display technology based on viewpoint positioning, single-view relief painting
CN108769642A (en) * 2018-03-27 2018-11-06 深圳负空间科技有限公司 An action shot based on high in the clouds renders and bore hole VR exchange methods
CN112184878A (en) * 2020-10-15 2021-01-05 洛阳众智软件科技股份有限公司 Method, device and equipment for automatically generating and rendering three-dimensional night scene light
CN112313699A (en) * 2019-05-24 2021-02-02 京东方科技集团股份有限公司 Method and device for controlling virtual reality display device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103019021A (en) * 2012-12-27 2013-04-03 Tcl集团股份有限公司 3D (three-dimensional) light field camera and method for processing images shot by same
CN103702099A (en) * 2013-12-17 2014-04-02 四川大学 Ultra-large visual-angle integrated-imaging 3D(Three-Dimensional)displaying method based on head tracking
US20160246165A1 (en) * 2015-02-25 2016-08-25 Delta Electronics, Inc. Three-dimension light field construction apparatus
CN106154567A (en) * 2016-07-18 2016-11-23 北京邮电大学 The imaging method of a kind of 3 d light fields display system and device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103019021A (en) * 2012-12-27 2013-04-03 Tcl集团股份有限公司 3D (three-dimensional) light field camera and method for processing images shot by same
CN103702099A (en) * 2013-12-17 2014-04-02 四川大学 Ultra-large visual-angle integrated-imaging 3D(Three-Dimensional)displaying method based on head tracking
US20160246165A1 (en) * 2015-02-25 2016-08-25 Delta Electronics, Inc. Three-dimension light field construction apparatus
CN106154567A (en) * 2016-07-18 2016-11-23 北京邮电大学 The imaging method of a kind of 3 d light fields display system and device

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108391117A (en) * 2018-02-01 2018-08-10 周金润 A kind of mobile phone bore hole 3D display technology based on viewpoint positioning, single-view relief painting
CN108769642A (en) * 2018-03-27 2018-11-06 深圳负空间科技有限公司 An action shot based on high in the clouds renders and bore hole VR exchange methods
CN112313699A (en) * 2019-05-24 2021-02-02 京东方科技集团股份有限公司 Method and device for controlling virtual reality display device
CN112184878A (en) * 2020-10-15 2021-01-05 洛阳众智软件科技股份有限公司 Method, device and equipment for automatically generating and rendering three-dimensional night scene light
CN112184878B (en) * 2020-10-15 2023-08-25 洛阳众智软件科技股份有限公司 Method, device and equipment for automatic generation and rendering of 3D night scene lights

Also Published As

Publication number Publication date
CN106878698B (en) 2019-05-24

Similar Documents

Publication Publication Date Title
CN102098524B (en) Tracking type stereo display device and method
CN101636747B (en) Two dimensional/three dimensional digital information acquisition and display device
CN106251403B (en) A kind of methods, devices and systems of virtual three-dimensional Scene realization
CN107071382B (en) Stereoscopic image display device
JP6023801B2 (en) Simulation device
WO2019105323A1 (en) Display module, head-mounted display device, and stereoscopic image display method and apparatus
CN105704479B (en) The method and system and display equipment of the measurement human eye interpupillary distance of 3D display system
EP3350989B1 (en) 3d display apparatus and control method thereof
KR101675961B1 (en) Apparatus and Method for Rendering Subpixel Adaptively
CN108259883B (en) Image processing method, head-mounted display, and readable storage medium
CN107105216B (en) Three-dimensional light field display device with continuous parallax and wide viewing angle based on pinhole array
CN105611278A (en) Image processing method and system for preventing naked eye 3D viewing dizziness and display device
US9197877B2 (en) Smart pseudoscopic-to-orthoscopic conversion (SPOC) protocol for three-dimensional (3D) display
CN205610834U (en) Stereo display system
CN106878698B (en) The method and system of mobile naked eye three-dimensional virtual reality based on optical path acquisition
CN106507096B (en) A kind of tracing type ground light field 3D display method and system with super large visual angle
WO2019076348A1 (en) Virtual reality (vr) interface generation method and apparatus
CN108076208B (en) Display processing method and device and terminal
CN105898287A (en) Device and method for machine visual analysis based on naked-eye stereoscopic display
CN102998095B (en) The detection method of a kind of naked-eye stereoscopic display and device
CN105676470A (en) 3D scene vision spatial resolution enhancing method and system
Boulos et al. Web GIS in practice VII: stereoscopic 3-D solutions for online maps and virtual globes
CN101655747B (en) Three-dimensional mouse
CN112532964B (en) Image processing method, device, apparatus and readable storage medium
CN115174886A (en) Mobile phone lens supporting stereoscopic virtual reality and augmented reality and display method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20190524