[go: up one dir, main page]

CN105812776A - Stereoscopic display system based on soft lens and method - Google Patents

Stereoscopic display system based on soft lens and method Download PDF

Info

Publication number
CN105812776A
CN105812776A CN201410852264.XA CN201410852264A CN105812776A CN 105812776 A CN105812776 A CN 105812776A CN 201410852264 A CN201410852264 A CN 201410852264A CN 105812776 A CN105812776 A CN 105812776A
Authority
CN
China
Prior art keywords
unit
image
point
locus
labelling point
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201410852264.XA
Other languages
Chinese (zh)
Inventor
何建行
刘君
包瑞
邵文龙
郁树达
邢旭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Mingyi Medical Charitable Foundation
SuperD Co Ltd
First Affiliated Hospital of Guangzhou Medical University
Original Assignee
Guangdong Mingyi Medical Charitable Foundation
SuperD Co Ltd
First Affiliated Hospital of Guangzhou Medical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Mingyi Medical Charitable Foundation, SuperD Co Ltd, First Affiliated Hospital of Guangzhou Medical University filed Critical Guangdong Mingyi Medical Charitable Foundation
Priority to CN201410852264.XA priority Critical patent/CN105812776A/en
Publication of CN105812776A publication Critical patent/CN105812776A/en
Pending legal-status Critical Current

Links

Landscapes

  • Length Measuring Devices By Optical Means (AREA)

Abstract

本发明属于医疗技术领域,提供一种基于软镜的立体显示系统及方法,该系统包括:显示单元、分光单元、跟踪设备及图像拍摄单元,该分光单元用于将该显示单元显示的图像空间上分成左视图和右视图,该跟踪设备用于获取第一目标对象的位置信息,该图像拍摄单元用于拍摄第二目标对象,该基于软镜的立体显示系统还包括图像播放处理设备,依据该第一目标对象的位置信息、该分光单元的光栅参数及该显示单元的显示参数,将接收到的该图像拍摄单元拍摄到的立体图像进行实时处理,处理后发送该显示单元实时显示。本发明的图像播放速度相比于现有技术大大提高,可满足实时立体显示的要求,具有方便医生手术操作及辅助医生提高手术成功率的优点。

The invention belongs to the field of medical technology, and provides a stereoscopic display system and method based on a soft mirror. The system includes: a display unit, a light splitting unit, a tracking device, and an image capturing unit. The light splitting unit is used to display the image space of the display unit. The upper part is divided into a left view and a right view, the tracking device is used to obtain the position information of the first target object, the image capture unit is used to capture the second target object, the stereoscopic display system based on the soft mirror also includes an image playback processing device, according to The position information of the first target object, the grating parameters of the spectroscopic unit and the display parameters of the display unit are processed in real time on the received stereoscopic image captured by the image capture unit, and then sent to the display unit for real-time display after processing. Compared with the prior art, the image playback speed of the present invention is greatly improved, can meet the requirement of real-time stereoscopic display, and has the advantages of being convenient for doctors to operate and assisting doctors to improve the success rate of operations.

Description

基于软镜的立体显示系统及方法Stereoscopic display system and method based on soft mirror

技术领域technical field

本发明涉及医疗设备技术领域,具体而言,涉及一种应用于临床医学上的基于软镜的立体显示系统及显示方法。The invention relates to the technical field of medical equipment, in particular to a soft mirror-based stereoscopic display system and display method applied in clinical medicine.

背景技术Background technique

内窥镜是一个配备有灯光的管子,它可以经口腔进入胃内或经其他孔道进入体内。利用内窥镜可以看到X射线不能显示的病变,因此它对医生非常有用。例如,借助内窥镜医生可以观察胃内的溃疡或肿瘤,据此制定出最佳的治疗方案。内窥镜是一种光学仪器,是由冷光源镜头、纤维光导线、图像传输系统、屏幕显示系统等组成,它能扩大手术视野。使用内窥镜的突出特点是手术切口小,切口瘢痕不明显,术后反应轻,出血、青紫和肿胀时间可大大减少,恢复也较传统手术快,非常符合美容外科美丽不留痕的要求。An endoscope is a tube equipped with a light that is passed through the mouth into the stomach or through other openings in the body. The endoscope can be used to see lesions that cannot be seen on X-rays, so it is very useful for doctors. For example, with the help of an endoscope, doctors can observe ulcers or tumors in the stomach and formulate the best treatment plan accordingly. Endoscope is an optical instrument, which is composed of cold light source lens, fiber optic wire, image transmission system, screen display system, etc. It can expand the surgical field of view. The prominent feature of using endoscope is that the surgical incision is small, the scar of the incision is not obvious, the postoperative reaction is mild, the time of bleeding, bruising and swelling can be greatly reduced, and the recovery is faster than traditional surgery, which is very in line with the requirements of beauty and no scars in cosmetic surgery.

内窥镜分为硬质材质和软质材质两种,称为医疗硬镜和软镜。最初的内窥镜是用硬质管做成的,发明于100多年前。虽然它们逐渐有所改进,但仍然未能被广泛使用。后来,在20世纪50年代内窥镜用软质管制作,因而能在人体内的拐角处轻易地弯曲。Endoscopes are divided into hard and soft materials, called medical hard mirrors and soft mirrors. The first endoscopes were made of rigid tubes and were invented more than 100 years ago. Although they have gradually improved, they are still not widely used. Later, in the 1950s, endoscopes were made of flexible tubes that could easily bend around the corners of the body.

医疗软镜包括胃镜、结肠镜、十二指肠镜、支气管镜、鼻咽喉镜、胆道镜等。Medical flexible mirrors include gastroscopes, colonoscopes, duodenoscopes, bronchoscopes, nasopharyngoscopes, choledochoscopes, etc.

目前市面上少数现有的医疗内窥镜具有显示3D画面的功能。但是仅限于佩戴3D眼镜式的辅助式3D显示。其采用的技术是使得左右眼射入不同偏振光,来产生视差,形成立体感观。这种技术的缺点是需要佩戴偏光片眼镜,一方面偏光片眼镜的使用,使得进入医生眼睛内的光线降低为原来的一半以下,浪费了腔体内本身十分宝贵的光线信息,降低了腔体内较暗环境下的信息识别率;另一方面,对于平日生活中不佩戴眼镜的医生而言,在手术中佩戴偏光片眼镜,极易产生不适感,且因同时佩戴口罩容易在呼吸时在眼镜片表面产生雾气,对手术安全产生极大的影响。At present, a few existing medical endoscopes on the market have the function of displaying 3D images. However, it is limited to the auxiliary 3D display of wearing 3D glasses. The technology it uses is to make the left and right eyes inject different polarized light to generate parallax and form a three-dimensional perception. The disadvantage of this technology is that you need to wear polarizer glasses. On the one hand, the use of polarizer glasses reduces the light entering the doctor's eyes to less than half of the original, wasting the very precious light information in the cavity itself, and reducing the relatively low light intensity in the cavity. information recognition rate in dark environments; on the other hand, for doctors who do not wear glasses in daily life, wearing polarized glasses during surgery is very likely to cause discomfort, and wearing a mask at the same time is easy Fog is generated on the surface, which has a great impact on surgical safety.

因此,如何克服上述问题,成为医学界目前面临的一大技术难题。Therefore, how to overcome the above-mentioned problems has become a major technical problem that the medical field is currently facing.

发明内容Contents of the invention

本发明的目的在于提供一种基于软镜的立体显示系统及显示方法,旨在解决由现有技术的局限和缺点引起的上述一个或多个技术问题。The purpose of the present invention is to provide a stereoscopic display system and display method based on a soft mirror, aiming to solve one or more of the above-mentioned technical problems caused by the limitations and shortcomings of the prior art.

本发明提供的一种基于软镜的立体显示系统,包括:显示单元、分光单元、跟踪设备及图像拍摄单元,所述分光单元位于所述显示单元的显示侧,用于将所述显示单元显示的图像空间上分成左视图和右视图,所述跟踪设备用于获取第一目标对象的位置信息,所述图像拍摄单元用于拍摄第二目标对象,其中,所述基于软镜的立体显示系统还包括图像播放处理设备,分别与所述跟踪设备、所述显示单元及所述图像拍摄单元连接,所述图像播放处理设备依据所述第一目标对象的位置信息、所述分光单元的光栅参数及所述显示单元的显示参数,将接收到的所述图像拍摄单元拍摄到的立体图像进行实时处理,处理后发送所述显示单元实时显示。A stereoscopic display system based on a soft mirror provided by the present invention includes: a display unit, a light splitting unit, a tracking device, and an image capture unit, the light splitting unit is located on the display side of the display unit, and is used to display the display unit The image space is divided into a left view and a right view, the tracking device is used to obtain the position information of the first target object, and the image capture unit is used to capture the second target object, wherein the stereoscopic display system based on the soft mirror It also includes an image playback processing device, which is respectively connected to the tracking device, the display unit, and the image capturing unit, and the image playback processing device is based on the position information of the first target object and the grating parameters of the spectroscopic unit and the display parameters of the display unit, process the received stereoscopic image captured by the image capturing unit in real time, and send the processed stereo image to the display unit for real-time display.

本发明还提供一种基于软镜的立体显示方法,所述方法包括以下步骤:S0拍摄第二目标对象的立体图像,并实时发送所述立体图像的信息;S1获取第一目标对象的位置信息;S2获取显示设备的分光单元的光栅参数和所述显示设备的显示单元的显示参数;S3依据所述位置信息和所述光栅参数及所述显示参数将接收到的所述图像拍摄单元拍摄到的立体图像进行实时处理;S4显示所述待播放的图像。The present invention also provides a method for stereoscopic display based on a soft mirror, said method comprising the following steps: S0 taking a stereoscopic image of a second target object, and sending the information of said stereoscopic image in real time; S1 obtaining the position information of the first target object ; S2 acquires the grating parameters of the spectroscopic unit of the display device and the display parameters of the display unit of the display device; S3 captures the received image capturing unit according to the position information, the grating parameters and the display parameters The stereoscopic image is processed in real time; S4 displays the image to be played.

本发明提供的基于软镜的立体显示系统及显示方法,图像播放速度相比于现有技术大大提高,可满足实时立体显示的要求,具有方便医生手术操作及辅助医生提高手术成功率的优点。The stereoscopic display system and display method based on the soft mirror provided by the present invention greatly improves the image playback speed compared with the prior art, can meet the requirements of real-time stereoscopic display, and has the advantages of being convenient for doctors to operate and assisting doctors to improve the success rate of surgery.

附图说明Description of drawings

图1示出了本发明实施方式一的基于软镜的立体显示系统的结构示意图。FIG. 1 shows a schematic structural diagram of a stereoscopic display system based on a soft mirror according to Embodiment 1 of the present invention.

图2为本发明实施方式一的基于软镜的立体显示系统的一个具体实施例的结构示意图。FIG. 2 is a schematic structural diagram of a specific example of a stereoscopic display system based on a soft mirror according to Embodiment 1 of the present invention.

图3示出了图2中的图像播放处理单元的结构示意图。FIG. 3 shows a schematic structural diagram of the image playback processing unit in FIG. 2 .

图4是本发明实施方式一的基于软镜的立体显示系统中分光单元与显示单元贴合的结构示意图。FIG. 4 is a schematic structural diagram of bonding the light splitting unit and the display unit in the stereoscopic display system based on the soft mirror according to Embodiment 1 of the present invention.

图5示出了本发明实施方式一的基于软镜的立体显示系统中跟踪设备的较佳实施例结构示意图。FIG. 5 shows a schematic structural diagram of a preferred embodiment of a tracking device in a stereoscopic display system based on a soft mirror according to Embodiment 1 of the present invention.

图6示出了图4中的获取单元的具体结构示意图。FIG. 6 shows a schematic structural diagram of the acquisition unit in FIG. 4 .

图7示出了图4中的重新构建单元第一变形例的具体结构示意图。FIG. 7 shows a schematic structural diagram of a first modified example of the reconstruction unit in FIG. 4 .

图8示出了图4中的重新构建单元第二变形例的具体结构示意图。FIG. 8 shows a schematic structural diagram of a second modified example of the reconstruction unit in FIG. 4 .

图9示出了图4中的重新构建单元第三变形例的具体结构示意图。FIG. 9 shows a schematic structural diagram of a third modified example of the reconstruction unit in FIG. 4 .

图10示出了图4的跟踪设备中对应第一目标对象设置标记点的定位支架的结构示意图。FIG. 10 shows a schematic structural diagram of a positioning bracket for setting a marker point corresponding to a first target object in the tracking device of FIG. 4 .

图11是本发明实施方式二的基于软镜的立体显示方法的流程示意图。FIG. 11 is a schematic flowchart of a stereoscopic display method based on a soft mirror according to Embodiment 2 of the present invention.

图12是图11中S1的具体流程示意图。FIG. 12 is a schematic flowchart of S1 in FIG. 11 .

图13是图12中S12的具体流程示意图。FIG. 13 is a schematic flow chart of S12 in FIG. 12 .

图14是图11中S13的第一变形例的具体流程示意图。FIG. 14 is a schematic flowchart of a first modified example of S13 in FIG. 11 .

图15是图11中S13的第二变形例的具体流程示意图。FIG. 15 is a schematic flowchart of a second modified example of S13 in FIG. 11 .

图16是图11中S13的第三变形例的具体流程示意图。FIG. 16 is a schematic flowchart of a third modified example of S13 in FIG. 11 .

图17是图11中的S3的具体流程示意图。FIG. 17 is a schematic flowchart of S3 in FIG. 11 .

具体实施方式detailed description

为了能够更清楚地理解本发明的上述目的、特征和优点,下面结合附图和具体实施方式对本发明进行进一步的详细描述。需要说明的是,在不冲突的情况下,本申请的实施方式及实施方式中的特征可以相互组合。In order to understand the above-mentioned purpose, features and advantages of the present invention more clearly, the present invention will be further described in detail below in conjunction with the accompanying drawings and specific embodiments. It should be noted that, in the case of no conflict, the embodiments of the present application and the features in the embodiments can be combined with each other.

在下面的描述中阐述了很多具体细节以便于充分理解本发明,但是,本发明还可以采用其他不同于在此描述的其他方式来实施,因此,本发明的保护范围并不受下面公开的具体实施方式的限制。In the following description, many specific details are set forth in order to fully understand the present invention. However, the present invention can also be implemented in other ways different from those described here. Therefore, the protection scope of the present invention is not limited by the specific details disclosed below. Implementation limitations.

实施方式一Implementation Mode 1

请参见图1,图1为本发明基于软镜的立体显示系统的结构示意图。如图1所示,本发明的基于软镜的立体显示系统包括:图像拍摄单元10、跟踪设备30、分光单元50及显示单元40。该图像拍摄单元10用于拍摄第二目标对象,并实时将拍摄到的该第二目标对象的图像发送给该图像播放单元。该跟踪设备30用于获取第一目标对象的位置信息,该分光单元50位于所述显示单元40的显示侧,用于将该显示单元40显示的图像空间上分成左视图和右视图。该基于软镜的立体显示系统还包括图像播放处理设备20,分别与该跟踪设备30和该显示单元40连接,该图像播放处理设备20依据该第一目标对象的位置信息、该分光单元50的光栅参数及显示单元40的显示参数实时处理待播放的图像,处理后发送该显示单元40进行显示。另外,该图像拍摄单元10设于软镜上,当软镜进入人或动物体内时,拍摄人或动物内部的图像,方便医生实时观察并及时手术。Please refer to FIG. 1 . FIG. 1 is a schematic structural diagram of a stereoscopic display system based on a soft mirror according to the present invention. As shown in FIG. 1 , the stereoscopic display system based on the soft mirror of the present invention includes: an image capturing unit 10 , a tracking device 30 , a spectroscopic unit 50 and a display unit 40 . The image capture unit 10 is used to capture a second target object, and send the captured image of the second target object to the image playback unit in real time. The tracking device 30 is used to obtain the position information of the first target object. The light splitting unit 50 is located on the display side of the display unit 40 and is used to spatially divide the image displayed by the display unit 40 into a left view and a right view. The stereoscopic display system based on the soft mirror also includes an image playback processing device 20, which is respectively connected to the tracking device 30 and the display unit 40, and the image playback processing device 20 is based on the position information of the first target object, the The raster parameters and display parameters of the display unit 40 process the image to be played in real time, and send it to the display unit 40 for display after processing. In addition, the image capture unit 10 is set on the soft mirror, and when the soft mirror enters the body of a human or animal, it can capture images inside the human or animal, which is convenient for doctors to observe in real time and operate in time.

由于跟踪设备30和显示单元40直接连接到图像播放处理设备20,图像播放处理设备20及时获取到第一目标对象的位置信息、光栅参数及显示参数,并据此进行图像处理,省去了现有技术中需要经过中央处理器的处理过程,因而图像播放的速度相比于现有技术大大提高,能满足实时立体显示的要求,方便医生手术操作及辅助医生提高手术成功率的优点。这是因为医生在手术时,能实时获得准确的立体图像,并及时进行手术,不会出现背景技术中提到的问题。上述光栅参数主要包括光栅的栅距(pitch)和光栅相对显示面板的倾斜角度、光栅相对显示面板的放置距离等参数。这些光栅参数可以是直接存储在图像播放处理设备内的存储器中,也可是其它检测设备实时检测并获取分光单元的光栅参数,将光栅参数值发送给图像播放处理设备20。上述显示单元参数包括显示单元的尺寸、显示单元的屏幕分辨率、显示单元的像素单元中子像素的排列顺序及排列结构等参数。子像素排列顺序即子像素是按照RGB排列或者RBG排列,还是成BGR排列,还是成其他顺序排列;子像素排列结构即子像素为竖直排列的还是横向排列的,如自上而下是按照RGB的方式循环排列,或者从左至右依次是按照RGB的方式循环排列等。Since the tracking device 30 and the display unit 40 are directly connected to the image playback processing device 20, the image playback processing device 20 obtains the position information, raster parameters and display parameters of the first target object in time, and performs image processing accordingly, eliminating the need In the existing technology, the processing process of the central processing unit is required, so the speed of image playback is greatly improved compared with the existing technology, which can meet the requirements of real-time stereoscopic display, facilitate the doctor's operation and assist the doctor to improve the success rate of the operation. This is because the doctor can obtain accurate stereoscopic images in real time during the operation, and perform the operation in time without the problems mentioned in the background art. The grating parameters above mainly include the pitch of the grating, the inclination angle of the grating relative to the display panel, and the placement distance of the grating relative to the display panel. These grating parameters can be directly stored in the memory of the image playback processing device, or other detection devices can detect and acquire the grating parameters of the spectroscopic unit in real time, and send the raster parameter values to the image playback processing device 20 . The above parameters of the display unit include parameters such as size of the display unit, screen resolution of the display unit, arrangement sequence and arrangement structure of sub-pixels in the pixel unit of the display unit. The sub-pixel arrangement order means whether the sub-pixels are arranged in RGB or RBG, or in BGR, or in other orders; the sub-pixel arrangement structure is whether the sub-pixels are arranged vertically or horizontally, such as from top to bottom according to Arranged circularly in the way of RGB, or arranged circularly in the way of RGB from left to right.

该图像拍摄单元10用于拍摄第二目标对象,并实时将拍摄到的该第二目标对象的图像发送给该图像播放单元。这里的第二目标对象主要是指由摄像机拍摄记录的各种场景,如手术的现场,病人的体内影像等等。通过图像拍摄单元10实时地拍摄立体图像,并将拍摄到的立体图像实时显示在显示单元上,无需经过额外的图像处理,及时并真实地显示拍摄到的各种场景,满足了用户对实时显示的需求,提高了用户体验。图像拍摄单元10可以包括单目摄像机、双目摄像机或者多目摄像机中的至少一种。The image capture unit 10 is used to capture a second target object, and send the captured image of the second target object to the image playback unit in real time. The second target object here mainly refers to various scenes shot and recorded by the camera, such as a scene of an operation, an internal image of a patient, and the like. The stereoscopic image is captured in real time by the image capture unit 10, and the captured stereoscopic image is displayed on the display unit in real time, without additional image processing, and various scenes captured are displayed in a timely and true manner, satisfying the user's demand for real-time display needs and improve the user experience. The image capture unit 10 may include at least one of a monocular camera, a binocular camera or a multi-camera.

当该图像拍摄单元10包括单目摄像机时,依据该单目摄像机拍摄并获取第二目标对象的立体图像。较佳地,该单目摄像机可以采用液晶透镜成像装置或者液晶微透镜阵列成像装置。在一个具体实施例中,该单目摄像机在不同时刻从不同角度获得被测物的两幅数字图像,并基于视差原理恢复出物体的三维几何信息,重建物体三维轮廓及位置。When the image capturing unit 10 includes a monocular camera, the stereoscopic image of the second target object is captured and acquired according to the monocular camera. Preferably, the monocular camera may use a liquid crystal lens imaging device or a liquid crystal microlens array imaging device. In a specific embodiment, the monocular camera obtains two digital images of the measured object from different angles at different times, and restores the three-dimensional geometric information of the object based on the parallax principle, and reconstructs the three-dimensional outline and position of the object.

当该图像拍摄单元10包括双目摄像机时,包括两个摄像机或者是一个摄像机有两个摄像头,通过双目摄像机对第二目标对象进行拍摄第二目标对象形成立体图像。具体地,双目摄像机主要是基于视差原理并由多幅图像获取物体三维几何信息。双目立体视觉系统一般由双摄像机从不同角度同时获得被测物(第二目标对象)的两幅数字图像,并基于视差原理恢复出物体的三维几何信息,重建物体三维轮廓及位置。When the image capture unit 10 includes a binocular camera, it includes two cameras or one camera has two cameras, and the second target object is captured by the binocular camera to form a stereoscopic image. Specifically, the binocular camera is mainly based on the principle of parallax and obtains three-dimensional geometric information of objects from multiple images. The binocular stereo vision system generally obtains two digital images of the measured object (second target object) from different angles simultaneously by dual cameras, and restores the three-dimensional geometric information of the object based on the parallax principle, and reconstructs the three-dimensional outline and position of the object.

当该图像拍摄单元10包括多目摄像机时,即三个以上(包括三个)的摄像机,这些摄像机呈矩阵排列,用于获取立体图像。由上述三个以上的摄像机从不同角度同时获取第二目标对象的多幅数字图像,基于视差原理恢复出物体的三维几何信息,重建物体三维轮廓及位置。When the image capture unit 10 includes a multi-eye camera, that is, more than three (including three) cameras, these cameras are arranged in a matrix for acquiring stereoscopic images. Simultaneously acquire multiple digital images of the second target object from different angles by the above three or more cameras, restore the three-dimensional geometric information of the object based on the parallax principle, and reconstruct the three-dimensional outline and position of the object.

该图像拍摄单元10还包括采集单元,该采集单元用于采集该第二目标对象的立体图像,并从该立体图像中提取出左视图信息和右视图信息。该采集单元一端与上述单目摄像机、双目摄像机或者上述多目摄像机连接,另一端连接到图像播放处理设备20上。通过采集单元在边拍摄立体图像时边提取立体图像的左视图信息和右视图信息,提高了图像处理的速度,保证了实时进行立体显示的显示效果。The image capture unit 10 also includes a capture unit, which is used to capture a stereo image of the second target object, and extract left view information and right view information from the stereo image. One end of the acquisition unit is connected to the above-mentioned monocular camera, binocular camera or the above-mentioned multi-eye camera, and the other end is connected to the image playback processing device 20 . The acquisition unit extracts the left view information and the right view information of the stereoscopic image while shooting the stereoscopic image, which improves the speed of image processing and ensures the display effect of real-time stereoscopic display.

上述跟踪设备30可以是摄像头和/或者红外传感器,主要用于追踪第一目标对象的位置,例如人的双眼或者人的头部或者人的脸部的位置或者人的上半身的位置。摄像头或红外传感器的数量不限制,可以是一个,也可以是多个。摄像头或者红外传感器可以安装在显示单元的边框上,或者单独放置在易于追踪到第一目标对象的位置。此外,如果采用红外传感器作为跟踪设备,还可在对应第一目标对象的位置设置红外发射器,通过接收到红外发射器发送的红外定位信号,利用红外发送器与第一目标对象的相对位置关系,计算出第一目标对象的位置信息。The above-mentioned tracking device 30 may be a camera and/or an infrared sensor, and is mainly used to track the position of the first target object, such as the position of the eyes of the person or the head or the face of the person or the position of the upper body of the person. The number of cameras or infrared sensors is not limited, and may be one or more. A camera or an infrared sensor can be mounted on the frame of the display unit, or placed separately at a position where it is easy to track the first target object. In addition, if an infrared sensor is used as the tracking device, an infrared transmitter can also be installed at the position corresponding to the first target object, and by receiving the infrared positioning signal sent by the infrared transmitter, the relative positional relationship between the infrared transmitter and the first target object can be used , to calculate the location information of the first target object.

具体地,该跟踪设备30包括摄像机,该摄像机拍摄该第一目标对象。摄像机的数量可以是一个或多个,可以设置在显示单元上,也可以单独设置。Specifically, the tracking device 30 includes a camera, and the camera shoots the first target object. The number of cameras can be one or more, and can be set on the display unit or set separately.

该跟踪设备30包括红外接收器,相应地,对应第一目标对象设置有红外发射器,该红外发射器可设置在第一目标对象的相应位置,也可以设置在其它与第一目标对象位置相对固定的物体上,该红外接收器接收来自对应第一目标对象所设置的红外发射器发送的红外信号。通过常见的红外定位方法实现对第一目标对象的定位。The tracking device 30 includes an infrared receiver. Correspondingly, an infrared emitter is arranged corresponding to the first target object. On the fixed object, the infrared receiver receives the infrared signal sent from the infrared emitter corresponding to the first target object. The positioning of the first target object is realized by a common infrared positioning method.

此外,上述跟踪设备30还可以采用GPS定位模块,由GPS定位模块发送定位信息至图像播放处理设备20。In addition, the tracking device 30 may also use a GPS positioning module, and the GPS positioning module sends positioning information to the image playback processing device 20 .

上述分光单元50设于显示单元40的出光侧,将显示单元40显示的具有视差的左视图和右视图分别发送到人的左眼和右眼,依据人的左眼和右眼合成立体图像,使人观看到立体显示的效果。较佳地,上述分光单元可以是视差屏障或透镜光栅。该视差屏障可以是液晶狭缝或固体狭缝光栅片或者电致变色狭缝光栅片等,该透镜光栅可以是液晶透镜或固体液晶透镜光栅。固体液晶透镜光栅主要是通过紫外光将液晶固化到薄片上,形成固态透镜,对光线进行分光后出射到人的左眼和右眼。较佳地,将上述显示单元40和分光单元50作为一个集成的显示设备60,该显示设备60是整个基于软镜的立体显示系统的显示部分,可以与前述图像播放处理设备及跟踪设备组装在一起,也可以是一个独立部分单独存在。例如,可以依据观看需要,单独将显示设备60放置在便于观看的位置,而图像播放处理设备20和跟踪设备30可以是各自具有独立功能的设备,使用时将这些设备组装起来实现本发明的实时立体显示功能即可。例如,该图像播放处理设备20可以是VMR3D播放设备,其本身具有3D播放处理功能,使用是将其组装到本发明的基于软镜的立体显示系统中,与其它设备建立连接。The above-mentioned spectroscopic unit 50 is arranged on the light-emitting side of the display unit 40, and sends the left view and right view with parallax displayed by the display unit 40 to the left and right eyes of the person respectively, and synthesizes a stereoscopic image according to the left and right eyes of the person, Make people watch the effect of stereoscopic display. Preferably, the light splitting unit may be a parallax barrier or a lens grating. The parallax barrier can be a liquid crystal slit or a solid slit grating sheet or an electrochromic slit grating sheet, and the lens grating can be a liquid crystal lens or a solid liquid crystal lens grating. The solid liquid crystal lens grating mainly solidifies the liquid crystal on the sheet by ultraviolet light to form a solid lens, which splits the light and emits it to the left and right eyes of the person. Preferably, the above-mentioned display unit 40 and the spectroscopic unit 50 are used as an integrated display device 60, which is the display part of the entire soft mirror-based stereoscopic display system, and can be assembled with the aforementioned image playback processing device and tracking device. together, or as an independent part. For example, according to viewing needs, the display device 60 can be placed separately at a convenient viewing position, while the image playback processing device 20 and the tracking device 30 can be devices with independent functions. Stereoscopic display function can be. For example, the image playback processing device 20 may be a VMR3D playback device, which itself has a 3D playback processing function, and is used to assemble it into the stereoscopic display system based on the soft mirror of the present invention, and establish a connection with other devices.

请参见图2,图2为本发明实施方式一的基于软镜的立体显示系统的一个具体实施例的结构示意图。如图2所示,本发明实施例1的基于软镜的立体显示系统中,图像播放处理设备20进一步包括:图像播放处理单元22及存储单元23,图像播放处理单元22主要是依据所述第一目标对象的位置信息、所述分光单元的光栅参数及显示单元的显示参数,将接收到的立体图像进行实时排图处理,处理后发送所述显示单元实时显示。存储单元23用于存储图像拍摄单元10传输过来的图像。在需要播放立体图像时,图像播放处理单元22调用存储单元23中存储的立体图像进行排图处理。Please refer to FIG. 2 . FIG. 2 is a schematic structural diagram of a specific example of a stereoscopic display system based on a soft mirror according to Embodiment 1 of the present invention. As shown in Figure 2, in the stereoscopic display system based on the soft mirror in Embodiment 1 of the present invention, the image playback processing device 20 further includes: an image playback processing unit 22 and a storage unit 23, and the image playback processing unit 22 is mainly based on the first The location information of the target object, the grating parameters of the spectroscopic unit and the display parameters of the display unit are used to perform real-time layout processing on the received stereoscopic images, and then send the processed images to the display unit for real-time display. The storage unit 23 is used for storing images transmitted from the image capture unit 10 . When it is necessary to play a stereoscopic image, the image playback processing unit 22 invokes the stereoscopic image stored in the storage unit 23 to perform image layout processing.

进一步地,该图像播放处理设备20还进一步包括:信号处理单元21,该信号处理单元21分别与存储单元23和图像播放处理单元22连接,其中信号处理单元21主要是对接收到的图像拍摄单元10拍摄的立体图像的信号进行处理,包括图像格式转换、图像压缩等处理。将压缩后的图像存储到存储单元23中。经过信号处理单元21处理后的信号可以对应于左视图、右视图信息分别输出图像信号,或者一并输出到图像播放处理单元22中。图像播放处理单元22依据所述第一目标对象的位置信息及所述分光单元的光栅参数,将接收到的信号处理单元21处理后的立体图像进行实时排图处理,处理后发送所述显示单元实时显示。这里图像播放处理单元22可以直接调用存储单元23中的立体图像并解压缩,也可以直接接收信号处理单元21处理后的立体图像,然后进行排图处理。Further, the image playback processing device 20 further includes: a signal processing unit 21, the signal processing unit 21 is respectively connected to the storage unit 23 and the image playback processing unit 22, wherein the signal processing unit 21 is mainly a unit for capturing images received 10 to process the signal of the stereoscopic image taken, including image format conversion, image compression and other processing. The compressed image is stored in the storage unit 23 . The signals processed by the signal processing unit 21 can output image signals corresponding to left view and right view information respectively, or output them to the image playback processing unit 22 together. The image playback processing unit 22 performs real-time layout processing on the received stereoscopic image processed by the signal processing unit 21 according to the position information of the first target object and the grating parameters of the spectroscopic unit, and sends the processed image to the display unit real-time display. Here, the image playback processing unit 22 can directly call and decompress the stereoscopic images in the storage unit 23, or directly receive the stereoscopic images processed by the signal processing unit 21, and then perform image layout processing.

请参见图3,图像播放处理单元22进一步包括:Referring to Fig. 3, the image playback processing unit 22 further includes:

排图参数确定模块201,依据获取到的所述第一目标对象的位置信息和所述分光单元的光栅参数及显示单的显示参数计算出在显示单元上的排图参数;The layout parameter determination module 201 calculates the layout parameters on the display unit according to the acquired position information of the first target object, the grating parameters of the spectroscopic unit and the display parameters of the display sheet;

视差图像排列模块202,用于依据所述排图参数排列显示单元上的视差图像;该视差图像是通过在空间上划分左眼图像以及右眼图像来生成的。The parallax image arrangement module 202 is configured to arrange the parallax images on the display unit according to the layout parameters; the parallax images are generated by spatially dividing the left-eye image and the right-eye image.

视差图像播放模块203,播放所述视差图像。在接收到排列后的视差图像后,进行播放,观看者在显示单元实时看到显示的立体图像。The parallax image playing module 203 plays the parallax image. After receiving the arranged parallax images, they are played, and the viewer sees the displayed stereoscopic images in real time on the display unit.

进一步地,该图像播放处理单元22还包括:立体图像获取模块204,获取所述图像拍摄单元10拍摄的立体图像信息,即立体图像的左视图和右视图信息。立体图像包括左视图和右视图,因此,对待播放的立体图像,需要先获得左视图和右视图的图像信息,才能进行排图处理。Further, the image playback processing unit 22 also includes: a stereoscopic image acquisition module 204, which acquires the stereoscopic image information captured by the image capturing unit 10, that is, the left view and right view information of the stereoscopic image. A stereoscopic image includes a left view and a right view. Therefore, for a stereoscopic image to be played, the image information of the left view and the right view needs to be obtained before image layout processing can be performed.

请继续参见图2,在本发明中的跟踪设备30进一步包括跟踪定位处理单元31和跟踪单元32。跟踪单元32用于跟踪第一目标对象的实时图像,其主要指摄像头、红外接收器等能够准确地捕捉人眼或头部视频信号的一类设备。跟踪定位处理单元31依据跟踪单元32跟踪到的第一目标对象的实时图像,通过对第一目标对象的特征点进行提取,计算第一目标对象的空间坐标。具体地,例如通过摄像头来实时记录人脸,经跟踪定位处理单元31进行人脸特征点提取,计算出人眼的空间坐标;还可以通过增加特征点的方式,如在人的头部增加特征点,如红外发射装置,再通过摄像头来捕捉该特征点的实时图像,跟踪定位处理单元31最后计算出人眼的空间坐标。Please continue to refer to FIG. 2 , the tracking device 30 in the present invention further includes a tracking and positioning processing unit 31 and a tracking unit 32 . The tracking unit 32 is used to track the real-time image of the first target object, which mainly refers to a type of equipment such as a camera and an infrared receiver that can accurately capture human eye or head video signals. The tracking and positioning processing unit 31 calculates the spatial coordinates of the first target object by extracting feature points of the first target object according to the real-time image of the first target object tracked by the tracking unit 32 . Specifically, for example, a camera is used to record a face in real time, and the tracking and positioning processing unit 31 performs feature point extraction on the face to calculate the spatial coordinates of the human eye; point, such as an infrared emitting device, and then capture a real-time image of the feature point through a camera, and the tracking and positioning processing unit 31 finally calculates the spatial coordinates of the human eye.

此外,跟踪定位处理单元31能够在人眼位置发生移动时快速实时地跟踪人眼位置,给出人眼的空间坐标,并将该坐标信息提供给图像播放处理单元22。In addition, the tracking and positioning processing unit 31 can quickly track the position of the human eye in real time when the position of the human eye moves, give the spatial coordinates of the human eye, and provide the coordinate information to the image playback processing unit 22 .

还有上述跟踪单元32可以包括摄像机或红外接收器,当跟踪单元32包括摄像机时,通过摄像机跟踪所述第一目标对象对应的特征点的位置变化。当跟踪单元32包括红外接收器时,所述红外接收器接收来自于对应所述第一目标对象且作为所述特征点的红外发射器发送的红外定位信号。In addition, the above-mentioned tracking unit 32 may include a camera or an infrared receiver. When the tracking unit 32 includes a camera, the camera tracks the position change of the feature point corresponding to the first target object. When the tracking unit 32 includes an infrared receiver, the infrared receiver receives an infrared positioning signal sent from an infrared transmitter corresponding to the first target object and serving as the feature point.

通过上述跟踪定位处理单元31和跟踪单元32提升了立体显示的观看效果,使立体显示随人的移动而自动调整,实时地给出最优的立体显示效果。Through the tracking and positioning processing unit 31 and the tracking unit 32, the viewing effect of the stereoscopic display is improved, the stereoscopic display is automatically adjusted with the movement of the person, and the optimal stereoscopic display effect is provided in real time.

实施例1Example 1

在本发明实施例1中,要得到较佳的实时立体显示效果,需要依据分光单元的光栅参数和显示单元的显示参数对分光单元与显示单元进行光学设计,该光学设计依据以下公式:In Embodiment 1 of the present invention, in order to obtain a better real-time stereoscopic display effect, it is necessary to carry out optical design on the light splitting unit and the display unit according to the grating parameters of the light splitting unit and the display parameters of the display unit, and the optical design is based on the following formula:

(( 11 )) -- -- -- nno ** IPDIPD mm ** tt == LL Ff

(( 22 )) -- -- -- ll -- pitchpitch pp -- pitchpitch == LL LL ++ Ff

(3)m*t=p-pitch(3) m*t=p-pitch

上述公式中,F为分光单元与显示单元之间的距离(即上述光栅参数中的光栅相对显示面板的放置距离),L为观看者与显示单元的距离,IPD为匹配瞳距,通常的人双瞳之间的距离,例如,一般取值为62.5mm,l-pitch为分光单元的节距(pitch),p-pitch为显示单元上的像素的排图节距,n为立体视图数量,m为分光单元所覆盖的像素数量,p为显示单元的点距,这里的点距主要是指一个像素单元的尺寸(属于显示参数的一种),该像素单元通常包括R、G、B三个子像素。为了消除摩尔纹,分光单元在贴合的时候一般会旋转一定的角度(即分光单元相较于显示单元有一定的倾斜角度),因此,实际上的分光单元的节距由以下公式给出:In the above formula, F is the distance between the spectroscopic unit and the display unit (that is, the placement distance of the grating in the above grating parameters relative to the display panel), L is the distance between the viewer and the display unit, and IPD is the matching pupil distance. The distance between the pupils, for example, generally takes a value of 62.5mm, l-pitch is the pitch (pitch) of the light splitting unit, p-pitch is the layout pitch of the pixels on the display unit, n is the number of stereoscopic views, m is the number of pixels covered by the light-splitting unit, and p is the dot pitch of the display unit. The dot pitch here mainly refers to the size of a pixel unit (a kind of display parameter), and the pixel unit usually includes R, G, B three sub-pixels. In order to eliminate moiré, the light splitting unit is generally rotated at a certain angle when it is attached (that is, the light splitting unit has a certain inclination angle compared to the display unit). Therefore, the actual pitch of the light splitting unit is given by the following formula:

(4)Wlens=l-pitch*sinθ(4) W lens = l-pitch*sinθ

其中,Wlens为分光单元的实际节距,θ为分光单元相对显示面板的倾斜角度(即上述光栅参数之一)。Wherein, W lens is the actual pitch of the light splitting unit, and θ is the inclination angle of the light splitting unit relative to the display panel (that is, one of the above grating parameters).

如上所述,对于分光单元与显示单元之间的距离F,当显示单元与分光单元之间的介质是空气时,F就等于分光单元与显示单元之间的实际距离;当显示单元与分光单元之间的介质是折射率为n(n大于1)的透明介质时,F等于分光单元与显示单元之间的实际距离除以该折射率n;当显示单元与分光单元之间存在不同的介质时,且介质的折射率分别为n1、n2、n3···(折射率均大于或等于1),F=s1/n1+s2/n2+s3/n3···,其中s1、s2、s3···为相应介质的厚度。As mentioned above, for the distance F between the light splitting unit and the display unit, when the medium between the display unit and the light splitting unit is air, F is equal to the actual distance between the light splitting unit and the display unit; when the display unit and the light splitting unit When the medium in between is a transparent medium with a refractive index n (n is greater than 1), F is equal to the actual distance between the spectroscopic unit and the display unit divided by the refractive index n; when there are different media between the display unit and the spectroscopic unit , and the refractive index of the medium is n1, n2, n3... (the refractive index is greater than or equal to 1), F=s1/n1+s2/n2+s3/n3..., where s1, s2, s3 ··· is the thickness of the corresponding medium.

通过上述光学计算公式,对分光单元和显示单元进行设置,可以减小摩尔纹,提高实时观看的立体显示效果。By setting the spectroscopic unit and the display unit through the above optical calculation formula, the moiré pattern can be reduced, and the stereoscopic display effect of real-time viewing can be improved.

此外,在一个变形实施例中,在分光单元与显示单元之间设置贴合单元,请参见图4,图4是本发明实施方式一的基于软镜的立体显示系统中分光单元与显示单元的贴合结构示意图。如图4所示,在分光单元50和显示单元40之间设有贴合单元,三者类似于“三明治结构”,贴合单元包括第一基板42和第二基板43,以及位于第一基板42与第二基板43之间的空气层41。该空气层41在第一基板42与第二基板43之间处于密封状态,防止空气逸出。第一基板42与显示面板贴合,可以是透明玻璃材料构成,也可以是透明树脂材料等构成。第二基板43与第一基板42相对设置,其背离第一基板42的一侧用于贴合分光单元50。由于在分光单元50和显示单元40之间设置贴合单元,且贴合单元采用上述结构,对于大屏幕的立体显示装置,既保证了光栅贴合的平整度,又减轻了整个立体显示装置的重量,防止采用纯玻璃时因过重而导致屏幕坠裂的风险。In addition, in a modified embodiment, a bonding unit is provided between the light splitting unit and the display unit, please refer to FIG. 4 , which shows the relationship between the light splitting unit and the display unit in the stereoscopic display system based on a soft mirror according to Embodiment 1 of the present invention. Schematic diagram of the fitting structure. As shown in FIG. 4, a bonding unit is provided between the light splitting unit 50 and the display unit 40, and the three are similar to a "sandwich structure". The bonding unit includes a first substrate 42 and a second substrate 43, and a 42 and the air layer 41 between the second substrate 43. The air layer 41 is in a sealed state between the first substrate 42 and the second substrate 43 to prevent air from escaping. The first substrate 42 is bonded to the display panel, and may be made of a transparent glass material, or a transparent resin material or the like. The second substrate 43 is disposed opposite to the first substrate 42 , and the side of the second substrate 43 facing away from the first substrate 42 is used for attaching the spectroscopic unit 50 . Since the bonding unit is arranged between the light splitting unit 50 and the display unit 40, and the bonding unit adopts the above-mentioned structure, for a large-screen stereoscopic display device, the flatness of grating bonding is guaranteed, and the weight of the entire stereoscopic display device is reduced. Weight, to prevent the risk of screen cracking due to excessive weight when pure glass is used.

实施例2Example 2

在本实施例2中,该跟踪设备30包括摄像机,该摄像机拍摄该第一目标对象。摄像机的数量可以是一个或多个,可以设置在显示单元上,也可以单独设置。还有,摄像机可以是单目摄像机、双目摄像机或者多目摄像机。In Embodiment 2, the tracking device 30 includes a camera, and the camera shoots the first target object. The number of cameras can be one or more, and can be set on the display unit or set separately. Also, the camera may be a monocular camera, a binocular camera or a multi-eye camera.

另外,该跟踪设备30还可以是包括红外接收器,相应地,对应第一目标对象设置有红外发射器,该红外发射器可设置在第一目标对象的相应位置,也可以设置在其它与第一目标对象位置相对固定的物体上,该红外接收器接收来自对应第一目标对象所设置的红外发射器发送的红外信号。通过常见的红外定位方法实现对第一目标对象的定位。In addition, the tracking device 30 may also include an infrared receiver. Correspondingly, an infrared emitter is provided corresponding to the first target object. On an object with a relatively fixed position on a target object, the infrared receiver receives an infrared signal sent from an infrared emitter corresponding to the first target object. The positioning of the first target object is realized by a common infrared positioning method.

此外,上述跟踪设备30还可以采用GPS定位模块,由GPS定位模块发送定位信息至图像播放处理单元20。In addition, the above-mentioned tracking device 30 may also use a GPS positioning module, and the GPS positioning module sends positioning information to the image playback processing unit 20 .

实施例3Example 3

请参见图5,图5示出了本发明实施方式一的基于软镜的立体显示系统中跟踪设备的较佳实施例结构示意图。如图5所示,本发明实施例3还提出另一种跟踪设备30,该跟踪设备30包括:Please refer to FIG. 5 , which shows a schematic structural diagram of a preferred embodiment of a tracking device in a stereoscopic display system based on a soft mirror according to Embodiment 1 of the present invention. As shown in Figure 5, Embodiment 3 of the present invention also proposes another tracking device 30, which includes:

标记点设置单元1,用于对应第一目标对象的空间位置设置标记点;这里的标记点可以设置第一目标对象上,也可以不设置于第一目标对象上,而是设置在与第一目标对象有相对位置关系,与第一目标对象同步移动的物体上也可。例如,第一目标对象是人眼,则可以在人眼的眼眶周围设置标记点;或者在人眼周围配置眼镜,将标记点设在眼镜的镜框上,又或者将标记点设在与人眼位置关系相对固定的人的耳朵上。该标记点可以是发送信号的红外发射传感器,LED灯,GPS传感器,激光定位传感器等各种部件,也可以是其它能够被摄像机捕捉的物理标识,例如是具有形状特征和/或颜色特征的物体。较佳地,为避免外界杂光的干扰,提高标记点追踪的鲁棒性,优选使用频谱较为狭窄的红外LED灯作为标记点,并使用只能通过红外LED所使用频谱的相应的红外摄像机对标记点进行捕捉。考虑到外界杂光多为不规则的形状且亮度分布不均匀,可以将标记点设置成可以发出规则形状的光斑,发光强度较高,亮度均匀。另外还可以设置多个标记点,每个标记点对应一个光斑,各个标记点组成规则的几何形状,如三角形,四边形等,从而易于跟踪到标记点,获得标记点的空间位置信息,并提高光斑提取的准确性。Marking point setting unit 1 is used to set the marking point corresponding to the spatial position of the first target object; the marking point here can be set on the first target object, or can not be set on the first target object, but be set on the first target object. The target object has a relative positional relationship, and it can also be on an object that moves synchronously with the first target object. For example, if the first target object is the human eye, marking points can be set around the eye sockets of the human eye; On the ear of a person whose positional relationship is relatively fixed. The marking point can be various components such as infrared emission sensors that send signals, LED lights, GPS sensors, laser positioning sensors, etc., or other physical signs that can be captured by the camera, such as objects with shape characteristics and/or color characteristics . Preferably, in order to avoid the interference of external stray light and improve the robustness of marker point tracking, it is preferable to use infrared LED lights with a relatively narrow spectrum as marker points, and use a corresponding infrared camera that can only pass through the spectrum used by infrared LEDs to Mark points to snap to. Considering that the external stray light is mostly irregular in shape and the brightness distribution is uneven, the marking points can be set to emit regular-shaped light spots, with high luminous intensity and uniform brightness. In addition, multiple marking points can be set, each marking point corresponds to a light spot, and each marking point forms a regular geometric shape, such as a triangle, a quadrilateral, etc., so that it is easy to track the marking point, obtain the spatial position information of the marking point, and improve the light spot. Extraction accuracy.

获取单元2,用于获取该标记点的位置信息;这可以是通过接收标记点发出的信号,来确定标记点的位置信息,也可以是采用摄像机来拍摄含有标记点的图像,对图像中的标记点进行提取。通过图像处理算法来获得标记点的位置信息。Acquisition unit 2 is used to obtain the position information of the marking point; this can be by receiving the signal sent by the marking point to determine the position information of the marking point, and it can also be to use a camera to shoot an image containing the marking point, and to determine the position information of the marking point in the image mark points for extraction. The location information of the marked points is obtained through image processing algorithms.

重新构建单元3,用于依据该标记点的位置信息,重新构建该第一目标对象的空间位置。当获得到该标记点的位置信息后,重新构建出标记点的空间位置,然后依据标记点与第一目标对象的相对位置关系,将标记点的空间位置转换到第一目标对象的空间位置(例如人的左右两眼的空间位置)。The reconstructing unit 3 is configured to reconstruct the spatial position of the first target object according to the position information of the marker point. After obtaining the position information of the mark point, the spatial position of the mark point is reconstructed, and then according to the relative positional relationship between the mark point and the first target object, the spatial position of the mark point is converted to the spatial position of the first target object ( For example, the spatial position of the left and right eyes of a person).

本发明实施例的跟踪设备30通过获取对应第一目标对象的标记点的位置信息,并依据该位置信息,重新构建出第一目标对象的空间位置。与现有技术中使用摄像机作为人眼捕捉设备需要对二维图像进行特征分析从而获取人眼位置或者使用其他利用人眼虹膜反射效果的人眼捕捉设备获取人眼位置相比较,具有稳定性好,准确度高、成本低廉且对跟踪设备与第一目标对象之间的距离远近没有要求的优点。The tracking device 30 in the embodiment of the present invention reconstructs the spatial position of the first target object by acquiring the position information of the marker point corresponding to the first target object and according to the position information. Compared with the existing technology that uses a camera as a human eye capture device and needs to perform feature analysis on a two-dimensional image to obtain the position of the human eye or use other human eye capture devices that use the reflection effect of the iris of the human eye to obtain the position of the human eye, it has good stability. , has the advantages of high accuracy, low cost, and no requirement on the distance between the tracking device and the first target object.

请参见图6,图6示出了图5中的获取单元的具体结构示意图。前述获取单元进一步包括:Please refer to FIG. 6 , which shows a schematic structural diagram of the acquisition unit in FIG. 5 . The aforementioned acquisition unit further includes:

预设模块21,用于预设一标准图像,所述标准图像中设有参考标记点,并获取所述参考标记点的空间坐标和平面坐标;标准图像例如可以是通过图像采集设备采集的一个标准图像,获取参考标记点的图像坐标,并使用其他精确的立体测量设备如激光扫描仪,结构光扫描仪(如Kinect等)等设备获取标准图像中参考标记点的空间坐标及平面坐标。The preset module 21 is used to preset a standard image, the reference mark point is set in the standard image, and the spatial coordinates and plane coordinates of the reference mark point are obtained; the standard image can be, for example, a collected by an image acquisition device Standard image, obtain the image coordinates of the reference mark points, and use other accurate stereo measurement equipment such as laser scanners, structured light scanners (such as Kinect, etc.) to obtain the spatial coordinates and plane coordinates of the reference mark points in the standard image.

获取模块22,用于获取包含所述第一目标对象及所述标记点的当前图像,及所述标记点在所述当前图像的平面坐标;An acquisition module 22, configured to acquire a current image including the first target object and the marker point, and the plane coordinates of the marker point in the current image;

匹配模块23,用于将所述当前图像中的标记点与所述标准图像的所述参考标记点进行匹配。这里要先将标记点在所述当前图像的平面坐标与参考标记点在标准图像的平面坐标之间建立对应关系,然后将标记点与参考标记点进行匹配。A matching module 23, configured to match the marker points in the current image with the reference marker points in the standard image. Here, it is first necessary to establish a corresponding relationship between the plane coordinates of the marker points in the current image and the plane coordinates of the reference marker points in the standard image, and then match the marker points with the reference marker points.

通过设置标准图像及参考标记点是便于在获取当前图像的空间位置时能有一个参照物,这进一步保证了本发明实施方式的目标跟踪装置的稳定性和准确性。By setting the standard image and the reference mark point, it is convenient to have a reference object when acquiring the spatial position of the current image, which further ensures the stability and accuracy of the target tracking device in the embodiment of the present invention.

进一步地,该跟踪设备30还包括:Further, the tracking device 30 also includes:

采集单元,用于采集所述标记点;a collection unit, configured to collect the marker points;

筛选单元,从所述标记点中筛选目标标记点。The screening unit is configured to screen target markers from the markers.

具体来说,当标记点的数量为多个时,采用摄像机采集对应第一目标对象的所有标记点,从所有标记点中选取与第一目标对象最相关的标记点,然后使用相应的图像处理算法对图像上的标记点进行提取,该提取需要根据标记点的特征来进行。总的来讲,对该标记点的特征进行提取的方法是对图像I使用特征提取函数H,获取图像中每个点的特征分数,并筛选出特征值足够高的标记点。这里可以用下述公式归纳表示:Specifically, when there are multiple marker points, a camera is used to collect all marker points corresponding to the first target object, and the marker point most relevant to the first target object is selected from all marker points, and then corresponding image processing is used to The algorithm extracts the marked points on the image, and the extraction needs to be carried out according to the characteristics of the marked points. Generally speaking, the method of extracting the features of the marked points is to use the feature extraction function H on the image I to obtain the feature score of each point in the image, and to filter out the marked points with sufficiently high feature values. Here it can be summed up by the following formula:

S(x,y)=H(I(x,y))S(x,y)=H(I(x,y))

F={arg(x,y)(S(x,y)>s0)}F={arg (x, y) (S(x, y)>s0)}

上述式中,H是特征提取函数,I(x,y)是每一个像素(x,y)所对应的图像值,可以是灰度值或三通道的颜色能量值等,S(x,y)是每一个像素(x,y)经过特征提取后的特征分数,s0是一个特征分数阈值,大于s0的S(x,y)可以被认为是标记点,F是标记点集合。较佳地,本发明实施例使用红外标记点以及红外摄像机所成图像的能量特征较为明显。由于使用窄带LED红外灯,以及相应的红外摄像机,摄像机所成图像的大多数像素能量很低,仅有标记点对应的像素具有高能量。因此相应的函数H(x,y)可以是对使用阈值分割算子后的图像B(x,y)进行区域增长获取若干子图像,并对获得到的子图像进行重心提取。同时,根据环境光中能在红外摄像机中成像的杂光,我们可在红外标记点提取过程中添加诸如标记点所成光斑面积,标记点在二维图像中的位置关系等约束条件对提取出的标记点进行筛选。In the above formula, H is the feature extraction function, I(x, y) is the image value corresponding to each pixel (x, y), which can be the gray value or the color energy value of the three channels, etc., S(x, y ) is the feature score of each pixel (x, y) after feature extraction, s0 is a feature score threshold, S(x, y) greater than s0 can be considered as a marker point, and F is a set of marker points. Preferably, the embodiment of the present invention uses infrared marker points and the energy characteristics of the image formed by the infrared camera are relatively obvious. Due to the use of narrow-band LED infrared lamps and corresponding infrared cameras, most of the pixels of the image formed by the camera have low energy, and only the pixels corresponding to the marking points have high energy. Therefore, the corresponding function H(x, y) can perform region growth on the image B(x, y) after using the threshold segmentation operator to obtain several sub-images, and extract the center of gravity of the obtained sub-images. At the same time, according to the stray light that can be imaged in the infrared camera in the ambient light, we can add constraints such as the area of the spot formed by the marker point and the positional relationship of the marker point in the two-dimensional image to the extraction process of the infrared marker point. mark points for filtering.

当摄像机数目大于1时,需要对不同摄像机在同一时刻或接近同一时刻获取的图像进行标记点匹配,从而为后续的标记点三维重建提供条件。标记点匹配的方法需要根据特征提取函数H而定。我们可以使用一些经典的基于图像灰度梯度图的特征点提取算子和与之相配的匹配方法如Harris,SIFT,FAST等方法获取并匹配标记点。也可以使用极限约束,标记点的先验条件等方式进行标记点匹配。这里使用极限约束进行匹配筛选的方法是:根据同一个点在两个不同摄像机图像上的投影都处于同一平面这个原理,对于某一个摄像机c0中的某一个标记点p0,我们都可以在其他摄像机c1中计算一条极线方程,标记点p0对应于该其它摄像机c1上的标记点p1符合下述关系:When the number of cameras is greater than 1, it is necessary to perform marker matching on images acquired by different cameras at or near the same time, so as to provide conditions for subsequent 3D reconstruction of markers. The method of mark point matching needs to be determined according to the feature extraction function H. We can use some classic feature point extraction operators based on image grayscale gradient maps and matching matching methods such as Harris, SIFT, FAST and other methods to obtain and match marker points. You can also use limit constraints, prior conditions for markers, etc. to match markers. The method of matching and screening using limit constraints here is: according to the principle that the projections of the same point on two different camera images are on the same plane, for a certain marker point p0 in a certain camera c0, we can use it in other cameras An epipolar line equation is calculated in c1, and the marker point p0 corresponds to the marker point p1 on the other camera c1 in accordance with the following relationship:

[p1;1]TF[p0;1]=0[p1;1] TF [p0;1]=0

上述式中,F是摄像机c0到摄像机c1的基础矩阵。通过使用上述关系,我们可以大大缩少标记点p1的候选个数,提高匹配准确度。In the above formula, F is the fundamental matrix of camera c0 to camera c1. By using the above relationship, we can greatly reduce the number of candidates for the marker point p1 and improve the matching accuracy.

此外,我们可以使用标记点的先验条件是标记点的空间顺序,标记点的尺寸等。比如根据两个摄像机的相互位置关系使其所拍摄的图像上的每一对对应同一空间点的两个像素在某一个维度比如y轴上相等,这个过程也被称为图像校准(rectification)。则此时标记点的匹配也就可以根据标记点的x轴顺序执行,即最小x对应最小x,依次类推,最大的x对应最大的x。In addition, we can use the prior conditions of the markers to be the spatial order of the markers, the dimensions of the markers, etc. For example, according to the mutual positional relationship of the two cameras, each pair of two pixels corresponding to the same spatial point on the captured image is equal in a certain dimension such as the y-axis. This process is also called image rectification. At this time, the matching of the marked points can also be performed according to the x-axis order of the marked points, that is, the smallest x corresponds to the smallest x, and so on, and the largest x corresponds to the largest x.

以下依据用于跟踪的摄像机数目的多少,来详细介绍本发明的目标跟踪装置。The object tracking device of the present invention will be introduced in detail below according to the number of cameras used for tracking.

请参见图7,图7示出了图5中的重新构建单元的具体结构示意图。如图7所示,在本实施例中,该跟踪设备30跟踪的第一目标对象对应的标记点不超过四个,且采用单目摄像机来获取标记点的位置信息时,重新构建单元进一步包括:Please refer to FIG. 7 , which shows a schematic structural diagram of the reconstruction unit in FIG. 5 . As shown in FIG. 7 , in this embodiment, the first target object tracked by the tracking device 30 corresponds to no more than four marker points, and when a monocular camera is used to obtain the position information of the marker points, the reconstruction unit further includes :

第一计算模块31,用于依据所述当前图像中的标记点的平面坐标与所述标准图像的所述参考标记点的平面坐标以及所述第一目标对象所在场景的假定条件计算所述当前图像与所述标准图像之间的单应变换关系;将当前图像的标记点与标准图像中的参考标记点进行匹配,并根据二者各自的平面坐标计算当前图像与标准图像之间的单应变换关系。所谓单应变换是对应几何中的单应性,是计算机视觉领域中常应用的一种变换方法。The first calculation module 31 is configured to calculate the current position according to the plane coordinates of the marker points in the current image, the plane coordinates of the reference marker points in the standard image, and the assumed conditions of the scene where the first target object is located. The homography transformation relationship between the image and the standard image; match the marker points of the current image with the reference marker points in the standard image, and calculate the homography between the current image and the standard image according to their respective plane coordinates transform relationship. The so-called homography transformation refers to the homography in corresponding geometry, and it is a transformation method often used in the field of computer vision.

第一重构模块32,用于根据所述单应变换关系计算出所述标记点在拍摄所述标准图像时刻的空间位置到当前时刻的空间位置的刚性变换,然后计算所述标记点在当前时刻的空间位置,并根据所述标记点在当前时刻的空间位置计算所述第一目标对象当前的空间位置。The first reconstruction module 32 is used to calculate the rigid transformation from the spatial position of the marker point at the moment when the standard image is taken to the spatial position at the current moment according to the homography transformation relationship, and then calculate the rigid transformation of the marker point at the current moment The spatial position at the moment, and calculate the current spatial position of the first target object according to the spatial position of the marker point at the current moment.

具体来讲,对于场景的假定条件,我们可以假定场景中的标记点的刚性变换时的某一维度的数值不变,比如三维空间场景中,空间坐标为x、y、z,x和y分别与摄像头的图像坐标(平面坐标)中x轴和y轴平行,z轴是垂直于摄像头的图像,假定条件可以是标记点在z轴上的坐标不变,也可以是标记点在x轴和/或y轴上的坐标不变。不同的场景假设条件,所使用的推测方法也不尽相同。又例如,在另一种假定条件下,假设第一目标对象的朝向与摄像头朝向之间的旋转角度在使用过程中始终保持不变,则可根据当前图像中的标记点互相之间的距离与标准图像上的标记点互相之间的距离之间的比例推测第一目标对象当前的空间位置。Specifically, for the assumed conditions of the scene, we can assume that the value of a certain dimension does not change during the rigid transformation of the marked points in the scene. For example, in a three-dimensional space scene, the spatial coordinates are x, y, z, and x and y are respectively It is parallel to the x-axis and y-axis in the image coordinates (plane coordinates) of the camera, and the z-axis is perpendicular to the image of the camera. The assumption can be that the coordinates of the marked point on the z-axis are unchanged, or that the marked point is on the x-axis and /or the coordinates on the y-axis are unchanged. Different scenario assumptions require different inference methods. For another example, under another assumption, assuming that the rotation angle between the orientation of the first target object and the orientation of the camera remains constant during use, the distances and The ratio of the distances between the marker points on the standard image estimates the current spatial position of the first target object.

通过以上的计算方法,可以实现单目摄像机对标记点的数量不超过四个时重新构建所述第一目标对象的空间位置,其操作简单,且跟踪结果也较准确,由于采用单目,降低了第一目标对象跟踪的成本。Through the above calculation method, the spatial position of the first target object can be reconstructed when the number of marker points of the monocular camera is not more than four, the operation is simple, and the tracking result is also relatively accurate. reduces the cost of tracking the first target object.

上述使用单个摄像机采集图像来恢复物体三维坐标的方法中,由于获取的图像信息较少,因此需要增加标记点的数目来提供更多的图像信息从而计算物体的三维坐标。根据机器视觉理论,要从单幅图像推算出场景的立体信息,需要至少确定图像中的五个标定点。因此,单目方案增加了标记点数量,也增加了设计的复杂性,但同时,只需要一个摄像机从而降低了图像采集的复杂性,降低了成本。In the above-mentioned method of recovering the three-dimensional coordinates of an object by collecting images with a single camera, since the acquired image information is less, it is necessary to increase the number of marker points to provide more image information to calculate the three-dimensional coordinates of the object. According to machine vision theory, to deduce the stereoscopic information of the scene from a single image, at least five calibration points in the image need to be determined. Therefore, the monocular solution increases the number of marking points and the complexity of the design, but at the same time, only one camera is needed, which reduces the complexity of image acquisition and reduces the cost.

请参见图8,图8示出了图5中的重新构建单元的第二变形实施例的具体结构示意图。如图7所示,在本实施例中,当所述标记点的数量为五个以上,且采用单目摄像机获取所述标记点的位置信息时,所述重新构建单元进一步包括:Please refer to FIG. 8 , which shows a schematic structural diagram of a second modified embodiment of the reconstruction unit in FIG. 5 . As shown in FIG. 7, in this embodiment, when the number of the marked points is more than five, and a monocular camera is used to obtain the position information of the marked points, the reconstruction unit further includes:

第二计算模块33,用于依据所述当前图像中的标记点的平面坐标与所述标准图像的所述参考标记点的平面坐标,计算所述当前图像与所述标准图像之间的单应变换关系。The second calculation module 33 is used to calculate the homography between the current image and the standard image according to the plane coordinates of the marker points in the current image and the plane coordinates of the reference marker points in the standard image transform relationship.

第二重构模块34,用于根据所述单应变换关系计算出所述标记点在拍摄所述标准图像时刻的空间位置到当前时刻的空间位置的刚性变换,然后计算所述标记点在当前时刻的空间位置,并根据所述标记点当前时刻的空间位置计算第一目标对象当前的空间位置。The second reconstruction module 34 is used to calculate the rigid transformation from the spatial position of the marker point at the moment when the standard image is taken to the spatial position at the current moment according to the homography transformation relationship, and then calculate the rigid transformation of the marker point at the current moment The spatial position at the moment, and calculate the current spatial position of the first target object according to the spatial position of the marker point at the current moment.

首先采集一幅标准图像,使用精确的深度摄像机或激光扫描仪等器件测量出参考标记点的空间位置,并获取此时的参考标记点的二维图像坐标(即平面坐标)。First, a standard image is collected, and the spatial position of the reference mark point is measured using an accurate depth camera or laser scanner, and the two-dimensional image coordinates (ie, plane coordinates) of the reference mark point at this time are obtained.

在使用过程中,摄像机不断捕捉含有第一目标对象的当前图像中的所有标记点的二维图像坐标,并根据此时二维图像坐标与标准图像参考标记点的二维坐标计算出当前状态下的标记点与拍摄标准图像时的标记点之间的刚性变换,在假设标记点之间相对位置不变的情况下,进而计算出出此时标记点相对于标准图像时的空间位置变换,从而计算出当前标记点的空间位置。During use, the camera continuously captures the two-dimensional image coordinates of all marker points in the current image containing the first target object, and calculates the current state based on the two-dimensional image coordinates and the two-dimensional coordinates of the standard image reference marker points at this time. The rigid transformation between the marked point and the marked point when shooting the standard image, assuming that the relative position between the marked points remains unchanged, and then calculate the spatial position transformation of the marked point relative to the standard image at this time, so that Calculate the spatial position of the current marker.

这里,使用五个点以上的点可以计算出当前标记点与拍摄标准图像时标记点的空间位置刚性变换[R|T],优选的,该五个以上的点不在一个平面上,且摄像头的投影矩阵P被事先标定好。计算[R|T]的具体方式如下:Here, the space position rigid transformation [R|T] between the current marker point and the marker point when shooting the standard image can be calculated by using more than five points. Preferably, the five or more points are not on the same plane, and the camera’s The projection matrix P is calibrated in advance. The specific way to calculate [R|T] is as follows:

每一个标记点在标准图像和当前图像的齐次坐标分别为X0,Xi。二者满足极限约束,即X0P-1[R|T]P=Xi。所有标记点组成一个未知参量为[R|T]的方程组。当标记点数量大于5时,可以对[R|T]求解;当标记点数量大于6时,可以对[R|T]求最优解,其方法可以使用奇异值分解SVD,和/或使用迭代的方法计算非线性最优解。当计算出标记点空间位置以后,我们可以根据事先标定好的标记点与第一目标对象(例如人眼)之间的相互位置关系推测出第一目标对象(如人眼)的空间位置。The homogeneous coordinates of each marked point in the standard image and the current image are X0, Xi respectively. Both satisfy the limit constraint, that is, X 0 P -1 [R|T]P=X i . All marked points form a system of equations whose unknown parameter is [R|T]. When the number of marked points is greater than 5, [R|T] can be solved; when the number of marked points is greater than 6, the optimal solution can be found for [R|T], and the method can use singular value decomposition SVD, and/or use The iterative method computes the nonlinear optimal solution. After calculating the spatial position of the marker point, we can infer the spatial position of the first target object (eg human eye) according to the mutual positional relationship between the pre-marked marker point and the first target object (eg human eye).

本实施例只用一个摄像机,使用五个或五个以上的标记点,就可以准确地构建出第一目标对象的空间位置,不仅操作简单,而且成本低廉。In this embodiment, only one camera and five or more marking points can be used to accurately construct the spatial position of the first target object, which is not only easy to operate, but also low in cost.

请参见图9,图9示出了图5中的重新构建单元的第三变形实施例的具体结构示意图。如图9所示,本实施例使用两个或两个以上的摄像机,一个或一个以上的标记点。采用双目摄像机或多目摄像机获取所述标记点的位置信息时,所述重新构建单元进一步包括:Please refer to FIG. 9 , which shows a schematic structural diagram of a third modified embodiment of the reconstruction unit in FIG. 5 . As shown in FIG. 9 , this embodiment uses two or more cameras and one or more marking points. When using a binocular camera or a multi-eye camera to obtain the position information of the marker point, the reconstruction unit further includes:

第三计算模块35,采用双目或多目三维重建原理,计算每一个标记点在当前时刻的空间位置;所谓双目或三目重建原理可以采用以下方法,例如采用左、右摄像头匹配的标记点之间的视差,计算每一个标记点在当前时刻的空间位置。或者是采用其它现有的常见方法来实现。The third calculation module 35 uses binocular or multi-eye three-dimensional reconstruction principles to calculate the spatial position of each marker point at the current moment; the so-called binocular or tri-eye reconstruction principles can use the following methods, for example, using the marks matched by the left and right cameras The parallax between points calculates the spatial position of each marker point at the current moment. Or use other existing common methods to achieve.

第三重构模块36,根据所述标记点当前时刻的空间位置计算第一目标对象当前的空间位置。The third reconstruction module 36 calculates the current spatial position of the first target object according to the current spatial position of the marker point.

具体地,首先使用多目摄像机校准的方法对每个摄像机之间的相互位置关系进行标定。然后在使用过程中,对每个摄像机获取到的图像提取标记点坐标,并对每个标记点进行匹配,即获取其在每个摄像机对应的标记点,然后使用匹配的标记点以及摄像机之间相互位置关系计算出标记点的空间位置。Specifically, firstly, a multi-camera calibration method is used to calibrate the mutual positional relationship between each camera. Then in the process of use, extract the coordinates of the marker points from the images acquired by each camera, and match each marker point, that is, obtain its corresponding marker point in each camera, and then use the matched marker points and between cameras The mutual positional relationship calculates the spatial position of the marker points.

在一个具体的例子中,使用多目摄像机(即摄像机数量大于等于2)来拍摄标记点,实现立体重建。已知一个标记点在某一摄像机所拍摄图像上的坐标u和该摄像机参数矩阵M,我们可以计算出一条射线,这个标记点在空间中处于此射线上。In a specific example, multi-cameras (that is, the number of cameras is greater than or equal to 2) are used to shoot marked points to achieve stereoscopic reconstruction. Knowing the coordinate u of a marker point on the image captured by a certain camera and the parameter matrix M of the camera, we can calculate a ray, and the marker point is located on this ray in space.

αjuj=MjXj=1…n(其中n为大于等于2的自然数)α j u j = M j Xj = 1...n (where n is a natural number greater than or equal to 2)

同理,依据上述公式,此标记点在其它摄像机上也可以计算出对应该其它摄像机的射线。理论上讲,这两条射线汇聚在一个点上,即此标记点的空间位置。实际上由于摄像机的数字化误差,摄像机内参和外参标定的误差等等,这些射线并不能汇聚于一点,因此需要使用三角测量(triangululation)的方法近似计算出标记点的空间位置。比如可以使用最小二乘判断准则确定距离所有光线最近的点作为物点。Similarly, according to the above formula, the marking point can also calculate the ray corresponding to the other camera on other cameras. Theoretically, these two rays converge on a point, which is the spatial position of this marked point. In fact, due to the digitization error of the camera, the calibration error of the internal and external parameters of the camera, etc., these rays cannot converge at one point, so it is necessary to use the method of triangulation to approximate the spatial position of the marker point. For example, the least square criterion can be used to determine the point closest to all rays as the object point.

Xx ′′ == argarg xx minmin ΣΣ jj == 11 mm [[ (( mm 11 jj Xx mm 33 jj Xx -- uu 11 jj )) 22 ++ (( mm 22 jj Xx mm 44 jj Xx -- uu 22 jj )) 22 ]]

当计算出标记点空间位置以后,我们可以根据事先标定好的标记点与第一目标对象(如人眼)之间的相互位置关系推测出第一目标对象(人眼)的空间位置。After calculating the spatial position of the marker point, we can infer the spatial position of the first target object (human eye) according to the mutual positional relationship between the pre-marked marker point and the first target object (such as the human eye).

在上述使用多目摄像机实现立体重建的方法中,较佳的方法是使用双目相机计算。其原理与前述多目摄像机重建原理一样,都是根据两摄像机的相互位置关系以及标记点在两摄像机成像的二维坐标计算标记点空间位置。其微小差别是双目摄像机平行摆放,根据简单的标定后对两个摄像机的图像做如前文所述的图像校准,使得两个互相匹配的二维标记点在y(或x)轴上相等,则此时标记点距摄像机的深度可由校准后的二维标记点在x(或y)轴上的差距计算得出。此方法可看成是多目立体重建在双目情况下的特殊方法,其简化了立体重建的步骤且在设备硬件上较容易实现。Among the above methods for realizing stereo reconstruction using multi-cameras, a better method is to use binocular cameras for calculation. The principle is the same as the aforementioned multi-camera reconstruction principle, which is to calculate the spatial position of the marker point based on the mutual positional relationship between the two cameras and the two-dimensional coordinates of the marker point imaged by the two cameras. The slight difference is that the binocular cameras are placed in parallel, and the images of the two cameras are calibrated as described above after simple calibration, so that the two matching two-dimensional marker points are equal on the y (or x) axis , then the depth from the marker point to the camera can be calculated from the distance between the calibrated two-dimensional marker points on the x (or y) axis. This method can be regarded as a special method for multi-eye stereo reconstruction in the case of binoculars, which simplifies the steps of stereo reconstruction and is easier to implement on the device hardware.

实施例4Example 4

请参见图10,图10示出了图5的跟踪装置中对应第一目标对象设置标记点的定位支架的结构示意图。如图10所示,本发明提供一种定位支架,该定位支架位于人眼(第一目标对象)前方,结构与眼镜相似,其佩戴类似于眼镜,包括:横梁11、固定部12、支撑部13以及控制部14,横梁11设置有标记点111;支撑部13设置于横梁11上;固定部12与横梁11的端部枢轴连接。其中标记点111设置的位置与人眼(第一目标对象)的位置相对应,通过获取标记点111的空间位置信息,然后据此计算人眼的空间位置信息。当人的头部发生移动时,相应地,与人眼对应的标记点111也发生移动,摄像机跟踪标记点111的移动,然后采用前述实施方式一的目标对象跟踪方法的方案获取标记点111的空间位置信息,利用标记点111与人眼的相对空间位置关系,重新构建出人眼(第一目标对象)的空间位置(即在空间中的三维坐标)。Please refer to FIG. 10 . FIG. 10 shows a schematic structural diagram of a positioning bracket for setting a marker point corresponding to a first target object in the tracking device in FIG. 5 . As shown in Figure 10, the present invention provides a positioning bracket, the positioning bracket is located in front of the human eye (the first target object), the structure is similar to glasses, and its wearing is similar to glasses, including: a beam 11, a fixing part 12, a supporting part 13 and the control part 14, the beam 11 is provided with a marking point 111; the support part 13 is provided on the beam 11; the fixing part 12 is pivotally connected to the end of the beam 11. The position of the marking point 111 corresponds to the position of the human eye (the first target object), and the spatial position information of the human eye is calculated by acquiring the spatial position information of the marking point 111 . When the head of the person moves, correspondingly, the marker point 111 corresponding to the human eye also moves, the camera tracks the movement of the marker point 111, and then uses the target object tracking method of the first embodiment to obtain the image of the marker point 111 The spatial location information uses the relative spatial location relationship between the marker point 111 and the human eye to reconstruct the spatial location (ie, the three-dimensional coordinates in space) of the human eye (the first target object).

在本实施例中,横梁11为一长条形,且具有一定的弧度,其弧度与人的额头弧度近似,以方便使用。横梁11包括上表面112、与其相对的下表面、设置在上表面112与下表面之间的第一表面114以及第二表面。In this embodiment, the crossbeam 11 is long and has a certain curvature, which is similar to the curvature of the forehead of a person, so that it is convenient to use. The beam 11 includes an upper surface 112 , a lower surface opposite thereto, a first surface 114 and a second surface disposed between the upper surface 112 and the lower surface.

在本实施例中,标记点111为三个LED灯,其间隔均匀地设置在横梁11的第一表面114上。可以理解的是,标记点111也可以为一个、两个或者更多个,且可以为任意光源,包括LED灯、红外灯或紫外灯等。并且,所述标记点111的排列方式与设置位置也可以根据需要进行调整。In this embodiment, the marking points 111 are three LED lights, which are evenly spaced on the first surface 114 of the beam 11 . It can be understood that there may be one, two or more marking points 111, and may be any light source, including LED lamps, infrared lamps, or ultraviolet lamps. Moreover, the arrangement and setting position of the marking points 111 can also be adjusted as required.

可以理解的是,横梁11也可以根据需要设计成直线形或者其他形状。It can be understood that the beam 11 can also be designed as a straight line or other shapes as required.

本实施例中,固定部12有两个,分别与横梁11的两端通过枢轴连接,且两个固定部12可相对向内折叠,同时,两个固定部12可分别向外展开至与横梁11呈100°左右的内角,具体的,可以根据实际操作需求调整内角的大小。可以理解的,固定部12也可以为一个。In this embodiment, there are two fixing parts 12, which are respectively pivotally connected to the two ends of the beam 11, and the two fixing parts 12 can be relatively inwardly folded, and at the same time, the two fixing parts 12 can be respectively expanded to the The beam 11 has an inner angle of about 100°, specifically, the size of the inner angle can be adjusted according to actual operation requirements. It can be understood that there may also be one fixing part 12 .

固定部12远离横梁11的一端沿支撑部13的延伸方向弯折设置,以用于将固定部12的端部固定于人的耳朵上。An end of the fixing part 12 away from the beam 11 is bent along the extending direction of the supporting part 13 for fixing the end of the fixing part 12 on the ear of a person.

本实施例中,支撑部13呈条状,设置在横梁11的下表面113的中部且向下延伸。进一步,支撑部13远离横梁11的端部设置有鼻托131,以用于将定位装置配合鼻梁,并将定位装置设置于人眼上方。可以理解的,在其他实施例中,若不设置鼻托131,则支撑部13可设置成倒“Y”型,并沿横梁11的中部且向下延伸,用以将定位装置配合鼻梁,并将定位装置设置于人眼上方。In this embodiment, the supporting portion 13 is in the shape of a strip, disposed in the middle of the lower surface 113 of the beam 11 and extending downward. Further, a nose pad 131 is provided at the end of the supporting part 13 far away from the cross beam 11, for fitting the positioning device with the bridge of the nose, and setting the positioning device above the human eyes. It can be understood that, in other embodiments, if the nose pad 131 is not provided, the support part 13 can be arranged in an inverted "Y" shape and extend downward along the middle of the beam 11 to match the positioning device with the bridge of the nose, and The positioning device is arranged above the human eyes.

控制部14呈圆角长方体,设置在固定部12上。控制部14用于给所述LED灯,红外灯或紫外灯提供电源、及/或者控制所述LED灯,红外灯或紫外灯的使用状态,其包括电源开关141、电源指示灯和充电指示灯。可以理解的是,控制部14不限定形状,其可以为任意形状,也可以为一集成芯片。并且,控制部14也可以设置在其他位置,如横梁11上。The control part 14 is in the form of a rectangular parallelepiped with rounded corners and is disposed on the fixing part 12 . The control unit 14 is used to provide power to the LED lamp, infrared lamp or ultraviolet lamp, and/or control the use state of the LED lamp, infrared lamp or ultraviolet lamp, which includes a power switch 141, a power indicator light and a charging indicator light . It can be understood that the shape of the control unit 14 is not limited, it can be in any shape, and can also be an integrated chip. Moreover, the control part 14 can also be arranged at other positions, such as on the beam 11 .

使用时,打开电源开关141,电源指示灯显示LED处于供电状态,LED灯被点亮;当电量不足时,充电指示灯提示电量不足;关闭电源开关,电源指示灯熄灭,表示LED处于关闭状态,LED灯熄灭。During use, turn on the power switch 141, the power indicator shows that the LED is in the power supply state, and the LED light is lit; The LED light turns off.

由于人的瞳距范围为58mm~64mm,可近似认为人的瞳距为定值,本发明提供的定位支架类似于眼镜架,且固定于人眼上方,类似于眼镜架,根据需要,将标记点设置在定位装置的预定位置,从而可以简单方便地根据标记点的位置确定人眼的位置。定位装置结构简单,设计与使用方便。Since the interpupillary distance ranges from 58mm to 64mm, it can be approximately considered as a fixed value. The positioning bracket provided by the present invention is similar to a spectacle frame, and is fixed above the human eye, similar to a spectacle frame. The point is set at the predetermined position of the positioning device, so that the position of the human eye can be determined simply and conveniently according to the position of the marked point. The positioning device is simple in structure and convenient in design and use.

实施方式二Implementation Mode Two

请参见图11至图14,图11是本发明实施方式二的基于软镜的立体显示方法的流程示意图,图12是图11中S1的具体流程示意图,图13是图12中S12的具体流程示意图,图14是图11中的S3的具体流程示意图。如图11至图14所示,本发明实施方式二的基于软镜的立体显示方法,主要包括以下步骤:Please refer to FIG. 11 to FIG. 14 , FIG. 11 is a schematic flowchart of a soft mirror-based stereoscopic display method according to Embodiment 2 of the present invention, FIG. 12 is a specific flowchart of S1 in FIG. 11 , and FIG. 13 is a specific flowchart of S12 in FIG. 12 Schematic diagram, FIG. 14 is a specific flowchart of S3 in FIG. 11 . As shown in Fig. 11 to Fig. 14, the stereoscopic display method based on the soft mirror according to the second embodiment of the present invention mainly includes the following steps:

S0图像拍摄步骤,拍摄第二目标对象的立体图像,并实时发送拍摄到的所述第二目标对象的立体图像的信息,该信息包括左视图信息和右视图信息。S0 image capturing step, capturing a stereoscopic image of the second target object, and sending information of the captured stereoscopic image of the second target object in real time, the information including left view information and right view information.

S1获取第一目标对象的位置信息;利用跟踪设备跟踪第一目标对象的位置,例如观看者所在的位置信息。S1 acquires location information of a first target object; uses a tracking device to track the location of the first target object, such as location information of a viewer.

S2获取立体显示装置的分光单元的光栅参数及显示单元的显示参数;分光单元的光栅参数主要包括光栅的栅距(pitch)和光栅相对显示面板的倾斜角度、光栅相对显示面板的放置距离等参数。S2 acquires the grating parameters of the light-splitting unit of the stereoscopic display device and the display parameters of the display unit; the grating parameters of the light-splitting unit mainly include the grating pitch (pitch) of the grating, the inclination angle of the grating relative to the display panel, and the placement distance of the grating relative to the display panel. .

S3依据所述位置信息和所述光栅参数及所述显示参数将接收到的所述图像拍摄单元拍摄到的立体图像进行实时处理。在待播放立体图像前,需要预先结合人眼的位置信息和光栅参数及显示单元的显示参数,对图像进行处理,以便提供给观看者最佳的立体显示效果。S3 Perform real-time processing on the received stereoscopic image captured by the image capturing unit according to the position information, the grating parameter and the display parameter. Before the stereoscopic image is to be played, it is necessary to pre-process the image in combination with the position information of human eyes, the grating parameters and the display parameters of the display unit, so as to provide the best stereoscopic display effect to the viewer.

S4显示该待播放的图像。S4 displays the image to be played.

本发明的基于软镜的立体显示方法,通过及时获取到第一目标对象的位置信息及光栅参数,并据此直接进行图像处理,提高了图像播放的速度,能满足实时立体显示的要求,方便医生手术操作及辅助医生提高手术成功率的优点。The three-dimensional display method based on the soft mirror of the present invention obtains the position information and grating parameters of the first target object in time, and directly performs image processing accordingly, improves the speed of image playback, can meet the requirements of real-time three-dimensional display, and is convenient The advantages of the doctor's surgical operation and assisting the doctor to improve the success rate of the operation.

另外,这里的第二目标对象主要指摄像机拍摄到的各种场景,可以是实际的人,或者正在直播的球赛或者借助一些设备拍摄的病人体内的影像等。通过实时地拍摄立体图像,并将拍摄到的立体图像实时显示在显示单元上,无需经过额外的图像处理,及时并真实地显示拍摄到的各种场景,满足了用户对实时显示的需求,提高了用户体验。In addition, the second target object here mainly refers to various scenes captured by the camera, which may be actual people, or a live broadcast of a football game, or an image of a patient's body captured by some equipment. By shooting stereoscopic images in real time and displaying the captured stereoscopic images on the display unit in real time, without additional image processing, various scenes captured can be displayed in a timely and true manner, which meets the needs of users for real-time display and improves user experience.

在一个具体的变形实施例中,上述步骤S0还包括:图像采集步骤,采集所述第二目标对象的立体图像,并从所述立体图像中提取出左视图信息和右视图信息。通过在边拍摄立体图像时边提取立体图像的左视图信息和右视图信息,提高了图像处理的速度,保证了实时进行立体显示的显示效果。In a specific variant embodiment, the above step S0 further includes: an image acquisition step, acquiring a stereoscopic image of the second target object, and extracting left view information and right view information from the stereoscopic image. By extracting the left view information and the right view information of the stereoscopic image while shooting the stereoscopic image, the image processing speed is improved, and the display effect of real-time stereoscopic display is guaranteed.

实施例5Example 5

请参见图12,本发明实施例5主要是对S1如何获取第一目标对象的位置信息作详细描述。这些第一目标对象例如为人眼、人的头部、人的面部或者人体的上半身等与人观看相关的部位。上述“S1获取第一目标对象的位置信息”主要包括以下步骤:Referring to FIG. 12 , Embodiment 5 of the present invention mainly describes in detail how S1 obtains the location information of the first target object. These first target objects are, for example, parts related to human viewing, such as human eyes, human head, human face, or upper body of human body. The above-mentioned "S1 obtaining the location information of the first target object" mainly includes the following steps:

S11对应第一目标对象的空间位置设置标记点;这里的标记点可以设置第一目标对象上,也可以不设置于第一目标对象上,而是设置在与第一目标对象有相对位置关系,并与第一目标对象同步移动的物体上也可。例如,目标对象是人眼,则可以在人眼的眼眶周围设置标记点;或者在人眼周围配置定位支架,将标记点设在定位支架的框上,又或者将标记点设在与人眼位置关系相对固定的人的耳朵上。该标记点可以是发送信号的红外发射传感器,LED灯,GPS传感器,激光定位传感器等各种部件,也可以是其它能够被摄像机捕捉的物理标识,例如是具有形状特征和/或颜色特征的物体。较佳地,为避免外界杂光的干扰,提高标记点追踪的鲁棒性,优选使用频谱较为狭窄的红外LED灯作为标记点,并使用只能通过红外LED所使用频谱的相应的红外摄像机对标记点进行捕捉。考虑到外界杂光多为不规则的形状且亮度分布不均匀,可以将标记点设置成可以发出规则形状的光斑,发光强度较高,亮度均匀。另外还可以设置多个标记点,每个标记点对应一个光斑,各个标记点组成规则的几何形状,如三角形,四边形等,从而易于跟踪到标记点,获得标记点的空间位置信息,并提高光斑提取的准确性。S11 setting a mark point corresponding to the spatial position of the first target object; the mark point here can be set on the first target object, or not set on the first target object, but set in a relative position relationship with the first target object, And also on an object that moves synchronously with the first target object. For example, if the target object is the human eye, marking points can be set around the orbit of the human eye; or a positioning bracket is arranged around the human eye, and the marking point is set on the frame of the positioning bracket; On the ear of a person whose positional relationship is relatively fixed. The marking point can be various components such as infrared emission sensors that send signals, LED lights, GPS sensors, laser positioning sensors, etc., or other physical signs that can be captured by the camera, such as objects with shape characteristics and/or color characteristics . Preferably, in order to avoid the interference of external stray light and improve the robustness of marker point tracking, it is preferable to use infrared LED lights with a relatively narrow spectrum as marker points, and use a corresponding infrared camera that can only pass through the spectrum used by infrared LEDs to Mark points to snap to. Considering that the external stray light is mostly irregular in shape and the brightness distribution is uneven, the marking points can be set to emit regular-shaped light spots, with high luminous intensity and uniform brightness. In addition, multiple marking points can be set, each marking point corresponds to a light spot, and each marking point forms a regular geometric shape, such as a triangle, a quadrilateral, etc., so that it is easy to track the marking point, obtain the spatial position information of the marking point, and improve the light spot. Extraction accuracy.

S12获取该标记点的位置信息;这可以是通过接收标记点发出的信号,来确定标记点的位置信息,也可以是采用摄像机来拍摄含有标记点的图像,对图像中的标记点进行提取。通过图像处理算法来获得标记点的位置信息。S12 Obtain the position information of the mark point; this may be to determine the position information of the mark point by receiving the signal sent by the mark point, or it may be to use a camera to capture an image containing the mark point, and extract the mark point in the image. The location information of the marked points is obtained through image processing algorithms.

S13依据该标记点的位置信息,重新构建该第一目标对象的空间位置。当获得到该标记点的位置信息后,重新构建出标记点的空间位置,然后依据标记点与第一目标对象的相对位置关系,将标记点的空间位置转换到第一目标对象的空间位置(例如人的左右两眼的空间位置)。S13 Reconstruct the spatial position of the first target object according to the position information of the marked point. After obtaining the position information of the mark point, the spatial position of the mark point is reconstructed, and then according to the relative positional relationship between the mark point and the first target object, the spatial position of the mark point is converted to the spatial position of the first target object ( For example, the spatial position of the left and right eyes of a person).

本发明实施方式二的通过获取对应第一目标对象的标记点的位置信息,并依据该位置信息,重新构建出第一目标对象的空间位置。与现有技术中使用摄像机作为人眼捕捉设备需要对二维图像进行特征分析从而获取人眼位置或者使用其他利用人眼虹膜反射效果的人眼捕捉设备获取人眼位置相比较,具有稳定性好,捕捉人眼的位置信息的准确度高、成本低廉且对跟踪设备与第一目标对象之间的距离远近没有要求等优点。Embodiment 2 of the present invention obtains the position information of the marked point corresponding to the first target object, and reconstructs the spatial position of the first target object according to the position information. Compared with the existing technology that uses a camera as a human eye capture device and needs to perform feature analysis on a two-dimensional image to obtain the position of the human eye or use other human eye capture devices that use the reflection effect of the iris of the human eye to obtain the position of the human eye, it has good stability. , the accuracy of capturing the position information of human eyes is high, the cost is low, and there is no requirement for the distance between the tracking device and the first target object.

请参见图13,上述步骤S12进一步包括:Please refer to Figure 13, the above step S12 further includes:

S121预设一标准图像,所述标准图像中设有参考标记点,并获取所述参考标记点的空间坐标和平面坐标;标准图像例如可以是通过图像采集设备采集的一个标准图像,获取参考标记点的图像坐标,并使用其他精确的立体测量设备如激光扫描仪,结构光扫描仪(如Kinect等)等设备获取标准图像中参考标记点的空间坐标及平面坐标。S121 preset a standard image, the reference mark point is set in the standard image, and the spatial coordinates and plane coordinates of the reference mark point are obtained; the standard image can be, for example, a standard image collected by an image acquisition device, and the reference mark is obtained The image coordinates of the points, and use other accurate stereo measurement equipment such as laser scanners, structured light scanners (such as Kinect, etc.) to obtain the spatial coordinates and plane coordinates of the reference marker points in the standard image.

S122获取包含所述目标对象及所述标记点的当前图像,及所述标记点在所述当前图像的平面坐标;S122 Acquire the current image including the target object and the marker point, and the plane coordinates of the marker point in the current image;

S123将所述当前图像中的标记点与所述标准图像的所述参考标记点进行匹配。这里要先将标记点在所述当前图像的平面坐标与参考标记点在标准图像的平面坐标之间建立对应关系,然后将标记点与参考标记点进行匹配。S123 Match the marker points in the current image with the reference marker points in the standard image. Here, it is first necessary to establish a corresponding relationship between the plane coordinates of the marker points in the current image and the plane coordinates of the reference marker points in the standard image, and then match the marker points with the reference marker points.

通过设置标准图像及参考标记点是便于在获取当前图像的空间位置时能有一个参照物,这进一步保证了本发明实施方式的目标跟踪方法的稳定性和准确性。By setting the standard image and the reference mark point, it is convenient to have a reference object when acquiring the spatial position of the current image, which further ensures the stability and accuracy of the target tracking method in the embodiment of the present invention.

进一步地,在上述步骤S11之前还包括:S10对用于获取所述标记点的位置信息的摄像机进行标定。Further, before the above step S11, the method further includes: S10 performing calibration on the camera used to obtain the position information of the marking point.

上述标定有分以下几种情况:The above calibration has the following situations:

(1)所述S10的摄像机为单目摄像机时,可以采用常见的张氏棋盘格标定算法,例如采用以下公式进行标定:(1) When the camera of the S10 is a monocular camera, the common Zhang’s checkerboard calibration algorithm can be used, for example, the following formula is used for calibration:

sm′=A[R|t]M′(1)sm'=A[R|t]M'(1)

式(1)中,A为内部参数,R为外部参数,t为平移向量,m’像点在图像中的坐标,M’为物点的空间坐标(即在空间中的三维坐标);其中A、R和t分别由以下公式确定:In formula (1), A is an internal parameter, R is an external parameter, t is a translation vector, m' is the coordinate of the image point in the image, and M' is the space coordinate of the object point (that is, the three-dimensional coordinate in space); where A, R and t are respectively determined by the following formulas:

A = fx 0 cx 0 fy cy 0 0 1 R = r 11 r 12 r 13 r 21 r 22 r 23 r 31 r 32 r 33 , 平移向量 t = t 1 t 2 t 3 . A = fx 0 cx 0 fy cy 0 0 1 and R = r 11 r 12 r 13 r twenty one r twenty two r twenty three r 31 r 32 r 33 , translation vector t = t 1 t 2 t 3 .

当然对于摄像机的标定算法有很多种,还可以采用其它业界常用的标定算法,本发明不作限定,主要是运用标定算法,以提高本发明的第一目标对象跟踪方法的准确性。Of course, there are many kinds of camera calibration algorithms, and other commonly used calibration algorithms in the industry can also be used. The present invention does not limit it. The calibration algorithm is mainly used to improve the accuracy of the first object tracking method of the present invention.

(2)所述S10的摄像机为双目摄像机或多目摄像机时,采用以下步骤进行标定:(2) When the camera of the S10 is a binocular camera or a multi-eye camera, the following steps are used for calibration:

S101先对所述双目摄像机或多目摄像机中的任一目摄像机进行标定,也是采用常见的张氏棋盘格标定算法,例如采用以下公式:S101 first calibrates any one of the binocular cameras or multi-cameras, using the common Zhang checkerboard calibration algorithm, for example, using the following formula:

sm′=A[R|t]M′(1)sm'=A[R|t]M'(1)

式(1)中,A为内部参数,R为外部参数,t为平移向量,m’像点在图像中的坐标,M’为物点的空间坐标;其中A、R和t分别由以下公式确定:In formula (1), A is an internal parameter, R is an external parameter, t is a translation vector, m' is the coordinate of the image point in the image, and M' is the space coordinate of the object point; where A, R and t are determined by the following formula Sure:

A = fx 0 cx 0 fy cy 0 0 1 R = r 11 r 12 r 13 r 21 r 22 r 23 r 31 r 32 r 33 , 平移向量 t = t 1 t 2 t 3 ; A = fx 0 cx 0 fy cy 0 0 1 and R = r 11 r 12 r 13 r twenty one r twenty two r twenty three r 31 r 32 r 33 , translation vector t = t 1 t 2 t 3 ;

S102计算所述双目摄像机或所述多目摄像机之间的相对旋转矩阵和相对平移量,采用以下公式:S102 Calculate the relative rotation matrix and relative translation between the binocular camera or the multi-camera, using the following formula:

相对旋转矩阵 R ′ = r 11 r 12 r 13 r 21 r 22 r 23 r 31 r 32 r 33 和相对平移量 T = T 1 T 2 T 3 . relative rotation matrix R ′ = r 11 r 12 r 13 r twenty one r twenty two r twenty three r 31 r 32 r 33 and the relative translation T = T 1 T 2 T 3 .

当然上述针对双目摄像机或多目摄像机的标定算法只是其中较常见的一种,还可以采用其它业界常用的标定算法,本发明不作限定,主要是运用标定算法,以提高本发明的第一目标对象跟踪方法的准确性。Of course, the above-mentioned calibration algorithm for binocular cameras or multi-eye cameras is only one of the more common ones, and other commonly used calibration algorithms in the industry can also be used. The present invention is not limited, mainly using calibration algorithms to improve the first goal of the present invention Accuracy of object tracking methods.

进一步地,在上述S11与S12之间还包括:Further, between the above S11 and S12, it also includes:

S14采集所述标记点;S14 collects the mark points;

S15从所述标记点中筛选目标标记点。S15 Screen target markers from the markers.

具体来说,当标记点的数量为多个时,采用摄像机采集对应第一目标对象的所有标记点,从所有标记点中选取与第一目标对象最相关的标记点,然后使用相应的图像处理算法对图像上的标记点进行提取,该提取需要根据标记点的特征来进行。总的来讲,对该标记点的特征进行提取的方法是对图像I使用特征提取函数H,获取图像中每个点的特征分数,并筛选出特征值足够高的标记点。这里可以用下述公式归纳表示:Specifically, when there are multiple marker points, a camera is used to collect all marker points corresponding to the first target object, and the marker point most relevant to the first target object is selected from all marker points, and then corresponding image processing is used to The algorithm extracts the marked points on the image, and the extraction needs to be carried out according to the characteristics of the marked points. Generally speaking, the method of extracting the features of the marked points is to use the feature extraction function H on the image I to obtain the feature score of each point in the image, and to filter out the marked points with sufficiently high feature values. Here it can be summed up by the following formula:

s(x,y)=H(I(x,y))s(x,y)=H(I(x,y))

F={arg(x,y)(S(x,y)>s0)}F={arg (x, y) (S(x, y)>s0)}

上述式中,H是特征提取函数,I(x,y)是每一个像素(x,y)所对应的图像值,可以是灰度值或三通道的颜色能量值等,S(x,y)是每一个像素(x,y)经过特征提取后的特征分数,s0是一个特征分数阈值,大于s0的S(x,y)可以被认为是标记点,F是标记点集合。较佳地,本发明实施例使用红外标记点以及红外摄像机所成图像的能量特征较为明显。由于使用窄带LED红外灯,以及相应的红外摄像机,摄像机所成图像的大多数像素能量很低,仅有标记点对应的像素具有高能量。因此相应的函数H(x,y)可以是对使用阈值分割算子后的图像B(x,y)进行区域增长获取若干子图像,并对获得到的子图像进行重心提取。该特征提取函数H(x,y),可以是Harris,SIFT,FAST等特征点函数,也可以是圆形光斑提取等图像处理函数。同时,根据环境光中能在红外摄像机中成像的杂光,我们可在红外标记点提取过程中添加诸如标记点所成光斑面积,标记点在二维图像中的位置关系等约束条件对提取出的标记点进行筛选。In the above formula, H is the feature extraction function, I(x, y) is the image value corresponding to each pixel (x, y), which can be the gray value or the color energy value of the three channels, etc., S(x, y ) is the feature score of each pixel (x, y) after feature extraction, s0 is a feature score threshold, S(x, y) greater than s0 can be considered as a marker point, and F is a set of marker points. Preferably, the embodiment of the present invention uses the infrared marker points and the energy characteristics of the image formed by the infrared camera to be relatively obvious. Due to the use of narrow-band LED infrared lamps and corresponding infrared cameras, most of the pixels of the image formed by the camera have low energy, and only the pixels corresponding to the marking points have high energy. Therefore, the corresponding function H(x, y) can perform region growth on the image B(x, y) after using the threshold segmentation operator to obtain several sub-images, and extract the center of gravity of the obtained sub-images. The feature extraction function H(x, y) may be a feature point function such as Harris, SIFT, or FAST, or an image processing function such as circular spot extraction. At the same time, according to the stray light that can be imaged in the infrared camera in the ambient light, we can add constraints such as the area of the spot formed by the marker point and the positional relationship of the marker point in the two-dimensional image to the extraction process of the infrared marker point. mark points for filtering.

当摄像机数目大于1时,需要对不同摄像机在同一时刻或接近同一时刻获取的图像进行标记点匹配,从而为后续的标记点三维重建提供条件。标记点匹配的方法需要根据特征提取函数H而定。我们可以使用一些经典的基于图像灰度梯度图的特征点提取算子和与之相配的匹配方法如Harris,SIFT,FAST等方法获取并匹配标记点。也可以使用极限约束,标记点的先验条件等方式进行标记点匹配。这里使用极限约束进行匹配筛选的方法是:根据同一个点在两个不同摄像机图像上的投影都处于同一平面这个原理,对于某一个摄像机c0中的某一个标记点p0,我们都可以在其他摄像机c1中计算一条极线方程,标记点p0对应于该其它摄像机c1上的标记点p1符合下述关系:When the number of cameras is greater than 1, it is necessary to perform marker matching on images acquired by different cameras at or near the same time, so as to provide conditions for subsequent 3D reconstruction of markers. The method of mark point matching needs to be determined according to the feature extraction function H. We can use some classic feature point extraction operators based on image grayscale gradient maps and matching matching methods such as Harris, SIFT, FAST and other methods to obtain and match marker points. You can also use limit constraints, prior conditions for markers, etc. to match markers. The method of matching and screening using limit constraints here is: according to the principle that the projections of the same point on two different camera images are on the same plane, for a certain marker point p0 in a certain camera c0, we can use it in other cameras An epipolar line equation is calculated in c1, and the marker point p0 corresponds to the marker point p1 on the other camera c1 in accordance with the following relationship:

[p1;1]TF[p0;1]=0[p1;1] TF [p0;1]=0

上述式中,F是摄像机c0到摄像机c1的基础矩阵。通过使用上述关系,我们可以大大缩少标记点p1的候选个数,提高匹配准确度。In the above formula, F is the fundamental matrix of camera c0 to camera c1. By using the above relationship, we can greatly reduce the number of candidates for the marker point p1 and improve the matching accuracy.

此外,我们可以使用标记点的先验条件是标记点的空间顺序,标记点的尺寸等。比如根据两个摄像机的相互位置关系使其所拍摄的图像上的每一对对应同一空间点的两个像素在某一个维度比如y轴上相等,这个过程也被称为图像校准(rectification)。则此时标记点的匹配也就可以根据标记点的x轴顺序执行,即最小x对应最小x,依次类推,最大的x对应最大的x。In addition, we can use the prior conditions of the markers to be the spatial order of the markers, the dimensions of the markers, etc. For example, according to the mutual positional relationship of the two cameras, each pair of two pixels corresponding to the same spatial point on the captured image is equal in a certain dimension such as the y-axis. This process is also called image rectification. At this time, the matching of the marked points can also be performed according to the x-axis order of the marked points, that is, the smallest x corresponds to the smallest x, and so on, and the largest x corresponds to the largest x.

以下依据用于跟踪的摄像机数目的多少,来详细介绍本发明的目标跟踪方法。The object tracking method of the present invention will be introduced in detail below according to the number of cameras used for tracking.

请参见图14,是图11中S13的第一变形例的具体流程示意图。如图13所示,在本实施例中,该第一目标对象跟踪方法跟踪的第一目标对象对应的标记点不超过四个,且采用单目摄像机来获取标记点的位置信息时,前述步骤S13进一步包括:Please refer to FIG. 14 , which is a schematic flowchart of a first modified example of S13 in FIG. 11 . As shown in Figure 13, in this embodiment, the first target object tracked by the first target object tracking method corresponds to no more than four marker points, and when a monocular camera is used to obtain the position information of the marker points, the preceding steps S13 further includes:

S131依据所述当前图像中的标记点的平面坐标与所述标准图像的所述参考标记点的平面坐标以及所述第一目标对象所在场景的假定条件计算所述当前图像与所述标准图像之间的单应变换关系;将当前图像的标记点与标准图像中的参考标记点进行匹配,并根据二者各自的平面坐标计算当前图像与标准图像之间的单应变换关系。所谓单应变换是对应几何中的单应性,是计算机视觉领域中常用的一种变换方法。S131 Calculate the difference between the current image and the standard image according to the plane coordinates of the marker points in the current image, the plane coordinates of the reference marker points in the standard image, and the assumed conditions of the scene where the first target object is located. The homography transformation relationship between the current image and the reference marker point in the standard image is matched, and the homography transformation relationship between the current image and the standard image is calculated according to their respective plane coordinates. The so-called homography transformation refers to the homography in corresponding geometry, and it is a transformation method commonly used in the field of computer vision.

S132根据所述单应变换关系计算出所述标记点在拍摄所述标准图像时刻的空间位置到当前时刻的空间位置的刚性变换,然后计算所述标记点在当前时刻的空间位置,并根据所述标记点在当前时刻的空间位置计算所述第一目标对象当前的空间位置。S132 Calculate the rigid transformation from the spatial position of the marker point at the moment when the standard image is captured to the spatial position at the current moment according to the homography transformation relationship, and then calculate the spatial position of the marker point at the current moment, and according to the The current spatial position of the first target object is calculated based on the spatial position of the marker point at the current moment.

具体来讲,对于场景的假定条件,我们可以假定场景中的标记点的刚性变换时的某一维度的数值不变,比如三维空间场景中,空间坐标为x、y、z,x和y分别与摄像头的图像坐标(平面坐标)中x轴和y轴平行,z轴是垂直于摄像头的图像,假定条件可以是标记点在z轴上的坐标不变,也可以是标记点在x轴和/或y轴上的坐标不变。不同的场景假设条件,所使用的推测方法也不尽相同。又例如,在另一种假定条件下,假设第一目标对象的朝向与摄像头朝向之间的旋转角度在使用过程中始终保持不变,则可根据当前图像中的标记点互相之间的距离与标准图像上的标记点互相之间的距离之间的比例推测第一目标对象当前的空间位置。Specifically, for the assumed conditions of the scene, we can assume that the value of a certain dimension does not change during the rigid transformation of the marked points in the scene. For example, in a three-dimensional space scene, the spatial coordinates are x, y, z, and x and y are respectively It is parallel to the x-axis and y-axis in the image coordinates (plane coordinates) of the camera, and the z-axis is perpendicular to the image of the camera. The assumption can be that the coordinates of the marked point on the z-axis are unchanged, or that the marked point is on the x-axis and /or the coordinates on the y-axis are unchanged. Different scenario assumptions require different inference methods. For another example, under another assumption, assuming that the rotation angle between the orientation of the first target object and the orientation of the camera remains constant during use, the distances and The ratio of the distances between the marker points on the standard image estimates the current spatial position of the first target object.

通过以上的计算方法,可以实现单目摄像机对标记点的数量不超过四个时重新构建所述第一目标对象的空间位置,其操作简单,且跟踪结果也较准确,由于采用单目,降低了第一目标对象跟踪的成本。Through the above calculation method, the spatial position of the first target object can be reconstructed when the number of marker points of the monocular camera is not more than four, the operation is simple, and the tracking result is also relatively accurate. reduces the cost of tracking the first target object.

上述使用单个摄像机采集图像来恢复物体三维坐标的方法中,由于获取的图像信息较少,因此需要增加标记点的数目来提供更多的图像信息从而计算物体的三维坐标。根据机器视觉理论,要从单幅图像推算出场景的立体信息,需要至少确定图像中的五个标定点。因此,单目方案增加了标记点数量,也增加了设计的复杂性,但同时,只需要一个摄像机从而降低了图像采集的复杂性,降低了成本。In the above-mentioned method of recovering the three-dimensional coordinates of an object by collecting images with a single camera, since the acquired image information is less, it is necessary to increase the number of marker points to provide more image information to calculate the three-dimensional coordinates of the object. According to machine vision theory, to deduce the stereoscopic information of the scene from a single image, at least five calibration points in the image need to be determined. Therefore, the monocular solution increases the number of marking points and the complexity of the design, but at the same time, only one camera is needed, which reduces the complexity of image acquisition and reduces the cost.

请参见图15,是图11中S13的第二变形例的具体流程示意图。如图14所示,在本实施例中,当所述标记点的数量为五个以上,且采用单目摄像机获取所述标记点的位置信息时,所述S13进一步包括:Please refer to FIG. 15 , which is a schematic flowchart of a second modified example of S13 in FIG. 11 . As shown in FIG. 14, in this embodiment, when the number of the marked points is more than five, and the position information of the marked points is obtained by using a monocular camera, the S13 further includes:

S133依据所述当前图像中的标记点的平面坐标与所述标准图像的所述参考标记点的平面坐标,计算所述当前图像与所述标准图像之间的单应变换关系;S133 Calculate a homography transformation relationship between the current image and the standard image according to the plane coordinates of the marker points in the current image and the plane coordinates of the reference marker points in the standard image;

S134根据所述单应变换关系计算出所述标记点在拍摄所述标准图像时刻的空间位置到当前时刻的空间位置的刚性变换,然后计算所述标记点在当前时刻的空间位置,并根据所述标记点当前时刻的空间位置计算第一目标对象当前的空间位置。S134 Calculate the rigid transformation from the spatial position of the marker point at the moment when the standard image is captured to the spatial position at the current moment according to the homography transformation relationship, and then calculate the spatial position of the marker point at the current moment, and according to the The current spatial position of the first target object is calculated based on the spatial position of the marker point at the current moment.

具体来说,首先采集一幅标准图像,使用精确的深度摄像机或激光扫描仪等器件测量出参考标记点的空间位置,并获取此时的参考标记点的二维图像坐标(即平面坐标)。Specifically, a standard image is collected first, the spatial position of the reference mark point is measured using an accurate depth camera or laser scanner, and the two-dimensional image coordinates (ie plane coordinates) of the reference mark point at this time are obtained.

在使用过程中,摄像机不断捕捉含有第一目标对象的当前图像中的所有标记点的二维图像坐标,并根据此时二维图像坐标与标准图像参考标记点的二维坐标计算出当前状态下的标记点与拍摄标准图像时的标记点之间的刚性变换,在假设标记点之间相对位置不变的情况下,进而计算出出此时标记点相对于标准图像时的空间位置变换,从而计算出当前标记点的空间位置。During use, the camera continuously captures the two-dimensional image coordinates of all marker points in the current image containing the first target object, and calculates the current state based on the two-dimensional image coordinates and the two-dimensional coordinates of the standard image reference marker points at this time. The rigid transformation between the marked point and the marked point when shooting the standard image, assuming that the relative position between the marked points remains unchanged, and then calculate the spatial position transformation of the marked point relative to the standard image at this time, so that Calculate the spatial position of the current marker.

这里,使用五个点以上的点可以计算出当前标记点与拍摄标准图像时标记点的空间位置刚性变换[R|T],优选的,该五个以上的点不在一个平面上,且摄像头的投影矩阵P被事先标定好。计算[R|T]的具体方式如下:Here, the space position rigid transformation [R|T] between the current marker point and the marker point when shooting the standard image can be calculated by using more than five points. Preferably, the five or more points are not on the same plane, and the camera’s The projection matrix P is calibrated in advance. The specific way to calculate [R|T] is as follows:

每一个标记点在标准图像和当前图像的齐次坐标分别为X0,Xi。二者满足极限约束,即X0P-1[R|T]P=Xi。所有标记点组成一个未知参量为[R|T]的方程组。当标记点数量大于5时,可以对[R|T]求解;当标记点数量大于6时,可以对[R|T]求最优解,其方法可以使用奇异值分解SVD,和/或使用迭代的方法计算非线性最优解。当计算出标记点空间位置以后,我们可以根据事先标定好的标记点与第一目标对象(例如人眼)之间的相互位置关系推测出第一目标对象(如人眼)的空间位置。The homogeneous coordinates of each marked point in the standard image and the current image are X0, Xi respectively. Both satisfy the limit constraint, that is, X 0 P -1 [R|T]P=X i . All marked points form a system of equations whose unknown parameter is [R|T]. When the number of marked points is greater than 5, [R|T] can be solved; when the number of marked points is greater than 6, the optimal solution can be found for [R|T], and the method can use singular value decomposition SVD, and/or use The iterative method computes the nonlinear optimal solution. After calculating the spatial position of the marker point, we can infer the spatial position of the first target object (eg human eye) according to the mutual positional relationship between the pre-marked marker point and the first target object (eg human eye).

本实施例只用一个摄像机,使用五个或五个以上的标记点,就可以准确地构建出第一目标对象的空间位置,不仅操作简单,而且成本低廉。In this embodiment, only one camera and five or more marking points can be used to accurately construct the spatial position of the first target object, which is not only easy to operate, but also low in cost.

请参见图16,图16图11中S13的第三变形例的具体流程示意图。如图16所示,本实施例3使用两个或两个以上的摄像机,一个或一个以上的标记点。采用双目摄像机或多目摄像机获取所述标记点的位置信息时,所述S3进一步包括:Please refer to FIG. 16 , which is a schematic flowchart of a third modified example of S13 in FIG. 11 . As shown in FIG. 16, Embodiment 3 uses two or more cameras and one or more marking points. When using a binocular camera or a multi-eye camera to obtain the position information of the marker point, the S3 further includes:

S35采用双目或多目三维重建原理,计算每一个标记点在当前时刻的空间位置;所谓双目或三目重建原理可以采用以下方法,例如采用左、右摄像头匹配的标记点之间的视差,计算每一个标记点在当前时刻的空间位置。或者是采用其它现有的常见方法来实现。S35 adopts binocular or multi-eye three-dimensional reconstruction principle to calculate the spatial position of each marker point at the current moment; the so-called binocular or trinocular reconstruction principle can use the following methods, such as using the parallax between the marker points matched by the left and right cameras , to calculate the spatial position of each marker point at the current moment. Or use other existing common methods to achieve.

S36根据所述标记点当前时刻的空间位置计算第一目标对象当前的空间位置。S36 Calculate the current spatial position of the first target object according to the current spatial position of the marker point.

具体地,首先使用多目摄像机校准的方法对每个摄像机之间的相互位置关系进行标定。然后在使用过程中,对每个摄像机获取到的图像提取标记点坐标,并对每个标记点进行匹配,即获取其在每个摄像机对应的标记点,然后使用匹配的标记点以及摄像机之间相互位置关系计算出标记点的空间位置。Specifically, firstly, a multi-camera calibration method is used to calibrate the mutual positional relationship between each camera. Then in the process of use, extract the coordinates of the marker points from the images acquired by each camera, and match each marker point, that is, obtain its corresponding marker point in each camera, and then use the matched marker points and between cameras The mutual positional relationship calculates the spatial position of the marker points.

在一个具体的例子中,使用多目摄像机(即摄像机数量大于等于2)来拍摄标记点,实现立体重建。已知一个标记点在某一摄像机所拍摄图像上的坐标u和该摄像机参数矩阵M,我们可以计算出一条射线,这个标记点在空间中处于此射线上。In a specific example, multi-cameras (that is, the number of cameras is greater than or equal to 2) are used to shoot marked points to achieve stereoscopic reconstruction. Knowing the coordinate u of a marker point on the image captured by a certain camera and the parameter matrix M of the camera, we can calculate a ray, and the marker point is located on this ray in space.

αjuj=MjXj=1…n(其中n为大于等于2的自然数)α j u j = M j Xj = 1...n (where n is a natural number greater than or equal to 2)

同理,依据上述公式,此标记点在其它摄像机上也可以计算出对应该其它摄像机的射线。理论上讲,这两条射线汇聚在一个点上,即此标记点的空间位置。实际上由于摄像机的数字化误差,摄像机内参和外参标定的误差等等,这些射线并不能汇聚于一点,因此需要使用三角测量(triangululation)的方法近似计算出标记点的空间位置。比如可以使用最小二乘判断准则确定距离所有光线最近的点作为物点。Similarly, according to the above formula, the marking point can also calculate the ray corresponding to the other camera on other cameras. Theoretically, these two rays converge on a point, which is the spatial position of this marked point. In fact, due to the digitization error of the camera, the calibration error of the internal and external parameters of the camera, etc., these rays cannot converge at one point, so it is necessary to use the method of triangulation to approximate the spatial position of the marker point. For example, the least square criterion can be used to determine the point closest to all rays as the object point.

Xx ′′ == argarg xx minmin ΣΣ jj == 11 mm [[ (( mm 11 jj Xx mm 33 jj Xx -- uu 11 jj )) 22 ++ (( mm 22 jj Xx mm 44 jj Xx -- uu 22 jj )) 22 ]]

当计算出标记点空间位置以后,我们可以根据事先标定好的标记点与第一目标对象(如人眼)之间的相互位置关系推测出第一目标对象(人眼)的空间位置。After calculating the spatial position of the marker point, we can infer the spatial position of the first target object (human eye) according to the mutual positional relationship between the pre-marked marker point and the first target object (such as the human eye).

在上述使用多目摄像机实现立体重建的方法中,较佳的方法是使用双目相机计算。其原理与前述多目摄像机重建原理一样,都是根据两摄像机的相互位置关系以及标记点在两摄像机成像的二维坐标计算标记点空间位置。其微小差别是双目摄像机平行摆放,根据简单的标定后对两个摄像机的图像做如前文所述的图像校准,使得两个互相匹配的二维标记点在y(或x)轴上相等,则此时标记点距摄像机的深度可由校准后的二维标记点在x(或y)轴上的差距计算得出。此方法可看成是多目立体重建在双目情况下的特殊方法,其简化了立体重建的步骤且在设备硬件上较容易实现。Among the above methods for realizing stereo reconstruction using multi-cameras, a better method is to use binocular cameras for calculation. The principle is the same as the aforementioned multi-camera reconstruction principle, which is to calculate the spatial position of the marker point based on the mutual positional relationship between the two cameras and the two-dimensional coordinates of the marker point imaged by the two cameras. The slight difference is that the binocular cameras are placed in parallel, and the images of the two cameras are calibrated as described above after simple calibration, so that the two matching two-dimensional marker points are equal on the y (or x) axis , then the depth from the marker point to the camera can be calculated from the distance between the calibrated two-dimensional marker points on the x (or y) axis. This method can be regarded as a special method for multi-eye stereo reconstruction in the case of binoculars, which simplifies the steps of stereo reconstruction and is easier to implement on the device hardware.

实施例6Example 6

请参见图17,图17是图11中的S3的具体流程示意图。如图17所示,基于前述实施方式二及前述实施例,本发明的基于软镜的立体显示方法的步骤S3进一步包括:Please refer to FIG. 17 , which is a schematic flowchart of S3 in FIG. 11 . As shown in FIG. 17 , based on the aforementioned second embodiment and the aforementioned examples, the step S3 of the stereoscopic display method based on the soft mirror of the present invention further includes:

S301排图参数确定步骤,依据获取到的所述第一目标对象的位置信息和所述分光单元的光栅参数及显示单元的显示参数计算出在显示单元上的排图参数;S301 Step of determining the layout parameters, calculating the layout parameters on the display unit according to the acquired position information of the first target object, the grating parameters of the light splitting unit and the display parameters of the display unit;

S302视差图像排列步骤,依据所述排图参数排列所述显示单元上的视差图像;S302 Parallax image arranging step, arranging the parallax images on the display unit according to the layout parameters;

S303视差图像播放步骤,播放所述视差图像。S303 Parallax image playing step, playing the parallax image.

通过上述的步骤,对待播放的立体图像进行重新排列,提高了立体显示的效果。Through the above steps, the stereoscopic images to be played are rearranged, and the stereoscopic display effect is improved.

进一步地,在步骤S301之前还包括:S304立体图像获取步骤,获取实时拍摄到的所述立体图像的信息。在边播放视差图像的同时,边获取实时拍摄到的立体图像信息,提高了图像处理的效率,不仅保证了实时播放,而且同时减少了因立体显示图像占用的数据存储量非常大而要求大内存的要求,降低了成本。Further, before the step S301, the method further includes: S304 a stereoscopic image acquisition step, acquiring the information of the stereoscopic image captured in real time. While playing parallax images, acquire real-time stereoscopic image information, which improves the efficiency of image processing, not only ensures real-time playback, but also reduces the need for large memory due to the large amount of data storage occupied by stereoscopic display images requirements, reducing costs.

以上所述仅为本发明的优选实施方式而已,并不用于限制本发明,对于本领域的技术人员来说,本发明可以有各种更改和变化。凡在本发明的精神和原则之内,所作的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。The above descriptions are only preferred embodiments of the present invention, and are not intended to limit the present invention. For those skilled in the art, the present invention may have various modifications and changes. Any modifications, equivalent replacements, improvements, etc. made within the spirit and principles of the present invention shall be included within the protection scope of the present invention.

Claims (25)

  1. null1. the three-dimensional display system based on soft lens,Including: display unit、Spectrophotometric unit、Tracking equipment and image capturing unit,Described spectrophotometric unit is positioned at the display side of described display unit,Left view and right view it is divided on the image space shown by described display unit,Described tracking equipment is for obtaining the positional information of first object object,Described image capturing unit is used for shooting the second destination object,It is characterized in that,The described three-dimensional display system based on soft lens also includes image player and processes equipment,Respectively with described tracking equipment、Described display unit and described image capturing unit connect,Described image player processes the equipment positional information according to described first object object、The grating parameter of described spectrophotometric unit and the display parameters of described display unit,The stereo-picture that the described image capturing unit received photographs is processed in real time,Send described display unit after process to show in real time.
  2. 2. the three-dimensional display system based on soft lens as claimed in claim 1, it is characterised in that described image capturing unit is located on described soft lens.
  3. 3. the three-dimensional display system based on soft lens as claimed in claim 1, it is characterised in that described image capturing unit includes monocular-camera, shoots and obtain the stereo-picture of described second destination object with a photographic head.
  4. 4. the three-dimensional display system based on soft lens as claimed in claim 1, it is characterised in that described image capturing unit includes binocular camera, shoots and obtain the stereo-picture of described second destination object with two photographic head.
  5. 5. the three-dimensional display system based on soft lens as claimed in claim 1, it is characterised in that described image capturing unit includes multi-lens camera, by the photographic head of more than three stereo-picture shooting and obtaining described second destination object arranged in arrays.
  6. 6. the three-dimensional display system based on soft lens as described in any one of claim 2 to 5, it is characterized in that, described image capturing unit farther includes collecting unit, described collecting unit is for gathering the stereo-picture of described second destination object, and extracts left view information and right view information from described stereo-picture.
  7. 7. the three-dimensional display system based on soft lens as claimed in claim 1, it is characterised in that described tracking equipment includes video camera, the change in location of first object object described in described Camera location.
  8. 8. the three-dimensional display system based on soft lens as claimed in claim 1, it is characterized in that, described tracking equipment includes infrared remote receiver, and described infrared remote receiver receives and comes from the infrared framing signal that the infrared transmitter set by corresponding described first object object sends.
  9. 9. the three-dimensional display system based on soft lens as claimed in claim 1, it is characterised in that described tracking equipment includes:
    Labelling point arranges unit, and the locus of corresponding described first object object arranges labelling point;
    Acquiring unit, obtains the positional information of described labelling point;
    Rebuild unit, according to the positional information of described labelling point, rebuild the locus of described first object object.
  10. 10. the three-dimensional display system based on soft lens as claimed in claim 9, it is characterised in that described acquiring unit farther includes:
    Presetting module, presets a standard picture, is provided with reference marker point, and obtains described reference marker point space coordinates in described standard picture and plane coordinates in described standard picture;
    Acquisition module, obtains the present image comprising described first object object with described labelling point, and described labelling point is at the plane coordinates of described present image;
    Matching module, mates the labelling point in described present image with the described reference marker point of described standard picture.
  11. 11. the three-dimensional display system based on soft lens as claimed in claim 10, it is characterised in that when the quantity of described labelling point is less than four, and when adopting positional information that monocular-camera obtains described labelling point, described in rebuild unit and also include:
    First computing module, for calculating the homograph relation between described present image and described standard picture according to the plane coordinates of the described reference marker point of the plane coordinates of the labelling point in described present image and described standard picture and the assumed conditions of described first object object place scene;
    First reconstructed module, for calculating the described labelling point rigid transformation in the locus to the locus of current time that shoot the described standard picture moment according to described homograph relation, then calculate described labelling point in the locus of current time, and calculate the current locus of described first object object in the locus of current time according to described labelling point.
  12. 12. the three-dimensional display system based on soft lens as claimed in claim 10, it is characterised in that when the quantity of described labelling point is more than five, and when adopting positional information that monocular-camera obtains described labelling point, described in rebuild unit and also include:
    Second computing module, is used for the plane coordinates of the plane coordinates according to the labelling point in described present image and the described reference marker point of described standard picture, calculates the homograph relation between described present image and described standard picture;
    Second reconstructed module, for calculating the described labelling point rigid transformation in the locus to the locus of current time that shoot the described standard picture moment according to described homograph relation, then calculate described labelling point in the locus of current time, and the locus according to described labelling point current time calculates the locus that first object object is current.
  13. 13. the three-dimensional display system based on soft lens as claimed in claim 10, it is characterised in that when the positional information adopting binocular camera or multi-lens camera to obtain described labelling point, described in rebuild unit and also include:
    3rd computing module, is used for adopting binocular or many orders three-dimensional reconstruction principle, calculates each labelling point in the locus of current time;
    Third reconstruction module, calculates, for the locus according to described labelling point current time, the locus that first object object is current.
  14. 14. the three-dimensional display system based on soft lens as described in any one of claim 2 to 13, it is characterized in that, described image player processes equipment and includes image player processing unit and memory element, described image player processing unit is for the display parameters according to the positional information of described first object object, the grating parameter of described spectrophotometric unit and described display unit, the stereo-picture received is carried out real-time row's figure process, sends described display unit after process and show in real time;Described memory element is for storing the image that described image capturing unit transmits, and wherein, described image player processing unit is connected with described memory element.
  15. 15. the three-dimensional display system based on soft lens as claimed in claim 14, it is characterised in that described image player processing unit includes:
    Stereo-picture acquisition module, obtains the information of the described stereo-picture of described image capturing unit shooting;
    Row's graph parameter determines module, according to row's graph parameter that the positional information of described first object object got and the grating parameter of described spectrophotometric unit calculate on described display unit;
    Parallax image arrangement module, for arranging the anaglyph of the described stereo-picture on described display unit according to described row's graph parameter;
    Anaglyph playing module, plays described anaglyph.
  16. 16. the three-dimensional display system based on soft lens as described in any one of claim 1 to 15, it is characterised in that be provided with laminating unit between described spectrophotometric unit and described display unit, by described laminating unit, described spectrophotometric unit is fitted on described display unit.
  17. 17. the three-dimensional display system based on soft lens as claimed in claim 16, it is characterised in that described laminating unit includes first substrate, second substrate and the air layer between described first substrate and described second substrate.
  18. 18. the stereo display method based on soft lens, it is characterised in that said method comprising the steps of:
    S0 shoots the stereo-picture of the second destination object, and sends the information of described stereo-picture in real time;
    S1 obtains the positional information of first object object;
    S2 obtains the grating parameter of the spectrophotometric unit of display device and the display parameters of the display unit of described display device;
    The stereo-picture that the described image capturing unit received photographs is processed in real time by S3 according to described positional information and described grating parameter and described display parameters;
    S4 shows described image to be played.
  19. 19. the stereo display method based on soft lens as claimed in claim 18, it is characterised in that described S0 also includes:
    Image acquisition step, gathers the stereo-picture of described second destination object, and extracts left view information and right view information from described stereo-picture.
  20. 20. the stereo display method based on soft lens as claimed in claim 19, it is characterised in that described S1 also includes:
    The locus of the corresponding described first object object of S11 arranges labelling point;
    S12 obtains the positional information of described labelling point;
    S13, according to the positional information of described labelling point, rebuilds the locus of described first object object.
  21. 21. the stereo display method based on soft lens as claimed in claim 20, it is characterized in that, described S12 farther includes: S121 presets a standard picture, is provided with reference marker point, and obtains described reference marker point space coordinates in described standard picture and plane coordinates in described standard picture;
    S122 obtains the present image comprising described first object object, described labelling point, and described labelling point is at the plane coordinates of described present image;
    Labelling point in described present image is mated by S123 with the described reference marker point of described standard picture.
  22. 22. the stereo display method based on soft lens as claimed in claim 21, it is characterised in that when the quantity of described labelling point is less than four, and when adopting the positional information of the monocular-camera described labelling point of acquisition, described S13 farther includes:
    S131 calculates the homograph relation between described present image and described standard picture according to the plane coordinates of the labelling point in described present image and the plane coordinates of described reference marker point of described standard picture and the assumed conditions of described first object object place scene;
    S132 calculates the described labelling point rigid transformation in the locus to the locus of current time that shoot the described standard picture moment according to described homograph relation, then calculate described labelling point in the locus of current time, and calculate the current locus of described first object object in the locus of current time according to described labelling point.
  23. 23. the stereo display method based on soft lens as claimed in claim 21, it is characterised in that when the quantity of described labelling point is more than five, and when adopting the positional information of the monocular-camera described labelling point of acquisition, described S13 farther includes:
    S133, according to the plane coordinates of the described reference marker point of plane coordinates and the described standard picture of the labelling point in described present image, calculates the homograph relation between described present image and described standard picture;
    S134 calculates the described labelling point rigid transformation in the locus to the locus of current time that shoot the described standard picture moment according to described homograph relation, then calculate described labelling point in the locus of current time, and the locus according to described labelling point current time calculates the locus that first object object is current.
  24. 24. the stereo display method based on soft lens as claimed in claim 21, it is characterised in that when the positional information adopting binocular camera or multi-lens camera to obtain described labelling point, described S13 farther includes:
    S135 adopts binocular or many orders three-dimensional reconstruction principle, calculates each labelling point in the locus of current time;
    S136 calculates, according to the locus of described labelling point current time, the locus that destination object is current.
  25. 25. the stereo display method based on soft lens as described in any one of claim 18 to 24, it is characterised in that described S3 farther includes:
    S304 stereo-picture obtaining step, obtains the information of the described stereo-picture that captured in real-time arrives;
    S301 arranges graph parameter and determines step, calculates row's graph parameter on the display unit according to the display parameters of the positional information of described first object object got and the grating parameter of described spectrophotometric unit and described display unit;
    S302 parallax image arrangement step, according to the anaglyph of the described stereo-picture on described row's graph parameter described display unit of arrangement;
    S303 anaglyph plays step, plays described anaglyph.
CN201410852264.XA 2014-12-29 2014-12-29 Stereoscopic display system based on soft lens and method Pending CN105812776A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410852264.XA CN105812776A (en) 2014-12-29 2014-12-29 Stereoscopic display system based on soft lens and method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410852264.XA CN105812776A (en) 2014-12-29 2014-12-29 Stereoscopic display system based on soft lens and method

Publications (1)

Publication Number Publication Date
CN105812776A true CN105812776A (en) 2016-07-27

Family

ID=56421459

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410852264.XA Pending CN105812776A (en) 2014-12-29 2014-12-29 Stereoscopic display system based on soft lens and method

Country Status (1)

Country Link
CN (1) CN105812776A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021063321A1 (en) * 2019-09-30 2021-04-08 北京芯海视界三维科技有限公司 Method and apparatus for realizing 3d display, and 3d display terminal
CN113947652A (en) * 2020-07-15 2022-01-18 北京芯海视界三维科技有限公司 Method and device for realizing target object positioning and display device

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1514300A (en) * 2002-12-31 2004-07-21 �廪��ѧ Method and system for multi-view X-ray stereoscopic imaging
CN101562756A (en) * 2009-05-07 2009-10-21 昆山龙腾光电有限公司 Stereo display device as well as display method and stereo display jointing wall thereof
CN101984670A (en) * 2010-11-16 2011-03-09 深圳超多维光电子有限公司 Stereoscopic displaying method, tracking stereoscopic display and image processing device
CN102208012A (en) * 2010-03-31 2011-10-05 爱信艾达株式会社 Scene matching reference data generation system and position measurement system
CN103139592A (en) * 2011-11-23 2013-06-05 韩国科学技术研究院 3d display system
CN103875243A (en) * 2011-10-14 2014-06-18 奥林巴斯株式会社 3d endoscope device
WO2014112782A1 (en) * 2013-01-18 2014-07-24 주식회사 고영테크놀러지 Tracking system and tracking method using same
CN204377059U (en) * 2014-12-29 2015-06-03 广东省明医医疗慈善基金会 Stereoscopic display system based on soft mirror

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1514300A (en) * 2002-12-31 2004-07-21 �廪��ѧ Method and system for multi-view X-ray stereoscopic imaging
CN101562756A (en) * 2009-05-07 2009-10-21 昆山龙腾光电有限公司 Stereo display device as well as display method and stereo display jointing wall thereof
CN102208012A (en) * 2010-03-31 2011-10-05 爱信艾达株式会社 Scene matching reference data generation system and position measurement system
CN101984670A (en) * 2010-11-16 2011-03-09 深圳超多维光电子有限公司 Stereoscopic displaying method, tracking stereoscopic display and image processing device
CN103875243A (en) * 2011-10-14 2014-06-18 奥林巴斯株式会社 3d endoscope device
CN103139592A (en) * 2011-11-23 2013-06-05 韩国科学技术研究院 3d display system
WO2014112782A1 (en) * 2013-01-18 2014-07-24 주식회사 고영테크놀러지 Tracking system and tracking method using same
CN204377059U (en) * 2014-12-29 2015-06-03 广东省明医医疗慈善基金会 Stereoscopic display system based on soft mirror

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021063321A1 (en) * 2019-09-30 2021-04-08 北京芯海视界三维科技有限公司 Method and apparatus for realizing 3d display, and 3d display terminal
CN113947652A (en) * 2020-07-15 2022-01-18 北京芯海视界三维科技有限公司 Method and device for realizing target object positioning and display device
WO2022012455A1 (en) * 2020-07-15 2022-01-20 北京芯海视界三维科技有限公司 Method and apparatus for implementing target object positioning, and display component
TWI816153B (en) * 2020-07-15 2023-09-21 大陸商北京芯海視界三維科技有限公司 Method, device and display device for achieving target object positioning

Similar Documents

Publication Publication Date Title
US12178403B2 (en) Light field capture and rendering for head-mounted displays
CN205610834U (en) Stereo display system
CN106773080B (en) Stereoscopic display device and display method
CN105809654B (en) Target object tracking, device and stereoscopic display device and method
CN109270688B (en) Head mounted display device and control method of head mounted display device
US20160295194A1 (en) Stereoscopic vision system generatng stereoscopic images with a monoscopic endoscope and an external adapter lens and method using the same to generate stereoscopic images
CN105996961B (en) 3D three-dimensional imagings capsule endoscope system and method based on structure light
CN114895471A (en) Head mounted display for virtual reality and mixed reality with inside-outside position tracking, user body tracking, and environment tracking
CN105942959A (en) Capsule endoscope system and three-dimensional imaging method thereof
CN102098524A (en) Tracking type stereo display device and method
WO2006086223A2 (en) Augmented reality device and method
CN204578692U (en) Three-dimensional display system
CN105342561B (en) The wearable molecular image navigation system of Wireless sound control
CN204377059U (en) Stereoscopic display system based on soft mirror
CN105812774A (en) Three-dimensional display system based on endoscope and method thereof
TW202215370A (en) Systems and methods for superimposing virtual image on real-time image
CN105812772A (en) Stereo display system and method for medical images
CN204377058U (en) Based on the three-dimensional display system of hard mirror
CN109710073A (en) A kind of VR medical operation teaching helmet
WO2019075922A1 (en) Single-lens horizontal-scroll stereoscopic endoscope system
CN204539353U (en) Medical image three-dimensional display system
CN105812776A (en) Stereoscopic display system based on soft lens and method
CN204377057U (en) Based on the three-dimensional display system of intubate mirror
US12270980B2 (en) Binocular device for visualizing optical radiation and method of visualizing optical radiation
JP2006305332A (en) Image processor and endoscope using the same

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20160727

RJ01 Rejection of invention patent application after publication