[go: up one dir, main page]

CN104980732B - The system and method for measuring the potential eye fatigue of stereoscopic motion picture - Google Patents

The system and method for measuring the potential eye fatigue of stereoscopic motion picture Download PDF

Info

Publication number
CN104980732B
CN104980732B CN201510315583.1A CN201510315583A CN104980732B CN 104980732 B CN104980732 B CN 104980732B CN 201510315583 A CN201510315583 A CN 201510315583A CN 104980732 B CN104980732 B CN 104980732B
Authority
CN
China
Prior art keywords
parallax
image
eye fatigue
images
frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201510315583.1A
Other languages
Chinese (zh)
Other versions
CN104980732A (en
Inventor
张冬青
安娜.B.班尼特兹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Thomson Licensing SAS
InterDigital Madison Patent Holdings SAS
Original Assignee
THOMSON LICENSING CORP
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by THOMSON LICENSING CORP filed Critical THOMSON LICENSING CORP
Priority to CN201510315583.1A priority Critical patent/CN104980732B/en
Priority claimed from CN200880129158.1A external-priority patent/CN102027752B/en
Publication of CN104980732A publication Critical patent/CN104980732A/en
Application granted granted Critical
Publication of CN104980732B publication Critical patent/CN104980732B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2213/00Details of stereoscopic systems
    • H04N2213/002Eyestrain reduction by processing stereoscopic signals or controlling stereoscopic devices

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Eye Examination Apparatus (AREA)

Abstract

提供一种用于测量在观看3D呈现(例如,立体运动画面)时由观众感受到的潜在眼睛疲劳的系统和方法。本公开的眼睛疲劳测量系统和方法基于3D呈现的立体图像的视差(或深度)和视差转换的测量。本公开的系统和方法规定:从第一片段获取第一图像和第二图像(202),估计第一图像中的至少一个点与第二图像中的至少一个对应点的视差(204),估计第一和第二图像的序列的视差转换(206),以及基于视差以及第一和第二图像的序列的视差转换确定在观看3D呈现时感受到的潜在眼睛疲劳(208)。

A system and method for measuring potential eyestrain experienced by a viewer while viewing a 3D presentation (eg, stereoscopic motion picture) is provided. The eyestrain measurement system and method of the present disclosure are based on the measurement of disparity (or depth) and disparity transformation of 3D rendered stereoscopic images. The disclosed systems and methods provide for acquiring a first image and a second image (202) from a first segment, estimating (204) the disparity of at least one point in the first image to at least one corresponding point in the second image, estimating The disparity transformation of the sequence of first and second images (206), and the potential eyestrain experienced while viewing the 3D presentation is determined based on the disparity and the disparity transformation of the sequence of first and second images (208).

Description

测量立体运动画面的潜在眼睛疲劳的系统和方法System and method for measuring potential eyestrain of stereoscopic motion picture

相关申请的交叉引用Cross References to Related Applications

本发明是申请日为2008年5月12日、申请号为200880129158.1的、名称为“用于测量立体运动画面的潜在眼睛疲劳的系统和方法”的发明专利申请的分案申请。The present invention is a divisional application of an invention patent application entitled "System and Method for Measuring Potential Eye Fatigue of Stereoscopic Motion Pictures" with the filing date of May 12, 2008 and the application number of 200880129158.1.

技术领域technical field

本公开一般地涉及计算机图形处理和显示系统,并且更具体地涉及一种用于测量由观看三维(3D)呈现(例如,立体运动画面)的观众感受到的潜在眼睛疲劳的系统和方法。The present disclosure relates generally to computer graphics processing and display systems, and more particularly to a system and method for measuring potential eyestrain experienced by a viewer viewing a three-dimensional (3D) presentation (eg, stereoscopic motion picture).

背景技术Background technique

立体成像是在视觉上组合一个场景的从稍微不同的视点取得的至少两个图像以便产生三维深度的错觉的处理。该技术依赖于如下事实:人眼分隔一定距离,并且因此没有观看完全相同的场景。通过为每只眼睛提供来自不同观察点(perspective)的图像,欺骗观看者的眼睛感觉到深度。典型地,在提供两个有区别的观察点的情况下,分别将组成图像称为“左”和“右”图像,也分别将其已知为基准图像和补充图像。然而,本领域技术人员将认识到,可以组合多于两个视点来形成立体图像。Stereoscopic imaging is the process of visually combining at least two images of a scene taken from slightly different viewpoints in order to create the illusion of three-dimensional depth. This technique relies on the fact that human eyes are separated by a distance, and therefore do not view the exact same scene. By providing each eye with an image from a different perspective, the viewer's eyes are tricked into perceiving depth. Typically, where two distinct points of view are provided, the constituent images are referred to as the "left" and "right" images, respectively, also known as the reference image and the supplementary image, respectively. However, those skilled in the art will recognize that more than two viewpoints may be combined to form a stereoscopic image.

在3D后期制作、VFX工作流程和3D呈现应用中,一项重要的处理是从由左眼视图图像和右眼视图图像组成的立体图像中推断深度图,以便创建立体运动画面。例如,最近商业化的自动立体3D显示器要求图像加深度图(image-plus-depth-map)输入格式,使得该显示器可以生成不同的3D视图来支持多个视角。In 3D post-production, VFX workflow, and 3D rendering applications, an important process is to infer a depth map from a stereoscopic image consisting of a left-eye view image and a right-eye view image, in order to create stereoscopic motion pictures. For example, recently commercialized autostereoscopic 3D displays require an image-plus-depth-map input format so that the display can generate different 3D views to support multiple viewing angles.

从立体图像对推断深度图的处理在计算机视觉研究的领域中被称为立体匹配,因为使用像素或块匹配以发现左眼和右眼视图图像中的对应点。从与场景中的相同点对应的各图像中的两个像素之间的相对距离推断深度值。The process of inferring a depth map from a stereo pair of images is known in the field of computer vision research as stereo matching because pixel or block matching is used to find corresponding points in the left and right eye view images. A depth value is inferred from the relative distance between two pixels in each image corresponding to the same point in the scene.

在许多计算机视觉应用(诸如,例如用于计算机辅助绘图(CAD)的快速对象建模和成型、用于人机交互(HCI)的对象分割和检测、视频压缩和视觉监控(visualsurveillance))中广泛使用数字图像的立体匹配,以便提供三维(3D)深度信息。立体匹配从场景中位于不同位置和朝向的两个或多个相机获得场景的图像。在大约相同的时间从每个相机获得这些数字图像,并且对应于空间中的3D点匹配每个图像中的点。通常,通过搜索一部分图像和使用约束(诸如外极线约束)来匹配来自不同图像的点,以便使一个图像中的点与另一图像中的点相关联。然后可以采用匹配后的图像和深度图来创建立体3D运动画面。Widespread in many computer vision applications such as, for example, fast object modeling and prototyping for computer-aided drafting (CAD), object segmentation and detection for human-computer interaction (HCI), video compression, and visual surveillance (visual surveillance) Stereo matching of digital images is used in order to provide three-dimensional (3D) depth information. Stereo matching obtains images of a scene from two or more cameras at different positions and orientations in the scene. These digital images are acquired from each camera at about the same time, and points in each image are matched corresponding to 3D points in space. Typically, points in one image are associated with points in another image by searching a portion of the image and matching points from different images using constraints, such as epipolar constraints. The matched image and depth map can then be used to create stereoscopic 3D motion pictures.

当前立体3D运动画面的主要问题之一是观众可能在观看运动画面一定时间之后感受到眼睛疲劳。因此,当导演制作3D影片时,他们不得不考虑如何以可以使由观众感受到的眼睛疲劳最小化的方式拍摄场景或剪辑影片,。这是制作3D运动画面比制作传统的2D运动画面更加复杂和耗时的部分原因。One of the main problems with current stereoscopic 3D motion pictures is that the viewer may experience eyestrain after watching the motion picture for a certain period of time. Therefore, when directors make 3D movies, they have to consider how to shoot scenes or edit movies in such a way that the eye strain experienced by the audience is minimized. This is part of the reason why making 3D motion graphics is more complex and time-consuming than making traditional 2D motion graphics.

制作3D运动画面的挑战在于导演和剪辑师非常难以在视觉上估计由观众感受到的潜在眼睛疲劳。存在促成该困难的若干因素。首先,因为眼睛疲劳是随着观看运动画面的过程的累计效果,所以导演或剪辑师不得不观看3D运动画面足够长以感受到眼睛疲劳。因为少量片段通常不导致眼睛疲劳。第二,眼睛疲劳还可能由两个片段之间的突然深度变化导致。当剪辑师在剪辑期间连接片段时,他们难以测量由突然深度变化导致的潜在眼睛疲劳。他们将需要使用耗时的反复试验(trial-and-error)过程来连接不同的片段,并且“感受”由深度转换导致的潜在眼睛疲劳。The challenge of producing 3D motion pictures is that it is very difficult for directors and editors to visually assess the potential eye strain experienced by the audience. There are several factors that contribute to this difficulty. First, because eye strain is a cumulative effect over the course of viewing motion, the director or editor has to watch 3D motion long enough to experience eye strain. Because a small number of fragments usually do not cause eyestrain. Second, eye strain can also be caused by sudden depth changes between two clips. When editors connect clips during editing, they struggle to gauge potential eye strain caused by sudden depth changes. They will need to use a time-consuming trial-and-error process to connect the different pieces and "feel" the potential eye strain caused by the depth transition.

因此,存在对于可以在观看诸如立体运动画面的3D呈现的同时测量感受到的潜在眼睛疲劳的技术的需要。此外,存在对于可以在剪辑3D运动画面的过程期间测量潜在眼睛疲劳的自动系统和方法的需要。Therefore, there is a need for a technique that can measure perceived potential eye strain while viewing a 3D presentation such as stereoscopic motion picture. Furthermore, there is a need for an automated system and method that can measure potential eyestrain during the process of editing 3D motion picture.

发明内容Contents of the invention

提供一种用于测量在观看3D呈现(例如,立体运动画面)时由观众感受到的潜在眼睛疲劳的系统和方法。本公开的系统和方法考虑:观看3D呈现时眼睛的会聚点和焦点之间的距离与所聚焦的对象的深度紧密相关,所聚焦的对象的深度也与对象的像素的视差相关。本公开的眼睛疲劳测量系统和方法基于3D呈现的立体图像的视差(或深度)和视差转换的测量。本公开的技术对于导演和剪辑师有效地制作舒适的3D电影是有用的。A system and method for measuring potential eyestrain experienced by a viewer while viewing a 3D presentation (eg, stereoscopic motion picture) is provided. The systems and methods of the present disclosure take into account that the distance between the eye's point of convergence and the focal point when viewing a 3D presentation is closely related to the depth of the focused object, which is also related to the disparity of the object's pixels. The eyestrain measurement system and method of the present disclosure are based on the measurement of disparity (or depth) and disparity transformation of 3D rendered stereoscopic images. The disclosed technology is useful for directors and editors to efficiently create comfortable 3D movies.

根据本公开的一个方面,提供一种测量在观看三维(3D)呈现时感受到的潜在眼睛疲劳的方法,所述方法包括从第一片段获取第一图像和第二图像;估计第一图像中的至少一个点与第二图像中的至少一个对应点的视差;估计第一和第二图像的序列的视差转换;以及基于视差以及第一和第二图像的序列的视差转换确定在观看3D呈现时感受到的潜在眼睛疲劳。According to one aspect of the present disclosure, there is provided a method of measuring potential eyestrain experienced while viewing a three-dimensional (3D) presentation, the method comprising acquiring a first image and a second image from a first segment; estimating disparity of at least one point of at least one corresponding point in the second image; estimating a disparity transformation of the sequence of first and second images; Potential eye fatigue experienced when

在另一方面中,估计视差转换步骤包括:估计之前片段的最后帧的视差;估计第一片段的第一帧的视差;以及确定之前片段的最后帧的视差和第一片段的第一帧的视差之间的差。In another aspect, the estimating disparity conversion step includes: estimating the disparity of the last frame of the previous segment; estimating the disparity of the first frame of the first segment; and determining the disparity of the last frame of the previous segment and the first frame of the first segment The difference between parallax.

在一个方面中,估计视差转换步骤包括:估计第一片段的多个帧的每个帧内的视差;以及确定第一片段的每个帧的视差之间的差。In one aspect, the estimating disparity transformation step includes: estimating disparity within each of the plurality of frames of the first segment; and determining a difference between the disparities of each frame of the first segment.

在另一方面中,确定潜在眼睛疲劳步骤还包括确定用于第一和第二图像的序列的每个帧的即时眼睛疲劳函数。In another aspect, the step of determining potential eyestrain further includes determining an instantaneous eyestrain function for each frame of the sequence of first and second images.

在另一方面中,确定潜在眼睛疲劳步骤还包括将衰减因子应用到用于第一和第二图像的序列的每个帧的即时眼睛疲劳函数。In another aspect, the step of determining potential eyestrain further includes applying an attenuation factor to the instantaneous eyestrain function for each frame of the sequence of first and second images.

在另一方面中,确定潜在眼睛疲劳步骤还包括在预定的时间段上累积衰减后的用于第一和第二图像的序列的每个帧的即时眼睛疲劳函数。In another aspect, the step of determining potential eyestrain further includes accumulating the attenuated instantaneous eyestrain function for each frame of the sequence of first and second images over a predetermined period of time.

在另一方面中,确定潜在眼睛疲劳步骤还包括使在第一和第二图像的序列上累积的眼睛疲劳函数饱和。In another aspect, the step of determining potential eyestrain further includes saturating the eyestrain function accumulated over the sequence of first and second images.

在另一方面中,本方法还包括:确定潜在眼睛疲劳是否可接受;以及如果潜在眼睛疲劳不可接受,则校正第一和第二图像的视差。In another aspect, the method further includes: determining whether the potential eyestrain is acceptable; and correcting the parallax of the first and second images if the potential eyestrain is not acceptable.

根据本公开的另一方面,提供一种用于测量在观看三维(3D)呈现时感受到的潜在眼睛疲劳的系统。所述系统包括:用于从第一片段获取第一图像和第二图像的部件;视差估计器,用于估计第一图像中的至少一个点与第二图像中的至少一个对应点的视差;视差转换估计器,用于估计第一和第二图像的序列的视差转换;以及眼睛疲劳估计器,用于基于视差以及第一和第二图像的序列的视差转换确定在观看3D呈现时感受到的潜在眼睛疲劳。According to another aspect of the present disclosure, a system for measuring potential eyestrain experienced while viewing a three-dimensional (3D) presentation is provided. The system includes: means for acquiring a first image and a second image from a first segment; a disparity estimator for estimating a disparity of at least one point in the first image to at least one corresponding point in the second image; a disparity transformation estimator for estimating the disparity transformation of the sequence of first and second images; and an eye strain estimator for determining the perceived sensation while viewing the 3D presentation based on the disparity and the disparity transformation of the sequence of first and second images potential eye fatigue.

根据本公开的另一方面,提供一种可由机器读取的程序存储设备,所述程序存储设备确实地体现可由机器执行的、用于执行用以测量在观看三维(3D)呈现时的潜在眼睛疲劳的方法步骤的指令的程序,所述方法包括:从第一片段获取第一图像和第二图像;估计第一图像中的至少一个点与第二图像中的至少一个对应点的视差;估计第一和第二图像的序列的视差转换;以及基于视差以及第一和第二图像的序列的视差转换确定在观看3D呈现时感受到的潜在眼睛疲劳。According to another aspect of the present disclosure, there is provided a program storage device readable by a machine, the program storage device tangibly embodying a program executable by a machine for performing a method for measuring a potential eye while viewing a three-dimensional (3D) presentation. A program of instructions for fatigued method steps, the method comprising: acquiring a first image and a second image from a first segment; estimating a disparity between at least one point in the first image and at least one corresponding point in the second image; estimating Parallax transformation of the sequence of first and second images; and determining potential eyestrain experienced while viewing the 3D presentation based on the disparity and the parallax transformation of the sequence of first and second images.

附图说明Description of drawings

将描述本公开的这些和其它方面、特征和优点,或者本公开的这些和其它方面、特征和优点将从优选实施例的以下详细描述变得明显,结合附图阅读优选实施例的以下详细描述。These and other aspects, features and advantages of the present disclosure will be described or will be apparent from the following detailed description of the preferred embodiments, read in conjunction with the accompanying drawings .

在附图中,其中贯穿各附图相同的参考标号表示类似的元件:In the drawings, where like reference numerals designate like elements throughout the several drawings:

图1图示在观看三维(3D)呈现的同时由观看者或者观众的成员体验的会聚点和焦点;FIG. 1 illustrates points of convergence and focal points experienced by a viewer or member of an audience while viewing a three-dimensional (3D) presentation;

图2图示当会聚点在用于显示3D呈现的屏幕的前面时,会聚距离、所感知的深度、视差和会聚角度的关系;Figure 2 illustrates the relationship of convergence distance, perceived depth, parallax and convergence angle when the convergence point is in front of the screen for displaying the 3D presentation;

图3图示当会聚点在屏幕后面时,会聚距离、所感知的深度、视差和会聚角度的关系;Figure 3 illustrates the relationship between convergence distance, perceived depth, parallax and convergence angle when the convergence point is behind the screen;

图4是根据本公开的一个方面的、用于测量在观看三维(3D)呈现时感受到的眼睛疲劳的系统的示例性图示;4 is an exemplary illustration of a system for measuring eye fatigue experienced while viewing a three-dimensional (3D) presentation, according to one aspect of the present disclosure;

图5是根据本公开的一个方面的、用于测量在观看三维(3D)呈现时感受到的眼睛疲劳的示例性方法的流程图;5 is a flowchart of an exemplary method for measuring eye fatigue experienced while viewing a three-dimensional (3D) presentation, according to one aspect of the present disclosure;

图6图示用于对眼睛疲劳感觉的饱和建模的S形函数;Figure 6 illustrates a sigmoid function for modeling saturation of eye fatigue perception;

图7图示根据本公开的一个方面的随着时间的眼睛疲劳测量方程式;以及FIG. 7 illustrates an eye strain measurement equation over time according to one aspect of the present disclosure; and

图8图示根据本公开的一个方面的即时眼睛疲劳函数和总体眼睛疲劳测量。FIG. 8 illustrates an instant eyestrain function and an overall eyestrain measure according to one aspect of the present disclosure.

应该理解的是:(各)附图是用于图示本公开的构思的目的,并且不必是用于图示本公开的唯一可能配置。It should be understood that the drawing(s) are for purposes of illustrating the concepts of the disclosure and are not necessarily the only possible configurations for illustrating the disclosure.

具体实施方式detailed description

应该理解的是:附图中示出的各元件可以以硬件、软件和其组合的各种形式实现。优选地,这些元件以硬件和在可以包括处理器、存储器和输入/输出接口的一个或多个适当地编程的通用设备上的软件的组合实现。It should be understood that each element shown in the drawings may be realized in various forms of hardware, software and combinations thereof. Preferably, these elements are implemented in a combination of hardware and software on one or more suitably programmed general purpose devices which may include a processor, memory and input/output interfaces.

本描述说明本公开的原理。因此,将认识到:本领域技术人员将能够设计出实施本公开的原理并被包括在本原理的精神和范围内的各种布置,尽管在这里没有明确地描述或示出所述布置。This description illustrates the principles of the disclosure. It will thus be appreciated that those skilled in the art will be able to devise various arrangements which, although not explicitly described or shown herein, embody the principles of the disclosure and are included within the spirit and scope of the principles.

在此叙述的所有示例和条件性语言意欲用于教学的目的以便帮助读者理解本公开的原理以及由本发明人贡献以促进现有技术的构思,并且应该被解释为不限制于这种具体叙述的示例和条件。All examples and conditional language recited herein are intended for pedagogical purposes in order to assist the reader in understanding the principles of the disclosure and concepts contributed by the inventors to advance the state of the art, and should be construed as not limiting to such specific recitations Examples and conditions.

另外,在这里叙述本公开的原理、方面和实施例及其特定示例的所有陈述意欲包括其结构和功能等效物。另外,意图是:这样的等效物包括当前已知的等效物以及将来开发的等效物二者,即所开发的执行相同功能的任何元件,而不论其结构如何。Additionally, all statements herein reciting principles, aspects, and embodiments of the disclosure, as well as specific examples thereof, are intended to encompass both structural and functional equivalents thereof. Additionally, it is intended that such equivalents include both currently known equivalents as well as equivalents developed in the future, ie, any elements developed that perform the same function, regardless of structure.

因此,例如,本领域技术人员将认识到:在此呈现的框图表示实施本原理的说明性电路的概念性视图。类似地,将认识到:任何流程图示(flow chart)、流程图(flowdiagram)、状态转换图、伪代码等表示实质上可以表示在计算机可读介质中并因此由计算机或处理器执行的各种处理,而不管是否明确地示出这样的计算机或处理器。Thus, for example, it will be appreciated by those skilled in the art that the block diagrams presented herein represent conceptual views of illustrative circuitry embodying the present principles. Similarly, it will be appreciated that any flow chart, flowdiagram, state transition diagram, pseudocode, etc. representations may substantially be represented in a computer-readable medium and thus executed by a computer or processor. processing regardless of whether such a computer or processor is explicitly shown.

可以通过使用专用硬件以及与适当的软件相关联的能够执行软件的硬件来提供图中示出的各种元件的功能。当利用处理器来提供所述功能时,可以利用单个专用处理器、利用单个共享处理器、或者利用其中一些可被共享的多个独立处理器来提供所述功能。另外,术语“处理器”或“控制器”的明确使用不应该被解释为排他性地指代能够执行软件的硬件,而是可以隐含地无限制地包括数字信号处理器(“DSP”)硬件、用于存储软件的只读存储器(“ROM”)、随机存取存储器(“RAM”)、和非易失性存储器。The functions of the various elements shown in the figures may be provided through the use of dedicated hardware as well as hardware capable of executing the software in association with appropriate software. When a processor is utilized to provide the described functionality, it may be provided with a single dedicated processor, with a single shared processor, or with multiple independent processors, some of which may be shared. Additionally, explicit use of the terms "processor" or "controller" should not be construed as referring exclusively to hardware capable of executing software, but may implicitly include, without limitation, digital signal processor ("DSP") hardware , read-only memory ("ROM"), random-access memory ("RAM"), and non-volatile memory for storing software.

还可以包括其它传统的和/或定制的硬件。类似地,图中示出的任何开关只是概念性的。它们的功能可以通过程序逻辑的运行、通过专用逻辑、通过程序控制和专用逻辑的交互、或者甚至手动地来执行,如从上下文更具体地理解的那样,可以由实施者选择具体技术。Other conventional and/or custom hardware may also be included. Similarly, any switches shown in the figures are conceptual only. Their function may be carried out through the operation of program logic, through dedicated logic, through the interaction of program control and dedicated logic, or even manually, the particular technique being selectable by the implementer as more specifically understood from the context.

在其权利要求中,被表示为用于执行指定功能的部件的任何元件意欲包含执行那个功能的任何方式,例如包括:a)执行那个功能的电路元件的组合或者b)与适当电路相组合的任何形式的软件,所述软件因此包括固件或微代码等,所述适当电路用于执行该软件以执行所述功能。由这种权利要求限定的本公开在于如下事实,即,以权利要求所要求的方式将由各种所叙述的部件提供的功能组合和集合到一起。因此认为可以提供那些功能的任何部件与在此示出的那些部件等效。In its claims, any element expressed as a means for performing a specified function is intended to encompass any way of performing that function, including, for example: a) a combination of circuit elements performing that function or b) a combination with an appropriate circuit Any form of software, thus including firmware or microcode or the like, that is executed by appropriate circuitry to perform the described functions. The present disclosure defined by such claims resides in the fact that the functionality provided by the various recited elements is combined and brought together in the manner required by the claims. Any means that can provide those functions are therefore considered equivalent to those shown herein.

通常,立体运动画面由左眼和右眼画面序列组成。对于电影院应用,观众通常不得不佩戴3D眼镜(例如,偏振或快门眼镜)观看3D呈现。对于不用眼镜的显示器,尽管3D显示的机制不同,但是3D感知的原理是相同的。也就是说,3D系统将使左眼观看左眼图像,而右眼仅观看右眼图像。人脑能够将这两个图像组合到一起,以便正确感知3D场景。然而,该系统可能潜在地导致眼睛肌肉的疲劳和眼睛疲劳,因为以下两个原因:Typically, a stereoscopic motion picture consists of a sequence of left-eye and right-eye pictures. For cinema applications, viewers typically have to wear 3D glasses (eg, polarized or shutter glasses) to view the 3D presentation. For displays without glasses, although the mechanism of 3D display is different, the principle of 3D perception is the same. That is, the 3D system will make the left eye see the left eye image, and the right eye only see the right eye image. The human brain is able to combine these two images in order to correctly perceive a 3D scene. However, this system can potentially cause eye muscle fatigue and eye strain for two reasons:

1.眼睛的会聚点和焦点是不同的。当我们观看真实3D场景时,我们眼睛的焦点与会聚点大致相同。然而,当我们观看3D影片时,我们眼睛的焦点10不得不总是在屏幕12上,而我们眼睛的会聚点14不得不在屏幕12的前面或后面,以便感知正确的3D场景,如图1所示。该差异是促进眼睛肌肉的疲劳因此导致眼睛疲劳的主要因素。1. The convergence point and focus of the eyes are different. When we watch a real 3D scene, the focal point of our eyes is roughly the same as the point of convergence. However, when we watch a 3D movie, the focal point 10 of our eyes has to be always on the screen 12, and the point of convergence 14 of our eyes has to be in front or behind the screen 12 in order to perceive the correct 3D scene, as shown in Figure 1 Show. This difference is a major factor that promotes fatigue of the eye muscles and thus causes eye fatigue.

2.3D运动画面中对象的深度改变。当对象的深度改变时,我们的眼睛不得不调整会聚点以感知正确的3D效果,同时保持焦点在屏幕上。如果深度改变是频繁和突然的,则我们的眼睛不得不频繁改变会聚点,导致眼睛肌肉的疲惫。2. Depth changes of objects in 3D motion pictures. When the depth of an object changes, our eyes have to adjust the point of convergence to perceive the correct 3D effect while maintaining focus on the screen. If the depth changes are frequent and sudden, our eyes have to change the point of convergence frequently, causing fatigue of the eye muscles.

所以,简而言之,眼睛疲劳主要由两个原因导致:1)眼睛的会聚点和焦点之间的距离(即,如图1所示的会聚距离16);以及2)会聚点14的改变。因此,测量眼睛疲劳需要考虑上面两个因素。So, in a nutshell, eyestrain is mainly caused by two causes: 1) the distance between the point of convergence and the focal point of the eyes (i.e., the convergence distance 16 as shown in FIG. 1 ); and 2) a change in the point of convergence 14 . Therefore, the measurement of eye fatigue needs to consider the above two factors.

本公开提供了一种用于测量在观看3D呈现(例如,立体运动画面)时由观众感受到的潜在眼睛疲劳的系统和方法。本公开的眼睛疲劳测量系统和方法基于视差(或深度)和视差转换的测量。本解决方案对于导演和剪辑师有效地制作好的和舒适的3D电影是有用的。The present disclosure provides a system and method for measuring potential eyestrain experienced by a viewer while viewing a 3D presentation (eg, stereoscopic motion picture). The eyestrain measurement system and method of the present disclosure are based on the measurement of disparity (or depth) and disparity transformation. This solution is useful for directors and editors to efficiently produce good and comfortable 3D movies.

本公开的系统和方法考虑:观看者的眼睛的会聚点和焦点之间的距离与3D呈现中所聚焦的对象的深度紧密相关,3D呈现中所聚焦的对象的深度也与对象的像素的视差有关。图2示出了观看3D呈现中涉及的变量的关系。可以看到:给定会聚距离(Cd)16、所感知的深度(Dp)18、观众距离(Ad)20、会聚角度(Ca)22、眼睛距离(Ed)24和视差(Ds)26,存在以下关系。The systems and methods of the present disclosure consider that the distance between the point of convergence and the focal point of the viewer's eyes is closely related to the depth of the object in focus in the 3D rendering, which is also the parallax of the object's pixels in the 3D rendering related. Figure 2 shows the relationship of the variables involved in viewing a 3D presentation. It can be seen that given the convergence distance (Cd) 16, perceived depth (Dp) 18, audience distance (Ad) 20, convergence angle (Ca) 22, eye distance (Ed) 24 and parallax (Ds) 26, there exists the following relationship.

1.Cd、Dp和Ad的关系:Ad=Cd+Dp1. The relationship between Cd, Dp and Ad: Ad=Cd+Dp

2.Cd、Ds、Ad和Ed的关系:Cd(1/Ds+1/Ed)=Ad/Ed2. The relationship between Cd, Ds, Ad and Ed: Cd(1/Ds+1/Ed)=Ad/Ed

3.Ca、Ed、Ad和Cd的关系:Cd=2atan(Ed/(2(Ad-Cd)))3. The relationship between Ca, Ed, Ad and Cd: Cd=2atan(Ed/(2(Ad-Cd)))

当会聚点如图3所示在屏幕后面时,这些关系保持正确,只要允许负视差和负会聚距离即可。对于特定观众,眼睛距离(Ed)24和观众距离(Ad)20在呈现期间是恒定的,而会聚距离(Cd)16、所感知的深度(Dp)18、会聚角度(Ca)22和视差(Ds)26在运动画面期间变化。基于这些关系,会聚点的计算可以简化为深度或视差的估计。这导致更简单的估计算法。因为存在许多视差估计算法,所以传统的视差测量可以用于眼睛疲劳估计。When the point of convergence is behind the screen as shown in Figure 3, these relationships hold true as long as negative parallax and negative convergence distances are allowed. For a particular audience, eye distance (Ed) 24 and audience distance (Ad) 20 are constant during presentation, while convergence distance (Cd) 16, perceived depth (Dp) 18, convergence angle (Ca) 22 and parallax ( Ds) 26 changes during a moving picture. Based on these relationships, the calculation of the convergence point can be reduced to the estimation of depth or disparity. This results in a simpler estimation algorithm. Because many disparity estimation algorithms exist, traditional disparity measurements can be used for eye strain estimation.

现在参照图4,示出了根据本公开的实施例的示例性系统组件。可以提供扫描设备103,用于将影片印件(film print)104(例如,相机原始底片(film negative))扫描为数字格式(例如Cineon格式或SMPTE DPX文件)。扫描设备103可以包括例如将从影片生成视频输出的电视电影或任何设备(诸如,例如具有视频输出的Arri LocProTM)。可替代地,可以直接使用来自后期制作过程或数字电影的文件106(例如,已经处于计算机可读形式的文件)。计算机可读文件的潜在来源是AVIDTM剪辑器、DPX文件、D5磁带。Referring now to FIG. 4 , exemplary system components are shown in accordance with an embodiment of the present disclosure. A scanning device 103 may be provided for scanning a film print 104 (eg camera negative) into a digital format (eg Cineon format or SMPTE DPX file). The scanning device 103 may comprise, for example, a telecine or any device that will generate a video output from film (such as, for example, an Arri LocPro with a video output). Alternatively, files 106 (eg, files already in computer-readable form) from post-production processes or digital cinema can be used directly. Potential sources of computer readable files are AVID Clipper, DPX files, D5 tapes.

所扫描的影片印件被输入到后期处理设备102(例如,计算机)。在具有以下硬件的各种已知计算机平台的任何计算机平台上实现该计算机,所述硬件诸如一个或多个中央处理单元(CPU)、诸如随机存取存储器(RAM)和/或只读存储器(ROM)的存储器110、以及诸如键盘和指针控制设备(例如,鼠标或操纵杆)和显示设备的(多个)输入/输出(I/O)用户接口112。计算机平台还包括操作系统和微指令代码。在此描述的各种处理和功能可以是经由操作系统执行的微指令代码的一部分或软件应用程序的一部分(或其组合)。在一个实施例中,软件应用程序确实地体现在程序存储设备上,其可以上载到任何合适的机器(诸如后期处理设备102)并且被执行。此外,各种其它外围设备可以通过各种接口和总线结构(诸如并行端口、串行端口或通用串行总线(USB))连接到计算机平台。其它外围设备可以包括额外的存储设备124和印片机(printer)128。印片机128可以用于冲印影片的修改版本126(例如,影片的立体版本),其中作为下面描述的技术的结果,可能已经使用3D建模的对象变更或替代场景或多个场景。The scanned film print is input to a post-processing device 102 (eg, a computer). The computer is implemented on any of the various known computer platforms having hardware such as one or more central processing units (CPUs), such as random access memory (RAM) and/or read-only memory ( ROM), and input/output (I/O) user interface(s) 112 such as a keyboard and pointer control device (eg, mouse or joystick) and display device(s). A computer platform also includes an operating system and microinstruction code. The various processes and functions described herein may be part of the microinstruction code or part of the software application (or a combination thereof) executed via the operating system. In one embodiment, the software application is tangibly embodied on a program storage device, which can be uploaded to any suitable machine (such as post-processing device 102) and executed. In addition, various other peripheral devices can be connected to the computer platform through various interfaces and bus structures, such as parallel ports, serial ports, or Universal Serial Bus (USB). Other peripherals may include additional storage devices 124 and a printer 128 . The printer 128 may be used to print a modified version 126 of the film (eg, a stereoscopic version of the film) in which the scene or scenes may have been altered or replaced using 3D modeled objects as a result of the techniques described below.

可替代地,已经处于计算机可读形式(例如,可以存储在外部硬盘驱动器124上的数字电影)文件/影片印件106可以直接输入到计算机102中。注意到,在此使用的术语“影片(film)”可以指影片印件或数字电影。Alternatively, the file/movie print 106 may be imported directly into the computer 102 already in a computer readable form (eg, a digital movie that may be stored on an external hard drive 124 ). Note that the term "film" as used herein may refer to film prints or digital films.

软件程序包括在存储器110中存储的眼睛疲劳测量和降低模块114,用于测量在观看3D呈现时由观众或观看者感受到的潜在眼睛疲劳。The software program includes an eyestrain measurement and reduction module 114 stored in memory 110 for measuring potential eyestrain experienced by a viewer or viewer while viewing a 3D presentation.

眼睛疲劳测量和降低模块114包括视差估计器116,被配置用于估计第一图像中的至少一个点与第二图像中的至少一个对应点的视差(第一和第二图像产生立体图像对),并且用于对于第一图像中的至少一个点与第二图像中的至少一个对应点的每一个从所估计的视差生成视差图。视差估计器116包括被配置为匹配第一和第二图像中的像素的像素匹配成本函数132,以及用于将平滑约束应用到视差估计的平滑成本函数134。视差估计器116还包括用于最小化所估计的视差的置信传播算法或函数136、以及动态程序设计算法或函数138,该动态程序设计算法或函数138利用应用于第一和第二图像的确定匹配函数的结果来初始化置信传播函数136以便加速置信传播函数136。要理解的是,置信传播和动态程序设计仅是用于确定视差的两个示例性方法,并且可以由视差估计器116采用其它视差估计方法和算法。The eyestrain measurement and reduction module 114 includes a disparity estimator 116 configured to estimate the disparity of at least one point in the first image and at least one corresponding point in the second image (the first and second images produce a stereoscopic image pair) , and for generating a disparity map from the estimated disparity for each of at least one point in the first image and at least one corresponding point in the second image. The disparity estimator 116 includes a pixel matching cost function 132 configured to match pixels in the first and second images, and a smoothing cost function 134 for applying a smoothing constraint to the disparity estimation. The disparity estimator 116 also includes a belief propagation algorithm or function 136 for minimizing the estimated disparity, and a dynamic programming algorithm or function 138 that utilizes the determined The result of the matching function is used to initialize the belief propagation function 136 in order to speed up the belief propagation function 136 . It is to be understood that belief propagation and dynamic programming are but two exemplary methods for determining disparity and that other disparity estimation methods and algorithms may be employed by disparity estimator 116 .

提供视差转换估计器118,用于确定立体图像的视差的转换或改变。提供眼睛疲劳估计器120,用于基于来自视差估计器116的视差图和来自视差转换估计器118的相关联图像的视差转换,估计潜在眼睛疲劳。A disparity transformation estimator 118 is provided for determining the transformation or change of the disparity of the stereoscopic images. An eyestrain estimator 120 is provided for estimating potential eyestrain based on the disparity map from the disparity estimator 116 and the disparity transformation of the associated images from the disparity transformation estimator 118 .

眼睛疲劳测量和降低模块114还包括视差校正器122,用于调整第一和第二图像之间的视差以降低潜在眼睛疲劳。还提供深度图生成器123,用于通过反转(invert)视差图的视差值而将视差图转换为深度图。在一个实施例中,深度图生成器123与用于降低潜在眼睛疲劳的视差校正器122协同工作,如下面将描述的。The eyestrain measurement and reduction module 114 also includes a parallax corrector 122 for adjusting the parallax between the first and second images to reduce potential eyestrain. A depth map generator 123 is also provided for converting the disparity map into a depth map by inverting the disparity values of the disparity map. In one embodiment, the depth map generator 123 works in conjunction with a parallax corrector 122 for reducing potential eye strain, as will be described below.

图5是根据本公开的一个方面的示例性方法的流程图,该示例性方法用于测量至少两个二维(2D)图像的潜在眼睛疲劳。最初,后期处理设备102获取至少两个二维(2D)图像,例如具有左眼和右眼视图的立体图像对(步骤202)。后期处理设备102可以通过获得处于计算机可读格式的数字标准图像文件(digital master image file)来获取至少两个2D图像。可以通过用数字相机捕获运动图像的时间序列来获取数字视频文件。可替代地,可以通过传统的胶片型相机捕获视频序列。在该方案中,经由扫描设备103扫描影片。5 is a flowchart of an exemplary method for measuring potential eyestrain of at least two two-dimensional (2D) images according to an aspect of the present disclosure. Initially, the post-processing device 102 acquires at least two two-dimensional (2D) images, such as a stereoscopic image pair with left-eye and right-eye views (step 202 ). Post-processing device 102 may acquire at least two 2D images by obtaining a digital master image file in a computer readable format. Digital video files can be acquired by capturing a time sequence of moving images with a digital camera. Alternatively, the video sequence may be captured by a conventional film-type camera. In this scenario, the film is scanned via the scanning device 103 .

要理解的是,不管是扫描影片还是影片已经处于数字格式,影片的数字文件将包括关于各帧的位置的指示或信息,例如,帧号、从影片开始起的时间等。数字图像文件的每个帧将包括一个图像,例如,I1、I2、…InIt will be appreciated that, whether the film is scanned or already in digital format, the digital file of the film will include indications or information about the location of each frame, eg, frame number, time since the start of the film, etc. Each frame of a digital image file will include an image, eg, I 1 , I 2 , . . . In .

可以通过具有相同设置的两个相机拍摄立体图像。相机的任一个被标定为具有相同的焦距、焦点高度和平行焦平面;或者必须基于已知的相机参数来偏转(warp)图像,好像它们由具有平行焦平面的相机拍摄。该偏转过程包括相机标定(camera calibration)和相机调校(camera rectification)。标定和调校过程调整立体图像的外极线,使得外极线精确地为图像的水平扫描线。因为对应点发现沿着外极线发生,所以调校过程将对应搜索简化为仅沿着扫描线搜索,这大大降低了计算成本。对应点是对应于图像中相同场景点的像素。Stereoscopic images can be captured by two cameras with the same settings. Either of the cameras are calibrated to have the same focal length, focal height and parallel focal plane; or the images must be warped based on known camera parameters as if they were taken by a camera with parallel focal planes. The deflection process includes camera calibration and camera rectification. The calibration and adjustment process adjusts the epipolar lines of the stereoscopic image so that the epipolar lines are precisely the horizontal scanning lines of the image. Because correspondence point finding occurs along the epipolar lines, the calibration procedure simplifies the correspondence search to only a search along scan lines, which greatly reduces computational cost. Corresponding points are pixels corresponding to the same scene point in the image.

接下来,在步骤204,经由视差估计器116对于场景(例如,帧)中的每个点估计视差图。用于每个场景点的视差被计算为左眼和右眼图像中的匹配点的相对距离,即,发现左和右图像中的对应于相同场景点的像素。例如,如果左眼图像中一个点的水平坐标是x,并且右眼图像中其对应点的水平坐标是x’,则视差d=x’–x。Next, at step 204 , a disparity map is estimated for each point in the scene (eg frame) via the disparity estimator 116 . The disparity for each scene point is calculated as the relative distance of the matching points in the left and right eye images, ie the pixels in the left and right images are found to correspond to the same scene point. For example, if the horizontal coordinate of a point in the image for the left eye is x, and the horizontal coordinate of its corresponding point in the image for the right eye is x', then the disparity d = x' - x.

在估计视差的一个实施例中,最初,获取立体图像对。计算视差成本函数,包括计算像素匹配成本函数132和计算平滑成本函数134。执行低成本立体匹配优化(例如,动态程序设计函数138),以便得到立体匹配两个图像的初始确定结果。然后低成本优化的结果用于初始化置信传播函数136,以便加速置信传播函数以最小化视差成本函数。要理解的是,在本领域中已知用于视差估计的其它方法,并且可以由本公开的系统和方法采用所述其它方法。In one embodiment of estimating disparity, initially a stereo pair of images is acquired. Computing the disparity cost function includes computing a pixel matching cost function 132 and computing a smoothing cost function 134 . A low-cost stereo matching optimization (eg, dynamic programming function 138) is performed to obtain an initial determination of stereo matching the two images. The results of the low-cost optimization are then used to initialize the belief propagation function 136 in order to speed up the belief propagation function to minimize the disparity cost function. It is to be understood that other methods for disparity estimation are known in the art and may be employed by the systems and methods of the present disclosure.

在步骤206,确定视差转换。基本存在两个类型的视差转换:影片片段(例如,场景、镜头等)之间的可能是突然的视差转换,以及影片片段(例如,场景、镜头等)内的通常连续的视差转换。如下所讨论的,片段意味着包含相同内容的帧的序列。In step 206, a disparity transformation is determined. There are basically two types of parallax transitions: potentially abrupt parallax transitions between film fragments (eg, scenes, shots, etc.), and generally continuous parallax transitions within film fragments (eg, scenes, shots, etc.). As discussed below, a segment means a sequence of frames containing the same content.

片段之间的视差转换通常是不连续和突然的。所以为了测量视差转换,首先估计之前片段的最后帧和当前片段的开始帧的视差图。换句话说,假设在第i片段的末尾处的视差图是Di,并且在第(i+1)片段的开始处的视差图是Di+1,那么视差差是Parallax transitions between clips are often discontinuous and abrupt. So to measure disparity transformation, first estimate the disparity map of the last frame of the previous segment and the start frame of the current segment. In other words, assuming that the disparity map at the end of the i-th segment is D i , and the disparity map at the beginning of the (i+1)-th segment is D i+1 , then the disparity difference is

δD=Di+1-Di (1)δD=D i+1 -D i (1)

为了测量深度改变的幅度,使用绝对视差差To measure the magnitude of the depth change, use the absolute parallax difference

|δD|=|Di+1-Di| (2)|δD|=|D i+1 -D i | (2)

为了得到总体视差改变,使用最大视差转换,其是To get the overall disparity change, the maximum disparity transformation is used, which is

或者使用平均视差转换,其是Or use the average disparity transformation, which is

其中W和H是视差图的宽度和高度。where W and H are the width and height of the disparity map.

对于片段中的画面,视差转换通常是连续的,但是少量时间中的大视差转换也将促进观众的眼睛疲劳。类似于片段之间的视差改变,可以将视差差用作测量,即,δD=Di+1-Di。然而,将在每个帧而不仅仅在片段的边界执行该测量。这里,确定一帧内部的像素上的平均视差值而不是跨越各帧的像素上的平均视差值,然后计算每个连续帧之间的差。Parallax transitions are usually continuous for frames in a clip, but large parallax transitions in a small amount of time will also contribute to viewer eye fatigue. Similar to the disparity change between segments, the disparity difference can be used as a measure, ie, δD=D i+1 −D i . However, this measurement will be performed every frame and not just at segment boundaries. Here, the average disparity value over pixels within a frame is determined instead of over pixels across frames, and then the difference between each successive frame is calculated.

由大视差和/或视差改变导致眼睛疲劳。因此,在步骤208,基于视差图和视差转换估计眼睛疲劳。本公开的系统和方法利用非常粗略的眼睛疲劳模型。该系统和方法假设存在将眼睛疲劳与在每帧处的视差和视差改变相关的函数,并且眼睛疲劳感跨越各帧得到累积,但是随着时间指数衰减。Eye strain caused by large parallax and/or parallax changes. Therefore, at step 208, eye strain is estimated based on the disparity map and the disparity transformation. The systems and methods of the present disclosure utilize a very rough model of eyestrain. The system and method assumes that there is a function relating eye fatigue to disparity and disparity change at each frame, and that eye fatigue is accumulated across frames, but decays exponentially over time.

首先,假设存在“即时眼睛疲劳函数”其将由视差导致的眼睛疲劳与平均视差和视差转换相关。然后,如果在第i帧之后视差保持为零,则眼睛疲劳测量可以表示为衰减模型,如下First, assume that there is an "instant eyestrain function" It compares eye strain caused by parallax with average parallax and parallax transformation relevant. Then, if the disparity remains zero after the i-th frame, the eye strain measurement can be expressed as an attenuation model as follows

其中是建模视差对眼睛疲劳的即时影响的函数。λ是衰减系数。该模型假设如果在屏幕上不再存在视差(即,会聚点变为焦点),则眼睛疲劳将快速渐弱。注意到:因为在各片段之间可能存在视差的突然改变,因此函数对于片段内的帧和各片段之间的帧可能是不同的,这是它们在上面被不同地处理的原因。函数的简单示例可以是之间的线性组合,如下:in is a function that models the immediate effect of parallax on eye fatigue. λ is the attenuation coefficient. The model assumes that if there is no longer parallax on the screen (ie, the point of convergence comes into focus), eyestrain will fade out quickly. Note: Because there may be sudden changes in disparity between fragments, the function It may be different for frames within a fragment and between fragments, which is why they are handled differently above. function A simple example could be with The linear combination between is as follows:

其中a和b是权重系数,其对于片段内的视差转换和片段之间的视差转换应该是不同的。a和b的值可以根据经验确定。where a and b are weight coefficients which should be different for disparity transformation within a fragment and for disparity transformation between fragments. The values of a and b can be determined empirically.

当视差保持跨越各帧改变时,那么,眼睛疲劳感应该随着时间得到累积。然而,眼睛疲劳不能永远上升,所以可以采用函数来建模眼睛疲劳感随着时间的渐弱效果。S函数用于对眼睛疲劳感的饱和建模,如下:As the disparity keeps changing across frames, then eye strain should build up over time. However, eye strain cannot rise forever, so a function can be used to model the fading effect of eye strain over time. The S function is used to model the saturation of eye fatigue as follows:

S(ES)=1/(1+exp(-ES)) (7)S(ES)=1/(1+exp(-ES)) (7)

在图6中示出该函数的形状。The shape of this function is shown in FIG. 6 .

给定这些要素,在帧i处总体眼睛疲劳测量可以递归地定义,如下:Given these elements, the overall eye strain measure at frame i can be recursively defined as follows:

其中ESi(ti)是在第i帧的眼睛疲劳测量,并且ti是第i帧的时间,并且λ是用于控制衰减速度的常数。该测量的计算可以通过随着时间的模拟而实现。where ES i (t i ) is the eye strain measurement at frame i , and ti is the time of frame i, and λ is a constant used to control the decay rate. Calculation of this measurement can be achieved by simulation over time.

图7是随着时间的眼睛疲劳测量方程式的图示。确定在第(i-t)帧的眼睛疲劳,并且将exp(-λt)的衰减因子应用到眼睛疲劳测量。结果然后与第i帧的即时眼睛疲劳函数组合。将S形函数应用于组合结果以确定在第i帧的眼睛疲劳。Figure 7 is a graphical representation of eye strain measurement equations over time. The eyestrain at frame (i-t) is determined, and an attenuation factor of exp(-λt) is applied to the eyestrain measure. The result is then combined with the instant eye fatigue function for frame i. Apply a sigmoid function to the combined results to determine eye strain at frame i.

参照图8,图示了即时眼睛疲劳函数和总体眼睛疲劳测量。在每个帧(例如,帧1、2等)上的曲线是该帧的即时眼睛疲劳函数,并且在测量点上的曲线是即时眼睛疲劳函数的累积。再次参照图5,在步骤210,确定由观看者经历的总体眼睛疲劳是否可接受。在一个实施例中,将总体眼睛疲劳测量对操作者(例如,导演和剪辑师)形象化为曲线图,以便根据曲线图确定眼睛疲劳是否太高。在另一实施例中,可以将总体眼睛疲劳测量与预定阈值比较,以便确定是否需要校正。例如,生成表示总体眼睛疲劳测量的曲线,然后将曲线上的每个点的值与预定阈值比较。在该实施例中,对于不同类型的场景和电影,预定阈值将是不同的。Referring to Figure 8, an instant eyestrain function and an overall eyestrain measure are illustrated. The curve at each frame (eg, frame 1, 2, etc.) is the instant eyestrain function for that frame, and the curve at the measurement points is the accumulation of the instant eyestrain function. Referring again to FIG. 5, at step 210, it is determined whether the overall eye strain experienced by the viewer is acceptable. In one embodiment, the overall eye strain measurement is visualized as a graph for operators (eg, directors and editors) to determine from the graph whether eye strain is too high. In another embodiment, the overall eye strain measurement may be compared to a predetermined threshold in order to determine whether correction is required. For example, a curve representing a measure of overall eye strain is generated, and the value of each point on the curve is compared to a predetermined threshold. In this embodiment, the predetermined threshold will be different for different types of scenes and movies.

如果确定眼睛疲劳太高,则在步骤212将对立体图像执行视差校正或调和(grading)以降低眼睛疲劳。调和是平滑跨越各帧的视差值的过程。通过平滑跨越各帧的视差,可以减少视差的突然改变,并且因此可以降低眼睛疲劳。用于减少视差的一个示例性方法被称为会聚调整,其通过向左或向右偏移右眼图像以调整会聚点来实现。通过向左或向右偏移右眼图像,可以人工减少或增加像素的视差,分别导致更小或更大的总体深度。If eye strain is determined to be too high, then at step 212 a parallax correction or grading will be performed on the stereoscopic images to reduce eye strain. Blending is the process of smoothing disparity values across frames. By smoothing the disparity across frames, sudden changes in disparity can be reduced, and thus eye fatigue can be reduced. One exemplary method for reducing parallax is called convergence adjustment, which is achieved by shifting the right-eye image left or right to adjust the point of convergence. By shifting the right-eye image to the left or right, the parallax of pixels can be artificially reduced or increased, resulting in smaller or greater overall depth, respectively.

图像偏移的问题在于所有像素的深度增加相同量,这与3D场景几何形状无关。然而,如果精确的深度图是可用的,则可能合成具有新的虚拟相机位置的场景的新视图,以便减少视差。在该实施例中,使用以下公式z=Bf/d(其中B是两个相机之间的距离,也称为基线,并且f是相机的焦距),通过深度图生成器123将上面确定的每个场景点的视差值d转换为深度值z,(即从场景点到相机的距离)。至少一个图像(例如,左眼视图图像)的每个的深度值存储在深度图中。深度图的深度值相应改变,以便合成新的视图同时减小图像的视差。例如,为了创建新的视图,首先改变深度值,然后重新绘制新的左或右(或两者)图像。重新绘制过程采用左(或右)图像和深度图,并且创建新的右(或左)图像。深度值是像素的3D信息,因此,可以使用诸如光线跟踪的技术,以便将3D点绘制为新的视图中的2D像素。新的视图将具有更小的视差或更小的视差转换,并且因此对于观看者或观众将导致降低的眼睛疲劳。The problem with image shifting is that the depth of all pixels is increased by the same amount, independent of the 3D scene geometry. However, if an accurate depth map is available, it is possible to synthesize new views of the scene with new virtual camera positions in order to reduce parallax. In this embodiment, each of the above-determined The disparity value d of a scene point is converted into a depth value z, (ie, the distance from the scene point to the camera). A depth value for each of at least one image (eg, a left-eye view image) is stored in a depth map. The depth values of the depth map are changed accordingly in order to synthesize new views while reducing image parallax. For example, to create a new view, first change the depth value, then redraw the new left or right (or both) image. The repainting process takes a left (or right) image and a depth map, and creates a new right (or left) image. Depth values are 3D information for pixels, so techniques such as ray tracing can be used in order to draw 3D points as 2D pixels in the new view. The new view will have less parallax or less parallax transition and thus will result in reduced eyestrain for the viewer or viewer.

现在参照图4,例如在存储设备124中存储对应图像和相关联的深度图,并且可以被检索用于3D回放。此外,运动画面或视频剪辑的所有校正图像可以与相关联的深度图存储在表示运动画面或剪辑的立体版本的单个数字文件130中。数字文件130可以存储在存储设备124中用于之后检索,例如,用于冲印原始影片的立体版本。Referring now to FIG. 4 , corresponding images and associated depth maps are stored, eg, in storage device 124 , and can be retrieved for 3D playback. Furthermore, all corrected images of a motion picture or video clip may be stored with associated depth maps in a single digital file 130 representing a stereoscopic version of the motion picture or clip. Digital file 130 may be stored on storage device 124 for later retrieval, for example, for developing a stereoscopic version of the original film.

尽管在此已经详细示出和描述了并入本公开的教导的实施例,但是本领域技术人员可以容易地设计仍并入这些教导的任何其它变化的实施例。已经描述了用于测量在观看3D呈现时感受到的潜在眼睛疲劳的系统和方法的优选实施例(其旨在是说明性的而不是限制性的),注意到根据上面的教导本领域技术人员可以进行修改和变化。因此,要理解的是在所公开的本公开的特定实施例中可以进行、在如由所附权利要求概述的本公开的范围内的改变。Although embodiments that incorporate the teachings of the present disclosure have been shown and described in detail herein, those skilled in the art can readily devise any other varied embodiments that still incorporate these teachings. Having described a preferred embodiment of a system and method for measuring potential eye strain experienced while viewing a 3D presentation (which is intended to be illustrative and not limiting), it is noted that those skilled in the art in light of the above teachings Modifications and variations are possible. It is therefore to be understood that changes may be made in the particular embodiments of the disclosure disclosed which are within the scope of the disclosure as outlined by the appended claims.

Claims (22)

1. the method for potential eye fatigue of one kind measurement when viewing three-dimensional (3D) is presented, methods described include:
The first image and the second image (202) are obtained from the first fragment;
Estimate the parallax (204) of at least one corresponding points at least one point and the second image in the first image;
Estimate the parallax conversion (206) of the sequence of the first and second images;And
The parallax conversion of sequence based on parallax and the first and second images determines that the potential eyes when watching 3D and presenting are tired Labor (208);
Wherein the first image and the second image produce the three-dimensional stereo pairs presented.
2. the method as described in claim 1, wherein estimating disparity switch process (206) also include:
The parallax of the last frame of fragment before estimation;
Estimate the parallax of the first frame of the first fragment;And
It is determined that the difference between the parallax of the parallax of the last frame of fragment and the first frame of the first fragment before.
3. the method as described in claim 1, wherein estimating disparity switch process (206) also include:
Estimate the parallax in each frame of multiple frames of the first fragment;And
Determine the difference between the parallax of each frame of the first fragment.
4. the method as described in claim 1, wherein determining that potential eye fatigue step (208) also includes determining to be used for the first He The instant eye fatigue function of each frame of the sequence of second image.
5. method as claimed in claim 4, wherein determining that potential eye fatigue step (208) is also included decay factor application To the instant eye fatigue function of each frame of the sequence for the first and second images.
6. method as claimed in claim 5, wherein determining that potential eye fatigue step (208) is additionally included in predetermined time period The instant eye fatigue function of each frame of the sequence for the first and second images after upper accumulation decay.
7. method as claimed in claim 6, wherein determining that potential eye fatigue step (208) also includes making first and second The eye fatigue function saturation accumulated in the sequence of image.
8. the method as described in claim 1, in addition to:
Determine whether potential eye fatigue is acceptable (210);And
If potential eye fatigue is unacceptable, the parallax (212) of the first and second images is corrected.
9. method as claimed in claim 8, wherein aligning step (212) also include at least the one of the first and second images of skew It is individual, to adjust convergent point of the eyes of beholder relative to the first and second images.
10. method as claimed in claim 8, wherein aligning step (212) also have what is reduced including the first fragment of synthesis The new view of parallax.
11. one kind is used for the system (100) for measuring the potential eye fatigue when viewing three-dimensional (3D) is presented, the system bag Include:
For obtaining the part of the first image and the second image from the first fragment;
Disparity estimator (116), for estimate the first image at least one point with it is at least one corresponding in the second image The parallax of point;
Parallax conversion estimator (118), the parallax conversion of the sequence for estimating the first and second images;And
Eye fatigue estimator (120), the parallax conversion for the sequence based on parallax and the first and second images determine Watch potential eye fatigue when 3D is presented;
Wherein the first image and the second image produce the three-dimensional stereo pairs presented.
12. system (100) as claimed in claim 11, wherein parallax conversion estimator (118) operation are come fragment before estimating Last frame parallax, estimate the parallax of the first frame of the first fragment, and before determining the last frame of fragment parallax and the Difference between the parallax of first frame of one fragment.
13. system (100) as claimed in claim 11, wherein parallax conversion estimator (118) are operated to estimate the first fragment Multiple frames each frame in parallax, and determine the first fragment each frame parallax between difference.
14. system (100) as claimed in claim 11, wherein eye fatigue estimator (120) are operated to determine to be used for first With the instant eye fatigue function of each frame of the sequence of the second image.
15. the operation of system (100) as claimed in claim 14, wherein eye fatigue estimator (120) should by decay factor Use the instant eye fatigue function of each frame of the sequence for the first and second images.
16. the operation of system (100) as claimed in claim 15, wherein eye fatigue estimator (120) came in the predetermined time The instant eye fatigue function of each frame of the sequence for the first and second images in section after accumulation decay.
17. system (100) as claimed in claim 16, wherein eye fatigue estimator (120) operation makes first and the The eye fatigue function saturation accumulated in the sequence of two images.
18. system (100) as claimed in claim 11, in addition to parallax offset mechanism (122), if eye fatigue estimator (120) determine that potential eye fatigue is unacceptable, then parallax offset mechanism (122) is operated to correct regarding for the first and second images Difference.
19. system (100) as claimed in claim 18, wherein parallax offset mechanism (122) are operated to offset the first and second figures Picture it is at least one, to adjust convergent point of the eyes of beholder relative to the first and second images.
20. system (100) as claimed in claim 18, wherein parallax offset mechanism (122) are operated to synthesize the tool of the first fragment There is the new view of the parallax of reduction.
21. a kind of program storage device that can be read by machine, described program storage device, which positively embodies, to be performed by machine , the program of the instruction of method and step for performing to measure the potential eye fatigue when viewing three-dimensional (3D) is presented, Methods described includes:
The first image and the second image (202) are obtained from the first fragment;
Estimate the parallax (204) of at least one corresponding points at least one point and the second image in the first image;
Estimate the parallax conversion (206) of the sequence of the first and second images;And
The parallax conversion of sequence based on parallax and the first and second images determines that the potential eyes when watching 3D and presenting are tired Labor (208);
Wherein the first image and the second image produce the three-dimensional stereo pairs presented.
22. program storage device as claimed in claim 21, wherein methods described also include:
Determine whether potential eye fatigue is acceptable (210);And
If potential eye fatigue is unacceptable, the parallax (212) of the first and second images is corrected.
CN201510315583.1A 2008-05-12 2008-05-12 The system and method for measuring the potential eye fatigue of stereoscopic motion picture Expired - Fee Related CN104980732B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510315583.1A CN104980732B (en) 2008-05-12 2008-05-12 The system and method for measuring the potential eye fatigue of stereoscopic motion picture

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN200880129158.1A CN102027752B (en) 2008-05-12 2008-05-12 For measuring the system and method for the potential eye fatigue of stereoscopic motion picture
CN201510315583.1A CN104980732B (en) 2008-05-12 2008-05-12 The system and method for measuring the potential eye fatigue of stereoscopic motion picture

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
CN200880129158.1A Division CN102027752B (en) 2008-05-12 2008-05-12 For measuring the system and method for the potential eye fatigue of stereoscopic motion picture

Publications (2)

Publication Number Publication Date
CN104980732A CN104980732A (en) 2015-10-14
CN104980732B true CN104980732B (en) 2018-01-16

Family

ID=54288312

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510315583.1A Expired - Fee Related CN104980732B (en) 2008-05-12 2008-05-12 The system and method for measuring the potential eye fatigue of stereoscopic motion picture

Country Status (1)

Country Link
CN (1) CN104980732B (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005177076A (en) * 2003-12-18 2005-07-07 Sophia Co Ltd Game machine
CN1838778A (en) * 2005-03-24 2006-09-27 精工爱普生株式会社 Stereoscopic image display device and method
JP4121881B2 (en) * 2003-03-28 2008-07-23 株式会社トプコン Three-dimensional observation state measuring apparatus and method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4121881B2 (en) * 2003-03-28 2008-07-23 株式会社トプコン Three-dimensional observation state measuring apparatus and method
JP2005177076A (en) * 2003-12-18 2005-07-07 Sophia Co Ltd Game machine
CN1838778A (en) * 2005-03-24 2006-09-27 精工爱普生株式会社 Stereoscopic image display device and method

Also Published As

Publication number Publication date
CN104980732A (en) 2015-10-14

Similar Documents

Publication Publication Date Title
CN102027752B (en) For measuring the system and method for the potential eye fatigue of stereoscopic motion picture
CN102047288B (en) System and method for depth extraction of images using forward and backward depth prediction
JP5153940B2 (en) System and method for image depth extraction using motion compensation
JP5156837B2 (en) System and method for depth map extraction using region-based filtering
CA2668941C (en) System and method for model fitting and registration of objects for 2d-to-3d conversion
CN102598683B (en) Stereoscopic video creation device and stereoscopic video creation method
CN101542536A (en) System and method for compositing 3D images
CN102668572B (en) Method and apparatus for optimal motion reproduction in stereoscopic digital cinema
CN104980732B (en) The system and method for measuring the potential eye fatigue of stereoscopic motion picture

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20190111

Address after: Paris France

Patentee after: Interactive Digital Madison Patent Holdings

Address before: I Si Eli Murli Nor, France

Patentee before: THOMSON LICENSING

Effective date of registration: 20190111

Address after: I Si Eli Murli Nor, France

Patentee after: THOMSON LICENSING

Address before: I Si Eli Murli Nor, France

Patentee before: THOMSON LICENSING

CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20180116

Termination date: 20190512