US20130063558A1 - Systems and Methods for Incorporating Two Dimensional Images Captured by a Moving Studio Camera with Actively Controlled Optics into a Virtual Three Dimensional Coordinate System - Google Patents
Systems and Methods for Incorporating Two Dimensional Images Captured by a Moving Studio Camera with Actively Controlled Optics into a Virtual Three Dimensional Coordinate System Download PDFInfo
- Publication number
- US20130063558A1 US20130063558A1 US13/232,938 US201113232938A US2013063558A1 US 20130063558 A1 US20130063558 A1 US 20130063558A1 US 201113232938 A US201113232938 A US 201113232938A US 2013063558 A1 US2013063558 A1 US 2013063558A1
- Authority
- US
- United States
- Prior art keywords
- camera
- calibration
- studio
- zoom
- studio camera
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/24—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
- G01B11/25—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object
- G01B11/254—Projection of a pattern, viewing through a pattern, e.g. moiré
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/2224—Studio circuitry; Studio devices; Studio equipment related to virtual studio applications
Definitions
- the invention relates generally to image capture and more particularly to systems and methods for incorporating two dimensional images into three dimensional coordinate systems.
- a tremendous amount of video content is produced professionally every year, with applications in television broadcast, movie production, and webcasts to name a few.
- the vast majority of this content is two dimensional—NTSC, PAL, or High Definition video, with many other formats, resolutions, sampling rates, and intended uses.
- Another very large content production source is 3D modeling and character animation, with applications in video games, movies, and television, virtual worlds, simulators, etc.
- This content is produced and stored in three dimensional format, though in applications such as movies and television the produced product is a two dimensional projection of the three dimensional asset, whether a character, prop, scene, etc.
- novel methods described herein provide for complete freedom of movement of the studio camera within the volume of interest, and allow un-restricted control of the studio camera optics (zoom and focus), with the complete calibration performed in-situ, with the only calibration devices needed a simple ruled rigid stick with two or more visible features (often retro-reflective markers) and a rigid ruled right-angle, with three or more ruled visible features (often retro-reflective markers).
- An embodiment of the present invention provides a method for incorporating two dimensional images captured by a moving studio camera with actively controlled optics into a virtual three dimensional coordinate system.
- the method includes encoding and accessing the zoom and focus settings for the at least one studio camera, identifying a marked calibration ruler, locating a marked calibration right angle on the calibration ruler, capturing calibration data over a functional range of the zoom and focus controls, calculating the internal camera calibration parameters and applying the parameters during video production, and tracking the moving studio camera.
- Another embodiment of the present invention provides a system for incorporating two dimensional images captured by at least one moving studio camera with actively controlled optics into a virtual three dimensional coordinate system.
- the system includes a device for encoding and accessing the zoom and focus settings for the moving at least one studio camera; a marked calibration ruler; a marked calibration right angle; a host module, adapted to capture calibration data over a functional range of the zoom and focus controls from the moving at least one studio camera data; a first logic processing module configured to receive and calculate internal camera calibration parameters and apply the parameters during video production, and a tracking module configured to track the moving camera.
- FIG. 1 depicts one embodiment of a studio, including a three dimensional imaging system, and two exemplar two dimensional studio cameras.
- FIG. 2 illustrates an embodiment of a two dimensional studio camera with dynamic optics.
- FIG. 3 depicts a frame of motion capture data after calibration of the motion capture system in accordance with one embodiment.
- FIG. 4 shows the motion capture system of FIG. 3 in a pre-calibration stage in accordance with an embodiment of the present invention.
- FIG. 5 show the motion capture system of FIG. 3 in a main calibration stage in accordance with an embodiment of the present invention.
- FIG. 6 depicts one embodiment of a frame of motion capture system calibration data (wand data) in accordance with an embodiment of the present invention.
- FIG. 7 depicts an embodiment of multiple frames of studio camera calibration data in accordance with an embodiment of the present invention.
- FIG. 8 shows a studio camera in operation in accordance with one embodiment.
- the term “or” is an inclusive “or” operator, and is equivalent to the term “and/or,” unless the context clearly dictates otherwise.
- the term “based on” is not exclusive and allows for being based on additional factors not described, unless the context clearly dictates otherwise.
- the meaning of “a,” “an,” and “the” include plural references.
- the meaning of “in” includes “in” and “on.”
- the term “coupled” implies that the elements may be directly connected together or may be coupled through one or more intervening elements. Further reference may be made to an embodiment where a component is implemented and multiple like or identical components are implemented. In such instances a reference may be made to (a . . . n), this is not a limitation on the placement or number of components but indicates rather that a multiple and perhaps indefinite number of components may be implemented and the placement of such may vary.
- FIG. 1 shows a studio 100 , wherein the studio 100 includes a three dimensional motion capture system and two exemplar two dimensional studio cameras 130 , 140 .
- a subject 120 moves within the studio 100 .
- the embodiments of the present invention are capable of being incorporated into any volume of space. Shown are a Virtual 3D coordinate system 110 ; the fixed cameras 150 ( a . . . n ).
- One exemplar fixed camera 160 is shown in detail, demonstrating that each fixed camera 160 , 150 ( a . . . n ) has a Camera Arbitrary 3D coordinate system 161 and an Image Plane 2D coordinate system 162 , for the (a . . . n) fixed cameras shown.
- 2 are the studio cameras 130 , 140 , each of which has a Camera Arbitrary 3D coordinate system 131 , 141 and an Image Plane 2D coordinate system 132 , 142 , for each exemplar device 130 , 140 .
- FIG. 2 depicts an embodiment of a two dimensional studio camera 200 with dynamic optics.
- the zoom 220 and focus 210 controls of the studio camera 200 are encoded electronically and sent back 230 to a data acquisition system.
- the studio camera 200 may be coupled to the data acquisition directly or indirectly.
- the focus 210 and zoom 220 may be controlled manually or through automation. Although the automation as shown implements a cog system 211 , 221 , this is not intended to be a limitation on the method or mean for controlling the zoom 220 and focus 210 controls.
- the data acquisition system describes where along their full range of travel each of the zoom and focus controls are. For example, if the zoom and focus are controlled by a linear stage, the current center spoke locations 213 and 223 might be transmitted.
- the data acquisition system may comprise a computer, which has a host module, and various software modules including but not limited to coordinate calculation modules, data acquisition modules, logic processing modules, tracking modules, and a database modules for storing data
- FIG. 3 shows a frame of motion capture data after calibration of the motion capture system 300 .
- Fixed cameras 310 ( a . . . n ) cooperate with an imaging device 320 .
- Retro-reflective markers 340 which are rigidly affixed to the studio camera 320 , are imaged by the fixed cameras 310 ( a . . . n ).
- Each of the motion capture devices 310 are calibrated to a common virtual coordinate system, so that it is possible to generate a ray from the feature locations on each of the fixed camera planes 315 ( a . . .
- retro-reflective markers 340 are attached rigidly to a studio camera.
- the 3D coordinates of this group of markers can be used to form an orthogonal 3D coordinate system, either by construction or mathematical manipulation.
- FIG. 4 shows a motion capture system in the pre-calibration stage 400 .
- Three or more retro-reflective markers 405 ( a . . . n ) are attached to a rigid orthogonal device 420 and this device is placed at the desired location 410 of the virtual coordinate system. This produces an initial estimate of the motion capture system's fixed camera's 430 ( a . . . n) external and internal camera parameters.
- FIG. 5 show a motion capture system in a main calibration stage 500 .
- the studio camera(s) 520 , 530 are introduced to the studio, any one of which has a motion capture system detectable rigid coordinate system 521 , 531 rigidly attached.
- the studio cameras may operate as automated systems 520 or may be manually operated 530 , further there may be multiple cameras operated in the same or multiple ways. There may be one or more studio cameras, although two studio cameras are shown in the embodiment, this is not intended to be a limitation on the scope of the present embodiments.
- a human operator 540 walks around the volume of interest, waving a calibration wand 550 over as much of the volume as practical.
- embodiments of the present invention contemplate the operator waving a calibration wand over less that the entire volume.
- the human operator may be replaced by an automated operator such as a robot, further, embodiments may include non-human subjects subject as animals or anything capable of motion.
- the studio camera operator, human or automatic images the calibration wand, changing zoom, focus, and location over time.
- Fixed cameras 510 ( a . . . n ) cooperate to calibrate the system.
- FIG. 6 shows a frame of motion capture system calibration data (wand data) 600 .
- At least two retro-reflective markers located on the calibration wand 650 are imaged by the motion capture system 610 ( a . . . n) and by the studio camera 620 .
- the motion capture system observes 2D projections ( 630 - 634 ) of the 3D markers in multiple imaging planes. The true 3D distances between the markers are known, and a current estimate of the camera parameters and the 3D locations of the markers are also known.
- the studio camera calibration 622 is similarly collected. An iterative process is used to adjust the estimate of the camera parameters and the estimate of the markers' locations until convergence.
- FIG. 7 shows multiple frames of studio camera calibration data 700 .
- the motion capture system is capable of producing highly accurate 3D locations for all of the wand data 750 .
- the projection of this 3D object onto the studio camera 710 having its own marker system 715 , imaging plane produces corresponding 2D data 730 - 736 over the functional range of the studio camera optics (zoom and focus).
- the encoded zoom and focus range data provide a first estimate of the studio camera calibration parameters for a given frame of data. A number of frames with the same encoded zoom and focus are automatically examined, and a set of simultaneous equations involving the 3D marker locations and their 2D projections is set up, from which the studio camera calibration parameters at this zoom and focus are calculated. This procedure is repeated at all of the other zoom and focus conditions acquired during calibration.
- FIG. 8 depicts a studio camera in operation 800 .
- the motion capture system 850 ( a . . . n ) observes the markers 834 attached to the camera 820 and determines the location of the Camera Arbitrary coordinate system 831 in the virtual coordinate system.
- the encoded zoom and focus data generated by the studio camera 820 is read, and the appropriate current frame camera calibration data is interpolated from the zoom and focus conditions captured during calibration.
- the fixed cameras 850 ( a . . . n) produce the virtual coordinate system 810 , the location of the Camera Arbitrary 3D coordinate system 831 , and the Camera Image Plane coordinate system 832 .
- the studio camera 820 produces the image data 830 .
- Embodiments of the invention combines this information, along with information from the zoom and focus encoders in the studio camera, and the calibration data, to transform the image data 830 to the final image 840 in the virtual coordinate system 810 .
- embodiments of the present system capture two-dimensional data and transform it into a three dimensional system.
- the studio cameras are modeled and the parameters of the camera models are determined. This process may be referred to as calibration.
- a motion tracking system is implemented to track the movement of the studio camera while synchronously collecting zoom and focus settings from that camera.
- the tracking is calibrated so that the principal point of the studio camera's lens is tracked.
- generator lock or genlock is not employed so that it is not necessary to synchronize the video output of one source, or a specific reference signal from a signal generator, with other pictures or images. Generally such synchronization enables the coincidence of signals in time at a combining or mixing or switching point.
- the novel embodiments of the present invention allow such without the need for genlock.
- the process involves at least two steps—Calibrating and Tracking.
- Calibration is the process that provides the offset from object on the camera to the principle point of the camera's lens and also provides the parameters describing the properties of the camera's lens.
- To calculate offset two objects are tracked—the object on the camera and the calibration wand—in addition, video data of the wand from the camera is collected.
- the system is then calibrated. In a preferred embodiment, the system is calibrated from about ten frames of data.
- the calibration process in then repeated at various lens settings. Although ten frames of data are used for calibration in a preferred embodiment, this number of frames is not intended to be a limitation on the scope of the embodiments and a fewer or greater number of frames may be used to complete the calibration.
- the tracking object has two segments.
- the first segment represents the tracking object.
- the second segment represents the studio camera's principal point. These segments have six degrees of freedom (6 DOF).
- the data for the calibration includes N frames (nFrames) of data, where N may be any number of frames. Preferably N is between 5 and 15 frames.
- Each frame has 2 or more wand points in the video and 2 or more wand points from the motion capture (“mocap”) data as well as the position of the tracking object from the motion capture system.
- the data includes the following:
- the calibration data is captured over the functional range of the zoom and focus controls from the studio camera. All of the calibration data is transmitted to a computer processing system. The system receives the data and calculates internal camera calibration parameters, then applies these parameters during video production.
- Each frame of data is transformed into the coordinate system of the tracking object. This gives data as though the camera were stationary very close to the origin.
- the parameters to be calculated are the 6 degree of freedom offsets to the principal point and the camera lens parameters.
- the lens parameters calculated for the studio camera model include the following:
- lens parameters are used to convert a camera pixel coordinate (X,Y) of a marker into the corrected slopes (U,V) of the ray from the camera to the marker according to the following:
- lens distortion parameters are calculated over the functional range of the studio camera zoom and focus controls. Given an XYZ location in the camera's field of view and a set of lens parameters the expected UV position of the centroid can be calculated.
- the solved set of 11 parameters represents a single calibration at a particular zoom and focus setting.
- the process is repeated with different zoom and focus settings.
- This multi-step process provides data for interpolating the calibration for any given setting.
- the object preferably has five or more retro-reflective markers, but may have as few as three.
- the tracking minimizes 2D motion capture camera residuals.
- a calibration setting is used that is an interpolation between the measured calibration settings.
- the second segment then, has a dynamic 6 DOF offset from the first segment depending on the interpolated calibration setting.
- the “Stationary Lock” feature detects this situation. If the position of the camera hasn't moved from a specified detection distance for a specified minimum duration, then it is considered stationary and the “Locked Position” is sent instead of the tracked position.
- This algorithm includes a minimum amount of time that the camera is stationary before it is locked so that the system doesn't keep popping in and out of its locked state when the camera is being moved very slowly.
- the cameras used in that tracking changes as the object moves across the volume.
- the very first frames of data are typically sub-par, as these are almost always views of partially obscured markers.
- the refinement tracks the history of visibility of a marker by a camera, and then weights a new camera contribution according to the length of it's time history. The weight for the new camera contribution ramps up, dynamically, from a small weight to the full weight over some selected period of frames.
- the camera data is weighted by temporal contiguity.
- weighting is completed in this way, the images are presented contiguous or close in time which results in improved imaging.
- This dynamic camera tracker system then provides a “performance animation” or “performance capture” capability that cannot be achieved by conventional means.
- the system provides the capability to track a performer and stream the data to get a live or real-time, interactive performance. This could be any part of a performer—full body, face and hands or any subset of that. This data is part of the same data stream that contains the camera tracking. It also includes the ability to track a prop.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Studio Devices (AREA)
Abstract
A method for incorporating two dimensional images such as those captured by a video camera, which is moving and whose optics, particularly zoom and focus, are controlled by a human or by automatic means, into a virtual three dimensional coordinate system is provided In one embodiment the method acquires calibration data over the functional range of the studio camera optics, and then in operation dynamically performs the appropriate transformations needed to map the video stream to the virtual coordinate system, even as the acquiring studio camera moves, zooms, and changes focus.
Description
- The invention relates generally to image capture and more particularly to systems and methods for incorporating two dimensional images into three dimensional coordinate systems.
- A tremendous amount of video content is produced professionally every year, with applications in television broadcast, movie production, and webcasts to name a few. The vast majority of this content is two dimensional—NTSC, PAL, or High Definition video, with many other formats, resolutions, sampling rates, and intended uses. Another very large content production source is 3D modeling and character animation, with applications in video games, movies, and television, virtual worlds, simulators, etc. This content is produced and stored in three dimensional format, though in applications such as movies and television the produced product is a two dimensional projection of the three dimensional asset, whether a character, prop, scene, etc.
- For some applications, it would be beneficial to be able to capture live video from commercially available and in-current-use video systems, such as TV or movie cameras, and be able to project that video stream into a three dimensional virtual coordinate system, which underpins any three dimensional modeling application, such as a virtual set, a virtual world, etc. However, there is an extremely limiting condition common to extant approaches. The novel systems and methods presented herein overcome these limitations, and address the need in modern video production practice to move studio cameras and change the zoom and focus of the cameras dynamically, while a performance is under production, including possibly being broadcast live. This novel approach is distinct from traditional approaches of setting studio camera locations, setting the optics at one particular zoom and focus, and acquiring calibration and production video at that one set of camera parameters, internal and external.
- Others deal with this issue only partially and in a way that does not allow or suggest the un-restricted production control practice of a freely moving, zooming, and focusing studio camera. Reynolds U.S. Pat. No. 5,889,550 teaches a method to deal with a moving camera, but requires and is dependent on acquiring video with fixed optics. Alexander, US Patent Publication No. 20070076096 teaches a method for calibrating moving, dynamic-optics imagers that is distinct from the novel method herein and is limited by two restrictions: the devices described are 3D imagers including an illumination system, and more importantly the method requires the use of a very significant pre-calibration device Alexander US Publication No. 20070104361, including a linear motorized track and a large planar checkerboard target. In contrast, the novel methods described herein provide for complete freedom of movement of the studio camera within the volume of interest, and allow un-restricted control of the studio camera optics (zoom and focus), with the complete calibration performed in-situ, with the only calibration devices needed a simple ruled rigid stick with two or more visible features (often retro-reflective markers) and a rigid ruled right-angle, with three or more ruled visible features (often retro-reflective markers).
- An embodiment of the present invention provides a method for incorporating two dimensional images captured by a moving studio camera with actively controlled optics into a virtual three dimensional coordinate system. The method includes encoding and accessing the zoom and focus settings for the at least one studio camera, identifying a marked calibration ruler, locating a marked calibration right angle on the calibration ruler, capturing calibration data over a functional range of the zoom and focus controls, calculating the internal camera calibration parameters and applying the parameters during video production, and tracking the moving studio camera.
- Another embodiment of the present invention provides a system for incorporating two dimensional images captured by at least one moving studio camera with actively controlled optics into a virtual three dimensional coordinate system. The system includes a device for encoding and accessing the zoom and focus settings for the moving at least one studio camera; a marked calibration ruler; a marked calibration right angle; a host module, adapted to capture calibration data over a functional range of the zoom and focus controls from the moving at least one studio camera data; a first logic processing module configured to receive and calculate internal camera calibration parameters and apply the parameters during video production, and a tracking module configured to track the moving camera.
- Other objectives, advantages, and applications of the embodiments of the present invention will be made clear by the following detailed description of a preferred embodiment of the present invention. The description makes reference to drawings in which:
- Non-limiting and non-exhaustive embodiments of the present invention are described with reference to the following drawings. In the drawings, like reference numerals refer to like parts throughout the various figures unless otherwise specified.
- For a better understanding of embodiments of the present invention, reference is made to the following Detailed Description, which is to be read in association with the accompanying drawings, wherein:
-
FIG. 1 depicts one embodiment of a studio, including a three dimensional imaging system, and two exemplar two dimensional studio cameras. -
FIG. 2 illustrates an embodiment of a two dimensional studio camera with dynamic optics. -
FIG. 3 depicts a frame of motion capture data after calibration of the motion capture system in accordance with one embodiment. -
FIG. 4 shows the motion capture system ofFIG. 3 in a pre-calibration stage in accordance with an embodiment of the present invention. -
FIG. 5 show the motion capture system ofFIG. 3 in a main calibration stage in accordance with an embodiment of the present invention. -
FIG. 6 depicts one embodiment of a frame of motion capture system calibration data (wand data) in accordance with an embodiment of the present invention. -
FIG. 7 depicts an embodiment of multiple frames of studio camera calibration data in accordance with an embodiment of the present invention. -
FIG. 8 shows a studio camera in operation in accordance with one embodiment. - The embodiments of the present invention are described more fully hereinafter with reference to the accompanying drawings, which form a part hereof, and which show, by way of illustration, specific exemplary embodiments by which the invention may be practiced. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. Among other things, the present invention may be embodied as systems, or devices. The following detailed description should not to be taken in a limiting sense.
- Throughout the specification and claims, the following terms take the meanings explicitly associated herein, unless the context clearly dictates otherwise. The phrase “in one embodiment” as used herein does not necessarily refer to the same embodiment, though it may. Furthermore, the phrase “in another embodiment” as used herein does not necessarily refer to a different embodiment, although it may. Thus, as described below, various embodiments of the invention may be readily combined, without departing from the scope or spirit of the invention.
- In addition, as used herein, the term “or” is an inclusive “or” operator, and is equivalent to the term “and/or,” unless the context clearly dictates otherwise. The term “based on” is not exclusive and allows for being based on additional factors not described, unless the context clearly dictates otherwise. In addition, throughout the specification, the meaning of “a,” “an,” and “the” include plural references. The meaning of “in” includes “in” and “on.” The term “coupled” implies that the elements may be directly connected together or may be coupled through one or more intervening elements. Further reference may be made to an embodiment where a component is implemented and multiple like or identical components are implemented. In such instances a reference may be made to (a . . . n), this is not a limitation on the placement or number of components but indicates rather that a multiple and perhaps indefinite number of components may be implemented and the placement of such may vary.
-
FIG. 1 shows astudio 100, wherein thestudio 100 includes a three dimensional motion capture system and two exemplar two 130,140. In operation adimensional studio cameras subject 120 moves within thestudio 100. Although referred to as a studio, the embodiments of the present invention are capable of being incorporated into any volume of space. Shown are a Virtual3D coordinate system 110; the fixed cameras 150 (a . . . n). One exemplarfixed camera 160 is shown in detail, demonstrating that eachfixed camera 160, 150 (a . . . n) has a Camera Arbitrary3D coordinate system 161 and an Image Plane2D coordinate system 162, for the (a . . . n) fixed cameras shown. Also shown, for j=1, 2 are the 130,140, each of which has a Camera Arbitrarystudio cameras 131, 141 and an Image Plane3D coordinate system 132, 142, for each2D coordinate system 130, 140.exemplar device -
FIG. 2 depicts an embodiment of a twodimensional studio camera 200 with dynamic optics. Thezoom 220 and focus 210 controls of thestudio camera 200 are encoded electronically and sent back 230 to a data acquisition system. Thestudio camera 200 may be coupled to the data acquisition directly or indirectly. Thefocus 210 andzoom 220 may be controlled manually or through automation. Although the automation as shown implements a 211, 221, this is not intended to be a limitation on the method or mean for controlling thecog system zoom 220 and focus 210 controls. The data acquisition system describes where along their full range of travel each of the zoom and focus controls are. For example, if the zoom and focus are controlled by a linear stage, the current center spoke 213 and 223 might be transmitted. The data acquisition system may comprise a computer, which has a host module, and various software modules including but not limited to coordinate calculation modules, data acquisition modules, logic processing modules, tracking modules, and a database modules for storing datalocations -
FIG. 3 shows a frame of motion capture data after calibration of themotion capture system 300. Fixed cameras 310 (a . . . n) cooperate with animaging device 320. Retro-reflective markers 340, which are rigidly affixed to thestudio camera 320, are imaged by the fixed cameras 310 (a . . . n). Each of themotion capture devices 310 are calibrated to a common virtual coordinate system, so that it is possible to generate a ray from the feature locations on each of the fixed camera planes 315 (a . . . n) through the volume, and to automatically examine locations where those vectors (nearly) intersect 330, and from that data estimate the 3D locations of those retro-reflective markers. In this case, five retro-reflective markers 340 are attached rigidly to a studio camera. The 3D coordinates of this group of markers can be used to form an orthogonal 3D coordinate system, either by construction or mathematical manipulation. -
FIG. 4 shows a motion capture system in thepre-calibration stage 400. Three or more retro-reflective markers 405 (a . . . n) are attached to a rigidorthogonal device 420 and this device is placed at the desiredlocation 410 of the virtual coordinate system. This produces an initial estimate of the motion capture system's fixed camera's 430(a . . . n) external and internal camera parameters. -
FIG. 5 show a motion capture system in amain calibration stage 500. The studio camera(s) 520, 530 are introduced to the studio, any one of which has a motion capture system detectable rigid coordinate 521, 531 rigidly attached. The studio cameras may operate assystem automated systems 520 or may be manually operated 530, further there may be multiple cameras operated in the same or multiple ways. There may be one or more studio cameras, although two studio cameras are shown in the embodiment, this is not intended to be a limitation on the scope of the present embodiments. Ahuman operator 540 walks around the volume of interest, waving acalibration wand 550 over as much of the volume as practical. Although preferably the operator calibrates the entire volume, nevertheless, embodiments of the present invention contemplate the operator waving a calibration wand over less that the entire volume. In alternate embodiments the human operator may be replaced by an automated operator such as a robot, further, embodiments may include non-human subjects subject as animals or anything capable of motion. The studio camera operator, human or automatic, images the calibration wand, changing zoom, focus, and location over time. Fixed cameras 510 (a . . . n) cooperate to calibrate the system. -
FIG. 6 shows a frame of motion capture system calibration data (wand data) 600. At least two retro-reflective markers located on thecalibration wand 650 are imaged by the motion capture system 610(a . . . n) and by thestudio camera 620. Although one studio camera is shown, this is not intended to be a limitation on the number of studio cameras and more cameras could be implemented in the motion capture calibration system. The motion capture system observes 2D projections (630-634) of the 3D markers in multiple imaging planes. The true 3D distances between the markers are known, and a current estimate of the camera parameters and the 3D locations of the markers are also known. Thestudio camera calibration 622 is similarly collected. An iterative process is used to adjust the estimate of the camera parameters and the estimate of the markers' locations until convergence. -
FIG. 7 shows multiple frames of studiocamera calibration data 700. The motion capture system is capable of producing highly accurate 3D locations for all of thewand data 750. The projection of this 3D object onto thestudio camera 710, having itsown marker system 715, imaging plane produces corresponding 2D data 730-736 over the functional range of the studio camera optics (zoom and focus). The encoded zoom and focus range data provide a first estimate of the studio camera calibration parameters for a given frame of data. A number of frames with the same encoded zoom and focus are automatically examined, and a set of simultaneous equations involving the 3D marker locations and their 2D projections is set up, from which the studio camera calibration parameters at this zoom and focus are calculated. This procedure is repeated at all of the other zoom and focus conditions acquired during calibration. -
FIG. 8 depicts a studio camera inoperation 800. As anoperator 805 moves, zooms, and changes focus on thestudio camera 820, the motion capture system 850 (a . . . n) observes themarkers 834 attached to thecamera 820 and determines the location of the Camera Arbitrary coordinatesystem 831 in the virtual coordinate system. The encoded zoom and focus data generated by thestudio camera 820 is read, and the appropriate current frame camera calibration data is interpolated from the zoom and focus conditions captured during calibration. The fixed cameras 850(a . . . n) produce the virtual coordinatesystem 810, the location of the Camera Arbitrary 3D coordinatesystem 831, and the Camera Image Plane coordinatesystem 832. Thestudio camera 820 produces theimage data 830. Embodiments of the invention combines this information, along with information from the zoom and focus encoders in the studio camera, and the calibration data, to transform theimage data 830 to thefinal image 840 in the virtual coordinatesystem 810. - In operation, embodiments of the present system capture two-dimensional data and transform it into a three dimensional system. In order to effectively incorporate video data into the virtual coordinate system the studio cameras are modeled and the parameters of the camera models are determined. This process may be referred to as calibration. Once the calibration of the studio camera over the functional range of its zoom and focus settings is complete, a motion tracking system is implemented to track the movement of the studio camera while synchronously collecting zoom and focus settings from that camera. The tracking is calibrated so that the principal point of the studio camera's lens is tracked. In alternate embodiments, generator lock or genlock is not employed so that it is not necessary to synchronize the video output of one source, or a specific reference signal from a signal generator, with other pictures or images. Generally such synchronization enables the coincidence of signals in time at a combining or mixing or switching point. The novel embodiments of the present invention allow such without the need for genlock. The process involves at least two steps—Calibrating and Tracking.
- Calibration is the process that provides the offset from object on the camera to the principle point of the camera's lens and also provides the parameters describing the properties of the camera's lens. To calculate offset two objects are tracked—the object on the camera and the calibration wand—in addition, video data of the wand from the camera is collected. The system is then calibrated. In a preferred embodiment, the system is calibrated from about ten frames of data. The calibration process in then repeated at various lens settings. Although ten frames of data are used for calibration in a preferred embodiment, this number of frames is not intended to be a limitation on the scope of the embodiments and a fewer or greater number of frames may be used to complete the calibration.
- The tracking object has two segments. The first segment represents the tracking object. The second segment represents the studio camera's principal point. These segments have six degrees of freedom (6 DOF).
- The data for the calibration includes N frames (nFrames) of data, where N may be any number of frames. Preferably N is between 5 and 15 frames. Each frame has 2 or more wand points in the video and 2 or more wand points from the motion capture (“mocap”) data as well as the position of the tracking object from the motion capture system. The data includes the following:
- Wand Mocap Data:
-
- Wand[iFrame][iMarker][iAxis]
- iFrame=0 to nFrames−1
- iMarker=0 to 1
- iAxis=0.2 (X,Y,Z)
- Wand[iFrame][iMarker][iAxis]
- Wand Video Data:
-
- Centroid [iFrame][iMarker] [iAxis]
- iFrame=0 to nFrames−1
- iMarker=0 to 1
- iAxis=0.1 (U,V)
- Centroid [iFrame][iMarker] [iAxis]
- Object Mocap Data
-
- Object[iFrame][iDOF]
- iFrame=0 to nFrames−1
- iDOF=0to 5
- Object[iFrame][iDOF]
- The calibration data is captured over the functional range of the zoom and focus controls from the studio camera. All of the calibration data is transmitted to a computer processing system. The system receives the data and calculates internal camera calibration parameters, then applies these parameters during video production.
- Each frame of data is transformed into the coordinate system of the tracking object. This gives data as though the camera were stationary very close to the origin. The parameters to be calculated are the 6 degree of freedom offsets to the principal point and the camera lens parameters.
- The lens parameters calculated for the studio camera model include the following:
- Horizontal Principal Point (PPu)
- Vertical Principal Point (PPv)
- Horizontal Scale (Su)
- Vertical Scale (Sw)
- First order distortion (K1)
- These lens parameters are used to convert a camera pixel coordinate (X,Y) of a marker into the corrected slopes (U,V) of the ray from the camera to the marker according to the following:
-
X′=(X−PP u)*S u -
Y′=(Y−PP v)*S v -
Factor=1.0+K 1*(X′ 2 +Y′ 2) -
U corrected =X′*Factor -
U corrected =Y′*Factor - These lens distortion parameters are calculated over the functional range of the studio camera zoom and focus controls. Given an XYZ location in the camera's field of view and a set of lens parameters the expected UV position of the centroid can be calculated.
- There are eleven calibration unknowns: Six (6) Principal Point Offsets and
- Five (5) Lens Parameters
- There are two equations available from each centroid and each frame.
-
dU=U corrected −X i/(−Z i) -
dV=V corrected −Y i/(−Z i) - This is the setup for a non-linear least-squares problem where the left-side errors dU and dV are being minimized. The right side is approximated using the current values of the unknowns and partial derivatives with respect to those unknowns. Corrections to the unknowns are iterated in any standard mathematical way (for example but not limited to Collatz conjecture and juggler sequences). The number of equations is:
-
N Equations=2*2*nFrames - Technically a minimum of three frames of data (12 equations) are needed to calculate the 11 unknowns. However, in practice, data that covers the video camera's field of view is needed so that small residual errors in the lens parameters are not exaggerated via extrapolation into uncalibrated regions of the camera image.
- The solved set of 11 parameters represents a single calibration at a particular zoom and focus setting. The process is repeated with different zoom and focus settings. This multi-step process provides data for interpolating the calibration for any given setting.
- Another stage of the process requires tracking the object attached to the camera, forming the Camera Arbitrary coordinate system. The object preferably has five or more retro-reflective markers, but may have as few as three. The tracking minimizes 2D motion capture camera residuals. When tracking, at any video frame, a calibration setting is used that is an interpolation between the measured calibration settings. The second segment, then, has a dynamic 6 DOF offset from the first segment depending on the interpolated calibration setting.
- Small amounts of noise in the tracking system can give unsatisfying results in the composited video. This is most noticeable when the studio camera is stationary. The “Stationary Lock” feature detects this situation. If the position of the camera hasn't moved from a specified detection distance for a specified minimum duration, then it is considered stationary and the “Locked Position” is sent instead of the tracked position. This algorithm includes a minimum amount of time that the camera is stationary before it is locked so that the system doesn't keep popping in and out of its locked state when the camera is being moved very slowly.
- When the system tracks a moving object, the cameras used in that tracking changes as the object moves across the volume. When a new camera is first able to observe, the very first frames of data are typically sub-par, as these are almost always views of partially obscured markers. The refinement tracks the history of visibility of a marker by a camera, and then weights a new camera contribution according to the length of it's time history. The weight for the new camera contribution ramps up, dynamically, from a small weight to the full weight over some selected period of frames.
- In one embodiment, the camera data is weighted by temporal contiguity. When weighting is completed in this way, the images are presented contiguous or close in time which results in improved imaging.
- This dynamic camera tracker system then provides a “performance animation” or “performance capture” capability that cannot be achieved by conventional means. The system provides the capability to track a performer and stream the data to get a live or real-time, interactive performance. This could be any part of a performer—full body, face and hands or any subset of that. This data is part of the same data stream that contains the camera tracking. It also includes the ability to track a prop.
- As noted previously the forgoing descriptions of the specific embodiments are presented for purposes of illustration and description. They are not intended to be exhaustive or to limit the invention to the precise forms disclosed and obviously many modifications and variations are possible in view of the above teachings. The embodiments were chosen and described in order to explain the principles of the invention and its practical applications, to thereby enable those skilled in the art to best utilize the invention and various embodiments thereof as suited to the particular use contemplated. It is intended that the scope of the invention be defined by the claims and their equivalents.
Claims (7)
1. A method for incorporating two dimensional images captured by at least one moving studio camera with actively controlled optics into a virtual three dimensional coordinate system, said method comprising:
encoding and accessing zoom and focus settings for the at least one studio camera,
identifying a marked calibration ruler,
locating a marked calibration right angle on the calibration ruler,
capturing calibration data over a functional range of the zoom and focus controls,
calculating the internal camera calibration parameters and applying the parameters during video production, and
tracking the moving studio camera.
2. The method of claim 1 , further comprising determining lens distortion parameters for the at least one studio camera over the functional range of the at least one studio camera zoom and focus controls
3. The method of claim 1 , wherein a stationary lock method is applied to improve video stability.
4. The method of claim 1 , in which the camera data is weighted by temporal contiguity.
5. The method of claim 1 , further comprising:
tracking a subject; and
streaming tracking data to produce a real-time, interactive performance.
6. A system for incorporating two dimensional images captured by at least one moving studio camera with actively controlled optics into a virtual three dimensional coordinate system, said system comprising:
a device for encoding and accessing zoom and focus settings for the moving at least one studio camera;
a marked calibration ruler;
a marked calibration right angle;
a host module, said host module adapted to capture calibration data over a functional range of the zoom and focus controls from the moving at least one studio camera data;
a first logic processing module configured to receive and calculate internal camera calibration parameters and apply the parameters during video production, and
a tracking module configured to track the moving camera.
7. The system of claim 1 , further comprising, a second logic processing module configured to calculate lens distortion parameters for the at least one studio camera over the functional range of the studio camera zoom and focus controls.
Priority Applications (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US13/232,938 US20130063558A1 (en) | 2011-09-14 | 2011-09-14 | Systems and Methods for Incorporating Two Dimensional Images Captured by a Moving Studio Camera with Actively Controlled Optics into a Virtual Three Dimensional Coordinate System |
| EP12778835.4A EP2756663A1 (en) | 2011-09-14 | 2012-09-14 | Systems and methods for incorporating two dimensional images captured by a moving studio camera with actively controlled optics into a virtual three dimensional coordinate system |
| PCT/US2012/055655 WO2013040516A1 (en) | 2011-09-14 | 2012-09-14 | Systems and methods for incorporating two dimensional images captured by a moving studio camera with actively controlled optics into a virtual three dimensional coordinate system |
| US14/659,467 US10271036B2 (en) | 2011-09-14 | 2015-03-16 | Systems and methods for incorporating two dimensional images captured by a moving studio camera with actively controlled optics into a virtual three dimensional coordinate system |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US13/232,938 US20130063558A1 (en) | 2011-09-14 | 2011-09-14 | Systems and Methods for Incorporating Two Dimensional Images Captured by a Moving Studio Camera with Actively Controlled Optics into a Virtual Three Dimensional Coordinate System |
Related Child Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/659,467 Continuation US10271036B2 (en) | 2011-09-14 | 2015-03-16 | Systems and methods for incorporating two dimensional images captured by a moving studio camera with actively controlled optics into a virtual three dimensional coordinate system |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| US20130063558A1 true US20130063558A1 (en) | 2013-03-14 |
Family
ID=47080787
Family Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US13/232,938 Abandoned US20130063558A1 (en) | 2011-09-14 | 2011-09-14 | Systems and Methods for Incorporating Two Dimensional Images Captured by a Moving Studio Camera with Actively Controlled Optics into a Virtual Three Dimensional Coordinate System |
| US14/659,467 Expired - Fee Related US10271036B2 (en) | 2011-09-14 | 2015-03-16 | Systems and methods for incorporating two dimensional images captured by a moving studio camera with actively controlled optics into a virtual three dimensional coordinate system |
Family Applications After (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/659,467 Expired - Fee Related US10271036B2 (en) | 2011-09-14 | 2015-03-16 | Systems and methods for incorporating two dimensional images captured by a moving studio camera with actively controlled optics into a virtual three dimensional coordinate system |
Country Status (3)
| Country | Link |
|---|---|
| US (2) | US20130063558A1 (en) |
| EP (1) | EP2756663A1 (en) |
| WO (1) | WO2013040516A1 (en) |
Cited By (37)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20120050543A1 (en) * | 2010-08-30 | 2012-03-01 | Gianluca Colla | Method and apparatus for measuring the focus performance of a camera and lens combination |
| US20130188017A1 (en) * | 2012-01-24 | 2013-07-25 | Chris Chen-Hsing Ma | Instant Calibration of Multi-Sensor 3D Motion Capture System |
| CN103400408A (en) * | 2013-07-29 | 2013-11-20 | 武汉浩翰动画有限公司 | Intelligent production and management system of cartoon game |
| CN103458181A (en) * | 2013-06-29 | 2013-12-18 | 华为技术有限公司 | Lens distortion parameter adjustment method and device and camera shooting device |
| CN104236407A (en) * | 2013-06-14 | 2014-12-24 | Ap系统股份有限公司 | Method for correcting platform scaleplate |
| US8938282B2 (en) | 2011-10-28 | 2015-01-20 | Navigate Surgical Technologies, Inc. | Surgical location monitoring system and method with automatic registration |
| WO2015022339A1 (en) * | 2013-08-13 | 2015-02-19 | Navigate Surgical Technologies, Inc. | System and method for focusing imaging devices |
| US20150288956A1 (en) * | 2014-04-08 | 2015-10-08 | Lucasfilm Entertainment Company, Ltd. | Calibration target for video processing |
| US9198737B2 (en) | 2012-11-08 | 2015-12-01 | Navigate Surgical Technologies, Inc. | System and method for determining the three-dimensional location and orientation of identification markers |
| USD751627S1 (en) | 2014-09-30 | 2016-03-15 | Lucasfilm Entertainment Company Ltd. | Camera calibration tool |
| US9307232B1 (en) | 2014-04-08 | 2016-04-05 | Lucasfilm Entertainment Company Ltd. | Calibration target for video processing |
| US9489738B2 (en) | 2013-04-26 | 2016-11-08 | Navigate Surgical Technologies, Inc. | System and method for tracking non-visible structure of a body with multi-element fiducial |
| US9554763B2 (en) | 2011-10-28 | 2017-01-31 | Navigate Surgical Technologies, Inc. | Soft body automatic registration and surgical monitoring system |
| US9566123B2 (en) | 2011-10-28 | 2017-02-14 | Navigate Surgical Technologies, Inc. | Surgical location monitoring system and method |
| US9585721B2 (en) | 2011-10-28 | 2017-03-07 | Navigate Surgical Technologies, Inc. | System and method for real time tracking and modeling of surgical site |
| WO2017144934A1 (en) | 2016-02-26 | 2017-08-31 | Trophy | Guided surgery apparatus and method |
| US9918657B2 (en) | 2012-11-08 | 2018-03-20 | Navigate Surgical Technologies, Inc. | Method for determining the location and orientation of a fiducial reference |
| US20180182168A1 (en) * | 2015-09-02 | 2018-06-28 | Thomson Licensing | Method, apparatus and system for facilitating navigation in an extended scene |
| WO2018226692A1 (en) * | 2017-06-05 | 2018-12-13 | Track160, Ltd. | Techniques for object tracking |
| JP2019109200A (en) * | 2017-12-20 | 2019-07-04 | オムロン株式会社 | Data generation device for calibration, data generation method for calibration, calibration system, and control program |
| US10404962B2 (en) * | 2015-09-24 | 2019-09-03 | Intel Corporation | Drift correction for camera tracking |
| US20200154004A1 (en) * | 2018-11-13 | 2020-05-14 | Twentieth Century Fox Film Corporation | Systems, methods and apparatuses for film virtual production |
| CN111210472A (en) * | 2019-12-31 | 2020-05-29 | 山东信通电子股份有限公司 | 3D positioning method, device, equipment and medium for video picture |
| US10755432B2 (en) * | 2017-09-27 | 2020-08-25 | Boe Technology Group Co., Ltd. | Indoor positioning system and indoor positioning method |
| CN113240732A (en) * | 2014-05-29 | 2021-08-10 | 株式会社高迎科技 | Optical tracking system and method for calculating posture and position of marker of optical tracking system |
| WO2021218727A1 (en) * | 2020-04-27 | 2021-11-04 | 深圳市瑞立视多媒体科技有限公司 | Method, apparatus, and device for limiting search range of camera, and storage medium |
| CN113767418A (en) * | 2019-02-25 | 2021-12-07 | Mo-Sys工程有限公司 | Lens calibration system |
| WO2022045899A1 (en) * | 2020-08-28 | 2022-03-03 | Weta Digital Limited | Motion capture calibration using cameras and drones |
| US11304777B2 (en) | 2011-10-28 | 2022-04-19 | Navigate Surgical Technologies, Inc | System and method for determining the three-dimensional location and orientation of identification markers |
| CN114972588A (en) * | 2022-05-26 | 2022-08-30 | 安徽江淮汽车集团股份有限公司 | Construction method, construction device and control method of vehicle-mounted virtual animation |
| US11438570B2 (en) * | 2018-05-09 | 2022-09-06 | Sony Corporation | Data processing apparatus and data processing method for generation of calibration data for performing image processing |
| WO2023026089A1 (en) * | 2021-08-26 | 2023-03-02 | Weta Digital Limited | A three-dimensional assembly for motion capture calibration |
| US11600022B2 (en) | 2020-08-28 | 2023-03-07 | Unity Technologies Sf | Motion capture calibration using drones |
| US11636621B2 (en) | 2020-08-28 | 2023-04-25 | Unity Technologies Sf | Motion capture calibration using cameras and drones |
| US12070365B2 (en) | 2012-03-28 | 2024-08-27 | Navigate Surgical Technologies, Inc | System and method for determining the three-dimensional location and orientation of identification markers |
| EP3692501B1 (en) * | 2017-10-08 | 2025-08-20 | Magik Eye Inc. | Calibrating a sensor system including multiple movable sensors |
| US12514678B2 (en) | 2013-03-18 | 2026-01-06 | Navigate Surgical Technologies, Inc. | System and method for determining the three-dimensional location and orientation of identification markers |
Families Citing this family (39)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10632740B2 (en) | 2010-04-23 | 2020-04-28 | Landa Corporation Ltd. | Digital printing process |
| US9902147B2 (en) | 2012-03-05 | 2018-02-27 | Landa Corporation Ltd. | Digital printing system |
| MX381618B (en) | 2012-03-05 | 2025-03-12 | Landa Corp Ltd | INK FILM STRUCTURES |
| US10434761B2 (en) | 2012-03-05 | 2019-10-08 | Landa Corporation Ltd. | Digital printing process |
| US10569534B2 (en) | 2012-03-05 | 2020-02-25 | Landa Corporation Ltd. | Digital printing system |
| US9643403B2 (en) | 2012-03-05 | 2017-05-09 | Landa Corporation Ltd. | Printing system |
| US10642198B2 (en) | 2012-03-05 | 2020-05-05 | Landa Corporation Ltd. | Intermediate transfer members for use with indirect printing systems and protonatable intermediate transfer members for use with indirect printing systems |
| CN104271356B (en) | 2012-03-05 | 2016-10-19 | 兰达公司 | Digital printing process |
| US9498946B2 (en) | 2012-03-05 | 2016-11-22 | Landa Corporation Ltd. | Apparatus and method for control or monitoring of a printing system |
| US9517618B2 (en) | 2012-03-15 | 2016-12-13 | Landa Corporation Ltd. | Endless flexible belt for a printing system |
| GB201401173D0 (en) | 2013-09-11 | 2014-03-12 | Landa Corp Ltd | Ink formulations and film constructions thereof |
| GB2536489B (en) | 2015-03-20 | 2018-08-29 | Landa Corporation Ltd | Indirect printing system |
| GB2537813A (en) | 2015-04-14 | 2016-11-02 | Landa Corp Ltd | Apparatus for threading an intermediate transfer member of a printing system |
| CN105072314A (en) * | 2015-08-13 | 2015-11-18 | 黄喜荣 | Virtual studio implementation method capable of automatically tracking objects |
| CN105303580A (en) * | 2015-11-16 | 2016-02-03 | 张璐 | Identification system and method of panoramic looking-around multi-camera calibration rod |
| GB201602877D0 (en) | 2016-02-18 | 2016-04-06 | Landa Corp Ltd | System and method for generating videos |
| JP7144328B2 (en) | 2016-05-30 | 2022-09-29 | ランダ コーポレイション リミテッド | digital printing process |
| GB201609463D0 (en) | 2016-05-30 | 2016-07-13 | Landa Labs 2012 Ltd | Method of manufacturing a multi-layer article |
| KR101765784B1 (en) * | 2016-07-04 | 2017-08-08 | 주식회사 그래픽코드 | Virtual Camera Tracking System |
| US10926532B2 (en) | 2017-10-19 | 2021-02-23 | Landa Corporation Ltd. | Endless flexible belt for a printing system |
| JP7225230B2 (en) | 2017-11-19 | 2023-02-20 | ランダ コーポレイション リミテッド | digital printing system |
| WO2019102297A1 (en) | 2017-11-27 | 2019-05-31 | Landa Corporation Ltd. | Digital printing system |
| US11707943B2 (en) | 2017-12-06 | 2023-07-25 | Landa Corporation Ltd. | Method and apparatus for digital printing |
| WO2019111223A1 (en) | 2017-12-07 | 2019-06-13 | Landa Corporation Ltd. | Digital printing process and method |
| CN110634149B (en) * | 2018-06-22 | 2022-03-18 | 湖南大学 | Non-rigid target characteristic point matching method for optical motion capture system |
| CN112399918B (en) | 2018-06-26 | 2023-01-31 | 兰达公司 | Intermediate Transmission Components of Digital Printing System |
| US10994528B1 (en) | 2018-08-02 | 2021-05-04 | Landa Corporation Ltd. | Digital printing system with flexible intermediate transfer member |
| US12001902B2 (en) | 2018-08-13 | 2024-06-04 | Landa Corporation Ltd. | Correcting distortions in digital printing by implanting dummy pixels in a digital image |
| JP7246496B2 (en) | 2018-10-08 | 2023-03-27 | ランダ コーポレイション リミテッド | Friction reduction means for printing systems and methods |
| US11787170B2 (en) | 2018-12-24 | 2023-10-17 | Landa Corporation Ltd. | Digital printing system |
| CN113692354B (en) | 2019-03-31 | 2024-03-26 | 兰达公司 | System and method for preventing or minimizing printing defects during printing |
| WO2021068070A1 (en) | 2019-10-08 | 2021-04-15 | Cast Group Of Companies Inc. | Electronic tracking device for camera and related system for controlling image output of the camera |
| CN114746813A (en) | 2019-11-25 | 2022-07-12 | 兰达公司 | Drying inks using infrared radiation in digital printing |
| US11321028B2 (en) | 2019-12-11 | 2022-05-03 | Landa Corporation Ltd. | Correcting registration errors in digital printing |
| JP7657229B2 (en) | 2019-12-29 | 2025-04-04 | ランダ コーポレイション リミテッド | Printing method and system |
| CN111970535B (en) | 2020-09-25 | 2021-08-31 | 魔珐(上海)信息科技有限公司 | Virtual live broadcast method, device, system and storage medium |
| EP4264377A4 (en) | 2021-02-02 | 2024-11-13 | Landa Corporation Ltd. | REDUCING DISTORTIONS IN PRINTED IMAGES |
| US12205328B2 (en) * | 2021-07-28 | 2025-01-21 | Htc Corporation | System for tracking camera and control method thereof |
| CN117880551B (en) * | 2024-03-12 | 2024-10-22 | 深圳市云实互联网信息科技有限公司 | Virtual live broadcast processing method, device, equipment and storage medium |
Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6522787B1 (en) * | 1995-07-10 | 2003-02-18 | Sarnoff Corporation | Method and system for rendering and combining images to form a synthesized view of a scene containing image information from a second image |
| US20060165310A1 (en) * | 2004-10-27 | 2006-07-27 | Mack Newton E | Method and apparatus for a virtual scene previewing system |
| US20110025853A1 (en) * | 2009-07-31 | 2011-02-03 | Naturalpoint, Inc. | Automated collective camera calibration for motion capture |
Family Cites Families (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5889550A (en) | 1996-06-10 | 1999-03-30 | Adaptive Optics Associates, Inc. | Camera tracking system |
| US20070076096A1 (en) | 2005-10-04 | 2007-04-05 | Alexander Eugene J | System and method for calibrating a set of imaging devices and calculating 3D coordinates of detected features in a laboratory coordinate system |
| WO2007058900A2 (en) | 2005-11-10 | 2007-05-24 | Alexander Eugene J | Calibrating an imaging device for generating three dimensional suface models of moving objects |
-
2011
- 2011-09-14 US US13/232,938 patent/US20130063558A1/en not_active Abandoned
-
2012
- 2012-09-14 EP EP12778835.4A patent/EP2756663A1/en not_active Withdrawn
- 2012-09-14 WO PCT/US2012/055655 patent/WO2013040516A1/en not_active Ceased
-
2015
- 2015-03-16 US US14/659,467 patent/US10271036B2/en not_active Expired - Fee Related
Patent Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6522787B1 (en) * | 1995-07-10 | 2003-02-18 | Sarnoff Corporation | Method and system for rendering and combining images to form a synthesized view of a scene containing image information from a second image |
| US20060165310A1 (en) * | 2004-10-27 | 2006-07-27 | Mack Newton E | Method and apparatus for a virtual scene previewing system |
| US20110025853A1 (en) * | 2009-07-31 | 2011-02-03 | Naturalpoint, Inc. | Automated collective camera calibration for motion capture |
Cited By (51)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US8581986B2 (en) * | 2010-08-30 | 2013-11-12 | Datacolor Holding Ag | Method and apparatus for measuring the focus performance of a camera and lens combination |
| US20120050543A1 (en) * | 2010-08-30 | 2012-03-01 | Gianluca Colla | Method and apparatus for measuring the focus performance of a camera and lens combination |
| US9585721B2 (en) | 2011-10-28 | 2017-03-07 | Navigate Surgical Technologies, Inc. | System and method for real time tracking and modeling of surgical site |
| US9452024B2 (en) | 2011-10-28 | 2016-09-27 | Navigate Surgical Technologies, Inc. | Surgical location monitoring system and method |
| US11304777B2 (en) | 2011-10-28 | 2022-04-19 | Navigate Surgical Technologies, Inc | System and method for determining the three-dimensional location and orientation of identification markers |
| US8938282B2 (en) | 2011-10-28 | 2015-01-20 | Navigate Surgical Technologies, Inc. | Surgical location monitoring system and method with automatic registration |
| US9566123B2 (en) | 2011-10-28 | 2017-02-14 | Navigate Surgical Technologies, Inc. | Surgical location monitoring system and method |
| US9554763B2 (en) | 2011-10-28 | 2017-01-31 | Navigate Surgical Technologies, Inc. | Soft body automatic registration and surgical monitoring system |
| US20130188017A1 (en) * | 2012-01-24 | 2013-07-25 | Chris Chen-Hsing Ma | Instant Calibration of Multi-Sensor 3D Motion Capture System |
| US12070365B2 (en) | 2012-03-28 | 2024-08-27 | Navigate Surgical Technologies, Inc | System and method for determining the three-dimensional location and orientation of identification markers |
| US9918657B2 (en) | 2012-11-08 | 2018-03-20 | Navigate Surgical Technologies, Inc. | Method for determining the location and orientation of a fiducial reference |
| US9198737B2 (en) | 2012-11-08 | 2015-12-01 | Navigate Surgical Technologies, Inc. | System and method for determining the three-dimensional location and orientation of identification markers |
| US12514678B2 (en) | 2013-03-18 | 2026-01-06 | Navigate Surgical Technologies, Inc. | System and method for determining the three-dimensional location and orientation of identification markers |
| US9489738B2 (en) | 2013-04-26 | 2016-11-08 | Navigate Surgical Technologies, Inc. | System and method for tracking non-visible structure of a body with multi-element fiducial |
| US9844413B2 (en) | 2013-04-26 | 2017-12-19 | Navigate Surgical Technologies, Inc. | System and method for tracking non-visible structure of a body with multi-element fiducial |
| CN104236407B (en) * | 2013-06-14 | 2017-10-10 | Ap系统股份有限公司 | The method of dressing plate scale |
| CN104236407A (en) * | 2013-06-14 | 2014-12-24 | Ap系统股份有限公司 | Method for correcting platform scaleplate |
| CN103458181A (en) * | 2013-06-29 | 2013-12-18 | 华为技术有限公司 | Lens distortion parameter adjustment method and device and camera shooting device |
| CN103400408A (en) * | 2013-07-29 | 2013-11-20 | 武汉浩翰动画有限公司 | Intelligent production and management system of cartoon game |
| US9456122B2 (en) | 2013-08-13 | 2016-09-27 | Navigate Surgical Technologies, Inc. | System and method for focusing imaging devices |
| WO2015022339A1 (en) * | 2013-08-13 | 2015-02-19 | Navigate Surgical Technologies, Inc. | System and method for focusing imaging devices |
| US9307231B2 (en) * | 2014-04-08 | 2016-04-05 | Lucasfilm Entertainment Company Ltd. | Calibration target for video processing |
| US9641830B2 (en) | 2014-04-08 | 2017-05-02 | Lucasfilm Entertainment Company Ltd. | Automated camera calibration methods and systems |
| US9307232B1 (en) | 2014-04-08 | 2016-04-05 | Lucasfilm Entertainment Company Ltd. | Calibration target for video processing |
| US20150288956A1 (en) * | 2014-04-08 | 2015-10-08 | Lucasfilm Entertainment Company, Ltd. | Calibration target for video processing |
| CN113240732A (en) * | 2014-05-29 | 2021-08-10 | 株式会社高迎科技 | Optical tracking system and method for calculating posture and position of marker of optical tracking system |
| USD751627S1 (en) | 2014-09-30 | 2016-03-15 | Lucasfilm Entertainment Company Ltd. | Camera calibration tool |
| US20180182168A1 (en) * | 2015-09-02 | 2018-06-28 | Thomson Licensing | Method, apparatus and system for facilitating navigation in an extended scene |
| US12293470B2 (en) | 2015-09-02 | 2025-05-06 | Interdigital Ce Patent Holdings, Sas | Method, apparatus and system for facilitating navigation in an extended scene |
| US11699266B2 (en) * | 2015-09-02 | 2023-07-11 | Interdigital Ce Patent Holdings, Sas | Method, apparatus and system for facilitating navigation in an extended scene |
| US10404962B2 (en) * | 2015-09-24 | 2019-09-03 | Intel Corporation | Drift correction for camera tracking |
| WO2017144934A1 (en) | 2016-02-26 | 2017-08-31 | Trophy | Guided surgery apparatus and method |
| US11348255B2 (en) | 2017-06-05 | 2022-05-31 | Track160, Ltd. | Techniques for object tracking |
| WO2018226692A1 (en) * | 2017-06-05 | 2018-12-13 | Track160, Ltd. | Techniques for object tracking |
| US10755432B2 (en) * | 2017-09-27 | 2020-08-25 | Boe Technology Group Co., Ltd. | Indoor positioning system and indoor positioning method |
| EP3692501B1 (en) * | 2017-10-08 | 2025-08-20 | Magik Eye Inc. | Calibrating a sensor system including multiple movable sensors |
| JP2019109200A (en) * | 2017-12-20 | 2019-07-04 | オムロン株式会社 | Data generation device for calibration, data generation method for calibration, calibration system, and control program |
| US11438570B2 (en) * | 2018-05-09 | 2022-09-06 | Sony Corporation | Data processing apparatus and data processing method for generation of calibration data for performing image processing |
| US10819886B2 (en) | 2018-11-13 | 2020-10-27 | Twentieth Century Fox Film Corporation | Systems, methods and apparatuses for film virtual production |
| US20200154004A1 (en) * | 2018-11-13 | 2020-05-14 | Twentieth Century Fox Film Corporation | Systems, methods and apparatuses for film virtual production |
| US10728429B2 (en) * | 2018-11-13 | 2020-07-28 | Twentieth Century Fox Film Corporation | Systems, methods and apparatuses for film virtual production |
| CN113767418A (en) * | 2019-02-25 | 2021-12-07 | Mo-Sys工程有限公司 | Lens calibration system |
| CN111210472A (en) * | 2019-12-31 | 2020-05-29 | 山东信通电子股份有限公司 | 3D positioning method, device, equipment and medium for video picture |
| WO2021218727A1 (en) * | 2020-04-27 | 2021-11-04 | 深圳市瑞立视多媒体科技有限公司 | Method, apparatus, and device for limiting search range of camera, and storage medium |
| WO2022045898A1 (en) * | 2020-08-28 | 2022-03-03 | Weta Digital Limited | Motion capture calibration using drones |
| US11636621B2 (en) | 2020-08-28 | 2023-04-25 | Unity Technologies Sf | Motion capture calibration using cameras and drones |
| US11600022B2 (en) | 2020-08-28 | 2023-03-07 | Unity Technologies Sf | Motion capture calibration using drones |
| WO2022045897A1 (en) * | 2020-08-28 | 2022-03-03 | Weta Digital Limited | Motion capture calibration using drones with multiple cameras |
| WO2022045899A1 (en) * | 2020-08-28 | 2022-03-03 | Weta Digital Limited | Motion capture calibration using cameras and drones |
| WO2023026089A1 (en) * | 2021-08-26 | 2023-03-02 | Weta Digital Limited | A three-dimensional assembly for motion capture calibration |
| CN114972588A (en) * | 2022-05-26 | 2022-08-30 | 安徽江淮汽车集团股份有限公司 | Construction method, construction device and control method of vehicle-mounted virtual animation |
Also Published As
| Publication number | Publication date |
|---|---|
| US10271036B2 (en) | 2019-04-23 |
| EP2756663A1 (en) | 2014-07-23 |
| US20150195509A1 (en) | 2015-07-09 |
| WO2013040516A1 (en) | 2013-03-21 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US10271036B2 (en) | Systems and methods for incorporating two dimensional images captured by a moving studio camera with actively controlled optics into a virtual three dimensional coordinate system | |
| KR100799088B1 (en) | High Speed Digital Pan Tilt Zoom Video | |
| Senior et al. | Acquiring multi-scale images by pan-tilt-zoom control and automatic multi-camera calibration | |
| CN101563709B (en) | Calibrating a camera system | |
| WO2016074639A1 (en) | Methods and systems for multi-view high-speed motion capture | |
| JP5725953B2 (en) | Imaging apparatus, control method therefor, and information processing apparatus | |
| JP2001094857A (en) | Method for controlling virtual camera, camera array and method for aligning camera array | |
| WO2018235163A1 (en) | Calibration device, calibration chart, chart pattern generation device, and calibration method | |
| JP2014529727A (en) | Automatic scene calibration | |
| US8223208B2 (en) | Device and method for calibrating an imaging device for generating three dimensional surface models of moving objects | |
| Hedborg et al. | Structure and motion estimation from rolling shutter video | |
| WO2003036565A2 (en) | System and method for obtaining video of multiple moving fixation points within a dynamic scene | |
| JP7657308B2 (en) | Method, apparatus and system for generating a three-dimensional model of a scene - Patents.com | |
| WO2020235110A1 (en) | Calibration device, chart for calibration, and calibration method | |
| GB2459033A (en) | Resolution reduced video monitoring method to enable real time video processing | |
| CN103729839B (en) | A kind of method and system of sensor-based outdoor camera tracking | |
| JP2003179800A (en) | Multi-view image generation apparatus, image processing apparatus and method, and computer program | |
| WO2007041690A2 (en) | Device for generating three dimensional surface models of moving objects | |
| BR112021008558A2 (en) | apparatus, disparity estimation method, and computer program product | |
| KR101745493B1 (en) | Apparatus and method for depth map generation | |
| CN112422848A (en) | Video splicing method based on depth map and color map | |
| JP2011146762A (en) | Solid model generator | |
| JP7040511B2 (en) | Information processing equipment and methods | |
| JP5925109B2 (en) | Image processing apparatus, control method thereof, and control program | |
| Beriault et al. | Flexible multi-camera network calibration for human gesture monitoring |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |