US20220132042A1 - Method for tracking movable object, tracking device, and method for controlling shooting parameters of camera - Google Patents
Method for tracking movable object, tracking device, and method for controlling shooting parameters of camera Download PDFInfo
- Publication number
- US20220132042A1 US20220132042A1 US17/506,696 US202117506696A US2022132042A1 US 20220132042 A1 US20220132042 A1 US 20220132042A1 US 202117506696 A US202117506696 A US 202117506696A US 2022132042 A1 US2022132042 A1 US 2022132042A1
- Authority
- US
- United States
- Prior art keywords
- time
- duration
- light emitting
- movable object
- emitting elements
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H04N5/23299—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/695—Control of camera direction for changing a field of view, e.g. pan, tilt or based on tracking of objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/56—Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/72—Combination of two or more compensation controls
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/73—Circuitry for compensating brightness variation in the scene by influencing the exposure time
-
- H04N5/2256—
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0138—Head-up displays characterised by optical features comprising image capture systems, e.g. camera
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B2027/0192—Supplementary details
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
Definitions
- the disclosure generally relates to tracking mechanism, in particular, to a method for tracking a movable object, a tracking device and a method for dynamically controlling shooting parameters of a camera.
- the mechanisms used for tracking objects in systems such as augmented reality (AR) system or virtual reality (VR) system generally include inside-out tracking and outside-in tracking.
- the inside-out tracking mechanism the to-be-tracked object may be disposed with light emitting elements for the camera on the head-mounted display to capture, and the pose of the to-be-tracked object may be accordingly determined.
- the shooting parameters of the camera are not properly designed, some issues may occur. For example, if the on durations of the camera and the light emitting elements are not properly designed, either unnecessary power consumption may be introduced, or the pose of the to-be-tracked object cannot be accurately determined. For another example, if the shooting parameters use for performing environment detection and object tracking are the same, either the pose of the to-be-tracked object cannot be accurately determined, or the environment cannot be accurately detected.
- the present invention is directed to a method for tracking a movable object, a tracking device and a method for dynamically controlling shooting parameters of a camera, which may be used to solve the above technical problem.
- the embodiments of the disclosure provide a method for tracking a movable object, adapted to a tracking device.
- the method includes: determining a first on duration of a camera of the tracking device, wherein the first on duration comprises a starting time and an ending time; determining a second on duration of a plurality of light emitting elements disposed on the movable object by adding a first guard time before the starting time and adding a second guard time after the ending time; turning on the light emitting elements based on the second on duration; and controlling the camera to capture a specific image of the light emitting elements in the first on duration and accordingly tracking the movable object.
- the embodiments of the disclosure provide a tracking device including a camera and a processor.
- the processor is coupled to the camera and configured to perform: determining a first on duration of the camera of the tracking device, wherein the first on duration comprises a starting time and an ending time; determining a second on duration of a plurality of light emitting elements disposed on the movable object by adding a first guard time before the starting time and adding a second guard time after the ending time; turning on the light emitting elements based on the second on duration; and controlling the camera to capture a specific image of the light emitting elements in the first on duration and accordingly tracking the movable object.
- the embodiments of the disclosure provide a method for dynamically controlling shooting parameters of a camera.
- the method includes: determining a plurality of time frames, wherein the time frames comprise a plurality of first time frames and a plurality of second time frames; controlling the camera of the tracking device to shoot a first image with a first exposure parameter in each of the first time frames and accordingly performing an environment detection;
- controlling the camera of the tracking device to shoot a second image with a second exposure parameter in each of the second time frames and accordingly performing a first specific detection, wherein the second exposure parameter is lower than the first exposure parameter.
- FIG. 1 shows a functional diagram of a tracking device and a movable object according to an embodiment of the disclosure.
- FIG. 2A shows a first combination of the first on duration and the second on duration according to an embodiment of the disclosure.
- FIG. 2B shows a second combination of the first on duration and the second on duration according to an embodiment of the disclosure.
- FIG. 3 shows a flow chart of the method for tracking the movable object according to an embodiment of the disclosure.
- FIG. 4 shows a third combination of the first on duration and the second on duration according to an embodiment of the disclosure.
- FIG. 5 shows a flow chart of the method for dynamically controlling shooting parameters of a camera according to an embodiment of the disclosure.
- FIG. 6 shows the time frames according to a first embodiment of the disclosure.
- FIG. 7 shows the time frames according to the second embodiment of the disclosure.
- the tracking device 100 may be any device capable of tracking the movable object 199 by using any known tracking mechanisms.
- the tracking device 100 may be a head-mounted display (HMD) used in an AR system or a VR system, and the movable object 199 may be a component of the AR/VR system that needs to be tracked, such as a handheld controller, but the disclosure is not limited thereto.
- HMD head-mounted display
- the tracking device 100 includes a camera 102 and a processor 104 .
- the camera 102 could be any cameras having charge coupled device (CCD) lens, complementary metal oxide semiconductor transistors (CMOS) lens, but the disclosure is not limited thereto.
- CCD charge coupled device
- CMOS complementary metal oxide semiconductor transistors
- the processor 104 is coupled to the camera 102 , and may be, for example, a general-purpose processor, a special purpose processor, a conventional processor, a digital signal processor (DSP), a plurality of microprocessors, one or more microprocessors in association with a DSP core, a controller, a microcontroller, Application Specific Integrated Circuits (ASICs), Field Programmable Gate Array (FPGAs) circuits, any other type of integrated circuit (IC), a state machine, and the like.
- DSP digital signal processor
- ASICs Application Specific Integrated Circuits
- FPGAs Field Programmable Gate Array
- the processor 104 may control the camera 102 to capture images of the movable object 199 and determine the pose of the movable object 199 accordingly.
- the movable object 199 may be disposed with a plurality of light emitting elements (such as regular light emitting diodes (LED) and/or infrared LEDs).
- the processor 104 may perform inside-out tracking based on the light distributions of the light emitting elements in the images captured by the camera 102 .
- the camera 102 may be configured to capture images in the corresponding on durations (referred to as first on durations). That is, the camera 102 would not capture/shoot in the durations other than the first on durations.
- the light emitting elements on the movable object 199 have their own on durations (referred to as second on durations). In this case, the light emitting elements would not emit light in the durations other than the second on durations.
- the first on durations and the second on durations need to (partially) overlap with each other for the camera 102 to capture the lights of the light emitting elements.
- the first on durations and the second on durations are not properly designed, some undesired effects may be introduced.
- FIG. 2A shows a first combination of the first on duration and the second on duration according to an embodiment of the disclosure.
- the first on duration 211 is longer than the second on duration 212 .
- the long exposure resulted from the first on duration 211 would brighten the background in the captured images, such that the objects needless to be tracked would also be captured in the images. Accordingly, the performance of tracking the movable object 199 would be degraded.
- the exposure duration of the camera 102 is usually very short, it is difficult to precisely control the light emitting elements to emit lights in the second on duration 212 .
- FIG. 2B shows a second combination of the first on duration and the second on duration according to an embodiment of the disclosure.
- the light emitting elements on the movable object 199 are designed to be always on.
- the first on duration 221 may be designed to be short, which prevents the background in the captured images from being unnecessarily brightened.
- the power consumption of the light emitting elements would be high.
- the embodiments of the disclosure provide a method for tracking the movable object, and detailed discussion would be provided in the following.
- FIG. 3 shows a flow chart of the method for tracking the movable object according to an embodiment of the disclosure.
- the method of this embodiment may be executed by the tracking device 100 in FIG. 1 , and the details of each step in FIG. 3 will be described below with the components shown in FIG. 1 .
- FIG. 4 would be used as an example, wherein FIG. 4 shows a third combination of the first on duration and the second on duration according to an embodiment of the disclosure.
- step S 310 the processor 104 determines a first on duration 411 of the camera 102 of the tracking device 100 .
- the processor 104 may determine the first on duration 411 to a length that would not be too long to overly brighten the background in the captured images, e.g., 16 ⁇ s, but the disclosure is not limited thereto.
- the first on duration 411 has a starting time T 1 and an ending time T 2 .
- step S 320 the processor 104 determines a second on duration 412 of the light emitting elements disposed on the movable object 199 by adding a first guard time G 1 before the starting time T 1 and adding a second guard time G 2 after the ending time T 2 .
- the lengths of the first guard time G 1 and the second guard time G 2 may be arbitrarily chosen based on the requirements of the designer.
- step S 330 the processor 104 turns on the light emitting elements based on the second on duration 412 .
- the processor 104 can directly or indirectly control the light emitting elements to emit lights in the second on duration 412 .
- the processor 104 may control the movable object 199 to turn on the light emitting elements in the second on duration 412 for emitting lights during the second on duration 412 .
- the processor 104 may synchronize with the movable object 199 and/or the light emitting elements to notify that the light emitting elements should be turned on in the second on duration 412 .
- step S 340 the processor 104 controls the camera 102 to capture a specific image of the light emitting elements in the first on duration 411 and accordingly tracking the movable object 199 .
- the processor 104 may determine the pose of the movable object 199 based on the light distribution of the light emitting elements in the specific image by using the inside-out tracking mechanism, but the disclosure is not limited thereto.
- the second on duration 412 is longer than the first on duration 411 , it would be easier to control the light emitting elements to emit light in the second on duration 412 .
- the power consumption of FIG. 4 would be much lower, which improves the battery life of the movable object 199 and/or the light emitting elements.
- the disclosure further proposes a method for dynamically controlling shooting parameters of a camera, which may be used to solve this problem.
- FIG. 5 shows a flow chart of the method for dynamically controlling shooting parameters of a camera according to an embodiment of the disclosure.
- the method of this embodiment may be executed by the tracking device 100 in FIG. 1 , and the details of each step in FIG. 5 will be described below with the components shown in FIG. 1 .
- step S 510 the processor 104 determines a plurality of time frames, wherein the time frames comprise a plurality of first time frames and a plurality of second time frames.
- step S 520 the processor 104 controls the camera 102 of the tracking device 100 to shoot a first image with a first exposure parameter in each of the first time frames and accordingly performing an environment detection.
- step S 530 the processor 104 controls the camera 102 of the tracking device 100 to shoot a second image with a second exposure parameter in each of the second time frames and accordingly performing a first specific detection, wherein the second exposure parameter is lower than the first exposure parameter.
- FIG. 5 For better understanding the concept of FIG. 5 , FIG. 6 and FIG. 7 would be used as example for discussions.
- the processor 104 may divide the time axis into a plurality of time frames, which are shown as rectangle boxes.
- the time frames may be divided into a plurality of first time frames F 11 -F 1 N and a plurality of second time frames F 21 -F 2 N.
- the first time frames F 11 -F 1 N and the second time frames F 21 -F 2 N are interleaved with each other. That is, a (2i+1)-th time frame of the time frames belongs to the first time frames F 11 -F 1 N, and a (2i+2)-th time frame of the time frames belongs to the second time frames F 21 -F 2 N, wherein i is an index.
- the (2i+1)-th time frame of the time frames may be the 1-st time frame, which would be the first time frame F 11 .
- the (2i+2)-th time frame of the time frames may be the 2-nd time frame, which would be the second time frame F 21 .
- the (2i+1)-th time frame of the time frames may be the 3-rd time frame, which would be the first time frame F 12 .
- the (2i+2)-th time frame of the time frames may be the 4-th time frame, which would be the second time frame F 22 .
- the processor 104 may control the camera 102 to shoot a first image with the first exposure parameter in each of the first time frames F 11 -F 1 N and accordingly performing an environment detection.
- the processor 104 may perform the environment detection based on a simultaneous localization and mapping (SLAM) mechanism, but the disclosure is not limited thereto.
- SLAM simultaneous localization and mapping
- the first specific detection may be used for tracking the movable object 199 disposed with the light emitting elements as taught in the above.
- the processor 104 since the processor 104 needs more feature points of the environment to better detect the environment, the processor 104 uses a higher first exposure parameter (e.g., higher exposure value and/or longer exposure time) to guarantee that the environment could be properly brightened in the first image. Accordingly, a better performance of the environment detection may be achieved in each of the first time frames F 11 -F 1 N.
- a higher first exposure parameter e.g., higher exposure value and/or longer exposure time
- the processor 104 can use a lower second exposure parameter (e.g., lower exposure value and/or shorter exposure time) to dim the lights from the environment in the second image, such that only the lights from the light emitting elements would be preserved in the second image. Accordingly, a better performance of the first specific detection may be achieved in each of the second time frames F 21 -F 2 N.
- a lower second exposure parameter e.g., lower exposure value and/or shorter exposure time
- the time frames may further include a plurality of third time frames
- the processor 104 may control the camera 102 to shoot a third image with a third exposure parameter in each of the third time frames and accordingly performing a second specific detection, wherein the third exposure parameter is higher than the second exposure parameter, and can be equal or not equal to the first exposure parameter.
- the first specific detection is used for tracking the movable object 199
- the second specific detection is used for the gesture detection or hand tracking.
- the processor 104 may divide the timeline into a plurality of time frames, which are shown as rectangle boxes.
- the time frames may be divided into a plurality of first time frames F 11 -F 1 N, a plurality of second time frames F 21 -F 2 N, and a plurality of third time frames F 31 -F 3 N, wherein the first time frames F 11 -F 1 N, the second time frames F 21 -F 2 N, and the third time frames F 31 -F 3 N are interleaved with each other.
- a (3i+1)-th time frame of the time frames belongs to the first time frames F 11 -F 1 N
- a (3i+2)-th time frame of the time frames belongs to the second time frames F 21 -F 2 N
- a (3i+3)-th time frame of the time frames belongs to the third time frames F 31 -F 3 N, wherein i is an index.
- the (3i+1)-th time frame of the time frames may be the 1-st time frame, which would be the first time frame F 11 .
- the (3i+2)-th time frame of the time frames may be the 2-nd time frame, which would be the second time frame F 21 .
- the (3i+3)-th time frame of the time frames may be the 3-rd time frame, which would be the third time frame F 31 .
- the (3i+1)-th time frame of the time frames may be the 4-th time frame, which would be the first time frame F 12 .
- the (3i+2)-th time frame of the time frames may be the 5-th time frame, which would be the second time frame F 22 .
- the (3i+3)-th time frame of the time frames may be the 6-th time frame, which would be the third time frame F 32 .
- the processor 104 may control the camera 102 to shoot a first image with the first exposure parameter in each of the first time frames F 11 -F 1 N and accordingly performing an environment detection. In one embodiment, the processor 104 may perform the environment detection based on the SLAM mechanism, but the disclosure is not limited thereto.
- the first specific detection may be used for tracking the movable object 199 disposed with the light emitting elements as taught in the above.
- the processor 104 since the processor 104 needs more feature points of the environment to better detect the environment, the processor 104 uses a higher first exposure parameter to guarantee that the environment could be properly brightened in the first image. Accordingly, a better performance of the environment detection may be achieved in each of the first time frames F 11 -F 1 N.
- the processor 104 can use a lower second exposure parameter to dim the lights from the environment in the second image, such that only the lights from the light emitting elements would be preserved in the second/third image. Accordingly, a better performance of the first specific detection may be achieved in each of the second time frames F 21 -F 2 N.
- the embodiments of the disclosure can add guard times before/after the first on duration of the camera of the tracking device to determine the second on duration of the light emitting elements, such that the light emitting elements can be easier controlled to emit light in the second on duration while reducing the power consumption.
- the embodiments of the disclosure can use a higher first exposure parameter (e.g., higher exposure value and/or exposure time) to retrieve more feature points of the environment in the first image when performing environment detection in each of the first time frames. Accordingly, a better performance of the environment detection may be achieved in each of the first time frames.
- a higher first exposure parameter e.g., higher exposure value and/or exposure time
- the embodiments of the disclosure can use a lower second exposure parameter (e.g., lower exposure value and/or exposure time) to make sure that only the lights from the light emitting elements would be preserved in the second image. Accordingly, a better performance of the first specific detection may be achieved in each of the second time frames.
- a lower second exposure parameter e.g., lower exposure value and/or exposure time
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Studio Devices (AREA)
- Closed-Circuit Television Systems (AREA)
- Automatic Focus Adjustment (AREA)
Abstract
Description
- This application claims the priority benefit of U.S. Provisional Application No. 63/105,366, filed on Oct. 26, 2020, and U.S. Provisional Application No. 63/145,497, filed on Feb. 4, 2021. The entirety of each of the above-mentioned patent applications is hereby incorporated by reference herein and made a part of this specification.
- The disclosure generally relates to tracking mechanism, in particular, to a method for tracking a movable object, a tracking device and a method for dynamically controlling shooting parameters of a camera.
- Nowadays, the mechanisms used for tracking objects in systems such as augmented reality (AR) system or virtual reality (VR) system generally include inside-out tracking and outside-in tracking. In the inside-out tracking mechanism, the to-be-tracked object may be disposed with light emitting elements for the camera on the head-mounted display to capture, and the pose of the to-be-tracked object may be accordingly determined.
- In this case, if the shooting parameters of the camera are not properly designed, some issues may occur. For example, if the on durations of the camera and the light emitting elements are not properly designed, either unnecessary power consumption may be introduced, or the pose of the to-be-tracked object cannot be accurately determined. For another example, if the shooting parameters use for performing environment detection and object tracking are the same, either the pose of the to-be-tracked object cannot be accurately determined, or the environment cannot be accurately detected.
- Accordingly, the present invention is directed to a method for tracking a movable object, a tracking device and a method for dynamically controlling shooting parameters of a camera, which may be used to solve the above technical problem.
- The embodiments of the disclosure provide a method for tracking a movable object, adapted to a tracking device. The method includes: determining a first on duration of a camera of the tracking device, wherein the first on duration comprises a starting time and an ending time; determining a second on duration of a plurality of light emitting elements disposed on the movable object by adding a first guard time before the starting time and adding a second guard time after the ending time; turning on the light emitting elements based on the second on duration; and controlling the camera to capture a specific image of the light emitting elements in the first on duration and accordingly tracking the movable object.
- The embodiments of the disclosure provide a tracking device including a camera and a processor. The processor is coupled to the camera and configured to perform: determining a first on duration of the camera of the tracking device, wherein the first on duration comprises a starting time and an ending time; determining a second on duration of a plurality of light emitting elements disposed on the movable object by adding a first guard time before the starting time and adding a second guard time after the ending time; turning on the light emitting elements based on the second on duration; and controlling the camera to capture a specific image of the light emitting elements in the first on duration and accordingly tracking the movable object.
- The embodiments of the disclosure provide a method for dynamically controlling shooting parameters of a camera. The method includes: determining a plurality of time frames, wherein the time frames comprise a plurality of first time frames and a plurality of second time frames; controlling the camera of the tracking device to shoot a first image with a first exposure parameter in each of the first time frames and accordingly performing an environment detection;
- controlling the camera of the tracking device to shoot a second image with a second exposure parameter in each of the second time frames and accordingly performing a first specific detection, wherein the second exposure parameter is lower than the first exposure parameter.
- The accompanying drawings are included to provide a further understanding of the invention, and are incorporated in and constitute a part of this specification. The drawings illustrate embodiments of the invention and, together with the description, serve to explain the principles of the invention.
-
FIG. 1 shows a functional diagram of a tracking device and a movable object according to an embodiment of the disclosure. -
FIG. 2A shows a first combination of the first on duration and the second on duration according to an embodiment of the disclosure. -
FIG. 2B shows a second combination of the first on duration and the second on duration according to an embodiment of the disclosure. -
FIG. 3 shows a flow chart of the method for tracking the movable object according to an embodiment of the disclosure. -
FIG. 4 shows a third combination of the first on duration and the second on duration according to an embodiment of the disclosure. -
FIG. 5 shows a flow chart of the method for dynamically controlling shooting parameters of a camera according to an embodiment of the disclosure. -
FIG. 6 shows the time frames according to a first embodiment of the disclosure. -
FIG. 7 shows the time frames according to the second embodiment of the disclosure. - Reference will now be made in detail to the present preferred embodiments of the invention, examples of which are illustrated in the accompanying drawings. Wherever possible, the same reference numbers are used in the drawings and the description to refer to the same or like parts.
- See
FIG. 1 , which shows a functional diagram of a tracking device and a movable object according to an embodiment of the disclosure. InFIG. 1 , thetracking device 100 may be any device capable of tracking the movable object 199 by using any known tracking mechanisms. In some embodiments, thetracking device 100 may be a head-mounted display (HMD) used in an AR system or a VR system, and the movable object 199 may be a component of the AR/VR system that needs to be tracked, such as a handheld controller, but the disclosure is not limited thereto. - In the embodiments of the disclosure, the
tracking device 100 includes acamera 102 and aprocessor 104. In various embodiments, thecamera 102 could be any cameras having charge coupled device (CCD) lens, complementary metal oxide semiconductor transistors (CMOS) lens, but the disclosure is not limited thereto. - The
processor 104 is coupled to thecamera 102, and may be, for example, a general-purpose processor, a special purpose processor, a conventional processor, a digital signal processor (DSP), a plurality of microprocessors, one or more microprocessors in association with a DSP core, a controller, a microcontroller, Application Specific Integrated Circuits (ASICs), Field Programmable Gate Array (FPGAs) circuits, any other type of integrated circuit (IC), a state machine, and the like. - In the embodiments of the disclosure, the
processor 104 may control thecamera 102 to capture images of the movable object 199 and determine the pose of the movable object 199 accordingly. In some embodiments, the movable object 199 may be disposed with a plurality of light emitting elements (such as regular light emitting diodes (LED) and/or infrared LEDs). - In this case, the
processor 104 may perform inside-out tracking based on the light distributions of the light emitting elements in the images captured by thecamera 102. - In the embodiments of the disclosure, the
camera 102 may be configured to capture images in the corresponding on durations (referred to as first on durations). That is, thecamera 102 would not capture/shoot in the durations other than the first on durations. Similarly, the light emitting elements on the movable object 199 have their own on durations (referred to as second on durations). In this case, the light emitting elements would not emit light in the durations other than the second on durations. - In the embodiments of the disclosure, the first on durations and the second on durations need to (partially) overlap with each other for the
camera 102 to capture the lights of the light emitting elements. However, if the first on durations and the second on durations are not properly designed, some undesired effects may be introduced. - See
FIG. 2A , which shows a first combination of the first on duration and the second on duration according to an embodiment of the disclosure. InFIG. 2A , the first onduration 211 is longer than the second onduration 212. In this case, the long exposure resulted from the first onduration 211 would brighten the background in the captured images, such that the objects needless to be tracked would also be captured in the images. Accordingly, the performance of tracking the movable object 199 would be degraded. In addition, since the exposure duration of thecamera 102 is usually very short, it is difficult to precisely control the light emitting elements to emit lights in the second onduration 212. - See
FIG. 2B , which shows a second combination of the first on duration and the second on duration according to an embodiment of the disclosure. InFIG. 2B , the light emitting elements on the movable object 199 are designed to be always on. In this case, the first onduration 221 may be designed to be short, which prevents the background in the captured images from being unnecessarily brightened. However, the power consumption of the light emitting elements would be high. - To solve the above problem, the embodiments of the disclosure provide a method for tracking the movable object, and detailed discussion would be provided in the following.
- See
FIG. 3 , which shows a flow chart of the method for tracking the movable object according to an embodiment of the disclosure. The method of this embodiment may be executed by thetracking device 100 inFIG. 1 , and the details of each step inFIG. 3 will be described below with the components shown inFIG. 1 . For better understanding the concept of the disclosure,FIG. 4 would be used as an example, whereinFIG. 4 shows a third combination of the first on duration and the second on duration according to an embodiment of the disclosure. - In step S310, the
processor 104 determines a first onduration 411 of thecamera 102 of thetracking device 100. In the embodiment, theprocessor 104 may determine the first onduration 411 to a length that would not be too long to overly brighten the background in the captured images, e.g., 16 μs, but the disclosure is not limited thereto. InFIG. 4 , the first onduration 411 has a starting time T1 and an ending time T2. - In step S320, the
processor 104 determines a second onduration 412 of the light emitting elements disposed on the movable object 199 by adding a first guard time G1 before the starting time T1 and adding a second guard time G2 after the ending time T2. In various embodiments, the lengths of the first guard time G1 and the second guard time G2 may be arbitrarily chosen based on the requirements of the designer. - Next, in step S330, the
processor 104 turns on the light emitting elements based on the second onduration 412. In one embodiment, if the light emitting elements on the movable object 199 can be independently controlled, theprocessor 104 can directly or indirectly control the light emitting elements to emit lights in the second onduration 412. In one embodiment, if the light emitting elements are powered and controlled by the movable object 199, theprocessor 104 may control the movable object 199 to turn on the light emitting elements in the second onduration 412 for emitting lights during the second onduration 412. - In one embodiment, the
processor 104 may synchronize with the movable object 199 and/or the light emitting elements to notify that the light emitting elements should be turned on in the second onduration 412. - In step S340, the
processor 104 controls thecamera 102 to capture a specific image of the light emitting elements in the first onduration 411 and accordingly tracking the movable object 199. - In one embodiment, the
processor 104 may determine the pose of the movable object 199 based on the light distribution of the light emitting elements in the specific image by using the inside-out tracking mechanism, but the disclosure is not limited thereto. - Unlike the scenario in
FIG. 2A , since the second onduration 412 is longer than the first onduration 411, it would be easier to control the light emitting elements to emit light in the second onduration 412. In addition, compared withFIG. 2B , the power consumption ofFIG. 4 would be much lower, which improves the battery life of the movable object 199 and/or the light emitting elements. - In addition, as mentioned in the above, if the shooting parameters use for performing environment detection and object tracking are the same, either the pose of the to-be-tracked object cannot be accurately determined, or the environment cannot be accurately detected. Therefore, the disclosure further proposes a method for dynamically controlling shooting parameters of a camera, which may be used to solve this problem.
- See
FIG. 5 , which shows a flow chart of the method for dynamically controlling shooting parameters of a camera according to an embodiment of the disclosure. The method of this embodiment may be executed by thetracking device 100 inFIG. 1 , and the details of each step inFIG. 5 will be described below with the components shown inFIG. 1 . - In step S510, the
processor 104 determines a plurality of time frames, wherein the time frames comprise a plurality of first time frames and a plurality of second time frames. - In step S520, the
processor 104 controls thecamera 102 of thetracking device 100 to shoot a first image with a first exposure parameter in each of the first time frames and accordingly performing an environment detection. - In step S530, the
processor 104 controls thecamera 102 of thetracking device 100 to shoot a second image with a second exposure parameter in each of the second time frames and accordingly performing a first specific detection, wherein the second exposure parameter is lower than the first exposure parameter. - For better understanding the concept of
FIG. 5 ,FIG. 6 andFIG. 7 would be used as example for discussions. - See
FIG. 6 , which shows the time frames according to a first embodiment of the disclosure. InFIG. 6 , theprocessor 104 may divide the time axis into a plurality of time frames, which are shown as rectangle boxes. In the present embodiment, the time frames may be divided into a plurality of first time frames F11-F1N and a plurality of second time frames F21-F2N. - In one embodiment, the first time frames F11-F1N and the second time frames F21-F2N are interleaved with each other. That is, a (2i+1)-th time frame of the time frames belongs to the first time frames F11-F1N, and a (2i+2)-th time frame of the time frames belongs to the second time frames F21-F2N, wherein i is an index.
- For example, when i is 0, the (2i+1)-th time frame of the time frames may be the 1-st time frame, which would be the first time frame F11. In addition, when i is 0, the (2i+2)-th time frame of the time frames may be the 2-nd time frame, which would be the second time frame F21. For another example, when i is 1, the (2i+1)-th time frame of the time frames may be the 3-rd time frame, which would be the first time frame F12. In addition, when i is 1, the (2i+2)-th time frame of the time frames may be the 4-th time frame, which would be the second time frame F22.
- In one embodiment, the
processor 104 may control thecamera 102 to shoot a first image with the first exposure parameter in each of the first time frames F11-F1N and accordingly performing an environment detection. In one embodiment, theprocessor 104 may perform the environment detection based on a simultaneous localization and mapping (SLAM) mechanism, but the disclosure is not limited thereto. - In one embodiment, the first specific detection may be used for tracking the movable object 199 disposed with the light emitting elements as taught in the above.
- In one embodiment, since the
processor 104 needs more feature points of the environment to better detect the environment, theprocessor 104 uses a higher first exposure parameter (e.g., higher exposure value and/or longer exposure time) to guarantee that the environment could be properly brightened in the first image. Accordingly, a better performance of the environment detection may be achieved in each of the first time frames F11-F1N. - On the other hand, when performing the first specific detection (e.g., tracking the movable object 199), the
processor 104 can use a lower second exposure parameter (e.g., lower exposure value and/or shorter exposure time) to dim the lights from the environment in the second image, such that only the lights from the light emitting elements would be preserved in the second image. Accordingly, a better performance of the first specific detection may be achieved in each of the second time frames F21-F2N. - In a second embodiment, the time frames may further include a plurality of third time frames, and the
processor 104 may control thecamera 102 to shoot a third image with a third exposure parameter in each of the third time frames and accordingly performing a second specific detection, wherein the third exposure parameter is higher than the second exposure parameter, and can be equal or not equal to the first exposure parameter. - In one embodiment, the first specific detection is used for tracking the movable object 199, the second specific detection is used for the gesture detection or hand tracking.
- See
FIG. 7 , which shows the time frames according to the second embodiment of the disclosure. InFIG. 7 , theprocessor 104 may divide the timeline into a plurality of time frames, which are shown as rectangle boxes. In the present embodiment, the time frames may be divided into a plurality of first time frames F11-F1N, a plurality of second time frames F21-F2N, and a plurality of third time frames F31-F3N, wherein the first time frames F11-F1N, the second time frames F21-F2N, and the third time frames F31-F3N are interleaved with each other. - That is, a (3i+1)-th time frame of the time frames belongs to the first time frames F11-F1N, a (3i+2)-th time frame of the time frames belongs to the second time frames F21-F2N, a (3i+3)-th time frame of the time frames belongs to the third time frames F31-F3N, wherein i is an index.
- For example, when i is 0, the (3i+1)-th time frame of the time frames may be the 1-st time frame, which would be the first time frame F11. When i is 0, the (3i+2)-th time frame of the time frames may be the 2-nd time frame, which would be the second time frame F21.
- When i is 0, the (3i+3)-th time frame of the time frames may be the 3-rd time frame, which would be the third time frame F31.
- For another example, when i is 1, the (3i+1)-th time frame of the time frames may be the 4-th time frame, which would be the first time frame F12. When i is 1, the (3i+2)-th time frame of the time frames may be the 5-th time frame, which would be the second time frame F22. When i is 1, the (3i+3)-th time frame of the time frames may be the 6-th time frame, which would be the third time frame F32.
- In one embodiment, the
processor 104 may control thecamera 102 to shoot a first image with the first exposure parameter in each of the first time frames F11-F1N and accordingly performing an environment detection. In one embodiment, theprocessor 104 may perform the environment detection based on the SLAM mechanism, but the disclosure is not limited thereto. - In one embodiment, the first specific detection may be used for tracking the movable object 199 disposed with the light emitting elements as taught in the above.
- In one embodiment, since the
processor 104 needs more feature points of the environment to better detect the environment, theprocessor 104 uses a higher first exposure parameter to guarantee that the environment could be properly brightened in the first image. Accordingly, a better performance of the environment detection may be achieved in each of the first time frames F11-F1N. - On the other hand, when performing the first specific detection (e.g., tracking the movable object 199), the
processor 104 can use a lower second exposure parameter to dim the lights from the environment in the second image, such that only the lights from the light emitting elements would be preserved in the second/third image. Accordingly, a better performance of the first specific detection may be achieved in each of the second time frames F21-F2N. - In summary, the embodiments of the disclosure can add guard times before/after the first on duration of the camera of the tracking device to determine the second on duration of the light emitting elements, such that the light emitting elements can be easier controlled to emit light in the second on duration while reducing the power consumption.
- In addition, the embodiments of the disclosure can use a higher first exposure parameter (e.g., higher exposure value and/or exposure time) to retrieve more feature points of the environment in the first image when performing environment detection in each of the first time frames. Accordingly, a better performance of the environment detection may be achieved in each of the first time frames.
- On the other hand, the embodiments of the disclosure can use a lower second exposure parameter (e.g., lower exposure value and/or exposure time) to make sure that only the lights from the light emitting elements would be preserved in the second image. Accordingly, a better performance of the first specific detection may be achieved in each of the second time frames.
- It will be apparent to those skilled in the art that various modifications and variations can be made to the structure of the present invention without departing from the scope or spirit of the invention. In view of the foregoing, it is intended that the present invention cover modifications and variations of this invention provided they fall within the scope of the following claims and their equivalents.
Claims (19)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/506,696 US20220132042A1 (en) | 2020-10-26 | 2021-10-21 | Method for tracking movable object, tracking device, and method for controlling shooting parameters of camera |
US18/395,648 US20240129634A1 (en) | 2020-10-26 | 2023-12-25 | Method for controlling shooting parameters of camera and tracking device |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US202063105366P | 2020-10-26 | 2020-10-26 | |
US202163145497P | 2021-02-04 | 2021-02-04 | |
US17/506,696 US20220132042A1 (en) | 2020-10-26 | 2021-10-21 | Method for tracking movable object, tracking device, and method for controlling shooting parameters of camera |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/395,648 Division US20240129634A1 (en) | 2020-10-26 | 2023-12-25 | Method for controlling shooting parameters of camera and tracking device |
Publications (1)
Publication Number | Publication Date |
---|---|
US20220132042A1 true US20220132042A1 (en) | 2022-04-28 |
Family
ID=81257173
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/506,696 Abandoned US20220132042A1 (en) | 2020-10-26 | 2021-10-21 | Method for tracking movable object, tracking device, and method for controlling shooting parameters of camera |
US18/395,648 Pending US20240129634A1 (en) | 2020-10-26 | 2023-12-25 | Method for controlling shooting parameters of camera and tracking device |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/395,648 Pending US20240129634A1 (en) | 2020-10-26 | 2023-12-25 | Method for controlling shooting parameters of camera and tracking device |
Country Status (3)
Country | Link |
---|---|
US (2) | US20220132042A1 (en) |
CN (2) | CN117119289A (en) |
TW (1) | TWI796809B (en) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1991014644A1 (en) * | 1990-03-28 | 1991-10-03 | Asea Brown Boveri Ab | Transfer and positioning of goods by means of container cranes |
US20120081566A1 (en) * | 2010-09-30 | 2012-04-05 | Apple Inc. | Flash synchronization using image sensor interface timing signal |
US20170366805A1 (en) * | 2014-12-31 | 2017-12-21 | Alt Llc | Method and system for displaying three-dimensional objects |
US20190318501A1 (en) * | 2018-04-16 | 2019-10-17 | Microsoft Technology Licensing, Llc | Tracking pose of handheld object |
WO2022019776A1 (en) * | 2020-07-22 | 2022-01-27 | Weta Digital Limited | Sealed active marker for performance capture |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101699237A (en) * | 2009-11-20 | 2010-04-28 | 中国航空工业空气动力研究院 | Three-dimensional model attitude angle video measuring system for wind tunnel model test |
JP2011259245A (en) * | 2010-06-09 | 2011-12-22 | Fujifilm Corp | Imaging device and imaging method |
CN102169366B (en) * | 2011-03-18 | 2012-11-07 | 汤牧天 | Multi-target tracking method in three-dimensional space |
JP2012226513A (en) * | 2011-04-19 | 2012-11-15 | Honda Elesys Co Ltd | Detection device and detection method |
EP2802973A1 (en) * | 2012-01-09 | 2014-11-19 | Epson Norway Research and Development AS | Low interference system and method for synchronization, identification and tracking of visual and interactive systems |
CN107767394A (en) * | 2016-08-16 | 2018-03-06 | 蒋博 | A kind of positioning of moving target and Attitude estimation method, apparatus and system |
WO2018187171A1 (en) * | 2017-04-04 | 2018-10-11 | Usens, Inc. | Methods and systems for hand tracking |
US10755422B2 (en) * | 2017-07-24 | 2020-08-25 | Htc Corporation | Tracking system and method thereof |
US20190243376A1 (en) * | 2018-02-05 | 2019-08-08 | Qualcomm Incorporated | Actively Complementing Exposure Settings for Autonomous Navigation |
US10878285B2 (en) * | 2018-04-12 | 2020-12-29 | Seiko Epson Corporation | Methods and systems for shape based training for an object detection algorithm |
US20210208673A1 (en) * | 2020-01-03 | 2021-07-08 | Facebook Technologies, Llc | Joint infrared and visible light visual-inertial object tracking |
-
2021
- 2021-10-21 US US17/506,696 patent/US20220132042A1/en not_active Abandoned
- 2021-10-25 CN CN202311106210.4A patent/CN117119289A/en active Pending
- 2021-10-25 CN CN202111240652.9A patent/CN114500823B/en active Active
- 2021-10-25 TW TW110139449A patent/TWI796809B/en active
-
2023
- 2023-12-25 US US18/395,648 patent/US20240129634A1/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1991014644A1 (en) * | 1990-03-28 | 1991-10-03 | Asea Brown Boveri Ab | Transfer and positioning of goods by means of container cranes |
US20120081566A1 (en) * | 2010-09-30 | 2012-04-05 | Apple Inc. | Flash synchronization using image sensor interface timing signal |
US20170366805A1 (en) * | 2014-12-31 | 2017-12-21 | Alt Llc | Method and system for displaying three-dimensional objects |
US20190318501A1 (en) * | 2018-04-16 | 2019-10-17 | Microsoft Technology Licensing, Llc | Tracking pose of handheld object |
WO2022019776A1 (en) * | 2020-07-22 | 2022-01-27 | Weta Digital Limited | Sealed active marker for performance capture |
Also Published As
Publication number | Publication date |
---|---|
CN114500823A (en) | 2022-05-13 |
CN114500823B (en) | 2023-10-13 |
TW202218407A (en) | 2022-05-01 |
CN117119289A (en) | 2023-11-24 |
TWI796809B (en) | 2023-03-21 |
US20240129634A1 (en) | 2024-04-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9007489B2 (en) | Method and apparatus for image background removal and replacement | |
US7962031B2 (en) | Pulsed control of camera flash | |
US10292242B2 (en) | Control device, control method, and program | |
CN109068057B (en) | Automatic light following system based on accurate positioning | |
CN110933337B (en) | HDR image sensor with LED flicker mitigation and motion blur reduction | |
US20140184896A1 (en) | Accessory, camera, accessory control program, and camera control program | |
CN113014748A (en) | Camera infrared lamp dynamic light supplementing system and method | |
US20220132042A1 (en) | Method for tracking movable object, tracking device, and method for controlling shooting parameters of camera | |
JP6459620B2 (en) | Light receiving unit, light emission control method and program | |
EP3836528A1 (en) | Night vision device and method | |
JP6654749B1 (en) | Camera device and IR light irradiation method | |
US9264596B2 (en) | Security camera system and method for eliminating video aberrations from excessive IR illumination | |
JP5423747B2 (en) | Accessories, camera, accessory control program, and camera control program | |
WO2020214078A1 (en) | External light source for mobile devices | |
US20220317771A1 (en) | Method for tracking trackers and host | |
KR20120077946A (en) | A network camera with enhanced efficiency for lighting device and a controlling method thereof | |
JP6859996B2 (en) | Light emitting unit, light emitting control method and program | |
JP2019132586A (en) | Sensor and electronic device | |
US10368004B2 (en) | Light source control method and camera apparatus thereof | |
TWI651663B (en) | Method, device, and non-transitory computer readable storage medium for object tracking | |
GB2589872A (en) | Night vision device and method | |
CN105516550A (en) | Night mode operation method of automatic camera head | |
JP2020030261A (en) | Imaging apparatus | |
JP2019003854A (en) | Imaging apparatus, controller, and lighting control system | |
JP2016201716A (en) | Monitoring camera apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: HTC CORPORATION, TAIWAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHEN, YUAN-TUNG;LIN, JYUN-JHONG;CHEN, CHIH CHIEN;REEL/FRAME:057922/0222 Effective date: 20211020 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |