CN113192125A - Multi-camera video concentration method and system in geographic scene with optimal virtual viewpoint - Google Patents
Multi-camera video concentration method and system in geographic scene with optimal virtual viewpoint Download PDFInfo
- Publication number
- CN113192125A CN113192125A CN202110327605.1A CN202110327605A CN113192125A CN 113192125 A CN113192125 A CN 113192125A CN 202110327605 A CN202110327605 A CN 202110327605A CN 113192125 A CN113192125 A CN 113192125A
- Authority
- CN
- China
- Prior art keywords
- camera
- video
- virtual viewpoint
- geographic
- scene
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 55
- 238000013507 mapping Methods 0.000 claims abstract description 24
- 238000013210 evaluation model Methods 0.000 claims abstract description 12
- 230000000694 effects Effects 0.000 claims abstract description 11
- 238000012544 monitoring process Methods 0.000 claims abstract description 8
- 239000011159 matrix material Substances 0.000 claims description 9
- 238000005457 optimization Methods 0.000 claims description 6
- 230000008569 process Effects 0.000 claims description 5
- 230000000007 visual effect Effects 0.000 claims description 5
- 230000008520 organization Effects 0.000 claims description 4
- 238000004364 calculation method Methods 0.000 claims description 3
- 230000005494 condensation Effects 0.000 claims description 3
- 238000009833 condensation Methods 0.000 claims description 3
- 239000000470 constituent Substances 0.000 claims description 3
- 238000005259 measurement Methods 0.000 claims description 3
- 238000012216 screening Methods 0.000 claims description 3
- 238000010276 construction Methods 0.000 claims description 2
- 230000004927 fusion Effects 0.000 abstract description 5
- 230000010354 integration Effects 0.000 abstract description 3
- 238000010586 diagram Methods 0.000 description 10
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 230000002035 prolonged effect Effects 0.000 description 2
- 238000010219 correlation analysis Methods 0.000 description 1
- 230000002349 favourable effect Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/05—Geographic models
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Geometry (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Remote Sensing (AREA)
- Computer Graphics (AREA)
- Processing Or Creating Images (AREA)
Abstract
A multi-camera video concentration method and system in a geographic scene with a virtual viewpoint being optimized acquire coordinate data of a same-name point pair by acquiring video sequence image information and acquiring the same-name point pair from a video image and a three-dimensional geographic scene model; establishing a mapping relation between a video image and a geographic space according to coordinate data of the same-name point pair, and positioning a camera view; constructing a camera observation domain model by analyzing the observable distance and the sight line deflection angle, and generating an observable set of a camera group; optimizing the observable set by constructing an evaluation model to obtain a virtual viewpoint group; and presetting display parameters of the moving target, and performing multi-camera video concentration according to the display parameters. The method has the remarkable effects that the mapping relation between the video target and the geographic scene is established, the effect of fusion expression of the monitoring video in the geographic scene is enhanced, and great convenience is provided for the quick retrieval and efficient understanding of the video geographic scene information integration.
Description
Technical Field
The invention relates to the technical field of real-time fusion of video streams and three-dimensional models, in particular to a method and a system for concentrating multi-camera videos in a geographic scene with optimal virtual viewpoints.
Background
With the requirement of the Virtual Geographic Environment (VGE) on the accuracy and the real-time performance of scene simulation, the multi-source heterogeneous data is introduced to enhance the visual expression and analysis functions of the VGE. The video data can not only realize the real scene display of the geographic environment, but also describe the space-time motion of moving objects (pedestrians, vehicles and the like) in the geographic scene. When a user views a video in a VGE, the virtual viewpoint is typically selected at a virtual location that is close to the original geographic location of the camera.
However, the conventional virtual viewpoint selecting method is often in actual deployment and use, and has the following difficulties and problems:
first, it is convenient to see a single camera and short duration videos in this way, but if a scene contains multiple videos, and the cameras have different sight directions and non-overlapping and discrete views, it is difficult for a user to view all videos through a single virtual viewpoint. If each path of video is individually arranged with one virtual viewpoint to be watched one by one, the watching duration can be greatly prolonged, and the user is not favorable for quickly watching the video content.
Secondly, video moving objects usually appear in cameras shooting different areas one after another, and each path of video image is viewed independently, so that the cross-camera global motion condition of the moving objects in a scene cannot be expressed.
Therefore, how to effectively select a small number of virtual viewpoints in a virtual scene to quickly view a multi-camera video target and show the cross-camera motion condition of the video target becomes a technical problem to be solved urgently.
Disclosure of Invention
Therefore, the invention provides a multi-camera video concentration method and system in a geographic scene with an optimal virtual viewpoint, which aims to solve the problem that in the prior art, when a user watches videos in a VGE, the virtual viewpoint is generally selected at a virtual position close to the original geographic position of a camera, so that the method is convenient for watching a single camera and short-duration videos, but if the scene contains multiple paths of videos, the sight directions of the cameras are different, and the sight fields are not overlapped and distributed discretely, the user is difficult to watch all the videos through a single virtual viewpoint; if each path of video is individually arranged with one virtual viewpoint to be watched one by one, the watching duration can be greatly prolonged, and the user is not facilitated to quickly view the video content; on the other hand, video moving objects usually appear in cameras shooting different areas in sequence, each path of video image is viewed independently, and the problem that the cross-camera global motion condition of the moving objects in a scene cannot be expressed is solved.
In order to achieve the above purpose, the invention provides the following technical scheme: in a first aspect, a method for concentrating multi-camera video in a geographic scene with a preferred virtual viewpoint is provided, which includes the following steps:
acquiring video sequence image information, acquiring homonymous point pairs in a video image and a three-dimensional geographic scene model, and acquiring coordinate data of the homonymous point pairs, wherein the coordinate data comprises image coordinates and geographic coordinates;
establishing a mapping relation between a video image and a geographic space according to coordinate data of the same-name point pair, and positioning a camera view;
constructing a camera observation domain model by analyzing the observable distance and the sight line deflection angle, and generating an observable set of a camera group;
optimizing the observable set by constructing an evaluation model to obtain a virtual viewpoint group;
presetting display parameters of the moving target, and performing multi-camera video concentration according to the display parameters.
As a preferred scheme of the multi-camera video concentration method in the geographic scene with the preferred virtual viewpoint, the video image is a first frame image of the intercepted monitoring video.
As a preferred scheme of the multi-camera video concentration method in the geographic scene with the preferred virtual viewpoint, the three-dimensional geographic scene model is a three-dimensional scene model constructed according to the measurement information of the real geographic scene, the number of the same-name point pairs collected on the video image and the virtual geographic scene is not less than three pairs, and the three pairs of the same-name point pairs are not completely collinear.
As a preferred scheme of the multi-camera video condensing method in the geographic scene with the preferred virtual viewpoint, the establishing of the mapping relationship between the video image and the geographic space includes the following steps:
a1) presetting an object point geographic space coordinate Q corresponding to a given image point image space coordinate Q, and expressing Q and Q as a homogeneous coordinate form:
q=[x y 1]T
Q=[X Y Z 1]T;
when a homography matrix M is recorded, the relationship between Q and Q is:
q=MQ;
the expression of the homography matrix M is:
a2) solving the geographic space coordinates of the object points corresponding to the image points in each image:
a3) assuming that there are L cameras in the current video camera network, the mapping matrix is labeled M for the k-th camera (k is 1,2 … L)k(ii) a Defining each camera position in geographic space asThe geospatial position of each camera field of view polygon is
Wherein the camera positionConsidered as a point in the geographic space, the camera view polygon is bounded by o boundary points Pk,numSequentially concatenating the constituent polygon records.
As a preferred scheme of a multi-camera video concentration method in a geographic scene with a preferred virtual viewpoint, in the process of positioning the camera view, two factors of a virtual line-of-sight distance and a camera-virtual viewpoint included angle are selected as constraint conditions;
the virtual line-of-sight distance refers to the geographic space distance between a virtual viewpoint and a certain point in the view field; the camera-virtual viewpoint included angle is an included angle formed by projection of a given point in a visual field, the virtual viewpoint and a camera position point on a horizontal plane;
defining a distance threshold TdisAnd an angle threshold TangAs a constraint, and assuming a distance threshold TdisAnd an angle threshold TangIt has been given that finding areas in the scene model that meet the constraints serves as virtual viewpoint ranges.
As a preferred scheme of the multi-camera video concentration method in a geographical scene with a preferred virtual viewpoint, the method for constructing a camera observation domain model by analyzing an observable distance and a view deflection angle and generating an observable set of camera groups comprises the following steps:
b1) recording camera positionPolygon with camera viewAnd isIn the distanceIs Pk,n1Pk,n2The farthest line segment is Pk,n3Pk,n4;
b2) Respectively at point Pk,n3,Pk,n4As a circle center, a distance threshold value TdisIs a radius on the line segment Pk,n3Pk,n4Towards the camera positionDrawing a semicircle at one side, and taking two semicircles as a line segment Pk,n1Pk,n2Near camera positionOne side intersection region as a virtual viewpoint distance reasonable region Ak,dis;
b3) Respectively at point Pk,n1,Pk,n2Being a corner point, an angle threshold TangFor deflection angle, clockwise and anticlockwise respectively deflect TangMaking four rays, and taking the four rays to be on the line segment Pk,n1Pk,n2Near camera positionOne side intersection region as a virtual viewpoint angle reasonable region Ak,ang;
b4) Virtual viewpoint range A of camerakIs Ak,disAnd Ak,angThe intersection of (a);
b5) let Obj be the total set of all video moving objects in all video cameras; note that there is N in the kth camerakEach video moving object has a track marked as Ck,i,Ck,iThe expression of (a) is as follows:
Obj={Ck,i,(k=1,2…L)}
Ck,i={Ik,i,j,Pk,i,j,(i=1,2,…,Nk)(j=1,2,…,n)};
wherein L represents the number of cameras, Ik,i,jAnd Pk,i,jRepresenting the sub-image of the ith video moving object in the jth video frame in the kth camera and the geographic space position of the sub-image are analyzed by the cross-camera association of the video moving objects, and the tracks of the single-camera video moving object are merged to obtain the tracks of the multi-camera video moving objectRealizing the related organization of the multi-camera video moving target:
Cubeio={Ck1,i1,Ck2,i2…Cko,iL,(k1,k2,…ko)∈(1,2…L)};
wherein L isoRepresenting the total number of video moving objects, Cube, after merging of cross-camera homonymous video moving objects in a surveillance video networkioRepresenting the global track of the video moving object with the sequence number io in the monitoring video network,and (4) representing the sub-track of the video moving object with the sequence number io in the ko camera.
As a preferred scheme of a multi-camera video concentration method in a geographical scene with a preferred virtual viewpoint, the observable collection is preferred by constructing an evaluation model to obtain a virtual viewpoint group, and the method specifically comprises the following steps:
c1) the number of cameras is recorded as L, and a set formed by all combination modes of the cameras is recorded as M:
mi={ni,j}
wherein m isiThe ith camera combination mode comprises all camera sets in the combination mode; n isi,jDigital camera combination mode miA j-th group of cameras under the condition that all cameras of the group are included;the first camera in the j camera set in the i camera set combination mode is referred to;
c2) by defining a distance threshold TdisAnd an angle threshold TangFor each cameraCombined mode miIn each camera group ni,jAll cameras inCalculating observable domains and calculating intersection; if a certain camera combination mode miAll the camera sets ni,jIf the intersection of any number of the camera observation domains is not empty, the combination mode m of the camera is recordediIs an observable combination, otherwise, the camera combination mode m is recordediIs an unobservable combination;
c3) based on multi-camera video target trajectory data, the following video condensation optimization targets are specified to realize the optimization of the camera group:
firstly, expressing coherence of homonymous targets across cameras, namely expressing the video cameras with the appearance of a single target in a consistent manner by using as few virtual viewpoints as possible;
expressing that the total number of virtual viewpoints used by all video targets is as small as possible;
c4) comprehensively evaluating the multi-camera video target expression effect of the camera combination corresponding to the virtual viewpoint group by value:
wherein n iscRepresents the total number of cameras, nvRepresenting the number of virtual viewpoints, N representing the total number of video moving objects, miRepresenting the number of virtual viewpoints which are related and expressed by each video moving object, wherein mu is a weight parameter;
c5) at a distance threshold TdisAnd an angle threshold TangWhen the value is constant, value values of all current camera observable sets are calculated through defining a parameter alpha, the maximum value is taken as a camera combination selection result, and multi-camera video concentration in a virtual scene is carried out.
As a preferred scheme of the method for concentrating the multi-camera video in the geographic scene with the preferred virtual viewpoint, the method for concentrating the multi-camera video according to the display parameters of the preset moving object comprises the following steps:
d1) recording that W virtual viewpoints (W is less than or equal to L) are needed for viewing the video moving objects of all L cameras under the current camera combination; meanwhile, setting a frame rate fps displayed by the video moving target subgraph in a three-dimensional scene as the number of the subgraphs displayed by a single video moving target per second; setting an object display interval time t0As time intervals for additional display of new video moving objects;
d2) for a certain virtual viewpoint W (W ≦ W), the first occurring moving object O is first displayed0In geospatial track T0And identifying the sequence of appearance of the video object among different cameras;
screening the video object subgraph according to the frame rate fps, converting the plane coordinate corresponding to the screened video object subgraph into a geographic coordinate, and meanwhile, according to a proportionality coefficient Pw、PhScaling a video object sub-graph, Pw、PhThe calculation formula is as follows:
whereinRandomly selecting an appropriate amount of average width and height of subgraphs from a video object subgraph library, mapping coordinates of three points, namely upper left, lower left and upper right, of the selected subgraphs in an original video frame to corresponding geographic positions in a virtual scene to obtain the length and height of the video object subgraphs in a three-dimensional space,is the average length and height of the video object sub-graph displayed in the virtual scene;
d3) displaying O in the camera view field in the virtual scene according to the frame rate fps in the dynamic display process0The video object sub-graph of the current frame is in the corresponding geographic position, and the old video object sub-graph is not displayed any more;
at t0,2t0…,nt0At the moment, the video object O is added1,O2…OnDynamically expressed in a three-dimensional scene model to realize the concentration of multi-camera video objects.
As a preferred scheme of the multi-camera video concentration method in the geographic scene with the preferred virtual viewpoint, for the condition that the same section of object track generated by overlapping the shooting areas of the cameras is obtained by a plurality of cameras, determining the camera for obtaining the object sub-image by comparing the included angles between the virtual viewpoint and the object track point and the connecting line of three points at the two camera positions:
and a view overlapping part C exists between the camera a and the camera b, for the video object passing through the view C, the included angle among the three points of the camera position, the track point and the virtual viewpoint V, namely the sizes of alpha and beta, is compared, if alpha is less than or equal to beta, the video object sub-graph acquired by the camera a is used, and otherwise, the video object sub-graph acquired by the camera b is used.
In a second aspect, there is provided a multi-camera video concentration system in a geographical scene with a preferred virtual viewpoint, where the multi-camera video concentration method in a geographical scene with a preferred virtual viewpoint in the first aspect or any possible implementation manner thereof is adopted, and the concentration system includes:
the same-name point acquisition module: the system comprises a video sequence image acquisition module, a video acquisition module, a three-dimensional geographic scene model acquisition module, a coordinate acquisition module and a display module, wherein the video sequence image acquisition module is used for acquiring video sequence image information, acquiring homonymous point pairs in a video image and the three-dimensional geographic scene model and acquiring coordinate data of the homonymous point pairs, and the coordinate data comprises image coordinates and geographic coordinates;
the mapping relation construction module: the method comprises the steps of establishing a mapping relation between a video image and a geographic space according to coordinate data of a same-name point pair, and positioning a camera view;
a camera group observable collection generation module: the system comprises a camera observation domain model and a camera group observation set, wherein the camera observation domain model is constructed by analyzing an observable distance and a sight line deflection angle to generate a camera group observable set;
a virtual viewpoint group generation module: the system is used for optimizing the observable collection by constructing an evaluation model to obtain a virtual viewpoint group;
the video target space-time motion expression module: the method comprises the steps of presetting display parameters of a moving target, and carrying out multi-camera video concentration according to the display parameters.
The invention has the following advantages: acquiring coordinate data of a same-name point pair by acquiring video sequence image information and acquiring the same-name point pair in a video image and a three-dimensional geographic scene model, wherein the coordinate data comprises an image coordinate and a geographic coordinate; establishing a mapping relation between a video image and a geographic space according to coordinate data of the same-name point pair, and positioning a camera view; constructing a camera observation domain model by analyzing the observable distance and the sight line deflection angle, and generating an observable set of a camera group; optimizing the observable set by constructing an evaluation model to obtain a virtual viewpoint group; and presetting display parameters of the moving target, and performing multi-camera video concentration according to the display parameters. The method has the remarkable effects that the mapping relation between the video target and the geographic scene is established, the effect of fusion expression of the monitoring video in the geographic scene is enhanced, and great convenience is provided for the quick retrieval and efficient understanding of the video geographic scene information integration.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below. It should be apparent that the drawings in the following description are merely exemplary, and that other embodiments can be derived from the drawings provided by those of ordinary skill in the art without inventive effort.
The structures, ratios, sizes, and the like shown in the present specification are only used for matching with the contents disclosed in the specification, so that those skilled in the art can understand and read the present invention, and do not limit the conditions for implementing the present invention, so that the present invention has no technical significance, and any structural modifications, changes in the ratio relationship, or adjustments of the sizes, without affecting the functions and purposes of the present invention, should still fall within the scope of the present invention.
Fig. 1 is a schematic diagram illustrating an integrated representation of multiple channels of video in a VGE according to an embodiment of the present invention;
fig. 2 is a schematic flowchart of a method for concentrating multi-camera video in a geographic scene with a preferred virtual viewpoint according to an embodiment of the present invention;
FIG. 3 is a schematic diagram of a camera and a geospatial coordinate system and an image space coordinate system provided in an embodiment of the invention;
FIG. 4(a) is a schematic diagram of a virtual line-of-sight distance provided in an embodiment of the present invention;
fig. 4(b) is a schematic diagram of a camera-virtual viewpoint included angle provided in the embodiment of the present invention;
FIG. 5(a) is a schematic diagram of the closest/farthest line segment of a camera to a video field provided in an embodiment of the present invention;
fig. 5(b) is a schematic diagram of a virtual viewpoint distance reasonable region provided in the embodiment of the present invention;
fig. 5(c) is a schematic diagram of a virtual viewpoint angle reasonable region provided in the embodiment of the present invention;
fig. 5(d) is a schematic diagram of a virtual viewpoint angle and distance plausible region provided in the embodiment of the present invention;
FIG. 6 is a schematic view of a set of observable cameras provided in an embodiment of the present invention;
FIG. 7 is a schematic diagram illustrating the concentration of multiple camera video objects in a geographic scene provided in an embodiment of the present invention;
FIG. 8 is a schematic view of a camera view overlap process provided in an embodiment of the invention;
fig. 9 is a schematic diagram of a multi-camera video concentrating system in a geographic scene with a preferred virtual viewpoint according to an embodiment of the present invention.
Detailed Description
The present invention is described in terms of particular embodiments, other advantages and features of the invention will become apparent to those skilled in the art from the following disclosure, and it is to be understood that the described embodiments are merely exemplary of the invention and that it is not intended to limit the invention to the particular embodiments disclosed. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Example 1
Referring to fig. 1, by introducing a video into the VGE and supporting video intelligent analysis with geospatial information, related functions such as video data organization management, spatial mapping, video-scene fusion expression and the like in the VGE can be realized.
Referring to fig. 2, a method for concentrating multi-camera video in a geographical scene with a preferred virtual viewpoint is provided, which comprises the following steps:
s1, acquiring video sequence image information, acquiring homonymy point pairs in a video image and a three-dimensional geographic scene model, and acquiring coordinate data of the homonymy point pairs, wherein the coordinate data comprises image coordinates and geographic coordinates;
s2, establishing a mapping relation between a video image and a geographic space according to coordinate data of the same-name point pair, and positioning a camera view;
s3, constructing a camera observation domain model by analyzing the observable distance and the sight line deflection angle, and generating an observable set of a camera group;
s4, optimizing the observable set by constructing an evaluation model to obtain a virtual viewpoint group;
and S5, presetting display parameters of the moving target, and concentrating the multi-camera video according to the display parameters.
Specifically, in step S1, the video image is a first frame image of the captured monitoring video. In step S1, the three-dimensional geographic scene model is a three-dimensional scene model constructed according to the real geographic scene measurement information, the number of the same-name point pairs collected on the video image and the virtual geographic scene is not less than three pairs, and the three pairs of the same-name point pairs are not all collinear.
In particular, referring to fig. 3, the relationship between the camera and the image space coordinate system and the geospatial coordinate system is illustrated. Recording a camera station center as C and an image space coordinate system as OiXiYiThe geospatial coordinate system is marked as OgXgYgZg. In step S2, the establishing of the mapping relationship between the video image and the geographic space includes the following steps:
a1) presetting an object point geographic space coordinate Q corresponding to a given image point image space coordinate Q, and expressing Q and Q as a homogeneous coordinate form:
q=[x y 1]T
Q=[X Y Z 1]T;
when a homography matrix M is recorded, the relationship between Q and Q is:
q=MQ;
the expression of the homography matrix M is:
a2) because M has 6 unknowns, at least 3 groups of known image point image space coordinates and object point geographic space coordinates are needed to realize the solution of M. After M is determined, solving the geographic space coordinates of the object points corresponding to the image points in each image:
a3) assuming that there are L cameras in the current video camera network, the mapping matrix is labeled M for the k-th camera (k is 1,2 … L)k(ii) a On this basis, each camera position in the geographic space is defined asThe geospatial position of each camera field of view polygon is
Wherein the camera positionConsidered as a point in the geographic space, the camera view polygon is bounded by o boundary points Pk,numSequentially concatenating the constituent polygon records.
In this embodiment, in step S2, in the process of positioning the camera view, two factors, namely, the virtual line-of-sight distance and the camera-virtual viewpoint included angle, are selected as constraint conditions;
referring specifically to fig. 4, the virtual line-of-sight distance refers to a geospatial distance between a virtual viewpoint and a certain point in the view field; the camera-virtual viewpoint included angle is an included angle formed by projection of a given point in a visual field, the virtual viewpoint and a camera position point on a horizontal plane;
defining a distance threshold TdisAnd an angle threshold TangAs a constraint, and assuming a distance threshold TdisAnd an angle threshold TangIt has been given that on this basis, areas meeting the constraints are found in the scene model as virtual viewpoint ranges.
Specifically, in step S3, the step of constructing a camera observation domain model by analyzing the observable distance and the gaze deflection angle, and generating the observable set of the camera group includes the following steps:
b1) referring specifically to FIG. 5(a), camera position is recordedPolygon with camera viewAnd isIn the distanceIs Pk,n1Pk,n2The farthest line segment is Pk,n3Pk,n4;
b2) With specific reference to FIG. 5(b), at points Pk,n3,Pk,n4As a circle center, a distance threshold value TdisIs a radius on the line segment Pk, n3Pk,n4Towards the camera positionDrawing a semicircle at one side, and taking two semicircles as a line segment Pk,n1Pk,n2Near camera positionOne side intersection region as a virtual viewpoint distance reasonable region Ak,dis;
b3) Referring specifically to FIG. 5(c), at points Pk,n1,Pk,n2Being a corner point, an angle threshold TangFor deflection angle, clockwise and anticlockwise respectively deflect TangMaking four rays, and taking the four rays to be on the line segment Pk,n1Pk,n2Near camera positionOne side intersection region as a virtual viewpoint angle reasonable region Ak,ang;
b4) Referring specifically to fig. 5(d), the virtual viewpoint range a of the camerakIs Ak,disAnd Ak,angThe intersection of (a);
b5) let Obj be the total set of all video moving objects in all video cameras; note N in the kth camerakEach video moving object has a track marked as Ck,i,Ck,iThe expression of (a) is as follows:
Obj={Ck,i,(k=1,2…L)}
Ck,i={Ik,i,j,Pk,i,j,(i=1,2,…,Nk)(j=1,2,…,n)};
wherein L represents the number of cameras, Ik,i,jAnd Pk,i,jAnd (3) representing the sub-image of the ith video moving object in the jth video frame in the kth camera and the geographic spatial position of the sub-image are subjected to cross-camera correlation analysis by the video moving object, and the single camera view is mergedObtaining multi-camera video motion target track by frequency motion target trackRealizing the related organization of the multi-camera video moving target:
Cubeio={Ck1,i1,Ck2,i2…Cko,iL,(k1,k2,…ko)∈(1,2…L)};
wherein L isoRepresenting the total number of video moving objects, Cube, after merging of cross-camera homonymous video moving objects in a surveillance video networkioRepresenting the global track of the video moving object with the sequence number io in the monitoring video network,and (4) representing the sub-track of the video moving object with the sequence number io in the ko camera.
Specifically, in step S4, the optimizing the observable set by constructing an evaluation model to obtain a virtual viewpoint group specifically includes the following steps:
c1) the number of cameras is recorded as L, and a set formed by all combination modes of the cameras is recorded as M:
mi={ni,j}
see in particular fig. 6, where miThe ith camera combination mode comprises all camera sets in the combination mode; n isi,jDigital camera combination mode miA j-th group of cameras under the condition that all cameras of the group are included;the first camera in the j camera set in the i camera set combination mode is referred to;
c2) by defining a distance threshold TdisAnd an angle threshold TangAssembling the patterns m for each cameraiIn each camera group ni,jAll cameras inCalculating observable domains and calculating intersection; if a certain camera combination mode miAll the camera sets ni,jIf the intersection of any number of the camera observation domains is not empty, the combination mode m of the camera is recordediIs an observable combination, otherwise, the camera combination mode m is recordediIs an unobservable combination;
c3) based on multi-camera video target trajectory data, the following video condensation optimization targets are specified to realize the optimization of the camera group:
firstly, expressing coherence of homonymous targets across cameras, namely expressing the video cameras with the appearance of a single target in a consistent manner by using as few virtual viewpoints as possible;
expressing that the total number of virtual viewpoints used by all video targets is as small as possible;
c4) comprehensively evaluating the multi-camera video target expression effect of the camera combination corresponding to the virtual viewpoint group by value:
wherein n iscRepresents the total number of cameras, nvRepresenting the number of virtual viewpoints, N representing the total number of video moving objects, miRepresenting the number of virtual viewpoints which are related and expressed by each video moving object, wherein mu is a weight parameter;
c5) at a distance threshold TdisAnd an angle threshold TangWhen the value is taken to be one, the value values of all current camera observable sets are calculated by defining a parameter alpha, and the maximum value is taken as the camera combination for selectionAnd taking the result, and concentrating the multi-camera video in the virtual scene.
Referring to fig. 7 in particular, step S4 sets the moving object display parameters to perform multi-camera video concentration under the condition that the camera display combination is selected, with the central point of the observable domain of each camera group as the virtual viewpoint, based on the observable set preference result.
In step S5, the preset display parameters of the moving object, according to the display parameters, perform multi-camera video concentration, including the following steps:
d1) recording that W virtual viewpoints (W is less than or equal to L) are needed for viewing the video moving objects of all L cameras under the current camera combination; meanwhile, setting a frame rate fps displayed by the video moving target subgraph in a three-dimensional scene as the number of the subgraphs displayed by a single video moving target per second; setting an object display interval time t0As time intervals for additional display of new video moving objects;
d2) for a certain virtual viewpoint W (W ≦ W), the first occurring moving object O is first displayed0In geospatial track T0And identifying the sequence of appearance of the video object among different cameras;
screening the video object subgraph according to the frame rate fps, converting the plane coordinate corresponding to the screened video object subgraph into a geographic coordinate, and meanwhile, according to a proportionality coefficient Pw、PhScaling a video object sub-graph, Pw、PhThe calculation formula is as follows:
whereinRandomly selecting the average width and height of a proper number of subgraphs from a video object subgraph library, and selectingThe extracted subgraph is mapped to the corresponding geographic position in the virtual scene corresponding to the three-point coordinates of the upper left, the lower left and the upper right in the original video frame to obtain the length and the height of the video object subgraph in the three-dimensional space,is the average length and height of the video object sub-graph displayed in the virtual scene;
d3) displaying O in the camera view field in the virtual scene according to the frame rate fps in the dynamic display process0The video object sub-graph of the current frame is in the corresponding geographic position, and the old video object sub-graph is not displayed any more;
on the other hand, at t0,2t0…,nt0At the moment, the video object O is added1,O2…OnDynamically expressed in a three-dimensional scene model to realize the concentration of multi-camera video objects.
Specifically, in step S5, for the case where the same object trajectory generated by overlapping the camera shooting areas is obtained by multiple cameras, the camera that obtains the object sub-graph is determined by comparing the included angles between the virtual viewpoint and the connecting line between the object trajectory point and the three points at the two camera positions:
specifically referring to fig. 8, a view overlapping part C exists between the camera a and the camera b, for a video object passing through the view C, the included angles between three points, i.e., the sizes of α and β, of the camera position, the track point and the virtual viewpoint V are compared, if α is not greater than β, the video object sub-graph acquired by the camera a is used, otherwise, the video object sub-graph acquired by the camera b is used.
Example 2
Referring to fig. 9, the present invention further provides a multi-camera video concentration system in a geographical scene with a preferred virtual viewpoint, which employs the multi-camera video concentration method in a geographical scene with a preferred virtual viewpoint in embodiment 1 or any possible implementation manner thereof, and the concentration system includes:
the homologous point acquisition module 1: the system comprises a video sequence image acquisition module, a video acquisition module, a three-dimensional geographic scene model acquisition module, a coordinate acquisition module and a display module, wherein the video sequence image acquisition module is used for acquiring video sequence image information, acquiring homonymous point pairs in a video image and the three-dimensional geographic scene model and acquiring coordinate data of the homonymous point pairs, and the coordinate data comprises image coordinates and geographic coordinates;
the mapping relation building module 2: the method comprises the steps of establishing a mapping relation between a video image and a geographic space according to coordinate data of a same-name point pair, and positioning a camera view;
the camera group observable collection generation module 3: the system comprises a camera observation domain model and a camera group observation set, wherein the camera observation domain model is constructed by analyzing an observable distance and a sight line deflection angle to generate a camera group observable set;
virtual viewpoint group generation module 4: the system is used for optimizing the observable collection by constructing an evaluation model to obtain a virtual viewpoint group;
specifically, a video image visual domain model is constructed to describe the range in which each camera in a virtual scene can be effectively observed, then a virtual viewpoint is generated, and the global motion condition of a video motion target in a multi-camera geographic scene is checked; on the basis of a camera observation domain model, exhaustively exhausting a camera observable set, and preferably selecting a combination with the best video target information expression effect as a virtual viewpoint generation region;
the video target space-time motion expression module 5: the method comprises the steps of presetting display parameters of a moving target, and carrying out multi-camera video concentration according to the display parameters.
Specifically, based on the optimal result of the observable set, under the condition that the camera display combination is selected, the central point of the observable domain of each camera group is used as a virtual viewpoint, the moving object display parameters are set, and multi-camera video concentration is performed.
It should be noted that, for the information interaction, execution process and other contents between the modules/units of the multi-camera video concentrating system in the geographic scene with the preferred virtual viewpoint, the same concept is based on as the method embodiment in the embodiment of the present application, and the technical effect brought by the same concept is the same as the method embodiment of the present application, and specific contents can be referred to the description in the foregoing method embodiment of the present application.
Acquiring video sequence image information, acquiring homonymous point pairs in a video image and a three-dimensional geographic scene model, and acquiring coordinate data of the homonymous point pairs, wherein the coordinate data comprises image coordinates and geographic coordinates; establishing a mapping relation between a video image and a geographic space according to coordinate data of the same-name point pair, and positioning a camera view; constructing a camera observation domain model by analyzing the observable distance and the sight line deflection angle, and generating an observable set of a camera group; optimizing the observable set by constructing an evaluation model to obtain a virtual viewpoint group; and presetting display parameters of the moving target, and performing multi-camera video concentration according to the display parameters. The method has the remarkable effects that the mapping relation between the video target and the geographic scene is established, the effect of fusion expression of the monitoring video in the geographic scene is enhanced, and great convenience is provided for the quick retrieval and efficient understanding of the video geographic scene information integration.
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or program product. Thus, various aspects of the invention may be embodied in the form of: an entirely hardware embodiment, an entirely software embodiment (including firmware, microcode, etc.) or an embodiment combining hardware and software aspects that may all generally be referred to herein as a "module" or "platform.
Although the invention has been described in detail above with reference to a general description and specific examples, it will be apparent to one skilled in the art that modifications or improvements may be made thereto based on the invention. Accordingly, such modifications and improvements are intended to be within the scope of the invention as claimed.
Claims (10)
1. A method for concentrating multi-camera video in a geographic scene with a preferred virtual viewpoint is characterized by comprising the following steps:
acquiring video sequence image information, acquiring homonymous point pairs in a video image and a three-dimensional geographic scene model, and acquiring coordinate data of the homonymous point pairs, wherein the coordinate data comprises image coordinates and geographic coordinates;
establishing a mapping relation between a video image and a geographic space according to coordinate data of the same-name point pair, and positioning a camera view;
constructing a camera observation domain model by analyzing the observable distance and the sight line deflection angle, and generating an observable set of a camera group;
optimizing the observable set by constructing an evaluation model to obtain a virtual viewpoint group;
presetting display parameters of the moving target, and performing multi-camera video concentration according to the display parameters.
2. The method for concentrating multi-camera video in geographic scene with preferred virtual viewpoint as claimed in claim 1, wherein the video image is the first frame image of the intercepted surveillance video.
3. The method as claimed in claim 2, wherein the three-dimensional geo-scene model is a three-dimensional scene model constructed according to real geo-scene measurement information, the number of the same-name point pairs collected on the video image and the virtual geo-scene is not less than three, and the three same-name point pairs are not all collinear.
4. The method of claim 3, wherein the step of establishing a mapping relationship between the video images and the geographic space and positioning the camera view comprises the steps of:
a1) presetting an object point geographic space coordinate Q corresponding to a given image point image space coordinate Q, and expressing Q and Q as a homogeneous coordinate form:
q=[x y 1]T
Q=[X Y Z 1]T;
when a homography matrix M is recorded, the relationship between Q and Q is:
q=MQ;
the expression of the homography matrix M is:
a2) solving the geographic space coordinates of the object points corresponding to the image points in each image:
a3) assuming that there are L cameras in the current video camera network, the mapping matrix of the kth camera (k 1, 2.. L) is labeled Mk(ii) a Defining each camera position in geographic space asThe geospatial position of each camera field of view polygon is
5. The method for concentrating multi-camera video in geographic scene with a preferred virtual viewpoint as claimed in claim 4, wherein the process of positioning the camera view selects two factors of virtual line-of-sight distance and camera-virtual viewpoint angle as constraints;
the virtual line-of-sight distance refers to the geographic space distance between a virtual viewpoint and a certain point in the view field; the camera-virtual viewpoint included angle is an included angle formed by projection of a given point in a visual field, the virtual viewpoint and a camera position point on a horizontal plane;
defining a distance threshold TdisAnd an angle threshold TangAs a constraint, and assuming a distance threshold TdisAnd an angle threshold TangIt has been given that finding areas in the scene model that meet the constraints serves as virtual viewpoint ranges.
6. The method for concentrating multiple cameras in a geographic scene with a preferred virtual viewpoint as claimed in claim 5, wherein the step of constructing a camera observation domain model by analyzing an observable distance and a view deflection angle to generate a camera group observable set comprises the following steps:
b1) recording camera positionPolygon with camera viewAnd isIn the distanceIs Pk,n1Pk,n2The farthest line segment is Pk,n3Pk,n4;
b2) Respectively at point Pk,n3,Pk,n4As a circle center, a distance threshold value TdisIs a radius on the line segment Pk,n3Pk,n4Towards the camera positionDrawing a semicircle at one side, and taking two semicircles as a line segment Pk,n1Pk,n2Near camera positionOne side intersection region as a virtual viewpoint distance reasonable region Ak,dis;
b3) Respectively at point Pk,n1,Pk,n2Being a corner point, an angle threshold TangFor deflection angle, clockwise and anticlockwise respectively deflect TangMaking four rays, and taking the four rays to be on the line segment Pk,n1Pk,n2Near camera positionOne side intersection region as a virtual viewpoint angle reasonable region Ak,ang;
b4) Virtual viewpoint range A of camerakIs Ak,disAnd Ak,angThe intersection of (a);
b5) let Obj be the total set of all video moving objects in all video cameras; note that there is N in the kth camerakEach video moving object has a track marked as Ck,i,Ck,iThe expression of (a) is as follows:
Obj={Ck,i,(k=1,2...L)}
Ck,i={Ik,i,j,Pk,i,j,(i=1,2,...,Nk)(j=1,2,...,n)};
wherein L represents the number of cameras, Ik,i,jAnd Pk,i,jRepresenting the sub-image of the ith video moving object in the jth video frame in the kth camera and the geographic space position of the sub-image are analyzed by the cross-camera association of the video moving objects, and the tracks of the single-camera video moving object are merged to obtain the tracks of the multi-camera video moving objectRealizing the related organization of the multi-camera video moving target:
Cubeio={Ck1,i1,Ck2,i2...Cko,iL,(k1,k2,...ko)∈(1,2...L)};
wherein L isoRepresenting the total number of video moving objects, Cube, after merging of cross-camera homonymous video moving objects in a surveillance video networkioRepresenting the global track of the video moving object with the sequence number io in the monitoring video network,and (4) representing the sub-track of the video moving object with the sequence number io in the ko camera.
7. The method for concentrating the multi-camera video in the geographic scene with the preferred virtual viewpoint as claimed in claim 6, wherein the observable set is preferred by constructing an evaluation model to obtain a virtual viewpoint group, specifically comprising the following steps:
c1) the number of cameras is recorded as L, and a set formed by all combination modes of the cameras is recorded as M:
mi={ni,j}
wherein m isiThe ith camera combination mode comprises all camera sets in the combination mode; n isi,jDigital camera combination mode miA j-th group of cameras under the condition that all cameras of the group are included;the first camera in the j camera set in the i camera set combination mode is referred to;
c2) by defining a distance threshold TdisAnd an angle threshold TangAssembling the patterns m for each cameraiIn each camera group ni,jAll cameras inCalculating observable domains and calculating intersection; if a certain camera combination mode miAll the camera sets ni,jIf the intersection of any number of the camera observation domains is not empty, the combination mode m of the camera is recordediIs an observable combination, otherwise, the camera combination mode m is recordediIs an unobservable combination;
c3) based on multi-camera video target trajectory data, the following video condensation optimization targets are specified to realize the optimization of the camera group:
firstly, expressing coherence of homonymous targets across cameras, namely expressing the video cameras with the appearance of a single target in a consistent manner by using as few virtual viewpoints as possible;
expressing that the total number of virtual viewpoints used by all video targets is as small as possible;
c4) comprehensively evaluating the multi-camera video target expression effect of the camera combination corresponding to the virtual viewpoint group by value:
wherein n iscRepresents the total number of cameras, nvRepresenting the number of virtual viewpoints, N representing the total number of video moving objects, miRepresenting the number of virtual viewpoints which are related and expressed by each video moving object, wherein mu is a weight parameter;
c5) at a distance threshold TdisAnd an angle threshold TangWhen the value is constant, value values of all current camera observable sets are calculated through defining a parameter alpha, the maximum value is taken as a camera combination selection result, and multi-camera video concentration in a virtual scene is carried out.
8. The method of claim 7, wherein the method for concentrating the multi-camera video in the geographic scene with the virtual viewpoint being preferred comprises the following steps:
d1) recording that W virtual viewpoints (W is less than or equal to L) are needed for viewing the video moving objects of all L cameras under the current camera combination; meanwhile, setting a frame rate fps displayed by the video moving target subgraph in a three-dimensional scene as the number of the subgraphs displayed by a single video moving target per second; setting an object display interval time t0As time intervals for additional display of new video moving objects;
d2) for a certain virtual viewpoint W (W ≦ W), the first occurring moving object O is first displayed0In geospatial track T0And identifying the sequence of appearance of the video object among different cameras;
screening the video object subgraph according to the frame rate fps, converting the plane coordinate corresponding to the screened video object subgraph into a geographic coordinate, and meanwhile, according to a proportionality coefficient Pw、PhScaling a video object sub-graph, Pw、PhThe calculation formula is as follows:
whereinRandomly selecting an appropriate amount of average width and height of subgraphs from a video object subgraph library, mapping coordinates of three points, namely upper left, lower left and upper right, of the selected subgraphs in an original video frame to corresponding geographic positions in a virtual scene to obtain the length and height of the video object subgraphs in a three-dimensional space,is the average length and height of the video object sub-graph displayed in the virtual scene;
d3) displaying O in the camera view field in the virtual scene according to the frame rate fps in the dynamic display process0The video object sub-graph of the current frame is in the corresponding geographic position, and the old video object sub-graph is not displayed any more;
at t0,2t0...,nt0At the moment, the video object O is added1,O2...OnDynamically expressed in a three-dimensional scene model to realize the concentration of multi-camera video objects.
9. The method according to claim 8, wherein for the case that the same object track generated by overlapping camera shooting areas is obtained by a plurality of cameras, the cameras for obtaining the object subgraph are determined by comparing the included angles between the virtual viewpoint and the object track point and the connecting line of three points at the two camera positions:
and a view overlapping part C exists between the camera a and the camera b, for the video object passing through the view C, the included angle among the three points of the camera position, the track point and the virtual viewpoint V, namely the sizes of alpha and beta, is compared, if alpha is less than or equal to beta, the video object sub-graph acquired by the camera a is used, and otherwise, the video object sub-graph acquired by the camera b is used.
10. A multi-camera video concentration system in a geographical scene with a preferred virtual viewpoint, which adopts the multi-camera video concentration method in a geographical scene with a preferred virtual viewpoint as claimed in any one of claims 1 to 9, wherein the concentration system comprises:
the same-name point acquisition module: the system comprises a video sequence image acquisition module, a video acquisition module, a three-dimensional geographic scene model acquisition module, a coordinate acquisition module and a display module, wherein the video sequence image acquisition module is used for acquiring video sequence image information, acquiring homonymous point pairs in a video image and the three-dimensional geographic scene model and acquiring coordinate data of the homonymous point pairs, and the coordinate data comprises image coordinates and geographic coordinates;
the mapping relation construction module: the method comprises the steps of establishing a mapping relation between a video image and a geographic space according to coordinate data of a same-name point pair, and positioning a camera view;
a camera group observable collection generation module: the system comprises a camera observation domain model and a camera group observation set, wherein the camera observation domain model is constructed by analyzing an observable distance and a sight line deflection angle to generate a camera group observable set;
a virtual viewpoint group generation module: the system is used for optimizing the observable collection by constructing an evaluation model to obtain a virtual viewpoint group;
the video target space-time motion expression module: the method comprises the steps of presetting display parameters of a moving target, and carrying out multi-camera video concentration according to the display parameters.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110327605.1A CN113192125B (en) | 2021-03-26 | 2021-03-26 | Multi-camera video concentration method and system in virtual viewpoint-optimized geographic scene |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110327605.1A CN113192125B (en) | 2021-03-26 | 2021-03-26 | Multi-camera video concentration method and system in virtual viewpoint-optimized geographic scene |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113192125A true CN113192125A (en) | 2021-07-30 |
CN113192125B CN113192125B (en) | 2024-02-20 |
Family
ID=76974146
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110327605.1A Active CN113192125B (en) | 2021-03-26 | 2021-03-26 | Multi-camera video concentration method and system in virtual viewpoint-optimized geographic scene |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113192125B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113840159A (en) * | 2021-09-26 | 2021-12-24 | 北京沃东天骏信息技术有限公司 | Video processing method, device, computer system and readable storage medium |
CN114067071A (en) * | 2021-11-26 | 2022-02-18 | 湖南汽车工程职业学院 | High-precision map making system based on big data |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013033442A1 (en) * | 2011-08-30 | 2013-03-07 | Digimarc Corporation | Methods and arrangements for identifying objects |
US20130088592A1 (en) * | 2011-09-30 | 2013-04-11 | OOO "ITV Group" | Method for searching for objects in video data received from a fixed camera |
CN110009561A (en) * | 2019-04-10 | 2019-07-12 | 南京财经大学 | A method and system for mapping surveillance video target to 3D geographic scene model |
CN110148223A (en) * | 2019-06-03 | 2019-08-20 | 南京财经大学 | Monitor video target concentration expression and system in three-dimensional geography model of place |
CN110516014A (en) * | 2019-01-18 | 2019-11-29 | 南京泛在地理信息产业研究院有限公司 | A method for mapping urban road surveillance video to a two-dimensional map |
CN111161130A (en) * | 2019-11-25 | 2020-05-15 | 北京智汇云舟科技有限公司 | Video correction method based on three-dimensional geographic information |
CN111582022A (en) * | 2020-03-26 | 2020-08-25 | 深圳大学 | A fusion method, system and electronic device of mobile video and geographic scene |
CN112381935A (en) * | 2020-09-29 | 2021-02-19 | 西安应用光学研究所 | Synthetic vision generation and multi-element fusion device |
-
2021
- 2021-03-26 CN CN202110327605.1A patent/CN113192125B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013033442A1 (en) * | 2011-08-30 | 2013-03-07 | Digimarc Corporation | Methods and arrangements for identifying objects |
US20130088592A1 (en) * | 2011-09-30 | 2013-04-11 | OOO "ITV Group" | Method for searching for objects in video data received from a fixed camera |
CN110516014A (en) * | 2019-01-18 | 2019-11-29 | 南京泛在地理信息产业研究院有限公司 | A method for mapping urban road surveillance video to a two-dimensional map |
CN110009561A (en) * | 2019-04-10 | 2019-07-12 | 南京财经大学 | A method and system for mapping surveillance video target to 3D geographic scene model |
CN110148223A (en) * | 2019-06-03 | 2019-08-20 | 南京财经大学 | Monitor video target concentration expression and system in three-dimensional geography model of place |
CN111161130A (en) * | 2019-11-25 | 2020-05-15 | 北京智汇云舟科技有限公司 | Video correction method based on three-dimensional geographic information |
CN111582022A (en) * | 2020-03-26 | 2020-08-25 | 深圳大学 | A fusion method, system and electronic device of mobile video and geographic scene |
CN112381935A (en) * | 2020-09-29 | 2021-02-19 | 西安应用光学研究所 | Synthetic vision generation and multi-element fusion device |
Non-Patent Citations (2)
Title |
---|
JIANQING ZHU; SHENGCAI LIAO; STAN Z. LI: "Multicamera Joint Video Synopsis", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY * |
解愉嘉: "地理场景中监控视频浓缩方法研究", 中国博士学位论文全文数据库 信息科技辑 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113840159A (en) * | 2021-09-26 | 2021-12-24 | 北京沃东天骏信息技术有限公司 | Video processing method, device, computer system and readable storage medium |
CN114067071A (en) * | 2021-11-26 | 2022-02-18 | 湖南汽车工程职业学院 | High-precision map making system based on big data |
CN114067071B (en) * | 2021-11-26 | 2022-08-30 | 湖南汽车工程职业学院 | High-precision map making system based on big data |
Also Published As
Publication number | Publication date |
---|---|
CN113192125B (en) | 2024-02-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Morse et al. | UAV video coverage quality maps and prioritized indexing for wilderness search and rescue | |
US7522186B2 (en) | Method and apparatus for providing immersive surveillance | |
US9189859B2 (en) | 3D image generation | |
RU2612378C1 (en) | Method of replacing objects in video stream | |
CN114693746B (en) | Intelligent monitoring system and method based on identity recognition and cross-camera target tracking | |
US20100103173A1 (en) | Real time object tagging for interactive image display applications | |
US20180075590A1 (en) | Image processing system, image processing method, and program | |
US20220141425A1 (en) | Target Tracking in a Multi-Camera Surveillance System | |
CN114399606A (en) | Interactive display system, method and equipment based on stereoscopic visualization | |
TW200818916A (en) | Wide-area site-based video surveillance system | |
CN101489116A (en) | Video surveillance system and method using ip-based networks | |
US20120027371A1 (en) | Video summarization using video frames from different perspectives | |
JP2006197373A (en) | Viewer information measuring instrument | |
CN113192125B (en) | Multi-camera video concentration method and system in virtual viewpoint-optimized geographic scene | |
WO2008132741A2 (en) | Apparatus and method for tracking human objects and determining attention metrics | |
CN114387679A (en) | System and method for line-of-sight estimation and attention analysis based on recurrent convolutional neural network | |
Cui et al. | Fusing surveillance videos and three‐dimensional scene: A mixed reality system | |
JP4881178B2 (en) | Odometer image generation device and odometer image generation program | |
CN107066975A (en) | Video identification and tracking system and its method based on depth transducer | |
CN111753112B (en) | Information generation method, device and storage medium | |
KR101686797B1 (en) | Method for analyzing a visible area of a closed circuit television considering the three dimensional features | |
den Hollander et al. | Automatic inference of geometric camera parameters and inter-camera topology in uncalibrated disjoint surveillance cameras | |
Zhang et al. | A 3D urban scene reconstruction enhancement approach based on adaptive viewpoint selection of panoramic videos | |
Bagdanov et al. | Acquisition of high-resolution images through on-line saccade sequence planning | |
Hu et al. | Constraints-based graph embedding optimal surveillance-video mosaicing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |