CN111147806A - Video content risk detection method, device and system - Google Patents
Video content risk detection method, device and system Download PDFInfo
- Publication number
- CN111147806A CN111147806A CN201811311966.1A CN201811311966A CN111147806A CN 111147806 A CN111147806 A CN 111147806A CN 201811311966 A CN201811311966 A CN 201811311966A CN 111147806 A CN111147806 A CN 111147806A
- Authority
- CN
- China
- Prior art keywords
- preset
- video data
- analyzed
- user
- camera
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 34
- 238000012806 monitoring device Methods 0.000 claims abstract description 74
- 238000000034 method Methods 0.000 claims abstract description 24
- 238000012549 training Methods 0.000 claims description 43
- 239000000284 extract Substances 0.000 claims description 19
- 238000012544 monitoring process Methods 0.000 description 15
- 230000006399 behavior Effects 0.000 description 9
- 230000005540 biological transmission Effects 0.000 description 4
- 230000007547 defect Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 230000006870 function Effects 0.000 description 3
- 230000033001 locomotion Effects 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 230000003068 static effect Effects 0.000 description 3
- 230000002159 abnormal effect Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 230000005856 abnormality Effects 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44008—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Human Computer Interaction (AREA)
- Image Analysis (AREA)
- Alarm Systems (AREA)
Abstract
The invention discloses a video content risk detection method, a device and a system, wherein the method comprises the following steps: the first camera sends the first video data to the monitoring device; the second camera sends the second video data to the monitoring device; the monitoring device determines a user to be analyzed; extracting a first background feature when a user to be analyzed is located at a first preset position, extracting a second background feature when the user to be analyzed is located at a second preset position, and calculating to obtain the matching degree between the background feature to be analyzed and a preset background collaborative model; extracting a first moment corresponding to the user to be analyzed when the first preset position appears from the first time information, extracting a second moment corresponding to the user to be analyzed when the second preset position appears from the second time information, and calculating the time interval between the first moment and the second moment; and if the matching degree is lower than a preset background threshold value or the time interval is greater than a preset time threshold value, generating a first comparison result and determining that a preset risk exists.
Description
Technical Field
The invention relates to the field of video monitoring, in particular to a video content risk detection method, device and system.
Background
An existing Automatic Teller Machine (ATM for short) is generally arranged in an Automatic bank, and after a bank card is inserted into the ATM, bank counter services such as money withdrawal, deposit, transfer and the like can be performed on the ATM. Due to the public, convenience and environmental specificity of self-service banking and automated teller machines. Criminal activity has increased in recent years for self-service banks and automated teller machines.
However, the conventional ATM video monitoring system mainly records the video, and after an event occurs, the recorded video is subjected to post-event evidence obtaining, so that disputes can be eliminated and cases can be cracked, but such a mechanism only provides a post-event evidence obtaining effect, and cannot achieve real-time or early warning.
Disclosure of Invention
The present invention aims to solve one of the above problems.
The invention mainly aims to provide a video content risk detection method, device and system.
In order to achieve the purpose, the technical scheme of the invention is realized as follows:
one aspect of the present invention provides a video content risk detection method, including: the method comprises the steps that a first camera carries out video acquisition on a first preset position in an environment to be detected to obtain first video data, and the first video data are sent to a monitoring device, wherein the first video data at least comprise first time information, and the first time information cannot be changed; the second camera carries out video acquisition on a second preset position in the environment to be detected to obtain second video data, and sends the second video data to the monitoring device, wherein the second video data at least comprises second time information which cannot be changed, the first camera and the second camera are arranged at different positions in the environment to be detected, and the first time information and the second time information are obtained based on the same reference timing; the monitoring device receives the first video data and the second video data, identifies the face corresponding to the user to be analyzed in the first video data and the second video data, and determines the user to be analyzed; the monitoring device extracts a first background feature of a user to be analyzed when the user to be analyzed is located at a first preset position from the first video data, extracts a second background feature of the user to be analyzed when the user to be analyzed is located at a second preset position from the second video data, and calculates to obtain a matching degree between the background feature to be analyzed and a preset background collaborative model, wherein the background feature to be analyzed comprises the first background feature and the second background feature; the monitoring device extracts a first moment corresponding to the user to be analyzed when the first preset position appears from the first time information, extracts a second moment corresponding to the user to be analyzed when the second preset position appears from the second time information, and calculates the time interval between the first moment and the second moment; the monitoring device compares the matching degree with a preset background threshold, if the matching degree is lower than the preset background threshold, or judges whether the time interval is greater than a preset time threshold, and if the matching degree is greater than the preset time threshold, a first comparison result is generated to determine that a preset risk exists.
Wherein, the method further comprises: and when the matching degree is not lower than a preset background threshold value and the time interval is not greater than a preset time threshold value, the monitoring device generates a second comparison result and determines that no preset risk exists.
Wherein, the method further comprises: the monitoring device receives training video data acquired by the first camera and the second camera in advance; the monitoring device respectively extracts training elements from the training video data, and obtains a preset background collaborative model and a preset time threshold value according to training of the training elements.
Wherein, the method further comprises: and the monitoring device executes alarm operation after determining that the user to be analyzed has the preset risk.
In another aspect, the present invention provides a video content risk detection system, including: the monitoring device comprises a first camera, a second camera and a monitoring device, wherein the first camera is used for carrying out video acquisition on a first preset position in an environment to be detected to obtain first video data and sending the first video data to the monitoring device, and the first video data at least comprises first time information which cannot be changed; the second camera is used for carrying out video acquisition on a second preset position in the environment to be detected, obtaining second video data and sending the second video data to the monitoring device, wherein the second video data at least comprise second time information which cannot be changed, the first camera and the second camera are arranged at different positions in the environment to be detected, and the first time information and the second time information are obtained based on the same reference timing; the monitoring device is used for receiving the first video data and the second video data, identifying the face corresponding to the user to be analyzed in the first video data and the second video data, and determining the user to be analyzed; extracting a first background feature of a user to be analyzed when the user to be analyzed is located at a first preset position from the first video data, extracting a second background feature of the user to be analyzed when the user to be analyzed is located at a second preset position from the second video data, and calculating to obtain a matching degree between the background feature to be analyzed and a preset background collaborative model, wherein the background feature to be analyzed comprises the first background feature and the second background feature; extracting a first moment corresponding to the user to be analyzed when the first preset position appears from the first time information, extracting a second moment corresponding to the user to be analyzed when the second preset position appears from the second time information, and calculating the time interval between the first moment and the second moment; and comparing the matching degree with a preset background threshold, if the matching degree is lower than the preset background threshold, or judging whether the time interval is greater than a preset time threshold, and if the matching degree is greater than the preset time threshold, generating a first comparison result and determining that a preset risk exists.
And the monitoring device is further used for generating a second comparison result when the matching degree is not lower than a preset background threshold and the time interval is not greater than a preset time threshold, and determining that no preset risk exists.
The monitoring device is also used for receiving training video data acquired by the first camera and the second camera in advance; and respectively extracting training elements from the training video data, and training according to the training elements to obtain a preset background collaborative model and a preset time threshold.
The monitoring device is also used for executing alarm operation after determining that the user to be analyzed has the preset risk.
In another aspect, the present invention provides a video content risk detection apparatus, including: the receiving module is used for receiving first video data obtained by a first camera performing video acquisition on a first preset position in an environment to be detected, wherein the first video data at least comprises first time information which cannot be changed, and receiving second video data obtained by a second camera performing video acquisition on a second preset position in the environment to be detected, wherein the second video data at least comprises second time information which cannot be changed, the first camera and the second camera are arranged at different positions in the environment to be detected, and the first time information and the second time information are obtained based on the same reference timing; the identification module is used for identifying the faces corresponding to the users to be analyzed in the first video data and the second video data and determining the users to be analyzed; the computing module is used for extracting a first background feature of a user to be analyzed when the user to be analyzed is located at a first preset position from the first video data, extracting a second background feature of the user to be analyzed when the user to be analyzed is located at a second preset position from the second video data, and computing to obtain a matching degree between the background feature to be analyzed and a preset background collaborative model, wherein the background feature to be analyzed comprises the first background feature and the second background feature; extracting a first moment corresponding to the user to be analyzed when the first preset position appears from the first time information, extracting a second moment corresponding to the user to be analyzed when the second preset position appears from the second time information, and calculating the time interval between the first moment and the second moment; and the judging module is used for comparing the matching degree with a preset background threshold, if the matching degree is lower than the preset background threshold, or judging whether the time interval is greater than a preset time threshold, and if the matching degree is greater than the preset background threshold, generating a first comparison result and determining that a preset risk exists.
The judging module is further configured to generate a second comparison result when the matching degree is not lower than a preset background threshold and the time interval is not greater than a preset time threshold, and it is determined that no preset risk exists.
Therefore, according to the video content risk detection method, device and system provided by the embodiment of the invention, at least two cameras are arranged at different positions, so that the time length required by the user to be analyzed to pass through the first preset position shot by the first camera and the second preset position shot by the second camera is judged, and the background characteristics of the user to be analyzed to pass through the first preset position and the second preset position are analyzed, so that the preset risk (such as illegal criminal intention) can be found in real time, and the defects of precautionary, counterfeiting, fraud and other criminal behaviors which cannot be prevented under the monitoring of a single camera in the past are overcome.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on the drawings without creative efforts.
Fig. 1 is a flowchart of a video content risk detection method according to an embodiment of the present invention;
fig. 2 is a schematic structural diagram of a video content risk detection system according to an embodiment of the present invention;
fig. 3 is a schematic structural diagram of a video content risk detection apparatus according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention are clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
In the description of the present invention, it is to be understood that the terms "center", "longitudinal", "lateral", "up", "down", "front", "back", "left", "right", "vertical", "horizontal", "top", "bottom", "inner", "outer", and the like, indicate orientations or positional relationships based on those shown in the drawings, and are used only for convenience in describing the present invention and for simplicity in description, and do not indicate or imply that the referenced devices or elements must have a particular orientation, be constructed and operated in a particular orientation, and thus, are not to be construed as limiting the present invention. Furthermore, the terms "first," "second," and the like are used for descriptive purposes only and are not to be construed as indicating or implying a relative importance or quantity or location.
In the description of the present invention, it should be noted that, unless otherwise explicitly specified or limited, the terms "mounted," "connected," and "connected" are to be construed broadly, e.g., as meaning either a fixed connection, a removable connection, or an integral connection; can be mechanically or electrically connected; they may be connected directly or indirectly through intervening media, or they may be interconnected between two elements. The specific meanings of the above terms in the present invention can be understood in specific cases to those skilled in the art.
Embodiments of the present invention will be described in further detail below with reference to the accompanying drawings.
Fig. 1 shows a flowchart of a video content risk detection method provided by an embodiment of the present invention, and referring to fig. 1, the video content risk detection method provided by the embodiment of the present invention includes:
s101, a first camera carries out video acquisition on a first preset position in an environment to be detected to obtain first video data, and the first video data are sent to a monitoring device, wherein the first video data at least comprise first time information, and the first time information cannot be changed; the second camera carries out video acquisition on a second preset position in the environment to be detected, second video data are obtained, and the second video data are sent to the monitoring device, wherein the second video data at least comprise second time information, the second time information cannot be changed, the first camera and the second camera are arranged at different positions in the environment to be detected, and the first time information and the second time information are obtained based on the same reference timing.
Specifically, first camera and second camera are the camera that sets up the different positions in waiting to detect the environment, for example when waiting to detect the environment and be self service bank, first camera can be for setting up the camera on the ATM, and the second camera can be for setting up the environment camera in the environment other than ATM in self service bank. Of course, in practical application of the present invention, more than two cameras may be provided, which is not limited in the present invention.
The first camera carries out video acquisition on the first preset position, and the second camera carries out video acquisition on the second preset position, wherein the acquisition is carried out from different positions, and the first camera and the second camera have different background characteristics. The first preset position and the second preset position are positions which are inevitably passed by a user when the user enters the environment to be detected to process the service, and the first preset position and the second preset position can be preset in the embodiment of the invention.
The first camera carries out video acquisition to a first preset position, and the second camera carries out video acquisition to a second preset position, for example, the first camera carries out video acquisition to a self-service bank entrance and exit, and the second camera carries out video acquisition to the environment in front of the ATM, carries out video acquisition through presetting two at least positions, can analyze whether the user action has the abnormality according to the time difference that the user arrives at two positions to be analyzed.
As an optional implementation manner of the embodiment of the present invention, when the first camera and the second camera perform video acquisition, current time information is added to video data, and the current time information cannot be changed, for example, the current time information may be encrypted by using an encryption method to obtain encrypted current time information, or the current time information may be signed by using a signature method to ensure that the current time information cannot be changed.
It should be noted that the first time information and the second time information are all time information contained in the video data, that is, the first time information and the second time information are a time stream formed by a plurality of time instants continuously, for example, recorded continuously in units of seconds.
The first time information and the second time information are obtained based on the same reference time, and both use, for example, internet time, that is, the first time information and the second time information are the same at the same time.
The first video data acquired by the first camera and the second video data acquired by the second camera are sent to the monitoring device in real time, or the acquired video data are sent to the monitoring device at regular time according to a preset period.
S102, the monitoring device receives the first video data and the second video data, identifies faces corresponding to users to be analyzed in the first video data and the second video data, and determines the users to be analyzed.
Specifically, the monitoring device may be disposed near the camera or in the background. For example, in an automated banking environment, the system may be disposed in an ATM or in a bank monitoring background, which is not limited in the present invention. After the monitoring device receives the first video data and the second video data, a user is identified from the first video data by adopting a face recognition technology, a user is identified from the second video data, and when the two users are determined to be the same user, the user is determined to be a user to be analyzed.
S103, the monitoring device extracts a first background feature of the user to be analyzed when the user to be analyzed is located at a first preset position from the first video data, extracts a second background feature of the user to be analyzed when the user to be analyzed is located at a second preset position from the second video data, and calculates to obtain a matching degree between the background feature to be analyzed and a preset background collaborative model, wherein the background feature to be analyzed comprises the first background feature and the second background feature; and the monitoring device extracts a first moment corresponding to the user to be analyzed when the first preset position appears from the first time information, extracts a second moment corresponding to the user to be analyzed when the second preset position appears from the second time information, and calculates the time interval between the first moment and the second moment.
In particular, the background feature may comprise any feature of a background identifier in the environment and any combination thereof to serve as an identifier for the background. For example, the information may include position information of the static object, shape information of the static object, quantity information of the static object, and motion rule of the dynamic object.
Specifically, a background cooperation model is preset in the monitoring device so as to analyze the background characteristics. As an optional implementation manner of the embodiment of the present invention, the monitoring device receives training video data acquired by the first camera and the second camera in advance; the monitoring device respectively extracts training elements from the training video data, and a preset background collaborative model is obtained according to training of the training elements. The background markers in the shooting range of each camera are analyzed to generate a background collaborative model, and a reasonable background threshold range is set according to a first preset position and a second preset position in different movement tracks of a normal user to judge, so that the intelligence and the accuracy of judgment are improved.
Inputting the extracted background features into a preset background collaborative model, and calculating a matching degree between the extracted background features and the background collaborative model, where the matching degree is a numerical value, and may be a percentage value, for example.
The monitoring device acquires first time information from the first video data, determines a first moment of the first time information when the user to be analyzed is at a first preset position, acquires second time information from the second video data, and determines a second moment of the second time information when the user to be analyzed is at a second preset position.
And S104, comparing the matching degree with a preset background threshold value by the monitoring device, if the matching degree is lower than the preset background threshold value, or judging whether the time interval is greater than a preset time threshold value, and if the matching degree is greater than the preset time threshold value, generating a first comparison result and determining that a preset risk exists.
Specifically, when the matching degree is lower than a preset background threshold, the background feature is considered to be not matched with the background collaborative model, and a preset risk may be considered to exist in the case that the background feature is not matched with the background collaborative model, for example: the video with the background features extracted has risks or users to be analyzed have risks, such as tampering of the video, hijacking of a camera, damage to normal collection of the camera by the users, and the like.
When the time length for the user to be analyzed to walk from the first preset position to the second preset position is longer than the preset time threshold, it can be determined that the behavior of the user to be analyzed is abnormal and does not conform to the normal behavior, and therefore the existence of the preset risk can be determined.
As an optional implementation manner of the embodiment of the present invention, the video content risk detection method further includes: the monitoring device receives training video data acquired by the first camera and the second camera in advance; the monitoring device respectively extracts training elements from the training video data, and obtains a preset time threshold value according to training of the training elements. The judgment is carried out by utilizing the preset time threshold obtained by training, and the intelligence and the accuracy of the judgment are improved.
As an optional implementation manner of the embodiment of the present invention, the preset time threshold may be a value or a value interval. Under the condition that the preset time threshold is a numerical value interval, the time interval is larger than the maximum value of the preset time threshold, or the time interval is smaller than the minimum value of the preset time threshold, the detection result can be generated, and the existence of the preset risk is determined. Therefore, the fact that the too long or too short time for the user to be analyzed to walk from the first preset position to the second preset position is abnormal can be judged, and the existence of the preset risk is determined.
It is worth mentioning that in the video content risk detection method, the monitoring device receives training video data acquired by the first camera and the second camera in advance; the monitoring device extracts training elements from the training video data respectively, and the preset background collaborative model and the preset time threshold obtained by training according to the training elements can be obtained in the same training or can be obtained in multiple training respectively, which is not limited in the present invention.
As an optional embodiment of the present invention, when the matching degree is not lower than the preset background threshold and the time interval is not greater than the preset time threshold, the monitoring apparatus generates a second comparison result to determine that there is no preset risk. Since the matching degree between the background features and the background collaborative model is high enough, no risk may be considered to exist, the time interval is not greater than the preset time threshold to conform to normal behavior, and no risk may be considered to exist, for example: the video is not at risk or the user to be analyzed is not at risk.
Optionally, as an optional embodiment of the present invention, the monitoring device performs an alarm operation after determining that the user to be analyzed has a preset risk. The alarm operation can be that an alarm device in the environment to be detected gives an alarm, for example, by sound and light, or an alarm device in a monitoring room of a background monitoring person, for example, by displaying on a monitoring display screen to give an alarm or sound, or sending a short message to a monitoring person or a policeman, or the like. The efficiency of risk handling for self-service banking and ATMs is further improved by alarming when a risk occurs.
Therefore, according to the video content risk detection method provided by the embodiment of the invention, at least two cameras are arranged at different positions, so that the time length required by the user to be analyzed to pass through the first preset position shot by the first camera and the second preset position shot by the second camera is judged, and the background characteristics of the user to be analyzed to pass through the first preset position and the second preset position are analyzed, so that the preset risk (such as illegal criminal intention) can be found in real time, and the defects of precautionary, intentional forgery, fake and other criminal behaviors under the monitoring of the single camera in the past are overcome.
As an optional embodiment of the present invention, first video data collected by a first camera is encrypted by a security chip disposed in the first camera, second video data collected by a second camera is encrypted by a security chip disposed in the second camera, the first camera sends the encrypted first video data to a monitoring device, and the second camera sends the encrypted second video data to the monitoring device; and after receiving the encrypted first video data and the encrypted second video data, the monitoring device decrypts the encrypted first video data and the encrypted second video data to obtain the first video data and the second video data. By carrying out encryption transmission on the video data, the security of video data transmission is improved, and the video data is prevented from being tampered after being cracked.
The method comprises the steps that first video data collected by a first camera are signed through a security chip arranged in the first camera to obtain first signature data, second video data collected by a second camera are signed through a security chip arranged in the second camera to obtain second signature data, the first camera sends the first video data and the first signature data to a monitoring device, and the second camera sends the second video data and the second signature data to the monitoring device; and after receiving the first video data, the first signature data, the second video data and the second signature data, the monitoring device checks the first signature data and the second signature data, and uses the first video data and the second video data to perform subsequent analysis after the first signature data and the second signature data pass the checking. By signing the video data, the authenticity of the video data source can be ensured, and the video data can be prevented from being tampered.
Fig. 2 is a schematic structural diagram of a video content risk detection system according to an embodiment of the present invention, where the video content risk detection system according to the embodiment of the present invention applies the above method, and only the structure of the video content risk detection system according to the embodiment of the present invention is briefly described below, and other things are not considered to be the best, with reference to the related description of the above video content risk detection method, referring to fig. 2, the video content risk detection system according to the embodiment of the present invention includes:
the first camera 201 is configured to perform video acquisition on a first preset position in an environment to be detected, obtain first video data, and send the first video data to the monitoring device, where the first video data at least includes first time information, and the first time information is unchangeable;
the second camera 202 is configured to perform video acquisition on a second preset position in the environment to be detected, obtain second video data, and send the second video data to the monitoring device, where the second video data at least includes second time information, the second time information is not changeable, the first camera and the second camera are disposed at different positions in the environment to be detected, and the first time information and the second time information are obtained based on the same reference timing;
the monitoring device 203 is used for receiving the first video data and the second video data, identifying faces corresponding to users to be analyzed in the first video data and the second video data, and determining the users to be analyzed; extracting a first background feature of a user to be analyzed when the user to be analyzed is located at a first preset position from the first video data, extracting a second background feature of the user to be analyzed when the user to be analyzed is located at a second preset position from the second video data, and calculating to obtain a matching degree between the background feature to be analyzed and a preset background collaborative model, wherein the background feature to be analyzed comprises the first background feature and the second background feature; extracting a first moment corresponding to the user to be analyzed when the first preset position appears from the first time information, extracting a second moment corresponding to the user to be analyzed when the second preset position appears from the second time information, and calculating the time interval between the first moment and the second moment; and comparing the matching degree with a preset background threshold, if the matching degree is lower than the preset background threshold, or judging whether the time interval is greater than a preset time threshold, and if the matching degree is greater than the preset time threshold, generating a first comparison result and determining that a preset risk exists.
Therefore, according to the video content risk detection system provided by the embodiment of the invention, at least two cameras are arranged at different positions, so that the time length required by the user to be analyzed to pass through the first preset position shot by the first camera and the second preset position shot by the second camera is judged, and the background characteristics of the user to be analyzed to pass through the first preset position and the second preset position are analyzed, so that the preset risk (such as illegal criminal intention) can be found in real time, and the defects of precautionary, intentional forgery, fake and other criminal behaviors under the monitoring of the single camera in the past are overcome.
As an optional embodiment of the present invention, the monitoring device 203 is further configured to generate a second comparison result when the matching degree is not lower than the preset background threshold and the time interval is not greater than the preset time threshold, so as to determine that there is no preset risk. Since the matching degree between the background features and the background collaborative model is high enough, no risk may be considered to exist, the time interval is not greater than the preset time threshold to conform to normal behavior, and no risk may be considered to exist, for example: the video is not at risk or the user to be analyzed is not at risk.
As an optional embodiment of the present invention, the monitoring device 203 is further configured to receive training video data acquired by the first camera and the second camera in advance; and respectively extracting training elements from the training video data, and training according to the training elements to obtain a preset background collaborative model and a preset time threshold. The background markers in the shooting range of each camera are analyzed to generate a background collaborative model, a reasonable background threshold range is set according to a first preset position and a second preset position in different movement tracks of a normal user to judge, and a preset time threshold obtained by training is utilized to judge, so that the intelligence and the accuracy of judgment are improved.
An alternative embodiment of the present invention is characterized in that the monitoring device 203 is further configured to perform an alarm operation after determining that the user to be analyzed has a preset risk. The efficiency of risk handling for self-service banking and ATMs is further improved by alarming when a risk occurs.
As an optional embodiment of the present invention, first video data collected by the first camera 201 is encrypted by a security chip disposed in the first camera, second video data collected by the second camera 202 is encrypted by a security chip disposed in the second camera, the first camera 201 sends the encrypted first video data to the monitoring apparatus, and the second camera 202 sends the encrypted second video data to the monitoring apparatus 203; after receiving the encrypted first video data and the encrypted second video data, the monitoring device 203 decrypts the encrypted first video data and the encrypted second video data to obtain the first video data and the second video data. By carrying out encryption transmission on the video data, the security of video data transmission is improved, and the video data is prevented from being tampered after being cracked.
A first video data collected by a first camera 201 is signed by a security chip arranged in the first camera to obtain a first signature data, a second video data collected by a second camera 202 is signed by a security chip arranged in the second camera to obtain a second signature data, the first camera 201 sends the first video data and the first signature data to a monitoring device, and the second camera 202 sends the second video data and the second signature data to a monitoring device 203; after receiving the first video data and the first signature data, and the second video data and the second signature data, the monitoring device 203 checks the first signature data and the second signature data, and performs subsequent analysis using the first video data and the second video data after the check passes. By signing the video data, the authenticity of the video data source can be ensured, and the video data can be prevented from being tampered.
On the basis of fig. 2, fig. 3 shows a schematic structural diagram of a video content risk detection apparatus provided in an embodiment of the present invention, where the video content risk detection apparatus is a monitoring apparatus in the system shown in fig. 2, and the video content risk detection apparatus provided in an embodiment of the present invention applies the above system and method, and only the structure of the video content risk detection apparatus provided in an embodiment of the present invention is briefly described below, and other things are not to the utmost, with reference to the above related description of the video content risk detection system and method, see fig. 3, the video content risk detection apparatus provided in an embodiment of the present invention includes:
a receiving module 2031, configured to receive first video data obtained by a first camera performing video acquisition on a first preset position in an environment to be detected, where the first video data at least includes first time information that is not changeable, and receive second video data obtained by a second camera performing video acquisition on a second preset position in the environment to be detected, where the second video data at least includes second time information that is not changeable, the first camera and the second camera are disposed at different positions in the environment to be detected, and the first time information and the second time information are obtained based on the same reference timing;
the identifying module 2032 is configured to identify faces corresponding to users to be analyzed in the first video data and the second video data, and determine the users to be analyzed;
a calculating module 2033, configured to extract, from the first video data, a first background feature when the user to be analyzed is located at a first preset position, extract, from the second video data, a second background feature when the user to be analyzed is located at a second preset position, and calculate to obtain a matching degree between the background feature to be analyzed and the preset background collaborative model, where the background feature to be analyzed includes the first background feature and the second background feature; extracting a first moment corresponding to the user to be analyzed when the first preset position appears from the first time information, extracting a second moment corresponding to the user to be analyzed when the second preset position appears from the second time information, and calculating the time interval between the first moment and the second moment;
the determining module 2034 is configured to compare the matching degree with a preset background threshold, and if the matching degree is lower than the preset background threshold, or determine whether the time interval is greater than a preset time threshold, and if the time interval is greater than the preset time threshold, generate a first comparison result and determine that a preset risk exists.
Therefore, according to the video content risk detection device provided by the embodiment of the invention, at least two cameras are arranged at different positions, so that the time length required by the user to be analyzed to pass through the first preset position shot by the first camera and the second preset position shot by the second camera is judged, and the background characteristics of the user to be analyzed to pass through the first preset position and the second preset position are analyzed, so that the preset risk (such as illegal criminal intention) can be found in real time, and the defects of precautionary, intentional forgery, fake and other criminal behaviors under the monitoring of the single camera in the past are overcome.
As an optional implementation manner of the embodiment of the present invention, the determining module 2034 is further configured to generate a second comparison result when the matching degree is not lower than the preset background threshold and the time interval is not greater than the preset time threshold, so as to determine that there is no preset risk. Since the matching degree between the background features and the background collaborative model is high enough, no risk may be considered to exist, the time interval is not greater than the preset time threshold to conform to normal behavior, and no risk may be considered to exist, for example: the video is not at risk or the user to be analyzed is not at risk.
Any process or method descriptions in flow charts or otherwise described herein may be understood as representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps of the process, and alternate implementations are included within the scope of the preferred embodiment of the present invention in which functions may be executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved, as would be understood by those reasonably skilled in the art of the present invention.
It should be understood that portions of the present invention may be implemented in hardware, software, firmware, or a combination thereof. In the above embodiments, the various steps or methods may be implemented in software or firmware stored in memory and executed by a suitable instruction execution system. For example, if implemented in hardware, as in another embodiment, any one or combination of the following techniques, which are known in the art, may be used: a discrete logic circuit having a logic gate circuit for implementing a logic function on a data signal, an application specific integrated circuit having an appropriate combinational logic gate circuit, a Programmable Gate Array (PGA), a Field Programmable Gate Array (FPGA), or the like.
It will be understood by those skilled in the art that all or part of the steps carried by the method for implementing the above embodiments may be implemented by hardware that is related to instructions of a program, and the program may be stored in a computer-readable storage medium, and when executed, the program includes one or a combination of the steps of the method embodiments.
In addition, functional units in the embodiments of the present invention may be integrated into one processing module, or each unit may exist alone physically, or two or more units are integrated into one module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode. The integrated module, if implemented in the form of a software functional module and sold or used as a separate product, may also be stored in a computer readable storage medium.
The storage medium mentioned above may be a read-only memory, a magnetic or optical disk, etc.
In the description herein, references to the description of the term "one embodiment," "some embodiments," "an example," "a specific example," or "some examples," etc., mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the invention. In this specification, the schematic representations of the terms used above do not necessarily refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.
Although embodiments of the present invention have been shown and described above, it is understood that the above embodiments are exemplary and should not be construed as limiting the present invention, and that variations, modifications, substitutions and alterations can be made in the above embodiments by those of ordinary skill in the art without departing from the principle and spirit of the present invention. The scope of the invention is defined by the appended claims and equivalents thereof.
Claims (10)
1. A video content risk detection method is characterized by comprising the following steps:
the method comprises the steps that a first camera carries out video acquisition on a first preset position in an environment to be detected to obtain first video data, and the first video data are sent to a monitoring device, wherein the first video data at least comprise first time information, and the first time information cannot be changed;
the method comprises the steps that a second camera carries out video acquisition on a second preset position in an environment to be detected to obtain second video data, and the second video data are sent to a monitoring device, wherein the second video data at least comprise second time information which cannot be changed, the first camera and the second camera are arranged at different positions in the environment to be detected, and the first time information and the second time information are obtained based on the same reference timing;
the monitoring device receives the first video data and the second video data, identifies a face corresponding to a user to be analyzed in the first video data and the second video data, and determines the user to be analyzed;
the monitoring device extracts a first background feature of the user to be analyzed when the user to be analyzed is located at the first preset position from the first video data, extracts a second background feature of the user to be analyzed when the user to be analyzed is located at the second preset position from the second video data, and calculates to obtain a matching degree between the background feature to be analyzed and a preset background collaborative model, wherein the background feature to be analyzed comprises the first background feature and the second background feature;
and
the monitoring device extracts a first moment corresponding to the user to be analyzed when the first preset position appears from the first time information, extracts a second moment corresponding to the user to be analyzed when the second preset position appears from the second time information, and calculates a time interval between the first moment and the second moment;
and the monitoring device compares the matching degree with a preset background threshold, if the matching degree is lower than the preset background threshold, or judges whether the time interval is greater than a preset time threshold, and if the matching degree is greater than the preset background threshold, a first comparison result is generated to determine that a preset risk exists.
2. The method of claim 1, further comprising:
and when the matching degree is not lower than the preset background threshold value and the time interval is not greater than the preset time threshold value, the monitoring device generates a second comparison result and determines that no preset risk exists.
3. The method of claim 1 or 2, further comprising:
the monitoring device receives training video data acquired by the first camera and the second camera in advance;
the monitoring device respectively extracts training elements from the training video data, and trains according to the training elements to obtain the preset background collaborative model and the preset time threshold.
4. The method of claim 1 or 2, further comprising:
and the monitoring device executes alarm operation after determining that the user to be analyzed has the preset risk.
5. A video content risk detection system, comprising:
the monitoring device comprises a first camera, a second camera and a monitoring device, wherein the first camera is used for carrying out video acquisition on a first preset position in an environment to be detected to obtain first video data and sending the first video data to the monitoring device, and the first video data at least comprises first time information which cannot be changed;
the second camera is used for carrying out video acquisition on a second preset position in the environment to be detected, obtaining second video data and sending the second video data to the monitoring device, wherein the second video data at least comprise second time information which cannot be changed, the first camera and the second camera are arranged at different positions in the environment to be detected, and the first time information and the second time information are obtained based on the same reference timing;
the monitoring device is used for receiving the first video data and the second video data, identifying faces corresponding to users to be analyzed in the first video data and the second video data, and determining the users to be analyzed; extracting a first background feature of the user to be analyzed when the user to be analyzed is located at the first preset position from the first video data, extracting a second background feature of the user to be analyzed when the user to be analyzed is located at the second preset position from the second video data, and calculating to obtain a matching degree between the background feature to be analyzed and a preset background collaborative model, wherein the background feature to be analyzed comprises the first background feature and the second background feature; extracting a first time corresponding to the user to be analyzed when the first preset position appears from the first time information, extracting a second time corresponding to the user to be analyzed when the second preset position appears from the second time information, and calculating a time interval between the first time and the second time; and comparing the matching degree with a preset background threshold, if the matching degree is lower than the preset background threshold, or judging whether the time interval is greater than a preset time threshold, and if the matching degree is greater than the preset background threshold, generating a first comparison result and determining that a preset risk exists.
6. The system according to claim 5, wherein the monitoring device is further configured to generate a second comparison result when the matching degree is not lower than the preset background threshold and the time interval is not greater than the preset time threshold, so as to determine that there is no preset risk.
7. The system according to claim 5 or 6, wherein the monitoring device is further configured to receive training video data acquired by the first camera and the second camera in advance; and respectively extracting training elements from the training video data, and training according to the training elements to obtain the preset background collaborative model and the preset time threshold.
8. The system according to claim 5 or 6, wherein the monitoring device is further configured to perform an alarm operation after determining that the user to be analyzed has a preset risk.
9. A video content risk detection apparatus, comprising:
the receiving module is used for receiving first video data obtained by a first camera performing video acquisition on a first preset position in an environment to be detected, wherein the first video data at least comprises first time information which cannot be changed, and receiving second video data obtained by a second camera performing video acquisition on a second preset position in the environment to be detected, wherein the second video data at least comprises second time information which cannot be changed, the first camera and the second camera are arranged at different positions in the environment to be detected, and the first time information and the second time information are obtained based on the same reference timing;
the identification module is used for identifying the faces corresponding to the users to be analyzed in the first video data and the second video data and determining the users to be analyzed;
a calculating module, configured to extract, from the first video data, a first background feature when the user to be analyzed is located at the first preset position, extract, from the second video data, a second background feature when the user to be analyzed is located at the second preset position, and calculate a matching degree between the background feature to be analyzed and a preset background collaborative model, where the background feature to be analyzed includes the first background feature and the second background feature; extracting a first time corresponding to the user to be analyzed when the first preset position appears from the first time information, extracting a second time corresponding to the user to be analyzed when the second preset position appears from the second time information, and calculating a time interval between the first time and the second time;
and the judging module is used for comparing the matching degree with a preset background threshold, if the matching degree is lower than the preset background threshold, or judging whether the time interval is greater than a preset time threshold, and if the matching degree is greater than the preset background threshold, generating a first comparison result and determining that a preset risk exists.
10. The apparatus according to claim 9, wherein the determining module is further configured to generate a second comparison result when the matching degree is not lower than the preset background threshold and the time interval is not greater than the preset time threshold, so as to determine that there is no preset risk.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811311966.1A CN111147806A (en) | 2018-11-06 | 2018-11-06 | Video content risk detection method, device and system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811311966.1A CN111147806A (en) | 2018-11-06 | 2018-11-06 | Video content risk detection method, device and system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111147806A true CN111147806A (en) | 2020-05-12 |
Family
ID=70515861
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811311966.1A Pending CN111147806A (en) | 2018-11-06 | 2018-11-06 | Video content risk detection method, device and system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111147806A (en) |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1529506A (en) * | 2003-09-29 | 2004-09-15 | �Ϻ���ͨ��ѧ | Video Object Segmentation Method Based on Motion Detection |
CN101246547A (en) * | 2008-03-03 | 2008-08-20 | 北京航空航天大学 | A Method for Detecting Moving Objects in Video Based on Scene Change Features |
CN101426080A (en) * | 2007-10-29 | 2009-05-06 | 三星电子株式会社 | Device and method for detecting and suppressing influence generated by camera moving in monitoring system |
CN102096977A (en) * | 2010-11-26 | 2011-06-15 | 上海电力带电作业技术开发有限公司 | Method for video monitoring and prewarning of intrusive external force |
CN103841367A (en) * | 2012-11-21 | 2014-06-04 | 深圳市赛格导航科技股份有限公司 | Monitoring system |
CN105761261A (en) * | 2016-02-17 | 2016-07-13 | 南京工程学院 | Method for detecting artificial malicious damage to camera |
CN106652376A (en) * | 2016-12-02 | 2017-05-10 | 重庆软汇科技股份有限公司 | Positioning monitor method and system |
CN108154518A (en) * | 2017-12-11 | 2018-06-12 | 广州华多网络科技有限公司 | A kind of method, apparatus of image procossing, storage medium and electronic equipment |
-
2018
- 2018-11-06 CN CN201811311966.1A patent/CN111147806A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1529506A (en) * | 2003-09-29 | 2004-09-15 | �Ϻ���ͨ��ѧ | Video Object Segmentation Method Based on Motion Detection |
CN101426080A (en) * | 2007-10-29 | 2009-05-06 | 三星电子株式会社 | Device and method for detecting and suppressing influence generated by camera moving in monitoring system |
CN101246547A (en) * | 2008-03-03 | 2008-08-20 | 北京航空航天大学 | A Method for Detecting Moving Objects in Video Based on Scene Change Features |
CN102096977A (en) * | 2010-11-26 | 2011-06-15 | 上海电力带电作业技术开发有限公司 | Method for video monitoring and prewarning of intrusive external force |
CN103841367A (en) * | 2012-11-21 | 2014-06-04 | 深圳市赛格导航科技股份有限公司 | Monitoring system |
CN105761261A (en) * | 2016-02-17 | 2016-07-13 | 南京工程学院 | Method for detecting artificial malicious damage to camera |
CN106652376A (en) * | 2016-12-02 | 2017-05-10 | 重庆软汇科技股份有限公司 | Positioning monitor method and system |
CN108154518A (en) * | 2017-12-11 | 2018-06-12 | 广州华多网络科技有限公司 | A kind of method, apparatus of image procossing, storage medium and electronic equipment |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10477156B2 (en) | Video analytics system | |
CN101266704B (en) | ATM secure authentication and pre-alarming method based on face recognition | |
TWI776796B (en) | Financial terminal security system and financial terminal security method | |
CN102004904B (en) | Automatic teller machine-based safe monitoring device and method and automatic teller machine | |
CN103714631A (en) | ATM intelligent monitoring system based on human face recognition | |
CN101609581A (en) | The anomalous video warning device of ATM | |
CN101201955A (en) | Intelligent Monitoring System and Method for ATM | |
CN109961587A (en) | A kind of monitoring system of self-service bank | |
US7552865B2 (en) | System and method for deep interaction modeling for fraud detection | |
US20230252803A1 (en) | Hand action tracking for card slot tampering | |
CN101692281A (en) | Safety monitoring method, safety monitoring device and automatic teller machine (ATM) system | |
CN111144181A (en) | Risk detection method, device and system based on background collaboration | |
TWI671701B (en) | System and method for detecting trading behavior | |
JP6935919B2 (en) | Automatic teller machines, terminal devices, control methods, control programs | |
CN113537034A (en) | Cash receiving loss prevention method and system | |
CN111145456A (en) | Risk detection method, device and system based on background collaboration | |
CN111147806A (en) | Video content risk detection method, device and system | |
US7451919B2 (en) | Self-service terminal | |
CN117893972A (en) | Abnormal behavior detection method based on video recording | |
CN111144180B (en) | Risk detection method and system for monitoring video | |
CN111144182B (en) | Method and system for detecting face risk in video | |
CN111147807A (en) | Risk detection method, device and system based on information synchronization | |
CN111144183B (en) | Risk detection method, device and system based on face concave-convex degree | |
CN109961588A (en) | A kind of monitoring system | |
CN111145455A (en) | Method and system for detecting face risk in surveillance video |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20200512 |
|
RJ01 | Rejection of invention patent application after publication |