WO2017156793A1 - Geographic location-based video processing method - Google Patents
Geographic location-based video processing method Download PDFInfo
- Publication number
- WO2017156793A1 WO2017156793A1 PCT/CN2016/077182 CN2016077182W WO2017156793A1 WO 2017156793 A1 WO2017156793 A1 WO 2017156793A1 CN 2016077182 W CN2016077182 W CN 2016077182W WO 2017156793 A1 WO2017156793 A1 WO 2017156793A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- additional information
- video data
- geographic location
- information
- Prior art date
Links
- 238000003672 processing method Methods 0.000 title claims abstract description 9
- 238000000034 method Methods 0.000 claims abstract description 43
- 230000008569 process Effects 0.000 claims abstract description 14
- 238000012545 processing Methods 0.000 claims description 31
- 238000012795 verification Methods 0.000 claims description 7
- 230000006855 networking Effects 0.000 claims description 2
- 238000007619 statistical method Methods 0.000 claims description 2
- 230000010365 information processing Effects 0.000 description 12
- 230000006870 function Effects 0.000 description 11
- 238000010586 diagram Methods 0.000 description 9
- 238000005516 engineering process Methods 0.000 description 8
- 238000004891 communication Methods 0.000 description 7
- 230000002093 peripheral effect Effects 0.000 description 3
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 238000005065 mining Methods 0.000 description 2
- 238000010295 mobile communication Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 206010070834 Sensitisation Diseases 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 230000005611 electricity Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000008313 sensitization Effects 0.000 description 1
- 239000004984 smart glass Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/63—Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
- H04N21/647—Control signaling between network components and server or clients; Network processes for video distribution between server and clients, e.g. controlling the quality of the video stream, by dropping packets, protecting content from unauthorised alteration within the network, monitoring of network load, bridging between two different networks, e.g. between IP and wireless
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/62—Control of parameters via user interfaces
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/7867—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/00127—Connection or combination of a still picture apparatus with another apparatus, e.g. for storage, processing or transmission of still picture signals or of information associated with a still picture
- H04N1/00249—Connection or combination of a still picture apparatus with another apparatus, e.g. for storage, processing or transmission of still picture signals or of information associated with a still picture with a photographic apparatus, e.g. a photographic printer or a projector
Definitions
- the present invention relates to video processing technologies, and in particular, to a geographic location based video processing technology.
- a geographic location-based video processing method is applied to a mobile shooting terminal, the mobile shooting terminal includes a camera, and the method includes:
- the additional information includes at least current positioning information of the mobile camera terminal
- the method further includes: detecting a rotation angle of the camera in real time during shooting; the additional information further includes a rotation angle of the camera.
- the method further includes: acquiring user identity information of the photographer corresponding to the current video frame; the additional information further includes the user identity information.
- the method further includes: acquiring a shooting time of the current video frame in real time during shooting; the additional information further includes the shooting time.
- the method further includes: obtaining an instruction input by a user; the additional information further including the shooting time.
- the method further includes receiving a label and/or textual input with the user; the additional information further including the label and/or textual introduction.
- the above method further comprises: encrypting at least part of the content of the additional information, the additional information including at least the content processed by the encryption.
- the method further includes: calculating corresponding verification information according to the video data of the current frame; and the verification information is included in the encrypted processed content.
- a geographic location-based video processing method applied to a cloud server system including:
- the additional information includes at least the current positioning information of the mobile shooting terminal
- the method further includes: parsing out from the video data packet An instruction corresponding to each video frame; and executing a data processing procedure corresponding to the instruction.
- the data processing process includes:
- the method further includes: parsing a label and/or a text description of the video from the video data package;
- each video frame of the video data includes positioning information, and may further include other additional information, and based on the additional information, the video data can be subjected to geo-location-based retrieval, thereby being able to perform on the basis of Provide a rich video application.
- FIG. 1 is a schematic structural diagram of a geographic location-based video information processing system according to an embodiment of the present invention.
- FIG. 2 is a block diagram showing the structure of a mobile photographing terminal of the video information processing system of FIG. 1.
- FIG. 3 is a schematic diagram showing the data structure of video data uploaded by the mobile camera terminal of FIG. 2.
- FIG. 4 is a schematic flow chart of processing video data uploaded by a mobile camera terminal by the video information processing system of FIG. 1.
- FIG. 5 is a schematic diagram of an interface of an application provided by the video information processing system of FIG. 1 based on video data with geographic locations.
- FIG. 6 is a schematic diagram of an interface of another application provided by the video information processing system of FIG. 1 based on video data with geographic location.
- FIG. 7 is a schematic flow chart of the video information processing system of FIG. 1 providing a panoramic preview function based on video data with geographic locations.
- FIG. 8 is a flow chart showing the video information processing system of FIG. 1 providing a video search service based on video data with geographic location.
- FIG. 9 is an expanded schematic diagram of the video information processing system of FIG. 1.
- FIG. 10 is a flow diagram of the video information processing system of FIG. 1 providing a video application based on video data with geographic location, tags, and/or textual introduction.
- FIG. 1 is a schematic structural diagram of a geographic location-based video information processing system according to a first embodiment of the present invention.
- the video information processing system 100 may include a mobile photography terminal 10, a cloud server system 20, and a client 30.
- the mobile photographing terminal 10 may be any mobile electronic terminal with a camera such as a mobile phone, a tablet computer, a drone, or the like.
- FIG. 2 it is a schematic structural diagram of the mobile photographing terminal 10.
- the mobile photographing terminal 10 includes a memory 102, a memory controller 104, and one or more (only one shown in the figure) processor 106, peripheral interface 108, network module 110, audio circuit 111, GPS (Global Positioning System) module 112, sensor 114, camera module 116, and power module 122 . These components communicate with one another via one or more communication bus/signal lines.
- GPS Global Positioning System
- the structure shown in FIG. 2 is merely illustrative and does not limit the structure of the mobile camera terminal 10.
- the mobile photography terminal 10 may also include more or less components than those shown in FIG. 2, or have a different configuration than that shown in FIG. 2.
- the memory 102 can be used to store software programs and modules, such as program instructions/modules corresponding to the methods and devices in the embodiments of the present invention.
- the processor 106 executes various functional applications by running software programs and modules stored in the memory 102. And data processing.
- Memory 102 can include high speed random access memory and can also include non-volatile memory, such as one or more magnetic storage devices, flash memory, or other non-volatile solid state memory.
- memory 102 can further include memory remotely located relative to processor 106, which can be connected to the server via a network. Examples of such networks include, but are not limited to, the Internet, intranets, local area networks, mobile communication networks, and combinations thereof. Access to the memory 102 by the processor 106 and other possible components can be performed under the control of the memory controller 104.
- Peripheral interface 108 couples various input/input devices to processor 106.
- the processor 106 runs various software within the memory 102, instructs the server to perform various functions, and performs data processing.
- peripheral interface 108, processor 106, and memory controller 104 can be implemented in a single chip. In other instances, they can be implemented by separate chips.
- the network module 110 is configured to receive and transmit network signals.
- the network signal described above may include a wireless signal.
- the network module 110 is substantially a radio frequency module that receives and transmits electromagnetic waves, and converts electromagnetic waves and electrical signals, thereby interacting with a communication network or other devices. Line communication.
- the radio frequency module can include various existing circuit components for performing these functions, such as an antenna, a radio frequency transceiver, a digital signal processor, an encryption/decryption chip, a Subscriber Identity Module (SIM) card, a memory, and the like.
- SIM Subscriber Identity Module
- the RF module can communicate with various networks such as the Internet, intranets, wireless networks or with other devices over a wireless network.
- the wireless network described above may include a cellular telephone network, a wireless local area network, or a metropolitan area network.
- the above wireless network can use various communication standards, protocols and technologies, including but not limited to Global System for Mobile Communication (GSM), Enhanced Data GSM Environment (EDGE), and wideband code.
- GSM Global System for Mobile Communication
- EDGE Enhanced Data GSM Environment
- wideband code wideband code
- W-CDMA Wideband code division multiple access
- CDMA Code division access
- TDMA Time Division Multiple Access
- WiFi Wireless Fidelity
- VoIP Voice Over Internet Protocol
- Wi-Max other protocols for mail, instant messaging, and short messages, as well as any other suitable communication protocol, may even include protocols that are not currently being developed.
- the audio circuit 111 is provided with an interface for providing recording of the mobile photographing terminal 10. Specifically, the audio circuit 111 receives an electrical signal from the microphone, converts the electrical signal into sound data, and transmits the sound data to the processor 102 for further processing.
- the GPS module 112 is configured to receive a positioning signal of a GPS satellite broadcast, and calculate its own position according to the positioning signal.
- the above positions can be expressed, for example, by longitude, latitude, and altitude. It will be appreciated that the manner in which positioning is achieved is not limited to GPS systems.
- other available satellite positioning systems include the Compass Navigation Satellite System (CNSS) or the Global Navigation Satellite System (GLONASS).
- positioning is not limited to satellite positioning technology, for example, Wireless positioning technology, such as positioning technology based on wireless base stations or positioning technology of WIFI.
- the GPS module 112 can be replaced with a corresponding module, or directly by the processor 102 executing a particular positioning procedure.
- sensors 114 include, but are not limited to, light sensors, attitude sensors, and other sensors.
- the ambient light sensor can sense the brightness of the ambient light, and thus can adjust the shooting.
- the attitude sensor may include, for example, an acceleration sensor, a gravimeter, a gyroscope, or the like, which can detect a spatial attitude of the moving photographing terminal 10, for example, a rotation angle in various directions, and the like. It can be understood that the rotation angle of the moving photographing terminal 10 in each direction corresponds to the photographing direction.
- Other sensors may include barometers, hygrometers, thermometers, and the like.
- the camera module 116 is used to take a photo or video.
- the photographed or video taken may be stored in the memory 104 and may be transmitted through the network module 110.
- the camera module 116 may specifically include components such as a lens module, an image sensor, and a flash.
- the lens module is used to image the captured object and map the formed image to the image sensor.
- the image sensor is used to receive light from the lens module to achieve sensitization to record image information.
- the image sensor can be implemented based on a Complementary Metal Oxide Semiconductor (CMOS), a Charge-coupled Device (CCD), or other image sensing principle.
- CMOS Complementary Metal Oxide Semiconductor
- CCD Charge-coupled Device
- the flash is used to compensate for exposure when shooting.
- the flash for moving the photographing terminal 10 may be a Light Emitting Diode (LED) flash.
- LED Light Emitting Diode
- the power module 122 is used to provide power to the processor 102 and other components.
- the power module 122 may include a power management system, one or more power sources (such as a battery or an alternating current), a charging circuit, a power failure detecting circuit, an inverter, a power status indicator, and any other power in the mobile shooting terminal 10. The generation, management, and distribution of related components.
- the software and program modules stored in the memory 104 may include an operating system 130 and an application running on the operating system 130.
- the operating system 130 can include various management systems Software components and/or drivers (such as memory management, storage device control, power management, etc.) that communicate with various hardware or software components to provide an operating environment for other software components.
- the application program may include: a shooting module 131, an additional information adding module 132, a video data encapsulating module 133, and a data sending module 134.
- the shooting module 131 is configured to invoke the camera module 116 to capture video data
- the additional information acquiring module 132 is configured to acquire additional information corresponding to the current video frame and add the additional information to the current video frame.
- the video data encapsulating module 133 is configured to package data of one or more video frames to which additional information is added; the data sending module 134 is configured to send the packaged video data to the cloud server system 20, so that the cloud server System 20 provides various information services based on the additional information based on the additional information within the received video data.
- multiple video frames may be included in the same video data packet, and each video frame includes both additional information and video data of the video frame, and the video data may be in any format (for example, H. 264 or MPEG4, etc. for storage).
- the additional information may include two types, one is editable additional information, and the user may modify, add or delete such information through a specific application, and the editable additional information may generally be used to store information input by the user; The other type is additional information that cannot be edited. Once the video frame is written, the user can no longer edit it.
- the non-editable additional information can generally be used to store the status information acquired in real time.
- the editable additional information may include: a label input by a user, a text introduction, and the like.
- the editable additional information described above may include the code of the instruction entered by the user. Instructions entered by the user may include sharing, reporting, and the like.
- the non-editable additional information described above may include positioning information, such as latitude, longitude, and altitude acquired by the GPS module 112.
- the non-editable additional information may include posture information of the mobile camera terminal 10, for example, a rotation angle of the mobile camera terminal 10 or the camera module 116 in each direction.
- the attitude information of the mobile photographing terminal 10 can be acquired by the sensor 114.
- the non-editable additional information may include: a shooting time of the current video frame.
- the non-editable additional information described above may include: user identification information of the video photographer.
- the user identification information herein may be, for example, an account of the user in a network account system, or other information capable of uniquely determining the user account in a network account system.
- the user of the video capture terminal 20, that is, the video photographer can be limited to only one person.
- the user may be a user account bound to the mobile camera terminal 10 or authorized to use the user account of the mobile camera terminal 10.
- the non-editable additional information may include: verification information of the video data of the current video frame.
- the verification information is calculated, for example, by using a hash algorithm according to the video data, and can be used to verify whether the video data is modified.
- the video data can be verified based on the verification information, so that the authenticity of the video data can be further confirmed, which provides a technical guarantee for the video as judicial evidence.
- editable additional information it can write only part of the video frame, for example, for multiple video frames generated within one second (which can also be other lengths of time), the editable additional information can be written to only one fixed video. Within a frame (for example, the first frame). This video frame with additional information that can be edited can be defined as a key video frame during that time. In this way, editable additional information can be written directly into the video frame, and the storage space occupied by the editable additional information can be minimized.
- non-editable additional information it is generally obtained in real time, so it can be in every frame Both are written. However, it is not limited to this manner, and it is still possible to write non-editable additional information only in a part of the video frame. For example, each second writes non-editable additional information in one video frame.
- the non-editable additional information may be encrypted and processed by the asymmetric encryption algorithm to be written to the video frame.
- each video capture terminal 10 may store the same public key, and the public key may be used to encrypt the non-editable additional information.
- the private key corresponding to the public key is only available in the cloud server system 20, that is, only the cloud server 10 can interpret the encrypted additional information written in the video frame.
- the video data uploaded by the mobile photographing terminal 10 includes the video data and the additional information described above.
- the cloud server system 2020 can include a video processing server 21, a database 22, a distributed file storage system 23, and an application server 24.
- the video processing server 21 is configured to receive the video data packet uploaded by the mobile camera terminal 10, and further process the received video data packet.
- the video processing server 21 performs further processing on the received video data packet, including the following steps:
- Step S101 extracting additional information of each frame of video from the video data packet.
- the video data packet is unpacked to obtain all the video frames, and then the additional information is parsed from the video frame according to a predefined protocol.
- Step S102 processing the video data into a format suitable for storage.
- the video data itself is subjected to a certain compression process, a conversion format, and the like.
- the processing in this step is only for the video data itself, and the processing does not affect the additional information. That is to say, even for the processed video data, the same additional information as before the processing is included in each frame.
- step S102 can be omitted, that is, each frame is extracted from the video data packet. After the video, the received video data packet is directly used as a storage format.
- Step S103 Store the video data in the distributed file storage system and obtain a corresponding storage index. That is, the video data or the video data packet acquired in step S102 is stored in the distributed file storage system, and the distributed file storage system returns a storage index, which is used to implement access to the video data.
- the additional information is stored in the database in association with the storage index.
- a relational database may be used to store additional information and a stored index, and different information within the additional information (eg, coordinates, shooting time, user ID, instruction code, gesture information, tags, etc.) may be stored in separate fields, respectively. It can be understood that if the additional information is encrypted, it needs to be decrypted first.
- the video data can be retrieved, counted, analyzed, outputted and the like based on the additional information, thereby providing various video applications to the user, and the specific process can be implemented by the application server 24.
- the client 30 may include, for example, a smart phone 31, a notebook computer 32, a desktop computer 33, a tablet computer 34, and any other smart terminal not shown in FIG. 1, such as smart glasses, augmented reality helmets, wearable smart devices, and the like. .
- the client 30 interacts with the application server 24 so that various video applications provided by the application server 24 can be used. The following will be described below in conjunction with specific scenarios.
- a certain mobile camera terminal 10 moves from position A to position B.
- the mobile camera terminal 10 is always shooting and uploading an additional information video to the cloud server system 20.
- the cloud server system 20 can forward the received data packet to the client.
- the mobile camera terminal 10 can also send the data packet directly to the client in a point-to-point manner.
- the client parses the positioning information from the received data packet, and can generate a trajectory line in the electronic map 301 according to the positioning information, and at the same time, the client can simultaneously output the video screen 302.
- the sub-map 301 knows the position of the mobile shooting terminal 10 in real time, and can view the currently captured video image in real time through the video screen 302, thereby realizing simultaneous viewing of the moving position and the real-time picture, and is particularly suitable for real-time tracking of a certain target.
- tracking is only meaningful for the same mobile camera terminal 10 and the photographer's identity is the same.
- the angle of view of the video can also be displayed in the electronic map 301.
- each point in the trajectory corresponds to a coordinate.
- the coordinates corresponding to the click point can be obtained according to the preset mapping relationship, and then It is possible to find a video frame whose positioning information is closest to the coordinates among all the received video frames, and to switch the video picture 302 to the video frame.
- the client is viewing the video captured by a particular mobile camera terminal 10, however, the form of attaching the video service in the electronic map is not limited to the above manner, for example, in another specific
- the cloud server system 20 can provide an online electronic map service to provide electronic map data to the client.
- a corresponding electronic map application is run in the client 30, which acquires electronic map data from the cloud server system 20 and displays it in the interface 61.
- a menu 62 can be displayed in the interface 61, with various additional functions in the menu 62, such as an entrance of "panoramic preview", when the user triggers the "panoramic preview” function, the electronic map The application acquires the coordinates of the user's click location, generates a preview request based on the coordinates, and sends the preview request to the cloud server system 20.
- the process of processing a panoramic preview request in the cloud server system 20 includes the following steps:
- Step S201 receiving a panoramic preview request.
- Step S202 parsing the preview coordinates from the panoramic preview request.
- the preview coordinates are obtained by the electronic map application when the user triggers the panoramic preview function.
- Step S203 retrieving the matched video data according to the preview coordinates and the additional information of the video data.
- additional information (including at least address information) of all the video data is stored in the database, and therefore, additional information matching the preview coordinates can be retrieved, and corresponding video data can be acquired via the additional information.
- the search results can be sorted according to a certain order, for example, shooting time, image sharpness, and the like. And then acquiring video data captured by at least one mobile shooting terminal from the sorted result as a retrieval result.
- the transmitted video data may include only one frame without transmitting all the video data, thereby reducing the network transmission amount.
- the video data that can be retrieved must be video data that has been authorized by the photographing user to be disclosed.
- Step S204 the retrieved data is sent to the client for display.
- the client After receiving the video data sent by the cloud server system 20, the client unpacks (if compressed), decrypts (if encrypted), and the like according to a predefined protocol, and then outputs the content in the interface.
- the direction selected by the user may also be included in the preview request.
- the retrieval step of step S203 in addition to retrieving the preview coordinates, it is necessary to retrieve the shooting direction of the video, and only the video data whose coordinates match the direction can be used as the matched video data.
- the panoramic preview function in the electronic map can be realized directly based on the video data captured by the mass mobile shooting terminal, without using a special street view shooting car to shoot Street view photos, effectively reducing the cost of building real-life browsing capabilities.
- the cloud server system 20 also provides a geographic location based video search service. As shown in FIG. 8, the flow of providing a video search service by the cloud server system 20 includes the following steps:
- Step S301 receiving a video search request sent by the client
- Step S302 parsing the search coordinates and the search time from the video search request
- Step S303 searching for matching video data according to the search coordinates and the search time
- Step S304 sending the searched video data to the client for display
- Step S305 modifying the order of the video data captured by the different video capturing terminals according to the feedback during the user watching the video.
- the corresponding video data can be obtained by providing the location and time of the event; and the video data is corrected according to the feedback result during the user viewing process.
- the sorting in turn ensures that the user is most interested in viewing the most interesting video content.
- the video data is retrieved and matched based on the geographic location information in the video data, and the specific video application scenario is implemented on the basis of the video data.
- the additional information within the video data is not limited to these application scenarios.
- the user can also input some instructions, such as sharing, traffic violation reporting, alarm, insurance, first aid, and the like.
- the input method of the command may be directly input through a button on the mobile photographing terminal 10, a touch screen, or the like, or may be input through the mobile electronic terminal 30 connected to the mobile photographing terminal 10.
- the code for these instructions can be added to the data frame of the video frame.
- the cloud server system 20 can parse out the instruction code therefrom, and execute the corresponding data processing process, or forward the instructions to the servers of other third-party organizations for further processing. deal with.
- the third party institutions herein may include: a self-media platform, a social network platform, a traffic violation handling agency, a police, an insurance institution, a first aid institution, and the like.
- the cloud server system 20 can automatically intercept a predetermined length (which can be started from the time of receiving the sharing instruction) or a video clip of a specified length, and share it to the mobile shooting terminal 10. From the media platform or social network platform (such as WeChat friends circle, QQ space, Weibo, etc.). In this way, when the user encounters interesting events, beautiful scenery or any other content that the user wants to share, it is possible to share the captured video with one click.
- a predetermined length which can be started from the time of receiving the sharing instruction
- a video clip of a specified length e.g., a video clip of a specified length
- the user may add a label, a text introduction, and the like to the shared content.
- the mobile photographing terminal 10 includes an input interface
- the user can directly input on the mobile photographing terminal 10, and the mobile photographing terminal 10 may not include the input interface.
- the input may be performed by the mobile electronic terminal bound to the mobile photographing terminal 10.
- These tags, text descriptions can be stored by the cloud server system 20 and used for video retrieval.
- the cloud server system 20 can automatically intercept a relevant video segment and send it to the server of the traffic violation processing institution, so that the user can realize one-click reporting of traffic violation and violation by the mobile shooting terminal 10. .
- the cloud server system 20 can automatically intercept a relevant video segment and send it to the alarm server of the alarm receiving institution. In this way, the user can realize a one-button alarm by moving the shooting terminal 10, and because the video frame data packet directly contains the positioning information, the police can conveniently locate the alarm position in time.
- the cloud server system 20 can automatically intercept a relevant video segment and send it to the server of the insurance institution. In this way, the user can realize the one-key insurance by moving the shooting terminal 10. Since the video can effectively restore the live situation on the spot, the insurance institution can realize the remote claim service without going to the scene.
- the cloud server system 20 can automatically intercept a relevant section.
- the video clip is sent to the server of the emergency department, so that the user can realize the function of one-touch call emergency service by moving the photographing terminal 10.
- the emergency department can quickly locate the call position and reduce the position communication time.
- the instruction may be directly embedded in the additional information of the video data, so that the cloud server system 20 executes the corresponding data processing process and functions, so that the video data can be more widely applied.
- the user when the user shares the video captured by the mobile shooting terminal 10, the user can input a label or a text introduction.
- the cloud server system 20 based on the text introduction, automatic discovery and mining of the hot event and related video can also be realized.
- the automatic mining process of hotspot events and related videos based on video tags and texts includes the following steps:
- Step S301 extracting keywords from the received video tags and text introduction.
- Video tags can generally be used directly as keywords.
- the text introduction can perform steps such as word segmentation processing and word frequency statistics to extract keywords.
- step S302 the frequency of occurrence of the keyword and/or other parameters are separately calculated based on the geographic location.
- the frequency of occurrence of a certain keyword and other parameters can be counted based on the geographical location.
- Other parameters herein may include, for example, density, increased speed, and the like.
- Step S303 when the frequency and/or other parameters of a certain keyword appearing in a certain place exceed a preset threshold, the content of the video related to the keyword based on the location is released in a content distribution system.
- the frequency and/or other parameters of a keyword appearing in a place are higher than the preset value, it can be regarded as a hot event or a hot spot.
- the video content of the place related to the keyword can be published in a content distribution system for the user to browse.
- the content publishing system here, for example, a video website, APP or other content publishing system.
- the released video content may be a video captured by a single mobile shooting terminal, or may be a video data captured by a plurality of mobile shooting terminals 10.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Signal Processing (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Databases & Information Systems (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Human Computer Interaction (AREA)
- Library & Information Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Computer Security & Cryptography (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
The present invention relates to a geographic location-based video processing method for use by a mobile filming terminal, wherein said mobile filming terminal comprises a camera, and said method comprises: invoking the camera to film so as to obtain video data; obtaining a current video frame to be sent according to the video data; obtaining additional information corresponding to the current video frame, wherein the additional information at least contains current positioning information of the mobile filming terminal; adding the additional information to a data packet of the current video frame; and sending the data packet of the current video frame to which the additional information has been added to either a cloud server or a remote client, thus allowing the cloud server or remote client to process the video frame according to the positioning information within the received data packet. In the above method, each video frame contains positioning information, and may also contain other additional information; on the basis of said additional information, the method allows for video data to be looked up according to geographic location, thereby providing plentiful applications of videos.
Description
本发明涉及视频处理技术,尤其涉及一种基于地理位置的视频处理技术。The present invention relates to video processing technologies, and in particular, to a geographic location based video processing technology.
随着网络通信技术以及网络基础设施的不断发展,使得在互联网上提供各种多媒体内容成为可能。另一方面,随着移动电子终端如智能手机的发展,网络上的已经上传了海量的各种用户拍摄的视频。这些视频记录了丰富的信息,而现有的视频检索系统一般都是基于视频上传者上传视频时设置的名称或者标签来实现的,视频内丰富的信息并不能被检索利用。With the continuous development of network communication technologies and network infrastructure, it is possible to provide various multimedia contents on the Internet. On the other hand, with the development of mobile electronic terminals such as smart phones, a large number of videos taken by various users have been uploaded on the network. These videos record a wealth of information, and existing video retrieval systems are generally implemented based on the names or tags set by the video uploader when uploading the video. The rich information in the video cannot be retrieved.
发明内容Summary of the invention
有鉴于此,有必要提供一种视频处理方法及系统,其可以解决现有技术中视频内包含的信息无法被检索利用的问题。In view of the above, it is necessary to provide a video processing method and system, which can solve the problem that information contained in a video in the prior art cannot be retrieved and utilized.
一种基于地理位置的视频处理方法,应用于移动拍摄终端中,所述移动拍摄终端包括摄像头,所述方法包括:A geographic location-based video processing method is applied to a mobile shooting terminal, the mobile shooting terminal includes a camera, and the method includes:
调用所述摄像头拍摄以获取视频数据;Calling the camera to capture video data;
根据所述视频数据获取待发送的当前视频帧;Obtaining a current video frame to be sent according to the video data;
获取与该当前视频帧对应的附加信息,所述附加信息至少包括所述移动拍摄终端当前的定位信息;Obtaining additional information corresponding to the current video frame, where the additional information includes at least current positioning information of the mobile camera terminal;
将所述附加信息添加至所述当前视频帧的数据包内;以及Adding the additional information to a data packet of the current video frame;
将添加有附加信息的当前视频帧的数据包发送给云端服务器或者远端的客户端,以使所述云端服务器或者远端的客户端根据接收到的数据包
内的定位信息对视频帧进行处理。Sending a data packet of the current video frame to which the additional information is added to the cloud server or the remote client, so that the cloud server or the remote client according to the received data packet
The positioning information within the processing of the video frame.
在一个实施例中,上述的方法还包括:在拍摄时实时检测所述摄像头的旋转角度;所述附加信息还包括所述摄像头的旋转角度。In one embodiment, the method further includes: detecting a rotation angle of the camera in real time during shooting; the additional information further includes a rotation angle of the camera.
在一个实施例中,上述的方法还包括:获取与所述当前视频帧对应的拍摄者的用户身份识别信息;所述附加信息还包括所述用户身份识别信息。In an embodiment, the method further includes: acquiring user identity information of the photographer corresponding to the current video frame; the additional information further includes the user identity information.
在一个实施例中,上述的方法还包括:在拍摄时实时获取所述当前视频帧的拍摄时间;所述附加信息还包括所述拍摄时间。In an embodiment, the method further includes: acquiring a shooting time of the current video frame in real time during shooting; the additional information further includes the shooting time.
在一个实施例中,上述的方法还包括:获取用户输入的指令;所述附加信息还包括所述拍摄时间。In one embodiment, the method further includes: obtaining an instruction input by a user; the additional information further including the shooting time.
在一个实施例中,上述的方法还包括:接收与用户输入的标签和/或文字介绍;所述附加信息还包括所述标签和/或文字介绍。In one embodiment, the method further includes receiving a label and/or textual input with the user; the additional information further including the label and/or textual introduction.
在一个实施例中,上述的方法还包括:对所述附加信息的至少部分内容进行加密处理,所述附加信息至少包括经过加密处理的内容。In one embodiment, the above method further comprises: encrypting at least part of the content of the additional information, the additional information including at least the content processed by the encryption.
在一个实施例中,上述的方法还包括:根据所述当前帧的视频数据计算对应的校验信息;所述经过加密处理的内容内包括所述校验信息。In an embodiment, the method further includes: calculating corresponding verification information according to the video data of the current frame; and the verification information is included in the encrypted processed content.
一种基于地理位置的视频处理方法,应用于云端服务器系统中,包括:A geographic location-based video processing method applied to a cloud server system, including:
接收移动拍摄终端发送的视频数据包;Receiving a video data packet sent by the mobile shooting terminal;
从所述视频数据包中解析出视频数据以及对应的附加信息,所述附加信息至少包括所述移动拍摄终端当前的定位信息;And parsing the video data and the corresponding additional information from the video data packet, where the additional information includes at least the current positioning information of the mobile shooting terminal;
存储所述视频数据并获取对应的索引;Storing the video data and obtaining a corresponding index;
将所述附加信息与所述索引关联存储;以及Storing the additional information in association with the index;
基于所述附加信息检索并获取对应的视频数据,并基于检索到的视频数据向客户端提供视频数据服务。Retrieving and acquiring corresponding video data based on the additional information, and providing a video data service to the client based on the retrieved video data.
在一个实施例中,上述的方法还包括:从所述视频数据包中解析出与
每个视频帧对应的指令;并执行与所述指令对应的数据处理过程。In an embodiment, the method further includes: parsing out from the video data packet
An instruction corresponding to each video frame; and executing a data processing procedure corresponding to the instruction.
在一个实施例中,所述数据处理过程包括:In one embodiment, the data processing process includes:
截取预定长度或者指定长度的视频片断,将其分享至与所述移动拍摄终端相绑定的自媒体平台或者社交网络平台;或者Capturing a video clip of a predetermined length or a specified length and sharing it to a self-media platform or a social networking platform bound to the mobile camera terminal; or
截取一段相关的视频片段发送至交通违章处理机构的服务器中;或者Intercepting a relevant video clip to be sent to the server of the traffic violation handling agency; or
自动截取一段相关的视频片段发送至报警受理机构的报警服务器中;或者Automatically intercepting a related video clip and sending it to the alarm server of the alarm receiving institution; or
自动截取一段相关的视频片段发送至急救机构的服务器中。Automatically intercepts a related video clip and sends it to the server of the emergency department.
在一个实施例中,上述的方法还包括:从所述视频数据包中解析出视频的标签和/或文字介绍;In an embodiment, the method further includes: parsing a label and/or a text description of the video from the video data package;
将所述标签和/或文字介绍存储;Store the label and/or text description;
从所述标签和/或文字介绍中提取出关键词;Extracting keywords from the label and/or text introduction;
基于地理位置对所述关键词进行统计分析;Perform statistical analysis on the keywords based on geographic location;
当某个地理位置的某个关键词的统计值超过预设阈值时获取相关的视频数据,并在一个内容发布系统中发布所述获取的相关的视频数据。When the statistical value of a certain keyword of a certain geographical location exceeds a preset threshold, relevant video data is acquired, and the acquired related video data is published in a content distribution system.
根据上述的技术方案,视频数据的每个视频帧内都包括定位信息,还可包括其他的附加信息,基于这些附加信息,可以使得视频数据能够实现基于地理位置的检索,从而可在此基础上提供丰富的视频应用。According to the above technical solution, each video frame of the video data includes positioning information, and may further include other additional information, and based on the additional information, the video data can be subjected to geo-location-based retrieval, thereby being able to perform on the basis of Provide a rich video application.
为让本发明的上述和其他目的、特征和优点能更明显易懂,下文特举较佳实施例,并配合所附图式,作详细说明如下。The above and other objects, features, and advantages of the present invention will become more apparent and understood by the appended claims appended claims
图1为本发明实施例提供的基于地理位置的视频信息处理系统的架构示意图。FIG. 1 is a schematic structural diagram of a geographic location-based video information processing system according to an embodiment of the present invention.
图2为图1的视频信息处理系统的移动拍摄终端的结构框图。
2 is a block diagram showing the structure of a mobile photographing terminal of the video information processing system of FIG. 1.
图3为图2的移动拍摄终端上传的视频数据的数据结构示意图。FIG. 3 is a schematic diagram showing the data structure of video data uploaded by the mobile camera terminal of FIG. 2.
图4为图1的视频信息处理系统处理移动拍摄终端上传的视频数据的流程示意图。4 is a schematic flow chart of processing video data uploaded by a mobile camera terminal by the video information processing system of FIG. 1.
图5为图1的视频信息处理系统基于带有地理位置的视频数据提供的应用的界面示意图。5 is a schematic diagram of an interface of an application provided by the video information processing system of FIG. 1 based on video data with geographic locations.
图6为图1的视频信息处理系统基于带有地理位置的视频数据提供的另一应用的界面示意图。6 is a schematic diagram of an interface of another application provided by the video information processing system of FIG. 1 based on video data with geographic location.
图7为图1的视频信息处理系统基于带有地理位置的视频数据提供全景预览功能的流程示意图。FIG. 7 is a schematic flow chart of the video information processing system of FIG. 1 providing a panoramic preview function based on video data with geographic locations.
图8为图1的视频信息处理系统基于带有地理位置的视频数据提供视频搜索服务的流程示意图。8 is a flow chart showing the video information processing system of FIG. 1 providing a video search service based on video data with geographic location.
图9为图1的视频信息处理系统的扩展示意图。9 is an expanded schematic diagram of the video information processing system of FIG. 1.
图10为图1的视频信息处理系统基于带有地理位置、标签和/或文字介绍的视频数据提供视频应用的流程示意图。10 is a flow diagram of the video information processing system of FIG. 1 providing a video application based on video data with geographic location, tags, and/or textual introduction.
为更进一步阐述本发明为实现预定发明目的所采取的技术手段及功效,以下结合附图及较佳实施例,对依据本发明的具体实施方式、结构、特征及其功效,详细说明如后。The specific embodiments, structures, features and functions of the present invention will be described in detail below with reference to the accompanying drawings and preferred embodiments.
参阅图1,其为本发明第一实施例提供的基于地理位置的视频信息处理系统的架构示意图。如图1所示,视频信息处理系统100可包括:移动拍摄终端10、云端服务器系统20、以及客户端30。FIG. 1 is a schematic structural diagram of a geographic location-based video information processing system according to a first embodiment of the present invention. As shown in FIG. 1, the video information processing system 100 may include a mobile photography terminal 10, a cloud server system 20, and a client 30.
移动拍摄终端10,具体地,可为任意带有摄像头的移动电子终端例如手机、平板电脑、无人机等等。参阅图2,其为移动拍摄终端10的结构示意图。移动拍摄终端10包括存储器102、存储控制器104,一个或多
个(图中仅示出一个)处理器106、外设接口108、网络模块110、音频电路111、GPS(Global Positioning System,全球定位系统)模块112、传感器114、摄像模块116、及电源模块122。这些组件通过一条或多条通讯总线/信号线相互通讯。The mobile photographing terminal 10, in particular, may be any mobile electronic terminal with a camera such as a mobile phone, a tablet computer, a drone, or the like. Referring to FIG. 2, it is a schematic structural diagram of the mobile photographing terminal 10. The mobile photographing terminal 10 includes a memory 102, a memory controller 104, and one or more
(only one shown in the figure) processor 106, peripheral interface 108, network module 110, audio circuit 111, GPS (Global Positioning System) module 112, sensor 114, camera module 116, and power module 122 . These components communicate with one another via one or more communication bus/signal lines.
本领域普通技术人员可以理解,图2所示的结构仅为示意,其并不对移动拍摄终端10的结构造成限定。例如,移动拍摄终端10还可包括比图2中所示更多或者更少的组件,或者具有与图2所示不同的配置。It will be understood by those skilled in the art that the structure shown in FIG. 2 is merely illustrative and does not limit the structure of the mobile camera terminal 10. For example, the mobile photography terminal 10 may also include more or less components than those shown in FIG. 2, or have a different configuration than that shown in FIG. 2.
存储器102可用于存储软件程序以及模块,如本发明实施例中的各方法及装置对应的程序指令/模块,处理器106通过运行存储在存储器102内的软件程序以及模块,从而执行各种功能应用以及数据处理。The memory 102 can be used to store software programs and modules, such as program instructions/modules corresponding to the methods and devices in the embodiments of the present invention. The processor 106 executes various functional applications by running software programs and modules stored in the memory 102. And data processing.
存储器102可包括高速随机存储器,还可包括非易失性存储器,如一个或者多个磁性存储装置、闪存、或者其他非易失性固态存储器。在一些实例中,存储器102可进一步包括相对于处理器106远程设置的存储器,这些远程存储器可以通过网络连接至上述服务器。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。处理器106以及其他可能的组件对存储器102的访问可在存储控制器104的控制下进行。 Memory 102 can include high speed random access memory and can also include non-volatile memory, such as one or more magnetic storage devices, flash memory, or other non-volatile solid state memory. In some examples, memory 102 can further include memory remotely located relative to processor 106, which can be connected to the server via a network. Examples of such networks include, but are not limited to, the Internet, intranets, local area networks, mobile communication networks, and combinations thereof. Access to the memory 102 by the processor 106 and other possible components can be performed under the control of the memory controller 104.
外设接口108将各种输入/输入装置耦合至处理器106。处理器106运行存储器102内的各种软件、指令上述服务器执行各种功能以及进行数据处理。在一些实施例中,外设接口108、处理器106以及存储控制器104可以在单个芯片中实现。在其他一些实例中,他们可以分别由独立的芯片实现。 Peripheral interface 108 couples various input/input devices to processor 106. The processor 106 runs various software within the memory 102, instructs the server to perform various functions, and performs data processing. In some embodiments, peripheral interface 108, processor 106, and memory controller 104 can be implemented in a single chip. In other instances, they can be implemented by separate chips.
网络模块110用于接收以及发送网络信号。上述网络信号可包括无线信号。在一个实施例中,网络模块110实质是射频模块,接收以及发送电磁波,实现电磁波与电信号的相互转换,从而与通讯网络或者其他设备进
行通讯。射频模块可包括各种现有的用于执行这些功能的电路元件,例如,天线、射频收发器、数字信号处理器、加密/解密芯片、用户身份模块(SIM)卡、存储器等等。射频模块可与各种网络如互联网、企业内部网、无线网络进行通讯或者通过无线网络与其他设备进行通讯。上述的无线网络可包括蜂窝式电话网、无线局域网或者城域网。上述的无线网络可以使用各种通信标准、协议及技术,包括但并不限于全球移动通信系统(Global System for Mobile Communication,GSM)、增强型移动通信技术(Enhanced Data GSM Environment,EDGE),宽带码分多址技术(wideband code division multiple access,W-CDMA),码分多址技术(Code division access,CDMA)、时分多址技术(Time Division Multiple Access,TDMA),无线保真技术(Wireless Fidelity,WiFi)(如美国电气和电子工程师协会标准IEEE 802.11a,IEEE 802.11b,IEEE802.11g和/或IEEE 802.11n)、网络电话(Voice Over Internet Protocol,VoIP)、全球微波互联接入(Worldwide Interoperability for Microwave Access,Wi-Max)、其他用于邮件、即时通讯及短消息的协议,以及任何其他合适的通讯协议,甚至可包括那些当前仍未被开发出来的协议。The network module 110 is configured to receive and transmit network signals. The network signal described above may include a wireless signal. In one embodiment, the network module 110 is substantially a radio frequency module that receives and transmits electromagnetic waves, and converts electromagnetic waves and electrical signals, thereby interacting with a communication network or other devices.
Line communication. The radio frequency module can include various existing circuit components for performing these functions, such as an antenna, a radio frequency transceiver, a digital signal processor, an encryption/decryption chip, a Subscriber Identity Module (SIM) card, a memory, and the like. The RF module can communicate with various networks such as the Internet, intranets, wireless networks or with other devices over a wireless network. The wireless network described above may include a cellular telephone network, a wireless local area network, or a metropolitan area network. The above wireless network can use various communication standards, protocols and technologies, including but not limited to Global System for Mobile Communication (GSM), Enhanced Data GSM Environment (EDGE), and wideband code. Wideband code division multiple access (W-CDMA), Code division access (CDMA), Time Division Multiple Access (TDMA), Wireless Fidelity (Wireless Fidelity, WiFi) (such as the Institute of Electrical and Electronics Engineers Standard IEEE 802.11a, IEEE 802.11b, IEEE 802.11g and / or IEEE 802.11n), Voice Over Internet Protocol (VoIP), Worldwide Interoperability for Microwave Access, Wi-Max, other protocols for mail, instant messaging, and short messages, as well as any other suitable communication protocol, may even include protocols that are not currently being developed.
音频电路111与提供移动拍摄终端10录音的接口。具体地,音频电路111从麦克风处接收电信号,将电信号转换为声音数据,并将声音数据传输给处理器102以进行进一步的处理。The audio circuit 111 is provided with an interface for providing recording of the mobile photographing terminal 10. Specifically, the audio circuit 111 receives an electrical signal from the microphone, converts the electrical signal into sound data, and transmits the sound data to the processor 102 for further processing.
GPS模块112用于接收GPS卫星播报的定位信号,并根据定位信号计算出自身的位置。上述位置例如可用经度、纬度以及海拔高度来表示。可以理解,实现定位的方式并不限于GPS系统。例如,其他可用的卫星定位系统还包括北斗卫星定位系统(Compass Navigation Satellite System,CNSS)或者格洛纳斯系统(Global Navigation Satellite System,GLONASS)。此外,定位也并不限于采用卫星定位技术,例如,还可采
用无线定位技术,例如基于无线基站的定位技术或者WIFI的定位技术。此时,GPS模块112可被替换成相应的模块,或者直接经由处理器102执行特定的定位程序来实现。The GPS module 112 is configured to receive a positioning signal of a GPS satellite broadcast, and calculate its own position according to the positioning signal. The above positions can be expressed, for example, by longitude, latitude, and altitude. It will be appreciated that the manner in which positioning is achieved is not limited to GPS systems. For example, other available satellite positioning systems include the Compass Navigation Satellite System (CNSS) or the Global Navigation Satellite System (GLONASS). In addition, positioning is not limited to satellite positioning technology, for example,
Wireless positioning technology, such as positioning technology based on wireless base stations or positioning technology of WIFI. At this point, the GPS module 112 can be replaced with a corresponding module, or directly by the processor 102 executing a particular positioning procedure.
传感器114的实例包括但并不限于:光传感器、姿态传感器、及其他传感器。其中,环境光传感器可以感测环境光的明暗,进而可以对拍摄进行调节。姿态传感器例如可包括加速度传感器、重力计、陀螺仪等等,其可检测移动拍摄终端10的空间姿态例如在各方向上的旋转角度等。可以理解,移动拍摄终端10的在各方向上的旋转角度既对应着拍摄方向。其他传感器可包括气压计、湿度计、温度计等。Examples of sensors 114 include, but are not limited to, light sensors, attitude sensors, and other sensors. Among them, the ambient light sensor can sense the brightness of the ambient light, and thus can adjust the shooting. The attitude sensor may include, for example, an acceleration sensor, a gravimeter, a gyroscope, or the like, which can detect a spatial attitude of the moving photographing terminal 10, for example, a rotation angle in various directions, and the like. It can be understood that the rotation angle of the moving photographing terminal 10 in each direction corresponds to the photographing direction. Other sensors may include barometers, hygrometers, thermometers, and the like.
摄像模块116用于拍摄照片或者视频。拍摄的照片或者视频可以存储至存储器104内,并可通过网络模块110发送。摄像模块116具体可包括镜头模组、影像感测器以及闪光灯等组件。镜头模组用于对被拍摄的目标成像,并将所成的像映射至影像感测器中。影像感测器用于接收来自镜头模组的光线,实现感光,以记录图像信息。具体地,影像感测器可基于互补金属氧化物半导体(Complementary Metal Oxide Semiconductor,CMOS)、电荷耦合元件(Charge-coupled Device,CCD)或者其他影像感测原理实现。闪光灯用于在拍摄时进行曝光补偿。一般来说,用于移动拍摄终端10的闪光灯可为发光二极体(Light Emitting Diode,LED)闪光灯。The camera module 116 is used to take a photo or video. The photographed or video taken may be stored in the memory 104 and may be transmitted through the network module 110. The camera module 116 may specifically include components such as a lens module, an image sensor, and a flash. The lens module is used to image the captured object and map the formed image to the image sensor. The image sensor is used to receive light from the lens module to achieve sensitization to record image information. Specifically, the image sensor can be implemented based on a Complementary Metal Oxide Semiconductor (CMOS), a Charge-coupled Device (CCD), or other image sensing principle. The flash is used to compensate for exposure when shooting. In general, the flash for moving the photographing terminal 10 may be a Light Emitting Diode (LED) flash.
电源模块122用于向处理器102以及其他各组件提供电力供应。具体地,电源模块122可包括电源管理系统、一个或多个电源(如电池或者交流电)、充电电路、电源失效检测电路、逆变器、电源状态指示灯以及其他任意与移动拍摄终端10内电力的生成、管理及分布相关的组件。The power module 122 is used to provide power to the processor 102 and other components. Specifically, the power module 122 may include a power management system, one or more power sources (such as a battery or an alternating current), a charging circuit, a power failure detecting circuit, an inverter, a power status indicator, and any other power in the mobile shooting terminal 10. The generation, management, and distribution of related components.
存储器104内存储有软件及程序模块可包括操作系统130以及运行在操作系统130上的应用程序。操作系统130其可包括各种用于管理系统任
务(例如内存管理、存储设备控制、电源管理等)的软件组件和/或驱动,并可与各种硬件或软件组件相互通讯,从而提供其他软件组件的运行环境。所述应用程序可包括:拍摄模块131、附加信息添加模块132、视频数据封装模块133以及数据发送模块134。The software and program modules stored in the memory 104 may include an operating system 130 and an application running on the operating system 130. The operating system 130 can include various management systems
Software components and/or drivers (such as memory management, storage device control, power management, etc.) that communicate with various hardware or software components to provide an operating environment for other software components. The application program may include: a shooting module 131, an additional information adding module 132, a video data encapsulating module 133, and a data sending module 134.
其中,拍摄模块131用于调用所述摄像模块116拍摄以获取视频数据;附加信息获取模块132用于获取与该当前视频帧对应的附加信息并将所述附加信息添加至所述当前视频帧内;视频数据封装模块133用于将一个或者多个添加有附加信息的视频帧的数据进行打包;数据发送模块134用于将打包后的视频数据发送给云端服务器系统20,以使所述云端服务器系统20根据接收到的视频数据内的附加信息提供各种基于该附加信息提供的信息服务。The shooting module 131 is configured to invoke the camera module 116 to capture video data, and the additional information acquiring module 132 is configured to acquire additional information corresponding to the current video frame and add the additional information to the current video frame. The video data encapsulating module 133 is configured to package data of one or more video frames to which additional information is added; the data sending module 134 is configured to send the packaged video data to the cloud server system 20, so that the cloud server System 20 provides various information services based on the additional information based on the additional information within the received video data.
如图3所示,在同一个视频数据包内,可包括多个视频帧,而每一个视频帧内同时包括该视频帧的附加信息与视频数据,视频数据可以采用任意的格式(例如H.264或者MPEG4等进行存储)。As shown in FIG. 3, multiple video frames may be included in the same video data packet, and each video frame includes both additional information and video data of the video frame, and the video data may be in any format (for example, H. 264 or MPEG4, etc. for storage).
附加信息可包括两类,一类是可编辑的附加信息,用户可以通过特定的应用实现对此类信息的修改、新增或者删除,可编辑的附加信息一般可以用来存储用户输入的信息;另一类是不可编辑的附加信息,一经写入视频帧,用户无法再对其进行编辑,不可编辑的附加信息一般可用来存储实时获取的状态信息。The additional information may include two types, one is editable additional information, and the user may modify, add or delete such information through a specific application, and the editable additional information may generally be used to store information input by the user; The other type is additional information that cannot be edited. Once the video frame is written, the user can no longer edit it. The non-editable additional information can generally be used to store the status information acquired in real time.
在一个具体的实施例中,上述的可编辑的附加信息可包括:用户输入的标签、文字介绍等信息。In a specific embodiment, the editable additional information may include: a label input by a user, a text introduction, and the like.
在一个具体的实施例中,上述的可编辑的附加信息可包括:用户输入的指令的代码。用户输入的指令可包括分享、举报等等。In a specific embodiment, the editable additional information described above may include the code of the instruction entered by the user. Instructions entered by the user may include sharing, reporting, and the like.
在一个具体的实施例中,上述的不可编辑的附加信息可包括:定位信息,例如通过GPS模块112获取到的经、纬度及高度。
In a specific embodiment, the non-editable additional information described above may include positioning information, such as latitude, longitude, and altitude acquired by the GPS module 112.
在一个具体的实施例中,上述的不可编辑的附加信息可包括:所述移动拍摄终端10的姿态信息,例如,移动拍摄终端10或者摄像模块116在各方向上的旋转角度。移动拍摄终端10的姿态信息可以通过传感器114获取。In a specific embodiment, the non-editable additional information may include posture information of the mobile camera terminal 10, for example, a rotation angle of the mobile camera terminal 10 or the camera module 116 in each direction. The attitude information of the mobile photographing terminal 10 can be acquired by the sensor 114.
在一个具体的实施例中,上述的不可编辑的附加信息可包括:所述当前视频帧的拍摄时间。In a specific embodiment, the non-editable additional information may include: a shooting time of the current video frame.
在一个具体的实施例中,上述的不可编辑的附加信息可包括:视频拍摄者的用户身份识别信息。此处的用户身份识别信息例如可以是用户在一个网络帐号系统中的帐号,或者其他能够在一个网络帐号系统中唯一地确定用户帐号的信息。在同一时刻,视频拍摄终端20的使用者,即视频拍摄者可以限定为仅一人。该使用者可以是与移动拍摄终端10绑定的用户帐号,或者被授权使用移动拍摄终端10的用户帐号。In a specific embodiment, the non-editable additional information described above may include: user identification information of the video photographer. The user identification information herein may be, for example, an account of the user in a network account system, or other information capable of uniquely determining the user account in a network account system. At the same time, the user of the video capture terminal 20, that is, the video photographer, can be limited to only one person. The user may be a user account bound to the mobile camera terminal 10 or authorized to use the user account of the mobile camera terminal 10.
在一个具体的实施例中,上述的不可编辑的附加信息可包括:所述当前视频帧的视频数据的校验信息。所述校验信息例如是采用哈希算法根据所述视频数据计算得到,可以用于校验所述视频数据是否被修改。如此无论该视频帧如何复制、传输,均可以基于该校验信息校验视频数据是否被修改,从而可以使得视频数据的真实性可以进一步确认,这给视频作为司法证据提供了技术上的保障。In a specific embodiment, the non-editable additional information may include: verification information of the video data of the current video frame. The verification information is calculated, for example, by using a hash algorithm according to the video data, and can be used to verify whether the video data is modified. Thus, no matter how the video frame is copied and transmitted, the video data can be verified based on the verification information, so that the authenticity of the video data can be further confirmed, which provides a technical guarantee for the video as judicial evidence.
对于可编辑的附加信息,其可以仅写入部分视频帧,例如,对于一秒(也可以为其他时间长度)内产生的多个视频帧,可编辑的附加信息可仅写入一个固定的视频帧(例如第一帧)内。这个有写有可编辑的附加信息的视频帧可以定义为该时间内的关键视频帧。采用这种方式,既可以直接在视频帧内写入可编辑的附加信息,也可以最大程度的减少可编辑的附加信息占据的存储空间。For editable additional information, it can write only part of the video frame, for example, for multiple video frames generated within one second (which can also be other lengths of time), the editable additional information can be written to only one fixed video. Within a frame (for example, the first frame). This video frame with additional information that can be edited can be defined as a key video frame during that time. In this way, editable additional information can be written directly into the video frame, and the storage space occupied by the editable additional information can be minimized.
对于不可编辑的附加信息,一般都是实时获取的,因此,可以在每帧
中都写入。然而,并不限于这种方式,仍然可以是仅在部分视频帧内写入不可编辑的附加信息。例如,每一秒在一个视频帧中写入不可编辑的附加信息。For non-editable additional information, it is generally obtained in real time, so it can be in every frame
Both are written. However, it is not limited to this manner, and it is still possible to write non-editable additional information only in a part of the video frame. For example, each second writes non-editable additional information in one video frame.
此外,为了防止不可编辑的附加信息被破坏、或者篡改,不可编辑的附加信息可以采用非对称加密算法进行加密处理后写入视频帧。例如,每个视频拍摄终端10内可以存储有相同的公钥,利用该公钥对不可编辑的附加信息进行加密处理。而与该公钥相对应的私钥只有云端服务器系统20中才有,也就是说,只有云端服务器10才能够解读写入视频帧内的加密后的附加信息。In addition, in order to prevent the non-editable additional information from being destroyed or falsified, the non-editable additional information may be encrypted and processed by the asymmetric encryption algorithm to be written to the video frame. For example, each video capture terminal 10 may store the same public key, and the public key may be used to encrypt the non-editable additional information. The private key corresponding to the public key is only available in the cloud server system 20, that is, only the cloud server 10 can interpret the encrypted additional information written in the video frame.
如上所述,在本实施例的视频信息处理系统中,移动拍摄终端10上传的视频数据内包括了视频数据以及上述的附加信息。As described above, in the video information processing system of the present embodiment, the video data uploaded by the mobile photographing terminal 10 includes the video data and the additional information described above.
而如图1所示,云端服务器系统2020可包括视频处理服务器21、数据库22、分布式文件存储系统23以及应用服务器24。As shown in FIG. 1, the cloud server system 2020 can include a video processing server 21, a database 22, a distributed file storage system 23, and an application server 24.
其中,视频处理服务器21用于接收移动拍摄终端10上传的视频数据包,并对接收到的视频数据包进行进一步的处理。The video processing server 21 is configured to receive the video data packet uploaded by the mobile camera terminal 10, and further process the received video data packet.
参阅图4,在一个具体的实施例中,视频处理服务器21对接收到的视频数据包进行进一步处理包括以下步骤:Referring to FIG. 4, in a specific embodiment, the video processing server 21 performs further processing on the received video data packet, including the following steps:
步骤S101,从视频数据包内提取出每帧视频的附加信息。首先,对视频数据包进行解包处理,得到所有的视频帧,然后根据预定义的协议从视频帧中解析出附加信息。Step S101, extracting additional information of each frame of video from the video data packet. First, the video data packet is unpacked to obtain all the video frames, and then the additional information is parsed from the video frame according to a predefined protocol.
步骤S102,将视频数据处理成适于存储的格式。例如,对视频数据本身进行一定的压缩处理,转换格式等等。但值得注意的是,此步骤中的处理仅仅是针对视频数据本身,处理过程并不影响附加信息。也就是说,即使是处理后的视频数据,每帧内仍然包括与处理前相同的附加信息。此外,步骤S102是可省略的,也就是说,在从视频数据包内提取出每帧的
视频后,直接将接收到的视频数据包作为存储格式。Step S102, processing the video data into a format suitable for storage. For example, the video data itself is subjected to a certain compression process, a conversion format, and the like. However, it is worth noting that the processing in this step is only for the video data itself, and the processing does not affect the additional information. That is to say, even for the processed video data, the same additional information as before the processing is included in each frame. In addition, step S102 can be omitted, that is, each frame is extracted from the video data packet.
After the video, the received video data packet is directly used as a storage format.
步骤S103,将视频数据存储于分布式文件存储系统中并获取对应的存储索引。即,将步骤S102中获取的视频数据,或者视频数据包存入分布式文件存储系统中,分布文件存储系统会返回存储索引,该存储索引用于实现对该视频数据在存取。Step S103: Store the video data in the distributed file storage system and obtain a corresponding storage index. That is, the video data or the video data packet acquired in step S102 is stored in the distributed file storage system, and the distributed file storage system returns a storage index, which is used to implement access to the video data.
步骤S104,将附加信息与存储索引关联存储于数据库中。例如,可以采用关系型数据库存储附加信息与存储索引,而附加信息内的不同信息(例如坐标、拍摄时间、用户ID、指令代码、姿态信息、标签等)可以分别存储于不同字段内。可以理解,如果附加信息经过了加密,还需要先进行解密处理。In step S104, the additional information is stored in the database in association with the storage index. For example, a relational database may be used to store additional information and a stored index, and different information within the additional information (eg, coordinates, shooting time, user ID, instruction code, gesture information, tags, etc.) may be stored in separate fields, respectively. It can be understood that if the additional information is encrypted, it needs to be decrypted first.
经过以上的处理过程,就可以基于这些附加信息对视频数据进行检索、统计、分析、输出等处理,从而提供各种视频应用程序给用户,而具体的过程可由应用服务器24实现。Through the above processing, the video data can be retrieved, counted, analyzed, outputted and the like based on the additional information, thereby providing various video applications to the user, and the specific process can be implemented by the application server 24.
客户端30可包括例如智能手机31、笔记本电脑32、台式电脑33、、平板电脑34、以及其他任意未显示在图1中的智能终端,例如智能眼镜、增强现实头盔、可穿戴智能设备等等。The client 30 may include, for example, a smart phone 31, a notebook computer 32, a desktop computer 33, a tablet computer 34, and any other smart terminal not shown in FIG. 1, such as smart glasses, augmented reality helmets, wearable smart devices, and the like. .
客户端30与应用服务器24进行交互,从而可以使用应用服务器24提供的各种视频应用程序。以下将结合具体的应场景描述如下。The client 30 interacts with the application server 24 so that various video applications provided by the application server 24 can be used. The following will be described below in conjunction with specific scenarios.
参阅图5,在一个具体的应用场景中,某个移动拍摄终端10从位置A移动到了位置B,在此过程中,移动拍摄终端10一直在拍摄并向云端服务器系统20上传带有附加信息视频帧的数据包。云端服务器系统20可以将接收到的数据包转发给客户端。当然,移动拍摄终端10也可以采用点对点的方式将数据包直接发送给客户端。客户端从接收到的数据包中解析出定位信息,根据定位信息可以在电子地图301中生成轨迹线,与此同时,客户端还可以同时输出视频画面302。如此,客户端的用户既可以通过电
子地图301实时了解移动拍摄终端10的位置,又可以通过视频画面302实时观看当前拍摄到的视频画面,实现了移动位置与实时画面的同时观看,尤其适用于对某个目标的实时追踪。但可以理解的是,这种追踪只针对同一个移动摄像终端10且拍摄者身份相同的时候才有意义。Referring to FIG. 5, in a specific application scenario, a certain mobile camera terminal 10 moves from position A to position B. In the process, the mobile camera terminal 10 is always shooting and uploading an additional information video to the cloud server system 20. The packet of the frame. The cloud server system 20 can forward the received data packet to the client. Of course, the mobile camera terminal 10 can also send the data packet directly to the client in a point-to-point manner. The client parses the positioning information from the received data packet, and can generate a trajectory line in the electronic map 301 according to the positioning information, and at the same time, the client can simultaneously output the video screen 302. In this way, the user of the client can pass the electricity
The sub-map 301 knows the position of the mobile shooting terminal 10 in real time, and can view the currently captured video image in real time through the video screen 302, thereby realizing simultaneous viewing of the moving position and the real-time picture, and is particularly suitable for real-time tracking of a certain target. However, it can be understood that such tracking is only meaningful for the same mobile camera terminal 10 and the photographer's identity is the same.
进一步地,当附加信息内包括移动拍摄终端10的姿态信息时,还可以在电子地图301中显示视频的视角。Further, when the posture information of the mobile photographing terminal 10 is included in the additional information, the angle of view of the video can also be displayed in the electronic map 301.
可以理解,在电子地图301中,轨迹线中的每一个点对应于一个坐标,当用户点击轨迹线上的某个点时,可以根据预设的映射关系获取到点击点所对应的坐标,然后就可以在接收到的所有的视频帧中寻找定位信息最接近于该坐标的视频帧,并可将视频画面302切换至该视频帧。It can be understood that, in the electronic map 301, each point in the trajectory corresponds to a coordinate. When the user clicks on a certain point on the trajectory, the coordinates corresponding to the click point can be obtained according to the preset mapping relationship, and then It is possible to find a video frame whose positioning information is closest to the coordinates among all the received video frames, and to switch the video picture 302 to the video frame.
在上述的应用场景中,客户端是在查看某个特定的移动拍摄终端10拍摄的视频,然而,在电子地图中附加视频服务的形式并不限于上述的这种方式,例如,在另一个具体的应用场景中,云端服务器系统20可以提供在线的电子地图服务,向客户端提供电子地图数据。In the above application scenario, the client is viewing the video captured by a particular mobile camera terminal 10, however, the form of attaching the video service in the electronic map is not limited to the above manner, for example, in another specific In the application scenario, the cloud server system 20 can provide an online electronic map service to provide electronic map data to the client.
参阅图6,在客户端30中运行有对应的电子地图应用程序,其从云端服务器系统20中获取电子地图数据并在界面61中加以显示。如图6所示,当被触发时,在界面61中可以显示菜单62,在菜单62中具有各种附加功能,例如“全景预览”的入口,当用户触发“全景预览”功能时,电子地图应用程序获取用户点击位置的坐标,根据该坐标生成预览请求,并将预览请求发送给云端服务器系统20。Referring to FIG. 6, a corresponding electronic map application is run in the client 30, which acquires electronic map data from the cloud server system 20 and displays it in the interface 61. As shown in FIG. 6, when triggered, a menu 62 can be displayed in the interface 61, with various additional functions in the menu 62, such as an entrance of "panoramic preview", when the user triggers the "panoramic preview" function, the electronic map The application acquires the coordinates of the user's click location, generates a preview request based on the coordinates, and sends the preview request to the cloud server system 20.
参阅图7,在云端服务器系统20内处理全景预览请求的流程包括以下步骤:Referring to FIG. 7, the process of processing a panoramic preview request in the cloud server system 20 includes the following steps:
步骤S201,接收全景预览请求。Step S201, receiving a panoramic preview request.
步骤S202,从所述全景预览请求中解析出预览坐标。Step S202, parsing the preview coordinates from the panoramic preview request.
预览坐标就是指用户触发全景预览功能时,电子地图应用程序获取的
用户点击位置的坐标,或者使其他方式标记的点的坐标。The preview coordinates are obtained by the electronic map application when the user triggers the panoramic preview function.
The coordinates at which the user clicks on the position, or the point at which the other way is marked.
步骤S203,根据所述预览坐标以及视频数据的附加信息检索匹配的视频数据。Step S203, retrieving the matched video data according to the preview coordinates and the additional information of the video data.
如上所述,在数据库中存储有所有视频数据的附加信息(至少包括地址信息),因此,可以检索与预览坐标匹配的附加信息,再经由这些附加信息获取对应的视频数据。As described above, additional information (including at least address information) of all the video data is stored in the database, and therefore, additional information matching the preview coordinates can be retrieved, and corresponding video data can be acquired via the additional information.
由于在同一位置,有可能具有多个移动拍摄终端拍摄的视频数据,因此,在搜索结果的呈现时,可以根据一定的顺序,例如,拍摄的时间、图像的清晰度等,对搜索结果进行排序,然后从排序后的结果内获取至少一个移动拍摄终端拍摄的视频数据作为检索结果。Since it is possible to have video data taken by a plurality of mobile shooting terminals at the same position, when the search results are presented, the search results can be sorted according to a certain order, for example, shooting time, image sharpness, and the like. And then acquiring video data captured by at least one mobile shooting terminal from the sorted result as a retrieval result.
值得注意的是,当用户在电子地图中预览全景时,发送的视频数据内可仅包括一帧,而无须将所有的视频数据都进行传输,从而可以降低网络传输量。It is worth noting that when the user previews the panorama in the electronic map, the transmitted video data may include only one frame without transmitting all the video data, thereby reducing the network transmission amount.
此外,在步骤S203的检索步骤中,能够进行检索的视频数据必须是那些已经被拍摄用户授权公开的视频数据。Further, in the retrieval step of step S203, the video data that can be retrieved must be video data that has been authorized by the photographing user to be disclosed.
步骤S204,将检索得到的数据发送给客户端以供展示。Step S204, the retrieved data is sent to the client for display.
客户端在接收到云端服务器系统20发送的视频数据后,根据预定义的协议对视频数据进行解包(如果有压缩)、解密(如果有加密)等处理,然后就可以在界面中进行输出。After receiving the video data sent by the cloud server system 20, the client unpacks (if compressed), decrypts (if encrypted), and the like according to a predefined protocol, and then outputs the content in the interface.
由于在用户在使用全景预览功能时,除了位置,还会涉及方向,因此,在预览请求中,还可以包括用户选择的方向。在此情形下,在步骤S203的检索步骤中,除了检索预览坐标,还必须检索视频的拍摄方向,只有当坐标与方向均匹配的视频数据才能作为匹配的视频数据。Since the direction is involved in the user when the user is using the panoramic preview function, the direction selected by the user may also be included in the preview request. In this case, in the retrieval step of step S203, in addition to retrieving the preview coordinates, it is necessary to retrieve the shooting direction of the video, and only the video data whose coordinates match the direction can be used as the matched video data.
根据上述的实施方式,可以直接基于海量移动拍摄终端拍摄的视频数据实现电子地图中的全景预览功能,而无须采用专门的街景拍摄车去拍摄
街景照片,有效降低实景浏览功能的构建成本。According to the above embodiment, the panoramic preview function in the electronic map can be realized directly based on the video data captured by the mass mobile shooting terminal, without using a special street view shooting car to shoot
Street view photos, effectively reducing the cost of building real-life browsing capabilities.
参阅图8,在另一个具体的应用场景中,云端服务器系统20还提供基于地理位置的视频搜索服务。如图8所示,云端服务器系统20提供视频搜索服务的流程包括以下步骤:Referring to FIG. 8, in another specific application scenario, the cloud server system 20 also provides a geographic location based video search service. As shown in FIG. 8, the flow of providing a video search service by the cloud server system 20 includes the following steps:
步骤S301,接收客户端发送的视频搜索请求;Step S301, receiving a video search request sent by the client;
步骤S302,从该视频搜索请求内解析出搜索坐标以及搜索时间;Step S302, parsing the search coordinates and the search time from the video search request;
步骤S303,根据所述搜索坐标以及搜索时间搜索匹配的视频数据;Step S303, searching for matching video data according to the search coordinates and the search time;
步骤S304,将搜索到的视频数据发送给客户端进行展示;以及Step S304, sending the searched video data to the client for display;
步骤S305,根据用户观看视频过程中的反馈修改不同视频拍摄终端拍摄的视频数据的排序。Step S305, modifying the order of the video data captured by the different video capturing terminals according to the feedback during the user watching the video.
根据这种方式,当某个地点发生热点事件时,无须进行专门的拍摄,只要提供事件发生的地点与时间,就可以获取到对应的视频数据;而根据用户观看过程中的反馈结果修正视频数据的排序又能够保证用户优先观看的是最感兴趣的视频内容。According to this method, when a hot event occurs in a certain place, no special shooting is required, and the corresponding video data can be obtained by providing the location and time of the event; and the video data is corrected according to the feedback result during the user viewing process. The sorting in turn ensures that the user is most interested in viewing the most interesting video content.
根据以上这各种实施方式,主要是基于视频数据内的地理位置信息实现视频数据的检索、匹配,并在此基础上实现具体的视频的应用场景。然而,视频数据内的附加信息并不限于这些应用场景。According to the various implementation manners above, the video data is retrieved and matched based on the geographic location information in the video data, and the specific video application scenario is implemented on the basis of the video data. However, the additional information within the video data is not limited to these application scenarios.
例如,在移动拍摄终端10拍摄并上传视频数据的过程中,用户还可以输入一些指令,例如分享、交通违章举报、报警、报保险、急救等。指令的输入方法可以是直接通过移动拍摄终端10上的按键、触控屏等方式输入,也可以通过与移动拍摄终端10连接的移动电子终端30输入。这些指令的代码可被添加至视频帧的数据包内。相应地,云端服务器系统20在接收到视频帧的数据包后,可以从其中解析出这些指令代码,并执行相应的数据处理过程,或者将这些指令转发给其他的第三方机构的服务器进行进一步的处理。
For example, in the process of the mobile shooting terminal 10 capturing and uploading video data, the user can also input some instructions, such as sharing, traffic violation reporting, alarm, insurance, first aid, and the like. The input method of the command may be directly input through a button on the mobile photographing terminal 10, a touch screen, or the like, or may be input through the mobile electronic terminal 30 connected to the mobile photographing terminal 10. The code for these instructions can be added to the data frame of the video frame. Correspondingly, after receiving the data packet of the video frame, the cloud server system 20 can parse out the instruction code therefrom, and execute the corresponding data processing process, or forward the instructions to the servers of other third-party organizations for further processing. deal with.
如图9所示,此处的第三方机构可包括:自媒体平台、社交网络平台、交通违章处理机构、警方、保险机构、急救机构等。As shown in FIG. 9, the third party institutions herein may include: a self-media platform, a social network platform, a traffic violation handling agency, a police, an insurance institution, a first aid institution, and the like.
例如,当接收到用户的分享指令时,云端服务器系统20可以自动截取预定长度(可从接收到分享指令时刻开始)或者指定长度的视频片断,将其分享至与移动拍摄终端10相绑定的自媒体平台或者社交网络平台(例如微信朋友圈、QQ空间、微博等等)中去。如此,当用户碰到有趣的事件、漂亮的风景时或者其他任意用户想要分享的内容,即可以实现一键将拍摄的视频分享出去。For example, when receiving the user's sharing instruction, the cloud server system 20 can automatically intercept a predetermined length (which can be started from the time of receiving the sharing instruction) or a video clip of a specified length, and share it to the mobile shooting terminal 10. From the media platform or social network platform (such as WeChat friends circle, QQ space, Weibo, etc.). In this way, when the user encounters interesting events, beautiful scenery or any other content that the user wants to share, it is possible to share the captured video with one click.
进一步地,用户对于分享的内容,还可以添加标签、文字介绍等。当移动拍摄终端10包括输入界面时,用户可以直接在移动拍摄终端10上输入,移动拍摄终端10也可不包括输入界面,此时,可以通过与移动拍摄终端10绑定的移动电子终端进行输入。这些标签、文字介绍可被云端服务器系统20存储并用于进行视频的检索。Further, the user may add a label, a text introduction, and the like to the shared content. When the mobile photographing terminal 10 includes an input interface, the user can directly input on the mobile photographing terminal 10, and the mobile photographing terminal 10 may not include the input interface. At this time, the input may be performed by the mobile electronic terminal bound to the mobile photographing terminal 10. These tags, text descriptions can be stored by the cloud server system 20 and used for video retrieval.
当接收到交通违章举报指令时,云端服务器系统20即可自动截取一段相关的视频片段发送至交通违章处理机构的服务器中,如此,用户通过移动拍摄终端10可以实现一键举报交通违法、违章现象。When receiving the traffic violation reporting instruction, the cloud server system 20 can automatically intercept a relevant video segment and send it to the server of the traffic violation processing institution, so that the user can realize one-click reporting of traffic violation and violation by the mobile shooting terminal 10. .
当接收到报警指令时,云端服务器系统20即可自动截取一段相关的视频片段发送至报警受理机构的报警服务器中。如此,用户通过移动拍摄终端10可以实现一键报警,而且由于视频帧数据包内直接包含了定位信息,可以方便警方及时定位报警的位置。When receiving the alarm instruction, the cloud server system 20 can automatically intercept a relevant video segment and send it to the alarm server of the alarm receiving institution. In this way, the user can realize a one-button alarm by moving the shooting terminal 10, and because the video frame data packet directly contains the positioning information, the police can conveniently locate the alarm position in time.
当接收到报保险指令时,云端服务器系统20即可自动截取一段相关的视频片段发送至保险机构的服务器中。如此,用户通过移动拍摄终端10可以实现一键报保险,由于视频可以有效还原现场的实况,保险机构无须去现场即可实现远程的理赔服务。When receiving the insurance instruction, the cloud server system 20 can automatically intercept a relevant video segment and send it to the server of the insurance institution. In this way, the user can realize the one-key insurance by moving the shooting terminal 10. Since the video can effectively restore the live situation on the spot, the insurance institution can realize the remote claim service without going to the scene.
当接收到急救指令时,云端服务器系统20即可自动截取一段相关的
视频片段发送至急救机构的服务器中,如此用户通过移动拍摄终端10即可实现一键呼叫急救服务的功能。而且由于视频帧数据包内直接包含了定位信息,可以让急救机构迅速的定位呼叫位置,减少位置沟通时间。When receiving the first aid instruction, the cloud server system 20 can automatically intercept a relevant section.
The video clip is sent to the server of the emergency department, so that the user can realize the function of one-touch call emergency service by moving the photographing terminal 10. Moreover, since the video frame data packet directly includes the positioning information, the emergency department can quickly locate the call position and reduce the position communication time.
根据上述的实施方式,在视频数据的附加信息内还可以直接嵌入指令,让云端服务器系统20执行对应的数据处理过程及功能,使得视频数据可被更加广泛的应用。According to the above embodiment, the instruction may be directly embedded in the additional information of the video data, so that the cloud server system 20 executes the corresponding data processing process and functions, so that the video data can be more widely applied.
如上所述,用户在分享移动拍摄终端10拍摄的视频时,可以输入标签或者文字介绍,在云端服务器系统20中,基于这些文字介绍,还可以实现热点事件及相关视频的自动发现与挖掘。As described above, when the user shares the video captured by the mobile shooting terminal 10, the user can input a label or a text introduction. In the cloud server system 20, based on the text introduction, automatic discovery and mining of the hot event and related video can also be realized.
参阅图10,基于视频标签与文字介绍的热点事件及相关视频的自动挖掘过程包括以下步骤:Referring to FIG. 10, the automatic mining process of hotspot events and related videos based on video tags and texts includes the following steps:
步骤S301,从接收到的视频标签以及文字介绍中提取出关键词。Step S301, extracting keywords from the received video tags and text introduction.
视频标签一般可以直接作为关键词。文字介绍可进行分词处理、词频统计等步骤,从中提取出关键词。Video tags can generally be used directly as keywords. The text introduction can perform steps such as word segmentation processing and word frequency statistics to extract keywords.
步骤S302,基于地理位置分别统计关键词的出现频率和/或其他参数。In step S302, the frequency of occurrence of the keyword and/or other parameters are separately calculated based on the geographic location.
由于与视频数据同时发送的附加信息内还包括定位信息,如此,就可以基于地理位置统计某个关键词出现的频率以及其他参数。此处的其他参数例如可包括密度、增加的速度等等。Since the additional information transmitted simultaneously with the video data also includes the positioning information, the frequency of occurrence of a certain keyword and other parameters can be counted based on the geographical location. Other parameters herein may include, for example, density, increased speed, and the like.
步骤S303,当某个地点某个关键词出现的频率和/或其他参数超过预设的阈值时,在一个内容发布系统中发布基于该地点与该关键词相关的视频的内容。Step S303, when the frequency and/or other parameters of a certain keyword appearing in a certain place exceed a preset threshold, the content of the video related to the keyword based on the location is released in a content distribution system.
当某个关键词在一个地点出现的频率和/或其他参数高于预设值时,可以视为该地点发生了某个热点事件,或者具有热点的内容。此时可以在一个内容发布系统中发布该地点与该关键词相关的视频内容以供用户浏览。此处的内容发布系统,例如是一个视频网站、APP或者其他内容发布
系统。When the frequency and/or other parameters of a keyword appearing in a place are higher than the preset value, it can be regarded as a hot event or a hot spot. At this time, the video content of the place related to the keyword can be published in a content distribution system for the user to browse. The content publishing system here, for example, a video website, APP or other content publishing
system.
发布的视频内容,即可以是单个移动拍摄终端所拍摄的视频,也可以是多个移动拍摄终端10所拍摄的视频数据经剪辑而成。The released video content may be a video captured by a single mobile shooting terminal, or may be a video data captured by a plurality of mobile shooting terminals 10.
根据上述的这种方式,可以自动发现生活中各种热门的事件、场景、内容,并直接以视频的内容呈现给用户,无须专门的拍摄团队去现场拍摄。According to the above method, various popular events, scenes and contents in life can be automatically found, and the content of the video is directly presented to the user without a special shooting team to shoot on the spot.
以上所述,仅是本发明的较佳实施例而已,并非对本发明作任何形式上的限制,虽然本发明已以较佳实施例揭示如上,然而并非用以限定本发明,任何本领域技术人员,在不脱离本发明技术方案范围内,当可利用上述揭示的技术内容做出些许更动或修饰为等同变化的等效实施例,但凡是未脱离本发明技术方案内容,依据本发明的技术实质对以上实施例所作的任何简介修改、等同变化与修饰,均仍属于本发明技术方案的范围内。
The above is only a preferred embodiment of the present invention, and is not intended to limit the scope of the present invention. The present invention has been disclosed in the preferred embodiments, but is not intended to limit the present invention, and any person skilled in the art. The present invention may be modified or modified to equivalent variations without departing from the technical scope of the present invention, without departing from the scope of the present invention. It is still within the scope of the technical solution of the present invention to make any of the above-mentioned embodiments, the equivalent modifications and modifications of the above embodiments.
Claims (12)
- 一种基于地理位置的视频处理方法,应用于移动拍摄终端中,所述移动拍摄终端包括摄像头,其特征在于,所述方法包括:A method for processing a video based on a geographic location, which is applied to a mobile camera. The mobile camera includes a camera. The method includes:调用所述摄像头拍摄以获取视频数据;Calling the camera to capture video data;根据所述视频数据获取待发送的当前视频帧;Obtaining a current video frame to be sent according to the video data;获取与该当前视频帧对应的附加信息,所述附加信息至少包括所述移动拍摄终端当前的定位信息;Obtaining additional information corresponding to the current video frame, where the additional information includes at least current positioning information of the mobile camera terminal;将所述附加信息添加至所述当前视频帧的数据包内;以及Adding the additional information to a data packet of the current video frame;将添加有附加信息的当前视频帧的数据包发送给云端服务器或者远端的客户端,以使所述云端服务器或者远端的客户端根据接收到的数据包内的定位信息对视频帧进行处理。Sending a data packet of the current video frame to which the additional information is added to the cloud server or the remote client, so that the cloud server or the remote client processes the video frame according to the positioning information in the received data packet. .
- 如权利要求1所述的基于地理位置的视频处理方法,其特征在于,还包括:在拍摄时实时检测所述摄像头的旋转角度;所述附加信息还包括所述摄像头的旋转角度。The method according to claim 1, further comprising: detecting a rotation angle of the camera in real time during shooting; the additional information further comprising a rotation angle of the camera.
- 如权利要求1所述的基于地理位置的视频处理方法,其特征在于,还包括:获取与所述当前视频帧对应的拍摄者的用户身份识别信息;所述附加信息还包括所述用户身份识别信息。The method for processing a geographic location based video according to claim 1, further comprising: acquiring user identification information of a photographer corresponding to the current video frame; the additional information further including the user identification information.
- 如权利要求1所述的基于地理位置的视频处理方法,其特征在于,还包括:在拍摄时实时获取所述当前视频帧的拍摄时间;所述附加信息还包括所述拍摄时间。The geographic location-based video processing method according to claim 1, further comprising: acquiring a shooting time of the current video frame in real time during shooting; the additional information further including the shooting time.
- 如权利要求1所述的基于地理位置的视频处理方法,其特征在于,还包括:获取用户输入的指令;所述附加信息还包括所述拍摄时间。 The method according to claim 1, further comprising: obtaining an instruction input by the user; the additional information further comprising the shooting time.
- 如权利要求1所述的基于地理位置的视频处理方法,其特征在于,还包括:接收与用户输入的标签和/或文字介绍;所述附加信息还包括所述标签和/或文字介绍。The method according to claim 1, further comprising: receiving a label and/or text input input by the user; the additional information further comprising the label and/or text introduction.
- 如权利要求1所述的基于地理位置的视频处理方法,其特征在于,还包括:对所述附加信息的至少部分内容进行加密处理,所述附加信息至少包括经过加密处理的内容。The geographic location-based video processing method of claim 1, further comprising: encrypting at least a portion of the content of the additional information, the additional information including at least the encrypted processed content.
- 如权利要求1所述的基于地理位置的视频处理方法,其特征在于,还包括:根据所述当前帧的视频数据计算对应的校验信息;所述经过加密处理的内容内包括所述校验信息。The method for processing a geographic location based video according to claim 1, further comprising: calculating corresponding verification information according to the video data of the current frame; wherein the content of the encrypted processing includes the verification information.
- 一种基于地理位置的视频处理方法,其特征在于,包括:A geographic location-based video processing method, comprising:接收移动拍摄终端发送的视频数据包;Receiving a video data packet sent by the mobile shooting terminal;从所述视频数据包中解析出视频数据以及对应的附加信息,所述附加信息至少包括所述移动拍摄终端当前的定位信息;And parsing the video data and the corresponding additional information from the video data packet, where the additional information includes at least the current positioning information of the mobile shooting terminal;存储所述视频数据并获取对应的索引;Storing the video data and obtaining a corresponding index;将所述附加信息与所述索引关联存储;以及Storing the additional information in association with the index;基于所述附加信息检索并获取对应的视频数据,并基于检索到的视频数据向客户端提供视频数据服务。Retrieving and acquiring corresponding video data based on the additional information, and providing a video data service to the client based on the retrieved video data.
- 如权利要求9所述的基于地理位置的视频处理方法,其特征在于,还包括:从所述视频数据包中解析出与每个视频帧对应的指令;并执行与所述指令对应的数据处理过程。 The geographic location-based video processing method according to claim 9, further comprising: parsing an instruction corresponding to each video frame from the video data packet; and performing data processing corresponding to the instruction process.
- 如权利要求10所述的基于地理位置的视频处理方法,其特征在于,所述数据处理过程包括:The method of processing a geographic location based video according to claim 10, wherein the data processing process comprises:截取预定长度或者指定长度的视频片断,将其分享至与所述移动拍摄终端相绑定的自媒体平台或者社交网络平台;或者Capturing a video clip of a predetermined length or a specified length and sharing it to a self-media platform or a social networking platform bound to the mobile camera terminal; or截取一段相关的视频片段发送至交通违章处理机构的服务器中;或者Intercepting a relevant video clip to be sent to the server of the traffic violation handling agency; or自动截取一段相关的视频片段发送至报警受理机构的报警服务器中;或者Automatically intercepting a related video clip and sending it to the alarm server of the alarm receiving institution; or自动截取一段相关的视频片段发送至急救机构的服务器中。Automatically intercepts a related video clip and sends it to the server of the emergency department.
- 如权利要求9所述的基于地理位置的视频处理方法,其特征在于,还包括:从所述视频数据包中解析出视频的标签和/或文字介绍;The method according to claim 9, further comprising: parsing a label and/or a text of the video from the video data package;将所述标签和/或文字介绍存储;Store the label and/or text description;从所述标签和/或文字介绍中提取出关键词;Extracting keywords from the label and/or text introduction;基于地理位置对所述关键词进行统计分析;Perform statistical analysis on the keywords based on geographic location;当某个地理位置的某个关键词的统计值超过预设阈值时获取相关的视频数据,并在一个内容发布系统中发布所述获取的相关的视频数据。 When the statistical value of a certain keyword of a certain geographical location exceeds a preset threshold, relevant video data is acquired, and the acquired related video data is published in a content distribution system.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610147581.0 | 2016-03-15 | ||
CN201610147581.0A CN105812660A (en) | 2016-03-15 | 2016-03-15 | Video processing method based on geographic position |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017156793A1 true WO2017156793A1 (en) | 2017-09-21 |
Family
ID=56468429
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2016/077182 WO2017156793A1 (en) | 2016-03-15 | 2016-03-24 | Geographic location-based video processing method |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN105812660A (en) |
WO (1) | WO2017156793A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110163050A (en) * | 2018-07-23 | 2019-08-23 | 腾讯科技(深圳)有限公司 | A kind of method for processing video frequency and device, terminal device, server and storage medium |
CN113222637A (en) * | 2021-02-26 | 2021-08-06 | 深圳前海微众银行股份有限公司 | Architecture method, device, equipment, medium and program product of store visitor information |
CN113704554A (en) * | 2021-07-13 | 2021-11-26 | 湖南中惠旅智能科技有限责任公司 | Video retrieval method and system based on electronic map |
CN114040006A (en) * | 2021-11-01 | 2022-02-11 | 达维信息技术(深圳)有限公司 | Multi-mobile-terminal data sharing method and system based on digital asset management |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108460037A (en) * | 2017-02-20 | 2018-08-28 | 北京金奔腾汽车科技有限公司 | A method of stroke video is preserved and retrieved based on geographical location |
CN110019628B (en) * | 2017-12-27 | 2023-12-29 | 努比亚技术有限公司 | Positioning method, mobile terminal and computer readable storage medium |
CN108109188B (en) * | 2018-01-12 | 2022-02-08 | 维沃移动通信有限公司 | An image processing method and mobile terminal |
CN108833767A (en) * | 2018-03-28 | 2018-11-16 | 深圳市语图科技有限公司 | A kind of positioning system and method applied to record motion profile |
CN112004046A (en) * | 2019-05-27 | 2020-11-27 | 中兴通讯股份有限公司 | Image processing method and device based on video conference |
CN111327860A (en) * | 2020-01-21 | 2020-06-23 | 成都纵横自动化技术股份有限公司 | Synchronous transmission method for figures and electronic equipment |
CN111353168A (en) * | 2020-02-27 | 2020-06-30 | 闻泰通讯股份有限公司 | Multimedia file management method, device, equipment and storage medium |
CN111444385B (en) * | 2020-03-27 | 2023-03-03 | 西安应用光学研究所 | Electronic map real-time video mosaic method based on image corner matching |
CN111770107B (en) * | 2020-07-07 | 2022-11-22 | 广州通达汽车电气股份有限公司 | Streaming media transmission method, system, storage medium and computer equipment for bearing dynamic data |
CN113505259A (en) * | 2021-06-28 | 2021-10-15 | 惠州Tcl云创科技有限公司 | Media file labeling method, device, equipment and medium based on intelligent identification |
CN115695924A (en) * | 2021-07-30 | 2023-02-03 | 瑞庭网络技术(上海)有限公司 | Data processing method, client, server, and computer-readable recording medium |
CN114326764A (en) * | 2021-11-29 | 2022-04-12 | 上海岩易科技有限公司 | Rtmp transmission-based smart forestry unmanned aerial vehicle fixed-point live broadcast method and unmanned aerial vehicle system |
CN114422856B (en) * | 2022-01-07 | 2024-06-04 | 北京达佳互联信息技术有限公司 | Video data verification method, device, equipment and storage medium |
CN115455275B (en) * | 2022-11-08 | 2023-02-03 | 广东卓维网络有限公司 | Video processing system integrated with inspection equipment |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6741790B1 (en) * | 1997-05-29 | 2004-05-25 | Red Hen Systems, Inc. | GPS video mapping system |
CN1551664A (en) * | 2003-04-14 | 2004-12-01 | ��ʽ����Ntt����Ħ | Mobile communication system, mobile communication terminal and its program |
CN101610376A (en) * | 2008-06-17 | 2009-12-23 | 三星电子株式会社 | Method and apparatus for recording and reproducing moving pictures with position information |
CN103716584A (en) * | 2013-11-30 | 2014-04-09 | 南京大学 | Context sensing-based intelligent mobile terminal field monitoring method |
CN103984710A (en) * | 2014-05-05 | 2014-08-13 | 深圳先进技术研究院 | Video interaction inquiry method and system based on mass data |
CN105022801A (en) * | 2015-06-30 | 2015-11-04 | 北京奇艺世纪科技有限公司 | Hot video mining method and hot video mining device |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101867730B (en) * | 2010-06-09 | 2011-11-16 | 马明 | Multimedia integration method based on user trajectory |
CN102289520A (en) * | 2011-09-15 | 2011-12-21 | 山西四和交通工程有限责任公司 | Traffic video retrieval system and realization method thereof |
CN103686239A (en) * | 2013-12-11 | 2014-03-26 | 深圳先进技术研究院 | Network sharing crime evidence collection system and method based on positioning video |
CN104679873A (en) * | 2015-03-09 | 2015-06-03 | 深圳市道通智能航空技术有限公司 | Aircraft tracing method and aircraft tracing system |
-
2016
- 2016-03-15 CN CN201610147581.0A patent/CN105812660A/en active Pending
- 2016-03-24 WO PCT/CN2016/077182 patent/WO2017156793A1/en active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6741790B1 (en) * | 1997-05-29 | 2004-05-25 | Red Hen Systems, Inc. | GPS video mapping system |
CN1551664A (en) * | 2003-04-14 | 2004-12-01 | ��ʽ����Ntt����Ħ | Mobile communication system, mobile communication terminal and its program |
CN101610376A (en) * | 2008-06-17 | 2009-12-23 | 三星电子株式会社 | Method and apparatus for recording and reproducing moving pictures with position information |
CN103716584A (en) * | 2013-11-30 | 2014-04-09 | 南京大学 | Context sensing-based intelligent mobile terminal field monitoring method |
CN103984710A (en) * | 2014-05-05 | 2014-08-13 | 深圳先进技术研究院 | Video interaction inquiry method and system based on mass data |
CN105022801A (en) * | 2015-06-30 | 2015-11-04 | 北京奇艺世纪科技有限公司 | Hot video mining method and hot video mining device |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110163050A (en) * | 2018-07-23 | 2019-08-23 | 腾讯科技(深圳)有限公司 | A kind of method for processing video frequency and device, terminal device, server and storage medium |
CN110163050B (en) * | 2018-07-23 | 2022-09-27 | 腾讯科技(深圳)有限公司 | Video processing method and device, terminal equipment, server and storage medium |
CN113222637A (en) * | 2021-02-26 | 2021-08-06 | 深圳前海微众银行股份有限公司 | Architecture method, device, equipment, medium and program product of store visitor information |
CN113704554A (en) * | 2021-07-13 | 2021-11-26 | 湖南中惠旅智能科技有限责任公司 | Video retrieval method and system based on electronic map |
CN113704554B (en) * | 2021-07-13 | 2024-03-29 | 湖南中惠旅智能科技有限责任公司 | Video retrieval method and system based on electronic map |
CN114040006A (en) * | 2021-11-01 | 2022-02-11 | 达维信息技术(深圳)有限公司 | Multi-mobile-terminal data sharing method and system based on digital asset management |
CN114040006B (en) * | 2021-11-01 | 2024-02-27 | 北京流通宝数据科技服务有限公司 | Multi-mobile terminal data sharing method and system based on digital asset management |
Also Published As
Publication number | Publication date |
---|---|
CN105812660A (en) | 2016-07-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2017156793A1 (en) | Geographic location-based video processing method | |
US12189827B2 (en) | Systems and methods for authenticating photographic image data | |
US10019774B2 (en) | Authentication and validation of smartphone imagery | |
CN105975570B (en) | Video searching method and system based on geographical location | |
JP6521864B2 (en) | Method using portable electronic device, portable electronic device, and computer program | |
CN105827959A (en) | Geographic position-based video processing method | |
CN105933651B (en) | Method and apparatus based on target route jumper connection video | |
US9792287B2 (en) | Method and apparatus for providing metadata search codes to multimedia | |
US10873846B2 (en) | Peripheral video presence detection | |
CN112866222A (en) | Data processing method and device and data processing device | |
CN115834584A (en) | Cross-network data transmission method, device, equipment and medium | |
JP2017220728A (en) | Verification system, information processor, and verification method | |
CN104967987A (en) | Method and server for mobile terminal positioning | |
US20150113039A1 (en) | Method and apparatus for defining hot spot based task for multimedia data | |
KR101465186B1 (en) | Method for data accessing among remote devices | |
US20250097696A1 (en) | Message communication method, and device | |
US20250078548A1 (en) | Authentication of an event through certification and verification of a computer file | |
WO2025108339A1 (en) | Satellite communication method and related apparatus | |
JP5793916B2 (en) | Distribution server device and distribution method | |
KR101476800B1 (en) | Apparatus and method for managing object based on panorama media | |
GB2563087A (en) | Methods, systems and devices for image acquisition | |
WO2020164726A1 (en) | Mobile communications device and media server | |
HK1200941B (en) | Method and apparatus for providing metadata search codes to multimedia |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 16893958 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 11.03.2019) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 16893958 Country of ref document: EP Kind code of ref document: A1 |