CN107729486B - Video searching method and device - Google Patents
Video searching method and device Download PDFInfo
- Publication number
- CN107729486B CN107729486B CN201710964471.8A CN201710964471A CN107729486B CN 107729486 B CN107729486 B CN 107729486B CN 201710964471 A CN201710964471 A CN 201710964471A CN 107729486 B CN107729486 B CN 107729486B
- Authority
- CN
- China
- Prior art keywords
- video
- index
- field
- subfield
- tag
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 52
- 238000012545 processing Methods 0.000 claims description 19
- 230000011218 segmentation Effects 0.000 claims description 12
- 230000010354 integration Effects 0.000 claims description 8
- 238000012986 modification Methods 0.000 claims description 4
- 230000004048 modification Effects 0.000 claims description 4
- 238000007792 addition Methods 0.000 claims description 3
- 238000012217 deletion Methods 0.000 claims description 3
- 230000037430 deletion Effects 0.000 claims description 3
- 238000010586 diagram Methods 0.000 description 4
- 230000004927 fusion Effects 0.000 description 4
- 230000008030 elimination Effects 0.000 description 2
- 238000003379 elimination reaction Methods 0.000 description 2
- 235000010627 Phaseolus vulgaris Nutrition 0.000 description 1
- 244000046052 Phaseolus vulgaris Species 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000007123 defense Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 235000005911 diet Nutrition 0.000 description 1
- 230000037213 diet Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/73—Querying
- G06F16/735—Filtering based on additional data, e.g. user or group profiles
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/73—Querying
- G06F16/738—Presentation of query results
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/7867—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Library & Information Science (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The application provides a video searching method and a device, which can enlarge the matching range of analysis results by matching the obtained analysis results with index fields generated by integrating information of all tags, thereby increasing the number of video searching results presented to users.
Description
Technical Field
The invention relates to the technical field of internet information, in particular to a video searching method and device.
Background
With the popularization and development of video network application, a plurality of video websites emerge, users can conveniently search videos on the video websites for watching, and the lives of the users are greatly enriched.
At present, a video website allows a user to search videos by adopting a video search method, which mainly includes parsing a search request of the user after receiving the search request, determining all video results meeting the search request of the user from an index field corresponding to a hit tag when the parsed result hits a certain tag in a preset configuration file, and further displaying all determined video results to the user. However, since the plurality of tags contained in the configuration file are independent from each other, and each tag has an index field corresponding to the tag, when the existing video search method is used for searching the video, all tags corresponding to the parsing result cannot be hit at the same time, so that the finally determined video result is only the video result belonging to the index field corresponding to the hit tag, the utilization rate of the tags in the configuration file is reduced, and the recall rate of the video result is further reduced.
Disclosure of Invention
In view of this, the present invention provides a video search method and apparatus, which improve the utilization rate of tags in configuration files, and further improve the recall rate of video results.
In order to achieve the purpose, the invention provides the following technical scheme:
a video search method, comprising:
when a video search request of a user is received, analyzing the video search request to obtain an analysis result;
matching fields corresponding to the analysis result from index fields according to the analysis result, wherein the index fields are generated after information integration is carried out on all the included labels;
and matching the video corresponding to the field from a video library according to the field corresponding to the analysis result to serve as a video search result.
Preferably, the process of generating the index field includes:
obtaining a plurality of labels, wherein each label comprises a label type and label related information;
classifying all the tags into corresponding entity types by using the tag type of each tag, wherein the number of the entity types is at least one;
removing the duplicate of the label with the same label related information in each entity type, and taking the label related information obtained after the duplicate removal as an index subfield;
generating an association relation between each video in the video library and the corresponding index subfield according to the video information of each video in the video library and the entity type, and storing the association relation in the corresponding index subfield;
and arranging and combining all index sub-fields storing the association relation to generate the index field.
Preferably, the generating an association relationship between each video in the video library and the corresponding index subfield according to the video information of each video in the video library and the entity type, and storing the association relationship in the corresponding index subfield includes:
when the target entity type is the first entity type, comparing a tag field in video information of each video in the video library with each index subfield contained in the target entity type, and judging whether the tag field is the same as the index subfield;
and if the label field is the same as the index subfield, generating an association relation between the video and the corresponding index subfield, and storing the association relation in the corresponding index subfield.
Preferably, the generating an association relationship between each video in the video library and the corresponding index subfield according to the video information of each video in the video library and the entity type, and storing the association relationship in the corresponding index subfield includes:
when the target entity type is a second entity type, performing word segmentation on a specific field in video information of each video in the video library to generate at least one specific subfield;
comparing all the specific subfields with each index subfield contained in the target entity type, and judging whether the specific subfields are the same as the index subfields;
and if the specific sub-field is the same as the index sub-field, generating an association relation between the video and the corresponding index sub-field, and storing the association relation in the corresponding index sub-field.
Preferably, the matching a field corresponding to the parsing result from an index field according to the parsing result includes:
and matching an index sub-field corresponding to the analysis result from an index field according to the analysis result.
Preferably, the matching out the video corresponding to the field from the video library according to the field corresponding to the parsing result as a video search result includes:
and matching the video with the index sub-field in association relationship from the video library according to the index sub-field corresponding to the analysis result, and taking the video as the video search result.
Preferably, after the acquiring the plurality of tags, the method further comprises:
receiving a tag processing instruction, and performing processing operation on a tag corresponding to the tag processing instruction, wherein the processing operation comprises any one or combination of more than one of addition, deletion, modification and query.
A video search apparatus, comprising:
the analysis module is used for analyzing the video search request to obtain an analysis result when the video search request of a user is received;
the field matching module is used for matching a field corresponding to the analysis result from an index field according to the analysis result, wherein the index field is generated after information integration is carried out on all the included labels;
and the video matching module is used for matching the video corresponding to the field from the video library according to the field corresponding to the analysis result and taking the video as a video search result.
Preferably, the apparatus further comprises:
the system comprises an acquisition module, a processing module and a display module, wherein the acquisition module is used for acquiring a plurality of labels, and each label comprises a label type and label related information;
the classification module is used for classifying all the labels into corresponding entity types by using the label type of each label, and the number of the entity types is at least one;
a deduplication module, configured to perform deduplication on tags with the same tag related information in each entity type, and use the tag related information obtained after deduplication as an index subfield;
the generating module is used for generating an association relation between each video in the video library and the corresponding index subfield according to the video information of each video in the video library and the entity type, and storing the association relation in the corresponding index subfield;
and the permutation and combination module is used for carrying out permutation and combination on all the index sub-fields storing the association relation to generate the index fields.
Preferably, the generating module comprises:
a first judging unit, configured to, when a target entity type is a first entity type, compare a tag field in video information of each video in the video library with each index subfield included in the target entity type, and judge whether the tag field is the same as the index subfield;
and the first generation unit is used for generating the association relation between the video and the corresponding index subfield and storing the association relation in the corresponding index subfield after the first judgment unit judges that the tag field is the same as the index subfield.
Preferably, the generating module comprises:
the word segmentation unit is used for segmenting a specific field in the video information of each video in the video library to generate at least one specific subfield when the target entity type is the second entity type;
a second judging unit, configured to compare all the specific subfields with each of the index subfields included in the target entity type, and judge whether the specific subfields are the same as the index subfields;
and a second generating unit, configured to generate an association relationship between the video and the corresponding index subfield and store the association relationship in the corresponding index subfield after the second determining unit determines that the specific subfield is the same as the index subfield.
Preferably, the field matching module includes:
and the field matching submodule is used for matching an index subfield corresponding to the analysis result from the index field according to the analysis result.
Preferably, the video matching module comprises:
and the video matching sub-module is used for matching a video which has an association relation with the index sub-field from the video library according to the index sub-field corresponding to the analysis result after the index sub-field corresponding to the analysis result is matched from the index field by the field matching sub-module according to the analysis result, and taking the video as the video search result.
It can be seen from the above technical solutions that, compared with the prior art, the present invention provides a video search method and apparatus, by matching an obtained parsing result with an index field generated by integrating information of all tags, the matching range of the parsing result can be expanded on the basis of improving the utilization rate of all tags, thereby increasing the number of video search results presented to a user, and as a result, matching of the parsing result is completed by using the index field generated by integrating information of all tags included, so that the problem that all tags corresponding to the parsing result cannot be hit at the same time due to mutual independence among a plurality of tags and respective index fields corresponding thereto can be effectively solved, thereby improving the recall rate of a video on the basis of fully utilizing all tags.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the provided drawings without creative efforts.
Fig. 1 is a flowchart of a method of searching a video according to an embodiment of the present invention;
fig. 2 is a flowchart of a method for generating an index field according to an embodiment of the present invention;
FIG. 3 is a flowchart of another method for generating an index field according to an embodiment of the present invention;
FIG. 4 is a flowchart of another video search method according to an embodiment of the present invention;
fig. 5 is a schematic structural diagram of a video search apparatus according to an embodiment of the present invention;
fig. 6 is a schematic structural diagram of an apparatus for generating an index field according to an embodiment of the present invention;
fig. 7 is a schematic structural diagram of another apparatus for generating an index field according to an embodiment of the present invention;
fig. 8 is a schematic structural diagram of another video search apparatus according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The embodiment of the invention discloses a video searching method, please refer to the attached figure 1, and the method specifically comprises the following steps:
s101: when a video search request of a user is received, analyzing the video search request to obtain an analysis result;
specifically, the video search request of the user may be text which is input by the user according to the video that the user wants to search and is related to the video content and meets a preset input rule, for example, any one or more combinations of a certain keyword, a video name, a video source country, and the like appearing in the video content, and the preset input rule may set different input rules according to different video websites, which is not limited herein.
The method for analyzing the video search request can be used for performing word segmentation on the received video search request to obtain an analysis result, for example, performing word segmentation on a disaster film of the video search request to obtain two words of the disaster and the film, wherein the two words are used as the analysis result of the video search and are used for performing related matching operation in a video website subsequently.
S102: matching fields corresponding to the analysis result from index fields according to the analysis result, wherein the index fields are generated after information integration is carried out on all the included labels;
specifically, the index field can be established in advance and is mainly used for storing all fields obtained after information integration is carried out on all tags, so that after an analysis result is obtained, matching can be directly carried out from the index field, the aim of carrying out matching one by one from all the included tags is indirectly achieved, the problem that all tags of corresponding analysis results cannot be hit at the same time due to the fact that only one tag can be queried in each search is effectively solved, and the video recall rate is improved; meanwhile, only one index field is established in the video website and is used for matching with the analysis result, so that the time required by the process of matching the index field with the analysis result is shortened, and the video searching efficiency is improved.
The matching method of matching a field corresponding to the parsing result obtained in S101 from the index field may be to match the same field as the parsing result from the index field. Taking two analysis results of 'disaster' and 'piece' as an example for specific explanation, if five fields of 'comedy', 'warmth', 'racing', 'disaster' and 'sports' are contained in the pre-established index field, the 'disaster' and the 'piece' are respectively matched with all the fields contained in the index field one by one, so as to obtain the field 'disaster' identical to the analysis result 'disaster', therefore, the field 'disaster' is used as the field of the corresponding analysis result in the video search for subsequent video matching operation.
The order of matching the analysis result with the index field may be sequential matching from front to back according to the arrangement order of all fields contained in the index field, or sequential reverse matching from back to front.
If the field corresponding to the analysis result cannot be matched in the index field, a search failure message can be generated so as to prompt the user that the video result which does not conform to the input video search request in the video website is not available. The generated search failure information may be information indicating that there is no search result, such as "the search result is 0", or information indicating that the video search has failed this time, such as "the video search has failed".
S103: matching the video corresponding to the field from a video library according to the field corresponding to the analysis result to serve as a video search result;
specifically, if the index field includes a field corresponding to the parsing result, it is proved that one or more videos corresponding to the parsing result can be matched from a video library in the video website. The video library can be pre-established and is mainly used for storing videos, and the videos in the video library and the fields in the index fields have corresponding relations, so that all videos having corresponding relations with the fields can be quickly matched from the video library according to the fields corresponding to the analysis results and serve as final video search results to be presented to users in the following.
For example, the analysis result "car" obtained after the analysis is used to match one by one from the index fields which are established in advance and contain five fields of "comedy", "warmth", "car", "disaster" and "sport", so as to match the "car" field in the index field as the field corresponding to the analysis result in the video search, and further match the "video a", "video B" and "video C" which have a corresponding relationship with the "car" field from the video library according to the "car" field, and the video a "," video B "and" video C "are used as the video search result and are finally presented to the user, thereby increasing the ratio of the number of video search results searched from the video library at one time, i.e. increasing the video recall rate.
It should be noted that the videos stored in the video library may be one or a combination of more of related link addresses of videos, related poster pictures of videos, video profiles, and the like. Accordingly, the search results ultimately presented to the user are a combination of one or more of the relevant link addresses for viewing the video, the relevant poster pictures of the video, the video vignette, and the like.
According to the video searching method disclosed by the embodiment of the invention, the obtained analysis result is matched with the index field generated by integrating the information of all the tags, so that the matching range of the analysis result is expanded on the basis of improving the utilization rate of all the tags, the number of video searching results presented to a user is increased, therefore, the matching of the analysis result is completed by utilizing the index field generated by integrating the information of all the tags, the problem that all the tags corresponding to the analysis result cannot be hit at the same time due to the fact that a plurality of tags are independent from each other and respectively have the index fields corresponding to the tags can be effectively solved, and the recall rate of the video is improved on the basis of fully utilizing all the tags.
According to the obtained analysis result, matching the field corresponding to the analysis result from the index field is an important step for quickly matching the corresponding video from the video library, and the pre-generated index field is an important factor influencing the step of matching the field corresponding to the analysis result. Therefore, how to generate the index field quickly and accurately is also a focus of attention in the present solution.
Therefore, as shown in fig. 2, for S102 in the embodiment corresponding to fig. 1, the embodiment of the present invention discloses a method for generating an index field, which specifically includes the following steps:
s201: obtaining a plurality of labels, wherein each label comprises a label type and label related information;
specifically, the tags are mainly used for reflecting the characteristics of the videos, wherein the types of the tags included in the tags are mainly used for indicating the categories to which the tags reflect the videos from multiple dimensions, such as "comedy", "say", "usa", "japanese", and the like; the tag related information contained in the tag may be a keyword set from multiple dimensions according to the content of the video, such as a keyword "speed and passion" set from the dimension of the video name, "2016" set from the dimension of the video showing year, a keyword "zhangxie" set from the dimension of the video director, and so on, each tag printed on the video contains the tag type and tag related information, thereby facilitating to quickly embody the search range to speed up the search.
Aiming at the same video, a plurality of different labels can be marked on the same video so as to realize the characteristic of the video from a plurality of dimensions; accordingly, the source of the tag marked on the video can be multiple, such as a bean tag, a diet grand tag, and the like.
The scheme of obtaining the plurality of tags is not limited, and the tags can be obtained from various tag sources by using a web crawler.
S202: classifying all the tags into corresponding entity types by using the tag type of each tag, wherein the number of the entity types is at least one;
specifically, a plurality of different entity types can be pre-established in the video website and are mainly used for storing the acquired tags, each entity type can be any one of a channel type, a version type, a region type, a language type, a media type, a common tag type and the like, and the stored entity type and the tag type of the acquired tag have an association relationship, so that each tag can be quickly classified into the entity type having the association relationship according to the tag type of the tag, and basic data is provided for subsequently establishing the index subfield.
For example, the obtained tags are "tag a", "tag B", "tag C", "tag D", and "tag E", respectively, where the tag type of "tag a" is "comedy", "the tag type of" tag B "is" usa "," the tag type of "tag C" is "disaster", "the tag type of" tag D "is" japanese "," the tag type of "tag E" is "hunt-south satellite", and the pre-established entity types include "channel type", "version type", "region type", "language type", and "media type", then "tag a" and "tag C" can be quickly categorized into the entity type "channel type" according to the association relationship between the tag type "comedy", "disaster" and the entity type "channel type", and "tag C" can be categorized into the entity type "channel type" according to the association relationship between the tag type "usa" and the entity type "region type", the label B is quickly classified into the entity type region type, the label D is quickly classified into the entity type language type according to the incidence relation between the label type Japanese and the entity type language type, and the label E is quickly classified into the entity type media type according to the incidence relation between the label type Hunan defense and the entity type media type.
S203: removing the duplicate of the label with the same label related information in each entity type, and taking the label related information obtained after the duplicate removal as an index subfield;
specifically, since multiple tags belonging to the same entity type may have the same tag-related information, the tags having the same tag-related information and acquired from multiple sources need to be deduplicated, so that the multiple tags are classified and fused, and only the next tag-related information is reserved as the index subfield.
For example, the entity type "language type" includes "tag a", "tag B", and "tag C", where the tag related information of "tag a" is "adventure", "tag related information of" tag B "is" friendship ", and" tag C "is" adventure ", and the tag related information" adventure "is deduplicated, so that two tag related fields of" adventure "and" friendship "are obtained and are respectively used as an index subfield for subsequently establishing the index field.
S204: when the target entity type is the first entity type, comparing a tag field in video information of each video in the video library with each index subfield contained in the target entity type, and judging whether the tag field is the same as the index subfield, if so, executing S205, and if not, executing S206;
specifically, each video in the video library has video information containing its own characteristics, and the video information contains one or more tag-related fields, i.e., tag fields, marked for the video information by a tag source; the first entity type is an entity type which contains each index subfield, namely label related information obtained after de-duplication, and is consistent with the label field in the video information contained in the video library, so that the aim of sequentially determining the index subfield corresponding to each video in the video library in all the entity types belonging to the first entity type can be quickly realized by judging whether the label field is the same as the index subfield. The target entity type is any entity type.
S205: generating an association relation between the video and the corresponding index subfield, storing the association relation in the corresponding index subfield, and executing S207;
specifically, if the tag field is judged to be the same as the index subfield, the association relationship between the index subfield and the video is established, so that all videos corresponding to the video search request of the user in the video library can be matched quickly in the following process.
S206: and generating video search failure information.
S207: arranging and combining all index sub-fields storing the association relation to generate the index fields;
specifically, all index sub-fields with the association relationship stored are arranged and combined to obtain a complete index field which is used as the connection between the subsequent analysis result and the videos in the video library, so that the number of videos presented to a user and the video search efficiency are improved on the basis of automatically integrating all the tags. The arrangement and combination mode of all index sub-fields storing the association relationship is not limited in the scheme, and the arrangement and combination can be performed according to the sequence obtained by the index sub-fields.
The above steps S204 to S205 are only a preferred implementation manner of the process of "generating an association relationship between each video in the video library and the corresponding index subfield according to the video information of each video in the video library and the entity type, and storing the association relationship in the corresponding index subfield" disclosed in the embodiment of the present invention, and a specific implementation manner related to this process may be arbitrarily set according to actual requirements, which is not limited herein.
In the embodiment of the invention, the obtained tags are classified into the corresponding entity types according to the tag types, so that the incidence relation between the video and the corresponding index subfields is conveniently established subsequently by adopting the same establishing mode aiming at the same entity type, the generation speed of the index field is accelerated, all tags in each entity type are subjected to de-duplication to obtain different tag related information as the index subfields, the classification and fusion of the tags can be realized, the comprehensiveness of the index fields generated by permutation and combination is ensured, and the video recall rate is indirectly improved. And when the target entity type is the first entity type and the tag field and the index sub-field are judged to be the same, establishing an association relation between the video corresponding to the tag field and the index sub-field, and storing the association relation, so that the video searching rate can be effectively increased, and the video searching efficiency can be effectively improved.
After S201 in the embodiment corresponding to fig. 2, the method further includes:
receiving a tag processing instruction, and performing processing operation on a tag corresponding to the tag processing instruction, wherein the processing operation comprises any one or combination of more than one of addition, deletion, modification and query;
specifically, the tag processing instruction may be an instruction set by a developer for a processing operation to be executed by the currently acquired tag, such as a "tag adding instruction", a "tag deleting instruction", a "tag modifying instruction", and a "tag querying instruction", which is beneficial to improving the accuracy of the index field.
In the embodiment of the invention, by receiving the tag processing instruction and processing the tag corresponding to the tag processing instruction, the accuracy of the index field can be improved, the video searching precision is further improved, and the occurrence probability of video searching failure is effectively reduced.
As for S102 in the embodiment corresponding to fig. 1, as shown in fig. 3, the embodiment of the present invention discloses another method for generating an index field, where the method specifically includes the following steps:
s301: a plurality of tags are obtained, each tag containing a tag type and tag related information.
S302: and classifying all the labels into corresponding entity types by using the label type of each label, wherein the number of the entity types is at least one.
S303: and carrying out deduplication on the labels with the same label related information in each entity type, and taking the label related information obtained after the deduplication as an index subfield.
S304: when the target entity type is a second entity type, performing word segmentation on a specific field in the video information of each video in the video library to generate at least one specific subfield, and executing S305;
specifically, the second entity type may be an entity type in which each included index subfield, that is, the tag related information obtained after deduplication, is inconsistent with a tag field in video information included in video in the video library, so that the corresponding index subfield cannot be matched by using the tag field in the video information. At this time, the specific fields in the video information of each video in the video library can be participled, and the generated specific subfields are used for matching with all index subfields in each entity type belonging to the second entity type one by one, so that the matching probability is increased, the number of videos in the video library corresponding to each index subfield is increased, and the video recall rate is indirectly increased.
The specific field within the video information may be a video name, a language field in the video, a play media field, etc.
S305: comparing all the specific subfields with each index subfield contained in the target entity type, and judging whether the specific subfields are the same as the index subfields, if so, executing S306, otherwise, executing S307;
specifically, each specific subfield generated after word segmentation is sequentially compared with each index subfield contained in the target entity type, so that the purpose of sequentially determining the index subfield corresponding to each video in the video library in all entity types belonging to the second entity type can be quickly achieved by judging whether the specific subfield is the same as the index subfield.
S306: and generating an association relationship between the video and the corresponding index subfield, storing the association relationship in the corresponding index subfield, and executing S308.
S307: and generating video search failure information.
S308: and arranging and combining all index sub-fields storing the association relation to generate the index field.
The above steps S304 to S306 are only a preferred implementation manner of the process of "generating an association relationship between each video in the video library and the corresponding index subfield according to the video information of each video in the video library and the entity type, and storing the association relationship in the corresponding index subfield" disclosed in the embodiment of the present invention, and a specific implementation manner related to this process may be arbitrarily set according to actual requirements, which is not limited herein.
In the embodiment of the invention, the obtained tags are classified into the corresponding entity types according to the tag types, so that the incidence relation between the video and the corresponding index subfields is conveniently established subsequently by adopting the same establishing mode aiming at the same entity type, the generation speed of the index field is accelerated, all tags in each entity type are subjected to de-duplication to obtain different tag related information as the index subfields, the classification and fusion of the tags can be realized, the comprehensiveness of the index fields generated by permutation and combination is ensured, and the video recall rate is indirectly improved. And when the target entity type is the second entity type and the specific subfield after word segmentation is judged to be the same as the index subfield, establishing and storing an association relation between the video corresponding to the specific subfield and the index subfield, so that the video searching rate can be effectively accelerated, and the video searching efficiency is effectively improved.
On the basis of the embodiment shown in fig. 2, another video search method is disclosed in the embodiment of the present invention, please refer to fig. 4, where the method specifically includes the following steps:
s401: when a video search request of a user is received, analyzing the video search request to obtain an analysis result.
S402: and matching index sub-fields corresponding to the analysis result from the index fields according to the analysis result, wherein the index fields are generated after information integration is carried out on all the included labels.
S403: matching videos having an association relation with the index sub-fields from the video library according to the index sub-fields corresponding to the analysis results to serve as video search results;
specifically, the index field is composed of a plurality of index sub-fields, and a relationship is established between each index sub-field and a video in the video library, so that matching can be performed from the index field according to an obtained analysis result, so as to determine the same index sub-field, and then according to the relationship stored in the index sub-field, all videos having a relationship with the index sub-field can be quickly matched from the video library to serve as a video search result which is finally required to be presented to a user.
According to the video searching method disclosed by the embodiment of the invention, the obtained analysis result is matched with the index field generated by integrating the information of all the tags, so that the index subfield corresponding to the obtained analysis result can be determined, and all videos corresponding to the index subfield can be quickly matched from the video library by utilizing the pre-stored association relation of the matched index subfield, thereby accelerating the video searching speed and effectively improving the searching experience of a user.
The embodiment of the invention discloses a video searching device, please refer to fig. 5, which includes:
the analysis module 501 is configured to, when a video search request of a user is received, analyze the video search request to obtain an analysis result;
a field matching module 502, configured to match a field corresponding to the analysis result from an index field according to the analysis result, where the index field is generated after information integration is performed on all included tags;
the video matching module 503 is configured to match a video corresponding to the field from a video library according to the field corresponding to the parsing result, and use the video as a video search result.
According to the video search device disclosed by the embodiment of the invention, the field matching module 502 is used for matching the analysis result obtained by the analysis module 501 with the index field generated by integrating the information of all the labels, so that the matching range of the analysis result can be expanded on the basis of improving the utilization rate of all the labels, the number of video search results presented to a user by the video matching module 503 is increased, and therefore, the matching of the analysis result is completed by using the index field generated by integrating the information of all the contained labels, the problem that all the labels corresponding to the analysis result cannot be hit at the same time due to the fact that a plurality of labels are independent from each other and respectively have the index field corresponding to the labels can be effectively solved, and the recall rate of the video is improved on the basis of fully utilizing all the labels.
Please refer to a method flowchart corresponding to fig. 1 for the working process of each module provided in the embodiment of the present invention, and detailed description of the working process is omitted.
On the basis of the embodiment shown in fig. 5, an embodiment of the present invention discloses an apparatus for generating an index field, please refer to fig. 6, which includes:
an obtaining module 601, configured to obtain multiple tags, where each tag includes a tag type and tag related information;
a classifying module 602, configured to classify all the tags into corresponding entity types according to the tag type of each tag, where the number of the entity types is at least one;
a deduplication module 603, configured to perform deduplication on tags in each entity type that have the same tag related information, and use the tag related information obtained after deduplication as an index subfield;
a generating module 604, configured to generate, according to the video information of each video in the video library and the entity type, an association relationship between each video in the video library and the corresponding index subfield, and store the association relationship in the corresponding index subfield;
and a permutation and combination module 605, configured to perform permutation and combination on all index sub-fields in which the association relationship is stored, so as to generate the index field.
The first generating module 604 specifically includes:
a first determining unit 6041, configured to, when a target entity type is a first entity type, compare a tag field in video information of each video in the video library with each index subfield included in the target entity type, and determine whether the tag field is the same as the index subfield;
a first generating unit 6042, configured to generate an association relationship between the video and the corresponding index subfield and store the association relationship in the corresponding index subfield after the first determining unit 6041 determines that the tag field is the same as the index subfield.
In the embodiment of the invention, the classification module 602 classifies the multiple tags acquired by the acquisition module 601 into the respective corresponding entity types according to the tag types, so that the incidence relation between the video and the corresponding index subfields is conveniently established in the same establishment mode for the same entity type subsequently, the generation speed of the index field is accelerated, the duplication elimination module 603 eliminates duplication on all tags in each entity type to obtain different tag-related information as the index subfields, the classification and fusion of the multiple tags can be realized, the comprehensiveness of the index field generated by combination is ensured, and the video recall rate is indirectly improved. When the target entity type is the first entity type and the first determining unit 6041 determines that the tag field is the same as the index subfield, the first generating unit 6042 establishes and stores an association relationship between the video corresponding to the tag field and the index subfield, so that the video search rate can be effectively increased, and the video search efficiency can be effectively improved.
Please refer to a method flowchart corresponding to fig. 2 for the working process of each module provided in the embodiment of the present invention, and detailed description of the working process is omitted.
On the basis of the embodiment shown in fig. 5, another apparatus for generating an index field is disclosed in the embodiment of the present invention, please refer to fig. 7, which includes:
an acquisition module 601, a classification module 602, a de-duplication module 603, a generation module 604 and an arrangement and combination module 605;
the generating module 604 specifically includes:
a word segmentation unit 6043, configured to, when the target entity type is the second entity type, perform word segmentation on a specific field in video information of each video in the video library, and generate at least one specific subfield;
a second determining unit 6044, configured to compare all the specific subfields with each of the index subfields included in the target entity type, and determine whether the specific subfields are the same as the index subfields;
a second generating unit 6045, configured to generate an association relationship between the video and the corresponding index subfield and store the association relationship in the corresponding index subfield after the second determining unit 5044 determines that the specific subfield is the same as the index subfield.
In the embodiment of the invention, the classification module 602 classifies the multiple tags acquired by the acquisition module 601 into the respective corresponding entity types according to the tag types, so that the incidence relation between the video and the corresponding index subfields is conveniently established in the same establishment mode for the same entity type subsequently, the generation speed of the index field is accelerated, the duplication elimination module 603 eliminates duplication on all tags in each entity type to obtain different tag-related information as the index subfields, the classification and fusion of the multiple tags can be realized, the comprehensiveness of the index field generated by combination is ensured, and the video recall rate is indirectly improved. When the target entity type is the second entity type and the second determining unit 6044 determines that the specific subfield after the word segmentation is the same as the index subfield, the second generating unit 6045 establishes and stores an association relationship between the video corresponding to the specific subfield and the index subfield, so that the video search rate can be effectively increased, and the video search efficiency can be effectively improved. Please refer to a method flowchart corresponding to fig. 3 for the working process of each module provided in the embodiment of the present invention, and detailed description of the working process is omitted.
On the basis of the embodiment shown in fig. 4, another video search apparatus is disclosed in the embodiment of the present invention, referring to fig. 8, which includes:
the system comprises an analysis module 501, a field matching module 502 and a video matching module 503;
wherein the field matching module 502 comprises: and the field matching submodule 5021 is used for matching an index subfield corresponding to the analysis result from the index field according to the analysis result.
The video matching module 503 includes: the video matching sub-module 5031 is configured to match, by the field matching sub-module 5021, a video having an association relationship with an index sub-field from the video library according to the index sub-field corresponding to the parsing result after the index sub-field corresponding to the parsing result is matched from the index field according to the parsing result, and the video is used as a video search result presented to a user.
In the video search device disclosed in the embodiment of the present invention, the field matching submodule 5021 matches the obtained parsing result with the index field generated by integrating the information of all tags, so as to determine the index subfield corresponding to the parsing result, so that the video matching submodule 5031 can quickly match all videos corresponding to the index subfield from the video library by using the pre-stored association relationship of the matched index subfield, thereby accelerating the video search rate and effectively improving the search experience of the user.
Please refer to a method flowchart corresponding to fig. 4 for the working process of each module provided in the embodiment of the present invention, and detailed description of the working process is omitted.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
Claims (11)
1. A video search method, comprising:
when a video search request of a user is received, analyzing the video search request to obtain an analysis result;
matching fields corresponding to the analysis result from index fields according to the analysis result, wherein the index fields are generated after information integration is carried out on all the included labels;
matching the video corresponding to the field from a video library according to the field corresponding to the analysis result to serve as a video search result;
wherein the process of generating the index field comprises:
obtaining a plurality of labels, wherein each label comprises a label type and label related information;
classifying all the tags into corresponding entity types by using the tag type of each tag, wherein the number of the entity types is at least one;
removing the duplicate of the label with the same label related information in each entity type, and taking the label related information obtained after the duplicate removal as an index subfield;
generating an association relation between each video in the video library and the corresponding index subfield according to the video information of each video in the video library and the entity type, and storing the association relation in the corresponding index subfield;
and arranging and combining all index sub-fields storing the association relation to generate the index field.
2. The method according to claim 1, wherein the generating and storing the association relationship between each video in the video library and the corresponding index subfield according to the video information of each video in the video library and the entity type comprises:
when the target entity type is the first entity type, comparing a tag field in video information of each video in the video library with each index subfield contained in the target entity type, and judging whether the tag field is the same as the index subfield;
and if the label field is the same as the index subfield, generating an association relation between the video and the corresponding index subfield, and storing the association relation in the corresponding index subfield.
3. The method according to claim 2, wherein the generating and storing the association relationship between each video in the video library and the corresponding index subfield according to the video information of each video in the video library and the entity type comprises:
when the target entity type is a second entity type, performing word segmentation on a specific field in video information of each video in the video library to generate at least one specific subfield;
comparing all the specific subfields with each index subfield contained in the target entity type, and judging whether the specific subfields are the same as the index subfields;
and if the specific sub-field is the same as the index sub-field, generating an association relation between the video and the corresponding index sub-field, and storing the association relation in the corresponding index sub-field.
4. The method according to claim 1, wherein matching a field corresponding to the parsing result from an index field according to the parsing result comprises:
and matching an index sub-field corresponding to the analysis result from an index field according to the analysis result.
5. The method according to claim 4, wherein said matching the video corresponding to the field from the video library according to the field corresponding to the parsing result as the video search result comprises:
and matching the video with the index sub-field in association relationship from the video library according to the index sub-field corresponding to the analysis result, and taking the video as the video search result.
6. The method of claim 1, after said obtaining a plurality of tags, further comprising:
receiving a tag processing instruction, and performing processing operation on a tag corresponding to the tag processing instruction, wherein the processing operation comprises any one or combination of more than one of addition, deletion, modification and query.
7. A video search apparatus, comprising:
the analysis module is used for analyzing the video search request to obtain an analysis result when the video search request of a user is received;
the field matching module is used for matching a field corresponding to the analysis result from an index field according to the analysis result, wherein the index field is generated after information integration is carried out on all the included labels;
the video matching module is used for matching the video corresponding to the field from the video library according to the field corresponding to the analysis result and taking the video as a video search result;
the device further comprises:
the system comprises an acquisition module, a processing module and a display module, wherein the acquisition module is used for acquiring a plurality of labels, and each label comprises a label type and label related information;
the classification module is used for classifying all the labels into corresponding entity types by using the label type of each label, and the number of the entity types is at least one;
a deduplication module, configured to perform deduplication on tags with the same tag related information in each entity type, and use the tag related information obtained after deduplication as an index subfield;
the generating module is used for generating an association relation between each video in the video library and the corresponding index subfield according to the video information of each video in the video library and the entity type, and storing the association relation in the corresponding index subfield;
and the permutation and combination module is used for carrying out permutation and combination on all the index sub-fields storing the association relation to generate the index fields.
8. The apparatus of claim 7, wherein the generating module comprises:
a first judging unit, configured to, when a target entity type is a first entity type, compare a tag field in video information of each video in the video library with each index subfield included in the target entity type, and judge whether the tag field is the same as the index subfield;
and the first generation unit is used for generating the association relation between the video and the corresponding index subfield and storing the association relation in the corresponding index subfield after the first judgment unit judges that the tag field is the same as the index subfield.
9. The apparatus of claim 8, wherein the generating module comprises:
the word segmentation unit is used for segmenting a specific field in the video information of each video in the video library to generate at least one specific subfield when the target entity type is the second entity type;
a second judging unit, configured to compare all the specific subfields with each of the index subfields included in the target entity type, and judge whether the specific subfields are the same as the index subfields;
and a second generating unit, configured to generate an association relationship between the video and the corresponding index subfield and store the association relationship in the corresponding index subfield after the second determining unit determines that the specific subfield is the same as the index subfield.
10. The apparatus of claim 7, wherein the field matching module comprises:
and the field matching submodule is used for matching an index subfield corresponding to the analysis result from the index field according to the analysis result.
11. The apparatus of claim 10, wherein the video matching module comprises:
and the video matching sub-module is used for matching a video which has an association relation with the index sub-field from the video library according to the index sub-field corresponding to the analysis result after the index sub-field corresponding to the analysis result is matched from the index field by the field matching sub-module according to the analysis result, and taking the video as the video search result.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710964471.8A CN107729486B (en) | 2017-10-17 | 2017-10-17 | Video searching method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710964471.8A CN107729486B (en) | 2017-10-17 | 2017-10-17 | Video searching method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107729486A CN107729486A (en) | 2018-02-23 |
CN107729486B true CN107729486B (en) | 2021-02-09 |
Family
ID=61211480
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710964471.8A Active CN107729486B (en) | 2017-10-17 | 2017-10-17 | Video searching method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107729486B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108763363B (en) * | 2018-05-17 | 2022-02-18 | 创新先进技术有限公司 | Method and device for checking record to be written |
CN109635157B (en) * | 2018-10-30 | 2021-05-25 | 北京奇艺世纪科技有限公司 | Model generation method, video search method, device, terminal and storage medium |
CN109977318B (en) * | 2019-04-04 | 2021-06-29 | 掌阅科技股份有限公司 | Book searching method, electronic device and computer storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4539552B2 (en) * | 2005-12-21 | 2010-09-08 | 日本ビクター株式会社 | Content search apparatus and content search program |
CN103310014A (en) * | 2013-07-02 | 2013-09-18 | 北京航空航天大学 | Method for improving accuracy of search result |
CN104123366A (en) * | 2014-07-23 | 2014-10-29 | 谢建平 | Search method and server |
CN104219575A (en) * | 2013-05-29 | 2014-12-17 | 酷盛(天津)科技有限公司 | Related video recommending method and system |
CN105187795A (en) * | 2015-09-14 | 2015-12-23 | 博康云信科技有限公司 | Video label positioning method and device based on view library |
-
2017
- 2017-10-17 CN CN201710964471.8A patent/CN107729486B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4539552B2 (en) * | 2005-12-21 | 2010-09-08 | 日本ビクター株式会社 | Content search apparatus and content search program |
CN104219575A (en) * | 2013-05-29 | 2014-12-17 | 酷盛(天津)科技有限公司 | Related video recommending method and system |
CN103310014A (en) * | 2013-07-02 | 2013-09-18 | 北京航空航天大学 | Method for improving accuracy of search result |
CN104123366A (en) * | 2014-07-23 | 2014-10-29 | 谢建平 | Search method and server |
CN105187795A (en) * | 2015-09-14 | 2015-12-23 | 博康云信科技有限公司 | Video label positioning method and device based on view library |
Non-Patent Citations (1)
Title |
---|
标签主题图的构建与实现研究;熊回香 等;《图书情报工作》;20140405(第7期);107-112 * |
Also Published As
Publication number | Publication date |
---|---|
CN107729486A (en) | 2018-02-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US12026194B1 (en) | Query modification based on non-textual resource context | |
US10198506B2 (en) | System and method of sentiment data generation | |
US8577882B2 (en) | Method and system for searching multilingual documents | |
CN110929125B (en) | Search recall method, device, equipment and storage medium thereof | |
CN112287914B (en) | PPT video segment extraction method, device, equipment and medium | |
CN111506771B (en) | Video retrieval method, device, equipment and storage medium | |
US10311113B2 (en) | System and method of sentiment data use | |
US20160063062A1 (en) | Code searching and ranking | |
EP2724256A1 (en) | System and method for matching comment data to text data | |
CN109492081B (en) | Text information searching and information interaction method, device, equipment and storage medium | |
CN112199526B (en) | Method and device for issuing multimedia content, electronic equipment and storage medium | |
CN111294660A (en) | Video clip positioning method, server, client and electronic equipment | |
CN107729486B (en) | Video searching method and device | |
US9135335B2 (en) | Filmstrip-based query suggestions | |
CN111400464A (en) | Text generation method, text generation device, server and storage medium | |
TWI709905B (en) | Data analysis method and data analysis system thereof | |
US10216989B1 (en) | Providing additional information for text in an image | |
US9720896B1 (en) | Synthesizing union tables from the web | |
CN110287338B (en) | Industry hotspot determination method, device, equipment and medium | |
CN109558468B (en) | Resource processing method, device, equipment and storage medium | |
CN113901263B (en) | Label generation method and device for video material | |
CN105808623A (en) | Search-based page access event association method and device | |
CN107577667B (en) | Entity word processing method and device | |
CN108255866B (en) | Method and device for checking links in website | |
CN111050194B (en) | Video sequence processing method, video sequence processing device, electronic equipment and computer readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |