[go: up one dir, main page]

WO2022003836A1 - Processing system and processing method - Google Patents

Processing system and processing method Download PDF

Info

Publication number
WO2022003836A1
WO2022003836A1 PCT/JP2020/025705 JP2020025705W WO2022003836A1 WO 2022003836 A1 WO2022003836 A1 WO 2022003836A1 JP 2020025705 W JP2020025705 W JP 2020025705W WO 2022003836 A1 WO2022003836 A1 WO 2022003836A1
Authority
WO
WIPO (PCT)
Prior art keywords
condition
scene
conditions
relationship
time zone
Prior art date
Application number
PCT/JP2020/025705
Other languages
French (fr)
Japanese (ja)
Inventor
遥 久保田
明 片岡
Original Assignee
日本電信電話株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電信電話株式会社 filed Critical 日本電信電話株式会社
Priority to PCT/JP2020/025705 priority Critical patent/WO2022003836A1/en
Priority to JP2022532896A priority patent/JP7439927B2/en
Publication of WO2022003836A1 publication Critical patent/WO2022003836A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/73Querying
    • G06F16/732Query formulation

Definitions

  • the present invention relates to a processing system and a processing method.
  • the video information can accurately reproduce the situation at the time of shooting. For this reason, video information is used in many fields regardless of personal use or business use. For example, video information is used for leisure and sports as an individual, and for security, business grasp, and trail as a business.
  • a search method using an index has been proposed in the past.
  • an index is added to each frame at the start / end timing or in between, and the corresponding scene is presented by selecting one of the indexes.
  • a method has been proposed in which an index is given by using the image information and the character information in the video information in a complex manner, and a scene search using a threshold value or the like based on the index is possible (for example, a non-patent document). 1).
  • a visualization method using a timeline has been proposed. This method is necessary from the timeline because it is possible to analyze by comparing the time width of each scene from the timeline by presenting the range delimited by the index etc. in chronological order on the timeline. It is also possible to determine a wide range.
  • the scene detection itself may become difficult.
  • the degree of matching of the conditions differs due to individual differences, individual differences, etc., it is necessary to relax the conditions in order to set the conditions that cover the whole, the total detection time becomes long, and the narrowing down becomes weak.
  • the present invention has been made in view of the above, and provides a processing system and a processing method capable of appropriately complementing a scene that could not be detected when detecting a scene in the video information.
  • the purpose is.
  • the processing system has a plurality of scenes in the video information based on parameters including time-series information associated with the video information.
  • the first storage unit that stores the condition and the relationship between each directional condition, and any condition, based on the relationship between each condition, from a plurality of conditions to any condition. It has a detection unit that obtains related conditions and detects a scene that corresponds to the obtained conditions.
  • FIG. 1 is a diagram showing an example of a scene detection screen in the video information according to the first embodiment.
  • FIG. 2 is a diagram showing an example of the functional configuration of the display processing system according to the first embodiment.
  • FIG. 3 is a diagram showing an example of a relationship graph.
  • FIG. 4 is a diagram illustrating a display example by the visualization information display unit shown in FIG. 2.
  • FIG. 5 is a flowchart showing a processing procedure of the conditions executed by the display processing system shown in FIG. 2 and the setting processing of the relationship between the conditions.
  • FIG. 6 is a flowchart showing a processing procedure of display processing of visualization information for a specified condition executed by the display processing system shown in FIG. FIG.
  • FIG. 7 is a flowchart showing a processing procedure of scene detection processing in video information executed by the display processing system shown in FIG. 2.
  • FIG. 8 is a diagram illustrating scene detection in video information in the conventional method.
  • FIG. 9 is a diagram illustrating scene detection in video information according to the first embodiment.
  • FIG. 10 is a diagram showing an example of a scene detection screen in the video information according to the second embodiment.
  • FIG. 11 is a diagram showing an example of the functional configuration of the display processing system according to the second embodiment.
  • FIG. 12 is a flowchart showing a processing procedure of the conditions executed by the display processing system shown in FIG. 11 and the setting processing of the relationship between the conditions.
  • FIG. 13 is a flowchart showing a condition executed by the display processing system shown in FIG. 11 and another processing procedure for setting the relationship between the conditions.
  • FIG. 14 is a diagram showing an example of a computer in which a display processing system is realized by executing a program.
  • FIG. 1 is a diagram showing an example of a scene detection screen in the video information according to the first embodiment.
  • the upper figure of FIG. 1 also shows an example of a scene detection screen by a conventional method.
  • a detection condition for determining whether or not the scene is applicable is set by combining parameters associated with the video. Set.
  • the search user performs a scene search using the label given to the creator of the detection condition.
  • a search user specifies a search condition (hereinafter referred to as a condition) for detecting a scene of video information
  • a condition a search condition for detecting a scene of video information
  • the time zone of the scene corresponding to the specified condition is colored as a detection result.
  • the timeline is displayed (see (1) and (2) in FIG. 1).
  • the detection conditional expression for example, AND of conditions A to C (see (3) in FIG. 1) and the like
  • the search condition expressions AND of conditions A to C
  • the search user needs to repeatedly set the condition specification and try the search a plurality of times.
  • a timeline will be described as an example of visualization information indicating a time zone of a scene corresponding to an arbitrary condition, but the visualization information is not limited to the timeline and is a time zone of a scene corresponding to the condition. It may be character information indicating.
  • the timeline L1 (first visualization information) (AND of the conditions A to C) showing the time zone of the scene corresponding to the condition specified by the user. )
  • the configuration conditions A, B, and C related to the designated conditions are also visualized by displaying the timelines (second visualization information) LA to LC (see (5) in FIG. 1).
  • the timeline L1 and the timelines LA to LC with the AND symbol M1 (third visualization information)
  • the relationship between the designated condition and the conditions A to D is visualized ((FIG. 1). 6)).
  • the timelines LC-1 and LC-2 are displayed, and the timeline LC and the timeline LC-1 and LC-2 are designated by the OR symbol M2.
  • the condition C is an OR of the condition C-1 and the condition C-2 (see (7) in FIG. 1).
  • the timeline L1 indicating the time zone of the scene satisfying the designated condition and the timeline showing the time zone of the scene corresponding to the constituent conditions A to C constituting the designated condition are also hierarchically displayed. ..
  • the search user can easily grasp the scene detection result of each related condition together with the scene detection result of the specified condition. Then, the user can refer to the scene detection results of more conditions with priorities, and by not limiting the use to grasp the relationship between each condition, it is general purpose. It will also be possible to utilize it for specific searches.
  • the search user when the search user cannot sufficiently detect the scene under the specified condition, it can be expected that the search user intuitively performs the re-search using the related condition by referring to the timeline of the specified condition and the related condition. .. Therefore, in the first embodiment, it is considered that the repeated trials of parameter adjustment or condition specification by the search user, which are conventionally required, can be reduced.
  • the search user can refer to the visualized timeline L1 and the timelines LA to LC-2 and their relationships with each other to specify conditions (for example, conditions A to C) that are more specific than the specified conditions. More abstract conditions (eg, conditions C-1, C-2) can be easily recognized. Then, the search user selects or combines these conditions A to C-2 to specify the conditions, so that the difficulty of setting the conditions by the search user is alleviated. Therefore, according to the first embodiment, when the search user detects the scene from the video information, it is possible to support the scene detection by the search user and appropriately supplement the scene that could not be detected.
  • conditions for example, conditions A to C
  • More abstract conditions eg, conditions C-1, C-2
  • the search user selects or combines these conditions A to C-2 to specify the conditions, so that the difficulty of setting the conditions by the search user is alleviated. Therefore, according to the first embodiment, when the search user detects the scene from the video information, it is possible to support the scene detection by the search user and appropriately supplement the scene that could
  • FIG. 2 is a diagram showing an example of the functional configuration of the display processing system according to the first embodiment.
  • the display processing system 10 sets a relationship between each condition together with a plurality of conditions for detecting a scene in the video information based on a parameter including time-series information associated with the video information. Hold.
  • the display processing system 10 visualizes the detection result of the scene under the arbitrary condition and the detection result of the scene corresponding to the condition related to the arbitrary condition in a state with the relationship between each condition. , It assists the user in detecting the scene, and makes it possible to appropriately supplement the scene that could not be detected.
  • the display processing system 10 is shown assuming that it functions as a terminal device, but the present invention is not limited to this, and the display processing system 10 may function as a server, and the searched video may be used.
  • the scene may be output to the user terminal.
  • the display processing system 10 includes a raw data storage unit 11 (first input unit), a data processing unit 12, a UI (User Interface) unit 13, a condition storage unit 14 (first storage unit), and a processing data storage unit 15 ( It has a first storage unit, a detection unit, a second storage unit, and a third storage unit). Each part will be described below. It should be noted that each of the above-mentioned parts may be held by a plurality of devices in a dispersed manner.
  • a predetermined program is read into a computer or the like including a ROM (Read Only Memory), a RAM (Random Access Memory), a CPU (Central Processing Unit), etc., and the CPU executes the predetermined program. It will be realized by.
  • the display processing system 10 has a communication interface for transmitting and receiving various information to and from other devices connected via a network or the like.
  • the display processing system 10 has a NIC (Network Interface Card) or the like, and communicates with other devices via a telecommunication line such as a LAN (Local Area Network) or the Internet.
  • the display processing system 10 has a touch panel, a voice input device, an input device such as a keyboard and a mouse, a display device such as a liquid crystal display, and a printing device such as a printer, and inputs and outputs information.
  • the raw data storage unit 11 receives and stores the video information to be searched and the input of raw data used in combination with the video information.
  • the raw data is sensor information obtained synchronously with the shooting of video information.
  • the sensor information is, for example, GPS (Global Positioning System) information, acceleration information, and temperature information.
  • the data processing unit 12 detects scenes in the video information corresponding to each condition based on the video information and raw data and each condition, and outputs the scene detection result according to each condition to the processing data storage unit 15. Further, the data processing unit 12 acquires raw data from the video information by processing the video information, and stores the video information and the raw data in the raw data storage unit 11 in association with each other. For example, the data processing unit 12 acquires raw data based on an object recognition result, position information by SLAM (Simultaneous Localization and Mapping), and the like.
  • SLAM Simultaneous Localization and Mapping
  • the UI unit 13 has a condition setting unit 131, a visualization information display unit 132 (second input unit, first display unit), and a video display unit 133.
  • the condition setting unit 131 receives an instruction to create a search condition (condition) for scene detection by operating the input device by the search user. Upon receiving an instruction to create a condition, the condition setting unit 131 creates a new condition using an arbitrary method, and stores the new condition and the relationship between the new condition and other existing conditions. Store in 14.
  • the visualization information display unit 132 receives the input of the search condition (designated condition) (first condition) designated for scene detection of video information.
  • the visualization information display unit 132 outputs the designated condition to the processing data storage unit 15. Then, the visualization information display unit 132 outputs visualization information related to scene detection based on the information including the scene detection result output from the processing data storage unit 15.
  • the visualization information display unit 132 receives the designation of the video reproduction scene by the operation of the input device by the search user who has referred to the visualization information
  • the visualization information display unit 132 outputs the reproduction range to the video display unit 133.
  • the video display unit 133 When the video display unit 133 receives the designation of the playback range from the visualization information display unit 132, the video display unit 133 reproduces the video in the designated range based on the video information stored in the raw data storage unit 11.
  • the condition storage unit 14 stores a plurality of conditions for detecting a scene in the video information and a relationship graph showing the relationship between each condition having directivity. When the relationship graph is registered or updated, the condition storage unit 14 outputs the registered or updated relationship graph to the processing data storage unit 15.
  • FIG. 3 is a diagram showing an example of a relationship graph.
  • the relation graph G1 comprehensively holds the relation of each condition by showing the logical product, the logical sum, or the time-series connection and the inclusion relation for each condition (FIG. 3). 3 (1)).
  • the upper condition is the result of a logical sum or AND between certain conditions, or if the upper condition includes a certain condition, the upper condition may occur due to multiple conditions occurring in chronological order.
  • the related condition is linked to the subordinate of the higher condition.
  • each condition may be linked, including conditions such as directionality (“inspection” ⁇ “recording”) and occurrence time (for example, within 10 seconds).
  • the condition of "inspection” includes the conditions of "stop” and "gaze”, and these two are associated with "inspection” in relation to the logical product.
  • abstract conditions are related under more specific conditions. As shown in the relationship graph G1, by making the relationship of conditions into multiple layers, it is possible to simplify the prioritization of the requirements to be referred to by the user. Then, by making the relation of the conditions into multiple layers as in the relation graph G1, it is also possible to discriminate indirectly related conditions such as "inspection” and "gaze 2 seconds".
  • the processing data storage unit 15 holds a relation graph, obtains a condition related to an arbitrary condition from a plurality of conditions based on the relation graph with respect to an arbitrary condition, and sets a scene corresponding to the obtained condition. To detect.
  • the processing data storage unit 15 holds the scene detection result according to each condition together with the relationship graph.
  • the scene detection result according to each condition is the detection result of the scene detected from the video information using each condition stored in the condition storage unit 14, and is obtained by the processing by the data processing unit 12.
  • the processing data storage unit 15 When the designated condition is input from the visualization information display unit 152, the processing data storage unit 15 has the designated condition and the related condition related to the designated condition (second condition) from among a plurality of conditions based on the relation graph. ) And ask. Then, the processing data storage unit 15 outputs, among the scene detection results according to each condition, the scene detection result corresponding to the designated condition and the scene detection result corresponding to the related condition to the visualization information display unit 132. At the same time, the processing data storage unit 15 outputs information indicating the relationship between the designated condition and the related condition to the visualization information display unit 132 based on the relationship graph.
  • the visualization information display unit 132 receives from the processing data storage unit 15 the scene detection result corresponding to the designated condition, the detection result of the scene corresponding to the related condition, and the information indicating the relationship between the designated condition and the related condition. , The first visualization information indicating the time zone of the scene corresponding to the designated condition, the second visualization information indicating the time zone of the scene corresponding to the related condition, and the third indicating the relationship between the designated condition and the related condition. Visualization information and output.
  • FIG. 4 is a diagram illustrating a display example by the visualization information display unit 132 shown in FIG.
  • the visualization information display unit 132 has a timeline showing the time zone of the scene corresponding to the user-designated condition “inspection”, and the related condition “stop” according to the relationship shown in the relationship graph G1.
  • “Gaze 5 seconds”, “Move”, “Overlook” is displayed as the third visualization information with a frame or symbol indicating that the timeline indicating the time zone of the corresponding scene is under the specified condition "Inspection”.
  • the visualization information display unit 132 displays a timeline relating to the condition "gaze 2 seconds” included in the related condition "gaze 5 seconds".
  • the visualization information display unit 132 hierarchically displays the timeline of the designated condition and its related information, so that the search user can refer to the related information regarding the designated condition "inspection” and display more related information. It can be intuitively used as a search condition (see (1) in FIG. 4).
  • the visualization information display unit 132 can use any visualization format, not limited to the timeline format. Further, the visualization information display unit 132 may change the display method such as the timeline or group the conditions, or may accept the change of the display method such as the timeline or the grouping between the conditions from the search user. good.
  • FIG. 5 is a flowchart showing a processing procedure of the conditions executed by the display processing system 10 shown in FIG. 2 and the setting processing of the relationship between the conditions.
  • the condition setting unit 131 receives an instruction to register a condition name or ID to be extracted for scene detection by operating an input device by a search user (step S1).
  • the condition setting unit 131 selects one or more parameters expressing the corresponding condition from the available parameters by using an arbitrary method, and sets numerical conditions such as a threshold value and a reference value (step S2).
  • the condition in which the name is specified in step S1 and the threshold value or the like is set in step S2 corresponds to the "new condition" in step S3 (described later).
  • the condition setting unit 131 extracts the condition related to the new condition from the condition storage unit 14 (step S3).
  • the new condition is a condition that is a logical sum or a logical product of the existing conditions in the condition storage unit 14, a condition that includes the existing condition, and a condition that is expressed by combining the existing conditions in chronological order.
  • the condition related to the new condition is a condition specified by the user as a related condition from the existing condition, a condition automatically detected by the display processing system 10 as being related by comparing the numerical conditions, or a new condition. It is a condition that is divided and a partial condition is created and linked.
  • the new condition does not necessarily have to be expressed by a combination of existing conditions. For example, for a new condition generated by a user, a new condition may be newly generated by searching for a related condition from an existing condition manually or by comparing numerical conditions, or by dividing a new condition.
  • the condition storage unit 14 registers the condition added by the condition setting unit 131, and updates the relationship graph between the condition and the existing condition (step S4) to set a more specific condition in the upper layer. do.
  • FIG. 6 is a flowchart showing a processing procedure of display processing of visualization information for a designated condition executed by the display processing system 10 shown in FIG.
  • the visualization information display unit 132 when the visualization information display unit 132 receives an input of a search condition (designated condition) designated for scene detection from a search user (step S11), the visualization information display unit 132 processes the specified condition.
  • Data storage unit 15 Output to.
  • the processing data storage unit 15 obtains a specified condition and a related condition related to the specified condition from a plurality of conditions, and detects and visualizes a scene detection result corresponding to the specified condition and a scene detection result corresponding to the related condition. In addition to outputting to the information display unit 132, information indicating the relationship between the designated condition and the related condition is output to the visualization information display unit 132 based on the relationship graph (step S12).
  • the visualization information display unit 132 displays and outputs each timeline of the designated condition and the related information together with the relationship between the conditions (step S13).
  • the visualization information display unit 132 receives the designation of the video reproduction scene from the input device by the search user who referred to the visualization information (step S14)
  • the visualization information display unit 132 outputs the reproduction range to the video display unit 133.
  • the video display unit 133 receives the designation of the reproduction range from the visualization information display unit 132
  • the video display unit 133 reproduces the video in the designated range based on the video information stored in the raw data storage unit 11 (step S15).
  • FIG. 7 is a flowchart showing a processing procedure of scene detection processing in video information executed by the display processing system 10 shown in FIG.
  • the display processing system 10 determines whether or not the data in the raw data storage unit 11 or the condition in the condition storage unit has been updated (step S21).
  • the display processing system 10 repeats the determination process in step S21 until the data in the raw data storage unit 11 or the condition in the condition storage unit is updated.
  • the display processing system 10 determines whether or not to extract information from the video information in the own system. Is determined (step S22).
  • the data processing unit 12 executes information extraction such as GPS information, acceleration information, and temperature information from the video information (step S23).
  • the data processing unit 12 refers to the detection condition of each condition stored in the condition storage unit 14 (step). S24). In response to this, the condition storage unit 14 outputs the stored detection condition to the data processing unit 12 (step S25).
  • the data processing unit 12 detects the corresponding scene in the video information based on each condition (step S26), and outputs the scene detection result to the processing data storage unit 15.
  • the processing data storage unit 15 saves the scene detection result of each condition (step S27), and ends the processing.
  • FIG. 8 is a diagram illustrating scene detection in video information in the conventional method.
  • the conventional method since the user can search the scene only in the label unit prepared by the creator of the detection condition (see (1) in FIG. 8), the time zone of the scene corresponding to the specified condition is colored. Only the timeline is displayed as the detection result.
  • FIG. 9 is a diagram illustrating scene detection in video information according to the first embodiment.
  • the display processing method according to the first embodiment holds a plurality of conditions for detecting a scene in video information and a relationship graph showing a relationship between each condition having directivity. Then, regarding the specified condition, based on the relationship between each condition, the related condition related to the specified condition is obtained from a plurality of conditions, and the scene corresponding to the obtained condition is detected. Then, in the first embodiment, as shown in FIG. 9, in addition to the timeline showing the time zone of the scene corresponding to the designated condition, the timeline showing the time zone of the scene corresponding to the related condition is used as the related condition. In addition to displaying each condition, the relationship between each condition is also displayed by linking the related conditions under the specified condition and displaying the timeline in a hierarchical manner.
  • the search user wants to detect a specific scene from the video information, he / she can refer to the detection status (0/1 or continuous value) of the specified condition and the related condition that are visualized hierarchically, and the related condition is used. It is possible to refer to the scene detection result of the specified condition while comparing it with the scene detection result. Then, when the detection result is insufficient, the search user can narrow down the scene by referring to the scene detection result of the related condition. For example, when the search user wants to preferentially acquire the scenes corresponding to the specified conditions, the search user may narrow down the scenes in the time zone within the frames W11, W12, and W13 in the timeline corresponding to the specified conditions (Fig.). 9 (1)).
  • the search user may find that the conditions A and C are indispensable from the scene detection results of the conditions A, B, and C in the frame W14, but are effective for the usage of the scene if the condition B is not satisfied. Since it can be discriminated, it is possible to set the condition only for the combination of the conditions A and C. Then, the search user can change the condition such that the condition C may be relaxed a little by referring to the inclusion condition of the condition C. As described above, according to the first embodiment, it is possible to deal with the case where it is assumed that the user arbitrarily groups a plurality of conditions or temporarily hides a specific condition.
  • the search user can easily grasp the scene detection result of each related condition together with the detection result of the designated condition. Then, according to the first embodiment, when the search user cannot sufficiently detect the scene under the specified condition, the search user intuitively uses the related condition by referring to the specified condition and the detection result of the related condition. It is expected that the search will be performed again. Therefore, in the first embodiment, it is considered that the user's repeated trials of parameter adjustment or condition specification, which have been conventionally required, can be reduced. Therefore, according to the first embodiment, when the search user detects the scene from the video information, it is possible to appropriately supplement the scene that could not be detected.
  • FIG. 10 is a diagram showing an example of a scene detection screen in the video information according to the second embodiment.
  • the second embodiment when a range of scenes to be detected in the future is specified by the search user, it is possible to compare and refer to each condition that can detect the scenes in this range from the video information.
  • the scene is specified from a plurality of conditions.
  • a condition for example, conditions A to C
  • the timelines LA to LC-2 highlighting the designated time zone among the time zones to which each condition corresponds are hierarchically output for each requested condition. For example, as highlighting, the part of the timelines LA to LC-2 corresponding to the specified time zone is surrounded by the frame W21, displayed in a color or brightness different from the others, or blinks. Can be considered.
  • the search user specifies the time zone in which the scene is to be detected, and the scene corresponding to this time zone is detected under any combination of the conditions. It can be compared and referred to (see (2) in FIG. 10). Then, the search user can create detection conditions from the next time onward for the time zone in which the scene is to be detected by comparing and referring to the timelines LA to LC-2 highlighting the designated time zone. A specific scene can be detected more efficiently.
  • the search user gives priority to more specific conditions or more limited conditions. It is possible to create a detection condition that is effective for the search. Therefore, according to the second embodiment, it is possible to detect a scene according to the desire of the search user by supporting the creation of a detection condition capable of detecting the scene desired by the search user.
  • FIG. 11 is a diagram showing an example of the functional configuration of the display processing system according to the second embodiment.
  • the display processing system 210 has a UI unit 213 and a processing data storage unit 215 (setting unit) as compared with the display processing system 10 shown in FIG. Further, in the display processing system 210, the UI unit 213 has a condition setting unit 2131 (first setting unit), a visualization information display unit 2132 (second display unit), and a video display unit 2133 (third input unit). , A fourth input unit).
  • the video display unit 2133 reproduces video information from any position of the search user. Then, the video display unit 2133 accepts the input of the time zone designation information for designating an arbitrary time zone in the video information by the operation of the input device by the search user. The video display unit 2133 outputs the time zone designated by the time zone designation information to the visualization information display unit 2132.
  • the visualization information display unit 2132 outputs the time zone specified by the time zone designation information to the processing data storage unit 215.
  • the visualization information display unit 2132 outputs visualization information according to the time zone designation information.
  • the processing data storage unit 215 obtains a condition in which the detected scene is in the time zone specified in the time zone designation information from a plurality of conditions based on the scene detection result according to each condition, and the scene detection result according to the obtained condition. And, the relationship information indicating the relationship between the obtained conditions is output to the visualization information display unit 2132.
  • the visualization information display unit 2132 outputs visualization information highlighting the time zone specified in the time zone designation information among the time zones of the scenes corresponding to each condition for each condition obtained by the processing data storage unit 215. At the same time, visualization information showing the relationship between each condition is output.
  • the condition setting unit 2131 registers the condition for detecting the scene in the time zone specified in the time zone specification information and the relationship between each detection condition by the operation of the input device by the search user referring to the output visualization information. Accept instructions or update instructions.
  • the condition setting unit 2131 registers or updates the conditions stored in the condition storage unit 14 and the relationship between the conditions in response to the received registration instruction or update instruction.
  • FIG. 12 is a flowchart showing a processing procedure of the condition executed by the display processing system 210 shown in FIG. 11 and the setting processing of the relationship between the conditions.
  • the video display unit 2133 reproduces video information from an arbitrary position designated by the search user (step S31). Then, the video display unit 2133 determines whether or not the time zone specified in the video information has been specified by the operation of the input device by the search user (step S32). If the video display unit 2133 has not received the time zone designation in the video information (step S32: No), the video display unit 2133 returns to the determination process of step S32.
  • the search user can also specify a plurality of time zones in the video information at the same time.
  • the specified time zone is output to the processing data storage unit 215 via the visualization information display unit 2132.
  • the processing data storage unit 215 obtains a condition in which a detection scene exists in a specified time zone from a plurality of conditions based on the scene detection result according to each condition, and the scene detection result according to the obtained condition and these conditions. The relationship between the two is output to the visualization information display unit 2132 (step S33).
  • the visualization information display unit 2132 outputs visualization information highlighting the specified time zone among the time zones of the screen corresponding to each condition for the scene detection result by the processing data storage unit 215, and outputs the visualization information highlighting the relationship between the conditions. Visualization information indicating the sex is output (step S34).
  • the visualization information display unit 2132 may display the visualization information display unit 2132 so that it can be distinguished in each time zone, such as by separating the colors for each time zone.
  • the condition setting unit 2131 determines whether or not a registration instruction or an update instruction has been received for the condition for detecting the scene in the designated time zone and the relationship between the detection conditions (step S35).
  • the display processing system 210 has not received the registration instruction or the update instruction for the condition and the relationship between the detection conditions (step S35: No)
  • the display processing system 210 ends the processing.
  • a registration instruction or an update instruction for a condition and a relationship between each detection condition is received (step S35: Yes)
  • the condition setting unit 2131 receives a registration instruction or an update instruction, and the condition storage unit 14 stores the condition. And the relationship between each condition are registered or updated (step S36).
  • the detection scene is in the designated time zone from a plurality of conditions.
  • the conditions are obtained, and for each of the obtained conditions, the visualization information highlighting the specified time zone among the time zones to which each condition applies is displayed together with the relationship between the conditions.
  • the condition when the search user wants to determine what kind of condition is effective for detecting the scene in the desired time zone, the condition is such that the scene in this time zone can be detected.
  • the usefulness of a condition can be determined from a more specific condition or a more limited condition by referring to the relationship between the conditions. For example, the search user can extract the conditions commonly detected in the corresponding scene and create a new condition as a template.
  • the search user can determine a useful condition as a condition for detecting a scene in a desired time zone when detecting a scene from the video information, so that the desired scene can be determined. It will be possible to detect it properly.
  • FIG. 13 is a flowchart showing a condition executed by the display processing system 210 shown in FIG. 11 and another processing procedure for setting the relationship between the conditions.
  • Step S41 and step S42 shown in FIG. 13 are the same processes as steps S31 and S32 shown in FIG. 12, respectively.
  • the processing data storage unit 215 detects the time zone specified from among a plurality of conditions based on the scene detection results according to each condition. A scene is obtained for a certain condition, and the scene detection result according to the obtained condition and the relationship information indicating the relationship between these conditions are output to the condition setting unit 2131 (second setting unit) (step S43).
  • the condition setting unit 2131 detects a scene in a designated time zone based on the condition output from the condition setting unit 2131 and the relationship information indicating the relationship between the conditions obtained by the condition setting unit 2131. Is determined (step S44).
  • the condition setting unit 2131 extracts the conditions commonly detected in the designated time zone and creates a new condition. Further, the condition setting unit 2131 searches for the lower condition in order from the upper condition among the conditions having the detection scene in the designated time zone, and sets the condition by combining the more specific upper condition than the lower condition. , Create a new condition for scene detection in the specified time zone.
  • the condition setting unit 2131 registers or updates the condition stored in the condition storage unit 14 and the relationship between each condition based on the determination result in step S44 (step S45).
  • the condition setting unit 2131 registers the new condition in the condition storage unit 14, and also registers or updates the relationship between the new condition and the existing condition.
  • Each component of the display processing systems 10 and 210 is a functional concept and does not necessarily have to be physically configured as shown in the figure. That is, the specific form of the distribution and integration of the functions of the display processing systems 10 and 210 is not limited to the one shown in the figure, and all or part of them may be functional in any unit according to various loads and usage conditions. Or it can be physically distributed or integrated.
  • each process performed in the display processing systems 10 and 210 may be realized by a CPU, a GPU (Graphics Processing Unit), and a program analyzed and executed by the CPU and the GPU, in whole or in any part thereof. .. Further, each process performed in the display processing system 10 may be realized as hardware by wired logic.
  • FIG. 14 is a diagram showing an example of a computer in which the display processing systems 10 and 210 are realized by executing a program.
  • the computer 1000 has, for example, a memory 1010 and a CPU 1020.
  • the computer 1000 also has a hard disk drive interface 1030, a disk drive interface 1040, a serial port interface 1050, a video adapter 1060, and a network interface 1070. Each of these parts is connected by a bus 1080.
  • Memory 1010 includes ROM 1011 and RAM 1012.
  • the ROM 1011 stores, for example, a boot program such as a BIOS (Basic Input Output System).
  • BIOS Basic Input Output System
  • the hard disk drive interface 1030 is connected to the hard disk drive 1090.
  • the disk drive interface 1040 is connected to the disk drive 1100.
  • a removable storage medium such as a magnetic disk or an optical disk is inserted into the disk drive 1100.
  • the serial port interface 1050 is connected to, for example, a mouse 1110 and a keyboard 1120.
  • the video adapter 1060 is connected to, for example, the display 1130.
  • the hard disk drive 1090 stores, for example, an OS (Operating System) 1091, an application program 1092, a program module 1093, and program data 1094. That is, the program that defines each processing of the display processing systems 10 and 210 is implemented as a program module 1093 in which the code that can be executed by the computer 1000 is described.
  • the program module 1093 is stored in, for example, the hard disk drive 1090.
  • the program module 1093 for executing the same processing as the functional configuration in the display processing systems 10 and 210 is stored in the hard disk drive 1090.
  • the hard disk drive 1090 may be replaced by an SSD (Solid State Drive).
  • the setting data used in the processing of the above-described embodiment is stored as program data 1094 in, for example, a memory 1010 or a hard disk drive 1090. Then, the CPU 1020 reads the program module 1093 and the program data 1094 stored in the memory 1010 and the hard disk drive 1090 into the RAM 1012 and executes them as needed.
  • the program module 1093 and the program data 1094 are not limited to those stored in the hard disk drive 1090, but may be stored in, for example, a removable storage medium and read by the CPU 1020 via the disk drive 1100 or the like. Alternatively, the program module 1093 and the program data 1094 may be stored in another computer connected via a network (LAN (Local Area Network), WAN (Wide Area Network), etc.). Then, the program module 1093 and the program data 1094 may be read from another computer by the CPU 1020 via the network interface 1070.
  • LAN Local Area Network
  • WAN Wide Area Network

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

A display processing system (10) has: a first memory unit that memorizes a plurality of conditions for detecting a scene in video information on the basis of a parameter which includes time series information associated with the video information and that memorizes directional relationships between the conditions; and a processed data storage unit (15) that obtains, with regard to a given condition, a condition relevant to the given condition, from the plurality of conditions on the basis of the relationships between the conditions and that detects a scene which corresponds to the obtained condition.

Description

処理システム及び処理方法Processing system and processing method
 本発明は、処理システム及び処理方法に関する。 The present invention relates to a processing system and a processing method.
 映像情報は、撮影時の状況を正確に再現可能である。このため、映像情報は、個人利用や事業利用を問わず、多くの分野で活用されている。例えば、映像情報は、個人としては、レジャーやスポーツにおいて活用され、事業としては、セキュリティ、業務把握、証跡のために使用される。 The video information can accurately reproduce the situation at the time of shooting. For this reason, video information is used in many fields regardless of personal use or business use. For example, video information is used for leisure and sports as an individual, and for security, business grasp, and trail as a business.
 そして、映像情報の活用において、連続的な映像情報から特定の場面のみを検出したいケースも多い。例えば、特定人物が被写体である場面の閲覧をしているかを確認するケース、特定作業の手順を確認するケース、特定時間帯のみ視聴しているかを確認するケースなどである。 And, in utilizing video information, there are many cases where you want to detect only a specific scene from continuous video information. For example, there are cases where it is confirmed whether a specific person is browsing a scene that is a subject, cases where a procedure for a specific work is confirmed, and cases where it is confirmed whether or not a specific person is viewing only during a specific time zone.
 しかしながら、目視での特定の場面の探索は、時間的・画像的な情報量が多く、一覧性が悪いため、また、予定表などとの対比では不十分(検索粒度が荒く、予定変更への対応も困難)であるため、時間がかかり見落としも多く非効率である。このため、映像情報に対する効率的な場面提示方法が要望されている。 However, the visual search for a specific scene has a large amount of time and image information, and the listability is poor, and the comparison with a schedule etc. is insufficient (the search particle size is rough, and the schedule is changed. It is also difficult to deal with), so it takes time and is often overlooked, which is inefficient. Therefore, there is a demand for an efficient scene presentation method for video information.
 この要望に対し、従来、画面切り替えに基づく場面分割方法が提案されている。しかしながら、この方法の場合、カットが挟まれないワンショット映像では利用が困難であった。 In response to this request, a scene division method based on screen switching has been proposed in the past. However, in the case of this method, it is difficult to use it in a one-shot video in which a cut is not sandwiched.
 また、従来、インデックスによる検索方法が提案されている。この方法は、開始終了タイミングまたはその間の各フレームにインデックスを付与し、いずれかのインデックスを選択することで該当場面を提示する。そして、映像情報内の画像情報と文字情報とを複合的に用いてインデックスを付与し、インデックスに基づいて閾値等を用いた場面検索を可能とする方法が提案されている(例えば、非特許文献1参照)。 In addition, a search method using an index has been proposed in the past. In this method, an index is added to each frame at the start / end timing or in between, and the corresponding scene is presented by selecting one of the indexes. Then, a method has been proposed in which an index is given by using the image information and the character information in the video information in a complex manner, and a scene search using a threshold value or the like based on the index is possible (for example, a non-patent document). 1).
 さらに、タイムラインによる可視化方法が提案されている。この方法は、インデックス等で区切られる範囲を時系列に対応させてタイムラインで提示することで、タイムライン上から各場面の時間幅を比較することで分析なども可能となり、タイムライン上から必要な範囲を判別することも可能である。 Furthermore, a visualization method using a timeline has been proposed. This method is necessary from the timeline because it is possible to analyze by comparing the time width of each scene from the timeline by presenting the range delimited by the index etc. in chronological order on the timeline. It is also possible to determine a wide range.
 しかしながら、従来の方法では、インデックス化における場面欠落の回避が課題となっている。 However, in the conventional method, avoiding missing scenes in indexing is an issue.
 生データに基づいて複数のパラメータから複合的にインデックスを付ける場合においては、意図した検出範囲と合致するような、場面欠落の少ないパラメータ設定が難しい。また、インデックス付与を自動化した場合や、利用ユーザとインデックス設定者とが異なる場合など、パラメータが検出結果にどのように影響しているかがユーザにとって不明瞭である場合、ユーザには、検出しきれなかった関連場面を補完することも困難である。 When indexing multiple parameters in a complex manner based on raw data, it is difficult to set parameters with few scene omissions that match the intended detection range. In addition, if it is unclear to the user how the parameters affect the detection result, such as when indexing is automated or when the user and the index setter are different, the user cannot detect it. It is also difficult to supplement the related scenes that did not exist.
 各パラメータが閾値以上か、という判定の場合を例に説明する。この場合、部分的な条件が達成されている場合でも良いケースが抽出できないという問題がある。具体的には、「車の映っているシーンが見たい」場合に、条件に「タイヤが映っている」があると、タイヤ以外はほぼ全体が映っている場面も検出が困難である。 The case of determining whether each parameter is above the threshold value will be described as an example. In this case, there is a problem that a good case cannot be extracted even if the partial condition is achieved. Specifically, if "I want to see the scene in which the car is reflected" and "the tire is reflected" in the condition, it is difficult to detect the scene in which almost the entire scene is shown except for the tire.
 また、意味的に連続した場面の過剰な分断が生じるという問題がある。具体的には、本来検出したかった場面中に一部の条件が達成されない箇所があると、この箇所で検出場面が分断されてしまう。 There is also the problem that excessive division of semantically continuous scenes occurs. Specifically, if there is a place where some conditions are not achieved in the scene that was originally desired to be detected, the detected scene is divided at this place.
 また、一部の条件の推定精度が悪い場合、場面検出自体が困難になるおそれがある。また、個人差・個体差などにより条件の合致度合が異なる場合、全体をカバーする条件を設定するには条件の緩和が必要となり、総検出時間が長くなる上に、絞り込みが弱くなる。 Also, if the estimation accuracy of some conditions is poor, the scene detection itself may become difficult. In addition, when the degree of matching of the conditions differs due to individual differences, individual differences, etc., it is necessary to relax the conditions in order to set the conditions that cover the whole, the total detection time becomes long, and the narrowing down becomes weak.
 これらの問題を回避するためには適切なパラメータ設定が必要になるが、どのようにパラメータを設定すればいいかの判断がユーザには難しく、検出しきれなかった関連場面を補完することが難しい。 Appropriate parameter settings are required to avoid these problems, but it is difficult for the user to determine how to set the parameters, and it is difficult to supplement related situations that could not be detected. ..
 本発明は、上記に鑑みてなされたものであって、映像情報内の場面を検出する場合に、検出しきれなかった場面を適切に補完することが可能になる処理システム及び処理方法を提供することを目的とする。 The present invention has been made in view of the above, and provides a processing system and a processing method capable of appropriately complementing a scene that could not be detected when detecting a scene in the video information. The purpose is.
 上述した課題を解決し、目的を達成するために、本発明に係る処理システムは、映像情報に対応付けられた時系列情報を含むパラメータに基づいて映像情報内の場面を検出するための複数の条件と、指向性を持つ各条件間の関係性とを記憶する第1の記憶部と、任意の条件に関し、各条件間の関係性に基づいて、複数の条件の中から、任意の条件に関連する条件を求め、求めた条件に該当する場面を検出する検出部と、を有する。 In order to solve the above-mentioned problems and achieve the object, the processing system according to the present invention has a plurality of scenes in the video information based on parameters including time-series information associated with the video information. With respect to the first storage unit that stores the condition and the relationship between each directional condition, and any condition, based on the relationship between each condition, from a plurality of conditions to any condition. It has a detection unit that obtains related conditions and detects a scene that corresponds to the obtained conditions.
 本発明によれば、映像情報内の場面を検出する場合に、検出しきれなかった場面を適切に補完することが可能になる。 According to the present invention, when detecting a scene in video information, it is possible to appropriately supplement the scene that could not be detected.
図1は、実施の形態1における映像情報内の場面検出画面の一例を示す図である。FIG. 1 is a diagram showing an example of a scene detection screen in the video information according to the first embodiment. 図2は、実施の形態1に係る表示処理システムの機能構成の一例を示す図である。FIG. 2 is a diagram showing an example of the functional configuration of the display processing system according to the first embodiment. 図3は、関係性グラフの一例を示す図である。FIG. 3 is a diagram showing an example of a relationship graph. 図4は、図2に示す可視化情報表示部による表示例を説明する図である。FIG. 4 is a diagram illustrating a display example by the visualization information display unit shown in FIG. 2. 図5は、図2に示す表示処理システムが実行する条件及び条件間の関係性の設定処理の処理手順を示すフローチャートである。FIG. 5 is a flowchart showing a processing procedure of the conditions executed by the display processing system shown in FIG. 2 and the setting processing of the relationship between the conditions. 図6は、図2に示す表示処理システムが実行する指定条件に対する可視化情報の表示処理の処理手順を示すフローチャートである。FIG. 6 is a flowchart showing a processing procedure of display processing of visualization information for a specified condition executed by the display processing system shown in FIG. 図7は、図2に示す表示処理システムが実行する映像情報内の場面検出処理の処理手順を示すフローチャートである。FIG. 7 is a flowchart showing a processing procedure of scene detection processing in video information executed by the display processing system shown in FIG. 2. 図8は、従来方法における映像情報内の場面検出を説明する図である。FIG. 8 is a diagram illustrating scene detection in video information in the conventional method. 図9は、実施の形態1における映像情報内の場面検出を説明する図である。FIG. 9 is a diagram illustrating scene detection in video information according to the first embodiment. 図10は、実施の形態2における映像情報内の場面検出画面の一例を示す図である。FIG. 10 is a diagram showing an example of a scene detection screen in the video information according to the second embodiment. 図11は、実施の形態2に係る表示処理システムの機能構成の一例を示す図である。FIG. 11 is a diagram showing an example of the functional configuration of the display processing system according to the second embodiment. 図12は、図11に示す表示処理システムが実行する条件及び条件間の関係性の設定処理の処理手順を示すフローチャートである。FIG. 12 is a flowchart showing a processing procedure of the conditions executed by the display processing system shown in FIG. 11 and the setting processing of the relationship between the conditions. 図13は、図11に示す表示処理システムが実行する条件及び条件間の関係性の設定処理の他の処理手順を示すフローチャートである。FIG. 13 is a flowchart showing a condition executed by the display processing system shown in FIG. 11 and another processing procedure for setting the relationship between the conditions. 図14は、プログラムが実行されることにより、表示処理システムが実現されるコンピュータの一例を示す図である。FIG. 14 is a diagram showing an example of a computer in which a display processing system is realized by executing a program.
 以下、図面を参照して、本発明の一実施形態を詳細に説明する。なお、この実施の形態により本発明が限定されるものではない。また、図面の記載において、同一部分には同一の符号を付して示している。 Hereinafter, an embodiment of the present invention will be described in detail with reference to the drawings. The present invention is not limited to this embodiment. Further, in the description of the drawings, the same parts are indicated by the same reference numerals.
[実施の形態1]
 図1は、実施の形態1における映像情報内の場面検出画面の一例を示す図である。図1の上図には、従来方法による場面検出画面の一例も示す。
[Embodiment 1]
FIG. 1 is a diagram showing an example of a scene detection screen in the video information according to the first embodiment. The upper figure of FIG. 1 also shows an example of a scene detection screen by a conventional method.
 図1の上図に示すように、従来方法では、映像情報内から特定の場面を検出する際、映像に紐づいたパラメータを組み合わせることで、該当場面であるか否かを判別する検出条件が設定される。このとき、検索ユーザは、検出条件の作成者に与えられたラベルにより場面検索を行う。 As shown in the upper figure of FIG. 1, in the conventional method, when a specific scene is detected from the video information, a detection condition for determining whether or not the scene is applicable is set by combining parameters associated with the video. Set. At this time, the search user performs a scene search using the label given to the creator of the detection condition.
 例えば、検索ユーザが、映像情報の場面検出のために検索条件(以降、条件とする。)を指定すると、従来方法では、検出結果として、指定した条件に該当する場面の時間帯が着色されたタイムラインが表示される(図1の(1),(2)参照)。 For example, when a search user specifies a search condition (hereinafter referred to as a condition) for detecting a scene of video information, in the conventional method, the time zone of the scene corresponding to the specified condition is colored as a detection result. The timeline is displayed (see (1) and (2) in FIG. 1).
 しかしながら、従来方法では、このタイムラインが、検出したい場面の検出結果として不十分であったとしても、検出条件式(例えば、条件A~CのANDである(図1の(3)参照)等が開示されないため、ユーザ自身が、場面補完や場面絞り込みを行うことは難しい。また、仮に検索条件式(条件A~CのAND)が開示され編集可能であったとしても、条件A~Cのどのパラメータがどう影響しているかを把握することは検索ユーザには困難である。このため、検索ユーザは、条件指定を繰り返し設定して、複数回の検索を試行する必要がある。 However, in the conventional method, even if this timeline is insufficient as the detection result of the scene to be detected, the detection conditional expression (for example, AND of conditions A to C (see (3) in FIG. 1) and the like) and the like. Is not disclosed, so it is difficult for the user to complement or narrow down the scenes. Even if the search condition expressions (AND of conditions A to C) are disclosed and editable, the conditions A to C are not disclosed. It is difficult for the search user to understand which parameter affects how. Therefore, the search user needs to repeatedly set the condition specification and try the search a plurality of times.
 これに対し、実施の形態1では、映像情報に対応付けられた生データに基づき場面を検出する条件について、条件ごとに該条件が該当する場面の時刻情報を保持するとともに(図1の(4)参照)、条件間の指向的な関係性(論理和、論理積、包含関係など)を階層的に表す関係性グラフを保持する。例えば、行動であるならば、「端末操作」は、「たちどまる」、「視点を止める」のAND条件で表現できるという関係性を有する。また、オブジェクト検出であれば、「車」の下に「タイヤ」、「ライト」が含まれるなどの関係性を有する。なお、各条件間は完全独立でなくてもよく、下層の条件は必ずしも上層の必要条件でなくてもよい。 On the other hand, in the first embodiment, regarding the condition for detecting the scene based on the raw data associated with the video information, the time information of the scene to which the condition corresponds is retained for each condition ((4) in FIG. 1 (4). )), Holds a relation graph that hierarchically represents the directional relations between conditions (logical sum, AND, inclusion relations, etc.). For example, in the case of an action, "terminal operation" has a relation that it can be expressed by the AND condition of "stopping" and "stopping the viewpoint". Further, in the case of object detection, there is a relationship such that "tire" and "light" are included under "car". It should be noted that the conditions of each condition do not have to be completely independent, and the conditions of the lower layer do not necessarily have to be the necessary conditions of the upper layer.
 そして、実施の形態1では、この関係性グラフを用いることで、検索ユーザによる指定条件のタイムラインとともに、指定条件に関連する関連条件についても、その関係性を伴った状態でタイムラインを表示する。なお、実施の形態1では、任意の条件に該当する場面の時間帯を示す可視化情報としてタイムラインを例に説明するが、可視化情報は、タイムラインに限らず、条件に該当する場面の時間帯を示す文字情報であってもよい。 Then, in the first embodiment, by using this relation graph, the timeline of the designated condition by the search user and the related condition related to the designated condition are displayed in a state with the relation. .. In the first embodiment, a timeline will be described as an example of visualization information indicating a time zone of a scene corresponding to an arbitrary condition, but the visualization information is not limited to the timeline and is a time zone of a scene corresponding to the condition. It may be character information indicating.
 具体的には、実施の形態1では、図1の下図のように、ユーザによる指定条件に該当する場面の時間帯を示したタイムラインL1(第1の可視化情報)(条件A~CのAND)に加え、指定条件に関連する構成条件A,B,Cについても、それぞれタイムライン(第2の可視化情報)LA~LCを表示して可視化する(図1の(5)参照)。そして、タイムラインL1とタイムラインLA~LCとをAND記号M1(第3の可視化情報)で連結することで、指定条件と条件A~Dとの間の関係性を可視化する(図1の(6)参照)。 Specifically, in the first embodiment, as shown in the lower figure of FIG. 1, the timeline L1 (first visualization information) (AND of the conditions A to C) showing the time zone of the scene corresponding to the condition specified by the user. ), The configuration conditions A, B, and C related to the designated conditions are also visualized by displaying the timelines (second visualization information) LA to LC (see (5) in FIG. 1). Then, by connecting the timeline L1 and the timelines LA to LC with the AND symbol M1 (third visualization information), the relationship between the designated condition and the conditions A to D is visualized ((FIG. 1). 6)).
 さらに、構成条件Cの下位の条件C-1,C-2についても、タイムラインLC-1,LC-2を表示し、タイムラインLCとタイムラインLC-1,LC-2とをOR記号M2で連結することで(図1の枠W1参照)、条件Cが条件C-1と条件C-2のORであることを可視化する(図1の(7)参照)。 Further, for the conditions C-1 and C-2 lower than the configuration condition C, the timelines LC-1 and LC-2 are displayed, and the timeline LC and the timeline LC-1 and LC-2 are designated by the OR symbol M2. By connecting with (see frame W1 in FIG. 1), it is visualized that the condition C is an OR of the condition C-1 and the condition C-2 (see (7) in FIG. 1).
 このように、実施の形態1では、指定条件を満たす場面の時間帯を示すタイムラインL1とともに、指定条件を構成する構成条件A~Cに該当する場面の時間帯を示すタイムラインも階層表示する。階層表示されたこれらのタイムラインに参照することによって、検索ユーザは、指定条件の場面検出結果と併せて、各関連条件の場面検出結果を容易に把握できるようになる。そして、ユーザは、より多くの条件の場面検出結果を、優先度を付して参照することが可能であり、また、各条件間の関係性を把握する際に用途に限定しないことで、汎用的な検索に活用することも可能になる。 As described above, in the first embodiment, the timeline L1 indicating the time zone of the scene satisfying the designated condition and the timeline showing the time zone of the scene corresponding to the constituent conditions A to C constituting the designated condition are also hierarchically displayed. .. By referring to these hierarchically displayed timelines, the search user can easily grasp the scene detection result of each related condition together with the scene detection result of the specified condition. Then, the user can refer to the scene detection results of more conditions with priorities, and by not limiting the use to grasp the relationship between each condition, it is general purpose. It will also be possible to utilize it for specific searches.
 そして、検索ユーザは、指定条件では十分に場面を検出できなかった場合に、指定条件とその関連条件のタイムラインを参照することで、関連条件を用いて直感的に再検索を行うことが見込める。このため、本実施の形態1では、従来必要であった検索ユーザによるパラメータ調整または条件指定の繰り返し試行を低減できると考えられる。 Then, when the search user cannot sufficiently detect the scene under the specified condition, it can be expected that the search user intuitively performs the re-search using the related condition by referring to the timeline of the specified condition and the related condition. .. Therefore, in the first embodiment, it is considered that the repeated trials of parameter adjustment or condition specification by the search user, which are conventionally required, can be reduced.
 例えば、検索ユーザは、可視化されたタイムラインL1とタイムラインLA~LC-2とその関係性を参照することで、指定条件よりもさらに具体化された条件(例えば、条件A~C)や、より抽象的な条件(例えば、条件C-1,C-2)を容易に認識することができる。そして、検索ユーザは、これらの条件A~C-2を選択また組み合わせて条件を指定することで、検索ユーザによる条件設定の困難さが緩和される。したがって、実施の形態1によれば、検索ユーザによる映像情報からの場面検出の際に、検索ユーザによる場面検出を支援し、検出しきれなかった場面を適切に補完することが可能になる。 For example, the search user can refer to the visualized timeline L1 and the timelines LA to LC-2 and their relationships with each other to specify conditions (for example, conditions A to C) that are more specific than the specified conditions. More abstract conditions (eg, conditions C-1, C-2) can be easily recognized. Then, the search user selects or combines these conditions A to C-2 to specify the conditions, so that the difficulty of setting the conditions by the search user is alleviated. Therefore, according to the first embodiment, when the search user detects the scene from the video information, it is possible to support the scene detection by the search user and appropriately supplement the scene that could not be detected.
[表示処理システム]
 次に、実施の形態1に係る表示処理システムについて説明する。図2は、実施の形態1に係る表示処理システムの機能構成の一例を示す図である。
[Display processing system]
Next, the display processing system according to the first embodiment will be described. FIG. 2 is a diagram showing an example of the functional configuration of the display processing system according to the first embodiment.
 実施の形態1に係る表示処理システム10は、映像情報に対応付けられた時系列情報を含むパラメータに基づいて映像情報内の場面を検出するための複数の条件とともに、各条件間の関係性を保持する。これによって、表示処理システム10は、任意の条件による場面の検出結果とともに、任意の条件に関連する条件に該当する場面の検出結果を、各条件間の関係性を伴った状態で可視化することで、ユーザによる場面検出を支援し、検出しきれなかった場面を適切に補完できるようにしている。なお、図1の例では、表示処理システム10が、端末装置として機能する場合を想定して図示しているが、これに限定されるものではなく、サーバとして機能してもよく、検索した映像シーンをユーザ端末に出力するようにしてもよい。 The display processing system 10 according to the first embodiment sets a relationship between each condition together with a plurality of conditions for detecting a scene in the video information based on a parameter including time-series information associated with the video information. Hold. As a result, the display processing system 10 visualizes the detection result of the scene under the arbitrary condition and the detection result of the scene corresponding to the condition related to the arbitrary condition in a state with the relationship between each condition. , It assists the user in detecting the scene, and makes it possible to appropriately supplement the scene that could not be detected. In the example of FIG. 1, the display processing system 10 is shown assuming that it functions as a terminal device, but the present invention is not limited to this, and the display processing system 10 may function as a server, and the searched video may be used. The scene may be output to the user terminal.
 表示処理システム10は、生データ格納部11(第1の入力部)、データ処理部12、UI(User Interface)部13、条件格納部14(第1の記憶部)及び処理データ格納部15(第1の記憶部、検出部、第2の記憶部、第3の記憶部)を有する。以下では、各部について説明する。なお、上述した各部は、複数の装置が分散して保持してもよい。 The display processing system 10 includes a raw data storage unit 11 (first input unit), a data processing unit 12, a UI (User Interface) unit 13, a condition storage unit 14 (first storage unit), and a processing data storage unit 15 ( It has a first storage unit, a detection unit, a second storage unit, and a third storage unit). Each part will be described below. It should be noted that each of the above-mentioned parts may be held by a plurality of devices in a dispersed manner.
 表示処理システム10は、例えば、ROM(Read Only Memory)、RAM(Random Access Memory)、CPU(Central Processing Unit)等を含むコンピュータ等に所定のプログラムが読み込まれて、CPUが所定のプログラムを実行することで実現される。また、表示処理システム10は、ネットワーク等を介して接続された他の装置との間で、各種情報を送受信する通信インタフェースを有する。例えば、表示処理システム10は、NIC(Network Interface Card)等を有し、LAN(Local Area Network)やインターネットなどの電気通信回線を介した他の装置との間の通信を行う。そして、表示処理システム10は、タッチパネル、音声入力デバイス、キーボードやマウス等の入力デバイス、液晶ディスプレイなどの表示装置、プリンタ等の印刷装置を有し、情報の入出力を行う。 In the display processing system 10, for example, a predetermined program is read into a computer or the like including a ROM (Read Only Memory), a RAM (Random Access Memory), a CPU (Central Processing Unit), etc., and the CPU executes the predetermined program. It will be realized by. Further, the display processing system 10 has a communication interface for transmitting and receiving various information to and from other devices connected via a network or the like. For example, the display processing system 10 has a NIC (Network Interface Card) or the like, and communicates with other devices via a telecommunication line such as a LAN (Local Area Network) or the Internet. The display processing system 10 has a touch panel, a voice input device, an input device such as a keyboard and a mouse, a display device such as a liquid crystal display, and a printing device such as a printer, and inputs and outputs information.
 生データ格納部11は、検索対象の映像情報及び映像情報に併用する生データの入力を受け付け、格納する。生データは、映像情報の撮影を同期的に取得されたセンサ情報などである。センサ情報は、例えば、GPS(Global Positioning System)情報、加速度情報、温度情報である。 The raw data storage unit 11 receives and stores the video information to be searched and the input of raw data used in combination with the video information. The raw data is sensor information obtained synchronously with the shooting of video information. The sensor information is, for example, GPS (Global Positioning System) information, acceleration information, and temperature information.
 データ処理部12は、映像情報及び生データと、各条件とに基づき、各条件に該当する映像情報内の場面を検出し、各条件による場面検出結果を処理データ格納部15に出力する。また、データ処理部12は、映像情報を処理することで、映像情報内から生データを取得し、映像情報と生データとを対応付けて生データ格納部11に格納する。例えば、データ処理部12は、オブジェクト認識結果やSLAM(Simultaneous Localization and Mapping)による位置情報などを基に、生データを取得する。 The data processing unit 12 detects scenes in the video information corresponding to each condition based on the video information and raw data and each condition, and outputs the scene detection result according to each condition to the processing data storage unit 15. Further, the data processing unit 12 acquires raw data from the video information by processing the video information, and stores the video information and the raw data in the raw data storage unit 11 in association with each other. For example, the data processing unit 12 acquires raw data based on an object recognition result, position information by SLAM (Simultaneous Localization and Mapping), and the like.
 UI部13は、条件設定部131、可視化情報表示部132(第2の入力部、第1の表示部)、映像表示部133を有する。 The UI unit 13 has a condition setting unit 131, a visualization information display unit 132 (second input unit, first display unit), and a video display unit 133.
 条件設定部131は、検索ユーザによる入力デバイスの操作によって、場面検出のための検索条件(条件)の作成指示を受け付ける。条件設定部131は、条件の作成指示を受けると、任意の方法を用いて新たな条件を作成し、新たな条件と、新たな条件と既存の他の条件との関係性とを条件格納部14に格納する。 The condition setting unit 131 receives an instruction to create a search condition (condition) for scene detection by operating the input device by the search user. Upon receiving an instruction to create a condition, the condition setting unit 131 creates a new condition using an arbitrary method, and stores the new condition and the relationship between the new condition and other existing conditions. Store in 14.
 可視化情報表示部132は、映像情報の場面検出のために指定された検索条件(指定条件)(第1の条件)の入力を受けつける。可視化情報表示部132は、指定条件を処理データ格納部15に出力する。そして、可視化情報表示部132は、処理データ格納部15から出力された場面検出結果を含む情報を基に、場面検出に関する可視化情報を出力する。可視化情報表示部132は、可視化情報を参照した検索ユーザによる入力デバイスの操作によって、映像再生場面の指定を受け付けると、再生範囲を映像表示部133に出力する。 The visualization information display unit 132 receives the input of the search condition (designated condition) (first condition) designated for scene detection of video information. The visualization information display unit 132 outputs the designated condition to the processing data storage unit 15. Then, the visualization information display unit 132 outputs visualization information related to scene detection based on the information including the scene detection result output from the processing data storage unit 15. When the visualization information display unit 132 receives the designation of the video reproduction scene by the operation of the input device by the search user who has referred to the visualization information, the visualization information display unit 132 outputs the reproduction range to the video display unit 133.
 映像表示部133は、可視化情報表示部132から再生範囲の指定を受け付けると、生データ格納部11に格納された映像情報を基に、指定範囲の映像を再生する。 When the video display unit 133 receives the designation of the playback range from the visualization information display unit 132, the video display unit 133 reproduces the video in the designated range based on the video information stored in the raw data storage unit 11.
 条件格納部14は、映像情報内の場面を検出するための複数の条件と、指向性を持つ各条件間の関係性を示す関係性グラフとを記憶する。条件格納部14は、関係性グラフが登録または更新された場合には、登録または更新された関係性グラフを処理データ格納部15に出力する。 The condition storage unit 14 stores a plurality of conditions for detecting a scene in the video information and a relationship graph showing the relationship between each condition having directivity. When the relationship graph is registered or updated, the condition storage unit 14 outputs the registered or updated relationship graph to the processing data storage unit 15.
 図3は、関係性グラフの一例を示す図である。図3に示すように、関係グラフG1は、各条件について、論理積、論理和、または、時系列な繋がりや包含関係などを示すことで、網羅的に各条件の関係性を保持する(図3の(1)参照)。例えば、上位の条件がある条件間の論理和や論理積の結果である場合、上位の条件がある条件を包含している場合、上位の条件が、ある複数条件が時系列で発生することで満たされる場合などには、上位の条件の配下に関連条件が紐づく。 FIG. 3 is a diagram showing an example of a relationship graph. As shown in FIG. 3, the relation graph G1 comprehensively holds the relation of each condition by showing the logical product, the logical sum, or the time-series connection and the inclusion relation for each condition (FIG. 3). 3 (1)). For example, if the upper condition is the result of a logical sum or AND between certain conditions, or if the upper condition includes a certain condition, the upper condition may occur due to multiple conditions occurring in chronological order. When it is satisfied, the related condition is linked to the subordinate of the higher condition.
 具体的には、関係グラフG1において、関係グラフG1では、条件「現状調査」に対して時系列につながる配下の条件として、「点検」及び「記録」がある。関係グラフG1では、方向性(「点検」→「記録」)と発生時間(例えば、10秒以内)といった条件を含めて、各条件間を紐づけてもよい。また、「点検」という条件には、「停留」及び「注視」の条件があり、この二つは論理積の関係で「点検」と紐づけられる。関係グラフG1では、より具体的な条件の配下に、抽象的な条件が関係づけられている。この関係グラフG1に示すように、条件の関係性を多層にすることで、参照する要件のユーザによる優先順位付けを簡易化することができる。そして、関係グラフG1のように、条件の関係性を多層にすることで、例えば「点検」及び「注視2秒」のように、間接的に関連する条件を判別することもできる。 Specifically, in the relation graph G1, in the relation graph G1, there are "inspection" and "record" as subordinate conditions connected to the time series with respect to the condition "current situation survey". In the relationship graph G1, each condition may be linked, including conditions such as directionality (“inspection” → “recording”) and occurrence time (for example, within 10 seconds). In addition, the condition of "inspection" includes the conditions of "stop" and "gaze", and these two are associated with "inspection" in relation to the logical product. In the relationship graph G1, abstract conditions are related under more specific conditions. As shown in the relationship graph G1, by making the relationship of conditions into multiple layers, it is possible to simplify the prioritization of the requirements to be referred to by the user. Then, by making the relation of the conditions into multiple layers as in the relation graph G1, it is also possible to discriminate indirectly related conditions such as "inspection" and "gaze 2 seconds".
 処理データ格納部15は、関係性グラフを保持し、任意の条件に関し、関係性グラフに基づいて、複数の条件の中から任意の条件に関連する条件を求め、求めた条件に該当する場面を検出する。 The processing data storage unit 15 holds a relation graph, obtains a condition related to an arbitrary condition from a plurality of conditions based on the relation graph with respect to an arbitrary condition, and sets a scene corresponding to the obtained condition. To detect.
 処理データ格納部15は、関係性グラフとともに、各条件による場面検出結果を保持する。各条件による場面検出結果は、条件格納部14が記憶する各条件を用いて映像情報から検出された場面の検出結果であり、データ処理部12による処理によって求められる。 The processing data storage unit 15 holds the scene detection result according to each condition together with the relationship graph. The scene detection result according to each condition is the detection result of the scene detected from the video information using each condition stored in the condition storage unit 14, and is obtained by the processing by the data processing unit 12.
 処理データ格納部15は、可視化情報表示部152から指定条件が入力されると、関係性グラフに基づいて、複数の条件の中から、指定条件と指定条件に関連する関連条件(第2の条件)とを求める。そして、処理データ格納部15は、各条件による場面検出結果のうち、指定条件に該当する場面検出結果と関連条件に該当する場面検出結果とを、可視化情報表示部132に出力する。これとともに、処理データ格納部15は、関係性グラフに基づいて、指定条件及び関連条件の関係性を示す情報を可視化情報表示部132に出力する。 When the designated condition is input from the visualization information display unit 152, the processing data storage unit 15 has the designated condition and the related condition related to the designated condition (second condition) from among a plurality of conditions based on the relation graph. ) And ask. Then, the processing data storage unit 15 outputs, among the scene detection results according to each condition, the scene detection result corresponding to the designated condition and the scene detection result corresponding to the related condition to the visualization information display unit 132. At the same time, the processing data storage unit 15 outputs information indicating the relationship between the designated condition and the related condition to the visualization information display unit 132 based on the relationship graph.
 そして、可視化情報表示部132は、処理データ格納部15から指定条件に該当する場面検出結果と、関連条件に該当する場面の検出結果と、指定条件及び関連条件の関係性を示す情報とを受け、指定条件に該当する場面の時間帯を示す第1の可視化情報と、関連条件に該当する場面の時間帯を示す第2の可視化情報と、指定条件及び関連条件との関係性を示す第3の可視化情報とを出力する。 Then, the visualization information display unit 132 receives from the processing data storage unit 15 the scene detection result corresponding to the designated condition, the detection result of the scene corresponding to the related condition, and the information indicating the relationship between the designated condition and the related condition. , The first visualization information indicating the time zone of the scene corresponding to the designated condition, the second visualization information indicating the time zone of the scene corresponding to the related condition, and the third indicating the relationship between the designated condition and the related condition. Visualization information and output.
 図4は、図2に示す可視化情報表示部132による表示例を説明する図である。可視化情報表示部132は、図4のように、ユーザによる指定条件「点検」に該当する場面の時間帯を示したタイムラインに加え、関係グラフG1に示す関係性にしたがって、関連条件「停留」、「注視5秒」、「移動」、「見渡し」が該当する場面の時間帯を示すタイムラインを指定条件「点検」の配下であることを示す枠や記号を第3の可視化情報として表示する。さらに、可視化情報表示部132は、関連条件「注視5秒」が含む条件「注視2秒」に関するタイムラインを表示する。このように、可視化情報表示部132が、指定条件とその関連情報のタイムラインを階層表示することによって、検索ユーザは、指定条件「点検」に関する関連情報を参照して、より多くの関連情報を直感的に検索条件として利用することができる(図4の(1)参照)。 FIG. 4 is a diagram illustrating a display example by the visualization information display unit 132 shown in FIG. As shown in FIG. 4, the visualization information display unit 132 has a timeline showing the time zone of the scene corresponding to the user-designated condition “inspection”, and the related condition “stop” according to the relationship shown in the relationship graph G1. , "Gaze 5 seconds", "Move", "Overlook" is displayed as the third visualization information with a frame or symbol indicating that the timeline indicating the time zone of the corresponding scene is under the specified condition "Inspection". .. Further, the visualization information display unit 132 displays a timeline relating to the condition "gaze 2 seconds" included in the related condition "gaze 5 seconds". In this way, the visualization information display unit 132 hierarchically displays the timeline of the designated condition and its related information, so that the search user can refer to the related information regarding the designated condition "inspection" and display more related information. It can be intuitively used as a search condition (see (1) in FIG. 4).
 なお、可視化情報表示部132は、タイムライン形式に限らず、任意の可視化形式を使用できる。また、可視化情報表示部132は、タイムラインなどの表示方法の変更や条件間のグルーピングなどを行ってもよく、検索ユーザからタイムラインなどの表示方法の変更や条件間のグルーピングなどを受け付けてもよい。 The visualization information display unit 132 can use any visualization format, not limited to the timeline format. Further, the visualization information display unit 132 may change the display method such as the timeline or group the conditions, or may accept the change of the display method such as the timeline or the grouping between the conditions from the search user. good.
[条件及び条件間の関係性の設定]
 次に、表示処理システム10が実行する処理について説明する。まず、表示処理システム10が実行する条件及び条件間の関係性の設定処理について説明する。図5は、図2に示す表示処理システム10が実行する条件及び条件間の関係性の設定処理の処理手順を示すフローチャートである。
[Setting conditions and relationships between conditions]
Next, the processing executed by the display processing system 10 will be described. First, the conditions executed by the display processing system 10 and the setting processing of the relationships between the conditions will be described. FIG. 5 is a flowchart showing a processing procedure of the conditions executed by the display processing system 10 shown in FIG. 2 and the setting processing of the relationship between the conditions.
 図5に示すように、条件設定部131が、検索ユーザによる入力デバイスの操作によって、場面検出のために抽出したい条件名またはIDの登録指示を受け付ける(ステップS1)。条件設定部131は、任意の方法を用いて、利用可能なパラメータから、該当条件を表現するものを1つ以上選び、閾値や基準値等の数値条件を設定する(ステップS2)。なお、ステップS1において、名前を指定し、ステップS2において閾値等を設定される条件が、ステップS3(後述)「新規条件」に対応する。 As shown in FIG. 5, the condition setting unit 131 receives an instruction to register a condition name or ID to be extracted for scene detection by operating an input device by a search user (step S1). The condition setting unit 131 selects one or more parameters expressing the corresponding condition from the available parameters by using an arbitrary method, and sets numerical conditions such as a threshold value and a reference value (step S2). The condition in which the name is specified in step S1 and the threshold value or the like is set in step S2 corresponds to the "new condition" in step S3 (described later).
 続いて、条件設定部131は、条件格納部14から、新規条件に関連する条件を抽出する(ステップS3)。新規条件は、条件格納部14に既存する条件の論理和や論理積である条件、既存条件を包含する条件、既存条件を時系列での組み合わせることで表されたる条件である。新規条件に関連する条件は、既存の条件から関連する条件としてユーザが指定した条件、表示処理システム10が数値条件を比較して関連すると判定して自動的に検出した条件、或いは、新規条件を分割して部分的な条件を作成し紐づけた条件などである。また、新規条件は、必ずしも既存条件の組合せで表現しなければいけないわけではない。例えば、ユーザによって生成された新規条件について、関連する条件を手動または数値条件比較で既存条件から探索することによって、或いは、新規条件の分割によって、新たに新規条件を生成してもよい。 Subsequently, the condition setting unit 131 extracts the condition related to the new condition from the condition storage unit 14 (step S3). The new condition is a condition that is a logical sum or a logical product of the existing conditions in the condition storage unit 14, a condition that includes the existing condition, and a condition that is expressed by combining the existing conditions in chronological order. The condition related to the new condition is a condition specified by the user as a related condition from the existing condition, a condition automatically detected by the display processing system 10 as being related by comparing the numerical conditions, or a new condition. It is a condition that is divided and a partial condition is created and linked. Moreover, the new condition does not necessarily have to be expressed by a combination of existing conditions. For example, for a new condition generated by a user, a new condition may be newly generated by searching for a related condition from an existing condition manually or by comparing numerical conditions, or by dividing a new condition.
 条件格納部14は、条件設定部131により追加された条件を登録し、その条件と既存条件との間の関係性グラフを更新することで(ステップS4)、より具体的な条件を上層に設定する。 The condition storage unit 14 registers the condition added by the condition setting unit 131, and updates the relationship graph between the condition and the existing condition (step S4) to set a more specific condition in the upper layer. do.
[指定条件に対する可視化情報の表示処理]
 次に、表示処理システム10が実行する指定条件に対する可視化情報の表示処理について説明する。図6は、図2に示す表示処理システム10が実行する指定条件に対する可視化情報の表示処理の処理手順を示すフローチャートである。
[Display processing of visualization information for specified conditions]
Next, the display processing of the visualization information for the designated condition executed by the display processing system 10 will be described. FIG. 6 is a flowchart showing a processing procedure of display processing of visualization information for a designated condition executed by the display processing system 10 shown in FIG.
 図6に示すように、可視化情報表示部132は、検索ユーザから、場面検出のために指定された検索条件(指定条件)の入力を受け付けると(ステップS11)、指定条件を処理データ格納部15に出力する。 As shown in FIG. 6, when the visualization information display unit 132 receives an input of a search condition (designated condition) designated for scene detection from a search user (step S11), the visualization information display unit 132 processes the specified condition. Data storage unit 15 Output to.
 処理データ格納部15は、複数の条件の中から、指定条件、指定条件に関連する関連条件を求め、指定条件に該当する場面検出結果と関連条件に該当する場面検出結果とを検出して可視化情報表示部132に出力するとともに、関係性グラフに基づいて、指定条件及び関連条件の関係性を示す情報を可視化情報表示部132に出力する(ステップS12)。 The processing data storage unit 15 obtains a specified condition and a related condition related to the specified condition from a plurality of conditions, and detects and visualizes a scene detection result corresponding to the specified condition and a scene detection result corresponding to the related condition. In addition to outputting to the information display unit 132, information indicating the relationship between the designated condition and the related condition is output to the visualization information display unit 132 based on the relationship graph (step S12).
 可視化情報表示部132は、図1及び図4に例示したように、指定条件及び関連情報のそれぞれのタイムラインを、各条件間の関係性とともに表示出力する(ステップS13)。 As illustrated in FIGS. 1 and 4, the visualization information display unit 132 displays and outputs each timeline of the designated condition and the related information together with the relationship between the conditions (step S13).
 そして、可視化情報表示部132は、可視化情報を参照した検索ユーザによる入力デバイスから、映像再生場面の指定を受け付けると(ステップS14)、再生範囲を映像表示部133に出力する。映像表示部133は、可視化情報表示部132から再生範囲の指定を受け付けると、生データ格納部11に格納された映像情報を基に、指定範囲の映像を再生する(ステップS15)。 Then, when the visualization information display unit 132 receives the designation of the video reproduction scene from the input device by the search user who referred to the visualization information (step S14), the visualization information display unit 132 outputs the reproduction range to the video display unit 133. When the video display unit 133 receives the designation of the reproduction range from the visualization information display unit 132, the video display unit 133 reproduces the video in the designated range based on the video information stored in the raw data storage unit 11 (step S15).
[映像情報内の場面検出処理]
 次に、表示処理システム10が実行する映像情報内の場面検出処理について説明する。図7は、図2に示す表示処理システム10が実行する映像情報内の場面検出処理の処理手順を示すフローチャートである。
[Scene detection processing in video information]
Next, the scene detection process in the video information executed by the display processing system 10 will be described. FIG. 7 is a flowchart showing a processing procedure of scene detection processing in video information executed by the display processing system 10 shown in FIG.
 図7に示すように、表示処理システム10は、生データ格納部11のデータ、または、条件格納部の条件に更新があったか否かを判定する(ステップS21)。表示処理システム10は、生データ格納部11のデータ、または、条件格納部の条件に更新があるまで、ステップS21の判定処理を繰り返す。 As shown in FIG. 7, the display processing system 10 determines whether or not the data in the raw data storage unit 11 or the condition in the condition storage unit has been updated (step S21). The display processing system 10 repeats the determination process in step S21 until the data in the raw data storage unit 11 or the condition in the condition storage unit is updated.
 そして、表示処理システム10は、生データ格納部11のデータ、または、条件格納部の条件に更新があった場合(ステップS21:Yes)、自システム内で映像情報から情報を抽出するか否かを判定する(ステップS22)。自システム内で映像情報から情報を抽出する場合(ステップS22:Yes)、データ処理部12は、映像情報からの、GPS情報、加速度情報、温度情報等の情報抽出を実行する(ステップS23)。 Then, when the data of the raw data storage unit 11 or the condition of the condition storage unit is updated (step S21: Yes), the display processing system 10 determines whether or not to extract information from the video information in the own system. Is determined (step S22). When extracting information from video information in the own system (step S22: Yes), the data processing unit 12 executes information extraction such as GPS information, acceleration information, and temperature information from the video information (step S23).
 自システム内で映像情報から情報を抽出しない場合(ステップS22:No)、または、ステップS23終了後、データ処理部12は、条件格納部14に格納された各条件の検出条件を参照する(ステップS24)。これに応じて、条件格納部14は、格納された検出条件をデータ処理部12に出力する(ステップS25)。 When the information is not extracted from the video information in the own system (step S22: No), or after the end of step S23, the data processing unit 12 refers to the detection condition of each condition stored in the condition storage unit 14 (step). S24). In response to this, the condition storage unit 14 outputs the stored detection condition to the data processing unit 12 (step S25).
 データ処理部12は、各条件に基づき、映像情報内の該当場面を検出し(ステップS26)、場面検出結果を処理データ格納部15に出力する。処理データ格納部15は、各条件の場面検出結果を保存して(ステップS27)、処理を終了する。 The data processing unit 12 detects the corresponding scene in the video information based on each condition (step S26), and outputs the scene detection result to the processing data storage unit 15. The processing data storage unit 15 saves the scene detection result of each condition (step S27), and ends the processing.
[実施の形態1の効果]
 図8は、従来方法における映像情報内の場面検出を説明する図である。従来方法では、ユーザは、検出条件の作成者によって用意されたラベル単位でのみ場面検索が可能であるため(図8の(1)参照)、指定した条件に該当する場面の時間帯が着色されたタイムラインのみが検出結果として表示される。
[Effect of Embodiment 1]
FIG. 8 is a diagram illustrating scene detection in video information in the conventional method. In the conventional method, since the user can search the scene only in the label unit prepared by the creator of the detection condition (see (1) in FIG. 8), the time zone of the scene corresponding to the specified condition is colored. Only the timeline is displayed as the detection result.
 しかしながら、従来方法では、このタイムラインが、検出したい条件の検出結果として不十分であったとしても、具体的な検出条件式はユーザに見えないため、ユーザ自身が、場面補完や場面絞り込みを行うことは難しい。また、仮に条件式が編集可能だとしても、どのパラメータがどう影響しているかの把握がユーザには困難なため、条件指定を繰り返し設定して、複数回の検索を試行する必要がある。 However, in the conventional method, even if this timeline is insufficient as the detection result of the condition to be detected, since the specific detection condition expression is not visible to the user, the user himself performs scene complementation and scene narrowing. That is difficult. Even if the conditional expression can be edited, it is difficult for the user to understand which parameter affects how, so it is necessary to repeatedly set the conditional specification and try the search multiple times.
 図9は、実施の形態1における映像情報内の場面検出を説明する図である。従来方法に対し、本実施の形態1に係る表示処理方法は、映像情報内の場面を検出するための複数の条件と、指向性を持つ各条件間の関係性を示す関係性グラフとを保持し、指定条件に関し、各条件間の関係性に基づいて、複数の条件の中から、指定条件に関連する関連条件を求め、求めた条件に該当する場面を検出する。そして、本実施の形態1では、図9に示すように、指定条件に該当する場面の時間帯を示したタイムラインに加え、関連条件に該当する場面の時間帯を示したタイムラインを関連条件ごとに表示するとともに、指定条件の配下に関連条件を紐づけて、タイムラインを階層化表示することで各条件間の関係性も表示している。 FIG. 9 is a diagram illustrating scene detection in video information according to the first embodiment. In contrast to the conventional method, the display processing method according to the first embodiment holds a plurality of conditions for detecting a scene in video information and a relationship graph showing a relationship between each condition having directivity. Then, regarding the specified condition, based on the relationship between each condition, the related condition related to the specified condition is obtained from a plurality of conditions, and the scene corresponding to the obtained condition is detected. Then, in the first embodiment, as shown in FIG. 9, in addition to the timeline showing the time zone of the scene corresponding to the designated condition, the timeline showing the time zone of the scene corresponding to the related condition is used as the related condition. In addition to displaying each condition, the relationship between each condition is also displayed by linking the related conditions under the specified condition and displaying the timeline in a hierarchical manner.
 検索ユーザは、映像情報内から特定の場面を検出したい場合に、階層的に可視化された指定条件及び関連条件の検出状況(0/1、または、連続値)を参照することで、関連条件による場面検出結果と比較させながら、指定条件の場面検出結果を参照することができる。そして、検出結果が不十分である場合、検索ユーザは、関連条件の場面検出結果を参照することで場面を絞り込むことができる。例えば、検索ユーザは、指定条件に該当する場面を優先的に取得したい場合には、指定条件に対応するタイムラインのうち枠W11,W12,W13内の時間帯で場面を絞り込めばよい(図9の(1)参照)。 When the search user wants to detect a specific scene from the video information, he / she can refer to the detection status (0/1 or continuous value) of the specified condition and the related condition that are visualized hierarchically, and the related condition is used. It is possible to refer to the scene detection result of the specified condition while comparing it with the scene detection result. Then, when the detection result is insufficient, the search user can narrow down the scene by referring to the scene detection result of the related condition. For example, when the search user wants to preferentially acquire the scenes corresponding to the specified conditions, the search user may narrow down the scenes in the time zone within the frames W11, W12, and W13 in the timeline corresponding to the specified conditions (Fig.). 9 (1)).
 また、指定条件に関連する関連条件の場面検出結果を可視化することによって、想定外の条件の組み合わせから有効な場面を検出することも可能である。例えば、検索ユーザは、枠W14内の条件A,B,Cの場面検知結果より、条件A,Cは必須だが、条件Bが満たされないだけであれば場面の使用用途には有効であることが判別できるため、条件A,Cの組み合わせのみに条件を設定することもできる。そして、検索ユーザは、条件Cの包含条件を参照することで、Cの条件を少し緩和してもよい、などの条件変更も可能である。このように、本実施の形態1によれば、ユーザの任意で複数条件のグルーピングや、特定条件の一時的な非表示などを行うことを想定した場合も、対応可能である。 It is also possible to detect effective scenes from unexpected combinations of conditions by visualizing the scene detection results of related conditions related to the specified conditions. For example, the search user may find that the conditions A and C are indispensable from the scene detection results of the conditions A, B, and C in the frame W14, but are effective for the usage of the scene if the condition B is not satisfied. Since it can be discriminated, it is possible to set the condition only for the combination of the conditions A and C. Then, the search user can change the condition such that the condition C may be relaxed a little by referring to the inclusion condition of the condition C. As described above, according to the first embodiment, it is possible to deal with the case where it is assumed that the user arbitrarily groups a plurality of conditions or temporarily hides a specific condition.
 したがって、本実施の形態1によれば、検索ユーザは、指定条件の検出結果と併せて、各関連条件の場面検出結果を容易に把握できるようになる。そして、本実施の形態1によれば、検索ユーザは、指定条件では十分に場面を検出できなかった場合に、指定条件とその関連条件の検出結果を参照することで、関連条件を用いて直感的に再検索を行うことが見込める。このため、本実施の形態1では、従来必要であったユーザによるパラメータ調整または条件指定の繰り返しの試行が低減できると考えられる。このため、実施の形態1によれば、検索ユーザによる映像情報からの場面検出の際に、検出しきれなかった場面を適切に補完することが可能になる。 Therefore, according to the first embodiment, the search user can easily grasp the scene detection result of each related condition together with the detection result of the designated condition. Then, according to the first embodiment, when the search user cannot sufficiently detect the scene under the specified condition, the search user intuitively uses the related condition by referring to the specified condition and the detection result of the related condition. It is expected that the search will be performed again. Therefore, in the first embodiment, it is considered that the user's repeated trials of parameter adjustment or condition specification, which have been conventionally required, can be reduced. Therefore, according to the first embodiment, when the search user detects the scene from the video information, it is possible to appropriately supplement the scene that could not be detected.
[実施の形態2]
 次に、実施の形態2について説明する。図10は、実施の形態2における映像情報内の場面検出画面の一例を示す図である。実施の形態2では、検索ユーザによって今後検出したい場面の範囲が指定された場合に、この範囲の場面を映像情報内から検出できる各条件の比較参照を可能とする。
[Embodiment 2]
Next, the second embodiment will be described. FIG. 10 is a diagram showing an example of a scene detection screen in the video information according to the second embodiment. In the second embodiment, when a range of scenes to be detected in the future is specified by the search user, it is possible to compare and refer to each condition that can detect the scenes in this range from the video information.
 例えば、検索ユーザが、今後検出したい場面として、映像情報内の任意の時間帯を指定すると(図10の(1)参照)、本実施の形態2では、複数の条件の中から、指定された時間帯に検出場面がある条件(例えば、条件A~C)を求める。そして、本実施の形態2では、求められた条件ごとに、各条件が該当する時間帯のうち、指定された時間帯を強調表示したタイムラインLA~LC-2を、階層出力する。例えば、強調表示として、タイムラインLA~LC-2のうち、指定された時間帯に対応する部分を、枠W21で囲むことや、他とは異なる色または明度で表示させることや、点滅表示させることが考えられる。 For example, when the search user specifies an arbitrary time zone in the video information as a scene to be detected in the future (see (1) in FIG. 10), in the second embodiment, the scene is specified from a plurality of conditions. A condition (for example, conditions A to C) in which a detection scene exists in a time zone is obtained. Then, in the second embodiment, the timelines LA to LC-2 highlighting the designated time zone among the time zones to which each condition corresponds are hierarchically output for each requested condition. For example, as highlighting, the part of the timelines LA to LC-2 corresponding to the specified time zone is surrounded by the frame W21, displayed in a color or brightness different from the others, or blinks. Can be considered.
 このように、本実施の形態2では、検索ユーザは、場面を検出したい時間帯を指定することで、各条件のうち、この時間帯に該当する場面が、どのような組合せの条件で検出されているかを比較参照することができる(図10の(2)参照)。そして、検索ユーザは、指定された時間帯を強調表示したタイムラインLA~LC-2を比較参照することによって、場面を検出したい時間帯に対し、次回以降の検出条件を作成することができ、より効率的に特定場面を検出することができる。 As described above, in the second embodiment, the search user specifies the time zone in which the scene is to be detected, and the scene corresponding to this time zone is detected under any combination of the conditions. It can be compared and referred to (see (2) in FIG. 10). Then, the search user can create detection conditions from the next time onward for the time zone in which the scene is to be detected by comparing and referring to the timelines LA to LC-2 highlighting the designated time zone. A specific scene can be detected more efficiently.
 そして、本実施の形態2では、場面検出条件A~C-2を、関係性グラフが有する階層性に応じて並べることで、検索ユーザは、より具体的な条件またはより限定的な条件から優先的に判断して、検索に有効な検出条件を作成することができる。したがって、実施の形態2によれば、検索ユーザが所望する場面を検出できる検出条件の作成を支援することによって、検索ユーザの希望に沿った場面検出を可能とする。 Then, in the second embodiment, by arranging the scene detection conditions A to C-2 according to the hierarchy of the relationship graph, the search user gives priority to more specific conditions or more limited conditions. It is possible to create a detection condition that is effective for the search. Therefore, according to the second embodiment, it is possible to detect a scene according to the desire of the search user by supporting the creation of a detection condition capable of detecting the scene desired by the search user.
[表示処理システム]
 次に、実施の形態2に係る表示処理システムについて説明する。図11は、実施の形態2に係る表示処理システムの機能構成の一例を示す図である。
[Display processing system]
Next, the display processing system according to the second embodiment will be described. FIG. 11 is a diagram showing an example of the functional configuration of the display processing system according to the second embodiment.
 実施の形態2に係る表示処理システム210は、図2に示す表示処理システム10と比して、UI部213及び処理データ格納部215(設定部)を有する。また、表示処理システム210は、UI部213は、条件設定部2131(第1の設定部)と、可視化情報表示部2132(第2の表示部)と、映像表示部2133(第3の入力部、第4の入力部)とを有する。 The display processing system 210 according to the second embodiment has a UI unit 213 and a processing data storage unit 215 (setting unit) as compared with the display processing system 10 shown in FIG. Further, in the display processing system 210, the UI unit 213 has a condition setting unit 2131 (first setting unit), a visualization information display unit 2132 (second display unit), and a video display unit 2133 (third input unit). , A fourth input unit).
 映像表示部2133は、検索ユーザの任意の位置から映像情報を再生する。そして、映像表示部2133は、検索ユーザによる入力デバイスの操作によって、映像情報内の任意の時間帯を指定する時間帯指定情報の入力を受け付ける。映像表示部2133は、時間帯指定情報によって指定された時間帯を可視化情報表示部2132に出力する。 The video display unit 2133 reproduces video information from any position of the search user. Then, the video display unit 2133 accepts the input of the time zone designation information for designating an arbitrary time zone in the video information by the operation of the input device by the search user. The video display unit 2133 outputs the time zone designated by the time zone designation information to the visualization information display unit 2132.
 可視化情報表示部2132は、時間帯指定情報によって指定された時間帯を処理データ格納部215に出力する。可視化情報表示部2132は、時間帯指定情報に応じた可視化情報を出力する。 The visualization information display unit 2132 outputs the time zone specified by the time zone designation information to the processing data storage unit 215. The visualization information display unit 2132 outputs visualization information according to the time zone designation information.
 処理データ格納部215は、各条件による場面検出結果を基に、複数の条件の中から、時間帯指定情報において指定された時間帯に検出場面がある条件を求め、求めた条件による場面検出結果と、求めた条件間の関係性を示す関係性情報を可視化情報表示部2132に出力する。 The processing data storage unit 215 obtains a condition in which the detected scene is in the time zone specified in the time zone designation information from a plurality of conditions based on the scene detection result according to each condition, and the scene detection result according to the obtained condition. And, the relationship information indicating the relationship between the obtained conditions is output to the visualization information display unit 2132.
 可視化情報表示部2132は、処理データ格納部215によって求められた条件ごとに、各条件に該当する場面の時間帯のうち時間帯指定情報において指定された時間帯を強調表示した可視化情報を出力するとともに、各条件間の関係性を示す可視化情報を出力する。 The visualization information display unit 2132 outputs visualization information highlighting the time zone specified in the time zone designation information among the time zones of the scenes corresponding to each condition for each condition obtained by the processing data storage unit 215. At the same time, visualization information showing the relationship between each condition is output.
 条件設定部2131は、出力された可視化情報を参照した検索ユーザによる入力デバイスの操作によって、時間帯指定情報において指定された時間帯の場面を検出する条件と各検出条件間の関係性とに対する登録指示または更新指示を受け付ける。条件設定部2131は、受け付けた登録指示または更新指示に応じて、条件格納部14が記憶する条件と各条件間の関係性とを登録または更新する。 The condition setting unit 2131 registers the condition for detecting the scene in the time zone specified in the time zone specification information and the relationship between each detection condition by the operation of the input device by the search user referring to the output visualization information. Accept instructions or update instructions. The condition setting unit 2131 registers or updates the conditions stored in the condition storage unit 14 and the relationship between the conditions in response to the received registration instruction or update instruction.
[指定された時間帯の場面の検出条件の登録処理]
 次に、表示処理システム210が実行する指定された時間帯の場面の検出条件の登録処理について説明する。図12は、図11に示す表示処理システム210が実行する条件及び条件間の関係性の設定処理の処理手順を示すフローチャートである。
[Registration process of detection conditions for scenes in the specified time zone]
Next, the process of registering the detection condition of the scene in the designated time zone executed by the display processing system 210 will be described. FIG. 12 is a flowchart showing a processing procedure of the condition executed by the display processing system 210 shown in FIG. 11 and the setting processing of the relationship between the conditions.
 図12に示すように、映像表示部2133は、検索ユーザによって指定された任意の位置から映像情報を再生する(ステップS31)。そして、映像表示部2133は、検索ユーザによる入力デバイスの操作によって、映像情報内の時間帯の指定を受けたか否かを判定する(ステップS32)。映像表示部2133は、映像情報内の時間帯の指定を受けていない場合(ステップS32:No)、ステップS32の判定処理に戻る。なお、検索ユーザは、映像情報内の複数の時間帯を同時に指定することも可能である。 As shown in FIG. 12, the video display unit 2133 reproduces video information from an arbitrary position designated by the search user (step S31). Then, the video display unit 2133 determines whether or not the time zone specified in the video information has been specified by the operation of the input device by the search user (step S32). If the video display unit 2133 has not received the time zone designation in the video information (step S32: No), the video display unit 2133 returns to the determination process of step S32. The search user can also specify a plurality of time zones in the video information at the same time.
 映像情報内の時間帯の指定を受けた場合(ステップS32:Yes)、指定された時間帯は、可視化情報表示部2132を介して、処理データ格納部215に出力される。処理データ格納部215は、各条件による場面検出結果を基に、複数の条件の中から、指定された時間帯に検出場面がある条件を求め、求めた条件による場面検出結果と、これらの条件間の関係性とを可視化情報表示部2132に出力する(ステップS33)。 When the time zone in the video information is specified (step S32: Yes), the specified time zone is output to the processing data storage unit 215 via the visualization information display unit 2132. The processing data storage unit 215 obtains a condition in which a detection scene exists in a specified time zone from a plurality of conditions based on the scene detection result according to each condition, and the scene detection result according to the obtained condition and these conditions. The relationship between the two is output to the visualization information display unit 2132 (step S33).
 可視化情報表示部2132は、処理データ格納部215による場面検出結果について、各条件に該当する画面の時間帯のうち指定された時間帯を強調表示した可視化情報を出力するとともに、各条件間の関係性を示す可視化情報を出力する(ステップS34)。なお、複数の時間帯が指定されている場合、可視化情報表示部2132は、時間帯ごとに色を分けるなど、各時間帯で区別できるように表示すればよい。 The visualization information display unit 2132 outputs visualization information highlighting the specified time zone among the time zones of the screen corresponding to each condition for the scene detection result by the processing data storage unit 215, and outputs the visualization information highlighting the relationship between the conditions. Visualization information indicating the sex is output (step S34). When a plurality of time zones are specified, the visualization information display unit 2132 may display the visualization information display unit 2132 so that it can be distinguished in each time zone, such as by separating the colors for each time zone.
 そして、条件設定部2131は、指定された時間帯の場面を検出する条件と各検出条件間の関係性とに対する登録指示または更新指示を受けたか否かを判定する(ステップS35)。表示処理システム210は、条件と各検出条件間の関係性とに対する登録指示または更新指示を受けていない場合(ステップS35:No)、処理を終了する。条件と各検出条件間の関係性とに対する登録指示または更新指示を受けた場合(ステップS35:Yes)、条件設定部2131は、登録指示または更新指示に応じて、条件格納部14が記憶する条件と各条件間の関係性とを登録または更新する(ステップS36)。 Then, the condition setting unit 2131 determines whether or not a registration instruction or an update instruction has been received for the condition for detecting the scene in the designated time zone and the relationship between the detection conditions (step S35). When the display processing system 210 has not received the registration instruction or the update instruction for the condition and the relationship between the detection conditions (step S35: No), the display processing system 210 ends the processing. When a registration instruction or an update instruction for a condition and a relationship between each detection condition is received (step S35: Yes), the condition setting unit 2131 receives a registration instruction or an update instruction, and the condition storage unit 14 stores the condition. And the relationship between each condition are registered or updated (step S36).
[実施の形態2の効果]
 このように、実施の形態2では、検索ユーザが、検出したい場面の範囲として、映像情報内の任意の時間帯を指定すると、複数の条件の中から、指定された時間帯に検出場面がある条件を求め、求めた条件ごとに、各条件が該当する時間帯のうち、指定された時間帯を強調表示した可視化情報を、各条件間の関係性とともに表示する。
[Effect of Embodiment 2]
As described above, in the second embodiment, when the search user specifies an arbitrary time zone in the video information as the range of the scene to be detected, the detection scene is in the designated time zone from a plurality of conditions. The conditions are obtained, and for each of the obtained conditions, the visualization information highlighting the specified time zone among the time zones to which each condition applies is displayed together with the relationship between the conditions.
 このため、実施の形態2によれば、検索ユーザは、所望の時間帯の場面を検出するためにどのような条件が有効かを判断したい場合には、この時間帯における場面を検出できる条件と、条件間の関係性を参照することで、より具体的な条件またはより限定的な条件から、条件の有用性を判別することができる。例えば、検索ユーザは、該当する場面で共通して検出される条件を抽出して新たな条件をテンプレート化することができる。 Therefore, according to the second embodiment, when the search user wants to determine what kind of condition is effective for detecting the scene in the desired time zone, the condition is such that the scene in this time zone can be detected. , The usefulness of a condition can be determined from a more specific condition or a more limited condition by referring to the relationship between the conditions. For example, the search user can extract the conditions commonly detected in the corresponding scene and create a new condition as a template.
 したがって、実施の形態2によれば、検索ユーザは、映像情報からの場面検出の際に、所望の時間帯における場面を検出するための条件として、有用な条件を判別できるため、所望の場面を適切に検出することが可能になる。 Therefore, according to the second embodiment, the search user can determine a useful condition as a condition for detecting a scene in a desired time zone when detecting a scene from the video information, so that the desired scene can be determined. It will be possible to detect it properly.
[実施の形態2の変形例]
 なお、表示処理システム210が、所望の時間帯の場面を検出可能である条件を判別して、所望の時間帯の場面するための新規条件や条件間の関係性を自動的に登録または更新してもよい。図13は、図11に示す表示処理システム210が実行する条件及び条件間の関係性の設定処理の他の処理手順を示すフローチャートである。
[Modified Example of Embodiment 2]
The display processing system 210 determines the conditions under which the scene in the desired time zone can be detected, and automatically registers or updates new conditions and relationships between the conditions for the scene in the desired time zone. You may. FIG. 13 is a flowchart showing a condition executed by the display processing system 210 shown in FIG. 11 and another processing procedure for setting the relationship between the conditions.
 図13に示すステップS41及びステップS42は、図12に示すステップS31及びステップS32とそれぞれ同じ処理である。映像情報内の時間帯の指定を受けた場合(ステップS42:Yes)、処理データ格納部215は、各条件による場面検出結果を基に、複数の条件の中から、指定された時間帯に検出場面がある条件を求め、求めた条件による場面検出結果と、これらの条件間の関係性を示す関係性情報とを条件設定部2131(第2の設定部)に出力する(ステップS43)。 Step S41 and step S42 shown in FIG. 13 are the same processes as steps S31 and S32 shown in FIG. 12, respectively. When the time zone specified in the video information is specified (step S42: Yes), the processing data storage unit 215 detects the time zone specified from among a plurality of conditions based on the scene detection results according to each condition. A scene is obtained for a certain condition, and the scene detection result according to the obtained condition and the relationship information indicating the relationship between these conditions are output to the condition setting unit 2131 (second setting unit) (step S43).
 条件設定部2131は、条件設定部2131から出力された条件と条件設定部2131によって求められた条件間の関係性を示す関係性情報とを基に、指定された時間帯の場面を検出する条件を判別する(ステップS44)。条件設定部2131は、指定された時間帯で共通して検出される条件を抽出して新たな条件を作成する。また、条件設定部2131は、指定された時間帯に検出場面がある条件のうち、上位の条件から順に下位の条件を探索し、下位よりも、より具体的な上位の条件を組み合わせた条件を、指定された時間帯の場面検出のための条件として新たに作成する。 The condition setting unit 2131 detects a scene in a designated time zone based on the condition output from the condition setting unit 2131 and the relationship information indicating the relationship between the conditions obtained by the condition setting unit 2131. Is determined (step S44). The condition setting unit 2131 extracts the conditions commonly detected in the designated time zone and creates a new condition. Further, the condition setting unit 2131 searches for the lower condition in order from the upper condition among the conditions having the detection scene in the designated time zone, and sets the condition by combining the more specific upper condition than the lower condition. , Create a new condition for scene detection in the specified time zone.
 条件設定部2131は、ステップS44における判別結果を基に、条件格納部14が記憶する条件と各条件間の関係性とを登録または更新する(ステップS45)。条件設定部2131は、新規条件を条件格納部14に登録するとともに、新規条件と既存の条件との関係性も登録または更新する。 The condition setting unit 2131 registers or updates the condition stored in the condition storage unit 14 and the relationship between each condition based on the determination result in step S44 (step S45). The condition setting unit 2131 registers the new condition in the condition storage unit 14, and also registers or updates the relationship between the new condition and the existing condition.
[実施の形態のシステム構成について]
 表示処理システム10,210の各構成要素は機能概念的なものであり、必ずしも物理的に図示のように構成されていることを要しない。すなわち、表示処理システム10,210の機能の分散及び統合の具体的形態は図示のものに限られず、その全部または一部を、各種の負荷や使用状況などに応じて、任意の単位で機能的または物理的に分散または統合して構成することができる。
[About the system configuration of the embodiment]
Each component of the display processing systems 10 and 210 is a functional concept and does not necessarily have to be physically configured as shown in the figure. That is, the specific form of the distribution and integration of the functions of the display processing systems 10 and 210 is not limited to the one shown in the figure, and all or part of them may be functional in any unit according to various loads and usage conditions. Or it can be physically distributed or integrated.
 また、表示処理システム10,210においておこなわれる各処理は、全部または任意の一部が、CPU、GPU(Graphics Processing Unit)、及び、CPU、GPUにより解析実行されるプログラムにて実現されてもよい。また、表示処理システム10においておこなわれる各処理は、ワイヤードロジックによるハードウェアとして実現されてもよい。 Further, each process performed in the display processing systems 10 and 210 may be realized by a CPU, a GPU (Graphics Processing Unit), and a program analyzed and executed by the CPU and the GPU, in whole or in any part thereof. .. Further, each process performed in the display processing system 10 may be realized as hardware by wired logic.
 また、実施の形態において説明した各処理のうち、自動的におこなわれるものとして説明した処理の全部または一部を手動的に行うこともできる。もしくは、手動的におこなわれるものとして説明した処理の全部または一部を公知の方法で自動的に行うこともできる。この他、上述及び図示の処理手順、制御手順、具体的名称、各種のデータやパラメータを含む情報については、特記する場合を除いて適宜変更することができる。 It is also possible to manually perform all or part of the processes described as being automatically performed among the processes described in the embodiment. Alternatively, all or part of the process described as being performed manually can be automatically performed by a known method. In addition, the above-mentioned and illustrated processing procedures, control procedures, specific names, and information including various data and parameters can be appropriately changed unless otherwise specified.
[プログラム]
 図14は、プログラムが実行されることにより、表示処理システム10,210が実現されるコンピュータの一例を示す図である。コンピュータ1000は、例えば、メモリ1010、CPU1020を有する。また、コンピュータ1000は、ハードディスクドライブインタフェース1030、ディスクドライブインタフェース1040、シリアルポートインタフェース1050、ビデオアダプタ1060、ネットワークインタフェース1070を有する。これらの各部は、バス1080によって接続される。
[program]
FIG. 14 is a diagram showing an example of a computer in which the display processing systems 10 and 210 are realized by executing a program. The computer 1000 has, for example, a memory 1010 and a CPU 1020. The computer 1000 also has a hard disk drive interface 1030, a disk drive interface 1040, a serial port interface 1050, a video adapter 1060, and a network interface 1070. Each of these parts is connected by a bus 1080.
 メモリ1010は、ROM1011及びRAM1012を含む。ROM1011は、例えば、BIOS(Basic Input Output System)等のブートプログラムを記憶する。ハードディスクドライブインタフェース1030は、ハードディスクドライブ1090に接続される。ディスクドライブインタフェース1040は、ディスクドライブ1100に接続される。例えば磁気ディスクや光ディスク等の着脱可能な記憶媒体が、ディスクドライブ1100に挿入される。シリアルポートインタフェース1050は、例えばマウス1110、キーボード1120に接続される。ビデオアダプタ1060は、例えばディスプレイ1130に接続される。 Memory 1010 includes ROM 1011 and RAM 1012. The ROM 1011 stores, for example, a boot program such as a BIOS (Basic Input Output System). The hard disk drive interface 1030 is connected to the hard disk drive 1090. The disk drive interface 1040 is connected to the disk drive 1100. For example, a removable storage medium such as a magnetic disk or an optical disk is inserted into the disk drive 1100. The serial port interface 1050 is connected to, for example, a mouse 1110 and a keyboard 1120. The video adapter 1060 is connected to, for example, the display 1130.
 ハードディスクドライブ1090は、例えば、OS(Operating System)1091、アプリケーションプログラム1092、プログラムモジュール1093、プログラムデータ1094を記憶する。すなわち、表示処理システム10,210の各処理を規定するプログラムは、コンピュータ1000により実行可能なコードが記述されたプログラムモジュール1093として実装される。プログラムモジュール1093は、例えばハードディスクドライブ1090に記憶される。例えば、表示処理システム10,210における機能構成と同様の処理を実行するためのプログラムモジュール1093が、ハードディスクドライブ1090に記憶される。なお、ハードディスクドライブ1090は、SSD(Solid State Drive)により代替されてもよい。 The hard disk drive 1090 stores, for example, an OS (Operating System) 1091, an application program 1092, a program module 1093, and program data 1094. That is, the program that defines each processing of the display processing systems 10 and 210 is implemented as a program module 1093 in which the code that can be executed by the computer 1000 is described. The program module 1093 is stored in, for example, the hard disk drive 1090. For example, the program module 1093 for executing the same processing as the functional configuration in the display processing systems 10 and 210 is stored in the hard disk drive 1090. The hard disk drive 1090 may be replaced by an SSD (Solid State Drive).
 また、上述した実施の形態の処理で用いられる設定データは、プログラムデータ1094として、例えばメモリ1010やハードディスクドライブ1090に記憶される。そして、CPU1020が、メモリ1010やハードディスクドライブ1090に記憶されたプログラムモジュール1093やプログラムデータ1094を必要に応じてRAM1012に読み出して実行する。 Further, the setting data used in the processing of the above-described embodiment is stored as program data 1094 in, for example, a memory 1010 or a hard disk drive 1090. Then, the CPU 1020 reads the program module 1093 and the program data 1094 stored in the memory 1010 and the hard disk drive 1090 into the RAM 1012 and executes them as needed.
 なお、プログラムモジュール1093やプログラムデータ1094は、ハードディスクドライブ1090に記憶される場合に限らず、例えば着脱可能な記憶媒体に記憶され、ディスクドライブ1100等を介してCPU1020によって読み出されてもよい。あるいは、プログラムモジュール1093及びプログラムデータ1094は、ネットワーク(LAN(Local Area Network)、WAN(Wide Area Network)等)を介して接続された他のコンピュータに記憶されてもよい。そして、プログラムモジュール1093及びプログラムデータ1094は、他のコンピュータから、ネットワークインタフェース1070を介してCPU1020によって読み出されてもよい。 The program module 1093 and the program data 1094 are not limited to those stored in the hard disk drive 1090, but may be stored in, for example, a removable storage medium and read by the CPU 1020 via the disk drive 1100 or the like. Alternatively, the program module 1093 and the program data 1094 may be stored in another computer connected via a network (LAN (Local Area Network), WAN (Wide Area Network), etc.). Then, the program module 1093 and the program data 1094 may be read from another computer by the CPU 1020 via the network interface 1070.
 以上、本発明者によってなされた発明を適用した実施の形態について説明したが、本実施の形態による本発明の開示の一部をなす記述及び図面により本発明は限定されることはない。すなわち、本実施の形態に基づいて当業者等によりなされる他の実施の形態、実施例及び運用技術等は全て本発明の範疇に含まれる。 Although the embodiment to which the invention made by the present inventor is applied has been described above, the present invention is not limited by the description and the drawings which form a part of the disclosure of the present invention according to the present embodiment. That is, other embodiments, examples, operational techniques, and the like made by those skilled in the art based on the present embodiment are all included in the scope of the present invention.
 10,210 表示処理システム
 11 生データ格納部
 12 データ処理部
 13 UI部
 14 条件格納部
 15,215 処理データ格納部
 131,2131 条件設定部
 132,2132 可視化情報表示部
 133,2133 映像表示部
10,210 Display processing system 11 Raw data storage unit 12 Data processing unit 13 UI unit 14 Condition storage unit 15,215 Processing data storage unit 131,2131 Condition setting unit 132,2132 Visualization information display unit 133,2133 Video display unit

Claims (5)

  1.  映像情報に対応付けられた時系列情報を含むパラメータに基づいて前記映像情報内の場面を検出するための複数の条件と、指向性を持つ各条件間の関係性とを記憶する第1の記憶部と、
     任意の条件に関し、各条件間の関係性に基づいて、前記複数の条件の中から、前記任意の条件に関連する条件を求め、求めた条件に該当する場面を検出する検出部と、
     を有することを特徴とする処理システム。
    A first storage that stores a plurality of conditions for detecting a scene in the video information based on a parameter including time-series information associated with the video information, and a relationship between each condition having directivity. Department and
    With respect to any condition, a detection unit that obtains a condition related to the arbitrary condition from the plurality of conditions based on the relationship between the conditions and detects a scene corresponding to the obtained condition.
    A processing system characterized by having.
  2.  検索対象の映像情報の入力を受け付ける第1の入力部と、
     前記第1の記憶部が記憶する各条件による前記検索対象の映像情報内の場面検出結果を記憶する第2の記憶部と、
     前記映像情報の場面検出のために指定された第1の条件の入力を受けつける第2の入力部と、
     前記場面検出結果に関する可視化情報を出力する第1の表示部と、
     を有し、
     前記検出部は、前記複数の条件の中から、前記第1の条件と前記第1の条件に関連する第2の条件とを求め、前記各条件による場面検出結果のうち、前記第1の条件に該当する場面検出結果と前記第2の条件に該当する場面検出結果とを前記第1の表示部に出力するとともに、前記第1の条件及び前記第2の条件の関係性を示す情報を前記第1の表示部に出力し、
     前記第1の表示部は、前記第1の条件に該当する場面の時間帯を示す第1の可視化情報と、前記第2の条件に該当する場面の時間帯を示す第2の可視化情報と、前記第1の条件及び前記第2の条件との関係性を示す第3の可視化情報とを出力することを特徴とする請求項1に記載の処理システム。
    The first input unit that accepts the input of the video information to be searched,
    A second storage unit that stores a scene detection result in the video information to be searched according to each condition stored in the first storage unit, and a second storage unit.
    A second input unit that receives input of the first condition specified for scene detection of the video information, and a second input unit.
    A first display unit that outputs visualization information regarding the scene detection result, and
    Have,
    The detection unit obtains the first condition and the second condition related to the first condition from the plurality of conditions, and among the scene detection results under each of the conditions, the first condition. The scene detection result corresponding to the above and the scene detection result corresponding to the second condition are output to the first display unit, and information indicating the relationship between the first condition and the second condition is output. Output to the first display
    The first display unit includes first visualization information indicating a time zone of a scene corresponding to the first condition, and second visualization information indicating a time zone of a scene corresponding to the second condition. The processing system according to claim 1, wherein the first condition and the third visualization information indicating the relationship with the second condition are output.
  3.  前記第1の記憶部が記憶する各条件による前記映像情報内の場面検出結果を記憶する第3の記憶部と、
     前記映像情報内の任意の時間帯を指定する時間帯指定情報の入力を受け付ける第3の入力部と、
     前記時間帯指定情報に応じた可視化情報を出力する第2の表示部と、
     前記第1の記憶部が記憶する条件と各条件間の関係性とを設定する第1の設定部と、
     さらに有し、
     前記検出部は、各条件による場面検出結果を基に、前記複数の条件の中から、前記時間帯指定情報において指定された時間帯に検出場面がある条件を求め、求めた条件による場面検出結果と、前記求めた条件間の関係性を示す関係性情報とを前記第2の表示部に出力し、
     前記第2の表示部は、前記検出部によって求められた条件ごとに、各条件に該当する場面の時間帯のうち前記時間帯指定情報において指定された時間帯を強調表示した可視化情報を出力するとともに、各条件間の関係性を示す可視化情報を出力し、
     前記第1の設定部は、前記時間帯指定情報において指定された時間帯の場面を検出する条件と各検出条件間の関係性とに対する登録指示または更新指示に応じて、前記第1の記憶部が記憶する条件と各条件間の関係性とを登録または更新することを特徴とする請求項1に記載の処理システム。
    A third storage unit that stores the scene detection result in the video information according to each condition stored in the first storage unit, and a third storage unit.
    A third input unit that accepts input of time zone designation information that specifies an arbitrary time zone in the video information, and
    A second display unit that outputs visualization information according to the time zone designation information, and
    A first setting unit that sets a condition stored in the first storage unit and a relationship between each condition, and a first setting unit.
    Have more
    Based on the scene detection results under each condition, the detection unit obtains a condition in which the detection scene exists in the time zone specified in the time zone designation information from the plurality of conditions, and the scene detection result under the obtained conditions. And the relationship information indicating the relationship between the obtained conditions are output to the second display unit.
    The second display unit outputs visualization information highlighting the time zone specified in the time zone designation information among the time zones of the scenes corresponding to each condition for each condition obtained by the detection unit. At the same time, it outputs visualization information showing the relationship between each condition.
    The first setting unit is the first storage unit in response to a registration instruction or an update instruction for a condition for detecting a scene in a time zone specified in the time zone designation information and a relationship between each detection condition. The processing system according to claim 1, wherein the conditions stored in the computer and the relationship between the conditions are registered or updated.
  4.  前記第1の記憶部が記憶する各条件による前記映像情報内の場面検出結果を記憶する第3の記憶部と、
     前記映像情報内の任意の時間帯を指定する時間帯指定情報の入力を受け付ける第4の入力部と、
     前記第1の記憶部が記憶する条件と各条件間の関係性とを設定する第2の設定部と、
     をさらに有し、
     前記検出部は、各条件による場面検出結果を基に、前記複数の条件の中から、前記時間帯指定情報において指定された時間帯に検出場面がある条件を求め、求めた条件による場面検出結果と、前記求めた条件間の関係性を示す関係性情報とを前記第2の設定部に出力し、
     前記第2の設定部は、前記検出部によって求められた条件と前記検出部によって求められた条件間の関係性を示す関係性情報とを基に、前記時間帯指定情報において指定された時間帯の場面を検出する条件を判別し、判別結果を基に前記第1の記憶部が記憶する条件と各条件間の関係性とを登録または更新することを特徴とする請求項1に記載の処理システム。
    A third storage unit that stores the scene detection result in the video information according to each condition stored in the first storage unit, and a third storage unit.
    A fourth input unit that accepts input of time zone designation information that specifies an arbitrary time zone in the video information, and
    A second setting unit that sets the conditions stored in the first storage unit and the relationship between the conditions, and
    Have more
    Based on the scene detection results under each condition, the detection unit obtains a condition in which the detection scene exists in the time zone specified in the time zone designation information from the plurality of conditions, and the scene detection result according to the obtained condition. And the relationship information indicating the relationship between the obtained conditions are output to the second setting unit.
    The second setting unit is a time zone designated in the time zone designation information based on the relationship information indicating the relationship between the condition obtained by the detection unit and the condition obtained by the detection unit. The process according to claim 1, wherein the condition for detecting the scene is determined, and the condition stored in the first storage unit and the relationship between the conditions are registered or updated based on the determination result. system.
  5.  処理システムが実行する処理方法であって、
     前記処理システムは、映像情報に対応付けられた時系列情報を含むパラメータに基づいて前記映像情報内の場面を検出するための複数の条件と、指向性を持つ各条件間の関係性とを記憶する記憶部を有し、
     任意の条件に関し、各条件間の関係性に基づいて、前記複数の条件の中から、前記任意の条件に関連する条件を検出する工程、
     を含んだことを特徴とする処理方法。
    It is a processing method executed by the processing system.
    The processing system stores a plurality of conditions for detecting a scene in the video information based on a parameter including time-series information associated with the video information, and a relationship between each condition having directivity. Has a storage unit
    A step of detecting a condition related to the arbitrary condition from the plurality of conditions based on the relationship between the arbitrary conditions.
    A processing method characterized by including.
PCT/JP2020/025705 2020-06-30 2020-06-30 Processing system and processing method WO2022003836A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/JP2020/025705 WO2022003836A1 (en) 2020-06-30 2020-06-30 Processing system and processing method
JP2022532896A JP7439927B2 (en) 2020-06-30 2020-06-30 Treatment system and treatment method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2020/025705 WO2022003836A1 (en) 2020-06-30 2020-06-30 Processing system and processing method

Publications (1)

Publication Number Publication Date
WO2022003836A1 true WO2022003836A1 (en) 2022-01-06

Family

ID=79314971

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2020/025705 WO2022003836A1 (en) 2020-06-30 2020-06-30 Processing system and processing method

Country Status (2)

Country Link
JP (1) JP7439927B2 (en)
WO (1) WO2022003836A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH07253986A (en) * 1994-03-15 1995-10-03 Sharp Corp Retrieval device for sound and moving picture
JP2008131170A (en) * 2006-11-17 2008-06-05 Nippon Hoso Kyokai <Nhk> Knowledge metadata generation device, digest generation device, knowledge metadata generation program, and digest generation program
JP2010026981A (en) * 2008-07-24 2010-02-04 Nippon Hoso Kyokai <Nhk> Particular scene learning system and program
JP2013037595A (en) * 2011-08-10 2013-02-21 Casio Comput Co Ltd Image searching device, animation searching device, image searching method, animation searching method and program

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH07253986A (en) * 1994-03-15 1995-10-03 Sharp Corp Retrieval device for sound and moving picture
JP2008131170A (en) * 2006-11-17 2008-06-05 Nippon Hoso Kyokai <Nhk> Knowledge metadata generation device, digest generation device, knowledge metadata generation program, and digest generation program
JP2010026981A (en) * 2008-07-24 2010-02-04 Nippon Hoso Kyokai <Nhk> Particular scene learning system and program
JP2013037595A (en) * 2011-08-10 2013-02-21 Casio Comput Co Ltd Image searching device, animation searching device, image searching method, animation searching method and program

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
MIYAZAKI, MASARU ET AL.: "Development of Digest Production System using Baseball Knowledgebase", PROCEEDINGS OF 6TH FORUM ON INFORMATION TECHNOLOGY (FIT2007); TOKYO, JAPAN; SEPTEMBER 12-14, 2017, vol. 16, no. 2, 22 August 2007 (2007-08-22) - 14 September 2017 (2017-09-14), pages 445 - 446, XP009534173 *

Also Published As

Publication number Publication date
JP7439927B2 (en) 2024-02-28
JPWO2022003836A1 (en) 2022-01-06

Similar Documents

Publication Publication Date Title
US7805451B2 (en) Ontology-integration-position specifying apparatus, ontology-integration supporting method, and computer program product
RU2449357C2 (en) Ranking diagram
US20110271224A1 (en) Pinning of tabs in tab groups
JP7509144B2 (en) Information processing device, information processing method, and program
JP2008097175A (en) Electronic file retrieving device
CN106249982B (en) Display control method, display control device, and control program
US10546013B2 (en) File management system facilitating the organization of content by using visualizations of the organizational state augmented by a configurable workflow
CN103477317B (en) Content display processing device, content display processing method and integrated circuit
JP2008108200A (en) Information extraction device, method, program and storage medium
JP2008102594A (en) Content search method and search device
US20110179003A1 (en) System for Sharing Emotion Data and Method of Sharing Emotion Data Using the Same
US20120317117A1 (en) Information Visualization System
US9245351B2 (en) Color evaluation apparatus, color evaluation method and computer program
US20090150429A1 (en) Data management apparatus and data processing method
JP5096850B2 (en) Search result display method, search result display program, and search result display device
WO2022003836A1 (en) Processing system and processing method
JP2006217046A (en) Video index image generator and generation program
US20150199419A1 (en) Information processing apparatus, information processing method and non-transitory computer readable medium
US9189497B2 (en) Information processing apparatus, control method therefor, and program for classifying a plurality of data
JP5302529B2 (en) Information processing apparatus, information processing method, program, and recording medium
JP2016122413A (en) Image processing apparatus, control method of image processing apparatus, and program
WO2022003842A1 (en) Processing system and processing method
US20240303887A1 (en) Systems and methods for identifying a design template matching a media item
EP4414445A1 (en) Cell image analysis system
JP5617535B2 (en) Information processing apparatus, information processing apparatus processing method, and program.

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20943632

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2022532896

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20943632

Country of ref document: EP

Kind code of ref document: A1