WO2020048324A1 - Video abstract generation method and apparatus, and electronic device and readable storage medium - Google Patents
Video abstract generation method and apparatus, and electronic device and readable storage medium Download PDFInfo
- Publication number
- WO2020048324A1 WO2020048324A1 PCT/CN2019/102073 CN2019102073W WO2020048324A1 WO 2020048324 A1 WO2020048324 A1 WO 2020048324A1 CN 2019102073 W CN2019102073 W CN 2019102073W WO 2020048324 A1 WO2020048324 A1 WO 2020048324A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- target
- picture
- target picture
- video
- time point
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/73—Querying
Definitions
- the present application relates to video surveillance technology, and in particular, to a method, a device, an electronic device, and a readable storage medium for generating a video summary.
- Video surveillance system as an important technical means of social security management, is increasingly used and deployed in the field of social security maintenance.
- the number of deployed surveillance devices increases and the scope of deployment expands, the amount of data stored in video recording data also increases. If you want to find out the specific target (person or vehicle, etc.) in which time period and place from the video recording data, it is often necessary to manually search and search a large amount of video recording data, which takes a long time and may be neglected. There are efficiency bottlenecks and incomplete risks in video positioning and integrated display.
- the present application provides a method and a device for generating a video summary.
- a method for generating a video digest including: receiving a target search request, where the target search request carries feature information of a target to be searched; and the search matches the feature information of the target to be searched A first target picture; generating a video summary according to the first target picture and the acquisition time corresponding to the first target picture.
- the method before searching for the first target picture that matches the feature information of the target to be searched, the method further includes: obtaining target picture information in video data of the video source device, where the target picture information includes the target picture, Collection time of the target picture and attribute information of the target picture; saving the target picture information to a picture information database.
- acquiring the target picture information in the video data of the video source device includes receiving the target picture information sent by the video source device.
- acquiring the target picture information in the video data of the video source device includes: receiving the target picture and the acquisition time of the target picture sent by the video source device; The picture is modeled, and attribute information of the target picture is extracted.
- acquiring the target picture information in the video data of the video source device includes: receiving the target picture sent by the video source device, a collection time of the target picture, and a first picture of the target picture.
- An attribute information modeling the target picture and extracting the second attribute information of the target picture; determining the target picture according to the first attribute information of the target picture and the second attribute information of the target picture Attribute information.
- acquiring the target picture information in the video data of the video source device includes: performing target detection on the video data provided by the video source device to obtain the video data in the video data.
- the target picture and the acquisition time of the target picture modeling the target picture and extracting attribute information of the target picture.
- the characteristic information of the target to be searched includes attribute information of the target to be searched; and searching for the first target picture that matches the characteristic information of the target to be searched includes: according to the attribute information of the target to be searched Searching the picture information database for the matching first target picture.
- the feature information of the target to be searched includes a target picture to be searched; searching for the first target picture that matches the characteristic information of the target to be searched includes: modeling the target picture to be searched, And extracting attribute information of the target picture to be searched for; and searching the picture information database for the matching first target picture according to the attribute information of the target picture to be searched for.
- feature information of the target to be searched includes target picture to be searched and third attribute information of the target picture to be searched; and searching for the first target picture that matches the characteristic information of the target to be searched includes: : Model the target picture to be searched, and extract fourth attribute information of the target picture to be searched; determine the attribute of the target picture to be searched according to the third attribute information and the fourth attribute information Information; searching for the matching first target picture in the picture information database according to the attribute information of the target picture to be searched.
- the target search request further carries a search time range; searching for the first target picture that matches the characteristic information of the target to be searched includes: comparing the picture according to the search time range
- the target pictures in the information database are filtered to obtain a second target picture whose acquisition time is within the range of the search time period; according to the feature information of the target to be searched, a matching target is searched from the second target picture.
- the first target picture is described.
- the target search request further carries a search channel number
- the target picture information further includes a channel number of the target picture; searching for the first target matching the feature information of the target to be searched
- the picture includes: filtering the target picture in the picture information database according to the search channel number to obtain a third target picture whose channel number is consistent with the search channel number; Feature information, searching the first target picture for matching from the third target picture.
- the target picture is a face picture
- the target search request is a face search request
- the target picture is a vehicle picture
- the target search request is a vehicle search request
- generating the video summary according to the first target picture and the acquisition time corresponding to the first target picture includes: sorting the first target picture in the order of the acquisition time from morning to night; Generating the video summary by using the first target picture.
- generating the video summary according to the first target picture and the acquisition time corresponding to the first target picture includes: for each first target picture, determining a target video clip corresponding to the first target picture, The target video clip is video data between the n-th second before the acquisition time corresponding to the first target picture and the m-th second after the acquisition time of the first target picture; according to each target The video clip generates a video summary.
- determining the target video clip corresponding to the first target picture includes: when there are multiple first target pictures with the same acquisition time, for the multiple first target pictures, For any first target picture in the target picture, determine a start time point and an end time point of the video clip corresponding to the first target picture, where the start time point is the acquisition time corresponding to the first target picture The nth second before, the end time point is the mth second after the acquisition time corresponding to the first target picture; searching whether the starting time point exists in the recording data of the video data channel to which the first target picture belongs If there are I frames at the end time point; if there are I frames at the start time point and I frames at the end time point, the remaining first target pictures in the plurality of first target pictures are discarded And determine the video clip corresponding to the first target picture as the target video clip.
- the method further includes: if there is no I frame at the start time point, increasing the start time point of the video clip corresponding to the first target picture by x seconds to obtain a new start point Point in time, and repeat the above search step until the new starting time point I frame is searched in the recording data of the video data channel to which the first target picture belongs, or the first target picture corresponds to The new start time point of the video clip is the same as the acquisition time; if there is no I frame at the end time point, the end time of the video clip corresponding to the first target picture The point is reduced by x seconds to obtain a new end time point, and the above search step is repeated until the I frame of the new end time point is searched in the recording data of the video data channel to which the first target picture belongs, or , The new end time point of the video clip corresponding to the first target picture is the same as the acquisition time; among the corresponding video clips in the plurality of first target pictures, the longest video clip is selected as the The target footage; discard the remaining length of the plurality of video clips
- generating the video summary according to each of the target video clips includes: filtering the target video clips according to a start time point and an end time point of each target video clip to remove time-repeated video data. Generating the video summary according to the filtered target video clip.
- filtering the target video clips according to the start time point and the end time point of each target video clip includes: matching the target video clips according to the start time point of each target video clip. Sorting each of the target video clips; for an adjacent first target video clip and a second target video clip, when the end time point of the first target video clip is greater than or equal to the start time of the second target video clip At the point of time, if the first target video clip and the second target video clip belong to the same video data channel, the first target video clip and the second target video clip are merged.
- the start time point is the start time point of the first target video clip
- the end time point is the end time point of the second target video clip
- the first target video clip and the second target video clip are Belonging to different video data channels, using the end time point of the first target video clip as the start time point of the second target video clip, or Use the start time point of the second target video clip as the end time point of the first target video clip; wherein the start time point of the first target video clip is smaller than the second The starting time point of the target video clip.
- a video digest generating device including: a receiving unit configured to receive a target search request, where the target search request carries characteristic information of a target to be searched; a search unit configured to search A first target picture matching the characteristic information of the target to be searched; a processing unit, configured to generate a video digest according to the first target picture and the acquisition time corresponding to the first target picture.
- the apparatus further includes: an obtaining unit, configured to obtain target picture information in the video data of the video source device, where the target picture information includes the target picture, the acquisition time of the target picture, and attribute information of the target picture; A unit, configured to save the target picture information to a picture information database.
- the obtaining unit is specifically configured to receive the target picture information sent by the video source device.
- the obtaining unit is specifically configured to receive the target picture and the collection time of the target picture sent by the video source device; model the target picture and extract attributes of the target picture information.
- the obtaining unit is specifically configured to receive the target picture, the acquisition time of the target picture, and first attribute information of the target picture sent by the video source device; and model the target picture, And extracting the second attribute information of the target picture; determining the attribute information of the target picture according to the first attribute information of the target picture and the second attribute information of the target picture.
- the obtaining unit is specifically configured to perform target detection on the video data provided by the video source device to obtain the target picture and the acquisition time of the target picture in the video data;
- the target picture is modeled, and attribute information of the target picture is extracted.
- the feature information of the target to be searched includes attribute information of the target to be searched; and the search unit is specifically configured to search the picture information database for the matching first part according to the attribute information of the target to be searched.
- a target picture is specifically configured to search the picture information database for the matching first part according to the attribute information of the target to be searched.
- the feature information of the target to be searched includes a target picture to be searched; the search unit is specifically configured to model the target picture to be searched and extract attribute information of the target picture to be searched; The attribute information of the target picture to be searched is searched in the picture information database for the matching first target picture.
- the feature information of the target to be searched includes target picture to be searched and third attribute information of the target picture to be searched;
- the search unit is specifically configured to model the target picture to be searched, and Extracting fourth attribute information of the target picture to be searched; determining attribute information of the target picture to be searched according to the third attribute information and the fourth attribute information; and according to the attribute information of the target picture to be searched in
- the picture information database searches for a matching first target picture.
- the target search request also carries a search time range; the search unit is specifically configured to filter the target pictures in the picture information database according to the search time range to obtain a collection time A second target picture within the range of the search period; and searching for a matching first target picture from the second target picture according to the feature information of the target to be searched.
- the target search request also carries a search channel number
- the target picture information further includes a channel number of the target picture
- the search unit is specifically configured to match the search channel number with the search channel number.
- the target pictures in the picture information database are filtered to obtain a third target picture with the same channel number as the search channel number; according to the feature information of the target to be searched, a matching first picture is searched from the third target picture.
- the target picture is a face picture
- the target search request is a face search request
- the target picture is a vehicle picture
- the target search request is a vehicle search request
- the processing unit is specifically configured to sort the first target picture in the order of the collection time from morning to night; and generate the video summary according to the sorted first target picture.
- the processing unit is specifically configured to determine, for each first target picture, a target video clip corresponding to the first target picture, where the target video clip is the acquisition time corresponding to the first target picture Recording data between the nth second before and the mth second after the acquisition time of the first target picture; generating a video summary according to each of the target video clips.
- the processing unit is specifically configured to, when there are multiple first target pictures with the same acquisition time, for any first target picture in the multiple first target pictures, determine that the first target picture corresponds to The start time point and end time point of the video clip, where the start time point is the n-th second before the acquisition time corresponding to the first target picture, and the end time point is the corresponding time point of the first target picture The m-th second after the acquisition time; searching whether the I frame at the start time point and the I frame at the end time point exist in the recording data of the video data channel to which the first target picture belongs; if the The I frame at the start time point and the I frame at the end time point, the remaining first target pictures in the plurality of first target pictures are discarded, and the video clip corresponding to the first target picture is determined as the target video Fragment.
- the processing unit is further configured to, if there is no I frame at the start time point, increase the start time point of the video clip corresponding to the first target picture by x seconds to obtain a new Start time point, and repeat the above search steps until the new start time point I frame is searched in the video data of the video data channel to which the first target picture belongs, or the first target
- the new start time point of the video clip corresponding to the picture is the same as the acquisition time; if there is no I frame at the end time point, the video clip corresponding to the first target picture is Decrease the end time point by x seconds to obtain a new end time point, and repeat the above search steps until the I frame of the new end time point is searched in the recording data of the video data channel to which the first target picture belongs.
- the new end time point of the video clip corresponding to the first target picture is the same as the acquisition time; among the corresponding video clips in the plurality of first target pictures, a video with the longest duration is selected Examples of the target video clip; discard the remaining length of the plurality of video clips when the first target image corresponding to the first target picture.
- the processing unit is specifically configured to filter the target video clip according to a start time point and an end time point of each of the target video clips to remove time-repeated video data; and according to the filtered target video
- the snippet generates the video summary.
- the processing unit is specifically configured to sort each of the target video clips according to the start time point of each of the target video clips; for an adjacent first target video clip and a second target video clip Clip, when the end time point of the first target video clip is greater than or equal to the start time point of the second target video clip, if the first target video clip and the second target video clip belong to the same video data Channel, the first target video clip and the second target video clip are merged, the start time point of the merged video clip is the start time point of the first target video clip, and the end time point is The end time point of the second target video clip; if the first target video clip and the second target video clip belong to different video data channels, the end time point of the first target video clip is used as The start time point of the second target video clip, or using the start time point of the second target video clip as the first target video clip Beam time point; wherein the first target segment of video start time point of the second target is less than the start time point of the video clip.
- an electronic device including a processor, a communication interface, a memory, and a communication bus, wherein the processor, the communication interface, and the memory pass through the processor.
- the communication bus completes communication with each other; a memory is configured to store a computer program; and a processor is configured to implement the steps of the above-mentioned video abstract generation method when the computer program stored on the memory is executed.
- a computer-readable storage medium characterized in that a computer program is stored in the computer-readable storage medium, and when the computer program is executed by a processor, the video digest generation is implemented. Method steps.
- the video abstract generating method in the embodiment of the present application receives a target search request, searches for a first target picture that matches the characteristic information of the target to be searched carried in the target search request, and according to each first target picture and its corresponding acquisition time Generate a video summary of the search target. This improves the efficiency and accuracy of locating targets in video recordings. On the basis of removing video recordings that do not match the target to be searched, the consistency of target video tracking is guaranteed.
- FIG. 1 is a schematic structural diagram of a video digest generating system according to an exemplary embodiment of the present application
- FIG. 2 is a schematic flowchart of a video abstract generating method according to an exemplary embodiment of the present application
- FIG. 3 is a schematic flowchart of generating a picture information database according to an exemplary embodiment of the present application.
- FIG. 4 is a schematic flowchart of a repeated picture filtering process according to an exemplary embodiment of the present application.
- FIG. 5 is a schematic flowchart of a video digest generating system according to another exemplary embodiment of the present application.
- FIG. 6 is a schematic flowchart of extracting attributes of a face picture according to an exemplary embodiment of the present application.
- FIG. 7 is a schematic flowchart of extracting attributes of a face picture according to another exemplary embodiment of the present application.
- FIG. 8 is a schematic flowchart of a video digest generating system according to still another exemplary embodiment of the present application.
- FIG. 9 is a schematic flowchart of extracting a picture attribute of a vehicle according to an exemplary embodiment of the present application.
- FIG. 10 is a schematic flowchart of extracting a picture attribute of a vehicle according to another exemplary embodiment of the present application.
- FIG. 11 is a schematic flowchart illustrating a video digest generating system according to another exemplary embodiment of the present application.
- FIG. 12 is a schematic structural diagram of a video digest generating apparatus according to an exemplary embodiment of the present application.
- FIG. 13 is a schematic structural diagram of a video digest generating apparatus according to another exemplary embodiment of the present application.
- Fig. 14 is a schematic diagram of a hardware structure of an electronic device according to an exemplary embodiment of the present application.
- the video digest generating system may include a video source device 110 and a search device 120.
- the video source device 110 may provide video data, and the video data may include real-time video data or video recording data (referred to as recording data herein).
- the search device 120 may receive a target search request, and search for a target picture (referred to herein as a text image) in the video data of the video source device 110 that matches the feature information of the target to be searched according to the characteristic information of the target to be searched carried in the target search request Is the first target picture), and a video summary of the target to be searched is generated according to the searched first target picture.
- a target picture referred to herein as a text image
- the video source device 110 may be a front-end video capture device (such as IPC (Internet Protocol Camera)) or a video recording storage device (such as NVR (Network Video Recorder, Network Hard Disk Video Recorder)). ));
- the search device 120 may be an NVR (with a target search function) or a device deployed in a video surveillance system dedicated to target search.
- the video source device 110 is an NVR, the video source device 110 and the search device 120 may be the same device.
- one video source device 110 may provide video data for multiple search devices 120, and one search device 120 may also obtain video data from multiple video source devices 110 (one-to-one is taken as an example in the figure).
- the video source device 110 is IPC, one video source device 110 can correspond to one video data channel; when the video source device 110 is a video recording storage device such as an NVR, one video source device 110 can provide multiple video data channels.
- Video data (recording data).
- FIG. 2 is a schematic flowchart of a video abstract generating method according to an embodiment of the present application.
- the video abstract generating method may be applied to a search device (taking an NVR as an example).
- the video summary generating method may include the following steps.
- Step S200 Receive a target search request, where the target search request carries characteristic information of a target to be searched.
- the target may include, but is not limited to, a human face, a vehicle, or a license plate.
- the feature information of the target to be searched may include, but is not limited to, a face picture, structured information of the face (such as whether to smile, whether to wear glasses, gender, age range, etc.), etc. Or more.
- the characteristic information of the target to be searched may be, but is not limited to, one or more of a vehicle picture, characteristic information of the vehicle (such as color, type, logo, brand, etc.).
- the feature information of the target to be searched may include, but is not limited to, one or more of a license plate picture, characteristic information (such as color, position, license plate number, etc.) of the license plate.
- the target may further include a human body, an animal, and the like.
- the feature information of the target to be searched may include, but is not limited to, one or more of a human body picture, human body characteristic information (such as height, weight, gender, skin color, clothing, etc.).
- the characteristic information of the target to be searched may include, but is not limited to, one or more of animal pictures, animal characteristic information (such as type, hair color, size, etc.) and the like.
- Step S210 Search for a first target picture that matches the feature information of the target to be searched.
- the search device when receiving a target search request, may search for a first target picture in the video data of the video source device that matches the characteristic information of the target to be searched for according to the characteristic information of the target to be searched.
- the first target picture and the acquisition time of the first target picture may be used as the first target picture information that matches the feature information of the target to be searched.
- the search result when the search device searches for the first target picture that matches the feature information of the target to be searched, the search result may be one or more first target pictures, or the search result may be empty.
- the search result is empty, that is, when no first target picture matching the feature information of the target to be searched is found, the search device may determine that the target search fails, and return a search failure response message.
- the collection time of the target picture may be the time when the front-end video capture device collects (eg, captures) the target picture, or the time when the target picture appears in the video image collected by the front-end video capture device.
- the acquisition time of the target picture may be carried in the target picture (for example, displaying the acquisition time of the target picture at a specific position in the target picture (such as the lower left corner or the lower right corner, etc.))
- the acquisition time of the target picture may be independent of the target picture, and its specific implementation is not described herein.
- Step S220 Generate a video summary according to the first target picture and the acquisition time corresponding to the first target picture.
- the search device after the search device determines the first target picture that matches the feature information of the target to be searched, it can generate a video summary of the target to be searched according to the first target picture and its corresponding acquisition time.
- generating the video summary according to the first target picture and the acquisition time corresponding to the first target picture includes: sorting multiple first target pictures in the order of the acquisition time from morning to night; Video summary from multiple first target images.
- the search device may directly sort the first target picture in the order of the collection time from morning to night to generate a video summary.
- the search device may directly according to the sorted first target pictures, Generate a video summary.
- generating a video summary according to the first target picture and the acquisition time corresponding to the first target picture includes: for each first target picture, determining a target video corresponding to the first target picture A segment, where the target video segment is video data from the n-th second before the acquisition time corresponding to the first target picture to the m-th second after the acquisition time of the first target picture; generated according to each target video segment Video summary.
- the search device when the search device searches for multiple first target pictures that match the feature information of the target to be searched, for each first target picture, the search device may Recording data between the nth second and the mth second after the acquisition time of the first target picture is determined as a target video clip corresponding to the first target picture, so as to obtain a video recording that matches the characteristic information of the target to be searched.
- the collection time of the first target picture is the time when the first target picture appears in the video data extracted from the first target picture (that is, the time when the first target picture is collected by the video acquisition device).
- the start time point of the target video clip corresponding to the first target picture (n-th second before the acquisition time of the first target picture) and the end time point (m-th after the acquisition time of the first target picture) Seconds) can be pre-configured in the search device or can be carried in the target search request (can be set by the user according to actual needs or default values are used).
- n and m are non-negative numbers.
- the search device when the video data provided by the video source device for the search device includes video data of multiple video data channels, when the search device determines the target video clip corresponding to the first target picture, it may first determine the first target picture belongs to Video data channel, and determine a target video clip corresponding to the first target picture in the video data of the video data channel. That is, among the recording data of the video data channel, the recording data between the nth second before and the mth second after the acquisition time of the first target picture is determined as a target video clip corresponding to the first target picture.
- the first target picture information may further include a channel number of the first target picture, that is, a channel number of a video data channel to which the first target picture belongs.
- the search device may fuse each target video clip, that is, according to the starting of each target video clip Sort and splice each target video clip at the start time and / or end time to generate a video summary of the target to be searched.
- a first target picture that matches the feature information of the target to be searched carried in the target search request is searched, and the n-th second to the m-th second after the acquisition time corresponding to each first target picture
- the recorded video data is determined as the target video clip corresponding to the first target picture, and further, the video summary of the target to be searched is obtained by fusing each target video clip. Therefore, the efficiency and accuracy of locating targets in video recordings are improved, and the consistency of target video tracking is ensured on the basis of removing video recordings that do not match the target to be searched.
- Video summary can be generated based on each target video clip and decoded and displayed.
- the method may further include: obtaining target picture information in the video data of the video source device, where the target picture information includes The target picture, the acquisition time of the target picture, and the attribute information of the target picture; the obtained target picture information is saved to the picture information database.
- the search device may obtain the target picture information in the video data of the video source device in advance, and save the obtained target picture information to the picture information database. Further, when a target search is required, it may directly The feature information of the search target searches for a matching first target picture from the picture information database to improve target search efficiency.
- the target picture information may include, but is not limited to, the target picture, feature information of the target picture, and acquisition time of the target picture.
- the picture information library can be a specified storage space in the search device, or it can be a third-party database.
- acquiring the target picture information in the video data of the video source device may include: receiving the target picture information sent by the video source device.
- the video source device when the video source device has a target picture acquisition function (such as a target picture capture function or a target detection function) and a target picture analysis function, the video source device can directly obtain the target picture, the acquisition time of the target picture, and the target picture The target picture information such as the attribute information of the camera, and send the target picture information to the search device.
- the search device can receive the target picture information sent by the video source device.
- the video source device can capture the target picture (and record the target picture capture time (that is, the acquisition time)), and perform the target picture capture Target image analysis to extract the attribute information of the target image. Furthermore, the video source device can send the target image information such as the target image, the acquisition time of the target image, and the attribute information of the target image to a search device, such as an NVR, which is stored by the search device.
- the target picture information such as the received target picture, the acquisition time of the target picture, and the attribute information of the target picture.
- obtaining the target picture information in the video data of the video source device may include: receiving the target picture and the acquisition time of the target picture sent by the video source device; modeling the target picture and extracting The attribute information of the target picture.
- the video source device when the video source device has a target picture acquisition function (such as a target picture capture function or a target detection function), the video source device can obtain the target picture and the acquisition time of the target picture, and convert the target picture and the target picture's The acquisition time is sent to the search device.
- the search device receives the target picture sent by the video source device and the acquisition time of the target picture, it can model the target picture and extract the attribute information of the target picture. Therefore, the search device can obtain the target picture, the acquisition time of the target picture, and the attribute information of the target picture in the video data of the video source device.
- the video source device can capture the target picture (and record the target picture capture time (that is, the acquisition time)), and send the captured target picture and the target picture collection time Give search devices such as NVR.
- the search device receives the target picture, it can model the target picture and extract the attribute information of the target picture.
- the search device can then store the target picture, the acquisition time of the target picture, and the attribute information of the target picture.
- obtaining target picture information in the video data of the video source device may include: receiving the target picture information sent by the video source device, the acquisition time of the target picture, and the first attribute information of the target picture; The target picture is modeled and the second attribute information of the target picture is extracted; the attribute information of the target picture is determined according to the first attribute information of the target picture and the second attribute information of the target picture.
- the video source device when the video source device has a target picture acquisition function (such as a target picture capture function or a target detection function) and a target picture analysis function, the video source device can directly obtain the target picture, the acquisition time of the target picture, and the target picture Attribute information (referred to herein as the first attribute information of the target picture), and sends the target picture, the acquisition time of the target picture, and the first attribute information of the target picture to the search device.
- the search device receives the target picture sent by the video source device, it can model the received target picture and extract the attribute information of the target picture (referred to herein as the second attribute information of the target picture).
- the search device may compare the first attribute information of the target picture with the second attribute information of the target picture. For the attribute information that exists in the first attribute information but does not exist in the second attribute information, or the first attribute The information does not exist, but the attribute information existing in the second attribute information is added to the attribute information of the target picture; for the attribute information existing in both the first attribute information and the second attribute information, the attributes in the second attribute information are added The information is added to the attribute information of the target picture, and then the attribute information of the target picture is obtained.
- the search device may also directly use the second attribute information of the target picture as the attribute information of the target picture.
- obtaining the target picture, the acquisition time of the target picture, and the attribute information of the target picture in the video data of the video source device may include: performing target detection on the video data provided by the video source device to Obtain the target picture and the acquisition time of the target picture in the video data; model the target picture and extract the attribute information of the target picture.
- the search device may directly perform target detection on the video data provided by the video source device to The target picture in the video data and the acquisition time of the target picture are obtained; after the search device obtains the target picture in the video data, the target picture can be further modeled and attribute information of the target picture can be extracted.
- the video source device may send the acquired video data to a search device, such as an NVR.
- a search device receives the video data sent by the video source device, it can perform target detection on the received video data to obtain the target picture in the video data and the acquisition time of the target picture (the time when the target picture appears in the video data) , And model the target picture to extract the attribute information of the target picture.
- the search device can then store the target picture, the acquisition time of the target picture, and the attribute information of the target picture.
- the search device obtains the target picture, the acquisition time of the target picture, and the attribute information of the target picture in the video data of the video source device
- the obtained target picture, the acquisition time of the target picture, and the target may be stored. Picture attribute information.
- the above-mentioned saving of the target picture information to the picture information database may include: for any target picture information of any video data channel, judging whether a picture information database is stored with the target
- the picture information includes other target pictures of the same target, where the other target picture information and the target picture information are from the same video data channel, and the difference between the acquisition time included in the other target picture information and the acquisition time included in the target picture information
- the value is less than the preset time threshold; when there is no other target picture information that contains the same target as the target picture information in the picture information database, the target picture information is saved to the picture information database.
- video data in a video data channel (such as video data obtained by an IPC) is usually video data in a fixed scene. Therefore, for any target picture information of any video data channel (including the search device obtaining target detection of the video data provided by the video source device, or the search device receiving the video source device), the search device is Before the picture information is saved in the picture information database, you can determine whether other target pictures that meet the following conditions are stored in the picture information database: the same target as the target picture; from the same video data channel; the acquisition time of other target pictures and the target picture The difference between the acquisition times included in the information is less than a preset threshold.
- the search device may save the target picture information to the picture information database.
- the search device determines that there is other target picture information in the picture information database that meets the above conditions, it refuses to save the target picture information to the picture information database, such as discarding the target picture information directly to reduce redundant picture storage. Therefore, the workload of searching the first target picture in the picture information database can be reduced, and the search efficiency can be improved.
- the feature information of the target to be searched for may include target picture to be searched and / or attribute information of the target picture to be searched.
- the feature information of the target to be searched includes the target picture to be searched; correspondingly, searching for the first target picture that matches the characteristic information of the target to be searched includes: modeling the target picture to be searched, And extracting the attribute information of the target picture to be searched; according to the attribute information of the target picture to be searched, a matching first target picture is searched in the picture information database.
- the search device may provide a target search function, according to the target picture to be searched carried in the received target search request, and search for a matching target picture in a map search mode.
- the search device may provide a target search request interface, and the target search request interface may include an input or / and selection area of a target picture to be searched, and a user enters or / and selects a target picture to be searched in the target search request interface, and Submit a target search request.
- the search device When the search device receives the target search request, it models the target picture to be searched and extracts the attribute information of the target picture to be searched. Furthermore, the search device can query the stored target picture information according to the attribute information of the target picture to be searched, and The target picture information corresponding to the attribute information of the target picture that matches the attribute information of the target picture to be searched is determined as the first target picture information.
- the search device receives the target search request, it can search for a matching first target face picture in the manner of map search, that is, the target
- the face picture carried in the search request is modeled to obtain a feature model of the face picture, and further, the similarity between the video data of the video source device and the feature model of the face picture is greater than or equal to a preset similarity threshold
- the face image of is determined as the first target face picture, and the acquisition time of the first target face picture and the first target face picture is used as the first target face picture information.
- the similarity threshold may be configured in a search device in advance, or may be carried in a target search request (can be set by a user according to actual needs or a default value is used).
- the feature information of the target to be searched includes target picture to be searched and third attribute information of the target picture to be searched; correspondingly, searching for a first target picture that matches the feature information of the target to be searched, Including: modeling the target picture to be searched and extracting the fourth attribute information of the target picture to be searched; determining the attribute of the target picture to be searched according to the third attribute information of the target picture to be searched and the fourth attribute information of the target picture to be searched Information; searching for a matching first target picture in the picture information database according to the attribute information of the target picture to be searched.
- the search device may further model the search target picture and extract the attributes of the target picture to be searched Information (herein referred to as the fourth attribute information of the target picture to be searched).
- the search device After the search device obtains the fourth attribute information of the target picture to be searched, it may determine the attribute information of the target picture to be searched according to the third attribute information of the target picture to be searched and the fourth attribute information of the target picture to be searched.
- the search device may compare the third attribute information of the target picture to be searched with the fourth attribute information of the target picture to be searched. For the attribute information that exists in the third attribute information but does not exist in the fourth attribute information, or the third attribute The information does not exist, but the attribute information existing in the fourth attribute information is added to the attribute information of the target picture to be searched; for the attribute information existing in both the third attribute information and the fourth attribute information, the fourth attribute information is added The attribute information of is added to the attribute information of the target picture to be searched, and then the attribute information of the target picture to be searched is obtained.
- the search device When the search device obtains the attribute information of the target picture to be searched, it can query the target picture information in the picture information database according to the attribute information of the target picture to be searched, and will correspond to the attribute information of the target picture that matches the attribute information of the target picture to be searched.
- the target picture information is determined as the first target picture information.
- the feature information of the target to be searched includes attribute information of the target to be searched.
- searching for a first target picture that matches the feature information of the target to be searched includes: searching for the matching first target picture in the picture information database according to the attribute information of the target to be searched.
- the feature information of the target to be searched may also be attribute information of the target picture to be searched.
- the search device receives the target search request, it may directly according to the attribute information of the target picture to be searched carried in the picture information database. Search for a matching first target picture.
- the target search when performing the target search, it may also carry a specific filtering attribute to instruct the search device to firstly After filtering the target image information, the first target image search is further performed.
- the specific filtering attribute may include, but is not limited to, a search time range or / and a search channel number.
- the target search request also carries a search time range; searching for the first target picture matching the characteristic information of the target to be searched for may include: comparing the picture information according to the search time range.
- the target pictures in the library are filtered to obtain a second target picture whose acquisition time is within the search time range; and a matching first target picture is searched from the second target picture according to the feature information of the target to be searched.
- the search device may first filter the target pictures in the picture information database according to the search time range carried in the target search request to obtain the acquisition time in the search time period.
- the second target picture within range. For example, assuming that the search time range is [t1, t2] (t2> t1), the second target picture refers to a target picture whose acquisition time t satisfies t1 ⁇ t ⁇ t2.
- the search device when it obtains the second target picture, it may search for a matching first target picture from the second target picture according to the feature information of the target to be searched.
- the target search request also carries a search channel number
- the target picture information further includes a channel number of the target picture (that is, a channel number of a video data channel to which the target picture belongs);
- the first target picture with matching feature information includes: filtering the target picture information in the picture information database according to the search channel number to obtain a third target picture with the same channel number as the search channel number; and according to the feature information of the target to be searched, The third target picture is searched for a matching first target picture.
- the search device may first filter the target pictures in the picture information database according to the search channel number and the channel number information of each target picture in the picture information database to obtain the channel number and The third target picture with the same channel number is searched, and then the first target picture is searched in the third target picture.
- the first target picture may be deduplicated.
- determining a target video clip corresponding to the first target picture may include: when there are multiple first target pictures with the same acquisition time, for Any one of the plurality of first target pictures determines a start time point and an end time point of a video clip corresponding to the first target picture, where the start time point corresponds to the first target picture The n-th second before the acquisition time, the end time point is the m-th second after the acquisition time corresponding to the first target picture; searching whether the start time exists in the recording data of the video data channel to which the first target picture belongs Point I frame, and whether there is an I frame at the end time; if there are I frames, discard the remaining first target pictures in the multiple first target pictures, and determine the video clip corresponding to the first target picture For the target video clip.
- the search device may determine the start of the video segment corresponding to each first target picture in the multiple first target pictures according to a preset policy. Time point and end time point. For any first target picture among the plurality of first target pictures, the start time point of the video clip corresponding to the first target picture is the n-th second and end time point before the acquisition time corresponding to the first target picture M seconds after the acquisition time corresponding to the first target picture.
- the search device After the search device determines the start time point and the end time point, it can search whether the I frame at the start time point exists in the recording data of the video data channel to which the first target picture belongs (that is, whether the video data channel exists The key frame at the start time point), and whether there is an I frame at the end time point (that is, whether there is a key frame at the end time point in the video data channel). If there are I frames, the search device may directly determine the video clip corresponding to the first target picture as the target video clip, and discard the remaining first target pictures among the plurality of first target pictures.
- the search device may increase the start time point of the video clip corresponding to the first target picture by x seconds, and search for the existence of an I frame at the new start time point.
- the I frame at the time point (the updated start time point), or the start time point of the video clip corresponding to the first target picture is the same as the acquisition time of the first target picture.
- x is a positive number.
- the search device may reduce the end time point of the video clip corresponding to the first target picture by x seconds, and Search whether there is an I frame at the end time point. If it does not exist, reduce the end time point of the video segment corresponding to the first target picture by x seconds again, and search for whether there is an I frame at the end time point, and repeat the operation. Until the I frame at the end time point (the updated end time point) is searched in the recording data of the video data channel to which the first target picture belongs, or the start time of the video clip corresponding to the first target picture The points are collected at the same time as the first target picture.
- the start time point may be 0 minutes and 58 seconds, and the end time point may be 1 minute and 3 seconds. It is searched whether the video data of the video data channel to which the first target picture belongs has an I frame at 0:58. If there is no I frame, you can increase the start time point by 1 second to obtain a new start time, 0 minutes and 59 seconds, and continue to search for whether there is an I frame in the video data of 0 minutes and 59 seconds. If the I frame still does not exist, increase the start time point by 1 second to obtain a new start time point, 1 minute and 0 seconds.
- Time is used as the start time. Similarly, it is searched whether the I frame exists in the video data of the video data channel to which the first target picture belongs in 1 minute and 3 seconds. If there is no I-frame, the end time point can be reduced by 1 second to obtain a new end time, 1 minute and 2 seconds, and it is continued to search whether there is an I-frame in the recording data of 1 minute and 2 seconds. If there is an I frame, the recording data between 1 minute and 0 seconds and 1 minute and 2 seconds is used as the recording segment corresponding to the first target picture.
- the I frame still does not exist, continue to reduce the end time point by 1 second to obtain a new end time, and continue to judge. Until the new end time point is 1 minute and 0 seconds. In this case, the video data of 1 minute and 0 seconds is used as the video clip corresponding to the first target picture.
- the search device After the search device determines the start time and end time of the video clips corresponding to the multiple first target pictures in the foregoing manner, it can determine that the duration of the corresponding video clip in the multiple first target pictures is the longest (end First target picture with the largest difference between the time point and the starting time point), determine the video clip corresponding to the first target picture as the target video clip, and discard the remaining first targets in the plurality of first target pictures image.
- a first one when the number of the first target pictures with the longest duration of the corresponding video clips in the plurality of first target pictures is greater than 1, a first one may be selected according to a preset strategy.
- a target picture, and a video clip corresponding to the selected first target picture is determined as the target video clip. For example, the first target picture with the earliest start time, the first target picture with the latest end time, or a random selection may be selected.
- deduplication processing of the multiple first target pictures may also be implemented manually.
- the search device may display the plurality of first target pictures in a specified interface, the user selects the first target picture to be retained, and discards the remaining first target pictures with the same acquisition time.
- generating a video summary according to each target recording segment may include: filtering the target recording segment according to a start time point and an end time point of each target recording segment to remove duplicate time Video data; video summary is generated based on the filtered target video clips.
- the target video clip can be filtered according to the start time point and the end time point of each target video clip to remove the video with a repeated time. data.
- filtering the target video clips according to the start time point and the end time point of each target video clip includes: sorting the target video clips according to the start time point of each target video clip; The first target video clip and the second target video clip, when the end time point of the first target video clip is greater than or equal to the start time point of the second target video clip, if the first target video clip and the second target video clip belong to the same
- the first target video clip and the second target video clip are combined, and the start time point of the merged video clip is the start time point of the first target video clip and the end time point is the second target video clip.
- the end time point of the second target video clip if the first target video clip and the second target video clip belong to different video data channels, the end time point of the first target video clip is used as the start time point of the second target video clip, or The start time point of the two target clips is used as the end time point of the first target video clip.
- the search device may sort each target video clip according to a start time point of each target video clip.
- the search device may sort each target video clip by using a bubble sorting method according to the start time point of each target video clip.
- the start time point of the first target video clip is less than the start time point of the second target video clip
- corresponding processing may be performed according to the video data channel to which the first target video clip and the second target video clip belong.
- the search device may merge the first target video clip and the second target video clip, and the start time point of the merged video clip is the first target.
- the start time point of the video clip, and the end time is the end time point of the second target video clip.
- the search device may use the end time point of the first target video clip as the start time point of the second target video clip, or the search device
- the start time point of the second target clip can be used as the end time point of the first target video clip, and its specific implementation will be described below in conjunction with a specific example.
- the search device is an NVR
- the video source device is an IPC
- the target is a face
- the target search is a face picture search.
- the NVR can use the structured video analysis technology on the GPU (Graphics Processing Unit) to perform face recognition on the real-time video stream transmitted by the IPC to obtain the face picture information in the real-time video stream .
- the NVR can receive face picture information from the IPC end.
- the face picture information may include, but is not limited to, a face picture, structured information of the face picture, acquisition time of the face picture, channel number of the face picture, and channel name of the face picture (for recording address information) Wait.
- the NVR maintains a buffer for each video data channel (referred to as the channel) between the IPC and the NVR, which is used to store the face picture information received within 3 seconds.
- the NVR deletes the face picture from the buffer.
- the NVR obtains new face picture information (including face picture information received directly from IPC or face picture information obtained through face detection) from the channel
- the face picture The information is compared with the face picture information in the buffer corresponding to the channel. If the face picture information of the same face exists in the buffer, the newly obtained face picture information is discarded, or the The face picture information is added to the buffer and saved to the picture information database.
- the NVR can obtain the newly obtained The face image information of the overwrites the earliest stored face image information in the buffer, and saves the newly obtained face image information to the picture information database.
- FIG. 3 The flow diagram of the face recognition and picture storage performed by the NVR can be shown in FIG. 3.
- the configurable recording fusion parameters may include, but are not limited to, the following parameters.
- Channel selection if single channel is selected, it means single-channel face image information search and video fusion; if the specified multiple channels are checked, face image information search and multi-channel video fusion are represented on multiple channels.
- Search time range Represents the time range information of the face image to be searched for.
- Similarity threshold If the threshold is 90%, the similarity result is 90% and above.
- Duplicate picture filtering mode There are two modes: automatic and manual. The processing methods of different modes are explained in point 3.
- the NVR receives a face search request, and the face search request carries a target face picture and a video fusion parameter.
- Image search NVR models the target face picture, searches and compares the face picture information in the picture information database that matches the channel number and search time range carried in the face search request, and calculates the similarity.
- List face pictures whose similarity with the target face picture is greater than or equal to the similarity threshold.
- the acquisition time corresponding to the face picture and the face picture constitutes face picture information (hereinafter referred to as the first face picture information).
- the first face picture information may be recorded in the first face picture information list.
- Face pictures are filtered repeatedly. For any one of the first face pictures, subtracting n seconds from the acquisition time of the first face picture to obtain the start of the video segment corresponding to the first face picture The time point x, and the acquisition time of the first face picture is added to m seconds to obtain the end time point y of the video clip corresponding to the first face picture.
- the NVR searches the video recordings of this channel respectively for the I frames at time points x and y.
- Duplicate picture filtering mode includes manual filtering mode or automatic filtering mode.
- the manual filtering mode is: the multiple first face pictures are output on the specified interface, the user checks the first face pictures to be retained, and the other first face pictures at the same acquisition time are discarded.
- the automatic filtering mode is: among the plurality of first face pictures, the first face picture corresponding to the longest video clip is retained, and other first face pictures with the same acquisition time are discarded. When there are multiple first face pictures corresponding to the longest video clip in the face picture, one of them is retained, and the rest are discarded.
- the first face image information after filtering the repeated pictures is formed into a final first face image information list.
- the schematic flowchart of the NVR generating the final first face picture information list can be shown in FIG. 4.
- each element includes the following information: the channel number of the channel to which the first face picture belongs, and the start of the video clip corresponding to the first face picture The time point (the starting time point of the I frame exists) and the end time point of the video clip corresponding to the first face picture (the end time point of the I frame exists); among them, each element is from early to late according to the starting time point Sort in order.
- Duplicate video clip filtering For the two adjacent elements of the video clip time period element set (hereinafter referred to as the first element and the second element respectively), it is assumed that the start time and the end time of the first element and the second element [A, B] and [C, D], where A ⁇ C ⁇ B. If the first element and the second element include the same channel number, the first element and the second element are combined into one element, and the start time point of the element is A and the end time point is D. If the channel numbers included in the first element and the second element are different, it is determined whether there is an I frame at time point B in the video recording of the channel to which the second element belongs.
- the start time point of the second element is updated to B, that is, the start time point and end time point of the first element and the second element are [A, B] and [B , D]. If there is no I frame at time point B, the start time point of the first element is updated to C, that is, the start time point and end time point of the first element and the second element are [A, C] and [ C, D].
- corresponding video data is obtained from the video video of the corresponding channel, and a video summary is generated based on the obtained video data.
- the video summary can be downloaded by the user, and the user can export a set of video clip elements used to generate the video summary.
- a target search request by receiving a target search request, a first target picture that matches the feature information of the target to be searched carried in the target search request is searched, and according to each first target picture and the corresponding collection of each first target picture
- the video summary of the target to be searched is generated in time, which improves the efficiency and accuracy of locating the target in the video recording.
- the consistency of the target video tracking is guaranteed.
- FIG. 5 is a schematic flowchart of a video abstract generating method according to another embodiment of the present application.
- the video abstract generating method may be applied to a retrieval device.
- the video digest generation method is directed to a face search request.
- the video summary generating method may include the following steps.
- Step S500 Acquire and store a face picture in the video data of the video source device, a collection time of the face picture, and attribute information of the face picture.
- the retrieval device may acquire and store the face picture information in the video data of the video source device.
- the face picture information may include, but is not limited to, a face picture, a collection time of the face picture, and attribute information of the face picture.
- the attribute information of the face picture may include, but is not limited to, one or more of the following: facial expression (such as whether to smile), whether to wear glasses, gender, age range, and ethnicity.
- the specific implementation method for obtaining the face picture, the collection time of the face picture, and the attribute information of the face picture in the video data of the video source device is similar to the foregoing method of obtaining the target picture information.
- the target picture information is replaced with the adult face picture information. Yes, I wo n’t repeat them here.
- the retrieval device obtains the face picture in the video data of the video source device, the acquisition time of the face picture, and the attribute information of the face picture, the retrieved face picture and the collection of the face picture can be stored. Time and attribute information of face pictures.
- Storing the face picture in the video data of the video source device, the collection time of the face picture, and the attribute information of the face picture may include: storing the face picture; recording the storage location of the face picture in the face picture information table, Collection time of face pictures and attribute information of face pictures.
- the retrieval device After the retrieval device obtains the face picture, the collection time of the face picture, and the attribute information of the face picture, it can store the obtained face picture, and store the storage location of the face picture, the collection time of the face picture, and the person.
- the face picture attribute information is recorded in the face picture information table, and its format can be shown in Table 1:
- Face image location information Face image collection time Face image attribute information Location information of face picture 1 Acquisition time of face picture 1 Attribute information of face picture 1 Location information of face picture 2 Acquisition time of face picture 2 Attribute information of face picture 2 ... ... ... ...
- the position information of the face picture may be a position offset and a length of the face picture in a storage space (such as a hard disk).
- the above implementation manner of storing the face picture, the collection time of the face picture, and the attribute information of the face picture in the video data of the video source device is only to store the face picture, the collection time of the face picture, and the face picture in this application.
- a specific example of the attribute information is not a limitation on the protection scope of the present application, that is, in the embodiments of the present application, the face picture, the time of collecting the face picture, and the person in the video data of the video source device may also be stored in other ways. Attribute information of face pictures.
- the face picture, the acquisition time of the face picture, and the attribute information of the face picture can be stored in the same database (that is, the face picture is directly stored in the database in a binary form).
- the face picture, the collection time of the face picture, and the attribute information of the face picture can be stored in the same data table. At this time, there is no need to additionally record the storage location of the face picture.
- the face picture can still be stored first to obtain the storage location of the face picture, but when the storage location of the face picture, the collection time of the face picture, and the attribute information of the face picture are no longer stored in the data Tables are stored in the form of other forms, such as a tree structure or a file. The specific implementation is not described here.
- Step S510 When a face retrieval request is received, a first target face picture matching the face retrieval filter condition is determined according to the face retrieval filter condition carried in the face retrieval request.
- the retrieval device can provide a face retrieval function, which retrieves matching face pictures according to the face retrieval filter conditions carried in the received face retrieval request, and at the same time can obtain the collection time of the matching face pictures.
- the retrieval device may provide a face retrieval request interface
- the face retrieval request interface may include a face retrieval filter condition input area or / and a face retrieval filter condition option, which is entered by the user in the face retrieval request interface or / And select a face search filter and submit a face search request.
- the face retrieval filter condition is attribute information of a face picture to be retrieved (this may be referred to as third attribute information of the face picture to be retrieved), which may include, but is not limited to, facial expressions of the face to be retrieved , Whether or not you wear glasses, gender, and age.
- the face retrieval filtering condition may include a face picture to be retrieved and third attribute information of the face picture to be retrieved.
- the retrieval device When the retrieval device receives a face retrieval request, it can obtain the face retrieval filter conditions carried in the face retrieval request, and query the stored face pictures, the collection time of the face pictures, and the person according to the face retrieval filter conditions. Attribute information of a face picture, and determine a face picture corresponding to the attribute information of a face picture matching a face retrieval filter condition as a face picture matching a face retrieval filter condition (referred to herein as a first target Face picture).
- the retrieval device may The search filter conditions query the attribute information of the face picture in the face picture information table to obtain the face picture information entry that matches the face search filter condition, and obtain the face picture in the face picture information entry Storage location (that is, the storage location of the first target face picture) and the acquisition time of the first target face picture.
- the retrieval device may obtain the first target face picture from the specified storage space according to the storage location of the first target face picture.
- the comparison of the face retrieval filter condition and the attribute information of the face picture recorded in the face picture information table may include: : Model the face picture to be retrieved and extract the fourth attribute information of the face picture to be retrieved; determine the person to be retrieved based on the third attribute information of the face picture to be retrieved and the fourth attribute information of the face picture to be retrieved Attribute information of the face; compare the attribute information of the face picture to be retrieved with the attribute information of the face picture recorded in the face picture information table.
- the retrieval device can model the face image to be retrieved and extract the attribute information of the face image to be retrieved (in this article Called the fourth attribute information of the face picture to be retrieved).
- the retrieval device After the retrieval device obtains the fourth attribute information, it can determine the attribute information of the face to be retrieved according to the third attribute information and the fourth attribute information.
- the retrieval device may compare the third attribute information and the fourth attribute information of the face picture to be retrieved. For the attribute information that exists in the third attribute information but does not exist in the fourth attribute information, or does not exist in the third attribute information , But the attribute information existing in the fourth attribute information is added to the attribute information of the face picture to be retrieved; for the attribute information existing in both the third attribute information and the fourth attribute information, the attribute information in the fourth attribute information is added The attribute information of the face picture to be retrieved is added to obtain the attribute information of the face picture to be retrieved.
- the retrieval device When the retrieval device obtains the attribute information of the face picture to be retrieved, it can query the stored face picture, the acquisition time of the face picture, and the attribute information of the face picture according to the attribute information of the face picture to be retrieved, and compare it with the attribute information of the face picture to be retrieved.
- the face picture corresponding to the attribute information of the face picture whose attribute information matches the face picture is determined as the first target face picture that matches the face retrieval filter condition.
- the retrieval device may directly query the database based on the attribute information of the face picture to be retrieved.
- the entry where the attribute information of the matching face picture is located, and the face picture information is obtained from the queried entry.
- Step S520 Generate a video summary according to the first target face picture and the collection time of the first target face picture, and play back the video summary.
- the search device may use the first target face picture and the first target face.
- the collection time of the picture generates a video summary with the retrieved face, and plays back the video summary of the detected face.
- the specific implementation method of this step is similar to the above step S220, except that the acquisition time corresponding to the first target picture and the first target picture is replaced with the acquisition time of the first target face picture and the first target face picture, and is not repeated here. To repeat.
- the video source device is IPC
- the retrieval device is NVR.
- the NVR is loaded with a smart chip with a smart analysis function.
- the video digest generation scheme implementation process is as follows.
- the IPC captures a face picture, and transmits the captured face picture and the acquisition time of the face picture to the NVR.
- the IPC can capture a face picture, and transmit the captured face picture and the acquisition time of the face picture (that is, the capture time of the face picture) to the NVR.
- the IPC will also transmit the real-time video stream to the NVR, and the NVR saves the video recording according to a preset policy.
- the NVR extracts feature values in the face picture, models the face picture according to the feature values of the face picture, and extracts attribute information of the face picture.
- the NVR when it receives the face picture transmitted by the IPC, it can intelligently analyze the face picture through a smart chip.
- the smart chip can use the algorithm library to extract the feature values of the face in the face picture, and use the algorithm library to model the face picture according to the extracted feature values and extract the attribute information of the face picture.
- the flow chart of the NVR obtaining the face picture information can be shown in FIG. 6.
- the NVR performs target detection on a video recording or a real-time video stream to obtain a face picture and a collection time of the face picture.
- the NVR can perform target detection on the video recording or real-time video stream through the smart chip to obtain the face picture and the acquisition time of the face picture in the video recording or real-time video stream (that is, the face picture is in the video data) Time of occurrence).
- the NVR extracts feature values in the face picture, models the face picture according to the feature values of the face picture, and extracts attribute information of the face picture.
- the NVR when it receives the face picture transmitted by the IPC, it can intelligently analyze the face picture through a smart chip.
- the smart chip can use the algorithm library to extract the feature values of the face in the face picture, and use the algorithm library to model the face picture according to the extracted feature values and extract the attribute information of the face picture.
- the flow chart of the NVR obtaining the face picture information can be shown in FIG. 7.
- the NVR stores the face picture, the collection time of the face picture, and the attribute information of the face picture. For specific implementation, refer to the subsequent description.
- the NVR stores the face picture to obtain the storage location of the face picture.
- a database table FaceTable Face table related to the face picture information is established, wherein the main fields in the FaceTable table are: the storage location of the face picture, the collection time of the face picture, and the face picture Attribute information.
- FaceTable database table
- the NVR can also store model data of face pictures, and its specific implementation is not described here.
- the NVR records the storage location of the face picture, the collection time of the face picture, and the attribute information of the face picture in the FaceTable.
- a face retrieval request is received, and the face retrieval request carries a face retrieval filtering condition.
- the NVR can provide a face search interface.
- the face search interface includes a face search filter input area or / and options. The user can fill in or / and select a face search filter condition through the face search interface, and submit a face search request.
- the NVR queries the storage location of the matching face pictures and the collection time of the face pictures from the FaceTable table according to the face retrieval filter conditions.
- the NVR can query the FaceTable table, compare the face retrieval filter conditions with the attribute information of the face pictures recorded in the FaceTable table, and compare the recorded face attribute information with the face records in the FaceTable entries that match the face retrieval filter conditions.
- the storage location of the picture and the collection time of the face picture are determined as the storage location of the matching face picture (that is, the storage location of the first target face picture) and the collection time of the face picture (that is, the first target face picture's Acquisition time).
- the NVR obtains the first target face picture according to the storage location of the first target face picture.
- the NVR can read the first target face picture from the hard disk according to the storage position (position offset + length) of the first target face picture, so that the NVR can obtain the first target face picture and the first target face picture. Acquisition time.
- the NVR can create a VideoTable1.
- the main fields in the VideoTable1 table are: the storage location of the video (the hard disk position offset + length) and the start and end time (start time and end time) of the video data.
- a new record ie, a new entry
- the VideoTable1 table is inserted into the VideoTable1 table, recording the video storage location and the start and end time.
- the NVR determines the 5 second before the acquisition time of the first target face picture as the start time of the target video clip, and the 5th time after the acquisition time of the first target face picture The second is determined as the end time of the target video clip, and the VideoTable1 table is queried according to the start time and end time of the first target face picture to obtain the target video clip.
- Face search filter conditions determine the first target face picture that matches the face search filter conditions, and then generate a video summary based on the first target face picture and the acquisition time of the first target face picture, and perform the video summary Playback avoids the need to extract matching face pictures from video data for each face retrieval, improves the efficiency and accuracy of face retrieval, and ensures the consistency of face video tracking.
- FIG. 8 is a schematic flowchart of a video abstract generating method according to another embodiment of the present application.
- the video abstract generating method may be applied to a retrieval device.
- the video digest generation method is directed to a vehicle search request.
- the video summary generating method may include the following steps.
- Step S800 Acquire and store a vehicle picture, a collection time of the vehicle picture, and attribute information of the vehicle picture in the video data of the video source device.
- the retrieval device can acquire and store the vehicle picture information in the video data of the video source device.
- the vehicle picture information may include, but is not limited to, a vehicle picture, a collection time of the vehicle picture, and attribute information of the vehicle picture.
- the attribute information of the vehicle picture may include, but is not limited to, one or more of the following: the location of the vehicle in the vehicle picture, the location of the license plate in the vehicle picture, the license plate number, the license plate color, the country type, and the body color , Vehicle brand, model (such as a large passenger car, truck or van, etc.), whether the driver is wearing a seat belt, and whether the driver is calling.
- the specific implementation method for obtaining the vehicle picture, the collection time of the vehicle picture, and the attribute information of the vehicle picture in the video data of the video source device is similar to the foregoing method of obtaining the target picture information, and the target picture information can be replaced with the vehicle picture information. This will not be repeated here.
- the retrieval device may store the acquired vehicle picture, the acquisition time of the vehicle picture, and the vehicle picture. Attribute information.
- Storing the vehicle picture in the video data of the video source device, the collection time of the vehicle picture, and the attribute information of the vehicle picture may include: storing the vehicle picture; recording the storage location of the vehicle picture, the collection time of the vehicle picture in the vehicle picture information table, and Attribute information of the vehicle picture.
- the retrieval device After the retrieval device obtains the vehicle picture, the collection time of the vehicle picture, and the attribute information of the vehicle picture, it can store the obtained vehicle picture, and record the storage location of the vehicle picture, the collection time of the vehicle picture, and the attribute information of the vehicle picture in Vehicle picture information table, its format can be shown in Table 2:
- the location information of the vehicle picture may be a position offset and a length of the vehicle picture in a storage space (such as a hard disk).
- the foregoing implementation manner of storing the vehicle picture, the collection time of the vehicle picture, and the attribute information of the vehicle picture in the video data of the video source device is only a specific example of storing the vehicle picture, the collection time of the vehicle picture, and the attribute information of the vehicle picture in this application.
- the example is not a limitation on the protection scope of the present application, that is, in the embodiment of the present application, the vehicle picture in the video data of the video source device, the collection time of the vehicle picture, and the attribute information of the vehicle picture may also be stored.
- the vehicle picture, the collection time of the vehicle picture, and the attribute information of the vehicle picture may be stored in the same database (that is, the vehicle picture is directly stored in the database in a binary form).
- the vehicle picture, the collection time of the vehicle picture, and the attribute information of the vehicle picture can be stored in the same data table. At this time, there is no need to additionally record the storage location of the vehicle picture.
- the vehicle picture may still be stored first to obtain the storage location of the vehicle picture, but the storage location of the vehicle picture, the time when the vehicle picture was collected, and the attribute information of the vehicle picture are no longer stored in the form of a data table. Instead, it is stored in other forms, such as a tree structure or a file. The specific implementation is not described here.
- Step S810 When a vehicle retrieval request is received, a first vehicle picture that matches the vehicle picture to be retrieved is determined according to the vehicle picture to be retrieved carried in the vehicle retrieval request.
- the retrieval device may provide a vehicle retrieval function, and according to the pictures of the vehicle to be retrieved carried in the received vehicle retrieval request, retrieve the matching vehicle pictures and the collection time of the vehicle pictures in the manner of map search.
- the retrieval device may provide a vehicle retrieval request interface
- the vehicle retrieval request interface may include an input or / and selection area of a picture of the vehicle to be retrieved, and a user enters or / and selects a picture of the vehicle to be retrieved in the vehicle retrieval request interface, and Submit a vehicle search request.
- the retrieval device When the retrieval device receives a vehicle retrieval request, it models the to-be-retrieved vehicle pictures and extracts attribute information of the to-be-retrieved vehicle pictures. Furthermore, the retrieval device can query the stored vehicle pictures and vehicle pictures based on the attribute information of the to-be-retrieved vehicle pictures. Collect the time and the attribute information of the vehicle picture, and determine the vehicle picture corresponding to the attribute information of the vehicle picture that matches the attribute information of the vehicle picture to be retrieved as the vehicle picture that matches the attribute information of the vehicle picture to be retrieved (referred to herein as For the first vehicle picture).
- the retrieval device may according to the attributes of the vehicle picture to be retrieved Information query the attribute information of the vehicle picture in the vehicle picture information table to obtain the vehicle picture information entry that matches the attribute information of the vehicle picture to be retrieved, and obtain the storage location of the vehicle picture in the vehicle picture information entry (i.e. Storage location of the first vehicle picture).
- the retrieval device may acquire the first vehicle picture from the specified storage space according to the storage location of the first vehicle picture.
- vehicle picture information is only a specific example in the case of storing vehicle picture information in the form of a vehicle picture information table, and is not a limitation on the protection scope of the present application. That is, in the embodiment of the present application, Other ways to achieve vehicle image information retrieval.
- the retrieval device may directly query the matching vehicle from the database according to the attribute information of the vehicle picture to be retrieved The entry where the attribute information of the picture is located, and the vehicle picture information is obtained from the queried entry.
- Step S820 Generate a video summary according to the first vehicle picture and the collection time of the first vehicle picture, and play back the video summary.
- the retrieval device may generate a searched vehicle with the first vehicle picture and the first vehicle picture collection time. Video summary, and playback of video summary of the vehicle to be detected.
- the specific implementation method of this step is similar to the above step S220, except that the acquisition time corresponding to the first target picture and the first target picture is replaced with the acquisition time of the first vehicle picture and the first vehicle picture, and details are not described herein again.
- the video source device is IPC
- the retrieval device is NVR.
- the NVR is loaded with a smart chip with a smart analysis function.
- the video digest generation scheme implementation process is as follows.
- the IPC captures the vehicle pictures and transmits the captured vehicle pictures and the collection time of the vehicle pictures to the NVR.
- the IPC can also capture a vehicle picture, and transmit the captured vehicle picture and the acquisition time of the vehicle picture (that is, the acquisition time of the vehicle picture) to the NVR.
- the IPC will also transmit the real-time video stream to the NVR, and the NVR saves the video recording according to a preset policy.
- the NVR extracts feature values from the vehicle pictures, models the vehicle pictures according to the feature values of the vehicle pictures, and extracts attribute information of the vehicle pictures.
- the NVR when it receives the vehicle picture transmitted by the IPC, it can intelligently analyze the vehicle picture through a smart chip.
- the smart chip can use an algorithm library to extract the feature values of the vehicle from the vehicle picture, and use the algorithm library to model the vehicle picture according to the extracted feature values, and extract the attribute information of the vehicle picture.
- the schematic diagram of the NVR's process of obtaining vehicle picture information can be shown in FIG. 9.
- the NVR performs object detection on the video recording or real-time video stream to obtain the vehicle picture and the acquisition time of the vehicle picture.
- the NVR can perform target detection on the video recording or real-time video stream through the smart chip to obtain the vehicle pictures and the collection time of the vehicle pictures in the video recording or real-time video stream (that is, the time when the vehicle pictures appear in the video data ).
- the NVR extracts feature values from the vehicle pictures, models the vehicle pictures according to the feature values of the vehicle pictures, and extracts attribute information of the vehicle pictures.
- the NVR can intelligently analyze the vehicle picture through a smart chip.
- the smart chip can use an algorithm library to extract the feature values of the vehicle from the vehicle picture, and use the algorithm library to model the vehicle picture according to the extracted feature values, and extract the attribute information of the vehicle picture.
- the schematic diagram of the process for the NVR to obtain vehicle picture information can be shown in FIG. 10.
- the NVR stores the vehicle picture, the time when the vehicle picture was collected, and the attribute information of the vehicle picture. For specific implementation, see the subsequent description.
- the NVR stores the vehicle picture to obtain the storage location of the vehicle picture.
- a vehicle table vehicle table
- the main fields in the VehicleTable table are: the storage location of the vehicle picture, the collection time of the vehicle picture, and the attribute information of the vehicle picture.
- the NVR can also store model data of vehicle pictures, and its specific implementation is not described here.
- the NVR records the storage location of the vehicle picture, the collection time of the vehicle picture, and the attribute information of the vehicle picture in the VehicleTable.
- a vehicle retrieval request is received, and the vehicle retrieval request carries a picture of a vehicle to be retrieved.
- the NVR can provide a vehicle search interface, and the vehicle search interface includes a picture input or / and selection area of a vehicle to be searched.
- the user can input or / and select pictures of the vehicle to be retrieved through the vehicle retrieval interface, and submit a vehicle retrieval request.
- the NVR models the vehicle pictures to be retrieved and extracts the attribute information of the vehicle pictures to be retrieved.
- the NVR queries the storage location of the matching vehicle picture and the collection time of the vehicle picture from the VehicleTable table according to the attribute information of the vehicle picture to be retrieved.
- the NVR can query the VehicleTable table, compare the attribute information of the vehicle picture to be retrieved with the attribute information of the vehicle picture recorded in the VehicleTable table, and match the attribute information of the recorded vehicle picture with the attribute information of the vehicle picture to be retrieved.
- the storage location of the vehicle picture and the collection time of the vehicle picture are determined as the matching storage location of the vehicle picture (that is, the storage location of the first vehicle picture) and the collection time of the vehicle picture (that is, the collection time of the first vehicle picture).
- the NVR obtains the first vehicle picture according to the storage location of the first vehicle picture.
- the NVR can read the first vehicle picture from the hard disk according to the storage position (position offset + length) of the first vehicle picture, so that the NVR can obtain the first vehicle picture and the collection time of the first vehicle picture.
- the NVR can create a VideoTable2.
- the main fields in the VideoTable2 table are: the storage location of the video (the hard disk position offset + length) and the start and end time (start time and end time) of the video data. After a completed video is stored on the hard disk, a new record (ie, a new entry) is inserted into the VideoTable2 table, recording the video storage location and the start and end time.
- the NVR determines the 5 second before the acquisition time of the first vehicle picture as the start time of the target video clip, and the 5 second after the acquisition time of the first vehicle picture is determined as the target video clip. Query the VideoTable2 table according to the start time and end time of the first vehicle picture to obtain the target video clip.
- FIG. 11 is a schematic flowchart of a video abstract generating method according to another embodiment of the present application.
- the video abstract generating method may be applied to a retrieval device.
- the video digest generation method is directed to a vehicle search request.
- the video digest generating method may include the following steps.
- Step S1100 Acquire and store a vehicle picture, a collection time of the vehicle picture, and attribute information of the vehicle picture in the video data of the video source device.
- step S800 For a specific implementation method of this step, reference may be made to step S800, and details are not described herein again.
- Step S1110 When a vehicle search request is received, a second vehicle picture matching the vehicle search filter condition is determined according to the vehicle search filter condition carried in the vehicle search request.
- the retrieval device may provide a vehicle retrieval function, and retrieve a matching vehicle picture and a collection time of the vehicle picture according to a vehicle retrieval filter condition carried in the received vehicle retrieval request.
- the retrieval device may provide a vehicle search request interface
- the vehicle search request interface may include a vehicle search filter condition input area or / and a vehicle search filter condition option, and a user enters or / and selects a vehicle search in the vehicle search request interface Filter conditions and submit a vehicle search request.
- the vehicle retrieval filter condition is attribute information of a picture of the vehicle to be retrieved (this may be referred to as the third attribute information of the picture of the vehicle to be retrieved), which may include, but is not limited to, the license plate number, body color, One or more of information such as model and vehicle brand.
- the vehicle retrieval filter condition may include the third attribute information of the image of the vehicle to be retrieved and the image of the vehicle to be retrieved.
- a retrieval device When a retrieval device receives a vehicle retrieval request, it can obtain the vehicle retrieval filter conditions carried in the vehicle retrieval request, and query the stored vehicle pictures, the collection time of the vehicle pictures, and the attribute information of the vehicle pictures according to the vehicle retrieval filter conditions, and The vehicle picture corresponding to the attribute information of the vehicle picture matching the vehicle search filter condition is determined as the vehicle picture matching the vehicle search filter condition (referred to herein as the second vehicle picture).
- the retrieval device can query the attributes of the vehicle picture in the vehicle picture information table according to the vehicle search filter Information to obtain a vehicle picture information entry that matches the vehicle search filter condition, and obtain the storage location of the vehicle picture in the vehicle picture information entry (that is, the storage location of the second vehicle picture) and the collection of the second vehicle picture time.
- the retrieval device may acquire the second vehicle picture from the designated storage space according to the storage location of the second vehicle picture.
- the above comparison of the vehicle retrieval filter condition and the attribute information of the vehicle picture recorded in the vehicle picture information table May include: modeling a picture of a vehicle to be retrieved and extracting fourth attribute information of the picture of the vehicle to be retrieved; and determining, based on the third attribute information of the picture of the vehicle to be retrieved and the fourth attribute information of the picture of the vehicle to be retrieved, Attribute information; compare the attribute information of the vehicle picture to be retrieved with the attribute information of the vehicle picture recorded in the vehicle picture information table.
- the retrieval device may model the picture of the vehicle to be retrieved and extract attribute information of the picture of the vehicle to be retrieved (this article (Referred to as the fourth attribute information of the picture of the vehicle to be retrieved).
- the retrieval device After the retrieval device obtains the fourth attribute information of the picture of the vehicle to be retrieved, it can determine the attribute information of the vehicle to be retrieved according to the third attribute information of the picture of the vehicle to be retrieved and the fourth attribute information of the picture of the vehicle to be retrieved.
- the retrieval device may compare the third attribute information of the picture of the vehicle to be retrieved with the fourth attribute information of the picture of the vehicle to be retrieved. For the attribute information that exists in the third attribute information but does not exist in the fourth attribute information, or the third attribute The information does not exist, but the attribute information existing in the fourth attribute information is added to the attribute information of the vehicle picture to be retrieved; for the attribute information existing in both the third attribute information and the fourth attribute information, the fourth attribute information is added to Is added to the attribute information of the picture of the vehicle to be retrieved, and further, the attribute information of the picture of the vehicle to be retrieved is obtained.
- the retrieval device When the retrieval device obtains the attribute information of the vehicle picture to be retrieved, it can query the stored vehicle picture, the collection time of the vehicle picture, and the attribute information of the vehicle picture according to the attribute information of the vehicle picture to be retrieved, and compare it with the attribute information of the vehicle picture to be retrieved The vehicle picture corresponding to the attribute information of the matched vehicle picture is determined as the second vehicle picture that matches the filter criteria of the vehicle search.
- the retrieval device may directly query the matching vehicle from the database according to the attribute information of the vehicle picture to be retrieved The entry where the attribute information of the picture is located, and the vehicle picture information is obtained from the queried entry.
- Step S1120 Generate a video summary according to the second vehicle picture and the collection time of the second vehicle picture, and play back the video summary.
- step S820 For a specific implementation method of this step, reference may be made to step S820, and details are not described herein again.
- the vehicle by obtaining and storing the vehicle picture, the collection time of the vehicle picture, and the attribute information of the vehicle picture in the video data of the video source device, when a vehicle retrieval request is received, the vehicle is retrieved according to the vehicle carried in the vehicle retrieval request.
- Extracting matching vehicle pictures from video data improves the efficiency and accuracy of vehicle retrieval, and ensures the consistency of vehicle video tracking.
- FIG. 12 is a schematic structural diagram of a video summary generating apparatus according to an embodiment of the present application.
- the video summary generating apparatus may be applied to the search device in the foregoing embodiment.
- the video summary generating apparatus The device may include the following units.
- the receiving unit 1210 is configured to receive a target search request, where the target search request carries characteristic information of a target to be searched.
- the search unit 1220 is configured to search for a first target picture that matches feature information of the target to be searched.
- the processing unit 1230 is configured to generate a video summary according to the first target picture and the acquisition time corresponding to the first target picture.
- the device further includes the following units.
- the obtaining unit 1240 is configured to obtain target picture information in the video data of the video source device, where the target picture information includes the target picture, a collection time of the target picture, and attribute information of the target picture.
- the saving unit 1250 is configured to save the target picture information to a picture information database.
- the obtaining unit 1240 is specifically configured to receive target picture information sent by the video source device.
- the obtaining unit 1240 is specifically configured to receive the target picture and the collection time of the target picture sent by the video source device; model the target picture, and Extracting attribute information of the target picture.
- the obtaining unit 1240 is specifically configured to receive the target picture, the collection time of the target picture, and first attribute information of the target picture sent by the video source device; Model the target picture, and extract the second attribute information of the target picture; determine the attribute information of the target picture according to the first attribute information of the target picture and the second attribute information of the target picture.
- the obtaining unit 1240 is specifically configured to perform target detection on the video data provided by the video source device to obtain the target picture and the target image in the video data. Acquisition time of the target picture; modeling the target picture, and extracting attribute information of the target picture.
- the feature information of the target to be searched includes attribute information of the target to be searched; and the searching unit 1220 is specifically configured to be stored in the picture information database according to the attribute information of the target to be searched Searching for a matching first target picture.
- the feature information of the target to be searched includes a target picture to be searched; the search unit 1220 is specifically configured to model the target picture to be searched and extract the target to be searched Attribute information of the target picture; and searching for the matching first target picture in the picture information database according to the attribute information of the target picture to be searched.
- the feature information of the target to be searched includes target picture to be searched and third attribute information of the target picture to be searched; and the searching unit 1220 is specifically configured to perform a search on the target picture to be searched Performing modeling, and extracting fourth attribute information of the target picture to be searched; determining attribute information of the target picture to be searched according to the third attribute information and the fourth attribute information; and according to the target to be searched
- the attribute information of the picture searches for a matching first target picture in the picture information database.
- the target search request further carries a search time period range;
- the search unit 1220 is specifically configured to target the target picture in the picture information database according to the search time range range. Perform filtering to obtain a second target picture whose acquisition time is within the range of the search period; and search for a matching first target picture from the second target picture according to the feature information of the target to be searched.
- the target search request further carries a search channel number
- the target picture information further includes a channel number of the target picture
- the search unit 1220 is specifically configured to perform the search according to the search.
- the channel number is used to filter the target pictures in the picture information database to obtain a third target picture whose channel number is consistent with the search channel number; according to the feature information of the target to be searched, from the third target picture Search for a matching first target picture.
- the target picture is a face picture
- the target search request is a face search request
- the target picture is a vehicle picture
- the target search request is a vehicle search request
- the processing unit 1230 is specifically configured to sort the first target picture in an order from early to late in the acquisition time; and generate the first target picture according to the sorted first target picture. Video summary.
- the processing unit 1230 is specifically configured to determine, for each first target picture, a target video clip corresponding to the first target picture, where the target video clip is the first target Recording data between the n-th second before the acquisition time corresponding to the picture and the m-th second after the acquisition time of the first target picture; generating a video summary according to each of the target video clips.
- the processing unit 1230 is specifically configured to, when there are multiple first target pictures with the same acquisition time, for any first target picture in the multiple first target pictures, Determine the start time point and end time point of the video clip corresponding to the first target picture, where the start time point is the n-th second before the acquisition time corresponding to the first target picture, and the end time point is The m-th second after the acquisition time corresponding to the first target picture; searching whether the I frame at the start time point exists in the recording data of the video data channel to which the first target picture belongs, and whether the end time point exists If there are I frames at the start time point and I frames at the end time point, discard the remaining first target pictures in the multiple first target pictures, and record the video corresponding to the first target pictures The clip is determined as the target video clip.
- the processing unit 1230 is further configured to: if there is no I-frame at the start time point, the start time of the video clip corresponding to the first target picture The point is increased by x seconds to obtain a new starting time point, and the above search steps are repeated until the I frame of the new starting time point is searched in the recording data of the video data channel to which the first target picture belongs.
- the new start time point of the video clip corresponding to the first target picture is the same as the acquisition time; if there is no I frame at the end time point, the corresponding The end time point of the video clip is reduced by x seconds to obtain a new end time point, and the above search steps are repeated until the new search result is found in the video data of the video data channel to which the first target picture belongs.
- I frame at the end time point, or the new end time point of the video clip corresponding to the first target picture is the same as the acquisition time; the corresponding video clips in the plurality of first target pictures respectively in, Selecting the longest video clip as the target video clip; discarding the first target picture corresponding to the remaining video clips of the plurality of first target pictures.
- the processing unit 1230 is specifically configured to filter the target video clip according to a start time point and an end time point of each target video clip to remove time-repeated video data. Generating the video summary according to the filtered target video clip.
- the processing unit 1230 is specifically configured to sort each of the target video clips according to the start time point of each of the target video clips; for an adjacent first target The video clip and the second target video clip, when the end time point of the first target video clip is greater than or equal to the start time point of the second target video clip, if the first target video clip and the second If the target video clip belongs to the same video data channel, the first target video clip and the second target video clip are merged, and the start time point of the combined video clip is the start time of the first target video clip Point, the end time point is the end time point of the second target video clip; if the first target video clip and the second target video clip belong to different video data channels, the The end time point is used as the start time point of the second target video clip, or the start time point of the second target video clip is used as the first The end time point of the target video clip; wherein the start time point of the first target video clip is smaller than the start time point of the second target video clip.
- FIG. 14 is a schematic diagram of a hardware structure of an electronic device according to an embodiment of the present application.
- the electronic device may include a processor 1401, a communication interface 1402, a memory 1403, and a communication bus 1404.
- the processor 1401, the communication interface 1402, and the memory 1403 complete communication with each other through the communication bus 1404.
- a computer program is stored in the memory 1403; the processor 1401 can execute the program stored in the memory 1403 to execute the video digest generating method described above.
- the memory 1403 mentioned herein may be any electronic, magnetic, optical, or other physical storage device, and may contain or store information such as executable instructions, data, and so on.
- the memory 1402 may be: RAM (Radom Access Memory), volatile memory, non-volatile memory, flash memory, storage drive (such as hard drive), solid state hard disk, any type of storage disk (such as optical disk , DVD, etc.), or similar storage media, or a combination thereof.
- An embodiment of the present application further provides a machine-readable storage medium storing a computer program, such as the memory 1403 in FIG. 14, and the computer program may be executed by the processor 1401 in the electronic device shown in FIG. 14 to implement the foregoing description Video digest generation method.
- a computer program such as the memory 1403 in FIG. 14
- the computer program may be executed by the processor 1401 in the electronic device shown in FIG. 14 to implement the foregoing description Video digest generation method.
Abstract
Description
人脸图片的位置信息Face image location information | 人脸图片的采集时间Face image collection time | 人脸图片的属性信息Face image attribute information |
人脸图片1的位置信息Location information of face picture 1 | 人脸图片1的采集时间Acquisition time of face picture 1 | 人脸图片1的属性信息Attribute information of face picture 1 |
人脸图片2的位置信息Location information of face picture 2 | 人脸图片2的采集时间Acquisition time of face picture 2 | 人脸图片2的属性信息Attribute information of face picture 2 |
…... | …... | …... |
车辆图片的位置信息Location information of vehicle pictures | 车辆图片的采集时间Collection time of vehicle pictures | 车辆图片的属性信息Vehicle picture attribute information |
车辆图片1的位置信息Location information of vehicle picture 1 | 车辆图片1的采集时间Collection time of vehicle picture 1 | 车辆图片1的属性信息Attribute information of vehicle picture 1 |
车辆图片2的位置信息Location information of vehicle picture 2 | 车辆图片2的采集时间Collection time of vehicle picture 2 | 车辆图片2的属性信息Attribute information of vehicle picture 2 |
…... | …... | …... |
Claims (40)
- 一种视频摘要生成方法,其特征在于,包括:A video summary generating method is characterized in that it includes:接收目标搜索请求,所述目标搜索请求中携带有待搜索目标的特征信息;Receiving a target search request, where the target search request carries characteristic information of a target to be searched;搜索与所述待搜索目标的特征信息匹配的第一目标图片;Searching for a first target picture that matches the feature information of the target to be searched;根据所述第一目标图片和所述第一目标图片对应的采集时间生成视频摘要。Generating a video summary according to the first target picture and the acquisition time corresponding to the first target picture.
- 根据权利要求1所述的方法,其特征在于,在搜索与所述待搜索目标的特征信息匹配的所述第一目标图片之前,还包括:The method according to claim 1, before searching for the first target picture matching the feature information of the target to be searched, further comprising:获取视频源设备的视频数据中的目标图片信息,所述目标图片信息包括目标图片、目标图片的采集时间以及目标图片的属性信息;Acquiring target picture information in video data of a video source device, where the target picture information includes the target picture, a collection time of the target picture, and attribute information of the target picture;将所述目标图片信息保存至图片信息库。Save the target picture information to a picture information database.
- 根据权利要求2所述的方法,其特征在于,获取所述视频源设备的所述视频数据中的所述目标图片信息,包括:The method according to claim 2, wherein the acquiring the target picture information in the video data of the video source device comprises:接收所述视频源设备发送的所述目标图片信息。Receiving the target picture information sent by the video source device.
- 根据权利要求2所述的方法,其特征在于,获取所述视频源设备的所述视频数据中的所述目标图片信息,包括:The method according to claim 2, wherein the acquiring the target picture information in the video data of the video source device comprises:接收所述视频源设备发送的所述目标图片及所述目标图片的采集时间;Receiving the target picture and the acquisition time of the target picture sent by the video source device;对所述目标图片进行建模,并提取所述目标图片的属性信息。Model the target picture, and extract attribute information of the target picture.
- 根据权利要求2所述的方法,其特征在于,获取所述视频源设备的所述视频数据中的所述目标图片信息,包括:The method according to claim 2, wherein the acquiring the target picture information in the video data of the video source device comprises:接收所述视频源设备发送的所述目标图片、所述目标图片的采集时间以及目标图片的第一属性信息;Receiving the target picture, the acquisition time of the target picture, and first attribute information of the target picture sent by the video source device;对所述目标图片进行建模,并提取所述目标图片的第二属性信息;Modeling the target picture, and extracting second attribute information of the target picture;根据所述目标图片的第一属性信息和所述目标图片的第二属性信息确定所述目标图片的属性信息。Determine the attribute information of the target picture according to the first attribute information of the target picture and the second attribute information of the target picture.
- 根据权利要求2所述的方法,其特征在于,获取所述视频源设备的所述视频数据中的所述目标图片信息,包括:The method according to claim 2, wherein the acquiring the target picture information in the video data of the video source device comprises:对所述视频源设备提供的所述视频数据进行目标检测,得到所述视频数据中的所述目标图片及所述目标图片的采集时间;Performing target detection on the video data provided by the video source device to obtain the target picture and the acquisition time of the target picture in the video data;对所述目标图片进行建模,并提取所述目标图片的属性信息。Model the target picture, and extract attribute information of the target picture.
- 根据权利要求2所述的方法,其特征在于,所述待搜索目标的特征信息包括待搜索目标的属性信息;The method according to claim 2, wherein the feature information of the target to be searched includes attribute information of the target to be searched;搜索与所述待搜索目标的特征信息匹配的所述第一目标图片,包括:Searching for the first target picture that matches the feature information of the target to be searched includes:根据所述待搜索目标的属性信息在所述图片信息库中搜索匹配的所述第一目标图片。Searching for the matching first target picture in the picture information database according to the attribute information of the target to be searched.
- 根据权利要求2所述的方法,其特征在于,所述待搜索目标的特征信息包括待搜索目标图片;The method according to claim 2, wherein the feature information of the target to be searched comprises a target picture to be searched;搜索与所述待搜索目标的特征信息匹配的所述第一目标图片,包括:Searching for the first target picture that matches the feature information of the target to be searched includes:对所述待搜索目标图片进行建模,并提取所述待搜索目标图片的属性信息;Modeling the target picture to be searched, and extracting attribute information of the target picture to be searched;根据所述待搜索目标图片的属性信息在所述图片信息库中搜索匹配的所述第一目标图片。Searching for the matching first target picture in the picture information database according to the attribute information of the target picture to be searched.
- 根据权利要求2所述的方法,其特征在于,所述待搜索目标的特征信息包括待搜索目标图片和所述待搜索目标图片的第三属性信息;The method according to claim 2, wherein the feature information of the target to be searched includes target picture to be searched and third attribute information of the target picture to be searched;搜索与所述待搜索目标的特征信息匹配的所述第一目标图片,包括:Searching for the first target picture that matches the feature information of the target to be searched includes:对所述待搜索目标图片进行建模,并提取所述待搜索目标图片的第四属性信息;Modeling the target picture to be searched, and extracting fourth attribute information of the target picture to be searched;根据所述第三属性信息和所述第四属性信息,确定所述待搜索目标图片的属性信息;Determining attribute information of the target picture to be searched according to the third attribute information and the fourth attribute information;根据所述待搜索目标图片的属性信息在所述图片信息库中搜索匹配的所述第一目标图片。Searching for the matching first target picture in the picture information database according to the attribute information of the target picture to be searched.
- 根据权利要求2所述的方法,其特征在于,所述目标搜索请求中还携带有搜索时间段范围;The method according to claim 2, wherein the target search request further carries a search time period range;搜索与所述待搜索目标的特征信息匹配的所述第一目标图片,包括:Searching for the first target picture that matches the feature information of the target to be searched includes:根据所述搜索时间段范围对所述图片信息库中的所述目标图片进行筛选,得到采集时间在所述搜索时间段范围内的第二目标图片;Filtering the target pictures in the picture information database according to the range of the search time period to obtain a second target picture whose acquisition time is within the range of the search time period;根据所述待搜索目标的特征信息,从所述第二目标图片中搜索匹配的所述第一目标图片。Searching for a matching first target picture from the second target picture according to the feature information of the target to be searched.
- 根据权利要求2所述的方法,其特征在于,所述目标搜索请求中还携带有搜索通道号,所述目标图片信息还包括有所述目标图片的通道号;The method according to claim 2, wherein the target search request further carries a search channel number, and the target picture information further includes a channel number of the target picture;搜索与所述待搜索目标的特征信息匹配的所述第一目标图片,包括:Searching for the first target picture that matches the feature information of the target to be searched includes:根据所述搜索通道号对所述图片信息库中的所述目标图片进行筛选,得到所述通道号与所述搜索通道号一致的第三目标图片;Filtering the target picture in the picture information database according to the search channel number to obtain a third target picture whose channel number is consistent with the search channel number;根据所述待搜索目标的特征信息,从所述第三目标图片中搜索匹配的所述第一目标图片。Searching for a matching first target picture from the third target picture according to the feature information of the target to be searched.
- 根据权利要求2所述的方法,其特征在于,所述目标图片为人脸图片,所述目 标搜索请求为人脸搜索请求。The method according to claim 2, wherein the target picture is a face picture, and the target search request is a face search request.
- 根据权利要求2所述的方法,其特征在于,所述目标图片为车辆图片,所述目标搜索请求为车辆搜索请求。The method according to claim 2, wherein the target picture is a vehicle picture, and the target search request is a vehicle search request.
- 根据权利要求1所述的方法,其特征在于,根据所述第一目标图片和所述第一目标图片对应的采集时间生成所述视频摘要,包括:The method according to claim 1, wherein generating the video summary according to the first target picture and the acquisition time corresponding to the first target picture comprises:按采集时间从早到晚的顺序排序所述第一目标图片;Sorting the first target picture in the order of the collection time from morning to night;根据排序后的所述第一目标图片,生成所述视频摘要。Generating the video summary according to the sorted first target picture.
- 根据权利要求1或14所述的方法,其特征在于,根据所述第一目标图片和所述第一目标图片对应的采集时间生成所述视频摘要,包括:The method according to claim 1 or 14, wherein generating the video summary according to the first target picture and the acquisition time corresponding to the first target picture comprises:对于每个第一目标图片,确定该第一目标图片对应的目标录像片段,其中,该目标录像片段为该第一目标图片对应的所述采集时间之前的第n秒到该第一目标图片的所述采集时间之后的第m秒之间的录像数据;For each first target picture, a target video clip corresponding to the first target picture is determined, where the target video clip is from the nth second before the acquisition time corresponding to the first target picture to the first target picture Video data between the m-th second after the acquisition time;根据各所述目标录像片段生成视频摘要。A video summary is generated according to each of the target video clips.
- 根据权利要求15所述的方法,其特征在于,对于所述每个第一目标图片,确定该第一目标图片对应的所述目标录像片段,包括:The method according to claim 15, wherein for each of the first target pictures, determining the target video clip corresponding to the first target picture comprises:当存在采集时间相同的多张第一目标图片时,对于该多张第一目标图片中的任一第一目标图片,确定该第一目标图片对应的录像片段的起始时间点和结束时间点,其中,该起始时间点为该第一目标图片对应的所述采集时间之前的第n秒,该结束时间点为该第一目标图片对应的所述采集时间之后的第m秒;When there are multiple first target pictures with the same acquisition time, for any first target picture in the multiple first target pictures, determine the start time point and end time point of the video clip corresponding to the first target picture Where the start time point is the n-th second before the acquisition time corresponding to the first target picture, and the end time point is the m-th second after the acquisition time corresponding to the first target picture;搜索该第一目标图片所属的视频数据通道的录像数据中是否存在该起始时间点的I帧,以及是否存在该结束时间点的I帧;Searching whether the I frame at the start time point exists in the recording data of the video data channel to which the first target picture belongs, and whether the I frame at the end time point exists;若存在该起始时间点的I帧和该结束时间点的I帧,则丢弃该多张第一目标图片中的其余第一目标图片,并将该第一目标图片对应的录像片段确定为所述目标录像片段。If there are I frames at the start time point and I frames at the end time point, the remaining first target pictures in the plurality of first target pictures are discarded, and the video clip corresponding to the first target picture is determined as the Describe the target video clip.
- 根据权利要求16所述的方法,其特征在于,所述方法还包括:The method according to claim 16, further comprising:若不存在该起始时间点的I帧,则将该第一目标图片对应的所述录像片段的所述起始时间点增加x秒得到新的起始时间点,并重复上述搜索步骤,直至在该第一目标图片所属的所述视频数据通道的所述录像数据中搜索到该新的起始时间点的I帧,或,该第一目标图片对应的所述录像片段的所述新的起始时间点与所述采集时间相同;If there is no I frame at the start time point, increase the start time point of the video clip corresponding to the first target picture by x seconds to obtain a new start time point, and repeat the above search steps until The I-frame at the new start time point is searched in the recording data of the video data channel to which the first target picture belongs, or the new one of the video clips corresponding to the first target picture The starting time point is the same as the collection time;若不存在该结束时间点的I帧,则将该第一目标图片对应的所述录像片段的所述结束时间点减少x秒得到新的结束时间点,并重复上述搜索步骤,直至在该第一目标图片所属的所述视频数据通道的所述录像数据中搜索到该新的结束时间点的I帧,或,该第 一目标图片对应的所述录像片段的所述新的结束时间点与所述采集时间相同;If there is no I frame at the end time point, the end time point of the video clip corresponding to the first target picture is reduced by x seconds to obtain a new end time point, and the above search steps are repeated until the An I-frame of the new end time point is searched in the recording data of the video data channel to which a target picture belongs, or the new end time point of the video clip corresponding to the first target picture and The acquisition times are the same;在该多张第一目标图片中分别对应的录像片段中,选择时长最长录像片段作为所述目标录像片段;Selecting, from the corresponding video clips in the multiple first target pictures, the longest video clip as the target video clip;丢弃该多张第一目标图片中的其余时长的录像片段对应的第一目标图片。Discard the first target pictures corresponding to the remaining video clips in the multiple first target pictures.
- 根据权利要求15所述的方法,其特征在于,根据各所述目标录像片段生成所述视频摘要,包括:The method according to claim 15, wherein generating the video summary according to each of the target video recording segments comprises:根据各所述目标录像片段的起始时间点和结束时间点对所述目标录像片段进行过滤,去除时间重复的录像数据;Filtering the target video clip according to the start time point and the end time point of each target video clip to remove time-repeated video data;根据过滤后的目标录像片段生成所述视频摘要。Generating the video summary according to the filtered target video clip.
- 根据权利要求18所述的方法,其特征在于,根据各所述目标录像片段的所述起始时间点和所述结束时间点对所述目标录像片段进行过滤,包括:The method according to claim 18, wherein filtering the target video clip according to the start time point and the end time point of each of the target video clips comprises:按照各所述目标录像片段的所述起始时间点对各所述目标录像片段进行排序;Sorting each of the target video clips according to the start time point of each of the target video clips;对于相邻的第一目标录像片段和第二目标录像片段,当所述第一目标录像片段的结束时间点大于等于所述第二目标录像片段的起始时间点时,For the adjacent first target video clip and the second target video clip, when the end time point of the first target video clip is greater than or equal to the start time point of the second target video clip,若所述第一目标录像片段和所述第二目标录像片段属于同一视频数据通道,则将所述第一目标录像片段和所述第二目标录像片段合并,合并后的录像片段的起始时间点为所述第一目标录像片段的起始时间点,结束时间点为所述第二目标录像片段的结束时间点;If the first target video clip and the second target video clip belong to the same video data channel, merge the first target video clip and the second target video clip, and start time of the combined video clip The point is the start time point of the first target video clip, and the end time point is the end time point of the second target video clip;若所述第一目标录像片段和所述第二目标录像片段属于不同视频数据通道,则以所述第一目标录像片段的所述结束时间点作为所述第二目标录像片段的所述起始时间点,或,以所述第二目标片段的所述起始时间点作为所述第一目标录像片段的所述结束时间点;If the first target video clip and the second target video clip belong to different video data channels, the end time point of the first target video clip is used as the start of the second target video clip A time point, or the start time point of the second target clip as the end time point of the first target video clip;其中,所述第一目标录像片段的所述起始时间点小于所述第二目标录像片段的所述起始时间点。Wherein, the start time point of the first target video clip is smaller than the start time point of the second target video clip.
- 一种视频摘要生成装置,其特征在于,包括:A video digest generating device, comprising:接收单元,用于接收目标搜索请求,所述目标搜索请求中携带有待搜索目标的特征信息;A receiving unit, configured to receive a target search request, where the target search request carries characteristic information of a target to be searched;搜索单元,用于搜索与所述待搜索目标的特征信息匹配的第一目标图片;A search unit, configured to search for a first target picture that matches feature information of the target to be searched;处理单元,用于根据所述第一目标图片和所述第一目标图片对应的采集时间生成视频摘要。A processing unit, configured to generate a video summary according to the first target picture and the acquisition time corresponding to the first target picture.
- 根据权利要求20所述的装置,其特征在于,所述装置还包括:The apparatus according to claim 20, further comprising:获取单元,用于获取视频源设备的视频数据中的目标图片信息,所述目标图片信息包括目标图片、目标图片的采集时间以及目标图片的属性信息;An obtaining unit, configured to obtain target picture information in video data of a video source device, where the target picture information includes the target picture, a collection time of the target picture, and attribute information of the target picture;保存单元,用于将所述目标图片信息保存至图片信息库。A saving unit, configured to save the target picture information to a picture information database.
- 根据权利要求21所述的装置,其特征在于,The device according to claim 21, wherein:所述获取单元,具体用于接收所述视频源设备发送的所述目标图片信息。The obtaining unit is specifically configured to receive the target picture information sent by the video source device.
- 根据权利要求21所述的装置,其特征在于,所述获取单元,具体用于The apparatus according to claim 21, wherein the obtaining unit is specifically configured to:接收所述视频源设备发送的所述目标图片及所述目标图片的采集时间;Receiving the target picture and the acquisition time of the target picture sent by the video source device;对所述目标图片进行建模,并提取所述目标图片的属性信息。Model the target picture, and extract attribute information of the target picture.
- 根据权利要求21所述的装置,其特征在于,所述获取单元,具体用于The apparatus according to claim 21, wherein the obtaining unit is specifically configured to:接收所述视频源设备发送的所述目标图片、所述目标图片的采集时间以及目标图片的第一属性信息;Receiving the target picture, the acquisition time of the target picture, and first attribute information of the target picture sent by the video source device;对所述目标图片进行建模,并提取所述目标图片的第二属性信息;Modeling the target picture, and extracting second attribute information of the target picture;根据所述目标图片的第一属性信息和所述目标图片的第二属性信息确定所述目标图片的属性信息。Determine the attribute information of the target picture according to the first attribute information of the target picture and the second attribute information of the target picture.
- 根据权利要求21所述的装置,其特征在于,所述获取单元,具体用于The apparatus according to claim 21, wherein the obtaining unit is specifically configured to:对所述视频源设备提供的所述视频数据进行目标检测,得到所述视频数据中的所述目标图片及所述目标图片的采集时间;Performing target detection on the video data provided by the video source device to obtain the target picture and the acquisition time of the target picture in the video data;对所述目标图片进行建模,并提取所述目标图片的属性信息。Model the target picture, and extract attribute information of the target picture.
- 根据权利要求21所述的装置,其特征在于,所述待搜索目标的特征信息包括待搜索目标的属性信息;The apparatus according to claim 21, wherein the feature information of the target to be searched includes attribute information of the target to be searched;所述搜索单元,具体用于根据所述待搜索目标的属性信息在所述图片信息库中搜索匹配的所述第一目标图片。The search unit is specifically configured to search for a matching first target picture in the picture information database according to attribute information of the target to be searched.
- 根据权利要求21所述的装置,其特征在于,所述待搜索目标的特征信息包括待搜索目标图片;The device according to claim 21, wherein the feature information of the target to be searched includes a target picture to be searched;所述搜索单元,具体用于The search unit is specifically used for对所述待搜索目标图片进行建模,并提取所述待搜索目标图片的属性信息;Modeling the target picture to be searched, and extracting attribute information of the target picture to be searched;根据所述待搜索目标图片的属性信息在所述图片信息库中搜索匹配的所述第一目标图片。Searching for the matching first target picture in the picture information database according to the attribute information of the target picture to be searched.
- 根据权利要求21所述的装置,其特征在于,所述待搜索目标的特征信息包括待搜索目标图片和所述待搜索目标图片的第三属性信息;The apparatus according to claim 21, wherein the feature information of the target to be searched includes target picture to be searched and third attribute information of the target picture to be searched;所述搜索单元,具体用于The search unit is specifically used for对所述待搜索目标图片进行建模,并提取所述待搜索目标图片的第四属性信息;Modeling the target picture to be searched, and extracting fourth attribute information of the target picture to be searched;根据所述第三属性信息和所述第四属性信息,确定所述待搜索目标图片的属性信息;Determining attribute information of the target picture to be searched according to the third attribute information and the fourth attribute information;根据所述待搜索目标图片的属性信息在所述图片信息库中搜索匹配的所述第一目标图片。Searching for the matching first target picture in the picture information database according to the attribute information of the target picture to be searched.
- 根据权利要求21所述的装置,其特征在于,所述目标搜索请求中还携带有搜索时间段范围;The apparatus according to claim 21, wherein the target search request further carries a search time period range;所述搜索单元,具体用于The search unit is specifically used for根据所述搜索时间段范围对所述图片信息库中的目标图片进行筛选,以得到采集时间在所述搜索时间段范围内的第二目标图片;Filtering the target pictures in the picture information database according to the search time period range to obtain a second target picture with a collection time within the search time range;根据所述待搜索目标的特征信息从所述第二目标图片中搜索匹配的第一目标图片。Searching for a matching first target picture from the second target picture according to the feature information of the target to be searched.
- 根据权利要求21所述的装置,其特征在于,所述目标搜索请求中还携带有搜索通道号,所述目标图片信息还包括有所述目标图片的通道号;The device according to claim 21, wherein the target search request further carries a search channel number, and the target picture information further includes a channel number of the target picture;所述搜索单元,具体用于The search unit is specifically used for根据所述搜索通道号对所述图片信息库中的目标图片进行筛选,得到所述通道号与所述搜索通道号一致的第三目标图片;Filtering the target pictures in the picture information database according to the search channel number to obtain a third target picture with the same channel number as the search channel number;根据所述待搜索目标的特征信息,从所述第三目标图片中搜索匹配的第一目标图片。Searching for a matching first target picture from the third target picture according to the feature information of the target to be searched.
- 根据权利要求21所述的装置,其特征在于,所述目标图片为人脸图片,所述目标搜索请求为人脸搜索请求。The device according to claim 21, wherein the target picture is a face picture, and the target search request is a face search request.
- 根据权利要求21所述的装置,其特征在于,所述目标图片为车辆图片,所述目标搜索请求为车辆搜索请求。The device according to claim 21, wherein the target picture is a vehicle picture, and the target search request is a vehicle search request.
- 根据权利要求20所述的装置,其特征在于,所述处理单元,具体用于The apparatus according to claim 20, wherein the processing unit is specifically configured to:按采集时间从早到晚的顺序排序所述第一目标图片;Sorting the first target picture in the order of the collection time from morning to night;根据排序后的所述第一目标图片,生成所述视频摘要。Generating the video summary according to the sorted first target picture.
- 根据权利要求20或33所述的装置,其特征在于,所述处理单元,具体用于The device according to claim 20 or 33, wherein the processing unit is specifically configured to对于每个第一目标图片,确定该第一目标图片对应的目标录像片段,其中,该目标录像片段为该第一目标图片对应的所述采集时间之前的第n秒到该第一目标图片的所述采集时间之后的第m秒之间的录像数据;For each first target picture, a target video clip corresponding to the first target picture is determined, where the target video clip is from the nth second before the acquisition time corresponding to the first target picture to the first target picture Video data between the m-th second after the acquisition time;根据各所述目标录像片段生成视频摘要。A video summary is generated according to each of the target video clips.
- 根据权利要求34所述的装置,其特征在于,所述处理单元,具体用于The apparatus according to claim 34, wherein the processing unit is specifically configured to:当存在采集时间相同的多张第一目标图片时,对于该多张第一目标图片中的任一第一目标图片,确定该第一目标图片对应的录像片段的起始时间点和结束时间点,其中, 该起始时间点为该第一目标图片对应的所述采集时间之前的第n秒,该结束时间点为该第一目标图片对应的所述采集时间之后的第m秒;When there are multiple first target pictures with the same acquisition time, for any first target picture in the multiple first target pictures, determine the start time point and end time point of the video clip corresponding to the first target picture Where the start time point is the n-th second before the acquisition time corresponding to the first target picture, and the end time point is the m-th second after the acquisition time corresponding to the first target picture;搜索该第一目标图片所属的视频数据通道的录像数据中是否存在该起始时间点的I帧,以及是否存在该结束时间点的I帧;Searching whether the I frame at the start time point exists in the recording data of the video data channel to which the first target picture belongs, and whether the I frame at the end time point exists;若存在该起始时间点的I帧和该结束时间点的I帧,则丢弃该多张第一目标图片中的其余第一目标图片,并将该第一目标图片对应的录像片段确定为所述目标录像片段。If there are I frames at the start time point and I frames at the end time point, the remaining first target pictures in the plurality of first target pictures are discarded, and the video clip corresponding to the first target picture is determined as the Describe the target video clip.
- 根据权利要求35所述的装置,其特征在于,所述处理单元,还用于The apparatus according to claim 35, wherein the processing unit is further configured to:若不存在该起始时间点的I帧,则将该第一目标图片对应的所述录像片段的所述起始时间点增加x秒得到新的起始时间点,并重复上述搜索步骤,直至在该第一目标图片所属的所述视频数据通道的所述录像数据中搜索到该新的起始时间点的I帧,或,该第一目标图片对应的所述录像片段的所述新的起始时间点与所述采集时间相同;If there is no I frame at the start time point, increase the start time point of the video clip corresponding to the first target picture by x seconds to obtain a new start time point, and repeat the above search steps until The I-frame at the new start time point is searched in the recording data of the video data channel to which the first target picture belongs, or the new one of the video clips corresponding to the first target picture The starting time point is the same as the collection time;若不存在该结束时间点的I帧,则将该第一目标图片对应的所述录像片段的所述结束时间点减少x秒得到新的结束时间点,并重复上述搜索步骤,直至在该第一目标图片所属的所述视频数据通道的所述录像数据中搜索到该新的结束时间点的I帧,或,该第一目标图片对应的所述录像片段的所述新的结束时间点与所述采集时间相同;If there is no I frame at the end time point, the end time point of the video clip corresponding to the first target picture is reduced by x seconds to obtain a new end time point, and the above search steps are repeated until the An I-frame of the new end time point is searched in the recording data of the video data channel to which a target picture belongs, or the new end time point of the video clip corresponding to the first target picture and The acquisition times are the same;在该多张第一目标图片中分别对应的录像片段中,选择时长最长录像片段作为所述目标录像片段;Selecting, from the corresponding video clips in the multiple first target pictures, the longest video clip as the target video clip;丢弃该多张第一目标图片中的其余时长的录像片段对应的第一目标图片。Discard the first target pictures corresponding to the remaining video clips in the multiple first target pictures.
- 根据权利要求34所述的装置,其特征在于,所述处理单元,具体用于The apparatus according to claim 34, wherein the processing unit is specifically configured to:根据各所述目标录像片段的起始时间点和结束时间点对所述目标录像片段进行过滤,去除时间重复的录像数据;Filtering the target video clip according to the start time point and the end time point of each target video clip to remove time-repeated video data;根据过滤后的目标录像片段生成所述视频摘要。Generating the video summary according to the filtered target video clip.
- 根据权利要求37所述的装置,其特征在于,所述处理单元,具体用于The apparatus according to claim 37, wherein the processing unit is specifically configured to:按照各所述目标录像片段的所述起始时间点对各所述目标录像片段进行排序;Sorting each of the target video clips according to the start time point of each of the target video clips;对于相邻的第一目标录像片段和第二目标录像片段,当所述第一目标录像片段的结束时间点大于等于所述第二目标录像片段的起始时间点时,For the adjacent first target video clip and the second target video clip, when the end time point of the first target video clip is greater than or equal to the start time point of the second target video clip,若所述第一目标录像片段和所述第二目标录像片段属于同一视频数据通道,则将所述第一目标录像片段和所述第二目标录像片段合并,合并后的录像片段的起始时间点为所述第一目标录像片段的起始时间点,结束时间点为所述第二目标录像片段的结束时间点;If the first target video clip and the second target video clip belong to the same video data channel, merge the first target video clip and the second target video clip, and start time of the combined video clip The point is the start time point of the first target video clip, and the end time point is the end time point of the second target video clip;若所述第一目标录像片段和所述第二目标录像片段属于不同视频数据通道,则以所 述第一目标录像片段的所述结束时间点作为所述第二目标录像片段的所述起始时间点,或,以所述第二目标片段的所述起始时间点作为所述第一目标录像片段的所述结束时间点;If the first target video clip and the second target video clip belong to different video data channels, the end time point of the first target video clip is used as the start of the second target video clip A time point, or the start time point of the second target clip as the end time point of the first target video clip;其中,所述第一目标录像片段的所述起始时间点小于所述第二目标录像片段的所述起始时间点。Wherein, the start time point of the first target video clip is smaller than the start time point of the second target video clip.
- 一种电子设备,其特征在于,包括处理器、通信接口、存储器和通信总线,其中,所述处理器,所述通信接口,所述存储器通过所述通信总线完成相互间的通信;An electronic device, comprising a processor, a communication interface, a memory, and a communication bus, wherein the processor, the communication interface, and the memory complete communication with each other through the communication bus;所述存储器,用于存放计算机程序;The memory is used to store a computer program;所述处理器,用于执行所述存储器上所存放的所述计算机程序时,实现权利要求1-19任一项所述的方法步骤。The processor is configured to implement the method steps according to any one of claims 1 to 19 when the computer program stored in the memory is executed.
- 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质内存储有计算机程序,所述计算机程序被处理器执行时实现权利要求1-19任一项所述的方法步骤。A computer-readable storage medium, characterized in that a computer program is stored in the computer-readable storage medium, and when the computer program is executed by a processor, the method steps according to any one of claims 1-19 are implemented.
Applications Claiming Priority (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811027494.7 | 2018-09-04 | ||
CN201811027515.5 | 2018-09-04 | ||
CN201811026894.6 | 2018-09-04 | ||
CN201811026894.6A CN110876090B (en) | 2018-09-04 | 2018-09-04 | Video abstract playback method and device, electronic equipment and readable storage medium |
CN201811027515.5A CN110876029B (en) | 2018-09-04 | 2018-09-04 | Video abstract playback method and device, electronic equipment and readable storage medium |
CN201811025858.8A CN110876092B (en) | 2018-09-04 | 2018-09-04 | Video abstract generation method and device, electronic equipment and readable storage medium |
CN201811025858.8 | 2018-09-04 | ||
CN201811027494.7A CN110929095A (en) | 2018-09-04 | 2018-09-04 | Video abstract playback method and device, electronic equipment and readable storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020048324A1 true WO2020048324A1 (en) | 2020-03-12 |
Family
ID=69722729
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/102073 WO2020048324A1 (en) | 2018-09-04 | 2019-08-22 | Video abstract generation method and apparatus, and electronic device and readable storage medium |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2020048324A1 (en) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103927364A (en) * | 2014-04-18 | 2014-07-16 | 苏州科达科技股份有限公司 | Storage method and system and display system for video abstract data |
CN105335387A (en) * | 2014-07-04 | 2016-02-17 | 杭州海康威视系统技术有限公司 | Retrieval method for video cloud storage system |
US20160342688A1 (en) * | 2013-06-25 | 2016-11-24 | Emc Corporation | Large scale video analytics architecture |
CN107436944A (en) * | 2017-07-31 | 2017-12-05 | 福州瑞芯微电子股份有限公司 | A kind of method and system of video search |
CN108337482A (en) * | 2018-02-08 | 2018-07-27 | 北京信息科技大学 | The storage method and system of monitor video |
-
2019
- 2019-08-22 WO PCT/CN2019/102073 patent/WO2020048324A1/en active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160342688A1 (en) * | 2013-06-25 | 2016-11-24 | Emc Corporation | Large scale video analytics architecture |
CN103927364A (en) * | 2014-04-18 | 2014-07-16 | 苏州科达科技股份有限公司 | Storage method and system and display system for video abstract data |
CN105335387A (en) * | 2014-07-04 | 2016-02-17 | 杭州海康威视系统技术有限公司 | Retrieval method for video cloud storage system |
CN107436944A (en) * | 2017-07-31 | 2017-12-05 | 福州瑞芯微电子股份有限公司 | A kind of method and system of video search |
CN108337482A (en) * | 2018-02-08 | 2018-07-27 | 北京信息科技大学 | The storage method and system of monitor video |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3830714B1 (en) | Systems and methods for generating metadata describing unstructured data objects at the storage edge | |
AU2017204338B2 (en) | Industry first method that uses users search query to show highly relevant poster frames for stock videos thereby resulting in great experience and more sales among users of stock video service | |
US9560323B2 (en) | Method and system for metadata extraction from master-slave cameras tracking system | |
EP3253042B1 (en) | Intelligent processing method and system for video data | |
JP6446971B2 (en) | Data processing apparatus, data processing method, and computer program | |
CN102483767B (en) | Object association means, method of mapping, program and recording medium | |
US7243101B2 (en) | Program, image managing apparatus and image managing method | |
JP5791364B2 (en) | Face recognition device, face recognition method, face recognition program, and recording medium recording the program | |
US8457466B1 (en) | Videore: method and system for storing videos from multiple cameras for behavior re-mining | |
US20080247610A1 (en) | Apparatus, Method and Computer Program for Processing Information | |
US20210382933A1 (en) | Method and device for archive application, and storage medium | |
WO2014106384A1 (en) | Method, apparatus and video monitoring system for providing monitoring video information | |
GB2528330A (en) | A method of video analysis | |
JP2011528150A (en) | Method and system for automatic personal annotation of video content | |
JP2012509522A (en) | Semantic classification for each event | |
US10037467B2 (en) | Information processing system | |
CN110543584B (en) | Method, device, processing server and storage medium for establishing face index | |
WO2022156234A1 (en) | Target re-identification method and apparatus, and computer-readable storage medium | |
JP2014067333A (en) | Image processing device, image processing method, and program | |
JPWO2018163398A1 (en) | Similar image search system | |
CN111881320A (en) | Video query method, device, equipment and readable storage medium | |
CN110876090B (en) | Video abstract playback method and device, electronic equipment and readable storage medium | |
CN108540760A (en) | Video monitoring recognition methods, device and system | |
CN109522799A (en) | Information cuing method, device, computer equipment and storage medium | |
WO2021196551A1 (en) | Image retrieval method and apparatus, computer device, and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19857715 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19857715 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19857715 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 040222) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19857715 Country of ref document: EP Kind code of ref document: A1 |