CN111405382B - Video abstract generation method and device, computer equipment and storage medium - Google Patents

Video abstract generation method and device, computer equipment and storage medium Download PDF

Info

Publication number
CN111405382B
CN111405382B CN201910551305.4A CN201910551305A CN111405382B CN 111405382 B CN111405382 B CN 111405382B CN 201910551305 A CN201910551305 A CN 201910551305A CN 111405382 B CN111405382 B CN 111405382B
Authority
CN
China
Prior art keywords
video
target
monitoring
target object
time
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910551305.4A
Other languages
Chinese (zh)
Other versions
CN111405382A (en
Inventor
汪渭春
周平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Hikvision System Technology Co Ltd
Original Assignee
Hangzhou Hikvision System Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Hikvision System Technology Co Ltd filed Critical Hangzhou Hikvision System Technology Co Ltd
Priority to CN201910551305.4A priority Critical patent/CN111405382B/en
Publication of CN111405382A publication Critical patent/CN111405382A/en
Application granted granted Critical
Publication of CN111405382B publication Critical patent/CN111405382B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8549Creating video summaries, e.g. movie trailer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/181Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • Closed-Circuit Television Systems (AREA)

Abstract

The application discloses a video abstract generation method and device, computer equipment and a storage medium, and relates to the field of video monitoring. The video abstract generating method provided by the embodiment of the application not only obtains the surveillance video of the first surveillance device, but also obtains the surveillance videos of the plurality of second surveillance devices in the target area of the first surveillance device, and obtains at least one second surveillance video including the target object from the plurality of second surveillance videos, so that the generated target video abstract not only includes the target object in the first video segment, but also includes the target object in each second video segment in the at least one second video segment.

Description

Video abstract generation method and device, computer equipment and storage medium
Technical Field
The present application relates to the field of video surveillance. In particular to a video abstract generating method, a video abstract generating device, computer equipment and a storage medium.
Background
With the advent of the networking era of video monitoring systems, the scale of the video monitoring systems is continuously enlarged, and the application of video monitoring is more and more extensive. When the related personnel want to search the target object from the monitoring video, the rapid search can be carried out through the video abstract of the monitoring video. The video abstract is the concentration of the surveillance video, and is the summarization of the surveillance video, and related personnel can know the related information of the target object in the surveillance video in a short time through the video abstract.
When the video abstract is generated in the related technology, the monitoring video of the current monitoring equipment is obtained firstly, then the monitoring video is segmented to obtain a sequence video frame, and the video abstract corresponding to the current monitoring equipment is generated according to the sequence video frame.
However, the amount of information included in the video summary generated in the related art is limited, and the amount of information acquired from the video summary by related personnel is small, so that the searching efficiency of the target object is low.
Disclosure of Invention
The embodiment of the application provides a video abstract generation method, a video abstract generation device, computer equipment and a storage medium, and can solve the problem of low searching efficiency of a target object. The technical scheme is as follows:
in one aspect, a method for generating a video summary is provided, where the method includes:
acquiring a first monitoring video of first monitoring equipment, and extracting a target object in the first monitoring video;
determining a plurality of second monitoring devices in a target area according to the first position of the first monitoring device, and acquiring a plurality of second monitoring videos of the plurality of second monitoring devices;
acquiring at least one second monitoring video from the plurality of second monitoring videos according to the target object, wherein each second monitoring video comprises the target object;
generating a first video clip corresponding to the target object according to the first monitoring video, and generating at least one second video clip corresponding to the target object according to the at least one second monitoring video;
and composing the first video segment and the at least one second video segment into a target video summary of the target object.
In one possible implementation, the composing the first video segment and the at least one second video segment into a target video summary of the target object includes:
acquiring a first time for generating the first video segment and at least one second time for generating the at least one second video segment; composing the first video clip and the at least one second video clip into the target video summary in chronological order of the first time and the at least one second time; or,
acquiring third time when the target object appears in the first monitoring video and acquiring each fourth time when the target object appears in each second monitoring video; and composing the first video segment and the at least one second video segment into the target video summary according to the time sequence of the third time and each fourth time.
In another possible implementation manner, the obtaining at least one second surveillance video from the plurality of second surveillance videos according to the target object includes:
determining a target time period comprising a third time according to the third time when the target object appears in the first monitoring video;
and selecting at least one second monitoring video which comprises the target object and has a fourth time appearing in the second monitoring video within the target time period from the plurality of second monitoring videos.
In another possible implementation manner, the generating a first video segment corresponding to the target object according to the first monitoring video includes:
selecting a video clip including the target object in the first monitoring video according to the target object, and taking the video clip as the first video clip; or,
according to the target object, extracting a plurality of key frames including the target object in the first monitoring video, and forming the first video clip by the plurality of key frames.
In another possible implementation manner, after the composing the first video segment and the at least one second video segment into a target video summary of the target object, the method further includes:
when the specified operation of the target video abstract is obtained, obtaining map data corresponding to the target area according to the identification of the target video abstract;
rendering the map data corresponding to the target area in a map display interface to obtain a target electronic map;
displaying the moving track of the target object on the target electronic map, and displaying the target video abstract on the target electronic map.
In another possible implementation manner, the displaying the target video summary on the target electronic map includes:
displaying a first playing window corresponding to the first video clip in a target electronic map according to the first position, and displaying at least one second playing window corresponding to the at least one second video clip in the target electronic map according to at least one second position, wherein the at least one second position is a position of at least one second monitoring device corresponding to the at least one second monitoring video;
and displaying the target video abstract on the target electronic map according to the moving track, the first playing window and the at least one second playing window.
In another possible implementation manner, the displaying the movement track of the target object on the target electronic map includes:
sequencing the third time and at least one fourth time according to the third time and the at least one fourth time to obtain a time sequence, wherein the third time is the time when the target object appears in the first monitoring video, and the at least one fourth time is the time when the target object appears in the at least one second monitoring video;
determining the position of the monitoring equipment corresponding to each two adjacent times from the first position and at least one second position according to each two adjacent times in the time sequence, wherein the at least one second position is the position of at least one second monitoring equipment corresponding to the at least one second monitoring video;
connecting the positions of the monitoring equipment corresponding to every two adjacent times according to a target connection mode to obtain a moving track of the target object between the positions of the monitoring equipment corresponding to every two adjacent times until each of the first position and the at least one second position is connected with other positions except the positions to obtain the moving track of the target object;
and displaying the movement track of the target object on the target electronic map.
In another possible implementation manner, the method further includes:
displaying the map display interface, wherein the map display interface comprises a calling interface, when the calling interface is triggered, a selection interface is displayed, the selection interface comprises a plurality of video abstract identifications, and when the selection operation of the target video abstract identification is obtained, the specified operation of the target video abstract is determined to be obtained; or,
when the triggering operation of the target video abstract is acquired, displaying a play list, wherein the play list comprises an identifier of an electronic map; when the selection operation of the identifier of the electronic map is acquired, determining to acquire the specified operation of the target video abstract; or,
and displaying the map display interface, and determining to acquire the specified operation of the target video abstract when the target video abstract is dragged into a display area of the map display interface.
In another possible implementation manner, the displaying the target video summary on the target electronic map according to the movement track, the first playing window, and the at least one second playing window includes:
sequentially playing the corresponding video clips in the first playing window and the at least one second playing window according to the time sequence of a first time and at least one second time, wherein the first time is the time for generating the first video clip, and the at least one second time is the time for generating the at least one second video clip;
when any video segment in the target video abstract is played, displaying a moving track between a position corresponding to the video segment and a position of a next video segment of the video segment on the target electronic map; or in the process of playing the video clip, displaying a moving track between a position corresponding to the video clip and a position of a next video clip of the video clip on the target electronic map.
In another possible implementation manner, the displaying the target video summary on the target electronic map according to the moving track, the first playing window and the at least one second playing window includes:
displaying the moving track on the target electronic map;
and sequentially playing the corresponding video clips in the first playing window and the at least one second playing window according to the time sequence of a first time and at least one second time, wherein the first time is the time for generating the first video clip, and the at least one second time is the time for generating the at least one second video clip.
In another possible implementation, the movement trajectory includes a plurality of sub-trajectories; one sub-track is a track from the position of one monitoring device to the position of another monitoring device; the method further comprises the following steps:
determining a target playing window of a currently played video clip;
determining first target monitoring equipment corresponding to the target playing window;
selecting a target sub-track from the moving track to a second target monitoring device, wherein the second target monitoring device is a monitoring device corresponding to a next video clip of the currently played video clip;
highlighting the target sub-track on the target electronic map.
In another possible implementation manner, the method further includes:
determining each frame of image comprising the target object, and marking the target object in each frame of image; and/or the presence of a gas in the atmosphere,
determining each frame of image including the target object, determining a place where the target object appears in each frame of image, acquiring position information of the place, and labeling the position information in each frame of image.
In another possible implementation manner, after the displaying the target video summary on the target electronic map according to the moving track, the first playing window and the at least one second playing window, the method further includes:
when the target electronic map is zoomed, determining the zoom scale of the target electronic map;
simultaneously zooming the target electronic map, the first playing window and the at least one second playing window according to the zooming scale of the target electronic map; or zooming the target electronic map according to the zooming scale of the target electronic map, and keeping the first playing window and the at least one second playing window unchanged.
In another aspect, an apparatus for generating a video summary is provided, the apparatus including:
the first acquisition module is used for acquiring a first monitoring video of first monitoring equipment and extracting a target object in the first monitoring video;
the second acquisition module is used for determining a plurality of second monitoring devices in a target area according to the first position of the first monitoring device and acquiring a plurality of second monitoring videos of the plurality of second monitoring devices;
a third obtaining module, configured to obtain at least one second surveillance video from the plurality of second surveillance videos according to the target object, where each second surveillance video includes the target object;
a generating module, configured to generate a first video segment corresponding to the target object according to the first monitoring video, and generate at least one second video segment corresponding to the target object according to the at least one second monitoring video;
a composing module for composing the first video segment and the at least one second video segment into a target video summary of the target object.
In a possible implementation manner, the composition module is further configured to obtain a first time at which the first video segment is generated and at least one second time at which the at least one second video segment is generated; composing the first video clip and the at least one second video clip into the target video summary in chronological order of the first time and the at least one second time; or,
the composition module is further configured to obtain a third time when the target object appears in the first surveillance video, and obtain each fourth time when the target object appears in each second surveillance video; and composing the first video segment and the at least one second video segment into the target video summary according to the time sequence of the third time and each fourth time.
In another possible implementation manner, the third obtaining module is further configured to determine, according to a third time when the target object appears in the first monitored video, a target time period including the third time; and selecting at least one second monitoring video which comprises the target object and has a fourth time of appearance in the second monitoring video within the target time period from the plurality of second monitoring videos.
In another possible implementation manner, the generating module is further configured to select, according to the target object, a video segment that includes the target object in the first surveillance video, and use the video segment as the first video segment; or, according to the target object, extracting a plurality of key frames including the target object in the first monitored video, and combining the plurality of key frames into the first video clip.
In another possible implementation manner, the apparatus further includes:
the fourth obtaining module is used for obtaining the map data corresponding to the target area according to the identification of the target video abstract when the specified operation of the target video abstract is obtained;
the rendering module is used for rendering the map data corresponding to the target area in a map display interface to obtain a target electronic map;
and the display module is used for displaying the moving track of the target object on the target electronic map and displaying the target video abstract on the target electronic map.
In another possible implementation manner, the display module is further configured to display a first playing window corresponding to the first video clip in a target electronic map according to the first position, and display at least one second playing window corresponding to the at least one second video clip in the target electronic map according to at least one second position, where the at least one second position is a position of at least one second monitoring device corresponding to the at least one second monitoring video; and displaying the target video abstract on the target electronic map according to the moving track, the first playing window and the at least one second playing window.
In another possible implementation manner, the presentation module is further configured to sort a third time and at least one fourth time according to the third time and the at least one fourth time to obtain a time sequence, where the third time is a time when the target object appears in the first surveillance video, and the at least one fourth time is a time when the target object appears in the at least one second surveillance video; determining the position of the monitoring equipment corresponding to each two adjacent times from the first position and at least one second position according to each two adjacent times in the time sequence, wherein the at least one second position is the position of at least one second monitoring equipment corresponding to the at least one second monitoring video; connecting the positions of the monitoring equipment corresponding to every two adjacent times according to a target connection mode to obtain a moving track of the target object between the positions of the monitoring equipment corresponding to every two adjacent times until each of the first position and the at least one second position is connected with other positions except the positions to obtain the moving track of the target object; and displaying the movement track of the target object on the target electronic map.
In another possible implementation manner, the apparatus further includes:
the first determining module is used for displaying the map display interface, the map display interface comprises a calling interface, when the calling interface is triggered, a selection interface is displayed, the selection interface comprises a plurality of video abstract identifiers, and when the selection operation of the target video abstract identifier is obtained, the specified operation of the target video abstract is determined to be obtained; or,
the first determining module is further configured to display a playlist when the triggering operation on the target video abstract is acquired, where the playlist includes an identifier of an electronic map; when the selection operation of the identifier of the electronic map is acquired, determining to acquire the specified operation of the target video abstract; or,
the first determining module is further configured to display the map display interface, and determine to acquire a specified operation on the target video abstract when the target video abstract is dragged into a display area of the map display interface.
In another possible implementation manner, the presentation module is further configured to sequentially play the corresponding video segments in the first playing window and the at least one second playing window according to a time sequence of a first time and at least one second time, where the first time is a time for generating the first video segment, and the at least one second time is a time for generating the at least one second video segment; when any video segment in the target video abstract is played, displaying a moving track between a position corresponding to the video segment and a position of a next video segment of the video segment on the target electronic map; or displaying a moving track between a position corresponding to the video clip and a position of a next video clip of the video clip on the target electronic map in the playing process of the video clip.
In another possible implementation manner, the display module is further configured to display the movement track on the target electronic map; and sequentially playing the corresponding video clips in the first playing window and the at least one second playing window according to the time sequence of a first time and at least one second time, wherein the first time is the time for generating the first video clip, and the at least one second time is the time for generating the at least one second video clip.
In another possible implementation, the movement trajectory includes a plurality of sub-trajectories; one sub-track is a track from the position of one monitoring device to the position of another monitoring device; the device further comprises:
the second determining module is used for determining a target playing window of the currently played video clip;
the third determining module is used for determining the first target monitoring equipment corresponding to the target playing window;
a selection module, configured to select a target sub-track from the moving tracks, where the target sub-track is from the first target monitoring device to a second target monitoring device, and the second target monitoring device is a monitoring device corresponding to a video clip next to the currently played video clip;
and the highlighting module is used for highlighting the target sub-track on the target electronic map.
In another possible implementation manner, the apparatus further includes:
the marking module is used for determining each frame of image comprising the target object and marking the target object in each frame of image; and/or determining each frame of image comprising the target object, determining the place where the target object appears in each frame of image, acquiring the position information of the place, and marking the position information in each frame of image.
In another possible implementation manner, the apparatus further includes:
the fourth determining module is used for determining the scaling of the target electronic map when the target electronic map is zoomed;
the zooming module is used for zooming the target electronic map, the first playing window and the at least one second playing window simultaneously according to the zooming scale of the target electronic map; or zooming the target electronic map according to the zooming scale of the target electronic map, and keeping the first playing window and the at least one second playing window unchanged.
In another aspect, a computer device is provided, the computer device comprising:
a processor and a memory, the memory having stored therein at least one instruction, at least one program, set of codes, or set of instructions, the instruction, the program, the set of codes, or the set of instructions being loaded and executed by the processor to implement the operations performed by any of the above video summary generation methods.
In another aspect, a computer-readable storage medium is provided, in which at least one instruction, at least one program, a set of codes, or a set of instructions is stored, which is loaded and executed by a processor to implement the operations performed by any one of the above video summary generation methods.
The technical scheme provided by the embodiment of the application has the following beneficial effects:
according to the video abstract generation method provided by the embodiment of the application, a target object in a first monitoring video is extracted by acquiring the first monitoring video of first monitoring equipment; determining a plurality of second monitoring devices in the target area according to the first position of the first monitoring device, and acquiring a plurality of second monitoring videos of the plurality of second monitoring devices; and acquiring at least one second monitoring video from the plurality of second monitoring videos according to the target object, wherein each second monitoring video comprises the target object. Generating a first video clip corresponding to the target object according to the first monitoring video, and generating at least one second video clip corresponding to the target object according to the at least one second monitoring video; and combining the first video segment and the at least one video segment into a target video abstract. According to the method, the surveillance video of the first surveillance device is obtained, the surveillance videos of the second surveillance devices in the target area of the first surveillance device are also obtained, at least one second surveillance video including the target object is obtained from the second surveillance videos, and therefore the generated target video abstract includes the target object in the first video segment and each second video segment in the at least one second video segment includes the target object.
Drawings
FIG. 1 is a schematic diagram of an implementation environment for generating a video summary according to an embodiment of the present application;
fig. 2 is a flowchart of a video summary generation method according to an embodiment of the present application;
fig. 3 is a flowchart of a video summary generating method according to an embodiment of the present application;
FIG. 4 is a schematic diagram illustrating a method for determining a plurality of second monitoring devices according to an embodiment of the present application;
fig. 5 is a schematic diagram illustrating a moving track and a playing window displayed on a target electronic map according to an embodiment of the present application;
fig. 6 is a schematic diagram illustrating another example of displaying a moving track and a playing window on a target electronic map according to the present disclosure;
fig. 7 is a schematic structural diagram of a video summary generation apparatus according to an embodiment of the present application;
fig. 8 is a block diagram of a computer device according to an embodiment of the present disclosure.
Detailed Description
In order to make the technical solutions and advantages of the present application more clear, the following describes the embodiments of the present application in further detail.
The embodiment of the present application provides an implementation environment for generating a video summary, and referring to fig. 1, the implementation environment includes: a computer device 101 and a plurality of monitoring devices 102. The computer device 101 and the plurality of monitoring devices 102 may be wirelessly connected. The monitoring device 102 may be disposed above a road traffic intersection or at a designated location of a designated site for monitoring a target object. The monitoring device 102 may send the computer device 101 the monitoring video it captured. The computer device 101 is configured to receive the surveillance video sent by the surveillance device 102, and generate a target video summary according to the surveillance video.
The computer device 101 may generate the target video summary upon receiving the surveillance video sent by the surveillance device 102. Or, the computer device 101 may also generate the target video summary when the relevant person searches for the target object, so that the generated target video summary is a video summary including the target object, and is more targeted.
In the related art, when a relevant person searches a target object from a surveillance video, a video abstract corresponding to a surveillance video of the current surveillance device 102 is mainly generated. However, the amount of information contained in the video summary corresponding to one monitoring device 102 is limited, and the amount of information acquired from the video summary by related personnel is small, so that the searching efficiency of the target object is low. In practical applications, related personnel often need to view video summaries of multiple monitoring devices 102 to obtain more information of a target object. Therefore, according to the solutions in the related art, the relevant personnel can only view the video abstract of each monitoring device 102 one by one, but each video abstract does not necessarily include the target object, and therefore, the method for finding the target object is time-consuming and labor-consuming, and the work efficiency is low.
In the embodiment of the present application, when the computer device 101 generates the video abstract, not only the first video segment including the target object but also at least one second video segment including the target object is generated, and the first video segment and the at least one second video segment form the target video abstract, so that each video segment in the target video abstract includes the target object, and the target video abstract includes abundant information, and can expand the dimension and range for searching the target object by related people, thereby improving the work efficiency. The first video segment is a video abstract of the first monitoring video, and the second video segment is a video abstract of the second monitoring video. In addition, in this embodiment of the application, after the computer device 101 generates the target video abstract, the target video abstract may be directly played or displayed on the target electronic map according to the requirement of the relevant person, which is not specifically limited in this embodiment of the application. The target object may be a vehicle, a person, a pet, or an article, among others.
In addition, the video abstract generation method provided by the application can be widely applied to various practical application scenes, and the video abstract generation method provided by the application is explained by combining several exemplary application scenes as follows:
(1) The method is applied to a license plate query scene, and a target object in the scene is a vehicle.
When a traffic accident occurs on a certain road, after the accident occurs, a vehicle owner drives the vehicle to leave the accident site, at the moment, related personnel can call a first monitoring video which is closest to the accident site and comprises a first monitoring device 102 of the vehicle, and at least one second monitoring video which comprises the vehicle in a target area near the first monitoring device 102 through a computer device 101, respectively generate a first video clip comprising the vehicle and at least one second video clip comprising the vehicle, and combine the first video clip and the at least one second video clip into a target video summary. And searching relevant information of the vehicle, such as the license plate number of the vehicle or the running track of the vehicle, from the target video abstract, and further locking the owner of the vehicle according to the license plate number or stopping the vehicle at a relevant position according to the running track of the vehicle.
(2) The method is applied to searching and tracking scenes, and the target object in the scene is a designated person or a pet.
When the related person tracks the designated person or pet, a first monitoring video of a first monitoring device 102 where the designated person or pet appears and at least one second monitoring video including the designated person or pet in a target area near the first monitoring device 102 can be called through a computer device 101, a first video segment including the designated person or pet and at least one second video segment including the designated person or pet are respectively generated, and the first video segment and the at least one second video segment form a target video abstract. Searching relevant information of the appointed person or the pet from the target video abstract, for example, an activity track of the appointed person or the pet, so as to track the appointed person according to the activity track; and searching the pet according to the activity track of the pet.
In one possible implementation, the application scenario may also be to find a pet owner or an article owner. When a person picks up a pet or an article, and finds the owner of the pet or the owner of the article, the related person can find the owner of the pet or the owner of the article from the target video abstract according to the method.
(3) The method is applied to finding a scene of lost articles, and target objects in the scene are articles.
According to the location where the article is lost found by the owner, the related personnel call the first surveillance video which is closest to the location and comprises the first surveillance device 102 of the lost article, and at least one second surveillance video which comprises the lost article in the target area near the first surveillance device 102, respectively generate a first video clip comprising the lost article and at least one second video clip comprising the lost article, and combine the first video clip and the at least one second video clip into a target video summary. And searching the relevant information of the lost item, such as the lost position of the lost item or the current position of the lost item, from the target video abstract, so as to search the item according to the lost position or the current position.
It should be noted that the video summary generating method may also be applied to other scenarios, and this is not specifically limited in this embodiment of the present application.
In one possible implementation, the computer device 101 may be a terminal or a server. In the embodiment of the present application, the computer apparatus 101 is not particularly limited. When the computer device 101 is a terminal, the terminal acquires a plurality of surveillance videos including a target object from the plurality of surveillance devices 102, and generates a target video summary of the target object according to the plurality of surveillance videos. When the computer device 101 is a server, the implementation environment further includes: a terminal and a plurality of monitoring devices 102. The terminal is provided with an application program client for generating the video abstract, and the server is a background server of the application program client. The terminal logs in the server through the application program client and communicates with the server. The server acquires a plurality of monitoring videos including a target object from a plurality of monitoring devices 102, generates a target video abstract of the target object according to the plurality of monitoring videos, and sends the target video abstract to the terminal. In the embodiment of the present application, a computer device 101 is taken as an example for description.
An embodiment of the present application provides a video summary generating method, referring to fig. 2, the method includes:
step 201: acquiring a first monitoring video of first monitoring equipment, and extracting a target object in the first monitoring video;
step 202: according to the first position of the first monitoring device, a plurality of second monitoring devices in the target area are determined, and a plurality of second monitoring videos of the plurality of second monitoring devices are obtained.
Step 202: acquiring at least one second monitoring video from a plurality of second monitoring videos according to the target object, wherein each second monitoring video comprises the target object;
step 203: generating a first video clip corresponding to the target object according to the first monitoring video, and generating at least one second video clip corresponding to the target object according to at least one second monitoring video;
step 204: and combining the first video clip and the at least one second video clip into a target video abstract of the target object.
In one possible implementation, composing a first video segment and at least one second video segment into a target video summary of a target object includes:
acquiring a first time for generating a first video clip and at least one second time for generating at least one second video clip; forming a target video summary of the first video clip and the at least one second video clip according to the time sequence of the first time and the at least one second time; or,
acquiring third time when the target object appears in the first monitoring video and acquiring each fourth time when the target object appears in each second monitoring video; and composing the first video clip and the at least one second video clip into a target video summary according to the time sequence of the third time and each fourth time.
In another possible implementation manner, acquiring at least one second surveillance video from a plurality of second surveillance videos according to a target object includes:
determining a target time period comprising a third time according to the third time when the target object appears in the first monitoring video;
and selecting at least one second monitoring video which comprises the target object and has the fourth time of the target object appearing in the second monitoring video within the target time period from the plurality of second monitoring videos.
In another possible implementation manner, generating a first video segment corresponding to the target object according to the first monitoring video includes:
selecting a video clip including the target object in the first monitoring video according to the target object, and taking the video clip as a first video clip; or,
according to the target object, a plurality of key frames including the target object in the first monitoring video are extracted, and the plurality of key frames form a first video clip.
In another possible implementation, after the first video segment and the at least one second video segment are combined into the target video summary of the target object, the method further includes:
when the specified operation of the target video abstract is obtained, obtaining map data corresponding to a target area according to the identification of the target video abstract;
rendering map data corresponding to the target area in a map display interface to obtain a target electronic map;
and displaying the moving track of the target object on the target electronic map, and displaying the target video abstract on the target electronic map.
In another possible implementation, displaying a target video summary on a target electronic map includes:
displaying a first playing window corresponding to the first video clip in the target electronic map according to the first position, and displaying at least one second playing window corresponding to at least one second video clip in the target electronic map according to at least one second position, wherein the at least one second position is the position of at least one second monitoring device corresponding to at least one second monitoring video;
and displaying the target video abstract on the target electronic map according to the moving track, the first playing window and the at least one second playing window.
In another possible implementation manner, displaying a moving track of a target object on a target electronic map includes:
sequencing the third time and at least one fourth time according to the third time and the at least one fourth time to obtain a time sequence, wherein the third time is the time when the target object appears in the first monitoring video, and the at least one fourth time is the time when the target object appears in at least one second monitoring video;
determining the position of the monitoring equipment corresponding to each two adjacent times from the first position and the at least one second position according to each two adjacent times in the time sequence, wherein the at least one second position is the position of at least one second monitoring equipment corresponding to at least one second monitoring video;
connecting the positions of the monitoring equipment corresponding to every two adjacent times according to a target connection mode to obtain a moving track of the target object between the positions of the monitoring equipment corresponding to every two adjacent times until each of the first position and the at least one second position is connected with other positions except the positions to obtain the moving track of the target object;
and displaying the moving track of the target object on the target electronic map.
In another possible implementation, the method further includes:
displaying a map display interface, wherein the map display interface comprises a calling interface, when the calling interface is triggered, a selection interface is displayed, the selection interface comprises a plurality of video abstract identifications, and when the selection operation of the target video abstract identifications is obtained, the specified operation of the target video abstract is determined to be obtained; or,
when the triggering operation of the target video abstract is acquired, displaying a play list, wherein the play list comprises an identifier of the electronic map; when the selection operation of the identifier of the electronic map is acquired, determining to acquire the specified operation of the target video abstract;
and displaying a map display interface, and determining to acquire the specified operation of the target video abstract when the target video abstract is dragged into a display area of the map display interface.
In another possible implementation manner, displaying a target video summary on a target electronic map according to a moving track, a first playing window and at least one second playing window includes:
according to the time sequence of the first time and the at least one second time, playing the corresponding video clips in the first playing window and the at least one second playing window in sequence, wherein the first time is the time for generating the first video clip, and the at least one second time is the time for generating the at least one second video clip;
when any video segment in the target video abstract is played, displaying a moving track between a position corresponding to the video segment and a position of a next video segment of the video segment on the target electronic map; or, in the process of playing the video clip, displaying a moving track between the position corresponding to the video clip and the position of the next video clip of the video clip on the target electronic map.
In another possible implementation manner, the displaying the target video summary on the target electronic map according to the moving track, the first playing window and the at least one second playing window includes:
displaying a moving track on the target electronic map;
and according to the time sequence of the first time and the at least one second time, sequentially playing the corresponding video clips in the first playing window and the at least one second playing window, wherein the first time is the time for generating the first video clip, and the at least one second time is the time for generating the at least one second video clip.
In another possible implementation, the movement trajectory includes a plurality of sub-trajectories; one sub-track is a track from the position of one monitoring device to the position of another monitoring device; the method further comprises the following steps:
determining a target playing window of a currently played video clip;
determining first target monitoring equipment corresponding to a target playing window;
selecting a target sub-track from the moving track to a second target monitoring device from a first target monitoring device, wherein the second target monitoring device is a monitoring device corresponding to a next video clip of the currently played video clip;
and highlighting the target sub-track on the target electronic map.
In another possible implementation manner, the method further includes:
determining each frame of image including a target object, and marking the target object in each frame of image; or,
determining each frame of image including the target object, determining the place where the target object appears in each frame of image, acquiring the position information of the place, and marking the position information in each frame of image.
In another possible implementation manner, after the target video summary is displayed on the target electronic map according to the moving track, the first playing window and the at least one second playing window, the method further includes:
when the target electronic map is zoomed, determining the zoom scale of the target electronic map;
simultaneously zooming the target electronic map, the first playing window and the at least one second playing window according to the zooming scale of the target electronic map; or zooming the target electronic map according to the zooming scale of the target electronic map, and keeping the first playing window and the at least one second playing window unchanged.
According to the video abstract generation method provided by the embodiment of the application, a target object in a first monitoring video is extracted by acquiring the first monitoring video of first monitoring equipment; determining a plurality of second monitoring devices in the target area according to the first position of the first monitoring device, and acquiring a plurality of second monitoring videos of the plurality of second monitoring devices; and acquiring at least one second monitoring video from the plurality of second monitoring videos according to the target object, wherein each second monitoring video comprises the target object. Generating a first video clip corresponding to the target object according to the first monitoring video, and generating at least one second video clip corresponding to the target object according to at least one second monitoring video; and composing the first video segment and the at least one video segment into a target video abstract. According to the method, not only the monitoring video of the first monitoring device is obtained, but also the monitoring videos of the plurality of second monitoring devices in the target area of the first monitoring device are obtained, and at least one second monitoring video including the target object is obtained from the plurality of second monitoring videos, so that the generated target video abstract includes the target object in the first video segment, and each second video segment in the at least one second video segment also includes the target object.
An embodiment of the present application provides a video summary generation method, which is applied to a computer device, and with reference to fig. 3, the method includes:
step 301: the computer equipment acquires a first monitoring video of the first monitoring equipment and extracts a target object in the first monitoring video.
In this step, the first monitoring video may be a video currently shot by the first monitoring device, or may be a video that is sent to the computer device by the first monitoring device in advance, and the computer device stores the first monitoring video in a video library; alternatively, the first monitoring video may also be a video shot by the first monitoring device in advance and stored locally. In the embodiment of the present application, the first surveillance video is not particularly limited. And, the first surveillance video is a video including the target object.
In a possible implementation manner, when the first monitoring video is a video currently shot by the first monitoring device, the computer device sends an obtaining instruction to the first monitoring device, the first monitoring device receives the obtaining instruction and directly sends the first monitoring video to the computer device, and the computer device obtains the first monitoring video sent by the first monitoring device.
In another possible implementation manner, when the first surveillance video is a video that is pre-stored in a video library by the computer device, the computer device obtains the first surveillance video from the video library. Correspondingly, the step of the computer device searching the first monitoring video from the video library may be: the computer equipment receives a query instruction, wherein the query instruction carries the position identification of the specified geographic position and the target characteristics of the target object and is used for searching the target object of the specified geographic position. And the computer equipment searches the monitoring video of the object which corresponds to the position identification and has the object according with the target characteristic from the video library according to the position identification and the target characteristic.
Or the computer equipment receives a query instruction, wherein the query instruction carries the position identification of the specified geographic position and the image of the target object and is used for searching the target image of the specified geographic position. And the computer equipment extracts the image characteristics in the image according to the image of the target object, and searches the monitoring video which corresponds to the position identification and has the image according with the image characteristics from the video library according to the position identification and the image characteristics.
Or the computer equipment receives an inquiry instruction, wherein the inquiry instruction carries the position identification of the specified geographic position, the specified time information of the specified time and the target characteristics of the target object, and is used for searching the target object which appears at the specified time and corresponds to the specified geographic position. And the computer equipment searches for the monitoring video which corresponds to the position identification, the appointed time information and the object which accords with the target characteristic from the video library according to the position identification, the appointed time information and the target characteristic.
In another possible implementation manner, when the first surveillance video is a video that is shot by the first surveillance device in advance and is stored locally, the computer device may obtain the first surveillance video according to the video identifier, and accordingly, the step of obtaining the first surveillance video of the first surveillance device by the computer device may be: the computer equipment sends an acquisition instruction to the first monitoring equipment, wherein the acquisition instruction carries the video identification and the target characteristics of the target object and is used for acquiring the first monitoring video. And the first monitoring device receives the acquisition instruction, searches for a first monitoring video which corresponds to the video identifier and has an object conforming to the target characteristic from the stored multiple monitoring videos, and sends the first monitoring video to the computer device. Or, the computer device may further obtain the first monitoring video according to the specified time information and the target feature, and correspondingly, the step may be: the computer equipment sends an acquisition instruction to the first monitoring equipment, wherein the acquisition instruction carries a video identifier, designated time information and target characteristics and is used for acquiring the first monitoring video. And the first monitoring equipment receives the acquisition instruction, searches for a first monitoring video which corresponds to the video identifier, is corresponding to the designated time information and has an object conforming to the target characteristic from a plurality of stored monitoring videos, and sends the first monitoring video to the computer equipment.
After the computer equipment acquires the first monitoring video of the first monitoring equipment, the target object in the first monitoring video is extracted. The target object may be a person, a vehicle, a pet, or an item. In the embodiment of the present application, the target object is not particularly limited. The computer device may extract the target object in the first surveillance video through the extraction model. For example, when the target object is a person, the extraction model may be a face recognition model; when the target object is a vehicle, the extraction model may be a license plate recognition model. In the embodiment of the present application, the extraction model is not particularly limited.
Step 302: the computer equipment determines a plurality of second monitoring equipment in the target area according to the first position of the first monitoring equipment, and obtains a plurality of second monitoring videos of the plurality of second monitoring equipment.
In a possible implementation manner, the computer device may determine, according to the first location of the first monitoring device, a plurality of second monitoring devices in the target area by taking the first monitoring device as a center. The first position of the first monitoring device may be a position of the first monitoring device on the electronic map, or may be a geographical position of the first monitoring device in actual road traffic. In the embodiments of the present application, this is not particularly limited. In this implementation manner, the second monitoring device determined by the computer device is the second monitoring device in the target area, so that blind searching for the target object can be avoided, the target object can be searched within a certain range, and the searching efficiency for the target object is improved.
It should be noted that when the target area is large, the number of the monitoring devices in the target area is large, and directly obtaining the monitoring videos of the multiple monitoring devices results in large workload for subsequently searching the target object and low searching efficiency.
Therefore, in another possible implementation manner, the computer device may further determine, based on the first location of the first monitoring device, a plurality of second monitoring devices associated with the first monitoring device in the target area, centering on the first monitoring device. Wherein the plurality of second monitoring devices and the first monitoring device associated with the first monitoring device may be monitoring devices controlled by one monitoring platform; or the plurality of second monitoring devices associated with the first monitoring device and the first monitoring device are monitoring devices in the same area. The same area may be a scenic spot, a mall, or an urban area of a city. Alternatively, the plurality of second monitoring devices associated with the first monitoring device are monitoring devices that establish a communication connection with the first monitoring device. In the embodiments of the present application, this is not particularly limited. In the implementation manner, the workload of subsequently searching for the target object can be reduced to a certain extent by determining the second monitoring device associated with the first monitoring device, and the searching efficiency is improved.
The size of the target area may be set and changed as needed, and in the embodiment of the present application, the size of the target area is not specifically limited. For example, the target area is 1 km, 2 km or 3 km. For example, referring to fig. 4, the computer device in fig. 4 uses the first monitoring device as a center, determines that within 2 kilometers of the first monitoring device, there are 7 second monitoring devices associated with the first monitoring device, and acquires second monitoring videos of the 7 second monitoring devices.
Step 303: and the computer equipment acquires at least one second monitoring video from the plurality of second monitoring videos according to the target object, wherein each second monitoring video comprises the target object.
In one possible implementation, the present step can be implemented by the following steps (1) to (2), including:
(1) And the computer equipment determines a target time period comprising a third time according to the third time when the target object appears in the first monitoring video.
When the computer device extracts the target object from the first surveillance video in step 301, a first time at which the target object appears in the first surveillance video is recorded. In this step, the computer device obtains the first time and determines a target time period including the first time.
For example, the first time that the target object appears in the first surveillance video is 11, the target time period may be 9.
(2) The computer device selects at least one second surveillance video, which includes the target object and in which a fourth time that the target object appears in the second surveillance video is within the target time period, from among the plurality of second surveillance videos.
This step can be realized by the following steps (2-1) to (2-2), including:
(2-1) the computer device selects a plurality of second surveillance videos including the target object from the plurality of second surveillance videos.
The method comprises the following steps: and the computer equipment performs feature extraction on the target object to obtain the target feature of the target object. For each second monitoring video, the computer equipment determines whether an object meeting the target characteristics exists in the second monitoring video according to the target characteristics of the target object. When an object meeting the target characteristic exists, determining that the second monitoring video comprises the target object; and when the object meeting the target characteristic does not exist, determining that the second monitoring video does not comprise the target object, thereby obtaining at least one second monitoring video comprising the target object.
(2-2) the computer device selects at least one second surveillance video within the target time period at a fourth time when the target object appears in the second surveillance video, from the plurality of second surveillance videos including the target object.
In this step, for each of a plurality of second surveillance videos including the target object, the computer device determines a time when the target object appears in the second surveillance video. And the computer equipment determines whether the time is in a target time period or not according to the time when the target object appears in the second monitoring video, and determines that the second monitoring video comprises the target object and the monitoring video of the fourth time, appearing in the second monitoring video, of the target object in the target time period when the time is in the target time period, so that at least one second monitoring video is obtained.
For example, the target time period is 9.
For example, the third time when the target object appears in the first surveillance video is 11, the target time period is 10.
It should be noted that, when the target object appears in the monitoring video of other monitoring devices, the computer device may determine a distance between the monitoring device corresponding to the other monitoring video and the first monitoring device, and expand the target area according to the distance, thereby obtaining the monitoring video in the target area. For example, the preset target area is 2 kilometers, when the target object appears in other surveillance videos, the computer device determines that the distance between the first surveillance device and the surveillance device corresponding to the other surveillance videos is 0.5 kilometer, and then the computer device may expand the target area to 3 kilometers according to the distance, and further obtain the surveillance videos within the range of 3 kilometers.
In this step, at least one second surveillance video acquired by the computer device is a surveillance video including the target object and the fourth time of the target object appearing in the second surveillance video is within the target time period, so that the user is more specific when searching for the target object, and the searching efficiency is improved.
Step 304: and the computer equipment generates a first video clip corresponding to the target object according to the first monitoring video and generates at least one second video clip corresponding to the target object according to the at least one second monitoring video.
In this step, the first video segment is a video abstract of the first surveillance video, and the second video segment is a video abstract of the second surveillance video.
Wherein the computer device may generate the first video segment by any one of the following two implementations. In a first implementation manner, the computer device selects a video segment including the target object in the first surveillance video according to the target object, and uses the video segment as the first video segment. In a second implementation manner, the computer device extracts a plurality of key frames including the target object from the first monitored video according to the target object, and combines the plurality of key frames into the first video clip. In the embodiment of the present application, the manner in which the computer device generates the first video segment is not particularly limited. In the implementation mode, the generated first video clip is a video clip including the target object, so that other irrelevant people or objects can be prevented from appearing in the first video clip, and the searching efficiency of the target object is improved.
For each second video segment, the manner in which the computer device generates the second video segment and the manner in which the computer device generates the first video segment may be the same or different, and in this embodiment, the manner in which the computer device generates the second video segment is not particularly limited.
Step 305: the computer device composes the first video segment and the at least one second video segment into a target video summary of the target object.
In this step, the computer device may obtain the target video summary of the target object through any one of the following two implementation manners.
In a first implementation, a computer device obtains a first time at which a first video segment is generated and at least one second time at which at least one second video segment is generated; and forming the first video clip and the at least one second video clip into a target video summary of the target object according to the time sequence of the first time and the at least one second time.
In step 304, the computer device records a first time at which the first video segment was generated when the first video segment was generated and a second time at which each of the second video segments was generated when the at least one second video segment was generated.
In this implementation, the computer device obtains the first time and the at least one second time, and forms the first video segment and the at least one second video segment into a target video summary of the target object according to the time sequence of the first time and the at least one second time from first to last, so that the user can subsequently know the generation sequence of the video segments according to the playing sequence of the target video summary.
For example, the number of the at least one second surveillance video is 2, the time of the target object appearing in the first surveillance video is 11. The target object generates a first video segment, which is recorded as video segment 1, then generates a video segment of a first second surveillance video, which is recorded as video segment 2, and finally generates a video segment of a second surveillance video, which is recorded as video segment 3. The computer device composes a target video summary in the order of video segments 1, 2, 3.
In a second implementation manner, the computer device acquires a third time when the target object appears in the first monitoring video and acquires a fourth time when the target object appears in the second monitoring video; and according to the time sequence of the third time and each fourth time, the first video clip and the at least one second video clip are combined into a target video abstract of the target object.
In step 301, when the computer device extracts the target object from the first surveillance video, recording a third time when the target object appears in the first surveillance video; in step 303, the at least one second surveillance video acquired by the computer device is a surveillance video including a target object, and the computer device may check the time when the target object appears in each second surveillance video to obtain at least one fourth time. The computer equipment acquires the third time and the at least one fourth time, and the first video clip and the at least one second video clip form the target video summary of the target object according to the time sequence of the third time and the at least one fourth time from first to last, so that the video clips in the target video summary can be played sequentially according to the time sequence of the target object in the surveillance video, and a user can know the sequence of the target object in the surveillance video and know the moving track of the target object.
For example, the number of the at least one second surveillance video is 2, the time of the target object appearing in the first surveillance video is 11. The target object generates a first video segment, which is recorded as video segment 1, then generates a video segment of a first second surveillance video, which is recorded as video segment 2, and finally generates a video segment of a second surveillance video, which is recorded as video segment 3. The computer device composes a target video summary in the order of video segments 2, 1, 3.
In the embodiment of the application, when the computer device generates the target video abstract, the computer device not only generates the video abstract of the first surveillance video, namely the first video segment, but also generates the video abstract of at least one second surveillance video including the target object, namely the at least one second video segment, which is beneficial for a user to acquire the surveillance information of the target object in the surveillance video from multiple dimensions, so that the computer device plays the maximum role of video surveillance in some practical application scenes.
It should be noted that, after the target video summary is generated, the computer device may send the target video summary to the user or end the operation, or directly play the target video summary, or display the target video summary on the electronic map. Accordingly, when the computer device presents the target video summary on the electronic map, the computer device performs step 306.
Step 306: when the computer equipment acquires the specified operation of the target video abstract, the computer equipment acquires the map data corresponding to the target area according to the identification of the target video abstract.
In one possible implementation manner, the computer device stores a correspondence between an identifier of the video summary and map data, where the map data is data corresponding to an area included in the video summary. When the computer equipment obtains the designated operation of the target video abstract, the computer equipment obtains the identifier of the target video abstract, and according to the identifier, the map data corresponding to the identifier is determined from the corresponding relation between the identifier of the video abstract and the map data, so that the map data corresponding to the target area is obtained.
The specifying operation may be preset, for example, the specifying operation may be a click operation, a double click operation, a slide operation, or the like. In the embodiment of the present application, the specified operation is not particularly limited.
It should be noted that, before this step, the computer device may determine to acquire the specified operation through any one of the following implementation manners.
In a first implementation, a computer device displays a map display interface, where the map display interface includes a call interface, and when the call interface is triggered, a selection interface is displayed, where the selection interface includes identifiers of a plurality of video summaries. When the computer device obtains the selection operation of the identifier of the target video abstract, the computer device determines the specified operation of obtaining the target video abstract.
After the computer device generates the target video abstract, the target video abstract can be stored. When a user wants to show the target video abstract on the electronic map, the computer equipment can be triggered to start the electronic map application, and a map showing interface is displayed, wherein the map showing interface comprises a calling interface. When the retrieval interface is triggered, a selection interface can pop up on the map display interface, and the identification of the stored video summaries is displayed in the selection interface. And selecting an identifier corresponding to the target video abstract from the identifiers of the plurality of video abstracts, and determining the specified operation for acquiring the target video abstract when the computer equipment acquires the selection operation of the identifier corresponding to the target video abstract.
And a check box can be displayed in the selection interface before or after each video abstract, and the corresponding video abstract is selected by selecting the check box. Or directly select the identity of the video summary. In addition, the display form of the calling interface in the map display interface can be set and changed according to the needs. For example, the presentation form of the call interface can be a button, a link, or a text. In the embodiments of the present application, this is not particularly limited.
In the related art, after an electronic map application is started, only a certain target position can be input in the electronic map, and then searching is performed. In the embodiment of the application, the retrieval interface is arranged on the map display interface of the electronic map application, and the target video abstract is retrieved through the retrieval interface, so that the target video abstract is displayed in the electronic map application, and the application range of the electronic map application is expanded.
In a second implementation manner, when the computer device obtains a trigger operation on the target video abstract, the computer device displays a playlist, wherein the playlist includes an identifier of the electronic map; when the computer equipment acquires the selection operation of the identifier of the electronic map, determining the designated operation of acquiring the target video abstract.
After the computer device generates the target video summary, the target video summary can be stored to a designated location. When a user operates the target video abstract at a designated position, the computer equipment acquires the trigger operation of the target video abstract and displays a playlist, wherein the playlist can be a playing mode or a display mode of the target video abstract. The playlist may include, in addition to the identification of the electronic map, an identification of other applications, such as an identification of a video application, an identification of an email application, or an identification of a compression application. And the computer equipment determines whether the specified operation on the target video abstract is acquired or not according to the selection operation of acquiring the identifiers of the different applications. When the computer equipment acquires the selection operation of the identifier of the electronic map, the specified operation of acquiring the target video abstract is determined.
In this implementation, the computer device associates the target video summary with the application in the playlist in advance, so that the target video summary can be subsequently operated by the application in the playlist accordingly.
The trigger operation may be set and changed as desired. For example, the trigger operation may be a right click, a long press, or the like. In the embodiments of the present application, this is not particularly limited.
For example, a computer device stores a plurality of video summaries in a folder, which when opened, displays the plurality of video summaries. The user selects a target video abstract from the plurality of video abstracts, right clicks the target video abstract, the computer device displays a playlist, and an identifier of the electronic map, an identifier of the video application, an identifier of the compression application, and the like are displayed in the playlist. The user can trigger different selection operations by selecting the identifiers of different applications, and then select different modes to play or display the target video abstract. For example, when a user clicks on an identification of a mapping application, the target video summary may subsequently be played in the corresponding mapping application identified. When the user clicks the identification of the mail application, the target video abstract can be added into the mail content for sending through the mail application corresponding to the identification. When the user clicks the identifier of the compression application, the target video summary may be subsequently compressed by identifying the corresponding compression application.
In the related art, the target video abstract can only be played through the video application, and in the embodiment of the application, the target video abstract is displayed on the electronic map by associating the target video abstract with the electronic map application, so that not only can a user know the position of a target object in the target video abstract in the electronic map more intuitively, but also the display form of the video abstract is enriched.
In a third implementation manner, the computer device displays a map display interface, and when the target video abstract is dragged into a display area of the map display interface, the computer device determines to acquire the specified operation on the target video abstract.
After the computer device generates the target video summary, the target video summary can be stored to a designated location. When a user wants to show the target video abstract on the electronic map, the computer equipment can be triggered to start the electronic map application, and a map showing interface is displayed. The computer equipment can simultaneously display a storage interface and a map display interface of the target video abstract stored at the designated position, and a user can directly drag the identifier of the target video abstract in the storage interface into the display area of the map display interface, namely when the computer equipment acquires the target video abstract dragged in the display area of the map display interface, the computer equipment determines to acquire the designated operation.
The computer equipment can determine the overlapping area between the dragged target video abstract and the display area, and when the overlapping area is larger than a preset area, the computer equipment determines that the target video abstract is dragged into the display area, and determines to acquire the specified operation.
The implementation mode directly drags the target video abstract into the map display interface, so that the target video abstract can be displayed in the map display interface subsequently, the operation is simple and quick, and brand new experience is brought to a user.
Step 307: and rendering the map data corresponding to the target area in the map display interface by the computer equipment to obtain the target electronic map.
In this step, the computer device renders the map data corresponding to the target area in the map display interface according to the acquired map data corresponding to the target area, so as to obtain the target electronic map. The target electronic map is an electronic map corresponding to the target area.
For example, the first monitoring device is a monitoring device in a certain market, and the target area is an area size corresponding to the market. The computer equipment displays the map corresponding to the market on a map display interface by acquiring the map data of the market, so that the moving track of the target object in the market is acquired. For another example, the first monitoring device is a monitoring device in a certain scenic spot, and the target area is the size of the area corresponding to the scenic spot. The computer equipment displays the map corresponding to the scenic spot on the map display interface by acquiring the map data of the scenic spot, so that the moving track of the target object in the scenic spot can be acquired subsequently.
Step 308: the computer device displays the movement track of the target object on the target electronic map.
In this step, the computer device may determine a movement trajectory of the target object in the target area in advance, and map the movement trajectory of the target object in the target area onto the target electronic map, so as to display the movement trajectory of the target object on the target electronic map.
The computer equipment can determine the movement track of the target object by the following steps: and the computer equipment sequences the third time and the at least one fourth time according to the third time and the at least one fourth time to obtain a time sequence. And determining the position of the monitoring equipment corresponding to each two adjacent times from the first position and the at least one second position according to each two adjacent times in the time sequence. And the computer equipment connects the positions of the monitoring equipment corresponding to every two adjacent times according to a target connection mode to obtain the movement track of the target object between the positions of the monitoring equipment corresponding to every two adjacent times until each of the first position and the at least one second position is connected with other positions except the position to obtain the movement track of the target object. Thereby displaying the moving track of the target object on the target electronic map. The third time is the time when the target object appears in the first monitoring device, and the at least one fourth time is the time when the target object appears in the at least one second monitoring video. The at least one second location is a location of at least one second monitoring device.
For example, the number of the at least one second surveillance video is 2, the third time of occurrence of the target object in the first surveillance video is 11. The computer device determines the position of the second monitoring device corresponding to the first second monitoring video according to the sequence of the third time and the 2 fourth times, then determines the position of the first monitoring device corresponding to the first monitoring video, and connects the two positions to obtain the moving track of the target object between the two positions. And then determining the position of second monitoring equipment corresponding to a second monitoring video, and connecting the position of the first monitoring equipment corresponding to the first monitoring video with the position of the second monitoring equipment corresponding to the second monitoring video to obtain the moving track of the target object between the two positions. And the computer equipment forms the moving track between every two positions into the moving track of the target object.
It should be noted that the first location may be a location of the first monitoring device on the target electronic map, and may also be a geographic location of the first monitoring device in actual road traffic. The second location may also be a location of the second monitoring device on the target electronic map, or a geographic location of the second monitoring device in actual road traffic. In addition, when the computer device displays the movement track of the target object on the target electronic map, the movement track can be displayed according to a default display scale, or the target electronic map can be set, and only the electronic map corresponding to the target area is displayed.
In the related art, the electronic map includes a large range, and a user cannot clearly know the related information of the target object in the target area through the electronic map. In the embodiment of the application, only the electronic map corresponding to the target area is displayed in the map display interface, so that a subsequent user can more clearly and intuitively know information such as a moving track or a place where the target object appears in the target area, and further track or search the target object.
The computer device forms a track route of the target object according to the positions of the first monitoring device and the 5 second monitoring devices in the target electronic map in fig. 4 and the sequence of the target object appearing in the monitoring video. Referring to fig. 5 and 6, respectively, the target electronic map in fig. 5 displays the movement track of the target object in the target area, which is displayed at the default display scale. Referring to fig. 6, only the movement locus of the target object in the target area is displayed in the target electronic map in fig. 6. Wherein the arrow direction indicates the moving direction of the target object.
Step 309: the computer equipment displays the target video abstract on the target electronic map.
In this step, the computer device may display the target video summary on the target electronic map through any one of the following implementation manners.
In a first implementation manner, the computer device may directly display a third playing window near a position corresponding to the first monitoring device or the second monitoring device on the target electronic map, and play each video segment in the target video summary near the third playing window.
In a second implementation, the computer device may further display a third playing window on the target electronic map, and play each video segment in the target video summary in the third playing window.
In a third implementation manner, the computer device may further display a plurality of third playing windows on the target electronic map, where each third playing window plays one video clip in the target video summary. For each third playing window, when the third playing window plays the corresponding video clip, the third playing window may be displayed in an enlarged manner. Or, when any one of the third playing windows is clicked, the clicked third playing window can be displayed in an enlarged manner, so that a user can search the target object from the video clip played in the playing window more clearly, and the searching efficiency is improved.
In a fourth implementation manner, the computer device may display, in the target electronic map, a first play window corresponding to the first video clip according to the first position, and display, in the target electronic map, at least one second play window corresponding to the at least one second video clip according to the at least one second position. And displaying the target video abstract on the electronic map according to the moving track of the target object, the first playing window and at least one second playing window.
In this implementation, when the first location is a location of the first monitoring device on the target electronic map, the computer device displays the first playback window directly near the first location. The first playing window may be located at the left side, the right side, the upper side, or the lower side of the first position, which is not particularly limited in the embodiment of the present application. The first playing window is used for playing the first video clip. When the first position is not the position of the first monitoring device on the target electronic map, the computer device determines the position of the first monitoring device in the target electronic map according to the first position, and displays a first playing window near the position.
In one possible implementation, a play button, a close button, a fast forward button, a rewind button, or the like may be displayed in the first play window. Different buttons will trigger different operations in the first play window. For example, when the play button is triggered for the first time, the first play window may pause playing the first video segment; when the play button is triggered for the second time, the first play window continues to play the first video clip. When the fast forward button is triggered, the first play window increases the speed of playing the first video segment. In addition, a fast-forward speed button can be displayed in the first playing window, and different speed buttons correspond to different fast-forward speeds, so that the playing time can be saved, and the working efficiency of a user can be improved. Each second play window may also display a play button, a close button, a fast forward button, a rewind button, or the like. For example, when the speed button displaying "2" is triggered, the first play window plays the first video clip at 2 times the normal speed; when the speed button displaying "1.5" is triggered, the first play window plays the first video segment at 1.5 times the normal speed.
In a fourth implementation manner, the computer device may display, on the target electronic map, a moving track between a position corresponding to a video segment that is played back and a next position thereof when the video segment that is played back in the playback window is finished, then play the video segment corresponding to the next position, and display, on the target electronic map, a moving track between the next position and a third position when the video segment corresponding to the next position is played back, where the third position is a next position of the next position. Namely the following first mode.
In a first mode, the computer device sequentially plays the video clips corresponding to the first time in the first playing window and the at least one second playing window according to the time sequence of the first time and the at least one second time; when the playing of any video segment in the target video abstract is finished, displaying a moving track between a position corresponding to the video segment and a position of a next video segment of the video segment on the target electronic map. In this way, when the video clip is played, the movement track between the position of the video clip and the position of the next video clip is displayed, and the user can intuitively know the next position where the target object appears.
Or, the computer device may also display, on the target electronic map, a movement track between a position corresponding to the video clip being played and a next position thereof in the process of playing the video clip in the playing window. And after the playing of the video clip being played is finished, playing the next video clip, and displaying the moving track between the next position and the third position on the target electronic map in the process of playing the next video clip. Namely the following second mode.
In a second mode, the computer device plays the corresponding video clips in the first playing window and the at least one second playing window in sequence according to the time sequence of the first time and the at least one second time; and in the video clip playing process, displaying a moving track between the position corresponding to the video clip and the position of the next video clip of the video clip on the target electronic map. In the method, the computer equipment plays the video clip and displays the moving track at the same time, so that the playing window continuously moves along with the moving track, and the method can bring more visual experience to related personnel.
Or, the computer device may also display the moving trajectory of the target object on the target electronic map, and then sequentially play the corresponding video segments according to the time sequence of the first time and the at least one second time. Namely, the following third mode.
In a third mode, a moving track is displayed on a target electronic map; and playing the corresponding video clips in the first playing window and the at least one second playing window in sequence according to the time sequence of the first time and the at least one second time. In the method, the moving track of the target object is firstly displayed on the target electronic map, so that a user can intuitively know the track route of the target object according to the moving track and track or search the target object purposefully. And then playing the corresponding video clips according to the time of the target object appearing in the monitoring video and the time sequence. With continued reference to fig. 5, the playing window shown in fig. 5 is playing a video clip corresponding to the third position in the moving track.
The moving track comprises a plurality of sub-tracks, and one sub-track is a track from the position of one monitoring device to the position of another monitoring device. In this implementation, the computer device may highlight the movement track, and accordingly, the step may be: the computer equipment determines a target playing window of a currently played video clip; determining a first target monitoring device corresponding to the target playing window; selecting a target sub-track from the moving track to a second target monitoring device from the first target monitoring device, wherein the second target monitoring device is a monitoring device corresponding to a next video clip of the currently played video clip; and highlighting the target sub-track on the target electronic map.
The highlighted style may be set and changed as desired. In the embodiments of the present application, this is not particularly limited. For example, the highlighted pattern may be a color deepening, a line thickening, or the like.
Moreover, when the computer device plays the video clip by any one of the three modes, the target object in the video clip can be marked, and/or the position information of the place where the target object appears can be marked. Accordingly, the steps may be: the computer equipment determines each frame of image comprising the target object and marks the target object in each frame of image; and/or the computer equipment determines each frame of image comprising the target object, determines the place where the target object appears in each frame of image, acquires the position information of the place, and marks the position information in each frame of image. By marking the target object on the image, the user can more intuitively view the target object from the video clip. By marking the position information of the place where the target object appears on the image, when the user wants to acquire the position information of the place, the user does not need to search again, and the acquisition efficiency is improved.
In addition, the computer equipment can also generate a navigation interface according to the position information of the place where the user appears, and the navigation interface can display the navigation path between every two places, so that the user can know the moving track of the target object between every two places.
In a possible implementation manner, the display style of the movement track can be preset on the display interface for displaying the target electronic map. And a plurality of candidate display styles of the movement track are displayed in a toolbar in the display interface in advance, a check box is arranged in front of each candidate display style, and the corresponding candidate display style is selected as the display style of the movement track by selecting the check box. For example, the display style may be a line, a graphic, or an arrow. In the embodiments of the present application, this is not particularly limited. Alternatively, the user may customize a plurality of display styles of the movement trajectory.
In another possible implementation manner, the time interval between the playing of two adjacent video clips can be preset on the display interface for displaying the map. A plurality of time intervals are displayed in a toolbar of the display interface in advance, a check box can be arranged before each time interval, and the corresponding time interval is selected as the time interval between two adjacent video clips by selecting the check box. For example, the time interval may be 1 second, 2 seconds, or 3 seconds. Or automatically playing the next video clip after the playing of any video clip is finished. Therefore, the situation that the playing needs manual operation of a user every time can be avoided, and automatic playing is realized.
In another possible implementation manner, an icon corresponding to the position of the first monitoring device and the at least one second monitoring device on the target electronic map may be triggered, and when the icon is triggered, the playing window corresponding to the position plays the video clip corresponding to the icon. For example, when an icon corresponding to the position of any one of the at least one second monitoring device on the target electronic map is triggered, the second playing window corresponding to the triggered icon plays the corresponding second video clip, so that different video clips can be played according to user requirements, and user experience is improved.
In another possible implementation, after or during the presentation of the target video summary on the target electronic map, the computer device may zoom the target electronic map. The playing window can be zoomed along with the zooming of the target electronic map, or the playing window is kept unchanged when the target electronic map is zoomed. Accordingly, the steps may be: when the target electronic map is zoomed, the computer equipment determines the zoom scale of the target electronic map; simultaneously zooming the target electronic map, the first playing window and the at least one second playing window according to the zooming scale of the target electronic map; or zooming the target electronic map according to the zooming scale of the target electronic map, and keeping the first playing window and the at least one second playing window unchanged. In the embodiments of the present application, this is not particularly limited.
When the playing window is zoomed along with the zooming of the target electronic map, a fixed proportion can be kept between the playing window and the target electronic map, and the watching habit of a user is met. When the playing window is not zoomed along with the zooming of the target electronic map and is kept unchanged, the situation that the playing window is too small and the played video segment is unclear when the target electronic map is zoomed down can be avoided; and when the target electronic map is amplified, the situation that the playing window is too large and exceeds a display interface for displaying the target electronic map, so that the picture of a part of video clips in the playing window cannot be watched, and the related information of the target object is omitted is avoided.
In the step, the computer equipment displays the target video abstract on the target electronic map, so that the user can visually see the corresponding position of the corresponding video clip on the target electronic map from the target electronic map, and the visual experience of the user is improved.
According to the video abstract generation method provided by the embodiment of the application, a target object in a first monitoring video is extracted by acquiring the first monitoring video of first monitoring equipment; determining a plurality of second monitoring devices in the target area according to the first position of the first monitoring device, and acquiring a plurality of second monitoring videos of the plurality of second monitoring devices; and acquiring at least one second monitoring video from the plurality of second monitoring videos according to the target object, wherein each second monitoring video comprises the target object. Generating a first video clip corresponding to the target object according to the first monitoring video, and generating at least one second video clip corresponding to the target object according to at least one second monitoring video; and combining the first video segment and the at least one video segment into a target video abstract. According to the method, not only the monitoring video of the first monitoring device is obtained, but also the monitoring videos of the plurality of second monitoring devices in the target area of the first monitoring device are obtained, and at least one second monitoring video including the target object is obtained from the plurality of second monitoring videos, so that the generated target video abstract includes the target object in the first video segment, and each second video segment in the at least one second video segment also includes the target object.
An embodiment of the present application provides a video summary generating apparatus, referring to fig. 7, the apparatus includes:
a first obtaining module 701, configured to obtain a first monitoring video of a first monitoring device, and extract a target object in the first monitoring video;
a second obtaining module 702, configured to determine, according to the first position of the first monitoring device, a plurality of second monitoring devices in the target area, and obtain a plurality of second monitoring videos of the plurality of second monitoring devices;
a third obtaining module 703, configured to obtain at least one second surveillance video from the multiple second surveillance videos according to the target object, where each second surveillance video includes the target object;
a generating module 704, configured to generate a first video segment corresponding to the target object according to the first monitoring video, and generate at least one second video segment corresponding to the target object according to at least one second monitoring video;
a composing module 705 for composing the first video segment and the at least one second video segment into a target video summary of the target object.
In a possible implementation manner, the composing module 705 is further configured to obtain a first time for generating the first video segment and at least one second time for generating at least one second video segment; forming a target video summary of the first video clip and the at least one second video clip according to the time sequence of the first time and the at least one second time; or,
the composition module 705 is further configured to obtain a third time when the target object appears in the first monitoring video, and obtain each fourth time when the target object appears in each second monitoring video; and composing the first video clip and the at least one second video clip into a target video summary according to the time sequence of the third time and each fourth time.
In another possible implementation manner, the third obtaining module 703 is further configured to determine, according to a third time when the target object appears in the first monitored video, a target time period including the third time; and selecting at least one second monitoring video which comprises the target object and has the fourth time of the target object appearing in the second monitoring video within the target time period from the plurality of second monitoring videos.
In another possible implementation manner, the generating module 704 is further configured to select, according to the target object, a video segment that includes the target object in the first surveillance video, and use the video segment as the first video segment; or extracting a plurality of key frames including the target object in the first monitoring video according to the target object, and forming the plurality of key frames into a first video clip.
In another possible implementation manner, the apparatus further includes:
the fourth acquisition module is used for acquiring map data corresponding to the target area according to the identification of the target video abstract when the specified operation of the target video abstract is acquired;
the rendering module is used for rendering the map data corresponding to the target area in the map display interface to obtain a target electronic map;
and the display module is used for displaying the moving track of the target object on the target electronic map and displaying the target video abstract on the target electronic map.
In another possible implementation manner, the display module is further configured to display a first playing window corresponding to the first video clip in the target electronic map according to the first position, and display at least one second playing window corresponding to at least one second video clip in the target electronic map according to at least one second position, where the at least one second position is a position of at least one second monitoring device corresponding to at least one second monitoring video; and displaying the target video abstract on the target electronic map according to the moving track, the first playing window and the at least one second playing window.
In another possible implementation manner, the presentation module is further configured to sort the third time and the at least one fourth time according to the third time and the at least one fourth time to obtain a time sequence, where the third time is a time when the target object appears in the first surveillance video, and the at least one fourth time is a time when the target object appears in the at least one second surveillance video; determining the position of the monitoring equipment corresponding to each two adjacent times from the first position and the at least one second position according to each two adjacent times in the time sequence, wherein the at least one second position is the position of at least one second monitoring equipment corresponding to at least one second monitoring video; connecting the positions of the monitoring equipment corresponding to every two adjacent times according to a target connection mode to obtain a moving track of the target object between the positions of the monitoring equipment corresponding to every two adjacent times until each of the first position and the at least one second position is connected with other positions except the positions to obtain the moving track of the target object; and displaying the moving track of the target object on the target electronic map.
In another possible implementation manner, the apparatus further includes:
the map display interface comprises a calling interface, when the calling interface is triggered, a selection interface is displayed, the selection interface comprises a plurality of video abstract identifiers, and when the selection operation of the target video abstract identifier is acquired, the specified operation of the target video abstract is determined to be acquired; or,
the first determining module is further used for displaying a playlist when the triggering operation of the target video abstract is acquired, wherein the playlist comprises an identifier of the electronic map; when the selection operation of the identification of the electronic map is obtained, determining to obtain the specified operation of the target video abstract; or,
the first determining module is further used for displaying a map display interface, and determining to acquire the specified operation on the target video abstract when the target video abstract is dragged into a display area of the map display interface.
In another possible implementation manner, the presentation module is further configured to sequentially play the corresponding video segments in the first playing window and the at least one second playing window according to a time sequence of a first time and at least one second time, where the first time is a time for generating the first video segment, and the at least one second time is a time for generating the at least one second video segment; when any video segment in the target video abstract is played, displaying a moving track between a position corresponding to the video segment and a position of a next video segment of the video segment on the target electronic map; or, in the process of playing the video clip, displaying a moving track between the position corresponding to the video clip and the position of the next video clip of the video clip on the target electronic map.
In another possible implementation manner, the display module is further configured to display the movement track on the target electronic map; and according to the time sequence of the first time and the at least one second time, playing the corresponding video clips in the first playing window and the at least one second playing window in sequence, wherein the first time is the time for generating the first video clip, and the at least one second time is the time for generating the at least one second video clip.
In another possible implementation, the movement trajectory includes a plurality of sub-trajectories; one sub-track is a track from the position of one monitoring device to the position of another monitoring device; the device still includes:
the second determining module is used for determining a target playing window of the currently played video clip;
the third determining module is used for determining the first target monitoring equipment corresponding to the target playing window;
the selection module is used for selecting a target sub-track from the moving track to a second target monitoring device from a first target monitoring device, wherein the second target monitoring device is a monitoring device corresponding to a next video clip of the currently played video clip;
and the highlighting module is used for highlighting the target sub-track on the target electronic map.
In another possible implementation manner, the apparatus further includes:
the marking module is used for determining each frame of image comprising the target object and marking the target object in each frame of image; and/or determining each frame of image including the target object, determining the place where the target object appears in each frame of image, acquiring the position information of the place, and marking the position information in each frame of image.
In another possible implementation manner, the apparatus further includes:
the fourth determining module is used for determining the scaling of the target electronic map when the target electronic map is zoomed;
the zooming module is used for zooming the target electronic map, the first playing window and the at least one second playing window simultaneously according to the zooming scale of the target electronic map; or zooming the target electronic map according to the zooming scale of the target electronic map, and keeping the first playing window and the at least one second playing window unchanged.
The video abstract generating device provided by the embodiment of the application extracts a target object in a first monitoring video by acquiring the first monitoring video of first monitoring equipment; determining a plurality of second monitoring devices in the target area according to the first position of the first monitoring device, and acquiring a plurality of second monitoring videos of the plurality of second monitoring devices; and acquiring at least one second monitoring video from the plurality of second monitoring videos according to the target object, wherein each second monitoring video comprises the target object. Generating a first video clip corresponding to the target object according to the first monitoring video, and generating at least one second video clip corresponding to the target object according to at least one second monitoring video; and composing the first video segment and the at least one video segment into a target video abstract. The device not only acquires the monitoring video of the first monitoring equipment, but also acquires the monitoring videos of a plurality of second monitoring equipment in the target area of the first monitoring equipment, and acquires at least one second monitoring video including the target object from the plurality of second monitoring videos, so that the generated target video abstract not only includes the target object in the first video segment, but also includes the target object in each second video segment in the at least one second video segment.
Fig. 8 is a block diagram of a computer device 800 according to an embodiment of the present disclosure. For example, the computer device 800 may be configured to perform the video summary generation methods provided in the various embodiments described above. Referring to fig. 8, the computer apparatus 800 includes: a processor 801 and a memory 802.
Processor 801 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and so forth. The processor 801 may be implemented in at least one hardware form of DSP (Digital Signal Processing), FPGA (Field-Programmable Gate Array), PLA (Programmable Logic Array). The processor 801 may also include a main processor and a coprocessor, where the main processor is a processor for Processing data in an awake state, and is also called a Central Processing Unit (CPU); a coprocessor is a low power processor for processing data in a standby state. In some embodiments, the processor 801 may be integrated with a GPU (Graphics Processing Unit) that is responsible for rendering and drawing content that the display screen needs to display. In some embodiments, the processor 801 may further include an AI (Artificial Intelligence) processor for processing computing operations related to machine learning.
Memory 802 may include one or more computer-readable storage media, which may be non-transitory. Memory 802 may also include high speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in memory 802 is used to store at least one instruction for execution by processor 801 to implement the video summary generation methods provided by method embodiments herein.
In some embodiments, computer device 800 may also optionally include: a peripheral interface 803 and at least one peripheral. The processor 801, memory 802, and peripheral interface 803 may be connected by buses or signal lines. Various peripheral devices may be connected to peripheral interface 803 by a bus, signal line, or circuit board. Specifically, the peripheral device includes: at least one of a radio frequency circuit 804, a display 805, a camera 806, an audio circuit 807, a positioning component 808, and a power source 809.
The peripheral interface 803 may be used to connect at least one peripheral related to I/O (Input/Output) to the processor 801 and the memory 802. In some embodiments, the processor 801, memory 802, and peripheral interface 803 are integrated on the same chip or circuit board; in some other embodiments, any one or two of the processor 801, the memory 802, and the peripheral interface 803 may be implemented on separate chips or circuit boards, which is not limited by the present embodiment.
The Radio Frequency circuit 804 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuitry 804 communicates with communication networks and other communication devices via electromagnetic signals. The radio frequency circuit 804 converts an electrical signal into an electromagnetic signal to transmit, or converts a received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 804 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and so forth. The radio frequency circuit 804 may communicate with other computer devices via at least one wireless communication protocol. The wireless communication protocols include, but are not limited to: the world wide web, metropolitan area networks, intranets, various generations of mobile communication networks (2G, 3G, 4G, and 5G), wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the radio frequency circuit 804 may further include NFC (Near Field Communication) related circuits, which are not limited in this application.
The display screen 805 is used to display a UI (User Interface). The UI may include graphics, text, icons, video, and any combination thereof. When the display 805 is a touch display, the display 805 also has the ability to capture touch signals on or above the surface of the display 805. The touch signal may be input to the processor 801 as a control signal for processing. At this point, the display 805 may also be used to provide virtual buttons and/or a virtual keyboard, also referred to as soft buttons and/or a soft keyboard. In some embodiments, the display 805 may be one, providing the front panel of the computer device 800; in other embodiments, the display 805 may be at least two, each disposed on a different surface of the computer device 800 or in a folded design; in still other embodiments, the display 805 may be a flexible display, disposed on a curved surface or on a folded surface of the computer device 800. Even further, the display 805 may be arranged in a non-rectangular irregular pattern, i.e., a shaped screen. The Display 805 can be made of LCD (Liquid Crystal Display), OLED (Organic Light-Emitting Diode), and other materials.
The camera assembly 806 is used to capture images or video. Optionally, camera assembly 806 includes a front camera and a rear camera. Generally, a front camera is disposed on a front panel of a computer apparatus, and a rear camera is disposed on a rear surface of the computer apparatus. In some embodiments, the number of the rear cameras is at least two, and each rear camera is any one of a main camera, a depth-of-field camera, a wide-angle camera and a telephoto camera, so that the main camera and the depth-of-field camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize panoramic shooting and VR (Virtual Reality) shooting functions or other fusion shooting functions. In some embodiments, camera assembly 806 may also include a flash. The flash lamp can be a monochrome temperature flash lamp or a bicolor temperature flash lamp. The double-color-temperature flash lamp is a combination of a warm-light flash lamp and a cold-light flash lamp, and can be used for light compensation at different color temperatures.
The audio circuitry 807 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 801 for processing or inputting the electric signals to the radio frequency circuit 804 to realize voice communication. For stereo capture or noise reduction purposes, the microphones may be multiple and located at different locations on the computer device 800. The microphone may also be an array microphone or an omni-directional pick-up microphone. The speaker is used to convert electrical signals from the processor 801 or the radio frequency circuit 804 into sound waves. The loudspeaker can be a traditional film loudspeaker and can also be a piezoelectric ceramic loudspeaker. When the speaker is a piezoelectric ceramic speaker, the speaker can be used for purposes such as converting an electric signal into a sound wave audible to a human being, or converting an electric signal into a sound wave inaudible to a human being to measure a distance. In some embodiments, the audio circuitry 807 may also include a headphone jack.
The Location component 808 is used to locate the current geographic Location of the computer device 800 to implement navigation or LBS (Location Based Service). The positioning component 808 may be a positioning component based on GPS in the united states, beidou in china, or galileo in the european union.
A power supply 809 is used to power the various components in the computer device 800. The power supply 809 can be ac, dc, disposable or rechargeable. When the power supply 809 includes a rechargeable battery, the rechargeable battery may be a wired rechargeable battery or a wireless rechargeable battery. The wired rechargeable battery is a battery charged through a wired line, and the wireless rechargeable battery is a battery charged through a wireless coil. The rechargeable battery can also be used to support fast charge technology.
In some embodiments, the computer device 800 also includes one or more sensors 810. The one or more sensors 810 include, but are not limited to: acceleration sensor 811, gyro sensor 812, pressure sensor 813, fingerprint sensor 814, optical sensor 815 and proximity sensor 816.
The acceleration sensor 811 may detect the magnitude of acceleration in three coordinate axes of a coordinate system established with the computer apparatus 800. For example, the acceleration sensor 811 may be used to detect components of the gravitational acceleration in three coordinate axes. The processor 801 may control the touch screen 805 to display the user interface in a landscape view or a portrait view according to the gravitational acceleration signal collected by the acceleration sensor 811. The acceleration sensor 811 may also be used for acquisition of motion data of a game or a user.
The gyro sensor 812 may detect a body direction and a rotation angle of the computer device 800, and the gyro sensor 812 may cooperate with the acceleration sensor 811 to acquire a 3D motion of the user with respect to the computer device 800. From the data collected by the gyro sensor 812, the processor 801 may implement the following functions: motion sensing (such as changing the UI according to a user's tilting operation), image stabilization at the time of photographing, game control, and inertial navigation.
Pressure sensors 813 may be disposed on the side bezel of computer device 800 and/or underneath touch display 805. When the pressure sensor 813 is arranged on the side frame of the computer device 800, the holding signal of the user to the computer device 800 can be detected, and the processor 801 performs left-right hand recognition or shortcut operation according to the holding signal collected by the pressure sensor 813. When the pressure sensor 813 is disposed at a lower layer of the touch display screen 805, the processor 801 controls the operability control on the UI interface according to the pressure operation of the user on the touch display screen 805. The operability control comprises at least one of a button control, a scroll bar control, an icon control and a menu control.
The fingerprint sensor 814 is used for collecting a fingerprint of the user, and the processor 801 identifies the identity of the user according to the fingerprint collected by the fingerprint sensor 814, or the fingerprint sensor 814 identifies the identity of the user according to the collected fingerprint. Upon identifying the user as a trusted identity, the processor 801 authorizes the user to perform relevant sensitive operations, including unlocking the screen, viewing encrypted information, downloading software, paying for and changing settings, etc. Fingerprint sensor 814 may be disposed on the front, back, or side of computer device 800. When a physical key or vendor Logo is provided on the computer device 800, the fingerprint sensor 814 may be integrated with the physical key or vendor Logo.
The optical sensor 815 is used to collect the ambient light intensity. In one embodiment, the processor 801 may control the display brightness of the touch screen 805 based on the ambient light intensity collected by the optical sensor 815. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 805 is increased; when the ambient light intensity is low, the display brightness of the touch display 805 is turned down. In another embodiment, the processor 801 may also dynamically adjust the shooting parameters of the camera assembly 806 according to the ambient light intensity collected by the optical sensor 815.
A proximity sensor 816, also known as a distance sensor, is typically provided on the front panel of the computer device 800. The proximity sensor 816 is used to capture the distance between the user and the front of the computer device 800. In one embodiment, the processor 801 controls the touch display 805 to switch from a bright screen state to a dark screen state when the proximity sensor 816 detects that the distance between the user and the front face of the computer device 800 is gradually reduced; when the proximity sensor 816 detects that the distance between the user and the front of the computer device 800 is gradually increasing, the touch display 805 is controlled by the processor 801 to switch from a breath-screen state to a bright-screen state.
Those skilled in the art will appreciate that the architecture illustrated in FIG. 8 is not intended to be limiting of the computer device 800, and may include more or fewer components than those illustrated, or some components may be combined, or a different arrangement of components may be used.
The embodiment of the present application further provides a computer-readable storage medium, which is applied to a terminal, and the computer-readable storage medium stores at least one instruction, at least one program, a code set, or a set of instructions, where the instruction, the program, the code set, or the set of instructions are loaded and executed by a processor to implement the operations performed by a computer device in the video summary generation method according to the foregoing embodiment.
It will be understood by those skilled in the art that all or part of the steps for implementing the above embodiments may be implemented by hardware, or may be implemented by a program instructing relevant hardware, where the program may be stored in a computer-readable storage medium, and the above-mentioned storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
The above description is only for facilitating the technical solution of the present application to be understood by those skilled in the art, and is not intended to limit the present application. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the present application shall be included in the protection scope of the present application.

Claims (12)

1. A method for displaying a video abstract, the method comprising:
acquiring a first monitoring video of first monitoring equipment, and extracting a target object in the first monitoring video;
determining a plurality of second monitoring devices in a target area according to the first position of the first monitoring device, and acquiring a plurality of second monitoring videos of the plurality of second monitoring devices;
acquiring at least one second monitoring video from the plurality of second monitoring videos according to the target object, wherein each second monitoring video comprises the target object;
generating a first video clip corresponding to the target object according to the first monitoring video, and generating at least one second video clip corresponding to the target object according to the at least one second monitoring video;
composing the first video clip and the at least one second video clip into a target video summary of the target object;
displaying a map display interface, wherein the map display interface comprises a calling interface;
when the calling interface is triggered, displaying a selection interface, wherein the selection interface comprises a plurality of video abstract identifiers;
when the selection operation of the identification of the target video abstract is obtained, obtaining map data corresponding to the target area;
rendering the map data corresponding to the target area in the map display interface to obtain a target electronic map;
displaying a first playing window corresponding to the first video clip in the target electronic map according to the first position, and displaying at least one second playing window corresponding to the at least one second video clip in the target electronic map according to at least one second position, wherein the at least one second position is a position of at least one second monitoring device corresponding to the at least one second monitoring video;
sequentially playing the corresponding video clips in the first playing window and the at least one second playing window according to the time sequence of first time and at least one second time, wherein the first time is the time for generating the first video clip, and the at least one second time is the time for generating the at least one second video clip;
in the process of playing any video segment in the target video abstract, displaying a moving track between a position corresponding to the video segment and a position of a next video segment corresponding to the video segment on the target electronic map;
in the process of playing any video segment in the target video abstract, when the icons corresponding to the positions of the first monitoring device and the at least one second monitoring device on the target electronic map are triggered, playing the corresponding video segment in the playing window corresponding to the triggered icon.
2. The method of claim 1, wherein the composing the first video segment and the at least one second video segment into a target video summary of the target object comprises:
acquiring the first time and the at least one second time; composing the first video clip and the at least one second video clip into the target video summary in chronological order of the first time and the at least one second time; or,
acquiring third time when the target object appears in the first monitoring video and acquiring each fourth time when the target object appears in each second monitoring video; and according to the time sequence of the third time and each fourth time, combining the first video clip and the at least one second video clip into the target video summary.
3. The method according to claim 1, wherein the obtaining at least one second surveillance video from the plurality of second surveillance videos according to the target object comprises:
determining a target time period comprising a third time according to the third time when the target object appears in the first monitoring video;
and selecting at least one second monitoring video which comprises the target object and has a fourth time of appearance in the second monitoring video within the target time period from the plurality of second monitoring videos.
4. The method according to claim 1, wherein the generating a first video segment corresponding to the target object according to the first surveillance video comprises:
selecting a video clip including the target object in the first monitoring video according to the target object, and taking the video clip as the first video clip; or,
according to the target object, extracting a plurality of key frames including the target object in the first monitoring video, and forming the first video clip by the plurality of key frames.
5. The method of claim 1, further comprising:
and displaying the movement track of the target object on the target electronic map.
6. The method according to claim 5, wherein the displaying the moving track of the target object on the target electronic map comprises:
sequencing the third time and at least one fourth time according to the third time and the at least one fourth time to obtain a time sequence, wherein the third time is the time when the target object appears in the first monitoring video, and the at least one fourth time is the time when the target object appears in the at least one second monitoring video;
determining the position of the monitoring equipment corresponding to each two adjacent times from the first position and at least one second position according to each two adjacent times in the time sequence, wherein the at least one second position is the position of at least one second monitoring equipment corresponding to the at least one second monitoring video;
connecting the positions of the monitoring equipment corresponding to every two adjacent times according to a target connection mode to obtain a moving track of the target object between the positions of the monitoring equipment corresponding to every two adjacent times until each of the first position and the at least one second position is connected with other positions except the positions to obtain a moving track of the target object;
and displaying the movement track of the target object on the target electronic map.
7. The method of claim 5, wherein the movement trajectory comprises a plurality of sub-trajectories; one sub-track is a track from the position of one monitoring device to the position of another monitoring device; the method further comprises the following steps:
determining a target playing window of a currently played video clip;
determining first target monitoring equipment corresponding to the target playing window;
selecting a target sub-track from the moving track to a second target monitoring device, wherein the second target monitoring device is a monitoring device corresponding to a next video clip of the currently played video clip;
and highlighting the target sub-track on the target electronic map.
8. The method according to any one of claims 1-7, further comprising:
determining each frame of image comprising the target object, and marking the target object in each frame of image; and/or the presence of a gas in the atmosphere,
determining each frame of image including the target object, determining a place where the target object appears in each frame of image, acquiring position information of the place, and labeling the position information in each frame of image.
9. The method of claim 1, further comprising:
when the target electronic map is zoomed, determining the zoom scale of the target electronic map;
simultaneously zooming the target electronic map, the first playing window and the at least one second playing window according to the zooming scale of the target electronic map; or zooming the target electronic map according to the zooming scale of the target electronic map, and keeping the first playing window and the at least one second playing window unchanged.
10. A video summary presentation apparatus, the apparatus comprising:
the first acquisition module is used for acquiring a first monitoring video of first monitoring equipment and extracting a target object in the first monitoring video;
the second acquisition module is used for determining a plurality of second monitoring devices in a target area according to the first position of the first monitoring device and acquiring a plurality of second monitoring videos of the plurality of second monitoring devices;
a third obtaining module, configured to obtain at least one second surveillance video from the plurality of second surveillance videos according to the target object, where each second surveillance video includes the target object;
a generating module, configured to generate a first video segment corresponding to the target object according to the first monitoring video, and generate at least one second video segment corresponding to the target object according to the at least one second monitoring video;
a composing module for composing the first video segment and the at least one second video segment into a target video summary of the target object;
the map display system comprises a first determination module, a second determination module and a display module, wherein the first determination module is used for displaying a map display interface which comprises a calling interface; when the calling interface is triggered, displaying a selection interface, wherein the selection interface comprises a plurality of video abstract identifiers; when the selection operation of the identification of the target video abstract is obtained, obtaining map data corresponding to the target area;
the rendering module is used for rendering the map data corresponding to the target area in the map display interface to obtain a target electronic map;
the display module is used for displaying a first playing window corresponding to the first video clip in the target electronic map according to the first position, and displaying at least one second playing window corresponding to the at least one second video clip in the target electronic map according to at least one second position, wherein the at least one second position is a position of at least one second monitoring device corresponding to the at least one second monitoring video; sequentially playing the corresponding video clips in the first playing window and the at least one second playing window according to the time sequence of a first time and at least one second time, wherein the first time is the time for generating the first video clip, and the at least one second time is the time for generating the at least one second video clip; in the process of playing any video segment in the target video abstract, displaying a moving track between a position corresponding to the video segment and a position of a next video segment corresponding to the video segment on the target electronic map;
displaying the target video abstract on the target electronic map; in the process of playing any video segment in the target video abstract, when the icons corresponding to the positions of the first monitoring device and the at least one second monitoring device on the target electronic map are triggered, playing the corresponding video segment in the playing window corresponding to the triggered icon.
11. A computer device, characterized in that the computer device comprises:
a processor and a memory, the memory having stored therein at least one instruction that is loaded and executed by the processor to implement the operations performed in the video summary presentation method of any of claims 1-9.
12. A computer-readable storage medium, having at least one instruction stored therein, which is loaded and executed by a processor to perform the operations of any one of claims 1 to 9.
CN201910551305.4A 2019-06-24 2019-06-24 Video abstract generation method and device, computer equipment and storage medium Active CN111405382B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910551305.4A CN111405382B (en) 2019-06-24 2019-06-24 Video abstract generation method and device, computer equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910551305.4A CN111405382B (en) 2019-06-24 2019-06-24 Video abstract generation method and device, computer equipment and storage medium

Publications (2)

Publication Number Publication Date
CN111405382A CN111405382A (en) 2020-07-10
CN111405382B true CN111405382B (en) 2022-12-02

Family

ID=71413080

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910551305.4A Active CN111405382B (en) 2019-06-24 2019-06-24 Video abstract generation method and device, computer equipment and storage medium

Country Status (1)

Country Link
CN (1) CN111405382B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111914118B (en) * 2020-07-22 2021-08-27 珠海大横琴科技发展有限公司 Video analysis method, device and equipment based on big data and storage medium
CN115455275B (en) * 2022-11-08 2023-02-03 广东卓维网络有限公司 Video processing system integrated with inspection equipment

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103794234B (en) * 2012-10-30 2016-09-21 北京航天长峰科技工业集团有限公司 Event trace quick-searching platform based on massive video
CN104581003A (en) * 2013-10-12 2015-04-29 北京航天长峰科技工业集团有限公司 Video rechecking positioning method
CN106878666B (en) * 2015-12-10 2020-04-28 杭州海康威视数字技术股份有限公司 Method, device and system for searching target object based on monitoring camera
CN106096577B (en) * 2016-06-24 2019-05-31 安徽工业大学 A kind of target tracking method in camera distribution map
CN106446002A (en) * 2016-08-01 2017-02-22 三峡大学 Moving target-based video retrieval method for track in map
CN107358622A (en) * 2017-06-19 2017-11-17 三峡大学 A kind of video information processing method and system based on visualization movement locus
CN107888877B (en) * 2017-11-13 2020-10-09 深圳市戴升智能科技有限公司 Method and system for vehicle tracking and road traffic information acquisition

Also Published As

Publication number Publication date
CN111405382A (en) 2020-07-10

Similar Documents

Publication Publication Date Title
CN109167894B (en) Camera control method and device, mobile terminal and storage medium
CN110087123B (en) Video file production method, device, equipment and readable storage medium
CN111093026B (en) Video processing method, electronic device and computer-readable storage medium
CN111147878B (en) Stream pushing method and device in live broadcast and computer storage medium
CN110276789B (en) Target tracking method and device
CN106165430A (en) Net cast method and device
CN107786827B (en) Video shooting method, video playing method and device and mobile terminal
CN111065001B (en) Video production method, device, equipment and storage medium
CN108848313B (en) Multi-person photographing method, terminal and storage medium
CN109922356B (en) Video recommendation method and device and computer-readable storage medium
CN111125442B (en) Data labeling method and device
CN110267054B (en) Method and device for recommending live broadcast room
CN110225390B (en) Video preview method, device, terminal and computer readable storage medium
CN113613028B (en) Live broadcast data processing method, device, terminal, server and storage medium
CN109618192B (en) Method, device, system and storage medium for playing video
CN113936699B (en) Audio processing method, device, equipment and storage medium
CN111405382B (en) Video abstract generation method and device, computer equipment and storage medium
CN110991260B (en) Scene marking method, device, equipment and storage medium
CN111491124B (en) Video processing method and device and electronic equipment
CN110470293B (en) Navigation method and mobile terminal
CN111383251B (en) Method, device, monitoring equipment and storage medium for tracking target object
CN111723124B (en) Data collision analysis method and device, electronic equipment and storage medium
CN114554112B (en) Video recording method, device, terminal and storage medium
CN111176338A (en) Navigation method, electronic device and storage medium
CN112004134A (en) Multimedia data display method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant